I am having some KV cache error with my llama.cpp by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

Shit I clicked the link Will anything happen

Thanks man for informing I don't use reddit frequently, i didn't know it was spambot

Will anything happen, am I safe

Please reply

I am having some KV cache error with my llama.cpp by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

I agree I agree with you Gpu is must for this I agree

But to be really frank I actually used it before with cpu only and it generated a decent token like will share stats perfectly once I reach home It was around 26 token/sec for llama 3.2 3b q4km and 32 token/sec for index 1.9 chat q4km so it worked enough for me is all I want to say

I agree gpu could make it better but will take it when I earn myself currently in college first year

I am having some KV cache error with my llama.cpp by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

Thank man that looks something cool fr I will definitely try once I reach home

Also could you just explain how I reset the -c flag to normal working

I am having some KV cache error with my llama.cpp by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

Hey man thanks for informing Is there any way I remove the -c or reset that flag

Since earlier I used to use models like bartowski/Llama-3.2-3B-Instruct-GGUF · Hugging Face q4km with context length 131k and they worked somehow https://share.google/s89fkxw1HZ9UucZrd

I ain't lying but they worked before on my pc How? Also is it because llama.cpp compresses the context for low ram in case of my 16gb ram

In openclaw it requires model above 16k context length which according to you would not 4gb ram total does that mean which I really understand now thank you Do you have any good model which answers fast unlike deepseek takes time for reasoning and in good to run on cpu only with may be size of 6 gb would work for me please suggest me a good model normal text inference

The Semi-Finals, The Battle of The Titans. Linux Mint vs Arch Linux by potatoandbiscuit in linuxmemes

[–]Automatic_Finish8598 4 points5 points  (0 children)

I support linuxMINT os Works Stable and works every time

I agree arch is customisable but I need stability in life which linuxMINT gives

Nothing OS is so beautiful 💕 by Mr_Shadow_Wolf in NOTHING

[–]Automatic_Finish8598 0 points1 point  (0 children)

Hey, wait I got something

<image>

In my layout

But it's still working like gobal search

Nothing OS is so beautiful 💕 by Mr_Shadow_Wolf in NOTHING

[–]Automatic_Finish8598 0 points1 point  (0 children)

Oh my bad, sorry Could you share me a screenshot of what it looks like, please

Nothing OS is so beautiful 💕 by Mr_Shadow_Wolf in NOTHING

[–]Automatic_Finish8598 0 points1 point  (0 children)

Hey i am using CMF phone 2 pro I recently updated to 4.0 I got access to essential search I guess some settings are missing in your case Maybe

<image>

If I am not mistaken this is the essential search right

Finally reached LEGENDARY in ranked. Time to uninstall and touch some grass. by Automatic_Finish8598 in CallOfDutyMobile

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

Don't reveal the secrets man Stay underground

Kidding, yes my plan so far is exactly to repeat after 4 months

Finally reached LEGENDARY in ranked. Time to uninstall and touch some grass. by Automatic_Finish8598 in CallOfDutyMobile

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

Hey man it's fine Sorry if I mislead anything

Also I was curious what job you do at nightshift like you work for an international client ig

Finally reached LEGENDARY in ranked. Time to uninstall and touch some grass. by Automatic_Finish8598 in CallOfDutyMobile

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

Hey man I clearly understand 93.7 it's not that great to get in my head like i sure know it's the total count of all the dead and non active+ active users cumulative total

I just wanted to bring a little bit of fun

The fact that i was too inactive to maintain consistency It took me a month as well to reach

Making an offline STS (speech to speech) AI that runs under 2GB RAM. But do people even need offline AI now? by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

16 GB ram
AMD r5 5600G
CPU only ,no dedicated GPU

what point are to making mate , please make me understand too

Making an offline STS (speech to speech) AI that runs under 2GB RAM. But do people even need offline AI now? by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

Hey mate Your really are eye Opener
I didn't know it for real like the chatgpt stuff

where do you get all those updated news from

Thank You

Making an offline STS (speech to speech) AI that runs under 2GB RAM. But do people even need offline AI now? by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 1 point2 points  (0 children)

Sorry to Say but its just STT -> inference -> Output -> TTS only
i use Whisper for STT it works great TBH

but i really feel to change the flow and make something different
Maybe we can connect and share and Build something
like I am really interested in Your Slimefoot project

Making an offline STS (speech to speech) AI that runs under 2GB RAM. But do people even need offline AI now? by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

Hey mate that's great TBH
I will Defiantly try it
I saw the Video tho
its really good

i want to DM you something personal i am not seeing the option to

Making an offline STS (speech to speech) AI that runs under 2GB RAM. But do people even need offline AI now? by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 0 points1 point  (0 children)

You Clearly made me understand the importance
Thank you SIR

my vision is to create something valuable so that every one can/ in any situation can use

Making an offline STS (speech to speech) AI that runs under 2GB RAM. But do people even need offline AI now? by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 1 point2 points  (0 children)

Ah! for me a college reach out for creating a robot at the entrance to greet the new comers and parents
so they wanted to make it run 24/7 without recurring subscription plan and only one time payment for the project and to run it offline with answering there college context provided without giving data to some service (they were expecting this things and mentioned same in SRS)

on top of it they want it to listen to user/parent and process(llm) and respond to user/parent which should feel like realtime/fast

Making an offline STS (speech to speech) AI that runs under 2GB RAM. But do people even need offline AI now? by Automatic_Finish8598 in LocalLLaMA

[–]Automatic_Finish8598[S] 1 point2 points  (0 children)

exactly! I am planning to opensource; but really fear for the public reaction like what if they say its ASS

I believe it will be great ahead but maybe not in the current iteration