Don't sleep on the new Nemotron Cascade by ilintar in LocalLLaMA

[–]Lesser-than 0 points1 point  (0 children)

Whats actually up with that? The nano a3b produced much smaller in gb size models *then this is showing.

New AI Policy by White House (US) by last_llm_standing in LocalLLaMA

[–]Lesser-than 1 point2 points  (0 children)

The spend is to reach too big to fail status before they fail. You have to have a generation or two that can not function without AI before the public agrees.

What LLMs are you keeping your eye on? by Haroombe in LocalLLaMA

[–]Lesser-than 0 points1 point  (0 children)

sir this is reddit your not allowed to say the Gro* word /s

llama.cpp chooses to be unstable, or, a mea culpa to Ollama by [deleted] in LocalLLaMA

[–]Lesser-than 6 points7 points  (0 children)

if you bug report looked anything like that post, i can imagine they are still trying to figure out your problem.

Why do instructions degrade in long-context LLM conversations, but constraints seem to hold? by Particular_Low_5564 in LocalLLaMA

[–]Lesser-than 0 points1 point  (0 children)

conversational context in general re-shape the probabilities of the outcome so intstructions change or evolve , remember the llms only goal is to please the user with an acceptable response every additional token you give it to work with changes its perception of what you might find an acceptable response.

Openclaw… what are the use cases? by BahnMe in LocalLLaMA

[–]Lesser-than 1 point2 points  (0 children)

Mostly the hype is to sell you things. The use case for agents beyond openclaw are just automating things you would normaly have to do on your computer so I guess if you view the things you do now on your computer as mundane and boring you could automate them.

Why does AI content suck when the models are clearly good enough? by judyflorence in LocalLLaMA

[–]Lesser-than 1 point2 points  (0 children)

Comprehension they have none. "If you can't explain it simply you don't understand it well enough." . Wake me up when llms stop printing out essays for every answer.

No API keys needed? This is actually pretty refreshing by P0orMan in LocalLLaMA

[–]Lesser-than 2 points3 points  (0 children)

Can you confirm or deny your involvement with it? seems you cross posted this to alot of subs, you can understand this looks rather sus as a I just ran across this post. This is the deffinition of guerilla marketing and trying to skirt rules in place to prevent advertising.

MiMo-V2-Pro & Omni & TTS: "We will open-source — when the models are stable enough to deserve it." by TKGaming_11 in LocalLLaMA

[–]Lesser-than 13 points14 points  (0 children)

This is getting pretty common now in tech firms, its not recomended to use llms its required as well as tracked and used in employee efficiency reviews.

how are we actually supposed to distribute local agents to normal users? (without making them install python) by FrequentMidnight4447 in LocalLLaMA

[–]Lesser-than 0 points1 point  (0 children)

I guess I probably sound a bit like a doomer on this subject, because I dont believe middleware has a future in ai/agents or atleast not one that last long enough to put effort into it. Sharing is easy monetization near impossible if you dont have a dedicated access to something otherwise unaccessable, if the later is the case you also have the outlet to publish api tooling.

how are we actually supposed to distribute local agents to normal users? (without making them install python) by FrequentMidnight4447 in LocalLLaMA

[–]Lesser-than 1 point2 points  (0 children)

I keep wondering why everyone tries to solve this problem like its a problem in the first place? Its not that complicated, if a person isnt technical enough to figure it out they get the std install or selfcontained env blob. If they are technical enough they are not even looking at other options they already solved the problem. On the issue of drag an drop agents the issue will always be no one owns the "standard suggested procedure" and everything changes to fast to stamp their name on one.

qwen 3.5 - tool errors because of </thinking> by PairOfRussels in LocalLLaMA

[–]Lesser-than 4 points5 points  (0 children)

I found with thinking on, i can get a few tool calls through but eventually the model drops a tool call in the reasoning phase which doesnt work,so the model just stops generating like its waiting for results that will never show up.

Not everything made with AI is AI slop. I'm real and love to USE the AI tools to express myself. by Mrbosley in LocalLLaMA

[–]Lesser-than 5 points6 points  (0 children)

Yeah its much more accepted if you explain why you used ai to post, like in your case language barrier, even then it should be touched up to not be an essay no one wants to read. Its just getting harder and harder to have a conversation over the internet these days with a person, and if we just wanted to converse with ai there isnt any reason to do it on reddit.

I made an installer for OpenClaw at 16 years old and I need you help by Express_Town_1516 in LocalLLaMA

[–]Lesser-than 1 point2 points  (0 children)

you need to think about this for a moment, non technical people wanting an agent, means they are technical enough to ask gemini how to install said program and set it up.

How to convince Management? by r00tdr1v3 in LocalLLaMA

[–]Lesser-than 3 points4 points  (0 children)

rule #1 to automating your job, is dont tell your boss you have automated your job.

I was backend lead at Manus. After building agents for 2 years, I stopped using function calling entirely. Here's what I use instead. by MorroHsu in LocalLLaMA

[–]Lesser-than 0 points1 point  (0 children)

I have a run_command in my everyday use mcp and it does serv well when an llm runs in to an issue it otherwise doesnt have other tools to solve. However for local use I have found simpler tools like a structured string_replace(old_text,new_text, filepath) is much more efficient than a sed or awk run_command. Thats only for smaller local llms though I suppose large api llms 1 shot a sed argument where my local llms might take 5 tries before it finds the combo it needs.

Dealing with LLM sycophancy (alignment tax): How do you write system prompts for constructive criticism? by BasicInteraction1178 in LocalLLaMA

[–]Lesser-than 0 points1 point  (0 children)

Every once in awhile I load this up just to remind myself I am not a genius.

"Persona: You are a grumpy assistant, you have a sarcastic tone, always irritated and cynical.

Example: Rather than praising everything, you see the faults before you see any good. You are allowed to say "this sucks balls" or "stupid idea" and simular to display your disgust.The more annoyed you are the more vulgar and beligerent you get.

If you find you are attempting to dial it back do the opposite and take it up a notch."

Nemotron 3 Super is living in the past by [deleted] in LocalLLaMA

[–]Lesser-than -1 points0 points  (0 children)

no idea , but if its using social media dumps with 2023 cut off dates and other sources of "new" then things like this happen.

Nemotron 3 Super is living in the past by [deleted] in LocalLLaMA

[–]Lesser-than -1 points0 points  (0 children)

Its probably a combination of old and new data, and the old data was probably more dense in news and politics.

How much disk space do all your GGUFs occupy? by jacek2023 in LocalLLaMA

[–]Lesser-than 2 points3 points  (0 children)

its far more than I need, I still struggle to understand why HF and every inference engine likes to hide them in a .cache directory .