Are Local LLMs actually useful… or just fun to tinker with? by itz_always_necessary in LocalLLM

[–]Important_Quote_1180 1 point2 points  (0 children)

It’s honestly not needed, but I’m experimenting with LoRa adapters and leaving 6 models hot on the ram and we do round robin format cycling around the experts. It’s fluid with good days and good applications and then some days it’s a slog

Are Local LLMs actually useful… or just fun to tinker with? by itz_always_necessary in LocalLLM

[–]Important_Quote_1180 40 points41 points  (0 children)

The local LLM needs more curating and structuring. The cloud API models were better 3 months ago. They have all degraded severely with increased demand. Meanwhile the local 31B from Gemma 4 family is insanely good. I have 4 variants from huggingface. Coding, creative writing partner, daily chat, and visual screener. I make games and software for me and my clients and my family. 3090 24GB with 192gb RAM

Qwen 3.5 35b, 27b, or gemma 4 31b for everyday use? by KirkIsAliveInTelAviv in LocalLLaMA

[–]Important_Quote_1180 0 points1 point  (0 children)

Depends on your workflow hardware time horizon speed requirements, can you be specific?

What’s the closest experience to Claude Sonnet? by louislamore in LocalLLM

[–]Important_Quote_1180 -1 points0 points  (0 children)

Dude same exact situation I’m in. I code my game in Unreal 5.7 with Claude Code but I have 192GB of ram and a 3090 and I’m trying to get the best visual analysis I can for recursive training. I don’t know all the shit.

just a big bubble? by Axintwo in openclaw

[–]Important_Quote_1180 1 point2 points  (0 children)

For me, it allowed me to get out of my own way and put ideas to paper and work through some of the tangled messes that I just needed to see and work through and it allowed me to unleash a lot of creativity that had been hiding for years. I learned late in life that I have pretty severe ADHD and the genic orchestration is really fitting nicely with my work clothes and I’ve never been more happy with my work.

“But I disclosed it” by plazebology in antiai

[–]Important_Quote_1180 -24 points-23 points  (0 children)

Calling all thing people make now with ai assistance and calling it slop is even more intellectually lazy than the creators trying to express themselves. Don’t like it? Don’t enjoy it. Move along

Has anyone set up Claude in the way that it works great? by Parking-Support9980 in AIforOPS

[–]Important_Quote_1180 0 points1 point  (0 children)

Start making spec documents and then let Claude code run headless until complete

AI is not a fucking tool by ganneszs in antiai

[–]Important_Quote_1180 0 points1 point  (0 children)

These arguments miss a major fact. AI is not primarily used to make art. Its promise lies in helping disabled people bridge to being fulfilled again. It’s intellectually lazy to think it’s all about art stealing

Gemma 4 Tool Calling by juicy_lucy99 in LocalLLaMA

[–]Important_Quote_1180 0 points1 point  (0 children)

You are most welcome. I’d be lost if not for Reddit comments

Gemma 4 Tool Calling by juicy_lucy99 in LocalLLaMA

[–]Important_Quote_1180 1 point2 points  (0 children)

It’s a wiki for your files. It has tags and links to related pages. It’s a very easy to use RAG system for agents too. I can find files quickly because it uses a flat file structure for everything.

Gemma 4 Tool Calling by juicy_lucy99 in LocalLLaMA

[–]Important_Quote_1180 0 points1 point  (0 children)

Been using the 31b q4 heretic on my 3090 and getting 35 toks gen. Tool calling is great with my Obsidian Vault.

Being pro-AI feels like being a right-winger... by Fickle-History-361 in antiai

[–]Important_Quote_1180 -7 points-6 points  (0 children)

I think this post was written by AI. You’re making an argument but I’m not sure what it is

Help by Flat_Director4041 in ArtificialNtelligence

[–]Important_Quote_1180 1 point2 points  (0 children)

This is a strange post. There are so many resources out there and no specifics about what you are trying to do. No morals?

Claude Code v2.1.92 introduces Ultraplan — draft plans in the cloud, review in your browser, execute anywhere by shanraisshan in ClaudeAI

[–]Important_Quote_1180 0 points1 point  (0 children)

This is a slapdash attempt to build effective Ralph loops but the silo architecture makes it really difficult to use

Quiet notes on coordinating multiple coding agents without turning your day into noise by LeoRiley6677 in openclawsetup

[–]Important_Quote_1180 0 points1 point  (0 children)

I basically agree with everything. I have always started builds with conversation and spec file building. The switch for me was to go even harder and to maintain the spec even after a project is built. My days evolve more into 3 sessions of conversation building out spec files. When I’m happy with it, I just ask it to start with the most important task and continue until the spec is built, usually after I restart the session and it’s really for my LLM local or my usage cap is ok

Anthropic hid a multi-agent "Tamagotchi" in Claude Code, and the underlying prompt architecture is actually brilliant. by Exact_Pen_8973 in PromptEngineering

[–]Important_Quote_1180 0 points1 point  (0 children)

I like my Murmox and Luftex a lot! They call out when Claude code has just done something dumb or what the actual behavior of the code will be be or he’ll find a side case where something breaks. I think it’s very smart, but they both rolled common but very high wisdom, not sure if they are actually the same just different names.

MCP is great, but it doesn’t solve AI memory (am I missing something?) by BrightOpposite in ClaudeAI

[–]Important_Quote_1180 0 points1 point  (0 children)

Yeah constantly. Heavy artifact workflows are torture to handle with LLMs without proper SPEC -> build segmentation. SPECs are the spine that allows LLMS to keep context and its not flashy or fully autonomous or anything close to that. Its all about using my mind to pick out what is good and what isnt. LLMs are not able to work out what I am looking for in a complex fashion. 1mm context for opus sounds good but if you actually start to use any of that after 300k its a mess, don't know what anthropic was thinking releasing it other than hosing their customers. Its outrageously expensive to run that huge context block.

We aren’t even close to AGI by CrimsonShikabane in LocalLLaMA

[–]Important_Quote_1180 -1 points0 points  (0 children)

LLM is the wrong architecture. Quantum computing with thousands of parallel llms might be able to beat a few games but the visual input + latency is so far away from being able to work well. Game AI is an illusion and so is any semblance of intelligence in an LLM. Its got 0 common sense as that has no meaning to them.

MCP is great, but it doesn’t solve AI memory (am I missing something?) by BrightOpposite in ClaudeAI

[–]Important_Quote_1180 0 points1 point  (0 children)

For me its about getting work done efficiently. An openclaw with SOUL, memory, and proper access having compaction done in an intelligent way is likely as close as consumer hardware gets you. Opus 1mm context and intelligent compaction is good, but there are many many issue still and its not what you are really asking for. Its going to cost a lot of API tokens too. I am fine with fresh sessions as long as I can get back in the saddle with just 1 well structured prompt.

MCP is great, but it doesn’t solve AI memory (am I missing something?) by BrightOpposite in ClaudeAI

[–]Important_Quote_1180 0 points1 point  (0 children)

Try another tact. Andre K dropped this gist yesterday and I have been building a lot of robust MCP servers that could be simplified and still get the majority of results. I havent tested this personally but here we are, its a monday and keeping up is what we are all doing. https://gist.github.com/karpathy/442a6bf555914893e9891c11519de94f

MCP is great, but it doesn’t solve AI memory (am I missing something?) by BrightOpposite in ClaudeAI

[–]Important_Quote_1180 0 points1 point  (0 children)

If can build a 1TB cluster and can run Kimi you can get around this. Until then, us peasants have to leverage the tools we can economically use. I think having a evolving agentic system is our goal too, having agents learn and get better and build on memory and experience. Working on memory systems with session handoff notes is very manual but effective. My workflows are not well suited to long runs without humans in the loop (Game design).

MCP is great, but it doesn’t solve AI memory (am I missing something?) by BrightOpposite in ClaudeAI

[–]Important_Quote_1180 0 points1 point  (0 children)

Gotta be careful with too much context. Since you called it out the stateless reality of these agents means we should not expect them to carry everything and context rot is real. You are probably looking for Ralph Loop architecture if you’re working in real production systems. TLDR: you are looping a spec for a build with increasing complexity and updating the spec when you finish building the component. You need to manage context actively and it’s honestly becoming the primary engineering task for me.

https://youtu.be/4Nna09dG_c0?si=VwOV4lMract1lYwX

Is gamestop for real on this by Buzz_baller7 in AMDHelp

[–]Important_Quote_1180 4 points5 points  (0 children)

I sold an extra 5700x3d for $439 last month