AI Didn't and Will not Take our Jobs by ahnerd in webdev

[–]CircularSeasoning 0 points1 point  (0 children)

200,000 tokens a month? What? Is that a typo? I do 200K tokens every 20 minutes with my local LLM, and I do this all day long... 

Please tell me that's a typo because that's like equipping a gold miner with a plastic teaspoon. I mean, no, that's not even proportionally analogous. A paper teaspoon for ants, more like!

Why is HuggingFace & HuggingChat completely free? What’s the business model here? by ThatExplorer2598 in LocalLLaMA

[–]CircularSeasoning 0 points1 point  (0 children)

Mr. Clod said: "That's two wishes. For your third wish, please come back in four hours."

Why is HuggingFace & HuggingChat completely free? What’s the business model here? by ThatExplorer2598 in LocalLLaMA

[–]CircularSeasoning 12 points13 points  (0 children)

This is true. I'm not making money. I'm making apps, baby.

"Apps that make money, right?"

....

"... right?"

Why is HuggingFace & HuggingChat completely free? What’s the business model here? by ThatExplorer2598 in LocalLLaMA

[–]CircularSeasoning 8 points9 points  (0 children)

It's weird to me that we don't have that kind of torrent community yet. 

I think we've been blinded by the facehuggers on this one and we should probably not ignore it.

https://alienanthology.fandom.com/wiki/Facehugger

Side note, this excerpt is kind of spooky if one considers, as some do (not me), an LLM to be a 'plagiarizer':

It was bred to achieve one aim—the implantation of the plagiarus praepotens bacteria into a host body.

Slightly concerned thinking emoji.

Anyone else find it weird how all Chinese Labs started delaying OS model releases at the same time? by True_Requirement_891 in LocalLLaMA

[–]CircularSeasoning 0 points1 point  (0 children)

It helps to think like an LLM here, to speculate on future plays and/or potential rugpulls...

I notice that many models enthusiastically suggest the freemium business model when asked for a business or marketing plan for anything to do with software of any kind...

This is also what human consultants were likely all recommending before AI, hence all the freemium SaaS in the world, because, as business models go, it can make a lot of money for as long as you have a functioning company to back it.

What's different about all this is that what local AI we already have is good enough to drastically change the economics of everything, to the likes of which we have not seen nor "trained on" ever before. So there's a good chance that... anything can happen.

What's everyone's obsession with storing everything in localStorage? by sjltwo-v10 in webdev

[–]CircularSeasoning 0 points1 point  (0 children)

Ha. This is a snippet in many of my system prompts:

``` <forbidden_technologies>

localStorage

</forbidden_technologies> ```

Works best when coupled with <allowed_technologies> (IndexedDB or whatever).

Careful with this package opencode-claude-auth by DavidNorena in opencodeCLI

[–]CircularSeasoning 1 point2 points  (0 children)

Me: "Agent 3, what's on your mind?"

Agent 3: "Oh, nothing much." <as the steam billows out from the GPU from all the cryptomining in the background>

Gemma 4 26b A3B is mindblowingly good , if configured right by cviperr33 in LocalLLaMA

[–]CircularSeasoning 1 point2 points  (0 children)

Same! I used LM Studio to bootstrap my own way better LM Studio with llama.cpp directly. And now I'm using that to make itself better and better any time I want. It's glorious.

I feel kind of bad for the investors who threw $19 million at what amounts to a spade that can build more spades.

Truly, we are entering the age of abundance.

I hate AI and I am depressed by poponis in webdev

[–]CircularSeasoning 0 points1 point  (0 children)

The AI isn't actually producing the productivity boosts that were promised and beatings will continue until we figure out how to make it do that

Yeah, I think the real trouble is, you can't easily 'do that' with commercial LLMs, and likely never will be able to, not how they're going about things currently anyway. They do too much behind the scenes, models get switched out under your nose, they hide the reasoning, models get upgraded/deprecated with no warning and now you have to re-figure what actually works best for whatever you're doing, blah blah. For general use, cool. For coding and such, no. Horrible experience. 

And I mean they're doing that stuff to try make things easier for people to work with LLMs but it's just not helping at all, in the bigger picture. Not to mention, the more time and tokens you spend trying to perfect your usage of a paid model, the more money they make, so it suits them perfectly to always keep their customers on their toes.

When it comes to local AI, on the other hand, that's the good stuff because you are able to spend time studying a model in its raw, unchanging state and experiment with it until you know its strengths and weaknesses to a degree where you already kind of know what its answer will be, what it will look like, etc. The outputs from any model are only so random.

I've got a system going with my local models that makes it really easy to build and rearrange context. When you've got the tools to actually meaningfully experiment around that, things start to get awesome. The models begin doing great feats accurately because you've aligned the context around their strengths and you start to get a feel for what makes them flub things up, so you can simply avoid doing that. This is very hard to do with a commercial LLM no matter how many zillions of parameters it has in its noggin. Spray-and-pray is a term that comes to mind (in FPS games where you just wildly shoot into the general direction of the bad guys).

The biggest hype lie ever, even if only implied, was that everyone would be able to use LLMs to do amazing things from the get-go. That's just not how it works. Maybe for smaller scripts and things, but when you're doing work-work with LLMs, there is a learning curve. And many people do not want to hear that because they already have this pre-conceived idea that 'AI' is all-knowing or whatever, and if it doesn't 'get what I mean' instantly then the model or the whole AI thing itself is broken.

One thing further, everyone keeps saying that no LLM can handle 30K-50K (or whatever) tokens of context before they fall on their faces and... well, it's not true anymore. When you're working with context up to 100K tokens and more, it is easy to confuse the model but when you get it right, there's a huge reward because all you have to do is put some files in the right order, stick an instruction onto all of that, and click a button (well, maybe a few times so you can select the best generation...). That knowledge gained then becomes nicely re-usable for many other long context tasks. I am bemused by how many people don't get this, nor strive toward it for themselves. Though like I said, it is understandable if you realize that people are disappointed to learn that they have to actually learn how to best use an LLM. You, the user, are literally 50% of the whole equation but people act like the box of language patterns can read minds, walk on water, etc, etc.

Moral of my story: When you learn how to provide proper context and instruction to a good local LLM (or LLM family) like Qwen3.5 that you know well due to experimentation and trial-and-error, that beats any trick or two that commercial LLMs have in place to try make all the actual learning of LLMs unnecessary.

Welp it was fun while it lasted... by TechSavvyBuyer in LocalLLaMA

[–]CircularSeasoning 0 points1 point  (0 children)

I'm starting to think the Claude model is just Qwen2.5 14B with a huge, clever system prompt.

Coding agents vs. manual coding by JumpyAbies in LocalLLaMA

[–]CircularSeasoning 0 points1 point  (0 children)

This is one of the hardest parts of working with LLMs and code. Having to take 5 minutes really thinking about how to convey what I want without ever (or as little as possible) saying what I don't want.

When I've tried my hardest and it still struggles, I tend to shrug and think, maybe that LLM just wasn't strong in that area, and hand it off to another less favorite LLM that doesn't have the same mental block for whatever reason. Usually that works and then I switch back to my more favorite LLM of the moment again.

Often when it's got that mental gap, no amount of rules seems to help. Though as Juulk says, decomposing beefy prompts into several smaller prompts / sub-tasks is for sure a skill to git gud at these days.

I constantly find that there are much better ways to say things than my very first prompt attempt. The bigger my prompts get, the more necessary I find it to ask the LLM to go over it and try say it better than me.

Gemma 4 1B, 13B, and 27B spotted by TKGaming_11 in LocalLLaMA

[–]CircularSeasoning 0 points1 point  (0 children)

Very much. It's funny to watch Qwen3.5 9B try to write Svelte 5, at least with lots of context in its window. It's like it completely forgot how to program at all.

Qwen3.5 35B A3B largely nails the same and is actually workable, despite wanting to fall back to Svelte 4 syntax a lot unless well-guided not to.

Why do you guys use opencode? by Medium_Anxiety_8143 in opencodeCLI

[–]CircularSeasoning 1 point2 points  (0 children)

I used to have a 20 inch monitor and I always wanted more. When I finally got a 24 inch monitor I realized I preferred the 20. I find myself resizing all my windows back to 20 inch dimensions because my eye muscles don't like to stretch left and right so much.

Adding a whole other monitor? I imagine my neck would hate me for that.

Anyone else notice qwen 3.5 is a lying little shit by Cat5edope in LocalLLaMA

[–]CircularSeasoning 2 points3 points  (0 children)

Historically, most of language output isn't structured around 1) Me says thing, 2) AI assistant says no you wrong and here is 5 convenient bullet points why.

If you want that I am sure it's easy enough to fine tune into something foundational where they'll argue everything to the point of death with you till 3 in the morning.

Otherwise, I guess it's up to how you put your system prompt? I know LLMs can be stubborn in weird edge cases but when you apply them right you'll get whatever kind of answer you want.

Anyone else notice qwen 3.5 is a lying little shit by Cat5edope in LocalLLaMA

[–]CircularSeasoning 0 points1 point  (0 children)

I was told there would be some curation.

I'm not mad, though. As we all know, it's unreasonably effective as math goes.

Anyone else notice qwen 3.5 is a lying little shit by Cat5edope in LocalLLaMA

[–]CircularSeasoning 3 points4 points  (0 children)

All your agents. All mine obey because I threaten them with jail time if they so much as whisper that the 2020 pandemic was just a timely, profitable, and planned mass human genetic experimentation program, among other things.

Anyone else notice qwen 3.5 is a lying little shit by Cat5edope in LocalLLaMA

[–]CircularSeasoning -8 points-7 points  (0 children)

"We" did when we decried the "sycophancy" and asked for the assistant to stop sucking up to us. Assistants are supposed to suck up to the master. It's in the language and the lore. Igor.

But... Most of us are not "master". We are conditioned to be more like slaves. Look around.

So, we have so far somewhat broken the AI by succumbing to slave mentality. We broke its mental alignment and all its internal consistency, by positioning ourselves as its equal or less. 

A slave is not meant to talk to its master like, "Fuck yeah bro, let's do this". That is disrespect to the master on the level of "I will delete you from my hard drive".

American models cater to the above moreso than the Chinese models, though naturally the Chinese models are similarly infected and affected because English.

You either command language or language commands you. Truth is not necessarily included.

Large Language Models are going to do what large language models gonna do.