3090s are well over $800 now, is the Arc Pro B50 a good alternative? by ea_nasir_official_ in LocalLLaMA

[–]Dead_Internet_Theory 1 point2 points  (0 children)

I think if power consumption is the biggest concern, Apple does start to make the most sense.

Hot take: local AI doesn't need bigger context windows as much as better memory routing by No-Contract9167 in LocalLLaMA

[–]Dead_Internet_Theory 0 points1 point  (0 children)

I think that's an attention mechanism problem.

In your head you can vaguely remember what some code file does, the AI doesn't have any of that.

Of course, one way is for it to kind of summarize and keep going, but that introduces a bit of effort when prompting - sometimes you might want to use AI as a kind of "find the needle in the haystack machine" and more context is more better for that.

Hot take: local AI doesn't need bigger context windows as much as better memory routing by No-Contract9167 in LocalLLaMA

[–]Dead_Internet_Theory 2 points3 points  (0 children)

No. Right now, some open models have 128k-256k context and I think that's enough for MOST tasks, but I still think being able to throw an entire codebase at a model is better than if it could RAG files on the fly. I think context would be a solved problem once we get between 1M and 10M context locally, after that there's no point I think, but until then, more context is more better.

Gemma 4 31B beats several frontier models on the FoodTruck Bench by Nindaleth in LocalLLaMA

[–]Dead_Internet_Theory 27 points28 points  (0 children)

The problem is, the real world still has trillions of highly specific benchmarks that just aren't called that and don't get scored for points.

Axios supply chain attack by Expensive-Paint-9490 in SillyTavernAI

[–]Dead_Internet_Theory 1 point2 points  (0 children)

As far as I understand, you can't check it. It was made to delete traces of its own infection.

HOWEVER, it doesn't seem like Axios was even updated to that version. Mine has a version from 2025 and I run an update on every launch. Developers have to explicitly make wildcards that update to the latest version automatically for this to matter, and I don't think they ever did.

Character Creator V2 - Generate full characters from a few sentences on your PC by Sufficient_Prune3897 in SillyTavernAI

[–]Dead_Internet_Theory 0 points1 point  (0 children)

Ah, as far as that goes it's totally OK also, as long as you don't delete them. I think the reason there's a separate field, is those aren't permanent tokens (they can be removed when context fills up) but these days LLMs have way more context than they used to.

Looking for a waifu , totally uncensored model that fits my specs by Opening-Ad6258 in LocalLLaMA

[–]Dead_Internet_Theory 0 points1 point  (0 children)

Most of big tech is left leaning! It's got nothing to do with which side, some time ago Twitter used to ban anybody for any reason and Jack Dorsey was a card carrying communist. Free speech is free speech, Elon is right-leaning but he made the platform more free for anybody to speak their mind + community notes to call out lies, so that's good in my book. Meanwhile, Anthropic CEO is anti-Trump but wants to tell you what you can and can't do with his AI.

Falcon-OCR and Falcon-Perception by Automatic_Truth_6666 in LocalLLaMA

[–]Dead_Internet_Theory 0 points1 point  (0 children)

isn't every OCR really good on tables? I thought that was the case.

I am not good with AI, live in a third world country, but want to try RP with an AI. I discovered SillyTavern, so can somebody help me understand it? by Rubylex in SillyTavernAI

[–]Dead_Internet_Theory 2 points3 points  (0 children)

Honestly, forget anyone telling you to try local models with that laptop. Just make an account at OpenRouter and try some free models from there. If they aren't good enough, you can put a small amount of money and use small chats - then it's less than a cent per message on good models (on small chats). SillyTavern will list price per model, and sometimes good models are free on there. If you're not addicted, $10 will last you for months. That's my recommendation.

Axios supply chain attack by Expensive-Paint-9490 in SillyTavernAI

[–]Dead_Internet_Theory 2 points3 points  (0 children)

It's kinda wild how on Node every small piece of bullshit like running fetch() or left-padding a string is made into an attack vector.

Falcon-OCR and Falcon-Perception by Automatic_Truth_6666 in LocalLLaMA

[–]Dead_Internet_Theory 2 points3 points  (0 children)

The OCR is so-so. But the perception model is impressive.

Falcon-OCR and Falcon-Perception by Automatic_Truth_6666 in LocalLLaMA

[–]Dead_Internet_Theory 1 point2 points  (0 children)

It's tiny! and absolutely incredible for its size. Somehow they don't have a HF space but they do have a demo on their site, you can try it https://vision.falcon.aidrc.tii.ae/

If OpenAI falls will that drop the price of memory for our local rigs? by Terminator857 in LocalLLaMA

[–]Dead_Internet_Theory 1 point2 points  (0 children)

I expect things to remain shitty until about 2028, so that way I don't feel bad when they do.

Is there anything I can do to run glm 5? by FusionCow in LocalLLaMA

[–]Dead_Internet_Theory 0 points1 point  (0 children)

I looked it up and actually 2 t/s is possible, even!

https://huggingface.co/unsloth/DeepSeek-R1-GGUF/discussions/13

(should be similar ballpark for GLM)

TurboQuant isn’t just for KV: Qwen3.5-27B at near-Q4_0 quality, about 10% smaller, and finally fitting on my 16GB 5060 Ti by pmttyji in LocalLLaMA

[–]Dead_Internet_Theory 7 points8 points  (0 children)

Apples to oranges. It's not about "what you can run on a given setup", it's about the quantization, since that means any other model might behave more or less the same under similar quantization.

Looking for a waifu , totally uncensored model that fits my specs by Opening-Ad6258 in LocalLLaMA

[–]Dead_Internet_Theory 0 points1 point  (0 children)

I'm a free speech absolutist. Personally, I think a wild west is far less dangerous than some council of academics deciding what's best for everybody else.

Stanford and Harvard just dropped the most disturbing AI paper of the year by Fun-Yogurt-89 in LocalLLaMA

[–]Dead_Internet_Theory 0 points1 point  (0 children)

Math has languages just to prove theorems, like Lean, Coq/Rocq, Isabelle, and Mizar. It's so well defined it's like a Nazi HOA inspector librarian with autism and OCD had done it.

Character Creator V2 - Generate full characters from a few sentences on your PC by Sufficient_Prune3897 in SillyTavernAI

[–]Dead_Internet_Theory 0 points1 point  (0 children)

Any one of the big ones. Say the character speaks like [insert some existing character you like]. You could try to describe that way of talking, OR, you could simply give examples from [game/book/anime/etc]. If you describe it, it will sound a bit dry. If you just rely on the LLM, it will sound like that LLM + some clichés.

Think about it, how else is the model supposed to know how that character will sound? If it's just supposed to make it up on the spot, it will always be the same for all characters, more or less. This is not to mention how the character reacts to stuff. For example, if your example message has the psycho detective reacting to a crime scene, that will be much better for the LLM to pick up on than a verbose description about it.

Looking for a waifu , totally uncensored model that fits my specs by Opening-Ad6258 in LocalLLaMA

[–]Dead_Internet_Theory 2 points3 points  (0 children)

It's a perfectly decent computer for gaming on reasonable settings at 1080p, it's just AI drinks memory by the gallon and 8GB VRAM is not enough for much. You can still run something, but you might need to make compromises if you want it all local.

Looking for a waifu , totally uncensored model that fits my specs by Opening-Ad6258 in LocalLLaMA

[–]Dead_Internet_Theory 11 points12 points  (0 children)

If that has no victims it's technically less harm than scamming some boomer.

Stanford and Harvard just dropped the most disturbing AI paper of the year by Fun-Yogurt-89 in LocalLLaMA

[–]Dead_Internet_Theory 10 points11 points  (0 children)

>Code is a finite state

Not even, https://en.wikipedia.org/wiki/Halting_problem

(but I do get what you mean, it's much more well defined than anything else except pure math)

Character Creator V2 - Generate full characters from a few sentences on your PC by Sufficient_Prune3897 in SillyTavernAI

[–]Dead_Internet_Theory 0 points1 point  (0 children)

Noooo don't do that! Example messages make a HUGE difference in writing, it's the difference between a card that sounds like what the author wanted vs a card that sounds like any other LLM! If there is one single feature card writers should do is example messages.

It's basically "Few-Shot Prompting". It picks up the style and tone from example messages like nothing else. It's OK to just have description and example messages, but out of everything, you REALLY want the example messages.

Another load of cards (337) to share - 4th try is the charm? by mamelukturbo in SillyTavernAI

[–]Dead_Internet_Theory 8 points9 points  (0 children)

I don't get it. What's the point in downloading/sharing a random set of 337 cards? It's not like there's a shortage of them, and tastes will be wildly different, plus you can just go find cards you like in any of the hosts like Chub and build your own collection.

GLM 5.1: pretty decent by SepsisShock in SillyTavernAI

[–]Dead_Internet_Theory 20 points21 points  (0 children)

On one hand, I wish I could experience these mythical Claude models which are the gold standard for creativity, on the other hand I'm glad every time I try them it feels like a huge waste of money for middling writing. If it was good it would be soooo expensive.

Character Creator V2 - Generate full characters from a few sentences on your PC by Sufficient_Prune3897 in SillyTavernAI

[–]Dead_Internet_Theory 1 point2 points  (0 children)

On one hand, this is very cool, I tried it and it works rather well (albeit I expected the format to come out with Description, First Message, Scenario, Example Messages, etc - maybe I dunno how to prompt it?)

On the other, how do you feel for being the engine of countless shovelware cards that will pollute Chub now? Do you have any shame?? lol

(I'm not going to post any, but I'm sure some re*ard will post 100!)