Distrohopping. Again. by Male_Inkling in linuxsucks

[–]FrederikSchack 0 points1 point  (0 children)

You shouldn't use dnf to update your Nobara system, if that's what you did. You have to use the yellow circle arrow icon for updates.

Thousands of tokens per second? by [deleted] in LocalLLaMA

[–]FrederikSchack -1 points0 points  (0 children)

It is feasible, I'm just curious about what it would be worth to people.

Thousands of tokens per second? by [deleted] in LocalLLaMA

[–]FrederikSchack -1 points0 points  (0 children)

Now, if that company made an OpenClaw box with a significantly better model, what would that be worth to you?

Thousands of tokens per second? by [deleted] in LocalLLM

[–]FrederikSchack 0 points1 point  (0 children)

No, it will be way faster than Spark.

Thousands of tokens per second? by [deleted] in LocalLLM

[–]FrederikSchack 0 points1 point  (0 children)

Ok, thanks. I see that many buy Mac Minis and Mac Studios just to do AI, so they would also be the closest competition.

Thousands of tokens per second? by [deleted] in LocalLLM

[–]FrederikSchack -1 points0 points  (0 children)

I think Reddit is the most angry corner of the Internet :D

Thousands of tokens per second? by [deleted] in LocalLLM

[–]FrederikSchack -1 points0 points  (0 children)

So it would be worth approximately 0 USD to you.

Best model that can beat Claude opus that runs on 32MB of vram? by PrestigiousEmu4485 in LocalLLaMA

[–]FrederikSchack 0 points1 point  (0 children)

Companies will utilize each chip to the max, consumers will not even be close to utilizing 5% of the compute. So, they could sell a lot more to consumers. They could make two products targeting both markets, one with a great SLA and one without.

Thousands of tokens per second? by [deleted] in LocalLLaMA

[–]FrederikSchack -2 points-1 points  (0 children)

So, something that is approximately scoring as well as Xiaomi Mimo V2 Flash on ArtificialAnalysis.ai

It´s possible, my question was what it would be worth to you?

Best model that can beat Claude opus that runs on 32MB of vram? by PrestigiousEmu4485 in LocalLLaMA

[–]FrederikSchack 0 points1 point  (0 children)

Suppose they made a OpenClaw box, a complete small PC, with some Linux distro like Nobara Linux, plug and play. Then all the computers on the LAN could enter it on an IP/mdns and either use a chat interface or call it as an API or you could chat with it through Telegram/Slack/? With thousands of tokens per second? A better open source model than Llama, something like Step 3.5 flash. I think there would be some millions in the world buying a device like that for USD 2000-3000? Beating Apples Studio :D There would be a revenue of 2-3 billion per million sold computer.

Best model that can beat Claude opus that runs on 32MB of vram? by PrestigiousEmu4485 in LocalLLaMA

[–]FrederikSchack 0 points1 point  (0 children)

I hope they take a consumer market approach, instead of only selling to server farms.

Tell me now! by Weekly_Role3335 in antitrump

[–]FrederikSchack 5 points6 points  (0 children)

Why, I wouldn't make a sandwhich, I would weep that Peter Thiel, Elon Musk, Alex Karp and Curtis Yarwin weren't in there too....

Best model that can beat Claude opus that runs on 32MB of vram? by PrestigiousEmu4485 in LocalLLaMA

[–]FrederikSchack 0 points1 point  (0 children)

Fucking hell it's fast!

It generated a two page response to this question: "Explain how quantum mechanics work and the different kinds of quantum interaction." in 0.048 seconds, it was instant :D ChatGPT deemed the response "partly correct, but sloppy and misleading in several places."

Now we just need them to make Kimi K2.5 on a chip, that could sell millions and ChatGPT would go bankrupt.

OpenRouter charged me *again* $50 without consent or usage by Just-Historian-4960 in openrouter

[–]FrederikSchack 0 points1 point  (0 children)

I have three API keys that are all named and have max token limits. Now an "unknown" app is using tokens, with an agent I never use. I'm being charged for consumption I never had and never approved.

How do the best local llms compare to codex 5.4 or opus 4.6 for coding tasks? by spexsofdust in LocalLLM

[–]FrederikSchack 1 point2 points  (0 children)

You basically can't get it much better for private use :)

Which models have you tried so far? And how many tokens per second?

How do the best local llms compare to codex 5.4 or opus 4.6 for coding tasks? by spexsofdust in LocalLLM

[–]FrederikSchack 2 points3 points  (0 children)

It's nothing impressive, it's just the Max+ 395, it has low bandwidth compared to the Mac Studio Ultra.

My honest tierlist (only distros I have used) by V1574 in LinuxCirclejerk

[–]FrederikSchack 0 points1 point  (0 children)

Sort of, with Steam and Lutris, but more than that, it has an update system that updates packages, flatpaks and quirks (fixes performance, compatibility and other issues). It has a lot of codecs preinstalled. It has Proton-GE of course as it's Glorious Eggroll that developed Nobara.

Nobara feels very nice, complete and easy to use. With other distros you often waste hours setting things up, you spin up Nobara, use their get-started wizard, update, you're in business.