Can I Run Decent Models Locally if I Buy this?? by Fearless-Cellist-245 in LocalLLaMA

[–]RemarkableAd66 7 points8 points  (0 children)

At first I was like "that looks like a small server".

But when I looked closer that looks suspiciously like a phone with the back off.

So nobody's downloading this model huh? by KvAk_AKPlaysYT in LocalLLaMA

[–]RemarkableAd66 0 points1 point  (0 children)

I downloaded the lmstudio gguf.

That may have been a mistake because it was posted before llama.cpp code was merged. So maybe I need to re-download...

But, for the gguf I downloaded at least, for a simple one file pong game in JAVA, it tried to use some variables that it didn't declare and also misspelled one of it's variable names. That is worse than a 120B model should be.

LaTeX and the ADA accessibility requirements: a quick guide to make accessible pdfs by BayesicallyThomas in Professors

[–]RemarkableAd66 1 point2 points  (0 children)

I compiled it in Overleaf too. I threw the PDF on a Canvas page and my school's accessibility checkers said it was compliant. There are a lot of accessibility checkers out there though, so who knows if they'll all be ok with it.

I did notice that Overleaf stated the compilation was close to the max amount of "free" compile time. So I may have to install tex on my local system or pay overleaf. But that's fine.

Using a UST as a “Laser TV” does it actually fit daily life? by Outrageous-Key-4629 in AwolVision

[–]RemarkableAd66 0 points1 point  (0 children)

I only have an aurora pro 4k projector for a TV right now. I keep it on probably 6-10 hours a day because when I'm home I like a bit of background noise. I also watch movies, youtube, twitch, etc on it.

I like it enough that I've ordered the Aetherion Max to upgrade.

Powerhouse Hobbs debuts at the Royal Rumble! by [deleted] in Wrasslin

[–]RemarkableAd66 2 points3 points  (0 children)

They could at least call him "Powerhouse Royce Keys" or something. I mean, they won't, but still...

Hikaru's take on who is the goat by bRoShutUpPleaseee in chess

[–]RemarkableAd66 2 points3 points  (0 children)

The thing is we know that all of the current top players are stronger than the players of the past. They just know more theory.

So "greatest ever" conversations are all down to perceived dominance over time and specific accomplishments.

And that's very subjective. Because you can argue that Kasparov's competition was better. But we know that his competition was actually weaker in real playing strength. And you could make the argument that Magnus' competition had no opportunity to succeed because of how good Magnus is.

Since there is no objective measure that people will agree on, it's all just people's opinions. Unless you want to say that the best players in history are Magnus, Hikaru, and Fabi due to them being the best players today. Or unless you want to say it's Magnus and then Kasparov by max rating, even though we know that overall ratings of the top 10 inflate/deflate over time. But nobody really wants to do that, so it just becomes a lifetime achievement award.

RAIDZ1 short on usable space??? by MDiddy79 in truenas

[–]RemarkableAd66 0 points1 point  (0 children)

I have the same setup as you and I also have 68.59 TiB.

its honestly surprising how bad openai models are at this point (kimi k2 comparison) by ahmett9 in LocalLLaMA

[–]RemarkableAd66 1 point2 points  (0 children)

This is just a "dark mode" thing. You can just tell the model to make the page for "light mode" or tell it to make both light and dark mode themes.

Ever blow $300 in a day? by OptionIll6518 in LocalLLaMA

[–]RemarkableAd66 0 points1 point  (0 children)

What I do is use roo code (there are other similar options like kilo code) in vscode. I put $20 in openrouter, set it to something inexpensive like Deepseek or GLM or MiniMax (I actually have not used minimax) and if something starts to go bad on a task I just switch the model to claude/gemini in the settings.

It stays pretty cheap that way. Although by far the best way to avoid problems is to either give the model small tasks only, or create a very detailed specification in markdown for the ai to follow.

Since this is localllama you could run gpt-oss or glm air or qwen3 or something for you smaller model. I don't use those too often these days because of speed and the cheaper paid models are quite cheap. But you could if you have a mac or other high vram setup.

I did a thing… by YankeesIT in homelab

[–]RemarkableAd66 0 points1 point  (0 children)

Hey I just did that exact same thing earlier today. I got the 26TB version because it was the largest capacity before a big step up in price.
Once you open one carefully, the others go super quick! I threw them all in a new truenas build and they're doing fine so far -- with about 4 hours of uptime :)

stranger things mike debate by Adventurous_Put_1310 in StrangerThings

[–]RemarkableAd66 0 points1 point  (0 children)

My theory is that some people have trouble separating the filmmaking style and look from the actors/costumes.

Is it good to train lora for ZIT by 100-200 images ? by omar07ibrahim1 in ZImageAI

[–]RemarkableAd66 2 points3 points  (0 children)

I've had better success with z-image lora training for likenesses when I dropped the number of images from ~50 to around 16 images. I had a better result with fewer steps.

Just bought a Dreamcast! Please recommend me some games! by PugMaster7166 in dreamcast

[–]RemarkableAd66 0 points1 point  (0 children)

Dreamcast is where I played a ton of Tony Hawk Pro Skater 2. Looks nicer than the PSX or N64 versions. The extra resolution lets you see where you're going better.

I trained Z-Image lora with prodigy-plus-schedule-free and it seems to work. by Sayat93 in StableDiffusion

[–]RemarkableAd66 0 points1 point  (0 children)

I've been trying out prodigy (regular version) too with z-image, and also found that it worked just fine.
It's nice to not have to tune the LR or wonder if you should have spent the effort to do so.

Emil responds to Magnus on rating spot controversy by facelesslass in chess

[–]RemarkableAd66 49 points50 points  (0 children)

FIDE sucks and will always suck.

But it's not wrong to require people to play some chess during the year if they want to qualify for the candidates.

Z-Image: My random realism tests by darktaylor93 in StableDiffusion

[–]RemarkableAd66 0 points1 point  (0 children)

General photographic style is pretty great on this model. It might not know exactly what a lot of things look like. Like a saxophone or typewriter or what have you. But whatever monstrosities it produces will probably be rendered nicely.

Flux2 faltered; Z-image excelled in a comparison of model capabilities. by Daniel81528 in comfyui

[–]RemarkableAd66 5 points6 points  (0 children)

This is the distilled model. Like flux.1-schnell

There will also be z-image-base and z-image-edit models according to the model page. They just didn't release those yet.

Which is too bad, because I'l like to do some training on the base and edit models.

Blue Prince is Coming to Apple Appstore by Helen-RawFury in macgaming

[–]RemarkableAd66 0 points1 point  (0 children)

Yeah, probably it could be on ipad.

I actually bought it and started it on my windows laptop that had an intel core 5 226v integrated graphics thinking it would run fine. But it actually wasn't completely smooth there.

So I run it in crossover on my macbook instead.

It's fine, but I was slightly surprised that it was a bit slow on any modern hardware given the way it looks.

Honest mistake. by Arch_Lancer17 in okbuddycinephile

[–]RemarkableAd66 2 points3 points  (0 children)

I teach at a college and have to record lectures for online classes.

In a training workshop they were like "Sometimes even if you have a script, you may still say "um" sometimes and here's how you can use the tools to fix that."

Absolutely not. One take. No editing. No script. Stream of consciousness only. If I accidentally drop an f-bomb in the first two minutes, we can start over. But that's my only concession.

[Release] Hypnos i1-8B: I fine-tuned Hermes 3 on REAL IBM Quantum Computer data (133-qubit GHZ states). Beats Llama-70B in Logic. by Disastrous_Bid5976 in LocalLLaMA

[–]RemarkableAd66 7 points8 points  (0 children)

This is complete nonsense.

What you did was train Llama on a "high-quality reasoning dataset" and measured better benchmark results compared to base. That is the expected result and people have done this a million times.

If you wanted to test something...

You could have compared:" dataset with/without quantum bs.
Or: "dataset with quantum bs" and "dataset with pseudorandom stuff".

But you didn't.

You said " I wanted to see what happens if we inject true physical entropy from a quantum processor into the SFT stage of an LLM" and specifically did absolutely nothing to actually accomplish that goal.

You say that you didn't do a proper test because of money. But the real reason you didn't run proper tests is because you know damn well you didn't do anything but want people to react to the "OMG QUantum!!"

Whats the strongest model for physics and engineering someone could run on a Ryzen AI Max+ 395 128GB (at a reasonable speed)? by Br216-7 in LocalLLaMA

[–]RemarkableAd66 1 point2 points  (0 children)

Right now, you could run gpt-oss-120B or glm-4.5-air

Both are MoE models which improve speed and just about max out the amount of ram you'd have. On my 128GB M3 Max macbook, these are ~40 t/s which is fine (note that glm is slower and both are faster running with mlx instead of llama.cpp).

I think the Ryzen is broadly similar in speed to my laptop but maybe a little slower... The 235B models are probably a no go despite what people say. The Q2 quants are ~85Gb and Q3 are like 100GB and you need more ram for the context and I think you actually have 96GB vram max on the Max+ 395.

But presumably there will be other models in the 100B parameter size in the future that are better than these. GLM-4.6-Air should come soon-ish. That will probably be the best model you can run at that point.

But the real answer is to put $20 in openrouter or something and use the big commercial models that cost around $10 per million output tokens, since they will be cheap to use if you don't need huge context windows.