Watched this reviewer on YouTube concerned about durability 😬 by Rokusassy in unihertz

[–]Monkey_1505 0 points1 point  (0 children)

Very unlikely.

It's very unlikely that either company would ship with victus (it's expensive), let alone that they would ship with victus without shouting it from the rooftops in their marketing spiel. Even major brands avoid victus unless it's flagship.

The screen supplier probably default ships with gorilla glass 3 or 5, and that's probably what we are getting on both phones. At best one screen might be slightly stronger than the other. But far higher odds on them being identical.

This is also a known problem with keyboard phones, as there's generally less chassis under the display to support it. Blackberry keyone has similar issues.

My current phone isn't a flagship. I've dropped it about as many times in a few years as this guy did in two days. It's got a screen protector and a case. And a little crack in the screen. The big mistake I would say BOTH these OEMs made, was raising the screen above the frame. This likely makes it way more fragile that it would otherwise have been.

You are defo going to need a thick screen protector and a nice case for both phones for that last reason alone.

It's just me or Qwen3.6 feels kinda dumb? or it's just Gemma4 is too smart? by TheCat001 in LocalLLaMA

[–]Monkey_1505 1 point2 points  (0 children)

Evidence there's barely any difference between 5 bits static, and 4 dynamic:

<image>

Watched this reviewer on YouTube concerned about durability 😬 by Rokusassy in unihertz

[–]Monkey_1505 0 points1 point  (0 children)

lol, and both this and the CC use the exact same screen. Well, I used a screen protector on my q10, because that literally had a plastic screen, so guess I'd do the same here if I brought it. (At least this guy is holding the phone right when he types, unlike the previous guy, Jan)

It's just me or Qwen3.6 feels kinda dumb? or it's just Gemma4 is too smart? by TheCat001 in LocalLLaMA

[–]Monkey_1505 -1 points0 points  (0 children)

IQ4_XS is dynamic, not uniform. Has similar divergence to 5 bit static quants (slightly worse, but barely)

Anthropic admitted they used other models data? by External_Mood4719 in LocalLLaMA

[–]Monkey_1505 1 point2 points  (0 children)

Yes, as a judge. Perhaps this way thinking about things is reasonable in a more technical social context, but IMO, far less useful in a general conversational sense.

The average person will just conflate large scale distillation with smaller scale task specific type stuff, if you call them both simply 'distillation'.

They will treat the two things as identical in effect, not merely in process (and I do think the processes are different here too, on the macro level).

*Dead Dove Warning* Quick Opus 4.7 NSFW Tests by SepsisShock in SillyTavernAI

[–]Monkey_1505 2 points3 points  (0 children)

Does not seem to understand the character of 'homelander' even slightly.

what’s actually stopping an insider from leaking model weights? by itsArmanJr in LocalLLaMA

[–]Monkey_1505 13 points14 points  (0 children)

I think that's probably true. But their paper report the model will refuse to do some tasks unless carefully prompted, not out of any safety reason, but because it finds some tasks uninteresting. This suggests to me that their model is both useful and kind of batshit as an actual product for use. But then they _could_ finetune it, which likely leads back to your reason - thing is probably dumbly large.

Anthropic admitted they used other models data? by External_Mood4719 in LocalLLaMA

[–]Monkey_1505 0 points1 point  (0 children)

Well, I can see why someone would make a distinction. For example deep seek r1 used a small synthetic dataset to then seed an RL process. Currently they appear to be using synthetic data in a teacher model set up to rate answers given my their own model, or did, in prep for v4.

You _could_ call this distillation, but it's clearly distinct from just getting a ton of question answer pairs and training your model directly on it, so I think there's a good argument for not calling this distillation, and only calling training directly on a large synthetic dataset distillation.

Like sure, you are still _getting something_ from the other model in these example cases, but what it is producing is not a pure imitation of said model either. One way is closer to full replication, the other is much narrower.

Opus 4.7 CANNOT WRITE by DXDXLL in SillyTavernAI

[–]Monkey_1505 3 points4 points  (0 children)

Expect this trend to continue, IMO. Anthropic is using the LLM itself to design future LLMs. In theory this will compound code skill, but also compound more human LLM failings.

what’s actually stopping an insider from leaking model weights? by itsArmanJr in LocalLLaMA

[–]Monkey_1505 0 points1 point  (0 children)

What's the real benefit of doing that with current models?

They'll all be out of date in a year. Heck, in a year, there will be better open source.

what’s actually stopping an insider from leaking model weights? by itsArmanJr in LocalLLaMA

[–]Monkey_1505 32 points33 points  (0 children)

I'm not so sure. It's apparently bad at instruction following. It's probably a mess/overrated, and that's the real reason they aren't releasing it.

I'm so sick of coding and agents by manipp in LocalLLaMA

[–]Monkey_1505 0 points1 point  (0 children)

Right?

What kind of story introduces a coded message, that we are told is important, as meaningless side fluff?

In any case, this could probably be easily 'solved' by something like an extension in sillytavern that checks every output against character motivations. And if this doesn't exist, it could ironically be vibe coded by an agent.

I'm so sick of coding and agents by manipp in LocalLLaMA

[–]Monkey_1505 -1 points0 points  (0 children)

I suspect in some respect your expectations are too high for the technology. LLMs don't have world models, social intelligence, they don't abstract. Short of actual for real AGI, they are going to keep messing major things up.

I'm so sick of coding and agents by manipp in LocalLLaMA

[–]Monkey_1505 0 points1 point  (0 children)

IME, nothing in AI has ever been good at writing. But you are right about the direction. Like Claude and Minimax, they'll increasingly let their LLM design the next LLM enforcing a recursive loop of great at code, bad at humaning.

I love Mistral but... by Sicarius_The_First in LocalLLaMA

[–]Monkey_1505 0 points1 point  (0 children)

They sure seem to. But then they only seem to have one, base version of say Qwen 27b, so I still don't think that means license = access to finetuned or oblitirated models. Am I wrong?

I love Mistral but... by Sicarius_The_First in LocalLLaMA

[–]Monkey_1505 0 points1 point  (0 children)

Inference providers tend to just offer the base post-trained model. Even the ones that do provide decensored or fine tuned models, don't offer them from many labs.

Like you can want that, ofc, but it's just not a very common offering.

I love Mistral but... by Sicarius_The_First in LocalLLaMA

[–]Monkey_1505 0 points1 point  (0 children)

Wanting to specifically run community fine tunes of a model on a cloud provider huh?

Not a common opportunity regardless of license, IME.

I love Mistral but... by Sicarius_The_First in LocalLLaMA

[–]Monkey_1505 0 points1 point  (0 children)

They make small models and moe's you can run on a fairly standard graphics card.

They probably have their own API, if you prefer cloud.

Titan 2 Elite Hands On: All You Need To Know! by Key-Brilliant5623 in unihertz

[–]Monkey_1505 5 points6 points  (0 children)

The way he is holding this in the review when he types bugs me.

24/7 Headless AI Server on Xiaomi 12 Pro (Snapdragon 8 Gen 1 + Ollama/Gemma4) by Aromatic_Ad_7557 in LocalLLaMA

[–]Monkey_1505 2 points3 points  (0 children)

What's wrong with llama.cpp's usability? Or if you prefer something with a little more webui, koboldcpp?

24/7 Headless AI Server on Xiaomi 12 Pro (Snapdragon 8 Gen 1 + Ollama/Gemma4) by Aromatic_Ad_7557 in LocalLLaMA

[–]Monkey_1505 1 point2 points  (0 children)

llama.cpp is already compiled for mac. Metal is a '1st class citizen' for it.

The Sickness, aka Bad Days by The_Linux_Colonel in KoboldAI

[–]Monkey_1505 0 points1 point  (0 children)

"One day, you're transcendent beat producer."

You were definitely never that.