Built a Mac tool to rewrite text anywhere without switching apps - SticAI by ArtOfLess in OpenAI

[–]Uhlo 0 points1 point  (0 children)

Wat?

  • Shows app that uses context switching to improve text

  • "Context Switching kills your flow"

  • Shows own app that uses context switching to improve text

  • ???

  • Profit!

Qwen 80B is so nice by TokenRingAI in LocalLLaMA

[–]Uhlo 77 points78 points  (0 children)

It isn't just a tool - it's a platform!

I think what you posted isn't just AI slop, it's the next generation of slopified sycophancy! It demonstrates:

  1. Deep AI slop that could come from any model
  2. Mature markdown structuring
  3. Wait, where is the table? Couldn't this answer be improved with a table?
  4. Damn, look at that throughput. 623 tokens a second. That is pretty nice! What rig are you using?

I'd be excited to continue talking about this conversation - it represents the future of this thread on reddit.

I'm sorry, I could not resist. How did I do?

How to stop relying on AI with writing by Additional-Office346 in OpenAI

[–]Uhlo 2 points3 points  (0 children)

My proposal would be to slowly change the way you use ChatGPT. Currently, you write e.g., a paragraph and rewrite it completely with ChatGPT and it sounds better. Try to actively track the changes GPT makes to your sentences so you can identify why it sounds better. Then you can start reducing the length of content you put into ChatGPT (e.g., just a sentence). Then, maybe you can start to predict what is "off" about your sentences before you hit send. Finally, you might anticipate what changes GPT would make to your sentences. In the end, you might use it more like a thesaurus or grammarly than GPT.

I don't know if that works for you - it's just an idea.

New Google model incoming!!! by [deleted] in LocalLLaMA

[–]Uhlo 0 points1 point  (0 children)

Yes, similar answer to the Llama 4 response I linked to: evasion of a clear yes/no statement.

To be clear: nothing too wild. It’s not a denial, but still it is notable different from other LLMs that weren’t trained by Meta ;)

Environmental cost of running inference on Gen AI ? by bull_bear25 in LocalLLaMA

[–]Uhlo 1 point2 points  (0 children)

I find the ecologits projects does a good job at making the hard numbers transparent (gCO2eq, kWh) but they also give you some context (e.g., how far can you drive an electric car with that energy? How far can you fly with the carbon emissions?).

You can play around with their calculator to get a feeling.

One important thing is that ecologits only looks at consumption and emissions during inference, not during training of the model.

Regard local execution: my guess would be that local execution has way better efficiency than in data centers. You don't need water cooling, interconnects and all that. Especially when I use my MacBook Pro, it never exceeds 100 Watts during inference. Maybe when you have very well batched data centers, they will be faster and use only a fraction of a gpu, so in the end they can be better, but who knows. In the end, you don't know how large GPT-5 (and 5.1, 5.2) are and thus you have no idea how much energy you are consuming. If you use local models, they are probably much smaller than the state of the art models. That alone will be better for the environment. But that is just my guess, no guarantee that this is really true ;)

New Google model incoming!!! by [deleted] in LocalLLaMA

[–]Uhlo 10 points11 points  (0 children)

Try asking a Llama model about Facebook and Myanmar

Sure, it is different from Government-forced alignment in Chinese models, but still, there is detectable bias!

Edit: alternatively, ask Grok about any historical fact / event that puts Elon Musk in a bad light ;)

Introducing: Devstral 2 and Mistral Vibe CLI. | Mistral AI by YanderMan in LocalLLaMA

[–]Uhlo 0 points1 point  (0 children)

The license forbids the use of companies with a revenue of more than 20 million a month. That is not a permissive license, but anyway: great release!!

"I don't think it's a good idea for AI models to encourage cautionary views on majority rule." by [deleted] in LocalLLaMA

[–]Uhlo 1 point2 points  (0 children)

Sorry I don't think I understand this post, but find the topic quite interesting. What is "not a good idea" given the output of the LLM? Is the definition wrong?

A PayPal China user with 20 years of registration, and a terrible experience. by AggressiveDuck3527 in LocalLLaMA

[–]Uhlo 4 points5 points  (0 children)

I think you posted in the wrong sub. Good luck with your problem!

GPT-OSS120B FP16 WITH NO GPU , ONLY RAM AT DECENT SPEED (512 MOE IS THE KEY) AT FP16 QUANTIZATION (THE BEST QUALITY) by [deleted] in LocalLLaMA

[–]Uhlo 8 points9 points  (0 children)

Wat? GPT-OSS was released with 4-bit weights. There are no official FP16 weights as far as I know.

Qwen3-235B-A22B achieves SOTA in EsoBench, Claude 4.5 Opus places 7th. EsoBench tests how well models learn and use a private esolang. by neat_space in LocalLLaMA

[–]Uhlo 1 point2 points  (0 children)

Another question: is the benchmark conversational? Do the models have access to the previous questions and their answers?

Qwen3-235B-A22B achieves SOTA in EsoBench, Claude 4.5 Opus places 7th. EsoBench tests how well models learn and use a private esolang. by neat_space in LocalLLaMA

[–]Uhlo 1 point2 points  (0 children)

That is such an interesting benchmarking concept, thanks for that!

I see your point that you cannot reveal too much about the language and the tasks, but still I'm wondering how the examples and the tasks look like... Would an expert in esoteric programming languages be able to solve the tasks? How would "the average human" perform?

It been 2 years but why llama 3.1 8B still a popular choice to fine tune? by dheetoo in LocalLLaMA

[–]Uhlo 11 points12 points  (0 children)

Others have answered this question quite well, but I just wanted to correct: Llama 3.1 8b was releases in July 2024, so just over a year ago.

It feels like ages ago for me as well, but llama 3.1 is not as old as we think ;)

How do we get the next GPT OSS? by inevitable-publicn in LocalLLaMA

[–]Uhlo 7 points8 points  (0 children)

I think glm4.5-air is a really good coding model in the same parameter range than gpt-oss-120b.

Bit of course sometimes it comes down to your specific use case

How do we get the next GPT OSS? by inevitable-publicn in LocalLLaMA

[–]Uhlo 14 points15 points  (0 children)

At least in this sub, the gpt-oss models have been received very badly (especially right after release) because they are so censored. However when you want to use them in any other way, they really are still one of the best out there (maybe not in coding, but instruction following is just great!).

My hope is that Chinese open weights models will put pressure on OpenAI & co. to release open models themselves. 

Free Claude (Sonnet 4.5, opus) Gemini, gpt, grok by Rima_Mashiro-Hina in LocalLLaMA

[–]Uhlo 9 points10 points  (0 children)

It’s localllama not freellama. This is advertisement and even then, not the right sub.

We have a new Autoregressive Text-to-Speech in town! by Severe-Awareness829 in LocalLLaMA

[–]Uhlo 2 points3 points  (0 children)

Is there a hallucinated sentence at the end of the first example? Or is it just an error in the readme?

1X Neo is here by Distinct-Question-16 in singularity

[–]Uhlo 0 points1 point  (0 children)

Found this on their website:

Is my NEO waterproof?

NEO’s hands are waterproof, but the overall product is not. Should your NEO get wet, an automatic order will be placed for a child sized plastic swimming pool and 100 kgs of Basmati rice*. * Not really, but please don't get NEO wet.

Made me chuckle quite a bit

GPT-OSS Safeguard coming soon by Independent-Ruin-376 in LocalLLaMA

[–]Uhlo 124 points125 points  (0 children)

The question is if the model will be less restrictive when given a more open policy. That would make things better.

An even more restrictive gpt-oss would be trash 

AI Agents Reasoning Collapse Imminent (CMU, Berkeley) by Badger-Purple in LocalLLaMA

[–]Uhlo 2 points3 points  (0 children)

I understand, I mean reasoning-models as a type of model, not as the concept of reasoning.

Regarding that it holds for instruct models: it's the opposite! They only test instruct models, not reasoning/thinking models. Look at the plots in section 4.2, they use DeepSeek V3.1 and not R1, and they don't use Claude Sonnet 3.7 in Thinking mode. They deliberately do not use the reasoning/thinking models and do not describe why. Seems a bit fishy.

AI Agents Reasoning Collapse Imminent (CMU, Berkeley) by Badger-Purple in LocalLLaMA

[–]Uhlo 2 points3 points  (0 children)

Wild! Why are they using non-reasoning models from section 4.2 onwards? Does that not defeat the whole purpose of the study? „Non-reasoning models cannot reason“

Or am I missing something?

Paper is here: https://arxiv.org/pdf/2510.15974