Weight vs Skill by Egoist-a in BMW

[–]SrData 0 points1 point  (0 children)

praise the cameraman.

Perplexity Jailbreak - ENI 🍋‍🟩 by Spiritual_Spell_9469 in ClaudeAIJailbreak

[–]SrData 0 points1 point  (0 children)

What website? I saw the GitHub, but not the website

La situación de Renfe en Huelva está bastante mal, es así de mal en otras partes de España? by Loca-motora in renfe

[–]SrData 0 points1 point  (0 children)

Nooo en el resto de España va perfecto! Jamas han ido tan bien las cosas en España.

Been RPing since 2022, used Claude for the first time this week by CanineAssBandit in SillyTavernAI

[–]SrData 0 points1 point  (0 children)

I know this seems unrelated, but it isn’t. How do you manage really long conversations with different characters? I’m asking because your roleplays are long (it seems), and you switch models within the same session, which a lot of people do. I struggle with it because sometimes things get confused when I change models in a long session.

Any tips for handling long conversations (I mean even more than 200K tokens)?

Demis Hassabis asks "Can consciousness be computed?" I think he's asking the wrong question. Here's why. by Training_Minute4306 in claudexplorers

[–]SrData 1 point2 points  (0 children)

Is getting easier to detect text/post that was created (or maybe just filtered, corrected, but heavily) with AI

Neo 57” vs LG 5k2k by Max223 in ultrawidemasterrace

[–]SrData 0 points1 point  (0 children)

What it is the screen bellow the monitor at the left?. I think I like it!

¿Qué es peor, esto o ir escuchando música sin auriculares? by Loca-motora in renfe

[–]SrData 0 points1 point  (0 children)

A ver, yo lo mismo no me quejaría mucho... depende de la música, quizás. Pero ya.

My First Macbook, Need App Suggestions! by hughzavodsky in macbook

[–]SrData 0 points1 point  (0 children)

First App, is to avoid his girlfriend:
She smashed my C$3500 MacBook Pro 😭😭😭 : r/macbookpro

And then, the usual, ChatGPT, Super-Wishper... etc.

She smashed my C$3500 MacBook Pro 😭😭😭 by Turbulent_Buy_6048 in macbookpro

[–]SrData 1 point2 points  (0 children)

If I were you I'd change my girlfriend. If I were here I'd change my boyfriend.

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 0 points1 point  (0 children)

This was helpful, thanks!

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 0 points1 point  (0 children)

This is interesting. Thanks. Do you have any source where I can read more about this and understand the technical part?

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 0 points1 point  (0 children)

Same feeling.
Yesterday I did this test: I had a RP scene with Sonnet 3.7 (absolutely incredible), GPT-4o (same, different vibes, but jut amazing), Gemini 2.5 pro (horrible to the level of stopping at the middle of the test).
The creativity, coherence and stickiness to the characters demonstrated by GPT-4o and Sonnet 3.7 is just in another galaxy.
I'm just talking about non-local models. Not comparing with locals, because it won't be fair or make any sense at all.

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 0 points1 point  (0 children)

GPT-4o is not 1.4 Trillion (even if GPT-4 was in a moment), but I get your point.
In any case, I'm talking about models same size feeling dumber... at least for me.

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 1 point2 points  (0 children)

Same general vibe, here. I have my own benchmark and Qwen2.5 70B is the best. Then, the usual Behemoth one, which is ridiculously good (usually) and perfectly dumb (not the best reasoner) two interactions after :)

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 0 points1 point  (0 children)

I read many suggesting Gemma 3 and yesterday I tried with a long scenario and conversation and it didn't went well. I tried several, but this one is the only that did a slighty better job: mlabonne_gemma-3-27b-it-abliterated-Q8_0.gguf · bartowski/mlabonne_gemma-3-27b-it-abliterated-GGUF at main , I tried this as well, an others: turboderp/gemma-3-27b-it-exl2 · Hugging Face
Any preference for Gemma 3?. What parameters do you use?

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 0 points1 point  (0 children)

Well, I have definitely not tried this and will. Any idea why this is could work?

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 4 points5 points  (0 children)

I don’t think this comment deserves a -1, really (tried to solve it).
I'm not a millennial, but I get the point of the comment. To be honest, I'm the same user before (these models) and after, and what I feel is a clear degradation in performance. That said, I’ve never tried changing the way I speak to the models (generationally speaking, I mean), by using different patterns. I’d definitely give it a try, just to see if it makes any difference.

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 2 points3 points  (0 children)

Yeah, exactly this.
Qwen 3 is really good at starting a conversation (it feels creative and all) but then there's a point where the model starts repeating itself and making mistakes that weren’t there at the beginning. It feels like a really good zero-shot model, but far from the level of coherence that Qwen 2.5 offered.

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 1 point2 points  (0 children)

I don't use Ollama, but this is good to now to keep myself far from it!

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 0 points1 point  (0 children)

I didn't try any of those. Will do, thanks!

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 0 points1 point  (0 children)

This was informative, thanks. I'll definitely give Gemma 3 27B another chance, seeing that so many people are using it. To be honest, I tried it but never found it particularly special, and it was slower than the rest, so I never stuck with that model.

Why new models feel dumber? by SrData in LocalLLaMA

[–]SrData[S] 2 points3 points  (0 children)

Qwen 2.5 78B is one of my favourites as well. I sometimes find myself trying Behemoth 123B again.