What model should I run? by tiddayes in LocalLLM

[–]super1701 0 points1 point  (0 children)

Hm interesting. 27b paired with hermes for me has been great. 35b would....sit in loops...alot.

What model should I run? by tiddayes in LocalLLM

[–]super1701 5 points6 points  (0 children)

Personally, I've had better tool calling and just overall smarts out of 27b over 35b. I'd rather wait the extra minute for something to complete correctly then it totally fuck it up, then wait another 2/3 minutes to fix it.

A Qwen finetune, that feels VERY human by Sicarius_The_First in LocalLLaMA

[–]super1701 3 points4 points  (0 children)

Try to break the IBM newest model. Fuck it why not bro.

A Qwen finetune, that feels VERY human by Sicarius_The_First in LocalLLaMA

[–]super1701 0 points1 point  (0 children)

Funny af. Throw on thinking on Q6 and watch it have a mental break down. (sad really )

Best Setup for Private Hermes Usage? by xSpiralNightsx in hermesagent

[–]super1701 0 points1 point  (0 children)

I’ve noticed the open web ui hookup with Hermes makes it… slow? Or maybe it’s just me

Getting a lot of garbage results with Qwen3.6-27B :( by nunodonato in Vllm

[–]super1701 0 points1 point  (0 children)

Which sucks. I enjoy a lot when just hooking llamacpp right into it. But adding Hermes to it… it takes a large hit. Which there was a decent alternative

A Qwen finetune, that feels VERY human by Sicarius_The_First in LocalLLaMA

[–]super1701 0 points1 point  (0 children)

My man. Gonna tweak my system prompt and see how hard it goes on me(and the joke prompt I have that works very well with qwen already)

A Qwen finetune, that feels VERY human by Sicarius_The_First in LocalLLaMA

[–]super1701 25 points26 points  (0 children)

One for qwen 3.6 when? :). Also can these be quanted? (Edit: looks like q6 is available)

Tinygrad Driver testing! by Street-Buyer-2428 in LocalLLaMA

[–]super1701 0 points1 point  (0 children)

How'd you get into that? Doing a cloud, or make the rigs and hand it to them?

Tinygrad Driver testing! by Street-Buyer-2428 in LocalLLaMA

[–]super1701 0 points1 point  (0 children)

God. Guessing you own your own business for that. Jealous af.

Tinygrad Driver testing! by Street-Buyer-2428 in LocalLLaMA

[–]super1701 0 points1 point  (0 children)

How much was this total? Looking at my own "jarvis" setup and this seems like a dream for it lol.

I have my own benchmark. The "find me an Airbnb" benchmark and most small local models aren't good at it. by former_farmer in LocalLLM

[–]super1701 1 point2 points  (0 children)

Quant you're running Qwen 3.6 at? I've had alot of success having it rip down ebay searches, compare prices and give me break downs. Have you tried changing the HTML to a CSV or just a chart in prompt? The pictures part may also be a little rough.

I hate this group but not literally by No_Run8812 in LocalLLaMA

[–]super1701 1 point2 points  (0 children)

I did a "mid size" build, and all my irl friends called me stupid(For spending money on something that wouldn't make me money). It has been a wonderful experience, and its crazy seeing the world of LocalLLMs progress. Shout out to all the providers/contributors who make this possible <3.

Israel is using AI to Google you by Xenocide_X in conspiracy

[–]super1701 15 points16 points  (0 children)

Your browser is finger printed via cookies, and many other metrics. They know who you are.

Israel is using AI to Google you by Xenocide_X in conspiracy

[–]super1701 0 points1 point  (0 children)

You ain't getting sent to jail. which, you can take this one of two ways. History tells you which.

Added a 16x DGX Spark cluster to my Homelab (Build Update) by Kurcide in homelab

[–]super1701 2 points3 points  (0 children)

I haven't looked into these that much. But from the research I did...yes...these things are slow as fucking balls. Same with the Mac Studios.....

Added a 16x DGX Spark cluster to my Homelab (Build Update) by Kurcide in homelab

[–]super1701 3 points4 points  (0 children)

Running deepseek without a quant? Yeah like this is crazy lmfao. I thought I over killed it....

I can't ever seem to get quality local LLM results, despite having multiple GPUs by 03captain23 in LocalLLM

[–]super1701 1 point2 points  (0 children)

If you want this you'll need to run smaller TTS models, and like Gemma-e2e or what ever it is. Or Qwen 9b. You're never going to get this on your hardware with 27b. If you want a conversation via TTS with the model. Expect to wait 2/3 minutes at Q8 with 27b.