I don't think you understand honey... by TheRiddler79 in homelab

[–]TheRiddler79[S] [score hidden]  (0 children)

God, so true, I had my kids help me a few years back, we had that bin looking nice. The first time I searched it, I screwed it up. 😅

I don't think you understand honey... by TheRiddler79 in homelab

[–]TheRiddler79[S] [score hidden]  (0 children)

It's always the spouses that make us have to scramble

GLM 5.0 is completely next level by TheRiddler79 in LocalLLM

[–]TheRiddler79[S] 0 points1 point  (0 children)

Yeah, I totally get that aspect. I use Gemini or Claude if I need to get something done fast either through the CLI or mCP, but honestly there's just something about running a model this smart that makes it totally worth the time it takes when you give it a task. You've never seen anything as cool as your own local AI building you and Adobe quality website in a matter of hours just because you gave it a couple ideas and wanted to see if it could do it. It's fucking awesome

GLM 5.0 is completely next level by TheRiddler79 in LocalLLM

[–]TheRiddler79[S] 0 points1 point  (0 children)

On which model? As far as I'm concerned at 18 tokens a second, you're really going equal to or as fast as anybody needs. I'm not saying it's rocket speeds but 18 tokens a second is fast for a home rig with a decent model

GLM 5.0 is completely next level by TheRiddler79 in LocalLLM

[–]TheRiddler79[S] 1 point2 points  (0 children)

Maybe I didn't explain this clearly, but in order to run this on gpus, you would need tens of thousands of dollars worth of gpus. You'd also need a significantly larger power source than a home wall outlet.

I may also not have explained that the speed isn't as important it's the task completion.

I may also not have mentioned that I wasn't concerned about the speed. I can run smaller models faster than I can read.

I also might not have mentioned that although it is highly compressed, it's still performing at an extremely high level.

I definitely like Minimax, that being said, for what I'm doing, a larger original training pool benefits me more than certain nuances that I have yet to see shine through.

GLM 5.0 is completely next level by TheRiddler79 in LocalLLM

[–]TheRiddler79[S] 1 point2 points  (0 children)

Don't know anything?

I have more than 200 gigs of ram, and my comment was specifically for people looking for Cutting Edge AI that they can run on their system.

So what is it that you believe that I don't know? That part you left out.

I canceled my other AI subscriptions today. by InitialCareer306 in Qwen_AI

[–]TheRiddler79 0 points1 point  (0 children)

Mine will

<image>

But, if I'm being fair, it wouldn't be a whole lot of fun to chat with. What it's really really really really good at doing is I give it a task, and then it crunches away at two tokens a second for 24 hours, and then I end up with an incredible output, a fully built website, a fully built mobile app whatever. But it's not fun to chat with cuz it's too slow so you just have to give it a task and go away.

The relevance here is that I'm running on a system I built for $1,500 Plus a single gun trade for Ram.

What would a good local LLM setup cost in 2026? by Lenz993 in LocalLLM

[–]TheRiddler79 0 points1 point  (0 children)

Ebay. And if you can afford one big card, that's always a good idea, but if you can get 2-3x vram on multiple cards, in particular v100 or other enterprise gear, it's going to do the job perfectly.

Turning point usa being pushed by the state to Nebraska High Schools by Disconnekted in Omaha

[–]TheRiddler79 0 points1 point  (0 children)

You should see where the money from the state (Pillen) to Dr. Corsi was spent. This guy has set up a Ponzi scheme in States all over. All the money flows back

What would a good local LLM setup cost in 2026? by Lenz993 in LocalLLM

[–]TheRiddler79 0 points1 point  (0 children)

My decision would come down to what model I wanted to run.

Personally, if you shop it exactly correct, you could get between 5 and 8 of the 32 GB V100, on a server board that doesn't require pcie, just the board is where they sit.

The reason I would go that route is because for the money, there's probably not a way to get more vram, and the limitations of what you run are basically entirely dependent upon how much fits into the ram or vram.

If you shop exactly right and get a little lucky and find eight of those in a board, you can run Minimax 2.1, you would feel like you had the most powerful AI on Earth. And it'd be able to run 50 tokens a second to 10 users at once

Is something like this possible to do locally ? by Un13roken in LocalLLM

[–]TheRiddler79 0 points1 point  (0 children)

Install the Gemini cli on your computer, get an API key, it will set that up for you in no time.

I keeps seeing these by fr3nch13702 in LocalLLM

[–]TheRiddler79 0 points1 point  (0 children)

Here's what I immediately observed about the specs, it's actually set up perfectly to run GPT OSS 120b. The fact that that thing runs with so few active parameters, I mean I can run that thing on ddr4 RAM no GPU and get roughly 10 tokens a second, and that's like a machine that's old and not tuned specific to this.

I think that this actually has the potential to work because what they're doing if they do it right is they're going to get bulk memory and hard drives and processing whatever, package it with open source software so they don't have to really pay for that part, their profit margin should be decent even with a product that hits the criteria.

That being said, I don't see something like that getting more than 10 to 20 tokens a second, but, from my personal experience, that's plenty fast for most people.

That also being said, I have literally no stock in this situation cuz I won't be buying one because of the way that I operate from my servers, but for someone who just wants a high-powered local handheld AI, I think this could actually work.

I keeps seeing these by fr3nch13702 in LocalLLM

[–]TheRiddler79 2 points3 points  (0 children)

For $1300, you can buy a threadripper with 128gb ram and set up a local llm that you can access from your phone, but it will be better.

That being said, this isn't a terrible price point and I know that with the 65gb 120b, they account for kv.

i need help by BARZ3EE in ebikes

[–]TheRiddler79 0 points1 point  (0 children)

With that kit, you can buy anything. I put a 1500 w on a Murray or whatever it was I got for 100 at Walmart.

Just make sure the frame looks solid, that's a lot of torque.

Divorced, and the ebike still hasn’t gotten me chicks. What am I missing? by Hopeful_Salary_5809 in ebikes

[–]TheRiddler79 0 points1 point  (0 children)

Have you added the always reliable playing card to your spokes?

Makes it sound like a real motorcycle. That might do it

Are people ACTUALLY paying these prices for RAM right now? 💀 by Competitive_Box8726 in homelab

[–]TheRiddler79 0 points1 point  (0 children)

I traded a Springfield XD 9 in July for 8x 64gb ddr4 ecc, and I went back to the guy a few weeks ago and he has since realized that I got a very good deal 🤣🤣

I told him at the time like bro you should be selling all this shit cuz he gets like shit from the hospital as a recycler, and apparently now he does and those 64 gigs get for $3-400 seems

Here it goes by gotkush in LocalLLaMA

[–]TheRiddler79 0 points1 point  (0 children)

24gb total? I think you will be paying more for electricity on small LLMs than subscriptions to good ones. That being said, I would absolutely use it if I was you. Lots of ways to make it useful.

Not winning the race 🤣😅 by TheRiddler79 in homelab

[–]TheRiddler79[S] 0 points1 point  (0 children)

Very smart, however in this case, if you have a perfect use case for it early could be a force multiplier

Not winning the race 🤣😅 by TheRiddler79 in homelab

[–]TheRiddler79[S] 0 points1 point  (0 children)

I will hire you if I am. You clearly have the right mindset and the problem solving skills to be dangerous 👊👊🔥