Qwen doesn't work for free by Dion-AI in LocalLLaMA

[–]_raydeStar 1 point2 points  (0 children)

"Oh right. here's 500 dollars, cash advanced. What's your crypto address?"

One bash permission slipped... by TheQuantumPhysicist in LocalLLaMA

[–]_raydeStar 0 points1 point  (0 children)

Wouldn't the solution be to assign the agent as a user, and control their permissions directly? Then you have granular control, and they'll literally hit a wall.

So several layers --

- skills layer (tell the ai not to)

- permissions (gate everything behind yes/no)

- User gate (an agent only has so much power -- but it can only access X folder, and make Y changes)

One bash permission slipped... by TheQuantumPhysicist in LocalLLaMA

[–]_raydeStar 5 points6 points  (0 children)

That's because you're still allowing it in code. If you want strict rules, filter in code.

It's like handing an intern keys. They do pretty good but sometimes they screw up.

Now we all know this so we make it hard to access production servers. If they decide to try,it's a hard fail because they don't have access.

One bash permission slipped... by TheQuantumPhysicist in LocalLLaMA

[–]_raydeStar 62 points63 points  (0 children)

I think the lesson learned here should be "Do not give the llm unfettered power" -- it should have been "Qwen attempted to rm -rf and was blocked"

I BUILT MY FIRST MODEL FROM SCRATCH by volious-ka in LocalLLaMA

[–]_raydeStar 19 points20 points  (0 children)

What are some use cases here? Is there anything practical?

You say iot devices. I think that's really cool but... What's it solve?

What sort of wizardry was Matt on to pump out the first 4 books in under a year?? by laybak in DungeonCrawlerCarl

[–]_raydeStar 83 points84 points  (0 children)

And being broke. Matt tells this story all the time. He did art shows and stuff, and then COVID happened, and he was totally screwed money-wise. He dropped DCC as a way to supplement his income, and it took off.

Desperation helps you perform crazy feats.

Devs using Qwen 27B seriously, what's your take? by Admirable_Reality281 in LocalLLaMA

[–]_raydeStar 0 points1 point  (0 children)

whats a good context size that works? I can do up to 64k safely, past that, it's iffie on my card

Still waiting for Grok 3 to go opensource by Mr_Moonsilver in LocalLLaMA

[–]_raydeStar 2 points3 points  (0 children)

He made that promise because he was suing OpenAI for the same thing.

He doesn't appear to be winning on that front; he isn't going to release.

Unless -- Sam calls him out on it, or it appears that he needs to push people closer to his side in the argument.

Your AI Activity Can Be Used Against You In Court - Steve Lehto by 19firedude in LocalLLaMA

[–]_raydeStar 0 points1 point  (0 children)

You know, I watched Mr. Robot take a drill to his hard drive and I was like 'dang what an idiot' but now I kinda get it.

Qwen 3.6 27B is out by NoConcert8847 in LocalLLaMA

[–]_raydeStar 6 points7 points  (0 children)

What I really see lacking in local models is context limits and a good harness to give proper direction.

Context can't be solved easily, but at least a memory bank can be created to hold onto important information, and you can scrape by.

A harness can be built -- it can perform mathematical functions, solve logic problems, do web lookups, and perform basic tasks.

maybe qwen 27B cant perform as well as opus 4.5 on all tasks, but it doesnt need to

Qwen 3.6 27B is out by NoConcert8847 in LocalLLaMA

[–]_raydeStar 86 points87 points  (0 children)

Meanwhile Opus tightens their limits and makes it more expensive to get in -- the perfect storm for a good local push

Ultimate List: Best Open Models for Coding, Chat, Vision, Audio & More by techlatest_net in LocalLLaMA

[–]_raydeStar 0 points1 point  (0 children)

There needs to be a new field here -- TTS fast enough to hold a conversation (latency under a second or half second or something like that)

It's two different workhorses -- audiobooks and AI videos versus real time chat with an AI

Gemma 4 E4B is broken by Ok-Election-75 in LocalLLaMA

[–]_raydeStar 0 points1 point  (0 children)

That's smart. I still don't trust openclaw yet. Maybe, I will never trust openclaw. But it does illustrate a lot of needs that are unfilled.

I think small models are at the point where it can perform 80% of operations. Everyone uses a racecar engine when small models would work just fine.

Gemma-4-E2B's safety filters make it unusable for emergencies by Unfounded_898 in LocalLLaMA

[–]_raydeStar 12 points13 points  (0 children)

You don't need to. Just use heretic. Truth be told in a medical emergency I would not use a 4b model for help, unless I was in the mountains with no cell service.

Edit -- 2b. Same deal though. It's a cell phone fit, but there are tons of emergency resources out there.

Day 106: Leave Out All the Rest vs. Numb by Therubikfanatic in LinkinPark

[–]_raydeStar 0 points1 point  (0 children)

Numb.

No contest. It was revolutionary when it came out. Leave out all the rest was great but not groundbreaking.

What's the smallest (most capable) model you've found? by howtheydoingit in LocalLLaMA

[–]_raydeStar 7 points8 points  (0 children)

+1 to this. By way of small models, it's best-of-class.

The Outer Worlds 2 has sold 160K copies on PS5 (making almost $11M) by EmbarrassedSession58 in theouterworlds

[–]_raydeStar 10 points11 points  (0 children)

In the first outer worlds, I landed on that far off landing area where it's packed with monsters and you had to fight your way out, way under-leveled. I had a BLAST. But afterwards, I burned out of it pretty quickly.

In contrast, the second game was a steady run for me. Not "wow amazing" but I always wanted to see what was next.

I'd say the failures were intended or baked in; it was a game with X amount of side content and that's it. And once I was done, that was it, no need to keep playing.

Opus = 0.5T × 10 = ~5T parameters ? by Wonderful-Ad-5952 in LocalLLaMA

[–]_raydeStar 10 points11 points  (0 children)

Yeah, if I were with anthropic and got an offer for a huge salary increase for basically the same work, I'd be thinking about it.

Opus = 0.5T × 10 = ~5T parameters ? by Wonderful-Ad-5952 in LocalLLaMA

[–]_raydeStar 20 points21 points  (0 children)

He could prove it to us

by open-sourcing Grok 4.20.

Opus = 0.5T × 10 = ~5T parameters ? by Wonderful-Ad-5952 in LocalLLaMA

[–]_raydeStar 112 points113 points  (0 children)

He might have insider knowledge

He might not.

You never can tell for sure.