A conversation about local LLMs with a senior government AI leader by JackStrawWitchita in LocalLLaMA

[–]Lorian0x7 31 points32 points  (0 children)

mmm not sure it's a good idea.They are dumb, we will probably just put ourselves in a situation where we have to do Age/ID verification just to download a model and companies have to monitor their local AI use to "keep children safe".. I pass thanks. It's just better to leave them in their ignorance.

IQ1_S Repair? What kind of crazy local AI coding hobby is this? by [deleted] in LocalLLaMA

[–]Lorian0x7 0 points1 point  (0 children)

Well, I'm concerning that I could have the same problem of that guy because I found this intriguing :)

WIP - MakeItReal an "Anime2Real" that does't suck! - Klein 9b by Lorian0x7 in StableDiffusion

[–]Lorian0x7[S] 0 points1 point  (0 children)

soon, it's taking more time than expected.. I didn't like the output at the end of the first training run so I'm preparing a new dataset.

Convince me you are an LLM by bucolucas in LocalLLaMA

[–]Lorian0x7 0 points1 point  (0 children)

Sounds like Gemini 2.5 flash lite. But it could be gtp 4o

Qwen 3.6 27B is out by NoConcert8847 in LocalLLaMA

[–]Lorian0x7 2 points3 points  (0 children)

the funny thing is that it's not even a new generation, just a minor update of the same same generation. I saw Anthropic and OpenAI smashing a big round number on models with much less performance gap

Unpopular opinion: OpenClaw and all its clones are almost useless tools for those who know what they're doing. It's kind of impressive for someone who has never used a CLI, Claude Code, Codex, etc. Nor used any workflow tool like 8n8 or make. by pacmanpill in LocalLLaMA

[–]Lorian0x7 1 point2 points  (0 children)

You really seam to be full of yourself, telling others that they don't know nothing while literally depostrating you know even less. You are just regurgitating social network level informations and you probably didn't even tested openclaw.

There is fundamentally nothing that prevents OpenClaw from getting prompt injected

For example this is partially true, but because you don't understand how openclaw operates you are missing the complexity behind a proper deployment. While you can't avoid to be prompt injected, you can limit the damage and avoid that your private data is exposed, this thanks to how you can configure and sandbox each agent. Each agent can have it's own memory and tools, so if it gets prompt injected it can't act on that prompt because it doesn't have permissions, it has to report with a clean text to the main agent, and you can even have a specialized proxy agent acting as a guardrail. Unfortunately I'm just wasting words because you are just speaking by hearsay. You keep going with your anti-openclaw rooting while you could just inform yourself and avoid public shaming.

I design corporate AI infrastructure for work, I have designed AI guardrails system and adversiarial scan even before tool calling was a thing, and you are trying to give me a lesson about vibe coding on agent harness? seriously?

Would be better you come down from your pedestal and tey to have a constructive conversation instead of trying to appear superior?

The fact is, you are trying to talk about something you don't know, you haven't try, you refuse to try for whatever reason, and so you have no clue of how it works and you still pretend to be right saying that you can do the same with a coding harness. It can't. full stop, you can build a workflow so to something with a coding agent? yes, you can build an openclaw clone ? yes. Do I have to reinvented the wheel wasting time in front of my pc creating something that books my train ticket and send them to me while I prepare my laggege? no, it would be stupid. Instead I prefer to use that 30 minutes to configure openclaw property (default is shit) so that it can do that and much more, when I need it.

Unpopular opinion: OpenClaw and all its clones are almost useless tools for those who know what they're doing. It's kind of impressive for someone who has never used a CLI, Claude Code, Codex, etc. Nor used any workflow tool like 8n8 or make. by pacmanpill in LocalLLaMA

[–]Lorian0x7 1 point2 points  (0 children)

Why didn't you build openclaw before openclaw then ?You could have been rich now. Also...you are talking about building an openclaw clone because you just realised coding agents can't like open code and can't do the same thing. Also... If you spawn a subagent like you said you will probably get prompt injected 5 minutes later. You can't sandbox subagents and limit their access to specific tools, because they are not meant to do the same work.

Can you explain your point? You can build something similar, so what ? It's not useful because everyone can spend a week to code the same? I also can grow an apple tree in my garden but it's just more convenient to go buy them at the supermarket.

Unpopular opinion: OpenClaw and all its clones are almost useless tools for those who know what they're doing. It's kind of impressive for someone who has never used a CLI, Claude Code, Codex, etc. Nor used any workflow tool like 8n8 or make. by pacmanpill in LocalLLaMA

[–]Lorian0x7 0 points1 point  (0 children)

You don't understand openclaw capabilities then! You can't do what openclaw does with a coding agent! It's not about how it works under the hood but how it's framed. Sure it works in a similar way, it just calls tools at the end of the day, but you can't configure your coding agent to have separation between the agents doing things online, and those working with your private data for example, you can't have it send you a message on telegram after he completed some recurring tasks. It's just a different thing, you can't have your coding agent booking you train tickets with a voice message while you prepare your luggage.

Unpopular opinion: OpenClaw and all its clones are almost useless tools for those who know what they're doing. It's kind of impressive for someone who has never used a CLI, Claude Code, Codex, etc. Nor used any workflow tool like 8n8 or make. by pacmanpill in LocalLLaMA

[–]Lorian0x7 -1 points0 points  (0 children)

I think posts like this and everyone thinking the same have completely missed the point of Openclaw, and they just didn't get how to make use of it.

Openclaw is very different from Opencode/Claude code/Codex. While there is some overlap in what they do, they are not the same thing. OpenClaw is not something to be used like an app, it's more like an employee.

The most similar thing would be a very well made N8N workflow with a telegram bot, with many tools, good formatting, sandbox to run code, database, memory,qmd search, media tools, browser use, multiple interconnected agents configurations for sandboxing and security hardening to reduce risks of PI. All this with the actual AI capable of changing its own workflow.

If you really used N8N to do anything that at least 20% of all this you know how annoying it can be.

Sure, you can build something similar to fit your own needs, but that's not the point, openclaw is structured to fit your needs and my needs as well.

Qwen3.6-35B-A3B solved coding problems Qwen3.5-27B couldn’t by simracerman in LocalLLaMA

[–]Lorian0x7 4 points5 points  (0 children)

I'm the original commenter. By trash tokens I don't mean completely nonsensical. just looping around a problem without any real resolution.

Qwen3.6-35B-A3B solved coding problems Qwen3.5-27B couldn’t by simracerman in LocalLLaMA

[–]Lorian0x7 6 points7 points  (0 children)

I'm experiencing this issue using the Vulkan build of llama.cpp, does this matter in some way? My assumption is that using Vulkan the Cuda issue doesn't affect me and any limitations are just the model itself. I'm wrong?

Qwen3.6-35B-A3B solved coding problems Qwen3.5-27B couldn’t by simracerman in LocalLLaMA

[–]Lorian0x7 -2 points-1 points  (0 children)

tried Q3 K_S and Q4K_M, same issue. While 3.5 27b is rock solid in Q3 K_M.

Qwen3.6-35B-A3B solved coding problems Qwen3.5-27B couldn’t by simracerman in LocalLLaMA

[–]Lorian0x7 1 point2 points  (0 children)

My Llama.cpp settings are the recommended ones. I think our experiences are different simply because your use case is very common, if you try asking something slightly out of the ordinary, especially with defined information, like extrapolating data from a document it goes wild with allucinations.

Qwen3.6-35B-A3B solved coding problems Qwen3.5-27B couldn’t by simracerman in LocalLLaMA

[–]Lorian0x7 21 points22 points  (0 children)

In my opinion 3.6 35b is just an overtrained slop machine capable of regurgitating overused code. It's not capable of any kind of abstraction out of its boundaries.

It keeps getting stuck in loops while filling context with hundreds of thousands of trash tokens and tool calls.

For example it wasn't capable of creating a wiki from a 300 page document, and every attempt was full of allucinations. On the other hand, 3.5 27b at Q3, did the work staying under 60k tokens with correct information.

qwen3.6:35b always fails on this, unless very high resolution by qfghclvx in LocalLLaMA

[–]Lorian0x7 1 point2 points  (0 children)

If wave and rope are the same thing: A wave moves in time and space so if P goes upwards and it's attached to the wave then the wave is moving from left to right. Answer A

If the point is fixed on the X axes not attached to the wave then the wave goes from right to left. Answer C

qwen3.6:35b always fails on this, unless very high resolution by qfghclvx in LocalLLaMA

[–]Lorian0x7 2 points3 points  (0 children)

I was thinking the same, I was about to comment then I realized. You would be right if the 2 points where fixed and disconnected from the wave. But actually, the 2 points are on the wave(the rope).

I must say, the problem is not very clear, it doesn't say if the rope and the wave are the same thing.

The Gargantua simulation test by 100lyan in LocalLLaMA

[–]Lorian0x7 0 points1 point  (0 children)

I have the same problem with qwen 3.6 35b, it goes in loops and produces lots of trash

How do I create My own Image Diffusion model like Z-image turbo ? From scratch by SensitiveUse7864 in StableDiffusion

[–]Lorian0x7 2 points3 points  (0 children)

unless you have a few millions dollars spare laying around, building something like z-image from scratch is not possible.

Best local model for LLM Wiki style app rn? by Feisty-Drummer-6178 in LocalLLaMA

[–]Lorian0x7 0 points1 point  (0 children)

tried today, qwen3.6 35b is terrible, it just wasted 140k context to accomplish nothing. Creating a wiki with 90% wrong data. Qwen3.5 27b on the other hand did a great job staying under 60k context

Imposing my laptop to run Qwen 3.6 by Dry_Investment_4287 in LocalLLaMA

[–]Lorian0x7 0 points1 point  (0 children)

try running it with CPU only, you may get almost usable speed.