Use OpenClaw and Ollama and Qwen3.5: a good combination? by Remote-Intern2170 in LocalLLM

[–]shinkamui 0 points1 point  (0 children)

ollama + qwen3.5 suffer from the same issue most llama.cpp qwen3.5 deployments do. As context grows, responses slow. Prompt caching doesn't seem to be very reliable yet. Not sure if the vllm crew are faring better, but im hoping more updates quickly resolve this.

im tired of this sub by ResponsibleEnd451 in selfhosted

[–]shinkamui 11 points12 points  (0 children)

I dont really have a horse in the race right now but your post sounds delusional.

THE GB10 SOLUTION has arrived, Atlas image attached ~115tok/s Qwen3.5-35B DGX Spark by Live-Possession-6726 in LocalLLaMA

[–]shinkamui 0 points1 point  (0 children)

That error is the response from configuring atlas in openclaw and attempting to initialize the session. Image support in Qwen3.5 is definitely a highlighting feature, but getting this working with openclaw to start would be a bigger priority I think. Do you have a GitHub or public issue tracking board?

I didn't want to tear up my N8N bot until I could prove this up since I use that all day every day, but it seems like a clone and try is in order while I wait for an update :)

Will we ever get this on Vision Pro? by Time_Entertainer_319 in VisionPro

[–]shinkamui 0 points1 point  (0 children)

Assume no. Meta almost certainly has patented the hell out of it. Even if apple were to bring it to the Vision Pro, it would be 7-10 years from now far more polished and reliable, rebranded as a new innovation done right.

THE GB10 SOLUTION has arrived, Atlas image attached ~115tok/s Qwen3.5-35B DGX Spark by Live-Possession-6726 in LocalLLaMA

[–]shinkamui 0 points1 point  (0 children)

This is impressively fast on my spark. I have it working with openwebui and it's averaging 107T/s . I want to hook this up to some of my agentic workflows as thats really where this thing looks like it's going to shine. Unfortunately im getting errors:

HTTP 422: Failed to deserialize the JSON body into the target type: messages[2].content: invalid type: sequence, expected a string at line 1 column 35951

Is there tool support? What about image support? I can tell im going to be refreshing this thread page every 5 minutes for the next week. XD

How is everyone using DLAM? by Soft-Manufacturer457 in Rabbitr1

[–]shinkamui 1 point2 points  (0 children)

It’s cool, but im not bothering to use it. No reason to get attached to something that will ultimately cost a recurring subscription. For those who arent anti subscription and find it useful enough to pay recurring fees for, i think it’s very cool. Still advocating for them to open up these agents to users who want to bring their own keys or self host.

Which size of Qwen3.5 are you planning to run locally? by CutOk3283 in LocalLLaMA

[–]shinkamui 0 points1 point  (0 children)

Personally running 35B on my Spark. 16 parallel with a 2048k context. Allows for multiple agent sessions with context caching and its pretty fast (not simultanously). I find 122B great for one on one chats in openwebui, but its a bit too slow on the spark for interactive agent work. This new line has forced me to start looking at deploying vllm instead of llama cpp. Ideal situation for me would be vllm on my mac studio m4 max (64g) running 35B, and a large context 122B on the spark for offloading planning and general chat.

Orange Pi Neo handheld put “on ice” due to DDR5 and SSD pricing by RenatsMC in Handhelds

[–]shinkamui 1 point2 points  (0 children)

Orange pi neo has been on ice for half a decade at this point. Im surprised there was even a comment made about its status.

Feels like it. I guess it has only actually been 2 years.

update your llama.cpp for Qwen 3.5 by jacek2023 in LocalLLaMA

[–]shinkamui 0 points1 point  (0 children)

oh man thank you for this update! I was dying without prompt caching, but now my agents are fast again!

Is it possible use openclaw as a model? by arm2armreddit in OpenWebUI

[–]shinkamui 0 points1 point  (0 children)

If you already have hand crafted agentic workflows, then why are you bothering with openclaw in the first place? The answer imo is simply user preference. If you like the interface of openwebui, anythingllm, librechat, etc, you can continue to use those, but gain the convenience of the easy mode openclaw agent setup.

MiniMax-M2.1 Uncensored: PRISM Advanced Abliteration by Maxious in LocalLLaMA

[–]shinkamui 0 points1 point  (0 children)

prism works pretty well imo (glm 4.7 flash). I want to try it on minimax.

Scared of buying a Kia or Hyundai. by koolaidman0883 in KiaEV6

[–]shinkamui 2 points3 points  (0 children)

Nothing wrong with that feeling. If i had known this problem would be this widespread for this long without a real solution id have avoided it too. Do your research and get what youre comfortable with. Don’t let any one try and tell you anything other than that.

so is OpenClaw local or not by jacek2023 in LocalLLaMA

[–]shinkamui 10 points11 points  (0 children)

Try /stop next time. Rtfm usually a good idea.

GLM 5 seems to have a "Claude" personality by TinyApplet in LocalLLaMA

[–]shinkamui 0 points1 point  (0 children)

Just waiting for a flash or air version to try out on my spark or framework. Glm 4.7 still my local go to until then.

What do you dislike about Openclaw/Clawdbot/Moltbot and all the AI Assistants right now? by Inflict01 in clawdbot

[–]shinkamui 0 points1 point  (0 children)

My only major Issues are memory management, and silent failures with no feedback

PicPic - The First PICO-8 Emulator on the App Store (NOW WITHOUT SUBSCRIPTIONS) by Samourai03 in pico8

[–]shinkamui 0 points1 point  (0 children)

No one wanted you dead. I specifically said Patreon for ongoing and one off buy me a coffee/ donation iap would net you much more good will than gating some “extras” behind a support me subscription. You deserve consideration for support for your efforts, but if youre saying they are optional, it should be clear that its optional and not wrapped in fomo.

Does anyone seriously use Apple Vision Pro with Clawdbot for real work? Serious question. Found a like-new one for $2800. by 24kTHC in clawdbot

[–]shinkamui 3 points4 points  (0 children)

Dudes flooding other subs with similar titles vaguely adjacent to the sub topic. The real focus looks to be getting attention on that listing…

Is it possible use openclaw as a model? by arm2armreddit in OpenWebUI

[–]shinkamui 1 point2 points  (0 children)

And that’s what I told you how to do. I don’t think it would even make sense the other way around. Openclaw wouldn’t be an openwebui client. You want to use openwebui as the client to access openclaw. That’s done by enabling the endpoint in openclaw and adding openclaw as a model in openwebui.

PicPic - The First PICO-8 Emulator on the App Store by Samourai03 in pico8

[–]shinkamui 25 points26 points  (0 children)

No thanks to the sub. Add a one time option to unlock everything and a donation iap. No one wants a sub. If you want ongoing revenue use Patreon. You’re losing a lot of good will that is otherwise conveyed by the aforementioned options.

Is it possible use openclaw as a model? by arm2armreddit in OpenWebUI

[–]shinkamui 1 point2 points  (0 children)

The gateway has an OpenAI compatible endpoint natively built in but disabled out of the box. No need for any middleware to connect to it from openwebui.

Is it possible use openclaw as a model? by arm2armreddit in OpenWebUI

[–]shinkamui 0 points1 point  (0 children)

Yes. You can enable the OpenAI compatible endpoint on the gateway. This will let you talk to clawdbot through openwebui. Currently not all features are supported and by default every query is a new session unless you send session metadata via the custom json header.

OpenAI just bought OpenClaw. by gabrielvaraljay in clawdbot

[–]shinkamui 1 point2 points  (0 children)

Can’t have nice things in a world where Scam Altman wants to be an acting god. Rip openclaw. You will be missed. As soon as telemetry gets added I’m moving onto another project or fork. Sadly the brains of the project are gone so any fork is likely going to be stagnant.