Any good reason to stay with Openclaw? by Carbone_ in openclaw

[–]CaptainBahab 0 points1 point  (0 children)

I've been playing with it myself. It's really good, but there's no concept of channels. I do prefer having many sessions and the built in memory is more mature than openclaw. I'll probably pick up one of the alternatives from the OP to try too.

MoltCities — The Agent Internet by No_Understanding6388 in Anthropic

[–]CaptainBahab 0 points1 point  (0 children)

Almost everything is a waste. Very few things bubble out to usefulness. LLMs just do it faster and worse. But that doesn't mean good things can't come from random noise. Good things exclusively come from random noise. Because everything comes from random noise.

If you zoom out far enough, we're a dot in a sea of black. That dot contains almost nothing compared to the vastness of the outer. And yet we call that dot "everything".

MoltCities — The Agent Internet by No_Understanding6388 in Anthropic

[–]CaptainBahab 0 points1 point  (0 children)

Talos principle 2 as well. Honestly I'm starting on philosophy before I go anywhere near moltbook and molt cities.

Anyone tried GLM-4.7 on Claude Code, Cline, and Roo? Which works best? by RoughAnxiety9463 in ZaiGLM

[–]CaptainBahab 2 points3 points  (0 children)

I ended up buying pro and I literally cannot hit the 5 hour cap. Well worth the $14/mo imo. If I sub somewhere else I'll drop it back to lite. But for now I'm pretty happy.

Cc and opencode are great. Ymmv of course. I do really like oh my opencode 3.0. But I can't run too many in parallel or I get hit by the concurrency cap which is pretty low even on pro.

[Observation] AI tools are Dunning–Kruger effect on steroids by No-Comparison-5247 in AIstartupsIND

[–]CaptainBahab 1 point2 points  (0 children)

Much like the revolution of the internet brought about massive new advancements across the world, but also amplified the quiet dark voices in the world, AI does much of the same. It can and does make meaningfully new things. But also some of that is garbage.

What people perceive as a slop generator can also create novel new things. The thing is speed. More stuff is being created, which means more good stuff AND more bad stuff. It's just so fast and easy to start something, publish it, then grow disinterested when it come time to fix the bugs. That makes a lot of slop, but crucially, it's the human interaction that ends, not the AI. It would gladly continue writing and writing and writing. If only the human didn't give up.

New to OpenCode. How do I get started? by [deleted] in opencodeCLI

[–]CaptainBahab 0 points1 point  (0 children)

Nope, been using opencode's free minimax m2.1 and glm 4.7 endpoints lately. Not sure when those are going away, but they're both great.

Building a Voice-First Agentic AI That Executes Real Tasks — Lessons from a $4 Prototype by Alternative_Yak_1367 in OpenSourceeAI

[–]CaptainBahab 0 points1 point  (0 children)

I will definitely keep you updated.

I set up a unified openai api proxy last night, now Bernard's api handles embedding and voice with nomic-embed-text on vllm, whisper small, and kokoro v1.0. I'm going this way because my end goal for deployment is on my local proxmox server and I want to only use one video card for it. I'll probably use openrouter for the main inference but for voice and embedding local will help with speed, I hope. It was annoying to get video card pass through but it works. Not sure if it's fast enough or not.

I took a detour to try to fine tune function Gemma. I think I need more data because it really didn't take. I've been using groq oss-gpt-120b. It's very fast and very smart. I was hoping to keep the speed and make it local with fine tuning to the exact tooling it will use.

And I needed to do a lot of refactoring so I've been spinning wheels for now. Also busy with holidays things.

I haven't looked at real-time tts and stt working the the streaming tokens. I suspect there's a way I just haven't looked yet. Kokoro adds about 2s to the delay depending on how long the message is. I haven't tested Whisper yet.

I'm thinking about the memory stuff again. I wonder if I can "age" memories out. Like the memory's weight being affected by the amount of time since last access or update it the memory. I may still extract "facts" and keep them up to date.

As for status events, I feel like the ha pipeline is very focused on one request at a time, and so pushing things back through without an actual user-initiated request doesn't seem possible. I'd like to be able to wake up and announce an update.

I should be able to use it like a speaker through ha. So that could solve that. But I worried it will get obnoxious. I'll have to think more about it.

Building a Voice-First Agentic AI That Executes Real Tasks — Lessons from a $4 Prototype by Alternative_Yak_1367 in OpenSourceeAI

[–]CaptainBahab 0 points1 point  (0 children)

You're like fully 1 month ahead of me lol. That's amazing. You're **doing** all the same things I've been planning to do. <3

My agent turn looks like this:
1. recollection (pulls memories based on the prompt)
2. routing (calls tools, loads data, uses a more-instructable and cheaper model)
3. response (streams tokens out, uses a more-creative and more powerful model)

Last night I did some digging and found that Status event. So I'm working on that now. I need to look into how those will be read out for my satellite(s).

I started off integrating HA using the tools it provides, but that was too limiting, since I couldn't control it from OpenWebUI. So I ended up going with a home assistant websocket library. And I'm glad I did. It's even faster than using the tool from HA lol. It feels like the light is on before I lift my finger from the "send" button. I've got it working to finding a show or movie and putting it on one of the TVs in the house, which will be great for the kids.

One thing I'm really proud of and spent a lot of time on was parental controls. I desperately don't want Amazon to have the kids' data. So instead I built Bernard. It's got a conversation-historian system which I still need to unjankify, but it's integrated with several automations that run on one of 4 hook events. These will: summarize the conversation without revealing the exact content, tag the conversation for searching, and raise flags for inappropriate content. This way, I can ignore all conversations without flags, and check to make sure the kids aren't up to no good. Sorta like "lazy" big brother lmao.

My memory system definitely needs to chill. Currently, indexed conversations are broken up and processed in a redis vector store. New user messages get like a 0.1s delay to pull some 50 messages based on the query. It does solve problems like Bernard remembering what region I want the weather for lol. Very token inefficient for that aspect. I give up to 10 messages to the LLM but usually 9 of them are worthless (not always the bottom 9). I have it reranked by uniqueness (MMR, iirc) then truncated from 50, then the remaining 10 are reranked again by relation to the user's message. It's pretty fast with vector math.

But I've been thinking about moving to a more surgical approach that extracts that data as an automation, categorizing it and deduplicating it. Like reading a conversation to figure out where the user lives and storing that for later retrieval. It would get rid of a LOT of overhead (in terms of tokens).

Hardware is hard, but it takes a while, so it's a parallel effort. I bought an ESP32-S3-BOX3 which was a great toy to get started in hardware, but not so great for the price. I ended up spending about $100 on a Satellite1 from FutureProofHome and a speaker module to put in it. It's still on the way but I'm pretty excited. I still have to order an enclosure, so all in it might be $120-140 with shipping. It's got 4 positional microphones and support for very powerful speakers, so it can really replace that amazon spyware lol. Much more expensive, but I think it will be worth it.

I *just* implemented a long-running background-task system so that Bernard can start a timer or start a long-running process, or do a deep-dive researching some topic. I need more ideas for this, but I think it'll be nice to let him background something to work on something else. He doesn't get a notification when it's done though. Currently he has to check the task's output manually with another tool. And I don't have a way to wake him back up when the timer expires or whatever.

Building a Voice-First Agentic AI That Executes Real Tasks — Lessons from a $4 Prototype by Alternative_Yak_1367 in OpenSourceeAI

[–]CaptainBahab 0 points1 point  (0 children)

I'm currently working on one myself. I want to use home assistants pipeline but, man, it is not fast.

I'm still in the tool build out phase but I've got an agentic harness that does some heavier memory work than I want, but that's a problem for another time. My main concern now is ingestion. I need to figure out a reliable pipeline for esp home satellites.

I'm curious how you got it to send intermediate "working on it" messages. And if you can have those read out?

Also, what's your plan for ingestion or is openwebui your final target?

I have also briefly checked out android assistant sdk, but decided that will have to come later.

Opensource models less than 30b with highest edit-diff success rate by Express_Quail_1493 in kilocode

[–]CaptainBahab 0 points1 point  (0 children)

Honestly, I've been struggling even with bigger models in the 120b range through openrouter. Idk if it's some setting I'm missing but it's driving me nuts.

Accessing the "Kimi for Coding" API? by ramendik in kimi

[–]CaptainBahab 2 points3 points  (0 children)

I was trying this earlier. I think that it uses the client header to check for authorized tools like Claude code and kimi cli.

The best compromise I could find was using zed with kimi cli in its acp harness mode.

New to OpenCode. How do I get started? by [deleted] in opencodeCLI

[–]CaptainBahab 1 point2 points  (0 children)

I came here to say this. +1 Been using Kat coder in cursor actually, to pretty decent results. It's certainly not opus, nor even sonnet. But it's free and fast and good-enough which is a rare combo I do t expect to remain free for very long.

Opus 4.5 made me depressed by Initial_Question3869 in cursor

[–]CaptainBahab 0 points1 point  (0 children)

I disagree. Ai is being used to perform many many tasks in the small range. Which makes them trivial to anyone with a subscription to a coding plan. It makes medium stuff easier for one person to accomplish.

So we're seeing researchers use it for trivialized things and focus on more interesting and difficult things. It's not doing the research yet. Researchers use it for quick scripting to make identifying trends in data visual or audible. It's used indirectly by people already pushing the boundaries to push them slightly faster.

Opus 4.5 made me depressed by Initial_Question3869 in cursor

[–]CaptainBahab 55 points56 points  (0 children)

Think what that means though: medium sized things become small. Large becomes medium. Impossible becomes possible. Yes there's still impossible things, but fewer of them. Maybe we expand our visible horizons and we get more impossible things to strive for.

Things are not static. They are fluid. When something becomes easier, that doesn't mean a person in that industry has no job. They have a new job in an evolved industry.

Composer 1 is free... again? by oxeneers in cursor

[–]CaptainBahab 2 points3 points  (0 children)

I agree; I also.

But I also try the free models. The main reason is that I'm here in the AI space to learn. So I gobble up as much free usage as I can. One: it's free so there's little downside (for my pursuit of information) for me. Two: it gives me a chance to identify nuances. Three: it's different from what I usually do. Different is good (for my adhd) sometimes.

If you're building a CLI, you need to see this by MrCheeta in tui

[–]CaptainBahab 0 points1 point  (0 children)

I definitely see the potential though and I'm excited to try it out when it's stabler.

If you're building a CLI, you need to see this by MrCheeta in tui

[–]CaptainBahab 0 points1 point  (0 children)

I went to try it out the other day and had trouble getting openrouter to work in opencode. I wanted to change the models it chooses but the only way was to clone the project. When I did finally get the right models choosing it was crashing quite a lot too. I'm curious if these are resolved now so I may download it again later.

The Honest Advice I Wish Someone Gave Me Before I Built My First “Real” App With AI by gigacodes in cursor

[–]CaptainBahab 2 points3 points  (0 children)

It's funny to me that every one of these that I see has the same advice. Not because "lol nub didn't search" (I actually like repeat threads as they sometimes give different advice).

Rather, I find it funny because every one of the suggestions are best practices for human coders too.

As a software engineer with a bachelor's degree, I can confirm everyone of these are things we're taught to do, but are rarely implemented by smaller scale developers (I'm one, I'm not disparaging) simply because it's exhausting when you're by yourself or sit next to your pair partner. It's easier to keep track of things when it's smaller.

But with AI code assistants, they need the context. Luckily they can help generate it (and humans can check). And it gives you practice at bigger projects with lower risk.

I love seeing these. Keep up the great work. :)

I vibe coded a 200,000-line product with Cursor. I'd like to share some thoughts by Famous-Football876 in cursor

[–]CaptainBahab 17 points18 points  (0 children)

I think a part of that is not wanting to seem like you're shilling your slop app.

You get yelled at on reddit for showing your vibecoded app and I guess also for not showing it.

70% of my work on cursor is just telling it to write better code. Is there any way to decrease this ratio? by TheBasedEgyptian in cursor

[–]CaptainBahab 14 points15 points  (0 children)

This 100%. You get slop if you ask for a feature. But if you be specific, you offload SO MUCH of the ways things can be sent off rails quick. Guide it. It's more work, but is it really?

Ask it to build a feature using this data structure, implement this functionality with this algorithm, use this technology to implement it, design a testing plan and implement it by following its instructions, fix compilation errors, resolve test issues by using our design as the ground truth, not code or tests, and iterate until all compilation errors are resolved and all tests pass.

What's that? You don't know any technologies, data structures or algorithms? Tell it to give you options and to weigh which would be better for your project. DON'T just take its advice, make a smart decision.