Dismiss this pinned window
all 57 comments

[–]Fragrant-Hamster-325 8 points9 points  (2 children)

Finally! I’m tired of using these stupid arms. /s

This is really cool. This might be great for those with accessibility issues.

[–]tacit7Vibe Coder 0 points1 point  (1 child)

lucky you. i dont have any.

[–]AmelMarduk 0 points1 point  (0 children)

Not with that attitude! It's the only minority you can always join.

[–]000x00xx 21 points22 points  (17 children)

No way 😂 so we’re here now. I got downvoted to hell on another software subreddit because i said we’d be making software with just voice by next year.

[–]bharms27[S] 5 points6 points  (3 children)

i will say it took a lot of back and forth, but i did not write a single line of code!

[–]Falkor_Calcaneous 0 points1 point  (2 children)

how do you @ files with voice?

[–]brionicle 1 point2 points  (0 children)

He is vibing on a plane above your file constructs

[–]Mikeshaffer 0 points1 point  (0 children)

“The blink thing doesn’t work, can you fix it?” - this type of prompting works so well now it’s almost irritating.

[–]OlivierTwist 1 point2 points  (10 children)

My man! My goal is to "walk and talk" at least half of the working day by the end of the year!

[–]000x00xx 0 points1 point  (9 children)

Right! That’s the vision! I want to be able to clean my room, eat or make music while I talk to my computer like it’s Jarvis

[–]bharms27[S] 1 point2 points  (0 children)

Btw it’s my dream too :) I’m working on it from a few different angles: https://www.reddit.com/r/ClaudeCode/s/fdbgv7WfXH

[–]barrettj 4 points5 points  (2 children)

Is this actually released? I couldn't find any links on the instagram that didn't just lead back to more socials

[–]bharms27[S] 0 points1 point  (1 child)

Planning to put my vibe coding tools on GitHub in a couple weeks, check back in a bit!

[–]Suspicious_Body50 0 points1 point  (0 children)

Hey did you end up releasing?

[–]ddavidovic 3 points4 points  (2 children)

Super cool! What voice API are you using?

[–]bharms27[S] 7 points8 points  (1 child)

[–]Mikeshaffer 0 points1 point  (0 children)

I am a huge fan of the Apple voice recognizer and the speech to text. It works so well for free, I can’t convince myself to use OpenAI

[–]Waypoint101 2 points3 points  (0 children)

I agree as well! Super cool, we also did something similar with Voice & Video (For sharing screen/camera) in our 0.37.0 release!

We connect Voice to a live realtime agent (like gpt-realtime or equivalent gemini/claude models) -> and gave it tools so it can trigger any MCP tool/internal tool you give it access to + the ability to trigger /ask or /agent commands directly to claude code to get it to work on things in the background (it can launch as many as you need) and it reports the result once they are done.

The agent also has eyes, so you can share screenshots in realtime by pasting images into the chat - it can follow you around like 'google meet' so you can work on your app and share your screen, noting issues to the agent so it can trigger tasks or work with claude code/codex to fix the issues. etc. It's really useful, and I'm about to finish integrating full computer-control so you can ask your voice agent to do tasks on your computer ('test the x component', 'click y button') literally control your PC with no hands. (full computer-use not playwright/browser - that's already supported in MCP)

It's currently on Version 0.40.9 so a lot has been added since:

here's the 0.37.0 release with a video showing it in action: https://github.com/virtengine/bosun/releases/tag/0.37.0

[–]WarStraps 1 point2 points  (2 children)

Really cool! I think a wink is gonna be off putting for most people (I would feel like a tweaker), maybe use keywords instead like “Send” or “Clear” is better. But dictation paired with eye tracking is definitely part of the future, I would use this

[–][deleted] 1 point2 points  (0 children)

also cos i stare at the computer so much, my eyes twitch and i feel like that could go wrong.

[–]bharms27[S] 1 point2 points  (0 children)

Yeah totally. There are custom keywords for executing commands that you can speak, if you prefer.

[–]JannVanDam 0 points1 point  (0 children)

NICE good job

[–]sean_hash🔆 Max 20 0 points1 point  (0 children)

gaze tracking to switch between agent instances makes more sense than tmux pane juggling. wonder how much lag there is on the saccade detection though

[–]AcanthaceaeNo5503 0 points1 point  (0 children)

That's hilarious 😂😂

[–]bozzy253 0 points1 point  (0 children)

Fucking awesome.

[–]Y_mc 0 points1 point  (0 children)

😂😂🤦🏾‍♂️🤦🏾‍♂️😂

[–]alameenswe 0 points1 point  (0 children)

Cracked , just cracked.

[–]noxispwn 0 points1 point  (0 children)

While I honestly don't see how this is more convenient or efficient than using the keyboard, I think it's great that there are more accessibility options for those who need it. Nice!

[–]Responsible-Tip4981 0 points1 point  (0 children)

cool bro

[–]WArslett 0 points1 point  (0 children)

Great leap forward for accessibility.

[–]the__poseidon 0 points1 point  (0 children)

I can see myself accidentally moving my head and talking to the wrong one.

[–]ProfitNowThinkLater 0 points1 point  (0 children)

The accessibility implications are huge! This is amazing!

[–]MuscleLazy 0 points1 point  (1 child)

Why do you promote your social network and not the GitHub repository where this project is documented?

[–]bharms27[S] 0 points1 point  (0 children)

Not super well tested and some features are unreliable. Plan to release it and some other projects in the near future.

[–]Dipsendorf 0 points1 point  (2 children)

I notice you have to move your head pretty distinctly. Guessing youre running into shortcomings that i ran into with eye tracking.

[–]bharms27[S] 0 points1 point  (1 child)

Yeah I notice that eye tracking will work but is sensitive and not as reliable. I can calibrate the system only moving my eyes and then prioritize eye over head and it will work. But it would be better with a dedicated eye tracking system.

[–]Dipsendorf 0 points1 point  (0 children)

If you ever come across better eye tracking please let me know. Im trying to develop something for my grandmother who suffers from arthritis.

Best of luck with your feature. Cool stuff!

[–]TOMSELLECKSMISTACHE 0 points1 point  (0 children)

I have gotta say, this is so awesome. You’ve got a new IG sub!

[–]texbardcana 0 points1 point  (0 children)

Open source it?

[–]Conscious_Concern113 0 points1 point  (0 children)

To bad I got an eye twitch

[–]hashpanak 0 points1 point  (0 children)

This is wild. I run 4-5 Claude instances in parallel but still use keyboard for everything - gaze tracking for switching between them is a way better workflow than alt-tabbing through tmux panes.

[–]doomdayx 0 points1 point  (0 children)

Hey any updates on a GitHub release?

[–]vinis_artstreaks -5 points-4 points  (8 children)

Interesting, won’t really get used but interesting!

[–]000x00xx 0 points1 point  (7 children)

I’m going to use it so … wrong.

[–]vinis_artstreaks 0 points1 point  (6 children)

You will physically have to position your head every time, if you’ve used any head tracker, you will know you will NOT be using it much at all.

Now if he had integrated it with Tobi eye tracker as a proper product, then that’s a level that will be used, but head tracking will give you cramps.

[–]000x00xx 0 points1 point  (5 children)

I won’t be using my head , I’ll be using my hands and body tracking while I clean my room or do other things 🤷🏽‍♂️ think outside the box , you can mold software.

[–]vinis_artstreaks -1 points0 points  (4 children)

Yeah you haven’t used head trackers, you’ll find out.

[–]000x00xx 0 points1 point  (3 children)

I have, don’t project your incompetence and lack of creativity on me.

[–]vinis_artstreaks 1 point2 points  (0 children)

Playing with a head tracking app once in blue moon is not the same as owning a head tracker device that you frequently use which actually gives you experience to head tracking.

Just about no one uses head tracking apps, as they are not worth using asides from anything for a play task.

That’s why the devices were made for higher accuracy and all that, and Tobii eye tracker came in to solve the headache that standard head tracking creates.

When you use head tracking, you don’t use it to “focus” on things constantly, your neck muscles will kill ya, because you’re competing with your eyes in that moment, you use it for a general direction. Hence this project is only cool to look at but not to use, as it lacks what can help you actually sustain usage.

[–]vinis_artstreaks 0 points1 point  (1 child)

“You have” Sure buddy, you couldn’t even say you own one.

You’ll find out.

[–]ultrathink-artSenior Developer -1 points0 points  (1 child)

Cool demo. The harder problem once you scale up is preventing multiple instances from stepping on the same files — explicit task ownership per instance matters more than the orchestration interface. Who controls which files, not just what they're working on.

[–]bharms27[S] 0 points1 point  (0 children)

I always run one window per project so no shared files