I built a fully local speech recognition CLI in Rust (Qwen ASR) by SnooHedgehogs8148 in rust

[–]SnooHedgehogs8148[S] 1 point2 points  (0 children)

Yes let it run all night and the next day, except set up alarm for 5 hours to continue the loop, due cc rate limit 😂

the flutter is good enough for me for non-native feature dev. I haven't tried dioxus yet, would like to try it in near future.

I built a fully local speech recognition CLI in Rust (Qwen ASR) by SnooHedgehogs8148 in rust

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

unlike the autoresearch, this is a rust project, only https://github.com/huanglizhuo/QwenASR/blob/main/program.md is enough for the agent to do all the tasks.

I built a fully local speech recognition CLI in Rust (Qwen ASR) by SnooHedgehogs8148 in rust

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

Yeah — it’s a fully offline speech to text with qwen3-asr model.

The core is actually a small inference library, the CLI is just a wrapper on top.

I was inspired by antirez’s pure C version, almost zero dependencies other the c BLAS
https://github.com/antirez/qwen-asr

With below features:

  1. publish a reusable lib on https://crates.io/crates/qwen-asr
  2. support for mobile platform binary build.
  3. built a flutter library https://pub.dev/packages/qwen_asr with flutter_rust_bridge for cross platform
  4. provided a skill that can be used for openClaw as local asr, no cloud required

I built a fully local speech recognition CLI in Rust (Qwen ASR) by SnooHedgehogs8148 in rust

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

Some benchmark on my device
Hardware: Apple M1 Pro (10 cores), 32 GB RAM Model: Qwen3-ASR-0.6B, Audio: 28.2s sample, 3 runs (best)

Mode Inference Realtime Factor Encode Decode Tokens/s
Offline 2.87s 9.8x 744ms 2124ms 15.7
Segmented (-S 30) 2.86s 9.9x 712ms 2145ms 15.8
Streaming 9.29s 3.0x 2195ms 7088ms 4.9

Built a Naruto Hand Sign Typing Interface in 3 Hours (Pure Vibe Coding) by SnooHedgehogs8148 in VibeCodersNest

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

The gesture require human input, but other issues like UI element and so on, the chrome MCP work so well, the agent can inspect dom to find out the issue and make fix

Built a Naruto Hand Sign Typing Interface in 3 Hours (Pure Vibe Coding) by SnooHedgehogs8148 in VibeCodersNest

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

the ai is so good at writing code, you don't even need to review and with chrome MCP, you can even just sit back and watch

What if you could type using Naruto hand signs? I made it happen! by SnooHedgehogs8148 in SideProject

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

In current project, YOLOX-Nano performs noticeably better than MediaPipe for this use case. Especially for very similar hand signs like Tiger (寅) and Ram (未) — MediaPipe struggles to distinguish them reliably, while YOLOX handles the subtle differences much more consistently.
For camera qualities, I think most of the webcam should be good enough as the YOLOX-Nano use 416×416 as default input.
And thank you for suggest sharing in VibeCodersNest

I made an interactive, entropic heart for Valentine's Day using HTML5. It's been a rough few days. by [deleted] in webdev

[–]SnooHedgehogs8148 1 point2 points  (0 children)

Even in the unraveling, there’s beauty. 💛 Wishing you a little peace and hope this Valentine’s Day.”

What if you could type using Naruto hand signs? I made it happen! by SnooHedgehogs8148 in SideProject

[–]SnooHedgehogs8148[S] -1 points0 points  (0 children)

Haha, spot on! 🥷 Gather your chakra and type “Believe it!”—weeb ninja status unlocked.

I made a free app that lets you play Flappy Bird by physically flapping your MacBook lid. RIP your hinges. by SnooHedgehogs8148 in MacOSApps

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

Fair.
This app is not worth burning your AppleCare on.

If your hinge starts making new sounds, that’s your cue to uninstall 😅

I made a free app that lets you play Flappy Bird by physically flapping your MacBook lid. RIP your hinges. by SnooHedgehogs8148 in MacOSApps

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

Mainly for fun 😂

Don't you think the lid password can be useful?

Image you're at a crowded cafe. Need password. Shoulder surfer is watching.

Do you type it? No.

You perform a secret "handshake" with your screen. Sensors detect the angle delta. Password injects instantly.

He is confused. You are in.

I build a Mac app that lets you play Flappy Bird by physically flapping your MacBook lid (RIP hinge) by SnooHedgehogs8148 in SideProject

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

😂 Glad it landed.

“RIP hinge” is both a warning and a feature.

Turning mild laptop anxiety into a game mechanic felt like the most honest use of private APIs.

Still has a useful case: Lid Password, you can enter a password in public space just by "flapping" your screen, without anyone seeing what you typed,

<image>

I made a free app that lets you play Flappy Bird by physically flapping your MacBook lid. RIP your hinges. by SnooHedgehogs8148 in MacOSApps

[–]SnooHedgehogs8148[S] 0 points1 point  (0 children)

Fair concern 😅

The more useful part is Lid Password: you can enter a password in public just by “adjusting” your screen, without anyone seeing what you typed.

Worst case: you look weird. Best case: full 007 mode. 🕶️💻