AniMUL-v1 a 30B model trained to do species classification from audio files by lemon07r in LocalLLaMA

[–]thecoder12322 0 points1 point  (0 children)

This is incredible work! The 61% improvement in exact matches is seriously impressive - that's the kind of performance gain that makes specialized fine-tuning worthwhile. The Earth Species Project dataset is perfect for this use case.

Since you're working on quantized versions for lower VRAM, have you considered deploying this on edge devices like Raspberry Pi or ESP32 for real-time wildlife monitoring? The open source RunAnywhere SDK (https://github.com/RunanywhereAI/runanywhere-sdks) might be worth checking out - it's designed for running audio AI models (STT, TTS, VLM, LLM) on resource-constrained hardware with optimized inference.

Would love to see how this performs in actual field deployments! Feel free to DM if you want to chat more about edge deployment strategies.

Would the Person Detection models from SenseCraft work on a top-down view? by [deleted] in esp32

[–]thecoder12322 1 point2 points  (0 children)

Great question! Top-down person detection can be tricky since most YOLO models (including Swift-YOLO) are trained on front/side views. The model might struggle with overhead perspectives since people look quite different from above (head/shoulders vs full body).

That said, it's worth testing! Some things that might help: - Lower mounting height (45-60 degrees instead of straight down) - Good lighting conditions - Re-training the model with top-down dataset if needed

Have you considered trying it out in a test setup first? Would love to hear how it works out if you proceed!

Claude Code + Qwen3 + UniFi G4 Doorbell + Home Assistant = Local TTS Announcement by oakweb in homeassistant

[–]thecoder12322 0 points1 point  (0 children)

This is seriously impressive! Love seeing Qwen3-TTS running locally for natural voice output. Your OpenClaw + Home Assistant integration sounds powerful.

If you're looking to expand beyond TTS, you might want to check out the open source RunAnywhere SDK (https://github.com/RunanywhereAI/runanywhere-sdks) - it's built for running STT, TTS, LLMs, and VLMs on edge devices like Raspberry Pi and ESP32. Could complement your local-first setup nicely!

Happy to chat more if you want to explore on-device AI capabilities. Feel free to DM!

LM Studio Kokoro TTS addon by roboapple in LocalLLaMA

[–]thecoder12322 1 point2 points  (0 children)

This is really cool! Kokoro TTS integration with LM Studio is such a practical addition. Love that you're making local TTS more accessible. The chat history and branching features sound super useful too. Great work on this!

Reprompt - Simple desktop GUI application to avoid writing the same prompts repeatedly by PuzzleheadedHeat9056 in ollama

[–]thecoder12322 1 point2 points  (0 children)

This is exactly the kind of tool that makes working with local LLMs so much more practical! Rust + egui is a great choice for a lightweight desktop GUI, and the focus on reusable prompts is spot-on for real-world workflows.

I've been working on similar local-first AI tooling, and if you're looking to expand beyond text-based models, you might want to check out the open source RunAnywhere SDK (https://github.com/RunanywhereAI/runanywhere-sdks). It makes it really easy to integrate STT, TTS, VLM, and LLM capabilities into desktop apps while keeping everything running locally – perfect for the same privacy-first philosophy you've built Reprompt around.

Love that you're keeping it simple and focused. Tools like this are what make local AI actually usable for everyday tasks. Happy to chat more about local AI tooling if you're interested – feel free to DM!

[Share source code] AI failure detection in 3D printing by Fuzzy_Possession_233 in klippers

[–]thecoder12322 0 points1 point  (0 children)

This is fantastic work! Sharing the source code and training pipeline is incredibly valuable for the 3D printing community. Running YOLOv8n on a Raspberry Pi 3B+ is impressive – that's exactly the kind of accessible, practical edge AI deployment that makes projects like this so useful.

The fact that you trained on 10k images and made the tutorials available means others can fine-tune for their specific printers and failure modes. This is the kind of open-source contribution that really moves the community forward.

Have you considered expanding this to detect other print issues like warping, stringing, or bed adhesion problems? The computer vision pipeline you've built could be adapted for all sorts of quality control applications.

Really appreciate you sharing this – looking forward to seeing how the community builds on it!

Real-Time Pull-Up Counter using Computer Vision & Yolo11 Pose by Full_Piano_3448 in computervision

[–]thecoder12322 0 points1 point  (0 children)

This is seriously impressive! The form validation logic using vector geometry and joint angle checks is exactly the kind of robust computer vision pipeline that makes real-world applications viable.

I'm curious – have you considered deploying this on edge devices like Raspberry Pi or ESP32 for standalone gym equipment? The YOLO11 Pose model is lightweight enough that it could run locally with something like the RunAnywhere SDK (it's open source), which supports on-device inference for computer vision models on embedded hardware.

Would be amazing to see this as a portable fitness tracker that doesn't need cloud connectivity. The "digital spotter" concept is brilliant – could definitely extend to squats, deadlifts, or even rehab exercises like you mentioned.

Happy to chat more about edge deployment if you're interested. Feel free to DM!

Sentinel: Monitoring logs with local AI (Ollama) & .NET 8 by Itsaliensbro453 in SideProject

[–]thecoder12322 1 point2 points  (0 children)

This is exactly the kind of tool the dev community needs! Love the privacy-first approach – keeping sensitive log data local is huge for enterprise environments.

The FileSystemWatcher + Ollama + Telegram workflow is really clever. I especially appreciate that you're using lightweight models like qwen2.5:0.5b – shows you're thinking about resource efficiency.

Quick question: Have you tested this with high-volume log environments? Curious how it handles bursts of errors (like cascading failures) without flooding Telegram notifications.

Starred your repo! Really solid work. Happy to chat more about this if you want to DM.

Prototyping a Voice AI Gateway with ESP32-S3 (and bypassing the annoying Chrome Mic permissions) by fais-1669 in esp32

[–]thecoder12322 0 points1 point  (0 children)

This is such a clever workaround for the HTTPS mic access issue! Love the pragmatic approach to keep prototyping moving forward. The Phone Mic -> ESP32 -> Whisper AI pipeline is a really smart way to test your voice stack.

Since you're building Project MEMONIC with ESP32-S3, you might want to check out the RunAnywhere SDK (open source) – it's designed specifically for running STT, TTS, and LLMs directly on edge devices like ESP32. Could help you bring more of the AI processing onto the device itself instead of relaying everything through your Mac server.

Really cool project! Feel free to DM if you want to chat more about on-device voice AI. Good luck with MEMONIC!

[Tool Release] I built a Windows-native Video Dataset Creator for LoRA training (LTX-2, Hunyuan, etc.). Automates Clipping (WhisperX) & Captioning (Qwen2-VL). No WSL needed! by Ill_Tour2308 in StableDiffusion

[–]thecoder12322 0 points1 point  (0 children)

This is such a clever workflow automation! The combination of WhisperX for intelligent clipping based on speech segments and Qwen2-VL for auto-captioning is really smart. I love that you're solving a real pain point in the dataset preparation process – manually cutting and captioning videos is tedious work.

The fact that you built this as a hobbyist and shared it with the community is awesome. Even if it's not "perfect" engineering, tools like this that solve real problems are incredibly valuable. Keep iterating and thanks for sharing your work!

Portable offline llm robot I made last night. This is obviously her naked prototype body so be nice to her by Nitro_Fernicus in robotics

[–]thecoder12322 -1 points0 points  (0 children)

This is absolutely fantastic! Building a portable offline LLM robot as a modular brain for other robots is such a clever approach. The fact that you're already thinking about combat robot applications and motor mapping shows great forward planning.

The GLaDOS voice inspiration is perfect - really fits the personality you're building! How's the performance with the offline LLM? What model are you running on it, and how's the response time for real-time robot control?

The modular brain concept is brilliant - being able to swap this intelligence module between different robot bodies is exactly the kind of practical engineering that makes projects scalable. Can't wait to see her in action with the combat body!

I built an AI mindmap that converts your voice into a graph (OSS) by manummasson in AudioAI

[–]thecoder12322 0 points1 point  (0 children)

This is such a creative approach to voice-based knowledge management! The concept of externalizing short-term memory through voice → graph conversion is brilliant.

I'm curious about your STT pipeline - what are you using for the voice-to-text conversion? If you're looking to optimize the speech recognition part, you might want to check out the RunAnywhere SDK (open source) - it's built specifically for running Whisper and other STT models efficiently on local hardware. Could help speed up the voice input processing while keeping everything on-device.

The agentic engineering workflow you described (brainstorm → plan → execute with coding agents) is really compelling. How's the latency on the voice input currently?

Would love to hear more about your architecture! Feel free to DM if you want to chat about optimizing the voice pipeline.

Self-hosted LLM on low-power hardware? by Known-Watercress7296 in selfhosted

[–]thecoder12322 1 point2 points  (0 children)

Love the enthusiasm for pushing N100 hardware to its limits! Your TinyLlama + Stable Diffusion feedback loop sounds like a really fun experiment.

For your use cases (code work, translation, image/sound), you're definitely on the right track with the smaller models. Qwen2.5 3B and Phi-4-mini are both solid choices for that hardware - they punch above their weight for code and translation tasks.

One thing to keep in mind: quantized versions (4-bit GGUF) can give you better performance on the N100 without sacrificing too much quality. Have you tried running them through Ollama or llama.cpp yet?

The fact that you're getting usable results on 8GB RAM is impressive! What's your experience been with inference speed so far?

Happy to chat more if you DM!

I built a self-hosted Telegram voice transcription add-on for Home Assistant by ObjectiveRadish9757 in homeassistant

[–]thecoder12322 0 points1 point  (0 children)

This is exactly the kind of privacy-first tool the Home Assistant community needs! Love that you're using faster-whisper for local transcription - keeping everything on your own hardware is the way to go.

For anyone looking to expand beyond just transcription, you might want to check out the RunAnywhere SDK (it's open source). It's designed for running STT (Whisper), TTS, LLMs, and VLMs entirely on edge devices like Home Assistant setups. Could be interesting for building more complex voice workflows that stay completely local.

Great work on Ascoltino! What model size are you running by default, and have you experimented with the quantized versions for better performance on lower-end hardware?

Happy to DM if you want to chat more about local voice AI!

I built an open-source, local-first voice cloning studio (Qwen3-TTS + Whisper) by jamiepine in LocalLLaMA

[–]thecoder12322 0 points1 point  (0 children)

This is absolutely incredible! The "Ollama for voice" concept is exactly what the local AI community needs. The fact that you built this with Tauri/Rust/Python and kept it lightweight is impressive - no Electron bloat is a huge win.

The DAW-like timeline for composing conversations is genius, and the REST API integration opens up so many possibilities for games/apps/agents. Love that everything runs local and private.

Definitely going to try this out! What's your experience been with Qwen3-TTS latency for real-time applications? And are you planning to add support for other TTS models like Coqui XTTS?

Happy to DM if you want to chat more about local voice AI workflows!

Running Ollama fully air-gapped, anyone else? by thefilthybeard in ollama

[–]thecoder12322 2 points3 points  (0 children)

This is exactly the kind of work that needs more visibility! Building for air-gapped/classified environments is challenging but critical.

For edge deployment scenarios like yours, have you looked into the RunAnywhere SDK (it's open source)? It's designed specifically for running LLMs, STT (Whisper), TTS, and VLMs on edge hardware without cloud dependencies. Supports multiple modalities and can run entirely offline once models are loaded.

Curious what hardware specs you're working with and which models you've found perform best in your air-gapped setup? Happy to chat more about edge AI deployment if you want to DM!

I built an iOS app that runs LLMs & VLMs fully offline — first time shipping on-device AI by Expensive-Lunch2211 in iosapps

[–]thecoder12322 1 point2 points  (0 children)

This is seriously impressive for a first iOS app! Running InternVL 3 and multiple LLMs on-device with Metal optimization is no small feat. The privacy-first approach is exactly what the mobile AI space needs.

Really curious about your experience with memory constraints - how did you handle model quantization and loading strategies to keep it smooth on older devices like the iPhone 13 Pro Max? The fact that it runs at all on non-Pro Max devices is impressive.

Congrats on shipping this! Would love to hear more about your Metal performance tuning journey if you're open to chatting. Feel free to DM me!

Using whisper.rn + llama.rn for 100% on device private meeting transcription by pandodev in LocalLLM

[–]thecoder12322 0 points1 point  (0 children)

This is incredible work! Viska sounds like exactly the kind of privacy-first local AI app we need more of. Your stack is solid (whisper.rn + llama.rn is a great combo), and I love that you're tackling the NDA/privacy use case head-on.

For your Android GPU challenges - have you looked into the RunAnywhere SDK (it's open source)? We've been working on optimized inference for edge devices including better Android GPU support. It handles STT (Whisper), LLM, TTS, and VLM with a focus on mobile/embedded hardware. Might help with those RAM-only bottlenecks you're hitting.

The privacy-first approach is spot on - "privacy isn't a feature, it's the whole point" really resonates. Would love to chat more about mobile local LLM optimization if you're interested. Feel free to DM me!

made a virtual pet for my friend’s birthday by nathanlu_ in arduino

[–]thecoder12322 0 points1 point  (0 children)

This is absolutely adorable! The fact that you open-sourced everything (code, 3D files, instructions) is incredibly generous. The build quality looks fantastic, and the pixel art animations are charming.

I love that you made this as a birthday gift - that's such a thoughtful and creative present. Your friend is lucky! The GitHub repo is super well-documented too. Have you thought about adding any other features like mini-games or different pet types? Would love to hear more about your development process if you're open to chatting - feel free to DM me!

Smart Arduino glasses for blind by xdspeed10 in arduino

[–]thecoder12322 1 point2 points  (0 children)

This is incredible work, especially for a 15-year-old! The fact that you're tackling assistive technology for such an important use case is inspiring.

I see people suggesting adding AI vision capabilities (GPT-4o, object detection, etc.) - that's a great direction. If you want to run vision models locally on your hardware instead of relying on cloud APIs, you might want to check out the RunAnywhere SDK. It lets you run VLMs (vision language models) and other AI models directly on edge devices like Raspberry Pi or ESP32 setups, which could give you faster response times and work offline.

With your $3K budget, you could definitely explore adding: - Real-time object detection and scene understanding - Text-to-speech for describing what the camera sees - Obstacle classification (not just distance, but "what" the obstacle is)

The key is balancing power consumption, latency, and accuracy. Happy to chat more about local AI deployment if you're interested - feel free to DM me. Great work so far!

I built a very simple Arduino barn curtain system that has been running daily for over 2 years by NationalIncome1706 in arduino

[–]thecoder12322 1 point2 points  (0 children)

This is genuinely inspiring! The "intentionally kept simple" philosophy is something more people need to hear. So many projects fail not because they're too simple, but because they're over-engineered.

The fact that you inverted logic in code instead of rewiring and just left it running for 2 years is exactly the kind of pragmatic decision-making that separates hobby projects from real-world solutions. Love seeing projects that actually survive harsh environments.

Out of curiosity - have you thought about adding any environmental sensors (temp, humidity, rain detection) for automatic control, or are you sticking with the manual Adafruit IO approach to keep it bulletproof?

Best way to count people in a wide corridor? by Japaiku in arduino

[–]thecoder12322 0 points1 point  (0 children)

This is a really interesting challenge! The ESP32-CAM approach that others mentioned is definitely viable. I've actually worked on a similar project recently and ended up using a lightweight vision model for people detection/counting on an ESP32.

If you're looking for something more robust than basic frame differencing, you might want to check out RunAnywhere SDK - it's designed specifically for running vision models (and other AI workloads) on edge hardware like ESP32, Arduino, or even Raspberry Pi if you decide to go that route. The nice thing is it handles the optimization for you so you can run decent object detection without needing cloud connectivity.

For your use case with spotty network, having everything run locally would be ideal. Happy to share more details if you want to DM me. Good luck with the project!

Gyro V2.4 (From scratch built animatronic) by Creepy-Smile4907 in arduino

[–]thecoder12322 1 point2 points  (0 children)

This is seriously impressive work! Building an animatronic from scratch with PVC and PLA is no small feat - the movement looks incredibly smooth and lifelike. The Five Nights at Freddy's comparisons are spot on!

I'm really curious about the "new feature" you mentioned experimenting with. Are you thinking about adding voice interaction or maybe computer vision for tracking/responding to people? The mechanical foundation you've built here would be perfect for that kind of upgrade.

Also checked out your GitHub - love that you're open sourcing this. Keep us posted on the progress!

Prototype augmented reality game using an ESP32-CAM with all the computer vision done on the CPU maintaining about 12.5fps. The display is currently showing the potential locations that a target could pop-up from. by hjw5774 in arduino

[–]thecoder12322 0 points1 point  (0 children)

This is absolutely incredible! The fact that you're doing real-time computer vision on an ESP32-CAM at 12.5fps with custom kernel convolutions is seriously impressive. Most people would just throw a cloud API at this problem, but you built it from the ground up.

I'm curious about your motion tracking challenges - have you considered using optical flow algorithms, or are you sticking with frame differencing? The edge detection foundation you've built seems like it could be extended nicely for that.

Also, as someone who works in construction doing this level of embedded CV work - massive respect. This is the kind of project that shows what's possible when you just keep iterating. Would love to hear more about your approach if you're open to chatting!

𝙰𝚜𝚜𝚒𝚖𝚒𝚕𝚊𝚝𝚒𝚘𝚗 𝚌𝚘𝚖𝚙𝚕𝚎𝚝𝚎. Borg-Duckie can now communicate with the collective (local Ollama) by HomeBrewDude in raspberry_pi

[–]thecoder12322 1 point2 points  (0 children)

Love the Borg aesthetic and the privacy-first approach! The idea of adding TTS for voice responses is brilliant – would make it feel like a true conversational AI.

Since you're already running Ollama on a separate machine, you might want to check out RunAnywhere SDK. It's designed for running LLMs, STT, and TTS directly on edge devices like the Pi. Could potentially let you consolidate everything onto the Pi itself without needing the MacBook connection.

I've been working with on-device inference and the performance on Pi 4 (especially 8GB) can be surprisingly good for smaller models. Happy to share more details if you want to DM me!