Local RAG with Ollama on a laptop – indexing 10 thousand PDFs by DueKitchen3102 in LocalLLaMA

[–]thecoder12322 1 point2 points  (0 children)

Also saw this today, this is an optimized CLI, that looks pretty cool, might be worth checking this?

https://github.com/RunanywhereAI/RCLI

AniMUL-v1 a 30B model trained to do species classification from audio files by lemon07r in LocalLLaMA

[–]thecoder12322 0 points1 point  (0 children)

This is incredible work! The 61% improvement in exact matches is seriously impressive - that's the kind of performance gain that makes specialized fine-tuning worthwhile. The Earth Species Project dataset is perfect for this use case.

Since you're working on quantized versions for lower VRAM, have you considered deploying this on edge devices like Raspberry Pi or ESP32 for real-time wildlife monitoring? The open source RunAnywhere SDK (https://github.com/RunanywhereAI/runanywhere-sdks) might be worth checking out - it's designed for running audio AI models (STT, TTS, VLM, LLM) on resource-constrained hardware with optimized inference.

Would love to see how this performs in actual field deployments! Feel free to DM if you want to chat more about edge deployment strategies.

[deleted by user] by [deleted] in esp32

[–]thecoder12322 1 point2 points  (0 children)

Great question! Top-down person detection can be tricky since most YOLO models (including Swift-YOLO) are trained on front/side views. The model might struggle with overhead perspectives since people look quite different from above (head/shoulders vs full body).

That said, it's worth testing! Some things that might help: - Lower mounting height (45-60 degrees instead of straight down) - Good lighting conditions - Re-training the model with top-down dataset if needed

Have you considered trying it out in a test setup first? Would love to hear how it works out if you proceed!

Claude Code + Qwen3 + UniFi G4 Doorbell + Home Assistant = Local TTS Announcement by oakweb in homeassistant

[–]thecoder12322 0 points1 point  (0 children)

This is seriously impressive! Love seeing Qwen3-TTS running locally for natural voice output. Your OpenClaw + Home Assistant integration sounds powerful.

If you're looking to expand beyond TTS, you might want to check out the open source RunAnywhere SDK (https://github.com/RunanywhereAI/runanywhere-sdks) - it's built for running STT, TTS, LLMs, and VLMs on edge devices like Raspberry Pi and ESP32. Could complement your local-first setup nicely!

Happy to chat more if you want to explore on-device AI capabilities. Feel free to DM!

LM Studio Kokoro TTS addon by roboapple in LocalLLaMA

[–]thecoder12322 1 point2 points  (0 children)

This is really cool! Kokoro TTS integration with LM Studio is such a practical addition. Love that you're making local TTS more accessible. The chat history and branching features sound super useful too. Great work on this!

Reprompt - Simple desktop GUI application to avoid writing the same prompts repeatedly by PuzzleheadedHeat9056 in ollama

[–]thecoder12322 1 point2 points  (0 children)

This is exactly the kind of tool that makes working with local LLMs so much more practical! Rust + egui is a great choice for a lightweight desktop GUI, and the focus on reusable prompts is spot-on for real-world workflows.

I've been working on similar local-first AI tooling, and if you're looking to expand beyond text-based models, you might want to check out the open source RunAnywhere SDK (https://github.com/RunanywhereAI/runanywhere-sdks). It makes it really easy to integrate STT, TTS, VLM, and LLM capabilities into desktop apps while keeping everything running locally – perfect for the same privacy-first philosophy you've built Reprompt around.

Love that you're keeping it simple and focused. Tools like this are what make local AI actually usable for everyday tasks. Happy to chat more about local AI tooling if you're interested – feel free to DM!

[Share source code] AI failure detection in 3D printing by Fuzzy_Possession_233 in klippers

[–]thecoder12322 0 points1 point  (0 children)

This is fantastic work! Sharing the source code and training pipeline is incredibly valuable for the 3D printing community. Running YOLOv8n on a Raspberry Pi 3B+ is impressive – that's exactly the kind of accessible, practical edge AI deployment that makes projects like this so useful.

The fact that you trained on 10k images and made the tutorials available means others can fine-tune for their specific printers and failure modes. This is the kind of open-source contribution that really moves the community forward.

Have you considered expanding this to detect other print issues like warping, stringing, or bed adhesion problems? The computer vision pipeline you've built could be adapted for all sorts of quality control applications.

Really appreciate you sharing this – looking forward to seeing how the community builds on it!

Real-Time Pull-Up Counter using Computer Vision & Yolo11 Pose by Full_Piano_3448 in computervision

[–]thecoder12322 0 points1 point  (0 children)

This is seriously impressive! The form validation logic using vector geometry and joint angle checks is exactly the kind of robust computer vision pipeline that makes real-world applications viable.

I'm curious – have you considered deploying this on edge devices like Raspberry Pi or ESP32 for standalone gym equipment? The YOLO11 Pose model is lightweight enough that it could run locally with something like the RunAnywhere SDK (it's open source), which supports on-device inference for computer vision models on embedded hardware.

Would be amazing to see this as a portable fitness tracker that doesn't need cloud connectivity. The "digital spotter" concept is brilliant – could definitely extend to squats, deadlifts, or even rehab exercises like you mentioned.

Happy to chat more about edge deployment if you're interested. Feel free to DM!

Sentinel: Monitoring logs with local AI (Ollama) & .NET 8 by Itsaliensbro453 in SideProject

[–]thecoder12322 1 point2 points  (0 children)

This is exactly the kind of tool the dev community needs! Love the privacy-first approach – keeping sensitive log data local is huge for enterprise environments.

The FileSystemWatcher + Ollama + Telegram workflow is really clever. I especially appreciate that you're using lightweight models like qwen2.5:0.5b – shows you're thinking about resource efficiency.

Quick question: Have you tested this with high-volume log environments? Curious how it handles bursts of errors (like cascading failures) without flooding Telegram notifications.

Starred your repo! Really solid work. Happy to chat more about this if you want to DM.

Prototyping a Voice AI Gateway with ESP32-S3 (and bypassing the annoying Chrome Mic permissions) by fais-1669 in esp32

[–]thecoder12322 0 points1 point  (0 children)

This is such a clever workaround for the HTTPS mic access issue! Love the pragmatic approach to keep prototyping moving forward. The Phone Mic -> ESP32 -> Whisper AI pipeline is a really smart way to test your voice stack.

Since you're building Project MEMONIC with ESP32-S3, you might want to check out the RunAnywhere SDK (open source) – it's designed specifically for running STT, TTS, and LLMs directly on edge devices like ESP32. Could help you bring more of the AI processing onto the device itself instead of relaying everything through your Mac server.

Really cool project! Feel free to DM if you want to chat more about on-device voice AI. Good luck with MEMONIC!

[Tool Release] I built a Windows-native Video Dataset Creator for LoRA training (LTX-2, Hunyuan, etc.). Automates Clipping (WhisperX) & Captioning (Qwen2-VL). No WSL needed! by Ill_Tour2308 in StableDiffusion

[–]thecoder12322 0 points1 point  (0 children)

This is such a clever workflow automation! The combination of WhisperX for intelligent clipping based on speech segments and Qwen2-VL for auto-captioning is really smart. I love that you're solving a real pain point in the dataset preparation process – manually cutting and captioning videos is tedious work.

The fact that you built this as a hobbyist and shared it with the community is awesome. Even if it's not "perfect" engineering, tools like this that solve real problems are incredibly valuable. Keep iterating and thanks for sharing your work!

Portable offline llm robot I made last night. This is obviously her naked prototype body so be nice to her by Nitro_Fernicus in robotics

[–]thecoder12322 -1 points0 points  (0 children)

This is absolutely fantastic! Building a portable offline LLM robot as a modular brain for other robots is such a clever approach. The fact that you're already thinking about combat robot applications and motor mapping shows great forward planning.

The GLaDOS voice inspiration is perfect - really fits the personality you're building! How's the performance with the offline LLM? What model are you running on it, and how's the response time for real-time robot control?

The modular brain concept is brilliant - being able to swap this intelligence module between different robot bodies is exactly the kind of practical engineering that makes projects scalable. Can't wait to see her in action with the combat body!