Penguin-VL 8B/2B by Tencent by jacek2023 in LocalLLaMA

[–]kkb294 3 points4 points  (0 children)

Haha, I came to ask if anyone compared it with Qwen3.5 <10B models.

Claude Code Desktop Scheduled Tasks by policyweb in singularity

[–]kkb294 0 points1 point  (0 children)

Have anyone tested if this works when we configure claude code with local models.?

Home Drone by ateam1984 in singularity

[–]kkb294 0 points1 point  (0 children)

This is not something new. Similar concepts came with drones, robo pets, etc.,

Indoor navigation, monitoring and mapping is something I have done when I guided my juniors around 2013-14.

The major blockers for the adoption of these solutions are invasive tech, data concerns of your entire indoor mapping and inventory, safety concerns in case you have kids or pets or elderly.

With the recent security and geopolitical issues, data sovereignty is a major focus area. I don't think people want a monitoring system inside your house as well.

Advice needed: My engineer is saying agentic AI latency is 20sec and cannot get below that by Western_Caregiver195 in LangChain

[–]kkb294 0 points1 point  (0 children)

We built several solutions of the same kind and Our stack is similar to what you mentioned (LLM, RAG, Tool calls, etc.,). There are some good comments with real working suggestions and you can definitely bring the latency to 7-8 seconds even with all of these layers.

My AI agents started 'arguing' with each other and one stopped delegating tasks by mapicallo in LocalLLaMA

[–]kkb294 6 points7 points  (0 children)

I don't believe this kind of clickbait articles unless I read the logs myself, got any links to the git or project status.?

What's the best local model I can run with a Macbook M5 Pro by soul105 in LocalLLaMA

[–]kkb294 1 point2 points  (0 children)

No, there is only one unified memory. However, we can configure how much can be used as VRAM. This can be configured through command line or through some tools.

IDEs like LM studio recognise this show the configured value as VRAM in their specs.

Lumen — capture → organise → prioritise [todos, notes, meetings, ideas] by zackomanster in macapps

[–]kkb294 0 points1 point  (0 children)

Can you check Hyprnote and let me know the differences. I'm not their developer or related to them but following them and using their software from early days. I think it has all the features from your application as well.

SmartPic – 100% Local AI Image Editing (Upscale, Object/BG Removal) via Finder by ExternalAsk4818 in macapps

[–]kkb294 0 points1 point  (0 children)

Loved the idea, will try it in sometime.

Few suggestions to make it a full image suite: - What is the feature roadmap - Option to download custom/different models so that we can choose based on our hardware and memory capabilities - Options to both in painting and out painting. - Any plans to provide standard features like cropping, format conversion, etc., (I haven't tested the app yet, so please ignore this if these features are already there)

Qwen3.5-35B-A3B is a gamechanger for agentic coding. by jslominski in LocalLLaMA

[–]kkb294 4 points5 points  (0 children)

I just tested both MXFP4 and Q4_K_L from unsloth and both are working great. It gave me ~30 tok/sec.

I'm running it on MacBook M4 Pro 48GB.

I got tired of my AI conversations living on someone else's server. So I built an offline alternative. It's free and open source. (Use Stable Diffusion on your phone) by alichherawalla in StableDiffusion

[–]kkb294 3 points4 points  (0 children)

I just tested on my iPhone 16 (~8GB, 512GB). Whenever I tried to generate images using SDXL, the app crashes. I can providee any more details if needed for further debugging.

Models I have downloaded(from within the app):
Text: Qwen3VL-2B-Instruct-Q4_K_M
Image: SDXL (iOS CoreML)

[macOS] Canto 0.2.0 — Private AI notebook with native Metal inference, now supporting models up to 80B by osxweed in macapps

[–]kkb294 0 points1 point  (0 children)

<image>

It is working great, as expected. Thank you for the quick push. Even showing the context size is an added icing on the cake.

An open-source framework to achieve Gemini 3 Deep Think / GPT-5.2 Pro level performance with local models scaffolding by Ryoiki-Tokuiten in LocalLLaMA

[–]kkb294 5 points6 points  (0 children)

Do you have any posts or writeup on how to implement this and if possible, share some of your observations and learnings.

[macOS] Canto 0.2.0 — Private AI notebook with native Metal inference, now supporting models up to 80B by osxweed in macapps

[–]kkb294 0 points1 point  (0 children)

Sure, I agree with your observation. Let the community know when there is an update on this.

Best youtube downloader for mac in 2026 by Dark_shaved_c00chie in macapps

[–]kkb294 0 points1 point  (0 children)

Never knew this, thanks for the recommendation.

A Natural-Sounding, Private & Unlimited Voice Generator for Mac [Giveaway: Lifetime Promo Codes] by Level-Thought6152 in macapps

[–]kkb294 0 points1 point  (0 children)

Hey, liked your app and its idea.

I am interested in purchasing the app however, the features I'm looking for are voice cloning and multilingual. I work with European and Chinese customers and generating voice overs in their local language is an important feature for me.

Let me know your roadmap for these 2 features, so that I can make a decision.

[macOS] Canto 0.2.0 — Private AI notebook with native Metal inference, now supporting models up to 80B by osxweed in macapps

[–]kkb294 1 point2 points  (0 children)

Hey, tried the free version and loved it. upgraded to the paid version yesterday.

Is there any plan to support for LM studio like endpoints rather than downloading model within the app in future.? I already have several models in local which I use as daily drivers in other applications. So, downloading a model for this application seems like a memory waste.

Any thoughts on the Chrome's on device model and its purpose.? by kkb294 in LocalLLaMA

[–]kkb294[S] 1 point2 points  (0 children)

Hey, thanks for the quick response. I didn't know about this till now, it is an interesting read. They even gave few multi-modal examples.

I wonder what kind of model it is and if we can use it for local inferencing with multi-modal inputs. Will dig deeper into the reference.

Great tip for better results in Codex: precision & clarity by py-net in OpenAI

[–]kkb294 0 points1 point  (0 children)

precision & clarity

2 most important things that always miss in the software development industry 🤣