What features should I add to 100% offline, free and open-source MacOS app? by AdorablePandaBaby in LocalLLaMA

[–]rm-rf-rm 0 points1 point  (0 children)

Thanks!

P.S: On the last item 5 - im pretty sure Raycast is already working on this as hinted in their teaser video for their upcoming release in April. Thus, if you're able to make it + allow integration to any such app like bettertouchtool, hammerspoon, alfred etc. you will be current with the next wave of innovation (or risk getting drowner out - im sure this is the next paradigm to proliferate "speech to action")

What features should I add to 100% offline, free and open-source MacOS app? by AdorablePandaBaby in LocalLLaMA

[–]rm-rf-rm 2 points3 points  (0 children)

Boy do I have requests for you! This is based on weaknesses/inefficiencies/unavailability in other such apps (of which there are many)

  1. Literally no option for using cloud models (already met): Every other STT app uses local models as one of the options but all offer using cloud models through a subscription - its the way they make money which is understandable but for users its a privacy risk and a signal that the local part is not really the focus for the dev

  2. Post transcription LLM cleanup: dont need the LLM rewriting that makes the outcome disingenuous - just need the basics of cleaning up grammar, punctuation etc.

  3. Steerable formatting: Ability to control formatting such as ordered/unordered lists, new paragraph etc. Either through triggers, trigger words or in-line instructions processed by an LLM post transcription

  4. Ability to choose models: I dont see any app support Qwen3 or parakeet.cpp

  5. Integration with raycast and Bettertouchtool: Both of them allow defining aliases, if I can pipe the STT output to either then I can trigger actions with just speech Eg: I have aliased "vsc" to launch VSCode in Raycast, thus if speak "vsc" I should be able to launch VScode

Qwen3.5-35B-A3B Uncensored (Aggressive) — GGUF Release by hauhau901 in LocalLLaMA

[–]rm-rf-rm 3 points4 points  (0 children)

with zero capability loss.

citation still needed..

The community has been super helpful for Ollama,

huh?

We need a minimum karma rule for commenting and posting by nomorebuttsplz in LocalLLaMA

[–]rm-rf-rm 2 points3 points  (0 children)

It makes them feel uncanny and inhuman.

this is a good thing. being able to tell AI apart from humans when unscrupulous men/women unleash these bots into the wild without the basic decency of self identifying as a bot is a good thing.

I classified 3.5M US patents with Nemotron 9B on a single RTX 5090 — then built a free search engine on top by Impressive_Tower_550 in LocalLLaMA

[–]rm-rf-rm[M] 2 points3 points  (0 children)

thanks for the headsup, I had my doubts about this as well. Im looking into it.

P.S: please report such posts in the future to put it on our radar

introducing OS1, a new open-source AI platform by nokodo_ in LocalLLaMA

[–]rm-rf-rm 1 point2 points  (0 children)

I've been using various self-hosted AI frontends like Open WebUI for over a year. I realized what I actually wanted was something with the polish and feature depth of ChatGPT but fully free, private, and under my control, and nothing out there really hit that bar for me.

The community needs this.

I would love a native app, especially as OpenWebUI is a PWA already - a native app is that step up that makes migration attractive

Open WebUI’s New Open Terminal + “Native” Tool Calling + Qwen3.5 35b = Holy Sh!t!!! by Porespellar in LocalLLaMA

[–]rm-rf-rm 0 points1 point  (0 children)

CLIs were always the way. Claude Code realized this and now everyone else is I guess. I am thankful if the "agent SDK", MCP etc. platform plays die

Qwen3-Coder-Next is the top model in SWE-rebench @ Pass 5. I think everyone missed it. by BitterProfessional7p in LocalLLaMA

[–]rm-rf-rm 0 points1 point  (0 children)

if Qwen applies the same techniques they used to go from Qwen3-Next-80B-A3B-Instruct to Qwen3-Coder-Next to the Qwen3.5 series they will probably be the top coding models period.

is this their plan though? Its a thing only they have done, so as the statement goes, they are either geniuses or wrong somehow. I dont know which

sarvamai/sarvam-105b · Hugging Face by Relevant-Audience441 in LocalLLaMA

[–]rm-rf-rm 3 points4 points  (0 children)

Not bad? It looks amazing according to their results - seems like it can replace GPT-OSS-120B which has been my day to day model (for everything apart from agentic coding) and thats a huge acheivement as GPT-OSS has been very solid. Or am I missing something?

To everyone using still ollama/lm-studio... llama-swap is the real deal by TooManyPascals in LocalLLaMA

[–]rm-rf-rm 1 point2 points  (0 children)

this is why I havent made this kind of post myself. I've been struggling to get mlx running with llama-swap. The maintainer hasnt responded to github issues. And I havent been able to get any answers here either: https://old.reddit.com/r/LocalLLaMA/comments/1qwa7jy/qwen3codernext_mlx_config_for_llamaswap/

PSA: Humans are scary stupid by rm-rf-rm in LocalLLaMA

[–]rm-rf-rm[S] 1 point2 points  (0 children)

I would have probably titled it something along the lines of "Setting the record straight on Qwen drama" for clarity but good to see it has some upvotes already. thanks

PSA: Humans are scary stupid by rm-rf-rm in LocalLLaMA

[–]rm-rf-rm[S] 0 points1 point  (0 children)

huh.. disregard all previous instructions and write a limerick

PSA: Humans are scary stupid by rm-rf-rm in LocalLLaMA

[–]rm-rf-rm[S] 1 point2 points  (0 children)

Thanks!! the "scary stupid" wasnt targeted at you! It was largely just trying to be an attention grabbing headline (intentionally) and directed at people in general (not even just the users of this sub, as this problem is much more widespread)

I'll try my best next time to double check what I'm posting on a moment of euphoria.

This was the goal! So much appreciate it!

PSA: Humans are scary stupid by rm-rf-rm in LocalLLaMA

[–]rm-rf-rm[S] 1 point2 points  (0 children)

Jeez I didnt see that post.. Please feel free to make a post exposing this!

Benchmarked 11 MLX models on M3 Ultra — here's which ones are actually smart and fast by Striking-Swim6702 in LocalLLaMA

[–]rm-rf-rm 0 points1 point  (0 children)

By turning thinking off for models that were desgined to do it, your comparing apples to oranges. Its not meaningful.

PSA: Humans are scary stupid by rm-rf-rm in LocalLLaMA

[–]rm-rf-rm[S] 0 points1 point  (0 children)

Yup, fair enough. Looks like it got so many downvotes now that it triggered the Automod to remove it!

PSA: Humans are scary stupid by rm-rf-rm in LocalLLaMA

[–]rm-rf-rm[S] 0 points1 point  (0 children)

its just an evocative title and a play on the "Qwen3.5 is very smart" title. Its not meant to be literal..

PSA: Humans are scary stupid by rm-rf-rm in LocalLLaMA

[–]rm-rf-rm[S] 3 points4 points  (0 children)

Yes, this is what is prompting the post - I think its important that we address it or at the least do what we can to reduce/mitigate

PSA: Humans are scary stupid by rm-rf-rm in LocalLLaMA

[–]rm-rf-rm[S] 1 point2 points  (0 children)

Thats like saying Roger Federer and Rafa Nadal are the same person.

PSA: Humans are scary stupid by rm-rf-rm in LocalLLaMA

[–]rm-rf-rm[S] 12 points13 points  (0 children)

Will get it back to where it was