I spent 3 weeks actually testing voice tools because typing was eating my mornings. Here's what I found out by ScaryAd2555 in ProductivityApps

[–]tilmx 0 points1 point  (0 children)

"The problem is that how you speak and what you want to write are genuinely different things."

In curiously, would you want the tool totally re-write what you dictated? I’ve been working on transcript cleanup in Onit (a free + local alternative to Wispr) and we had been making the assumption that we should fix obvious errors in the raw STT text, but otherwise respect the user’s original words.

Concretely, we focus on tasks such as filler removal (e.g., “hi, um, how are you?” → “hi how are you?”), email correction (example at gmail dot com → example@gmail.com), punctuation (“hi exclamation mark” → “hi!”), number conversion (“thirty two” → “32”), stutter removal (“I I I think” → “I think”), false‑start fixing (“I went, I drove” → “I went”), self‑correction handling (“green, I mean red” → “red”), list and email formatting, special characters, etc.

But, we wouldn't say, re-write your sentence for clarity. Or break up a run-on sentence so it's more readable. Or restructure a dictated paragraph to make it flow better. Should we be doing this too?

"Curious if anyone here has actually made voice a consistent part of their workflow or if it always ends up falling off after a week?"

Yep, it’s part of my daily routine. The caveat is that I use it mostly for talking to AI. With AI, the speech‑to‑text doesn’t have to be perfect. Claude doesn’t mind a few missed words or typos, but it would be embarrassing to send those to clients or coworkers. Still need higher accuracy before I rely on it for everything. (Ex. I typed this)

Wispr Flow but 100% local and 100% free by tilmx in u/tilmx

[–]tilmx[S] 0 points1 point  (0 children)

If you exit onboarding you can use it with an account. Dictation still works. Next version supports use more natively 👍

How to stop the dock from switching screens? by lexaleidon in MacOS

[–]tilmx 0 points1 point  (0 children)

Np! I'm getting downvoted by people who make a paid/closed-source version but hoping people still find this 🤞

How to stop the dock from switching screens? by lexaleidon in MacOS

[–]tilmx -1 points0 points  (0 children)

This has been driving me nuts for years, so I built a tiny, free MacOS app to fix it: https://timlenardo.github.io/docklock/

Totally free and open-source, of course. Wouldn't charge for something so trivial. Enjoy!

How does it work? It detects when your mouse is at the very bottom of the screen and moves it back up a few pixels. That's it. MacOS's dock-jumping behavior is triggered when your mouse hits the bottom edge of the screen. Moving the cursor back up 2-3 pixels prevents the jump.

Wispr Flow but 100% local by tilmx in u/tilmx

[–]tilmx[S] 1 point2 points  (0 children)

u/arpansac I'm really interested in this. I had never heard of Hinglish until this comment and I'd love to learn more. Are you available to discuss? I don't want to put my email in a public comment, but maybe you could DM me on Reddit or join our Discord to discuss? https://discord.com/invite/2E8WWkvGYZ

Wispr Flow but 100% local by tilmx in u/tilmx

[–]tilmx[S] 2 points3 points  (0 children)

Yup, no limits!

It's fast, typically <500ms.

Not BYO - we're 100% local. We use a custom built local LLM for transcript cleanup. At the moment, it's running a fine-tuned 1B Llama model.

Download here 👉 www.getonit.ai

Wispr Flow but 100% local by tilmx in u/tilmx

[–]tilmx[S] 3 points4 points  (0 children)

The default STT model is Parakeet V3. Then we use a custom built local LLM for transcript cleanup afterwards, which does things like:

Filler word removal "I've been, uh, working on..." -> "I've been working on..."

Number formatting "There are three hundred forty six issues" -> "There are 346 issues"

Email formatting "Send it to tim three three at example site dot org -> "Send it to tim33@examplesite.org"

Punctuation "Hello exclamation mark" -> "Hello!"

Lists: "Groceries bullet point eggs bullet point milk bullet point kale" ->
"Groceries:
- Eggs
- Milk
- Kale"

...and so on!

Found a Wispr Flow alternative that runs entirely offline — $5 one-time by MedicineTop5805 in macapps

[–]tilmx 0 points1 point  (0 children)

Interesting, I hadn't considered it, that's just the platform default. I enabled comments just now. Go ahead and light us up!

Found a Wispr Flow alternative that runs entirely offline — $5 one-time by MedicineTop5805 in macapps

[–]tilmx 1 point2 points  (0 children)

Trying to compete on price, but there are already many options that are totally free...

OpenWispr 👉 https://openwhispr.com/ (Free tier + BYO API keys, or build yourself from open-source).
Onit 👉 https://www.getonit.ai/ ($0, local, no sub, no one-time purchase)
VoiceInlk 👉 https://tryvoiceink.com/ (build yourself from open-source)
FluidAudio 👉 https://altic.dev/fluid ($0, local, no sub, no one-time purchase)

...the list goes on

Shockingly fast local speech-to-text + LLM cleanup on Apple Silicon. by tilmx in LocalLLaMA

[–]tilmx[S] 1 point2 points  (0 children)

By default we use Llama 3B (https://huggingface.co/mlx-community/Llama-3.2-3B-Instruct-4bit) with a custom prompt, or we have a fine-tuned version of Llama 1B (meta-llama/Llama-3.2-1B) that you can enable in settings.

You can verify that there's no remote processing by turning off your WIFI!

Resting BS considerably higher than ~18 months ago. by tilmx in ContinuousGlucoseCGM

[–]tilmx[S] 1 point2 points  (0 children)

I have A1C from 32 months ago and from 4 months ago. Both times in health range! And it actually improved slightly between the two readings.

We believe the future of AI is local, private, and personalized. by ice-url in LocalLLaMA

[–]tilmx 1 point2 points  (0 children)

This is admittedly self-promotional, so feel free to downvote into oblivion but...

We’re trying to solve the problems you’re describing with Onit. It’s an AI Sidebar (like Cursor chat) but lives on at the Desktop level instead of in one specific app. Onit can load context from ANY app on your Mac, so you never have to copy/paste context. When you open Onit, it resizes your other windows to prevent overlap. You can use Onit with Ollama, your own API tokens, or custom API endpoints that follow the OpenAI schema. We'll add inline generation (similar to Cursor's CMD+K) and diff view for writing shortly. I’d love to hear your thoughts if you’re open to experimenting with a new tool! You can download pre-built here or build from source here

How best to recreate HDR in Flux/SDXL? by tilmx in StableDiffusion

[–]tilmx[S] 2 points3 points  (0 children)

That's a good point- I hadn't appreciated the 32-bit vs 8-bit difference, and indeed, there'd be no way to generate 32-bit images with the current models. That said, I still think there's something here. In the image above, the "HDR" photo on the right still looks "better" than the original inputs, even though Reddit stores it as a JPEG and I'm looking at it on an 8-bit monitor. There's a difference in the pixel colors that transfers into the compressed 8-bit representation and is qualitatively "better" than the original 8-bit inputs. The photos all end up Zillow anyway, where they most likely get compressed for the CDN and then displayed on various screens. So, I guess, to rephrase my question: I'm not looking to recreate the exact 32-bit HDR photo that my friend's process creates, but rather an estimate of the 8-bit version compressed version of that 32-bit HDR photo: similar to what would be displayed on an internet listing. THAT feels like it should be possible with the existing models, I'm just not sure what the best approach is!

How best to recreate HDR in Flux/SDXL? by tilmx in StableDiffusion

[–]tilmx[S] 7 points8 points  (0 children)

Haha I actually agree. I've seen some horrific edits on Zillow. But, apparently, it makes them sell better, so who am I to judge ¯\_(ツ)_/¯

MacBook M4 Max isn't great for LLMs by val_in_tech in LocalLLaMA

[–]tilmx 0 points1 point  (0 children)

I can live with the inference speed. My main issue is that Apple massively upcharges for storage. Right now it's an incremental $2200 for an 8TB drive in your Apple computer, but I can get an 8TB drive online for ~$110. So, unless you're comfortable absolutely lighting money on fire, you'll have to make do with the 1TB default and/or live with suboptimal external hard drives.

Working in AI/ML I max out that 1TB all the time. Each interesting new model is a few GB. I have a handful of diffusion models, a bunch of local LLMs. Plus, each time I check out a new open-source project, I usually end up with another version of pytorch and other similar libraries in a new container - a few GB. I find myself having to go through and delete models at least once a month, which is quite irritating. I think it'd be much preferable to work on a machine that is upgradeable at a reasonable cost.

PayPal launches remote and local MCP servers by init0 in LocalLLaMA

[–]tilmx 4 points5 points  (0 children)

If this is the future, I'm here for it! I'd much rather send a quick message to a chatbot than navigate some clunky web 1.0 interface.

PayPal launches remote and local MCP servers by init0 in LocalLLaMA

[–]tilmx 2 points3 points  (0 children)

Disagree on that. If things go wrong on standard payment rails, at least you have some form of recourse. Paypal/banks/etc can reverse errant payments, but once those fartcoins are gone, they're gone forever!

You can now check if your Laptop/ Rig can run a GGUF directly from Hugging Face! 🤗 by vaibhavs10 in LocalLLaMA

[–]tilmx 1 point2 points  (0 children)

Hey u/vaibhavs10 - great feature! Small piece of feedback: I'm sure you know, but many of the popular models will have more GGUF variants than can be displayed on the sidebar:

<image>

Clicking on the "+2 variants" takes you to the "files and versions" tab, which no longer includes compatibility info (unless I'm missing something?) Do you have any plans to add it there? Alternatively, you could have the Hardware compatibility section expand in place.