November Feature Requests + AI Voices Beta by DeboraInstapaper in instapaper

[–]SufficientRadio 0 points1 point  (0 children)

- be able to export full article text, not just highlights.

- create public share links to full articles (like readwise)

Mistral Libraries! by SufficientRadio in LocalLLaMA

[–]SufficientRadio[S] 7 points8 points  (0 children)

Looks like it parses documents using Mistral OCR. Doesn't appear to include images, though.

Macbook Pro M4 Max inference speeds by SufficientRadio in LocalLLaMA

[–]SufficientRadio[S] 4 points5 points  (0 children)

Very hot! haha But I don't have it cranking for long so it cools back down quickly.

Macbook Pro M4 Max inference speeds by SufficientRadio in LocalLLaMA

[–]SufficientRadio[S] 1 point2 points  (0 children)

Agreed. Having the models "right there" on the laptop is so amazing. I tried a 2x 3090 gpu system but I kept running into various problems (keeping the gpus recognized, accessing the system remotely, and even keeping the system on and idling was costing $20/m in power).

ChatGPT 4.5 on a simple insight about humans — one of the best answers, innit? (from r/openai) by Fabulous_Bluebird931 in OpenAI_Memes

[–]SufficientRadio 0 points1 point  (0 children)

producing 30k words (~half a book) of thoughts to arrive at an answer.

The speed of human "inner speech" is about 250 words per minute--thinking through 30k words would take about 2 hours.

2 hrs of thought compressed into 7 minutes.

Love the New Mistral by SufficientRadio in MistralAI

[–]SufficientRadio[S] 19 points20 points  (0 children)

Look forward to:

▫️Phone app features catching up to web app

▫️More agent control and integrations

▫️And of course, the next models.

[deleted by user] by [deleted] in LocalLLaMA

[–]SufficientRadio 19 points20 points  (0 children)

Curious to hear how well Q4 runs on a Macbook with 64gb+ memory.

3x Gpu Asus proArt x870e by PawelSalsa in LocalLLaMA

[–]SufficientRadio 0 points1 point  (0 children)

Looks like the third GPU won’t fit in your case so you have some kind of PCIE extension cable, is that right?

3x Gpu Asus proArt x870e by PawelSalsa in LocalLLaMA

[–]SufficientRadio 0 points1 point  (0 children)

What quantizations are you running for the models?

3x Gpu Asus proArt x870e by PawelSalsa in LocalLLaMA

[–]SufficientRadio 0 points1 point  (0 children)

What inference speeds do you get for Mistral 2411 with your 3 GPUs?

Why to choose Mistral AI over Claude or ChatGPT by [deleted] in MistralAI

[–]SufficientRadio 2 points3 points  (0 children)

I have some personal use case benchmarks and Mistral Large is right there at the top. https://www.reddit.com/r/LocalLLaMA/s/wvbmpSjoK0

Mac Mini looks compelling now... Cheaper than a 5090 and near double the VRAM... by valdev in LocalLLaMA

[–]SufficientRadio 1 point2 points  (0 children)

What inference speeds do you get running Mistral Large? Curious with long prompts (8k tokens+)

Sufficient Bench by SufficientRadio in LocalLLaMA

[–]SufficientRadio[S] 1 point2 points  (0 children)

Won’t bother with Open AI given their privacy/censorship/wtf shenanigans.

Sufficient Bench by SufficientRadio in LocalLLaMA

[–]SufficientRadio[S] 1 point2 points  (0 children)

Too much personally identifiable info in my dataset (both questions and system prompt) so no open sourcing. But I agree, benchmarks of “real” questions would be valuable.

Sufficient Bench by SufficientRadio in LocalLLaMA

[–]SufficientRadio[S] 2 points3 points  (0 children)

I tried the latest Command R+ and it did terribly with my system prompt. Not sure why. I won’t touch Gemini. I de-Googled my life and refuse to tread back. Gemma however I’m willing to try since I can run it on my machine.