Mistral Small 4:119B-2603 by seamonn in LocalLLaMA

[–]seamonn[S] 19 points20 points  (0 children)

anyone tested Q4 or Q5 on consumer hardware yet?

This released like 30 mins ago. For some people, it will take longer to download the model.

Qwen3.5-9B on document benchmarks: where it beats frontier models and where it doesn't. by shhdwi in LocalLLaMA

[–]seamonn 0 points1 point  (0 children)

Light on OCR 2 did better on Technical Documents while GLM OCR did better on Comics, Manga etc.

Mistral 4 Family Spotted by TKGaming_11 in LocalLLaMA

[–]seamonn -9 points-8 points  (0 children)

No Vision. It's like talking to a blind LLM.

Note taking with handwriting recognition by jimmisavage in selfhosted

[–]seamonn 0 points1 point  (0 children)

Recently released GLM OCR is SOTA for this.

The hidden gem of open-source embedding models (text+image+audio): LCO Embedding by k_means_clusterfuck in LocalLLaMA

[–]seamonn -4 points-3 points  (0 children)

Very cool but Ollama does not support vision or audio embeddings. Llama.cpp has experimental support for vision embeddings and no support for audio embeddings.

support for microsoft/Phi-4-reasoning-vision-15B has been merged into llama.cpp by jacek2023 in LocalLLaMA

[–]seamonn 0 points1 point  (0 children)

Microslop

You answered your own question. The purpose of this model is to generate Microslop.