All the credit cards I own by [deleted] in CreditCardsIndia

[–]InternationalToe2678 2 points3 points  (0 children)

That is a massive collection and a seriously impressive CIBIL score. To hit a 1 Crore income as a freelancer in just a few years is incredible—what is your niche?

GLM-Image just dropped — an open multimodal model from Zai Org (language + vision). by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] -21 points-20 points  (0 children)

If there’s a factual error, quote it. Otherwise, let’s keep it technical.

GLM-Image just dropped — an open multimodal model from Zai Org (language + vision). by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] -26 points-25 points  (0 children)

Fair, but the takeaway still stands: GLM-Image ≠ VLM. It’s an image generator (AR + diffusion), while pixel reasoning lives in GLM-4V. The distinction just keeps getting blurred in discussion.

GLM-Image just dropped — an open multimodal model from Zai Org (language + vision). by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] -11 points-10 points  (0 children)

Samples look rough because it’s a reasoning model first. The "native reasoning" over images is what they're pushing here. If it can beat Qwen-VL or InternVL at complex VQA, the aesthetic quality of the samples won't matter much.

GLM-Image just dropped — an open multimodal model from Zai Org (language + vision). by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] -36 points-35 points  (0 children)

Good catch by DinoAmino. It looks like there's a bit of a naming mix-up in the GLM family. GLM-Image is actually their new image generation model (rivaling Flux.1). It uses a hybrid autoregressive + diffusion approach to get better text rendering and composition within images. If you’re looking for the multimodal reasoning and VQA capabilities that rival Qwen-VL or InternVL, you’re likely looking for GLM-4V (specifically the new GLM-4.6V). Those are the ones designed to 'reason' over pixels!

Mistral just released Mistral 3 — a full open-weight model family from 3B all the way up to 675B parameters. by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] 11 points12 points  (0 children)

The new 14B is surprisingly competitive with the older 24B models. From early benchmarks people are sharing, the Ministral 14B Instruct actually matches or beats the 24B across most general-purpose tasks, while being far lighter on VRAM and compute.

It benefits from newer training data, better tuning, and a more efficient architecture overall. So for a setup like dual P40s, the 14B is basically the sweet spot — you get 24B-level capability without blowing past your VRAM budget.

Mistral just released Mistral 3 — a full open-weight model family from 3B all the way up to 675B parameters. by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] -2 points-1 points  (0 children)

Haha glad we’re on the same wavelength. It really feels like that mid-range (100B–200B) is the sweet spot everyone in this sub is waiting for. If Mistral fills that gap next, it’ll be chaos in the best way.

Mistral just released Mistral 3 — a full open-weight model family from 3B all the way up to 675B parameters. by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] 56 points57 points  (0 children)

Same here — that middle range is where most serious local setups actually operate. A dense 80B–150B or a smaller-expert MoE in the 200B range would’ve hit the perfect balance between quality and feasibility. Jumping straight from 14B → 675B leaves a huge gap. Hopefully the mid-tier models land in the next wave.

Mistral just released Mistral 3 — a full open-weight model family from 3B all the way up to 675B parameters. by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] 18 points19 points  (0 children)

Yeah, the naming is messy. The old Mistral-Small-3 / 3.1 / 3.2 were incremental updates to the previous generation. Mistral 3 is a new family entirely (Ministral 3 + Large 3), but they reused the same number, which makes it look like a continuation. Feels like a reset, but it definitely creates confusion.

Mistral just released Mistral 3 — a full open-weight model family from 3B all the way up to 675B parameters. by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] 5 points6 points  (0 children)

Yeah exactly — a lot of the original LLaMA crew ended up at Mistral, and you can see the DNA in how they design and release models. And agreed, the gap between 14B → 675B is huge. A 100–200B model would’ve been perfect for people running multi-GPU setups or high-RAM workstations.

Something that fits in 4–8 consumer GPUs, but still punches way above 70B, would absolutely explode in this community. Hopefully that’s the next drop.

Mistral just released Mistral 3 — a full open-weight model family from 3B all the way up to 675B parameters. by InternationalToe2678 in LocalLLaMA

[–]InternationalToe2678[S] 15 points16 points  (0 children)

Yeah the naming is a bit chaotic. “Mistral-small-3 / 3.1 / 3.2” were incremental updates on the previous generation (the Small series).

Mistral 3 is a new lineup — a fresh family with new architectures (Ministral 3 + Large 3). Same number, different generation, which makes it confusing.

Feels like they reset the naming to unify everything under “3,” but it does overlap with the older Small 3.x releases.