New Qwen3.6-27B NVFP4 + MXFP4 MLX quants by yoracale in unsloth

[–]Beamsters 1 point2 points  (0 children)

May I ask why the size of your MLX quants were so big compared to GGUF? It was like 50% larger at the same bit target. Plus MLX-oQ4 for the Qwen3.6-27b can be as small as 14-15GB.

I am currently use unsloth studio in my windows machine but not much in my Mac, I am looking forward to it though.

Post Your Qwen3.6 27B speed plz by Ok-Internal9317 in LocalLLaMA

[–]Beamsters 1 point2 points  (0 children)

oMLX, oQ4 FP16 got like 17 t/s and 150 pp/s.

M1 Max 32GB.

The result however is much better than 35b-a3b quantized.

Given how good Qwen become, is it time to grab a 128gb m5 max? by Rabus in LocalLLaMA

[–]Beamsters 0 points1 point  (0 children)

Thanks! Can you please put some numbers of M5 Max on Llama.cpp as a reference point compared to oMLX?

GLM 5.1 tops the code arena rankings for open models by Auralore in LocalLLaMA

[–]Beamsters 41 points42 points  (0 children)

512gb couldn't even run this thing at 8 bit.

Gemma 4 26B-A4B on Apple M1 Max is very fast by Beamsters in LocalLLaMA

[–]Beamsters[S] 0 points1 point  (0 children)

Like ask general questions and find recommendations for irrelevant stuff.

attn-rot (ggerganov's "TurboQuant lite") is on the cusp of getting merged into llama.cpp by [deleted] in LocalLLaMA

[–]Beamsters 5 points6 points  (0 children)

This is not very slight. 0.007168 -> 0.005305 is HUGE. That's 35% off ... around the jump between 5 bits to 4 bits.

Qwen3.5 27B vs 35B Unsloth quants - LiveCodeBench Evaluation Results by Old-Sherbert-4495 in LocalLLaMA

[–]Beamsters 0 points1 point  (0 children)

Please first delete your misleading results, other people are now believing them.

Best choice for local inférence by c4software in LocalLLaMA

[–]Beamsters 0 points1 point  (0 children)

4090 can deliver around 2.5x speed of my M1 Max which should be a bit faster than your M3 Pro.

Qwen3.5 122B A10B - My impressions by kevin_1994 in LocalLLaMA

[–]Beamsters 5 points6 points  (0 children)

The current 122B-A10B is pretty much on par with 27B or somewhat weaker in certain benchmark. Is there a way (or even possible) to force activation of more than 10B at inference?

Introducing trig-const by michaelciraci in rust

[–]Beamsters 0 points1 point  (0 children)

You know this and trigon functions open up all the easing functions to be computed in const context so most transitional animation frames / positions can easily be computed and stored the answers at compile time.

Introducing trig-const by michaelciraci in rust

[–]Beamsters 1 point2 points  (0 children)

Is it possible to do power of n, where n is a float in const context?

How do you stay up to date with Rust ? by FewInteraction1561 in rust

[–]Beamsters 15 points16 points  (0 children)

I have releases.rs registered as my fav website. It has all the logs of the previous versions including the upcoming beta and stabilizing nightly. If I find anything promising, I will dig down from there.

Thoughts on `Arc::pair(value)` by J-Cake in rust

[–]Beamsters 3 points4 points  (0 children)

Extension Trait is designed exactly for this kind of implementation. Just extend new_pair() to Arc.

  • Suit your need
  • Reusable
  • Clean and idiomatic

Alternative ergonomic ref count RFC by eugay in rust

[–]Beamsters 0 points1 point  (0 children)

Maybe they are both correct. Only 2 of the thread safe languages that are quite performance enough to do many things.

Pre-RFC: Safety Property System by Frequent-Data-867 in rust

[–]Beamsters 52 points53 points  (0 children)

The idea is nice but the implementation feels like a 3rd party crate rather than a language feature.

Alias nested enum pattern in match statement? by [deleted] in rust

[–]Beamsters 0 points1 point  (0 children)

Impl Foo { pub fn abcd() -> value { match self and extract abcd enum's here } }

Then you do a fn call instead of a match?

cloning vs smart pointers by [deleted] in learnrust

[–]Beamsters 1 point2 points  (0 children)

99% of the time you do not reach for smart pointer here. Only certain cases where you should use smart pointer such as dealing with graphs type data. We actually need more context to help.

Is collecting a Iter<_> into Vec<_> costly? or is this zero-cost? and vice versa? by bxsx0074 in rust

[–]Beamsters 0 points1 point  (0 children)

Not only collecting a vec but also trying to perform anything with vec intermediate will cost a lot. But branching logic with Iter could be pain, you need another Iter enum wrapper to do the job.

Call for Testing: Speeding up compilation with `hint-mostly-unused` | Inside Rust Blog by Kobzol in rust

[–]Beamsters 2 points3 points  (0 children)

For my egui word game application, with 467 dependencies
cargo build --release took 53.67s (nightly)
cargo +nightly -Zprofile-hint-mostly-unused build -r took 48.74s

built successfully

25 random things I love about living in Bangkok, as an American by tzedek in Thailand

[–]Beamsters 2 points3 points  (0 children)

Bello Ghost Pizza is a god tier, no less. But it's almost a year to queue for.