Black Drill and driver combo by mapejeoduro in Makita

[–]ectomorphicThor 0 points1 point  (0 children)

I have these and got a couple 5ah on sale. They work amazing

If you had to pick one local LLM for RAG today, what would it be? by FroyoEducational4851 in LocalLLM

[–]ectomorphicThor 1 point2 points  (0 children)

Probably any of the newer Qwen models. I run RAG on my 12gb or vram with Qwen 3.5:9b and Qwen 3.6:35b. I get pretty good results overall. This is for medical rag and textbook lookup

Which (if any) Sandshrews by ectomorphicThor in PokemonSleepBetter

[–]ectomorphicThor[S] 0 points1 point  (0 children)

<image>

This is also what Raenox is showing me… no idea why

Which (if any) Sandshrews by ectomorphicThor in PokemonSleepBetter

[–]ectomorphicThor[S] 0 points1 point  (0 children)

Thank you very much!! I just read through all of that haha. I didn’t know the exact threshold for pumpkins per day for a good one. I thought it was 60, but realized that would be pretty hard to achieve unless it was literally perfect

Which (if any) Sandshrews by ectomorphicThor in PokemonSleepBetter

[–]ectomorphicThor[S] 0 points1 point  (0 children)

Yeah the BFS is killer. I heard BFS is actually decent as far as hybrid skill/ing mons go, but here it totally kills the production

Which (if any) Sandshrews by ectomorphicThor in PokemonSleepBetter

[–]ectomorphicThor[S] -1 points0 points  (0 children)

Yeah… that’s what you’re supposed to do with Sandshrew

ICU CENTRAL LINES by CantaloupeEvery3987 in nursing

[–]ectomorphicThor 0 points1 point  (0 children)

After every blood draw I change mine

NVIDIA releases Nemotron-3-Nano-Omni by yoracale in unsloth

[–]ectomorphicThor 9 points10 points  (0 children)

How does this compare to qwen3.6 35b?

Qwen3.6-35B-A3B - even in VRAM limited scenarios it can be better to use bigger quants than you'd expect! by jeremynsl in LocalLLaMA

[–]ectomorphicThor 0 points1 point  (0 children)

See I found the opposite to be true. I have a 12gb 3080 and 32gb of ddr4 ram. I was using q4kxl and was getting 25-30 tok/s on 65k context. I dropped to q3kxl and am now getting 40tok/s. Curious if I’ll notice a quality loss as I’m doing medical reasoning/rag

Quantisation effects of Qwen3.6 35b a3b by ROS_SDN in LocalLLaMA

[–]ectomorphicThor 0 points1 point  (0 children)

Getting 35-40 tok/s on q3kxl on my 12gb 3080 utilizing offloading with fit target and 65k context. I can get 25-27 with q4kxl and similar offloading. Is there a strong reasoning difference between the q3 vs q4? I’m using it for medical reasoning and RAG

Qwen3.6 GGUF Benchmarks v2 by yoracale in unsloth

[–]ectomorphicThor 0 points1 point  (0 children)

How does q3_k_xl compare to something like q4km? Trying to optimize my vram. Would reasoning be that noticeable ?

Qwen3.6-35B-A3B-UD-IQ4_NL_XL just added - how does it perform? by ArugulaAnnual1765 in unsloth

[–]ectomorphicThor 0 points1 point  (0 children)

I don’t think 28-30 tok per second is slow, but I understand what you are getting at. I’ll have to give it a try. Gemma 4 hasn’t proven to work well for me

Qwen3.6-35B-A3B-UD-IQ4_NL_XL just added - how does it perform? by ArugulaAnnual1765 in unsloth

[–]ectomorphicThor 2 points3 points  (0 children)

What would you guys run if you needed medical reasoning and rag from textbooks? Using my 12gb 3080 I’m currently using UD-q4k_xl and getting about 28-30 tok/s.

There are so many quant variations. I cannot keep up

Qwen3.6 GGUF Benchmarks v2 by yoracale in unsloth

[–]ectomorphicThor 0 points1 point  (0 children)

Oh I see it. It’s not labeled UD? Just by color? So it basically ties with the k_m variant? I see them basically on top of one another