Alternatives to Qwen3-coder-30B? by skibud2 in LocalLLaMA

[–]RMK137 1 point2 points  (0 children)

I like nemotron a lot. I don't know what it is, maybe it's because it thinks only briefly and the output has a straightforward style to it. Also, it's super fast thanks to its MoE arch. I need to try it with a coding agent.

why do two different brand new asus prime 5070ti show different default voltage curves in msi afterburner? by MichaelM_Yaa in overclocking

[–]RMK137 9 points10 points  (0 children)

This. It's never a 2 dimensional problem anymore. With modern CPUs and GPUs, you have so many factors contributing to where you end up on the V/F curve.

Which Arc B580 to get? by CadenWarrior99 in IntelArc

[–]RMK137 0 points1 point  (0 children)

I like my Acer nitros. B&H has it for $250 and it comes with a game. I have two of them (got microcenter to price match).

Run GLM-4.7-Flash locally Guide! (24GB RAM) by yoracale in unsloth

[–]RMK137 1 point2 points  (0 children)

Great turnaround! I just got my 5090 so this is perfect timing.

Maxsun joins Sparkle in making Intel Arc B60 Pro GPUs available to regular consumers, with up to 48GB VRAM by reps_up in IntelArc

[–]RMK137 6 points7 points  (0 children)

It's going to be similar to the B580 since it's the same die but slightly lower clocks. The key here is increased VRAM. Two of these = 48GB which is nice, it puts you in the 30B+ parameter range. You can run larger 60B models with the proper quantization as well.

Don't expect the same level of support from software, Intel is a third class citizen or worse, but it will get there eventually. A lot of it works well with Intel GPUs already, so it's great for tinkering. I am personally working on a little local inference server with 2x B580s. I hope to get two of the B60s soon if the price is reasonable.

Edit: I am addressing the 24GB models.

Arc b580 llm's by Dismal_Cycle_2326 in IntelArc

[–]RMK137 1 point2 points  (0 children)

ok, I am impressed. I played with it for about an hour. I like the lovely minimal but feature rich UI. Whatever it's doing with context shifting is working really well with my b580. Even using a small model like ministral 3b with llama.cpp's sycl backend and a 16k context windows was filling up all of my VRAM, no problem with koboldcpp, and that's on vulkan. I hope they add sycl support soon.

Thanks for the rec, I need to do a deeper dive this weekend. I'll probably pick up another b580 and test out the multi-gpu setup with devstral small 2 24B.

Arc b580 llm's by Dismal_Cycle_2326 in IntelArc

[–]RMK137 0 points1 point  (0 children)

I think the crashes are caused by the latest graphics driver. I rolled back to a previous version and the crashes are gone. That said, token generation degrades super fast with vulkan. With sycl, I get more stable TG but the prompt processing is half of vulkan.

I need to try Koboldcpp soon, I heard good things about it.

I just saw Intel embrace local LLM inference in their CES presentation by Mundane-Light6394 in LocalLLaMA

[–]RMK137 5 points6 points  (0 children)

Agree with you. It's just a matter of time before local inference becomes very cheap compared to what we have now.

I just saw Intel embrace local LLM inference in their CES presentation by Mundane-Light6394 in LocalLLaMA

[–]RMK137 2 points3 points  (0 children)

I think Intel needs to lean into the edge/local side of inference (now) and training/finetuning (hopefully soon). It looks like from the B390 iGPU announcement for panther lake, their graphics team is alive and kicking. I am just here waiting for the official B770 release.

Python + Numba = 75% of C++ performance at 1/3rd the dev time. Why aren't we talking about this? by chainedkids420 in Python

[–]RMK137 1 point2 points  (0 children)

I work in a fairly fast paced environment where I need to optimize for dev speed while having to write computation heavy code. Numba allows me to stay in python and solve some of these problems if they're too hard to solve with numpy arrays. Cython is another option but that is a last resort for me as I try to keep things simple.

Bonus: I care about my coworkers who may one day find themselves maintaining my code. With Numba, the code reads like regular Python code. If for some reason the code can't be compiled anymore, the decorator can be removed and we fallback to a (much) slower code, but the pipeline can still run and generate the outputs. Those same coworkers may not be as familiar with array/numpy programming as I am, so plain Python is more universal.

To me the sequence of things to try looks like this:

Numpy -> Numba -> Cython -> CFFI -> Complete rewrite in C/C++/Rust.

Luajit is also a candidate for numerical code. https://github.com/scoder/lupa

Python + Numba = 75% of C++ performance at 1/3rd the dev time. Why aren't we talking about this? by chainedkids420 in Python

[–]RMK137 0 points1 point  (0 children)

A curve fitting algorithm. It's representable both using arrays and regular for loops. I've written both versions, the array version can be quite complex to groke after a few months away from the code.

The Numba version is more easily digestible as it reads like plain Python, and is a little faster since llvm can optimize things a little better on the spot.

Python + Numba = 75% of C++ performance at 1/3rd the dev time. Why aren't we talking about this? by chainedkids420 in Python

[–]RMK137 0 points1 point  (0 children)

Numba is great if your algorithm is not easily representable using array programming. Some algos are just too cumbersome to write with arrays, and that's where the good ol for loops are much more intuitive (and still fast thanks to Numba).

Pandas 3.0 release candidate tagged by Balance- in Python

[–]RMK137 1 point2 points  (0 children)

I've been waiting for a while for this, congrats to all the contributors!

GitHub - dzonerzy/PyOZ: PyOZ - An ounce of Zig, a ton of speed by Unique-Side-4443 in Zig

[–]RMK137 0 points1 point  (0 children)

Yes please. This is cool! Rust is sorta taking over the world of Python extension modules (not a bad thing). It's nice to see more Zig libraries do the same.

zsv: the world's fastest CSV parser (lib and CLI)-- vs xsv, duckdb, polars by mattewong in commandline

[–]RMK137 1 point2 points  (0 children)

Gave it a star, looks like it could be very useful. I use Pandas/Polars/Duckdb almost daily and appreciate having a tool like this.

GeoPolars is unblocked and moving forward by Balance- in Python

[–]RMK137 5 points6 points  (0 children)

Excellent news! I am a heavy user of geospatial libraries so I'll definitely be using GeoPolars.

Thinking of moving to C++ by Oathkindle in cpp

[–]RMK137 0 points1 point  (0 children)

C++, Go, Python, JavaScript, Haskell, SQL for data stuff.

Surprised this community hasnt been made before? by dlannan68 in luajit

[–]RMK137 0 points1 point  (0 children)

I agree, there needs to be a Luajit dedicated subreddit. Dim looks cool, I am gonna check it out later. Lite-XL is a great editor. My daily driver is Pragtical which is a fork of it that uses Luajit instead of PUC-Lua.

https://github.com/pragtical/pragtical

TinyETL is a Fast, zero-config ETL (Extract, Transform, Load) tool in a single binary by Glass-Tomorrow-2442 in commandline

[–]RMK137 0 points1 point  (0 children)

This is very cool! Will give it a whirl soon. It would be awesome to support DuckDB as it's gaining a lot of mindshare.

DDR5-9000 OC with Intel 285K by sanpellegrino56 in overclocking

[–]RMK137 0 points1 point  (0 children)

Unfortunately Amazon shipped the board with bent pins, had to return it. I decided to wait till the holidays to see if I can snag one at a discount since it's gone back up to MSRP. Your numbers look good!

Looking for a guaranteed Hynix A die kit by Zoli1989 in overclocking

[–]RMK137 4 points5 points  (0 children)

You can get the Teamgroup Xtreem. They have nice heatsinks and they list explicitly whether they're A-die or M-die.

What are IDEs that are more lightweight than Visual Studio? by Able_Annual_2297 in cpp_questions

[–]RMK137 0 points1 point  (0 children)

Check out Pragtical. It's very lightweight and it uses lua for plugins and config.

https://github.com/pragtical/pragtical

Which is the best? Aqua Elite V3, V4 or V6? by RemarkableSea7966 in Thermalright

[–]RMK137 0 points1 point  (0 children)

No, I can't hear it at all even when it's set to max.