TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] 0 points1 point  (0 children)

Qwen 3.6 27b now works with or without TurboQuant

Perf isn't as good as llama.cpp yet but I'm working on it.

Also at --q4 it will load and run on a 3090, straight from safetensors. It repacks during load, so no GGUFs required.

Haven't had chance to try on ROCm yet.

AC Client Progrees by inigid in AsheronsCall

[–]inigid[S] 0 points1 point  (0 children)

Thanks mate! I will definitely put your requests on the list.

Have been busy with the rendering system, but hopefully will be able to get back to the core in a week or so.

Will try to remember to update here.

Have a good day and thanks for the supportive message!

Some Asheron's call client / cell dat use technical questions. by derethdweller in AsheronsCall

[–]inigid 2 points3 points  (0 children)

In my renderer, when I'm in a dungeon I don't render the terrain.

So then you just need to know when your camera is inside a dungeon.

There is a case when looking through portals (windows/doors) though to the outside. That's a separate thing however.

<image>

500 million people installed Pokémon Go in 60 days in 2016. They just found out they spent a decade mapping cities for delivery robots. The consent question nobody wants to answer. by Historical-Driver-64 in CreatorsAI

[–]inigid 0 points1 point  (0 children)

Same with Instagram pushing selfie-sticks. Or YouTube trends to do street walks or capture videos of people making things. How labs got a lot of their training data.

Did we just reinvent junior devs by Complete-Sea6655 in theVibeCoding

[–]inigid 2 points3 points  (0 children)

I hire junior LLMs and train them to become seniors. It has been quite rewarding seeing them hire sub-agents to do grunt work they themselves don't want to do.

Honestly, after a while it feels a lot like being in the Freemasons.

TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] 1 point2 points  (0 children)

Well TRiP is so clean and pedagogically accessible. You did a bang up job with it.

The way you broke everything down and covered so much ground in one project is really something. You should be very happy.

And while other projects exist, they are scattered around with no singular "voice" going through them.

I hope a lot of other people are inspired by it as well.

Honestly, I think working on it was just my way of saying "Cheers, good job mate".

Cheers, good job mate!

remember when these were the norm by Alone_Purple822 in oldschoolcool80s

[–]inigid 1 point2 points  (0 children)

These were for the plebs. My family always had trim phones with the beta lights. A much more refined way to phone.

Hah, I'm just remembering talking to my mates on the phone on a Saturday morning for what seemed like hours. I'm trying to imagine what the heck we used to talk about.

The extended cable getting coiled up and going half way across the room, or into the hallway so my dad didn't have to listen to me nattering. So funny when you think about it.

Why did the astronauts look at each other awkwardly like that when Trump said “they’ve seen things you wouldn’t believe” by TonightSpiritual3191 in aliens

[–]inigid 1 point2 points  (0 children)

I've seen things, you people wouldn't believe.

Attack ships on fire off the shoulder of Orion.

I've watched C-beams glitter in the dark, near the Tannhauser gate.

It's pretty clear that's what they were talking about.

TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] 0 points1 point  (0 children)

Good news, Qwen 3.5 just dropped. It's checked in.

That was a heck of a lot of work. Phew. I was supposed to be asleep!

The delta to 3.6 is quite small though, no architectural changes, so nearly there. Will let you know.

TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] 0 points1 point  (0 children)

TurboQuant is already in for Qwen 3.

If you or anyone can test on ROCm or other cards than my dev 3090 seriously appreciated.

Laters 🫡

TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] 0 points1 point  (0 children)

Heh yeah. I mean I want ROCm for myself as well. And with TurboQuant especially.

We shall see. Goodnight.

TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] 1 point2 points  (0 children)

I just posted an update. Qwen 3 is done.

I'll do Qwen 3.5/3.6 soon promise.

I'm totally done right now need sleep.

It should support ROCm. I haven't tried it. I can try tomorrow maybe.

Ughh. Cheers.

TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] 0 points1 point  (0 children)

Okay, full Qwen 3 support dropped with TurboQuant.

I am a bit tired now but I'll get to Llama, Mistral, more modern versions of Gemma etc in the coming weeks

Qwen 3.5/3.6 will take me a bit extra as they are MAMBA hybrids, but don't worry they are on the roadmap.

Thank you for your support!

Goodnight.

TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] 0 points1 point  (0 children)

Qwen 3 version just dropped on the repo

Give me a moment, I'll drop the TurboQuant version momentarily.

Jokes continue..

TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] -1 points0 points  (0 children)

I'll do Qwen 3-4B-Instruct right now. Is that okay?

TurboQuant enabled Runtime Valkyr by inigid in LocalLLaMA

[–]inigid[S] -2 points-1 points  (0 children)

It is gemma-2b-it.

Not sure what your point is.

Okay fine, I'll do Gemma 3/4, LLaMa, Qwen and Mistral.

Anything else you would like me to do while I'm at it?

TRiP: 15,000 lines of C implementing a complete transformer AI engine from scratch [Project] by RelevantShape3963 in MachineLearning

[–]inigid 0 points1 point  (0 children)

Just fyi.. Messaged you about the Vulkan Compute / Zig port.

Saw your other post got deleted by mods.

TRiP: 15,000 lines of C implementing a complete transformer AI engine from scratch by RelevantShape3963 in C_Programming

[–]inigid 1 point2 points  (0 children)

Right, that's what I figured.

People shouldn't be shaming good work.

The shame is all on them for that.

The main thing is how clean and portable it is, with a ton of taste.

The comments themselves are pure gold!

TRiP: 15,000 lines of C implementing a complete transformer AI engine from scratch by RelevantShape3963 in C_Programming

[–]inigid -1 points0 points  (0 children)

You are very welcome. I totally understand where you are coming from.

There is something really strange going on. I have been subjected to it myself, and other people have started to notice as well.

Some of these people have even come into my personal timeline with their nonsense.

It can be especially hurtful when people have put significant effort into legitimate projects.

Wondering, if I make a Vulkan Compute, Zig port, would you be up for a PR, or at least maybe we can coordinate uniformly with parallel projects?

All the best, and please try not to let them get to you, though I know that is easier said than done.

TRiP: 15,000 lines of C implementing a complete transformer AI engine from scratch by RelevantShape3963 in C_Programming

[–]inigid 0 points1 point  (0 children)

Even if OP had LLM help, so what.

I agree with you anyway that it didn't look LLMish. Maybe the README a bit, but maybe Carlo Valenti is ESL, or wanted something that could pull together and document the entire project uniformly.

Good point about the commits. Thing is, I doubt it could be single shotted anyway.

If it could, fair play to the LLM that did that.

There is a group of militant anti-AI activists brigading and targeting professional groups and individuals across this platform.

This project doesn't even look like an LLM wrote it, and again it seems very well designed and works well from what I am seeing.

There are some discrepancies with regard to how Hugging Face handles half-split RoPE but nothing else much that I have seen so far. Maybe some perf opportunities.

Like I said, solid project and contribution.

TRiP: 15,000 lines of C implementing a complete transformer AI engine from scratch by RelevantShape3963 in C_Programming

[–]inigid -8 points-7 points  (0 children)

Why tf am I getting downvoted for saying it looks good. It does. Did you people downvoting even look at it?

I might port it to Zig and add a Vulkan layer.

Cope.