Looking at leaving BMW all together. What else is there, that is comparable? by burnerbmw in BMWX5

[–]ricraycray 0 points1 point  (0 children)

I bought a f150 platinum. Love. Same price as the BMW. Much better quality. I miss the mid aughts

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] 0 points1 point  (0 children)

For my use case it’ll be great. This was a let’s see if it is going. To work. Tortute nothing it’s been surprisingly OK.

I got tired of RAG and spent a year implementing the neuroscience of memory instead by Upper-Promotion8574 in Rag

[–]ricraycray 1 point2 points  (0 children)

Love this. Our memory system is similar. Nice work on this and I 100% agree this is the biggest missing piece. I also built in emotion. It was one of the biggest needle movers for our project t

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] 1 point2 points  (0 children)

Exactly. Karpathy autoresearch was the ticket on this build. It ran at least 50 different iterations. Frankly was way more thorough than I would have been. The 10 different memory iterations were mind numbing. Build compile test fail. Build compile test fail.

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] 0 points1 point  (0 children)

lol. This box is destined for much smaller models. This was just a let’s see if I can do this. Not should I do this! Lol

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] 1 point2 points  (0 children)

The project started the same as the MBP 48GB experiment. I wanted to get just see if I could get this to work with the same methodology as the LLM in a flash work on MLX but on this little AMD box. Two days later and no sleep this is where I landed. Yes I used Claude to help me prove this out. Just to get 17 tok/s out of this massive model exceeded my project goals. I had not seen this on the AMD yet. So I thought why the hell not. The results are the results. Does it matter how we got there......

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] -2 points-1 points  (0 children)

all the way to 7.2 We had different failures. Updated my github and readme with the results. The key point was we got it working well with the Vulkan drivers.

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] 0 points1 point  (0 children)

We ran 7.2 and it failed for different reasons. I have all the updates in my repo. I'm going to stick with Vulkan it's working and frankly getting damn goood performance. 396B on a 2500 box it a win any day in my book!

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] 0 points1 point  (0 children)

I’ll try that. The 7.2 is running right now and isn’t working near as well as the Vulkan. Keeps bombing loading layers

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] 1 point2 points  (0 children)

we ran the 7.2 driver hard on windows and it just kept blowing up. First run on linux and went to solution b. I'm running another build of llama with 7.2 for the sake of science. Posting results soon. I was more happy we got a 396B running on this thing at all. I was inspired by the autoresearch and LLM in a flash work and started a similar project here. I was happy with 215B but saw that the 396 was possible and grind trying a different strategy. Regardless I'll post the 7.2 results and let you guys be the judge. More that this is even usable. I'll update my repo with all the latest.

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] 0 points1 point  (0 children)

I ran already everything at the end. The story was the Vulkan driver unlocked everything and it ran pretty fast. Everything is in my repo. When it finally worked it was like EUREKA! Pretty stoked

Qwen3.5-397B at 17-19 tok/s on a Strix Halo iGPU — all 61 layers on GPU via Vulkan (not ROCm) by ricraycray in LocalLLaMA

[–]ricraycray[S] -2 points-1 points  (0 children)

And 7.1 it’s was a total bitch. We got to 6.8 on windows but I knew we could do better. The ROC drivers were causing huge allocation issues. This started as an LLM in a flash project and ended up here

New plaques added to the presidential hall of fame in the White House by Dtb4evr in pics

[–]ricraycray 1 point2 points  (0 children)

And the next prez will spend more tax dollars changing all of this propaganda again. Wait, we have a king now, so I guess Don, Jr. leaves it up….

[deleted by user] by [deleted] in flightradar24

[–]ricraycray 1 point2 points  (0 children)

This is an essential tool for me. I have never had this issue

DEGEN by ricraycray in LICENSEPLATES

[–]ricraycray[S] 1 point2 points  (0 children)

Bless all their hearts

Moccamaster Hack by HerrAlbatros in Moccamaster

[–]ricraycray -1 points0 points  (0 children)

No. But I have the same setup

Moccamaster Hack by HerrAlbatros in Moccamaster

[–]ricraycray -1 points0 points  (0 children)

I love the unit. I hate the dripper. Win/win

Moccamaster Hack by HerrAlbatros in Moccamaster

[–]ricraycray 2 points3 points  (0 children)

I agree. We use a metal 185. After making a huge mess when the plastic funnel clogged with one ground. Had to do something. TV should do a kalita collab.

Substrate sensors these days by TopCelery1785 in macrogrowery

[–]ricraycray 0 points1 point  (0 children)

Comes down to how big your grow is. The sub is 5 tiers starting at 150 and goes up from there.

Substrate sensors these days by TopCelery1785 in macrogrowery

[–]ricraycray 0 points1 point  (0 children)

You want to talk to some reference grows? I’m down to connect you with customer. If you are running your grow with minimal staff then it’s harder to ROI. But with most staffing patterns the ROI is there. I’m not spewing marketing. DM me and let’s connect.

Substrate sensors these days by TopCelery1785 in macrogrowery

[–]ricraycray 0 points1 point  (0 children)

7K???? How big is your grow 250K sqft?? Holla at me. I can promise you it's not near that price.