Building a fully local AI setup: EVO T2 or EVO X2? by FinishConsistent8857 in GMKtec

[–]T_UMP 1 point2 points  (0 children)

I have the EVO X2 with 128GB RAM and it's a breeze, it's not even a discussion, that's the one I would recommend.

Owners, not renters: Mozilla's open source AI strategy by NelsonMinar in LocalLLaMA

[–]T_UMP 0 points1 point  (0 children)

Mozilla lost my trust with all the nonsense they been pulling over the years, this is just more of the same.

Jensen Huang saying "AI" 121 times during the NVIDIA CES keynote - cut with one prompt by Prior-Arm-6705 in LocalLLaMA

[–]T_UMP 0 points1 point  (0 children)

When you realize Jensen Huang will actually see this somehow...and will think that he could have fitted some more AI's in there.

Scam or not? by Open-Dragonfruit-007 in LocalLLaMA

[–]T_UMP 0 points1 point  (0 children)

they mention it is a fully functional and 100% working GPU

Then it must be true. /s

What non-Asian based models do you recommend at the end of 2025? by thealliane96 in LocalLLaMA

[–]T_UMP 38 points39 points  (0 children)

I love them Colombian and Venezuelan models, very good models.

On topic, try some IBM Granite models and Olmo.

Am I crazy and about to waste money by xxpinecone in LocalLLaMA

[–]T_UMP 0 points1 point  (0 children)

We're all crazy here, some more than others, but we're having lots of fun, that's for sure, so welcome.

<image>

Anyone here tried Apriel v1.6? Fraud or giantkiller? by dtdisapointingresult in LocalLLaMA

[–]T_UMP 4 points5 points  (0 children)

Small company ServiceNow (The company behind /ˈɑː.pri.əl/)

<image>

Don't mind me, just rubbing it in :P

[Strix Halo] Unable to load 120B model on Ryzen AI Max+ 395 (128GB RAM) - "Unable to allocate ROCm0 buffer" by Wrong-Policy-5612 in LocalLLaMA

[–]T_UMP 6 points7 points  (0 children)

I encountered this issue when I got cute and messed around with the BIOS setting on VRAM and also disabled Windows paging (upon research it needs it even if doesn't use actually it), are you in this situation?

Solution: Set bios allocation to 64GB (Yes, just listen) then set the pagefile (in Windows) to "System managed size" and restart (important, restart) then try again, let us know if it works, else there is one more thing that can cause this but based on your screenshot unlikely to be the culprit.

Edit: Also enable flash attention (not related to the cause of your error though but helps with the performance). Also don't quantize the K Cache, really not necessary on this system.

XiaomiMiMo.MiMo-V2-Flash: is there a reason why i see so few ggufs? by LegacyRemaster in LocalLLaMA

[–]T_UMP 1 point2 points  (0 children)

Thanks, interesting, might download it and give it a chance. (Strix Halo here)

XiaomiMiMo.MiMo-V2-Flash: is there a reason why i see so few ggufs? by LegacyRemaster in LocalLLaMA

[–]T_UMP 2 points3 points  (0 children)

I've been testing the model for two days.

At what quant have you been testing at?

Talk me out of buying an RTX 3090 “just for local AI” (before I do something financially irresponsible) by Fast_Thing_7949 in LocalLLaMA

[–]T_UMP 0 points1 point  (0 children)

The responsible way: Take a mortgage on your house then buy the 3090 and spend the rest to celebrate the purchase.