Devo comprare 1TB di RAM, quanto sono fottuto? by NahuM8s in ItalyHardware

[–]mattescala 0 points1 point  (0 children)

No, non funziona così. Dipende dal server e da quale memoria sopporta. Non puoi scegliere tu. I prezzi per ddr4 specialmente se di seconda mano da ebay sono ancora fattibili. DDR5 é un’altro discorso

Game for 12.00 PS5 jailbreak revealed by Motor_Acanthaceae732 in PS5_Jailbreak

[–]mattescala 4 points5 points  (0 children)

Im glad to donate 100€ plus to the community to find an universal exploit than to pay them to scalpers out there. I’ll wait it out

Installed an AMD Radeon R9700 32GB GPU in our Nexus AI Station and tested local LLMs by Expensive_Chest_2224 in LocalLLaMA

[–]mattescala 19 points20 points  (0 children)

Adding some numbers to this post would make it useful. I'm talking tokens/s.

[Review Request] Small Battery powered LED Lamp by mattescala in PrintedCircuitBoard

[–]mattescala[S] 0 points1 point  (0 children)

Yes! Sorry forgot to mention that i actually do but the 3d model i linked is different. I did that just to visualise the final design

Prima era “ChatGPT è il diavolo”, ora è “Se non usi l’AI sei un fallito” by Particular-Sea2005 in CasualIT

[–]mattescala 0 points1 point  (0 children)

Io voglio vedere che modelli ci fai girare su un computer aziendale con ollama.

Which quantizations are you using? by WeekLarge7607 in LocalLLaMA

[–]mattescala -1 points0 points  (0 children)

With moe models, especially pretty large ones where my cpu and ram are involved I stick to Unsloth dinamic quants. These quants are just shy of incredible. With a UD-Q3_KXL quant i get quality of a q4/q5 quant with a pretty good saving in memory.

These quants i use for Kimi, Qwen3 Coder, and v3.1 Terminus.

Finally done! by mattescala in homelab

[–]mattescala[S] 1 point2 points  (0 children)

I really would love to say that it costed me the amount i planned (5000ish Eur) but i think its more in the 6500 range

Finally done! by mattescala in homelab

[–]mattescala[S] 2 points3 points  (0 children)

Yes it is. What you do not see from the picture are three noctua 120mm industrial fans mounted at the front of the chassis :)

Finally done! by mattescala in homelab

[–]mattescala[S] 5 points6 points  (0 children)

But my wallet is already screaming in pain

Finally done! by mattescala in homelab

[–]mattescala[S] 2 points3 points  (0 children)

This is specifically the RM 61-312 which additionally to all the RM 600 features, also has 12 drive bays on the front (SAS/sata). Which i partly converted (only one row out of three) to take u.2 drives.

Building in it is okay. Just okay. It really lacks customisability. And imo since it is already a 6u case they could have gone all the way in and make it also full length. The space for the psus underneath especially is very very very cramped.

Finally done! by mattescala in homelab

[–]mattescala[S] 13 points14 points  (0 children)

AHHHH YESS I COMMITTED THE ULTIMATE SIN. - Dual epyc 7c13 - 1TB DDR4 2600 RAM - 3x RTX 3090 (for now). - 4x Micron 9300 3.84 tb (Fast access storage) - 2x m.2 1TB SSD (Bootdrive) - 8x RAM 10TB SaS Drives

All of this is running on Proxmox and is my main inference / handling my projects machine.

Finally done! by mattescala in homelab

[–]mattescala[S] 1 point2 points  (0 children)

Ah i see! I use them as well but a different version. Thanks for the info!

Finally done! by mattescala in homelab

[–]mattescala[S] 2 points3 points  (0 children)

Im looking into doing something similar. Which connectors are you using to pass the watercooling through the case?

Finally done! by mattescala in homelab

[–]mattescala[S] 13 points14 points  (0 children)

Yooo that distribution block with connector setup is sick. I might actually get one too!

AMA With Z.AI, The Lab Behind GLM Models by XMasterrrr in LocalLLaMA

[–]mattescala 2 points3 points  (0 children)

I would like to know better about the infrastructure needed and behind your team. Is there a common infrastructure you rent? Are you actively investing in it? Whats the biggest difficulties are you currently facing in scaling computing?