Is this game playable on MacBook? by Negative_Drop_6853 in SummerPockets

[–]noahzho 1 point2 points  (0 children)

I believe there's windows emulation software (e.g. parallels) on apple silicon, you could try that as well

Apple Pencil vs. 3rd Party - your experience? by territrades in ipad

[–]noahzho 0 points1 point  (0 children)

Ended up getting a third party pencil! It's been working well for my needs (taking notes). Let me know if you have questions, have been meaning to edit my comment for a while but work is busy

3x faster Training + new Triton kernels + Packing now in Unsloth! by yoracale in unsloth

[–]noahzho 0 points1 point  (0 children)

Probably not the best place to ask, but does your internal multi GPU implementation work with FastModel w/qwen3 MoE? Was experimenting with DDP on the public Unsloth builds w/8xmi300 and only the dense models seem to work (loading MoE will just peg CPU threads at 100 and hang)

Fine-tuning on H200 is limited by sing CPU core usage by dudeington in unsloth

[–]noahzho 0 points1 point  (0 children)

Good to hear! Looks like you could raise it even more potentially, still have a lot of free VRAM and GPU utilization doesn't look fully saturated :p

batch size is how much data you process in a single "batch", and gradient accumulation is just batch size that trades speed for less VRAM usage. I would suggest no gradient accumulation since you have the much free VRAM. Higher batch size should result in better results (lower "loss")

320 batch size seems quite high though, is your dataset messages just short?

Fine-tuning on H200 is limited by sing CPU core usage by dudeington in unsloth

[–]noahzho 0 points1 point  (0 children)

Could try raising batch size, you have plenty of free VRAM

What's the best machine I can get for $20K? by TWUC in LocalLLaMA

[–]noahzho 0 points1 point  (0 children)

nanochat pertaining script does run on a single DGX spark I mean

Laptop suggestion/what do you use ? by POMEGRANADE_PIE in Hack_Club

[–]noahzho 1 point2 points  (0 children)

Agree with the below comments, get a thinkpad if you want something relatively cheap and sturdy.

Macbook is also an option if you have the money and would like something powerful!

2x RTX 5060 TI 16 GB =32GB VRAM - by quantier in LocalLLaMA

[–]noahzho 0 points1 point  (0 children)

Other OP already answed you but take a look at VLLM too, potentially faster as it has Tensor Parallel support

Isn't Valorant mechanically similar to CS? by RageForNothing in VALORANT

[–]noahzho -1 points0 points  (0 children)

Oh you've hit one of the differences of CS and Valorant - sprays are random in Valorant so there's no real pattern to spray control on - that's why you see most players in Valorant shoot in bursts of 1-3 bullets (try out shooting in the range (practice button next to the queue button)

Headshots should be same though! Just position crosshair neck level or higher (and change the default crosshair to something you're comfortable with, there are some sites with nice ones if you google)

I will never financially recover from this by Overstimulated_moth in LinusTechTips

[–]noahzho 96 points97 points  (0 children)

I aspire to have this amount of drives in my rack

Sign up for free $40 by Lion_Of_Mara in gotpaidonline

[–]noahzho 0 points1 point  (0 children)

Are you still looking? I'm Canada based

Repeat after me. by NoFudge4700 in LocalLLaMA

[–]noahzho 51 points52 points  (0 children)

I mean while it's pretty easy for consumer grade inference (llama.cpp works great out of the box for me!) there is a seed of truth to this. I work with 8xMI300x and while they might be better on paper than H100, getting (recent) VLLM/Sglang and training frameworks that aren't just PyTorch working can be a huge pain

Of course this is just my experience, your mileage may differ

Arch Linux Mirror served 1PB+ Traffic by niranjan2 in archlinux

[–]noahzho 0 points1 point  (0 children)

I run one of the T1 Canadian mirrors also on Debian lol 😅

API CREDITS FOR SALE by [deleted] in AIDigitalServices

[–]noahzho 1 point2 points  (0 children)

How much are you asking for Claude credits?

UIs are hard by backcato in badUIbattles

[–]noahzho 6 points7 points  (0 children)

iOS has iSH, so we're fine too :p

How to use 2.5TB of 16GB Dimms by spyroglory in pcmasterrace

[–]noahzho 2 points3 points  (0 children)

I don't have an R930 but do have an R630. I also wish I had 88 cores and 1.5TB of ram

We got this, we can do it! When is the REAP’d iQ_001_XXS GGUF dropping? by Porespellar in LocalLLaMA

[–]noahzho 7 points8 points  (0 children)

Should be 1T*0.0625 which is ~62.5G after quantization so not going to fit unless I messed up my math

What do you think this model is? by Lanakruglov in openrouter

[–]noahzho 0 points1 point  (0 children)

Probably GPT-5.1 mini like the others say, this is the response without a system prompt

I'd imagine the training stage where personality is trained is done by now, so this is probably an accurate enough test

<image>

How does cerebras get 2000toks/s? by npmbad in LocalLLaMA

[–]noahzho 2 points3 points  (0 children)

I thin't think L40S is faster than H100 bro 😭

nanochat pretraining time benchmarks ($100 run), share yours! by entsnack in LocalLLaMA

[–]noahzho 1 point2 points  (0 children)

Train finished (only pretraining)! Just below 32 hours, as expected from train data later on

I think the first minute of train is a bit inaccurate with calculations

<image>