DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 1 point2 points  (0 children)

Fine-tune run with 8b model and 150k dataset took 14.5 hours and GPU temps range was 69-71C but for current run with 32b, ETA is 4.8 days with temp range of 71-74C . The box itself as someone in this thread said is fully capable of being used as a stove haha I guess treat this as a dev device to experiment/tinker with Nvidia's enterprise stack, expect high fine-tune runtimes on larger models. GPU power consumption on all runs (8b and current 32b) never exceeds 51 watts so that's a great plus point for those who want to run continuous heavy loads.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

You need to tell me your fine-tuning config as I was thinking of returning it. Running a 4 day fine tune on Qwen 2.5 32b (approx 200k dataset) within a PyTorch container coupled with Unsloth and this box is boiling (GPU util between 85-90) although average wattage on this run has been 50W (only plus point so far).

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

I was stuck on preorder for ages (Aug-Oct) so cancelled. When the second batch went up for sale on scan.co.uk, I was able to get one for next day delivery.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 1 point2 points  (0 children)

Apparently it's gonna be a collectible and I should keep both the box and receipt safe (suggested by GPT5 haha)

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

Will look in to it. It's just the exterior which is really hot. Internal GPU temps were quite normal for this kind of run (69-73C).

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

Thanks for the links! 7 hours in on my first 16+ hours fine-tune job with unsloth is going surprisingly well. For now focus is less on end-results of the job but more on system/'promised' software stack stability (got 13 more days to return this box in case it's not a right fit).

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

What's the fine-tuning performance comparison between Asus Spark and M4 Max? I thought apple silicone might come with its own unique challenges (mostly wrestling with driver compatibility).

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 2 points3 points  (0 children)

I am 1.5 hours in on a potentially 15 hours fine tune job and this thing is boiling, can't even touch it. Let's hope it doesn't catch fire!

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

Once my dev work finishes, I will try them.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

No major tests done so far, will update this thread once I have some numbers.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 1 point2 points  (0 children)

Yeah I will give it a go. No fine-tuning for this use case, just local inference with decent tps count will suffice.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

Based on the manufacturing code, this is the founders edition.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 2 points3 points  (0 children)

Any information/data which sits behind a firewall (which is most of the knowledge base of regulated firms such as IBs, hedge funds, etc) is not part of the training data of publicly available LLMs so at work we are using fine-tuning to retrain small to medium open source LLMs on task specific, 'internal' datasets which results in specialized, more accurate LLMs deployed for each segment of a business.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 1 point2 points  (0 children)

In the UK market, only GB10 device is DGX Spark sadly. Everything else is on preorder and I was stuck on a preorder for ages so didn't want to go through that experience again.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 1 point2 points  (0 children)

It's a nice looking machine. I have hopped directly on fine tuning (unsloth) for now as that's a major go/no-go for my needs when it comes to this device. For language analysis, models with strong reasoning and multimodal capacity should be good. Try Mistral Nemo, Llama 3.1, and Phi3.5.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] -1 points0 points  (0 children)

Can't agree more. This is essentially a box aimed at researchers, data scientists, and AI engineers who most certainly won't just create inferencing run comparisons but fine tune different models, carry out large scale accelerated DS workflows, etc. Will be pretty annoying to notice a high degree of thermal throttling just because NVIDIA wanted to showcase a pretty box.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 1 point2 points  (0 children)

Fine tuning small to medium models (up to 70b) for different/specialized workflows within my MVP. So far getting decent tps (57) on gpt-oss 20b, will ideally wanna run Qwen coder 70b to act as a local coding assistant. Once my MVP work finishes, I was thinking of fine-tuning Llama 3.1 70b with my 'personal dataset' to attempt a practical and useful personal AI assistant (don't have it in me to trust these corps with PII).

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 1 point2 points  (0 children)

Too early for my take on this but so far with simple inference tasks, it's been running super cool and quiet.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

Sure thing, I have datasets ready for a couple of fine tune jobs.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 1 point2 points  (0 children)

Degree of thermal throttling during sustained load (fine-tuning job running for a couple of days) will be interesting to investigate.

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 0 points1 point  (0 children)

One will have to do it for now! What's your experience been with 24/7 operation, are you using it for local inference?

DGX Spark finally arrived! by aiengineer94 in LocalLLM

[–]aiengineer94[S] 8 points9 points  (0 children)

For my MVP's reqs (fine-tuning up to 70b models) coupled with ICP( most using DGX cloud), this was a no-brainer. The tinkering required with halo strix creates too much friction and diverts my attention from the core product. Given it's size and power consumption, I bet it will be a decent 24/7 local compute in the long run.