In partnership with

Attention spans are shrinking. Get proven tips on how to adapt:

Mobile attention is collapsing.

In 2018, mobile ads held attention for 3.4 seconds on average.
Today, it’s just 2.2 seconds.

That’s a 35% drop in only 7 years. And a massive challenge for marketers.

The State of Advertising 2025 shows what’s happening and how to adapt.

Get science-backed insights from a year of neuroscience research and top industry trends from 300+ marketing leaders. For free.

I Bought the NVIDIA DGX Spark. Here's Why I Can't Recommend It (Yet).

I Bought the NVIDIA DGX Spark. Here's Why I Can't Recommend It (Yet).

The $4,000 "desktop AI supercomputer" that doesn't deliver on its promises

NVIDIA launched the DGX Spark with bold claims: a petaflop of AI computing power in a sleek gold box that fits on your desk. At $3,999, they promised to "put a Grace Blackwell powered AI supercomputer on every desk."

After spending time with it, here's my honest assessment: wait for the fixes, or look elsewhere.


🔥 The Thermal Throttling Problem

Let's start with what John Carmack—yes, the legendary Doom creator—publicly called out on X: the DGX Spark maxes out at only 100 watts, less than half of its 240-watt rating.

The result? Roughly half the advertised performance.

Multiple reports on NVIDIA's developer forums confirm GPU crashes and unexpected shutdowns under sustained load. My own experience mirrors this—the device runs hot, and I've encountered spontaneous reboots during long runs.

For a device marketed to researchers running intensive AI workloads, this is a dealbreaker.

⚠️ The Driver Nightmare

Here's something nobody warned me about: peripheral drivers randomly dying.

Keyboard and mouse connections become unstable. The underlying DGX OS (based on Ubuntu 24.04) struggles with basic hardware compatibility. For a $4,000 "supercomputer," you'd expect the basics to work reliably.

🐍 PyTorch & Blackwell Compatibility Issues

This one hurts the most.

The GB10 chip uses CUDA capability 12.1, but PyTorch's official releases only support up to 12.0. Every time you run a workload, you'll see warnings like:

UserWarning: Found GPU0 NVIDIA GB10 which is of cuda capability 12.1.
Minimum and Maximum cuda capability supported by this version of PyTorch is (8.0) - (12.0)

Getting proper PyTorch training to work requires workarounds, custom Docker containers, and fighting with version mismatches. Simon Willison documented his struggle—he eventually got PyTorch 2.7 working for CUDA on ARM, but couldn't get 2.8 running.

For a device from NVIDIA, the company that owns CUDA, this ecosystem fragmentation is inexcusable.

📊 The "1 Petaflop" Marketing Trick

NVIDIA's headline claim of 1 PFLOP performance uses FP4 with structured sparsity—a technique that only works when neural network operations are specifically optimized for it.

In real-world applications? You're looking at roughly 125 TFLOPS dense equivalent at best, and users report getting far less than that.

The 273 GB/s memory bandwidth from LPDDR5X is also shared between CPU and GPU. Compare that to an Apple Mac Studio M3 Ultra with 819 GB/s, or even HBM-equipped datacenter GPUs with 2-3x more bandwidth.

🍎 Mac Mini and Mac Studio: The Uncomfortable Comparison

Here's what NVIDIA doesn't want you to hear:

Token generation (decode) Mac Studio M3 Ultra is 3.4x faster than DGX Spark
30-70B parameter models Mac Mini M4 Pro performs comparably in Ollama benchmarks
Large models (Qwen3:32b, Llama3.1:70b) DGX Spark's performance "isn't inspiring for the price premium"

The DGX Spark excels at prefill (processing your prompt), but stumbles on decode (generating responses)—which is what actually matters for interactive use.

☁️ Cloud GPU vs. DGX Spark: The Math Doesn't Work

Let's do the math on the "escape cloud costs" narrative:

  • DGX Spark: $3,999 upfront + electricity + cooling + maintenance
  • Cloud A100 hourly rates: ~$1-2/hour for on-demand

If you run 4 hours daily, 5 days a week, that's ~$100-200/month in cloud costs. It would take 2-3 years of constant use to break even—by which point NVIDIA will have released something that actually works.

And cloud GPUs give you:

  • No thermal throttling
  • Full software compatibility
  • No driver headaches
  • Actual production-ready performance

🎓 The "For Researchers" Positioning Falls Flat

NVIDIA markets this to "AI researchers, data scientists, and students." But:

  • Researchers need reliable, reproducible environments. Not devices that crash mid-training.
  • Data scientists need ecosystem compatibility. Not PyTorch version warnings.
  • Students need value. A well-configured RTX 4090 workstation offers better price-performance for learning.

✅ What Would Make This Worth It

To be fair, the DGX Spark has genuine potential:

  • 128GB unified memory CAN run 200B parameter models for inference
  • The NVLink clustering (connecting two units) is genuinely innovative
  • The form factor is beautiful engineering

But NVIDIA needs to:

  1. Release firmware updates addressing thermal throttling
  2. Ensure official PyTorch builds support CUDA 13/GB10 out of the box
  3. Fix the driver stability issues
  4. Be honest about real-world performance metrics

🎯 My Recommendation

Don't buy the DGX Spark in its current state.

If you need local AI development:

  • Mac Studio M3/M4 Ultra for general AI + creative work
  • AMD Strix Halo for a more affordable alternative
  • Cloud GPUs for production workloads

If you're committed to the NVIDIA ecosystem:

  • Wait 6-12 months for firmware fixes
  • Wait for PyTorch/ecosystem maturity
  • Consider used datacenter GPUs for better value

The DGX Spark is a cautionary tale about the gap between marketing claims and real-world performance. NVIDIA built gorgeous hardware, then shipped it before the software was ready.

For $4,000, you deserve better.


Have you tried the DGX Spark? Reply to this email and let me know your experience.

— Deep
Senior SRE @ Adobe | AI/ML Engineering | ResearchAudio.io

Keep Reading

No posts found