You are viewing a single comment's thread from:

RE: Nvidia DGX Spark Released & Big Improvements on my Strix Halo

in #ai28 days ago (edited)

I paid $1800 for my Strix and I run a 120b q8 model at 50 tokens/sec. I have run Qwen3 235B as well.

Cuda is becoming less of a deal breaker with rocm improving rapidly.

I don’t take the Strix or the Spark seriously, in my opinion they are both toys.

Sort:  

well yeah, they are toys compared to Enterprise infrastructure. But both of those run on normal 120V power outlets so they can be used in everyday homes.

A DGX runs n 240V C19/C20 cables so not really an option for 'normal' people.

So get 2 Spark devices connected over their ConnectX-7 ports and you can run any 'consumer-grade' model.

And CUEA being less of a deal breaker is true but moving from 99% market share to 94% is less as well, doesn't mean that they all AI Enterprise developers still utilize CUDA :)

But enough Devil's Advocate, The spark is cool and is a super efficient tool to run consumer AI models but so is the AMD Strix.

Loading...
 18 days ago Reveal Comment
Loading...