Readit News logoReadit News
aurareturn commented on Nvidia DGX Spark   nvidia.com/en-us/products... · Posted by u/janandonly
wmf · 4 hours ago
It's the same as Strix Halo and M4 Max that people are going gaga about, so either everyone is wrong or it's fine.
aurareturn · 3 hours ago
M4 max has more than double the bandwidth.

Strix Halo has the same and I agree it’s overrated.

aurareturn commented on Nvidia DGX Spark   nvidia.com/en-us/products... · Posted by u/janandonly
hereme888 · 3 hours ago
FP4-sparse (TFLOPS) | Price | $/TF4s

5090: 3352 | 1999 | 0.60

Thor: 2070 | 3499 | 1.69

Spark: 1000 | 3999 | 4.00

____________

FP8-dense (TFLOPS) | Price | $/TF8d (4090s have no FP4)

4090 : 661 | 1599 | 2.42

4090 Laptop: 343 | vary | -

____________

Geekbench 6 (compute score) | Price | $/100k

4090: 317800 | 1599 | 503

5090: 387800 | 1999 | 516

M4 Max: 180700 | 1999 | 1106

M3 Ultra: 259700 | 3999 | 1540

____________

Apple NPU TOPS (not GPU-comparable)

M4 Max: 38

M3 Ultra: 36

aurareturn · 3 hours ago
It's not good value when you put it like that. It doesn't have a lot of compute and bandwidth. What it has is the ability to run DGX software for CUDA devs I guess. Not a great inference machine either.
aurareturn commented on Nvidia results show spending on A.I. infrastructure remains robust   nytimes.com/2025/08/27/te... · Posted by u/cuttothechase
cuttothechase · 4 hours ago
Would this work once the funds start to dry up and the business has to self sustain, at some point?

Assumption is that business become profitable, and not just break even profitable but profitable at their expected multiples, isn't it?

aurareturn · 3 hours ago
Not sure if it's profitable but I don't think intelligence can be compared to just CPUs in dotcom.

Given enough compute, you can a very smart LLM to work on any number of things by itself.

aurareturn commented on Nvidia DGX Spark   nvidia.com/en-us/products... · Posted by u/janandonly
BoorishBears · 4 hours ago
They're abysmal compared to anything dedicated at any reasonable batch size because of both bandwidth and compute, not sure why you're wording this like it disagrees with what I said.

I've run inference workloads on a GH200 which is an entire H100 attached to an ARM processor and the moment offloading is involved speeds tank to Mac Mini-like speeds, which is similarly mostly a toy when it comes to AI.

aurareturn · 4 hours ago
Again, prompt processing isn't the major problem here. It's bandwidth. 256GB/s bandwidth (maybe ~210 in real world) limits the tokens per second well before prompt processing.

Not entirely sure how your ARM statement matters here. This is unified memory.

aurareturn commented on Nvidia DGX Spark   nvidia.com/en-us/products... · Posted by u/janandonly
BoorishBears · 4 hours ago
No. These are practically useless for AI.

Their prompt processing speeds are absolutely abysmal: if you're trying to tinker from time to time, a GPU like a 5090 or renting GPUs is a much better option.

If you're just trying to prep for impending mainstream AI applications, few will be targeting this form factor: it's both too strong compared to mainstream hardware, and way too weak compared to dedicated AI-focused accelerators.

-

I'll admit I'm taking a less nuanced take than some would prefer, but I'm also trying to be direct: this is not ever going to be a better option than a 5090.

aurareturn · 4 hours ago

  Their prompt processing speeds are absolutely abysmal
They are not. This is Blackwell with Tensor cores. Bandwidth is the problem here.

aurareturn commented on Nvidia DGX Spark   nvidia.com/en-us/products... · Posted by u/janandonly
cherioo · 4 hours ago
The mainstream options seem to be

Ryzen AI Max 395+, ~120 tops (fp8?), 128GB RAM, $1999

Nvidia DGX Spark, ~1000 tops fp4, 128GB RAM, $3999

Mac Studio max spec, ~120 tflops (fp16?), 512GB RAM, 3x bandwidth, $9499

DGX Spark appears to potentially offer the most token per second, but less useful/value as everyday pc.

aurareturn · 4 hours ago

  Mac Studio max spec, ~120 tflops (fp16?), 384GB RAM, 3x bandwidth, $9499
512GB.

DGX has 256GB/s bandwidth so it wouldn't offer the most tokens/s.

aurareturn commented on Nvidia results show spending on A.I. infrastructure remains robust   nytimes.com/2025/08/27/te... · Posted by u/cuttothechase
boroboro4 · 5 hours ago
Same argument can be somewhat applied to CPUs circa 90s. The growth did stop/stalled in the end.

I think the expectation of ever growing compute is not totally crazy. It will come with lower margins eventually though, and more players in the market. It also might get much more moderate, including from hardware limitations. Efficiency wise h200->b200 isn’t as crazy as a100->h100.

aurareturn · 4 hours ago

  Same argument can be somewhat applied to CPUs circa 90s. The growth did stop/stalled in the end.
One difference is that the more compute you have, the more work you can do. You can basically run tokens 24/7 building apps or trying to do research or running a massive model in thinking mode.

aurareturn commented on Denmark summons top US diplomat over alleged Greenland influence operation   bbc.com/news/articles/c0j... · Posted by u/vinni2
AnimalMuppet · 9 hours ago
I don't think that Vance is a complete robot. If he wins, why would he resign? Just to please Trump? What does Trump's approval mean to him once Vance is the president?
aurareturn · 5 hours ago

  If he wins, why would he resign
Presumably the public/voters votes for him with an understanding that he resigns.

aurareturn commented on Denmark summons top US diplomat over alleged Greenland influence operation   bbc.com/news/articles/c0j... · Posted by u/vinni2
vintermann · 15 hours ago
Yes, but they won't, is what I'm saying.
aurareturn · 14 hours ago
The enemy of my enemy is my friend. :)
aurareturn commented on US Intel   stratechery.com/2025/u-s-... · Posted by u/maguay
nonethewiser · 15 hours ago
> They want to wait when it's so obvious that Taiwan would lose and lose quickly that Taiwan simply gives itself up.

I agree they prefer a peaceful resolution but that would be in the form of Taiwan wanting to become part of China and deciding to reunify. It makes no sense to surrender without being forced to.

aurareturn · 14 hours ago
It makes sense if it's inevitable.

u/aurareturn

KarmaCake day4520November 14, 2022View Original