Remix.run Logo
whimsicalism 2 days ago

also cheaper memory bandwidth. where are you claiming that M5 wins?

Infernal 2 days ago | parent [-]

I'm not sure where else you can get a half TB of 800GB/s memory for < $10k. (Though that's the M3 Ultra, don't know about the M5). Is there something competitive in the nvidia ecosystem?

whimsicalism 2 days ago | parent [-]

I wasn't aware that M3 Ultra offered a half terabyte of unified memory, but an RTX5090 has double that bandwidth and that's before we even get into B200 (~8TB/s).

650REDHAIR 2 days ago | parent [-]

You could get x1 M3 Ultra w/ 512gb of unified ram for the price of x2 RTX 5090 totaling 64gb of vram not including the cost of a rig capable of utilizing x2 RTX 5090.

bigyabai 2 days ago | parent [-]

Which would almost be great, if the M3 Ultra's GPU wasn't ~3x weaker than a single 5090: https://browser.geekbench.com/opencl-benchmarks

I don't think I can recommend the Mac Studio for AI inference until the M5 comes out. And even then, it remains to be seen how fast those GPUs are or if we even get an Ultra chip at all.

adastra22 2 days ago | parent [-]

Again, memory bandwidth is pretty much all that matters here. During inference or training the CUDA cores of retail GPUs are like 15% utilized.

my123 a day ago | parent | next [-]

Not for prompt processing. Current Macs are really not great at long contexts

2 days ago | parent | prev [-]
[deleted]