Remix.run Logo
seydor 9 hours ago

How come we don't have AI@Home

throwup238 9 hours ago | parent | next [-]

The network bandwidth between nodes is a bigger limitation than compute. The newest Nvidia cards come with 400gbit busses now to communicate between them, even on a single motherboard.

Compared to SETI or Folding @Home, this would work glacially slow for AI models.

fourthark 8 hours ago | parent [-]

Seems like training would be a better match, where you need tons of compute but don’t care about latency.

ronsor 35 minutes ago | parent [-]

No, the problem is that with training, you do care about latency, and you need a crap-ton of bandwidth too! Think of the all_gather; think of the gradients! Inference is actually easier to distribute.

6 hours ago | parent | prev [-]
[deleted]