| ▲ | mrinterweb 7 hours ago | ||||||||||||||||
As far as I understand all the inference purpose-build silicon out there is not being sold to competitors and kept in-house. Google's TPU, Amazon's Inferentia (horrible name), Microsoft's Maia, Meta's MTIA. It seems that custom inference silicon is a huge part of the AI game. I doubt GPU-based inference will be relevant/competitive soon. | |||||||||||||||||
| ▲ | nightshift1 6 hours ago | parent | next [-] | ||||||||||||||||
According to this semianalysis article, the Google/Broadcom TPU are being sold to others like Anthropic. https://newsletter.semianalysis.com/p/tpuv7-google-takes-a-s... | |||||||||||||||||
| ▲ | nomel 6 hours ago | parent | prev | next [-] | ||||||||||||||||
> It seems that custom inference silicon is a huge part of the AI game. Is there any public info about % inference on custom vs GPU, for these companies? | |||||||||||||||||
| |||||||||||||||||
| ▲ | almostgotcaught 6 hours ago | parent | prev [-] | ||||||||||||||||
> soon When people say things like this I always wonder if they really think they're smarter than all of the people at Nvidia lolol | |||||||||||||||||
| |||||||||||||||||