Remix.run Logo
apfsx 4 hours ago

Google’s TPU’s are not powering Gemini or whatever X equivalent LLM you want to compare to.

rfw300 4 hours ago | parent | next [-]

This isn't true. Gemini is trained and run almost entirely on TPUs. Anthropic also uses TPUs for inference, see, e.g., https://www.anthropic.com/news/expanding-our-use-of-google-c... and https://www.anthropic.com/engineering/a-postmortem-of-three-.... OpenAI also uses TPUs for inference at least in some measure: https://x.com/amir/status/1938692182787137738?t=9QNb0hfaQShW....

skirmish 4 hours ago | parent | prev | next [-]

I can assure you that most internal ML teams are using TPUs both for training and inference, they are just so much easier to get. Whatever GPUs exist are either reserved for Google Cloud customers, or loaned temporarily to researchers who want to publish easily externally reproducible results.

jsolson 3 hours ago | parent | prev | next [-]

This comment is incorrect: https://storage.googleapis.com/deepmind-media/Model-Cards/Ge...

stingraycharles 4 hours ago | parent | prev | next [-]

They are, even Apple famously uses Google Cloud for their cloud based AI stuff solely because of Apple not wanting to buy NVidia.

Google Cloud does have a lot of NVidia, but that’s for their regular cloud customers, not internal stuff.

rdudek 4 hours ago | parent | prev [-]

What is powering Gemini?

jsolson 3 hours ago | parent [-]

TPUs: https://storage.googleapis.com/deepmind-media/Model-Cards/Ge...