| ▲ | moralestapia 6 hours ago | |
Thanks for having a brain. Not sure who started that "split into 10 chips" claim, it's just dumb. This is Llama 3B hardcoded (literally) on one chip. That's what the startup is about, they emphasize this multiple times. | ||
| ▲ | aurareturn 6 hours ago | parent [-] | |
It’s just dumb to think that one chip per model is their plan. They stated that their plan is to chain multiple chips together. I was indeed wrong about 10 chips. I thought they would use llama 8B 16bit and a few thousand context size. It turns out, they used llama 8B 3bit with around 1k context size. That made me assume they must have chained multiple chips together since the max SRAM on TSMC n6 for reticle sized chip is only around 3GB. | ||