| ▲ | Tepix 4 hours ago | |
No, they're not selling this as an "AI accelerator": Here is the quote: "The company says operators deploying 5G Advanced and future 6G networks increasingly rely on server CPUs for virtualized RAN and edge AI inference, as they do not want to re-architect their data centers in a bid to accommodate AI accelerators." Edge AI usually means very small models that run fine on CPUs. | ||
| ▲ | hedora 4 hours ago | parent [-] | |
A very small model is going to be, what, 8GB? That'll easily blow through the caches. You're going to end up bottlenecked on DRAM either way. So, I wonder if this is going to be any faster than the previous generation for edge AI. | ||