| ▲ | leo_e 6 hours ago | |||||||
Impressive numbers on paper, but looking at their site, this feels dangerously close to vaporware. The bottleneck for inference right now isn't just raw FLOPS or even memory bandwidth—it's the compiler stack. The graveyard of AI hardware startups is filled with chips that beat NVIDIA on specs but couldn't run a standard PyTorch graph without segfaulting or requiring six months of manual kernel tuning. Until I see a dev board and a working graph compiler that accepts ONNX out of the box, this is just a very expensive CGI render. | ||||||||
| ▲ | mg 5 hours ago | parent | next [-] | |||||||
Six months of one developer tuning the kernel? That seems like not much compared to the hundreds of billions of dollars US companies currently invest into their AI stack? OpenAI pays thousands of engineers and researchers full time. | ||||||||
| ||||||||
| ▲ | vlovich123 37 minutes ago | parent | prev | next [-] | |||||||
Inference accelerators are not where Nvidia is maintaining their dominance afaik. | ||||||||
| ▲ | m00dy 3 hours ago | parent | prev [-] | |||||||
very good point leo_e indeed no mention of PyTorch in their website...honestly it looks a bit scammy as well | ||||||||