| ▲ | wmf 2 days ago | |
inference accelerators ... It’s why Apple is so far ahead of their peers in local device AI compute, and will be for some time. Not really. llama.cpp was just using the GPU when it took off. Apple's advantage is more VRAM capacity. this introduction isn’t just for large clusters It doesn't work for large clusters at all; it's limited to 6-7 Macs and most people will probably use just 2 Macs. | ||