| ▲ | wongarsu 3 hours ago | |
I've also always though that it's an interesting opportunity for custom hardware. Two bit addition is incredibly cheap in hardware, especially compared to anything involving floating point. You could make huge vector instructions on the cheap, then connect it to the fastest memory you can buy, and you have a capable inference chip. You'd still need full GPUs for training, but for inference the hardware would be orders of magnitude simpler than what Nvidia is making | ||
| ▲ | monocasa 3 minutes ago | parent | next [-] | |
These are trits, which provide their own efficiencies. Interestingly, a trit x float multiplier is cheaper than a trit x integer multiplier in hardware if you're willing to ignore things like NaNs. 0 and one are trivial, just a mux for identity and zero. But because floats are sign-magnitude, multiply by negative one is just an inverter for the sign bit, where as for integers you need a bitwise inverter and full incrermenter. | ||
| ▲ | regularfry 3 hours ago | parent | prev [-] | |
You only need GPUs if you assume the training is gradient descent. GAs or anything else that can handle nonlinearities would be fine, and possibly fast enough to be interesting. | ||