| ▲ | MediaSquirrel 8 hours ago | |||||||
More data -> better, faster on-device models The actual plan was to distill Gemini 2.5 Pro into the best on-device voice dictation model. Pretty sure it would have worked. Alas. | ||||||||
| ▲ | nomel 8 hours ago | parent [-] | |||||||
Reasons for running local aside... What is the practical latency difference you see between on-device and, say, whisper, in streaming mode, over the internet? Comparable? Seems that internet latency would be mostly negligible (assuming reasonable internet/cell coverage), or at least compensated for by the higher end hardware on the other side? | ||||||||
| ||||||||