▲ | solid_fuel 4 days ago | |
The problem with UX driven by this kind of interface is latency. Right now, this kind of flow goes more like: "What burgers do you have?" (Thinking...) (4 seconds later:) (expands to show a set of pictures) "Sigh. I'll have the thing with chicken and lettuce" (Thinking...) (3 seconds later:) > "Do you mean the Crispy McChicken TM McSandwich TM?" "Yes" (Thinking...) (4 seconds later:) > "Would you like anything else?" "No" (Thinking...) (5 seconds later:) > "Would you like to supersize that?" "Is there a human I can speak with? Or perhaps I can just point and grunt to one of the workers behind the counter? Anyone?" It's just exasperating, and it's not easy to overcome until local inference is cheap and common. Even if you do voice recognition on the kiosk, which probably works well enough these days, there's still the round trip to OpenAI and then the inference time there. And of course, this whole scenario gets even worse and more frustrating anywhere with subpar internet. | ||
▲ | 3 days ago | parent [-] | |
[deleted] |