| ▲ | numpad0 2 hours ago | |
The actually issue according to another comment [0] is this[1]: > Around iOS 17 (Sept. 2023) Apple updated their autocorrect to use a transformer model which should've been awesome and brought it closer to Gboard (Gboard is a privacy terror but honestly, worth it). > What it actually did/failed to improve is make your phone keyboard: > Suck at suggesting accurate corrections to misspelled words > "Correct" misspelled words with an even worse misspelling > "Correct" your correctly spelled word with an incorrectly spelled word Which makes me wonder: is Transformer model good with manipulating short texts and texts with errors at all ? It's kind of known that open weight LLMs don't perform well for CJK conversion tasks[2], and I've also been disappointed by their general lack of typo tolerances myself as well. They're BAD for translating ultrashort sentences and singled out words as well[3]. They're great for vibecoding, though. Which makes me think, are they usable for anything under <100 bytes at all? Does it seem like they have a minimum usable input entropy or something? 0: https://news.ycombinator.com/item?id=47006171 1: https://thismightnotmatter.com/a-little-website-i-made-for-a... 2: The process of yielding "㍑" from "rittoru" 3: No human can translate, e.g. "translate left" in isolation correctly as "move left arm", but LLMs seem to be more all over the place than humans | ||