▲ | FrostAutomata 8 hours ago | |
ASCII art is extremely difficult for LLMs due to how characters are converted into tokens without preserving their relative positions. | ||
▲ | beklein 5 hours ago | parent | next [-] | |
Great point, but you need to have a good understanding in how LLMs work to understand this limitation. If you don't have an intuitive understanding think like it's one of these draw on my back games, just each new token is a new human in the loop, known words are like simple shapes you felt before 100 times on your back and are easy for you to reproduce and change, random ASCII strings are harder to _grasp_ and will produce a fuzzy output... all models are wrong, but some are useful. | ||
▲ | light_hue_1 4 hours ago | parent | prev | next [-] | |
This isn't the reason. Models are pretty good at understanding relative positions. We put that in them and reward it a lot. The issue is the same as why we don't use LLMs for image generation. Even though they can nominally do that. Image generation seems to need some amount of ability to revise the output in place. And it needs a big picture view to make local decisions. It doesn't lend itself to outputting pixel by pixel or character by character. | ||
▲ | 7 hours ago | parent | prev [-] | |
[deleted] |