| ▲ | manucardoen 10 hours ago |
| It's not good at making ASCII art. This, for example, is when I ask it for a realistic depiction of the Eiffel tower on fire: .
.'.
|o|
.'o'.
|.-.|
' '
/ \
| |
__/_________\__
|###############|
|###############|
|###############|
|###############|
|###############|
/###############/|
/###############/ |
/###############/ /|
/###############/ / |
/###############/ / /|
/###############/ / / |
/###############/ / / /|
/###############/ / / / |
/###############/ / / / /|
'-----------------' / / / /
| FIRE & ASH |/ / / /
'---------------' / / /
~ ~ ~ ~ ~ ~ ~ ~ / /
~~ /
~~
~~ ⬆ SMOKE
~~
|
|
| ▲ | cucubeleza 5 hours ago | parent | next [-] |
| well, I asked to GPT to draw Arnold Schwarzenegger: ______
/ \
| O O |
| __ |
\______/
||||
/--||--\
/ || \
| || |
| / \ |
\__/ \__/
|| ||
|| ||
/ | | \
/_/ \_\
|
| |
|
| ▲ | bezbac 3 hours ago | parent | prev | next [-] |
| I've read that the results improve if you ask them to write a program that creates the desired ASCII art. Haven't tried it myself yet so far. |
|
| ▲ | FrostAutomata 8 hours ago | parent | prev | next [-] |
| ASCII art is extremely difficult for LLMs due to how characters are converted into tokens without preserving their relative positions. |
| |
| ▲ | beklein 5 hours ago | parent | next [-] | | Great point, but you need to have a good understanding in how LLMs work to understand this limitation.
If you don't have an intuitive understanding think like it's one of these draw on my back games, just each new token is a new human in the loop, known words are like simple shapes you felt before 100 times on your back and are easy for you to reproduce and change, random ASCII strings are harder to _grasp_ and will produce a fuzzy output... all models are wrong, but some are useful. https://www.youtube.com/watch?v=bA_DQHoxj34 | |
| ▲ | light_hue_1 5 hours ago | parent | prev | next [-] | | This isn't the reason.
Models are pretty good at understanding relative positions. We put that in them and reward it a lot. The issue is the same as why we don't use LLMs for image generation. Even though they can nominally do that. Image generation seems to need some amount of ability to revise the output in place. And it needs a big picture view to make local decisions. It doesn't lend itself to outputting pixel by pixel or character by character. | |
| ▲ | 7 hours ago | parent | prev [-] | | [deleted] |
|
|
| ▲ | yesbabyyes 10 hours ago | parent | prev | next [-] |
| This is something I and a few of my colleagues have noticed, as we asked several models to draw ASCII art of a wasp, which is one of our logos. The results are hilarious, and only seem to get worse as you ask it to do better. |
|
| ▲ | bn-l 8 hours ago | parent | prev [-] |
| Art is highly subjective |
| |