| ▲ | feisty0630 12 hours ago | |
Interesting that it reads a bit like it came from a Markov chain rather than an LLM. Perhaps limited training data? | ||
| ▲ | kgeist 8 hours ago | parent [-] | |
Early LLMs used to have this often. I think's that where the "repetition penalty" parameter comes from. I suspect output quality can be improved with better sampling parameters. | ||