| ▲ | dragonwriter 2 days ago | |
> A Markov Chain trained by only a single article of text will very likely just regurgitate entire sentences straight from the source material. Strictly speaking, this is true of one particular way (the most straightforward) to derive a Markov chain from a body of text; a Markov chain is just a probabilistic model of state transitions where the probability of each possible next state is dependent only on the current state. Having the states be word sequence of some number of words, overlapping by all but one word, and having the probabilities being simply the frequency with which the added word in the target state follows the sequence in the source state in the training corpus is one way you can derive a Markov chain from a body of text, but not the only one. | ||
| ▲ | JPLeRouzic a day ago | parent [-] | |
> A Markov Chain trained by only a single article of text will very likely just regurgitate entire sentences straight from the source material. I see this as a strength; try training an LLM on a 42KB text to see if it can produce a coherent output. | ||