▲ | entropyneur 4 days ago | ||||||||||||||||||||||
How does one look at gpt-1 output and think "this has potential"? You could easily produce more interesting output with a Markov chain at the time. | |||||||||||||||||||||||
▲ | empiko 4 days ago | parent | next [-] | ||||||||||||||||||||||
This was an era where language modeling was only considered as a pretraining step. You were then supposed to fine tune it further to get a classifier or similar type of specialized model. | |||||||||||||||||||||||
▲ | iNic 4 days ago | parent | prev [-] | ||||||||||||||||||||||
At the time getting complete sentences was extremely difficult! N-gram models were essentially the best we had | |||||||||||||||||||||||
|