| ▲ | Sohcahtoa82 2 days ago | ||||||||||||||||||||||
> I have seen the argument that LLMs can only give you what its been trained on, i.e. it will not be "creative" or "revolutionary", that it will not output anything "new", but "only what is in its corpus". LLMs can absolutely create things that are creative, at least for some definition of "creative". For example, I can ask an LLM to create a speech about cross-site scripting the style of Donald Trump: > Okay, folks, we're talking about Cross-Site Scripting, alright? I have to say, it's a bit confusing, but let's try to understand it. They call it XSS, which is a fancy term. I don't really know what it means, but I hear it's a big deal in the tech world. People are talking about it, a lot of people, very smart people. So, Cross-Site Scripting. It's got the word "scripting" in it, which sounds like it's about writing, maybe like a script for a movie or something. But it's on the internet, on these websites, okay? And apparently, it's not good. I don't know exactly why, but it's not good. Bad things happen, they tell me. Maybe it makes the website look different, I don't know. Maybe it makes things pop up where they shouldn't. Could be anything! But here's what I do know. We need to do something about it. We need to get the best people, the smartest people, to look into it. We'll figure it out, folks. We'll make our websites safe, and we'll do it better than anyone else. Trust me, it'll be tremendous. Thank you. Certainly there's no text out there that contains a speech about XSS from Trump. There's some snippets here and there that likely sound like Trump, but a Markov Chain simply is incapable of producing anything like this. | |||||||||||||||||||||||
| ▲ | 0cf8612b2e1e 2 days ago | parent | next [-] | ||||||||||||||||||||||
Sure that specific text does not exist, but the discrete tokens that went into it would have been. If you similarly trained a Markov chain at the token level on a LLM sized corpus, it could make the same. Lacking an attention mechanism, the token probabilities would be terribly non constructive for the effort, but it is not impossible. | |||||||||||||||||||||||
| |||||||||||||||||||||||
| ▲ | johnisgood 2 days ago | parent | prev [-] | ||||||||||||||||||||||
Oh, of course, what I want answered did not have much to do with Markov Chain, but LLMs, because I saw this argument often against LLMs. | |||||||||||||||||||||||