| ▲ | erichocean 2 hours ago | |||||||
> Models do not (broadly speaking) learn over time. They can be tuned by their operators, or periodically rebuilt with new inputs or feedback from users and experts. Models also do not remember things intrinsically: when a chatbot references something you said an hour ago, it is because the entire chat history is fed to the model at every turn. Longer-term “memory” is achieved by asking the chatbot to summarize a conversation, and dumping that shorter summary into the input of every run. This is the part of the article that will age the fastest, it's already out-of-date in labs. | ||||||||
| ▲ | lamasery an hour ago | parent | next [-] | |||||||
I'm struggling to reckon how that can even possibly be true, unless we're counting automation of the "dumping that shorter summary into the input of every run" thing. I can imagine it being true with models so small that each user could afford to have their own, but not with big shared models like what're getting used for all the major services. Is that what you mean? | ||||||||
| ||||||||
| ▲ | qsera an hour ago | parent | prev | next [-] | |||||||
Source? | ||||||||
| ▲ | dgb23 an hour ago | parent | prev [-] | |||||||
In what way? | ||||||||