▲ | fragmede 3 days ago | |||||||||||||||||||||||||||||||||||||
A system that self-updates its weights is so obvious the only question is who will be the first to get there? | ||||||||||||||||||||||||||||||||||||||
▲ | soulofmischief 3 days ago | parent | next [-] | |||||||||||||||||||||||||||||||||||||
It's not always as useful as you think from the perspective of a business trying to sell an automated service to users who expect reliability. Now you have to worry about waking up in the middle of the night to rewind your model to a last known good state, leading to real data loss as far as users are concerned. Data and functionality become entwined and basically you have to keep these systems on tight rails so that you can reason about their efficacy and performance, because any surgery on functionality might affect learned data, or worse, even damage a memory. It's going to take a long time to solve these problems. | ||||||||||||||||||||||||||||||||||||||
▲ | danenania 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
I’m not sure that self-updating weights is really analogous to “continuous learning” as humans do it. A memory data structure that the model can search efficiently might be a lot closer. Self-updating weights could be more like epigenetics. | ||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||
▲ | HarHarVeryFunny 2 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
Sure, it's obvious, but it's only one of the missing pieces required for brain-like AGI, and really upends the whole LLM-as-AI way of doing things. Runtime incremental learning is still going to be based on prediction failure, but now it's no longer failure to predict the training set, but rather requires closing the loop and having (multi-modal) runtime "sensory" feedback - what were the real-world results of the action the AGI just predicted (generated)? This is no longer an auto-regressive model where you can just generate (act) by feeding the model's own output back in as input, but instead you now need to continually gather external feedback to feed back into your new incremental learning algorithm. For a multi-modal model the feedback would have to include image/video/audio data as well as text, but even if initial implementations of incremental learning systems restricted themselves to text it still turns the whole LLM-based way of interacting with the model on it's head - the model generates text-based actions to throw out into the world, and you now need to gather the text-based future feedback to those actions. With chat the feedback is more immediate, but with something like software development far more nebulous - the model makes a code edit, and the feedback only comes later when compiling, running, debugging, etc, or maybe when trying to refactor or extend the architecture in the future. In corporate use the response to an AGI-generated e-mail or message might come in many delayed forms, with these then needing to be anticipated, captured, and fed back into the model. Once you've replaced the simple LLM prompt-response mode of interaction with one based on continual real-world feedback, and designed the new incremental (Bayesian?) learning algorithm to replace SGD, maybe the next question is what model is being updated, and where does this happen? It's not at all clear that the idea of a single shared (between all users) model will work when you have millions of model instances all simultaneously doing different things and receiving different feedback on different timescales... Maybe the incremental learning now needs to be applied to a user-specific model instance (perhaps with some attempt to later share & re-distribute whatever it has learnt), even if that is still cloud based. So... a lot of very fundamental changes need to be made, just to support self-learning and self-updates, and we haven't even discussed all the other equally obvious differences between LLMs and a full cognitive architecture that would be needed to support more human-like AGI. | ||||||||||||||||||||||||||||||||||||||
▲ | tmountain 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
I’m no expert, but it seems like self updating weights requires a grounded understanding of the underlying subject matter, and this seems like a problem current LLM systems. | ||||||||||||||||||||||||||||||||||||||
▲ | imtringued 2 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
I wonder when there will be proofs in theoretical computer science that an algorithm is AGI-complete, the same way there are proofs of NP-completeness. Conjecture: A system that self updates its weights according to a series of objective functions, but does not suffer from catastrophic forgetting (performance only degrades due to capacity limits, rather than from switching tasks) is AGI-complete. Why? Because it could learn literally anything! | ||||||||||||||||||||||||||||||||||||||
▲ | emporas 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
But then it is a specialized intelligence, specialized to altering it's weights. Reinforcement Learning doesn't work as well when the goal is not easily defined. It does wonders for games, but anything else? Someone has to specify the goals, a human operator or another A.I. The second A.I. better be an A.G.I. itself, otherwise it's goals will not be significant enough for us to care. | ||||||||||||||||||||||||||||||||||||||
▲ | fuckaj 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||||||||
True. In the same way as making noises down a telephone line is the obvious way to build a million dollar business. | ||||||||||||||||||||||||||||||||||||||
▲ | 3 days ago | parent | prev [-] | |||||||||||||||||||||||||||||||||||||
[deleted] |