▲ | robviren 7 days ago | ||||||||||||||||||||||
I feel it is interesting but not what would be ideal. I really think if the models could be less linear and process over time in latent space you'd get something much more akin to thought. I've messed around with attaching reservoirs at each layer using hooks with interesting results (mainly over fitting), but it feels like such a limitation to have all model context/memory stuck as tokens when latent space is where the richer interaction lives. Would love to see more done where thought over time mattered and the model could almost mull over the question a bit before being obligated to crank out tokens. Not an easy problem, but interesting. | |||||||||||||||||||||||
▲ | dkersten 7 days ago | parent | next [-] | ||||||||||||||||||||||
Agree! I’m not an AI engineer or researcher, but it always struck me as odd that we would serialise the 100B or whatever parameters of latent space down to maximum 1M tokens and back for every step. | |||||||||||||||||||||||
▲ | CuriouslyC 7 days ago | parent | prev | next [-] | ||||||||||||||||||||||
They're already implementing branching thought and taking the best one, eventually the entire response will be branched, with branches being spawned and culled by some metric over the lifetime of the completion. It's just not feasible now for performance reasons. | |||||||||||||||||||||||
▲ | vonneumannstan 7 days ago | parent | prev [-] | ||||||||||||||||||||||
>I feel it is interesting but not what would be ideal. I really think if the models could be less linear and process over time in latent space you'd get something much more akin to thought. Please stop, this is how you get AI takeovers. | |||||||||||||||||||||||
|