| ▲ | photochemsyn a day ago | |
Long-term memory on top of the base model, but is this idea for local users or for the data-center hosted model used by many different people? P.S. This quote from the paper sounds just like LLM output: > "This memory module provides significantly higher expressive power, allowing the model to summarize large volumes of information without losing important context. The model isn't simply taking notes; it's understanding and synthesizing the entire story. Crucially, Titans doesn’t just passively store data. It actively learns how to recognize and retain important relationships and conceptual themes that connect tokens across the entire input." | ||