Remix.run Logo
usernametaken29 6 hours ago

> δ-mem compresses past information into a fixed-size state matrix updated by delta-rule learning

This doesn’t solve the capacity problem of memory. You can cram more into one context window, but then again you need to associate them with input queries. That’s very hard because slight variations in input create hugely different activations. So really, it doesn’t improve caching. This paper might do a thing or two approximating the compression limit for context windows, but there’s a fundamental limit on how much information can go into it. What you really need is contextual search, as in, different events and objects with the same abstractions and semantic lead to same response, so you can cache effectively… on this front the paper does little to improve “memory” in a meaningful way

jsemrau 4 hours ago | parent | next [-]

I am currently working on deep context query which uses dynamically generated regex to pull only the relevant context blocks. By using lightweight RegEx pattern matching to detect semantic intent and filter structured context sections accordingly, you avoid the attention degradation that comes from stuffing semantically redundant information into the window

https://jdsemrau.substack.com/p/tokenmaxxing-and-optimizing-...

structuredPizza 3 hours ago | parent [-]

The more real world use cases we see, the more we see the use of a well thought out regex as a bridge from probabilistic to deterministic.

jandrese 4 hours ago | parent | prev | next [-]

So instead of a FIFO approach to memory management it instead continually degrades the existing data the more you put in? Details start getting lost or mangled more and more over time?

trollbridge an hour ago | parent [-]

That’s basically what happens.

As you hit the limits and try to compact the context, etc., things get more erratic.

kordlessagain 4 hours ago | parent | prev | next [-]

Like Ferricula: https://deepbluedynamics.com/ferricula (site/docs still in progress).

4 hours ago | parent | prev [-]
[deleted]