Discussion about this post

User's avatar
Neural Foundry's avatar

The signal dilution problem across 32+ layers is a massive engineering challenge that doesn't get enough attention in TTT discussions. Cross-attention at layer 0 is clever because it sidesteps the architectural surgery issue, but the vanishing gradient risk scales exponentially with model size. I built something similar last year using LoRA adapters mid-network and ran into the exact same 'gullible memory' issue when dealing with conflicting user inputs across sessions. The 44.7% vs 34% result is especially interesting becuase it suggests the memory module is functionally acting as a learnable denoising filter rather than just storage.That BABILong setup is brutal btw, forcing strict separation between context and memory actually proves the mechanism works.

Expand full comment
John Michael Thomas's avatar

This looks excellent.

The only comment I'd add is that limitation #2 (The Gullible Memory risk) isn't just about adversarial inputs. It also makes this approach somewhere between less than ideal and unusable for any inputs which change over time. Which, fortunately or unfortunately, a very large amount of information does.

For example, if it learns about a businesses core products & services, and the business discontinues some products and focuses on others, the memory seems likely to continue to retrieve old information. And since that old information was likely built up over time, then the od may very well drown out the new. At a minimum, this would dilute newer, more accurate information; at a maximum it would in some sense hallucinate about the past - even though it has information about the present in its memory as well.

So, eventually, memory architectures will need some mechanism to either automatically purge memories that are no longer relevant, or assign and adjust priorities to different memories over time. If I tell a human that our strategy has changed, they handle this automatically - they understand both that there's a current state and a historical state (both of which are valuable to remember, but for different reasons). LLM memory will eventually need to figure out how to do this as well.

Quick thought: What if the information placed into memory was time stamped, and when the memory contains multiple conflicting hits on a topic, it automatically prioritizes the most recent information over the past information? This seems like it would more closely match how us humans handle changing information over time - and might also provide a way to more easily counter adversarial information planted in the past.

Expand full comment
4 more comments...

No posts

Ready for more?