We present EvolveMem, a self-evolving memory architecture that exposes its full retrieval configuration as a structured action space optimized by an LLM-powered diagnosis module. </p>\n","updatedAt":"2026-05-15T02:25:02.956Z","author":{"_id":"684ff37fa383bc5d6b0ff77f","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/0JPr-cd_rxQz3k6rmzBOF.png","fullname":"JiaqiLiu","name":"JiaaqiLiu","type":"user","isPro":false,"isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":5,"isUserFollowing":false}},"numEdits":1,"identifiedLanguage":{"language":"en","probability":0.8672400116920471},"editors":["JiaaqiLiu"],"editorAvatarUrls":["https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/0JPr-cd_rxQz3k6rmzBOF.png"],"reactions":[],"isReport":false}}],"primaryEmailConfirmed":false,"paper":{"id":"2605.13941","authors":[{"_id":"6a068311b1a8cbabc9f098d6","name":"Jiaqi Liu","hidden":false},{"_id":"6a068311b1a8cbabc9f098d7","name":"Xinyu Ye","hidden":false},{"_id":"6a068311b1a8cbabc9f098d8","name":"Peng Xia","hidden":false},{"_id":"6a068311b1a8cbabc9f098d9","name":"Zeyu Zheng","hidden":false},{"_id":"6a068311b1a8cbabc9f098da","name":"Cihang Xie","hidden":false},{"_id":"6a068311b1a8cbabc9f098db","name":"Mingyu Ding","hidden":false},{"_id":"6a068311b1a8cbabc9f098dc","name":"Huaxiu Yao","hidden":false}],"publishedAt":"2026-05-13T00:00:00.000Z","submittedOnDailyAt":"2026-05-15T00:00:00.000Z","title":"EvolveMem:Self-Evolving Memory Architecture via AutoResearch for LLM Agents","submittedOnDailyBy":{"_id":"684ff37fa383bc5d6b0ff77f","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/0JPr-cd_rxQz3k6rmzBOF.png","isPro":false,"fullname":"JiaqiLiu","user":"JiaaqiLiu","type":"user","name":"JiaaqiLiu"},"summary":"Long-term memory is essential for LLM agents that operate across multiple sessions, yet existing memory systems treat retrieval infrastructure as fixed: stored content evolves while scoring functions, fusion strategies, and answer-generation policies remain frozen at deployment. We argue that truly adaptive memory requires co-evolution at two levels: the stored knowledge and the retrieval mechanism that queries it. We present EvolveMem, a self-evolving memory architecture that exposes its full retrieval configuration as a structured action space optimized by an LLM-powered diagnosis module. In each evolution round, the module reads per-question failure logs, identifies root causes, and proposes targeted configuration adjustments; a guarded meta-analyzer applies them with automatic revert-on-regression and explore-on-stagnation safeguards. This closed-loop self-evolution realizes an AutoResearch process: the system autonomously conducts iterative research cycles on its own architecture, replacing manual configuration tuning. Starting from a minimal baseline, the process converges autonomously, discovering effective retrieval strategies including entirely new configuration dimensions not present in the original action space. On LoCoMo, EvolveMem outperforms the strongest baseline by 25.7% relative and achieves a 78.0% relative improvement over the minimal baseline. On MemBench, EvolveMem exceeds the strongest baseline by 18.9% relative. Evolved configurations transfer across benchmarks with positive rather than catastrophic transfer, indicating that the self-evolution process captures universal retrieval principles rather than benchmark-specific heuristics. Code is available at https://github.com/aiming-lab/SimpleMem.","upvotes":21,"discussionId":"6a068311b1a8cbabc9f098dd","ai_summary":"EvolveMem enables adaptive memory systems for LLM agents through self-evolving retrieval mechanisms that autonomously optimize configuration parameters via diagnostic modules and iterative research cycles.","ai_keywords":["long-term memory","LLM agents","retrieval infrastructure","self-evolving memory","structured action space","LLM-powered diagnosis module","meta-analyzer","guarded meta-analyzer","AutoResearch","iterative research cycles","convergence","retrieval strategies","benchmark transfer"]},"canReadDatabase":false,"canManagePapers":false,"canSubmit":false,"hasHfLevelAccess":false,"upvoted":false,"upvoters":[{"_id":"66f6bc97980d52c75c300511","avatarUrl":"/avatars/f7c23c4b09701580b533212ec9b6e306.svg","isPro":false,"fullname":"Yongliang Wu","user":"Liang0223","type":"user"},{"_id":"68a8a0aea2ba0b535569cc95","avatarUrl":"/avatars/d37e55b73627e6486bf485699e2121b1.svg","isPro":false,"fullname":"yexinyu","user":"xinyuyee","type":"user"},{"_id":"684ff37fa383bc5d6b0ff77f","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/0JPr-cd_rxQz3k6rmzBOF.png","isPro":false,"fullname":"JiaqiLiu","user":"JiaaqiLiu","type":"user"},{"_id":"643e9ee6f6bb3c31a26e7bc4","avatarUrl":"/avatars/acfaa7d6a23dada24c86b954c3be116a.svg","isPro":false,"fullname":"Peng Xia","user":"richardxp888","type":"user"},{"_id":"620783f24e28382272337ba4","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/620783f24e28382272337ba4/zkUveQPNiDfYjgGhuFErj.jpeg","isPro":false,"fullname":"GuoLiangTang","user":"Tommy930","type":"user"},{"_id":"699d9c2823c56fbd7d515355","avatarUrl":"/avatars/240639ba7142274b1c6854d21ccbf4c2.svg","isPro":false,"fullname":"Noah Thompson","user":"noahthompson","type":"user"},{"_id":"69830d5915cf39e7565f0e4c","avatarUrl":"/avatars/c783104c693b7f01068a0ab724e964a5.svg","isPro":false,"fullname":"Diego Fernández","user":"my-42","type":"user"},{"_id":"698f8ee4801eed96d9d3555a","avatarUrl":"/avatars/f98f9b01a6cceb0f725185fde6187be6.svg","isPro":false,"fullname":"Q6v1uxi6","user":"q6v1uxi6","type":"user"},{"_id":"698f93bbe1b31e94e9dcf727","avatarUrl":"/avatars/ac8e672f70d76f400bcb1e0952df6b6f.svg","isPro":false,"fullname":"Ylnmt1b25eu45","user":"ylnmt1b25eu45","type":"user"},{"_id":"69bcdf2f685c38830c62723f","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/noauth/RY6sEuQB49XqNalWZocqk.jpeg","isPro":false,"fullname":"준서 윤","user":"ljingyi","type":"user"},{"_id":"69a2c54ad628b8476143a650","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/noauth/UTLTAUjbCA1hl6Mo4ECZ7.png","isPro":false,"fullname":"Илья Алексеев","user":"xiao-wenhao79","type":"user"},{"_id":"69bb774375169b73cc37051d","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/noauth/Zs--Z-sM9LG-CWy2M_NLA.png","isPro":false,"fullname":"結翔 鈴木","user":"zyichen64","type":"user"}],"acceptLanguages":["en"],"dailyPaperRank":0,"markdownContentUrl":"https://huggingface.co/buckets/huggingchat/papers-content/resolve/2605/2605.13941.md"}">
EvolveMem:Self-Evolving Memory Architecture via AutoResearch for LLM Agents
Abstract
EvolveMem enables adaptive memory systems for LLM agents through self-evolving retrieval mechanisms that autonomously optimize configuration parameters via diagnostic modules and iterative research cycles.
AI-generated summary
Long-term memory is essential for LLM agents that operate across multiple sessions, yet existing memory systems treat retrieval infrastructure as fixed: stored content evolves while scoring functions, fusion strategies, and answer-generation policies remain frozen at deployment. We argue that truly adaptive memory requires co-evolution at two levels: the stored knowledge and the retrieval mechanism that queries it. We present EvolveMem, a self-evolving memory architecture that exposes its full retrieval configuration as a structured action space optimized by an LLM-powered diagnosis module. In each evolution round, the module reads per-question failure logs, identifies root causes, and proposes targeted configuration adjustments; a guarded meta-analyzer applies them with automatic revert-on-regression and explore-on-stagnation safeguards. This closed-loop self-evolution realizes an AutoResearch process: the system autonomously conducts iterative research cycles on its own architecture, replacing manual configuration tuning. Starting from a minimal baseline, the process converges autonomously, discovering effective retrieval strategies including entirely new configuration dimensions not present in the original action space. On LoCoMo, EvolveMem outperforms the strongest baseline by 25.7% relative and achieves a 78.0% relative improvement over the minimal baseline. On MemBench, EvolveMem exceeds the strongest baseline by 18.9% relative. Evolved configurations transfer across benchmarks with positive rather than catastrophic transfer, indicating that the self-evolution process captures universal retrieval principles rather than benchmark-specific heuristics. Code is available at https://github.com/aiming-lab/SimpleMem.
Community
We present EvolveMem, a self-evolving memory architecture that exposes its full retrieval configuration as a structured action space optimized by an LLM-powered diagnosis module.
Upload images, audio, and videos by dragging in the text input, pasting, or clicking here.
Tap or paste here to upload images
Cite arxiv.org/abs/2605.13941 in a model README.md to link it from this page.
Cite arxiv.org/abs/2605.13941 in a dataset README.md to link it from this page.
Cite arxiv.org/abs/2605.13941 in a Space README.md to link it from this page.
Discussion (0)
Sign in to join the discussion. Free account, 30 seconds — email code or GitHub.
Sign in →No comments yet. Sign in and be the first to say something.