Preprint
Article

This version is not peer-reviewed.

LLM Agent Memory: A Survey from a Unified Representation–Management Perspective

Submitted:

03 March 2026

Posted:

04 March 2026

Read the latest preprint version here

Abstract
Large language models (LLMs) face significant challenges in sustaining long-term memory for agentic applications due to limited context windows. To address this limitation, many work has proposed diverse memory mechanisms to support long-term, multi-turn interactions, leveraging different approaches tailored to distinct memory storage objects, such as KV caches. In this survey, we present a unified taxonomy that organizes memory systems for long-context scenarios by decoupling memory abstractions from model-specific inference and training methods. We categorize LLM memory into three primary paradigms: natural language tokens, intermediate representations and parameters. For each paradigm, we organize existing methods by three management stages, including memory construction, update, and query, so that long-context memory mechanisms can be described in a consistent way across system designs, with their implementation choices and constraints made explicit. Finally, we outline key research directions for long-context memory system design.
Keywords: 
;  ;  

1. Introduction

Large Language Models (LLMs) are deployed in long-context and interactive settings, such as multi-turn dialogue, task-oriented assistants, and agent-based systems, where task completion requires information to be retained and reused across extended temporal spans rather than within a single prompt (Qian et al., 2023; Gao et al., 2023;Wang et al., 2023). As input sequences grow to thousands or even millions of tokens, performance often degrades on tasks that require entity tracking, logical consistency, or recall of task-relevant facts across long interaction histories (Gao et al., 2024; Zhang et al., 2024; Wu et al., 2025). These observations indicate that the core difficulty in long-context and multi-turn scenarios lies not only in how much information the model can observe, but in how information is selectively retained, retrieved, and integrated during inference (Shinwari and Usama, 2025; Maharana et al., 2024;Wan and Ma, 2025). This has motivated the introduction of memory as a key abstraction in LLM-based systems, enabling task-relevant information to persist beyond the immediate context window and to be reused for future reasoning and decision-making.
Many existing studies have explored diverse LLM memory mechanisms to support long-context reasoning and multi-turn interaction (Shinn et al., 2023; Zhong et al., 2023; Modarressi et al., 2023; Zhong et al., 2024; Qian et al., 2024). Broadly, memory enables task-relevant information to persist beyond the immediate context window and be reused across interactions, often drawing loose analogies to human cognition, where short-term working memory supports immediate reasoning and long-term memory enables recall over time (Baddeley, 2007; Budson and Kensinger, 2023). The analogy serves as a useful intuition that effective memory depends not only on capacity, but also on selective access and integration (Gao et al., 2024; Wu et al., 2025). Corresponding surveys have reviewed this literature from multiple perspectives, including LLM-based agents and long-term interaction (Zhang et al., 2025), long-context modeling and long-term memory (Wu et al., 2025; Huang et al., 2023; Jiang et al., 2024), personalization (Liu et al., 2025), and system-level efficiency such as inference-time memory management (Pan and Li, 2025; LI et al., 2025; Luohe et al.). These works underscore the central role of memory across application, modeling, and system dimensions.
In this survey, we take a system-level, operation-centric view of LLM memory. Instead of cataloging methods by tasks or modules, we organize the literature around a unified abstraction that connects task knowledge requirements to memory representations through shared management interfaces. Figure 1 sketches this logic from applications (top) to recurring representation choices (middle) and a common management interface over long interactions (bottom). Under this view, diverse methods can be compared by three questions: what is stored, where it is stored, and how it is operated. We structure the survey along two recurring dimensions:
  • Memory representation describes where and in what form information resides: token-level memory in the input context, intermediate latent memory as inference-time states (e.g., Key–Value caches), and parameter-level memory in model weights via adaptation or editing.
  • Memory management describes how memory is operated over time to satisfy task requirements under practical constraints. Across representations, we observe a shared interface of three core operations: memory construction (what to store and how to structure it), memory update (how to maintain, consolidate, or remove stored content), and memory query (how to select and integrate relevant information during inference).
Following this organization, Sections 2–4 review memory mechanisms by aligning each representation with the same construction–update–query interface and its key trade-offs. Section 5 synthesizes insights across memory types and outlines open challenges for reliable long-context and multi-turn LLM memory.

2. Natural Language Tokens as Memory

Natural language tokens are the most explicit memory format that leverages the context window for non-invasive information reuse. However, it is limited by O ( n 2 ) computational costs and the tendency to overlook information buried in the middle of long sequences (Liu et al., 2024). To address these bottlenecks, Retrieval-Augmented Generation and Agentic Memory transform fragmented histories into structured, retrievable knowledge.
Preprints 201223 i001

2.1. Retrieval-Augmented Generation (RAG)

RAG here refers to retrieval-based memory systems centered around an external a vector database, rather than full agent frameworks with explicit environment interaction. RAG offers a cost-effective way to extend long-context behavior by retrieving evidence and inserting it into the input.

Memory Construction.

Construction in RAG specifies what evidence is stored and how it is indexed for retrieval. For what to store, early RAG systems mainly ingest unstructured corpora and domain datasets (Li et al., 2023; Yan et al., 2024). More recent settings extend to semi-structured documents such as PDFs, where table-aware processing is often required (e.g., table-to-text normalization or Text-to-SQL-style querying (Zha et al., 2023; Luo et al., 2023)). RAG can also build on structured sources such as knowledge graphs, where KnowledGPT (Wang et al., 2023) and G-Retriever (He et al., 2024) improve graph grounding and evidence selection via soft prompting and PCST-style subgraph optimization. Some approaches leverage the model’s internal knowledge to reduce retrieval overhead or bootstrap contexts, for example by selectively invoking retrieval (Wang et al., 2023), generating aligned contexts (Yu et al., 2022), or building unbounded memory pools (Cheng et al., 2023).
For how to index, most RAG systems first define the index unit, which determines the retrieval granularity: coarser units provide broader context but introduce redundancy, while finer units (e.g., tokens or phrases) improve precision at the risk of losing supporting context (Shi et al., 2023; Yu et al., 2023; Chen et al., 2023; Jin et al., 2023; Wang et al., 2023). A common approach splits documents into fixed-length spans (typically 100-512 tokens) and builds an index over these units (Teja, 2023). Variants such as recursive splitting, sliding windows, and Small2Big expansion preserve local coherence by attaching surrounding context (Langchain, 2023; Yang, 2023). Beyond unit definition, many systems enrich indexed entries with auxiliary fields (e.g., page numbers or generated cues) to improve matching (Gao et al., 2022). Indexes can further adopt hierarchical or graph-based structures to support multi-step evidence aggregation across documents (Wang et al., 2023). Finally, embeddings govern retrieval, encoding units as sparse or dense representations and ranking candidates via lexical matching (e.g., BM25) or embedding similarity, often using dense retrievers such as AngIE, Voyage, or BGE (Li and Li, 2023; VoyageAI, 2023; BAAI, 2023). Hybrid sparse-dense retrieval is commonly used to improve robustness, especially for zero-shot queries or rare entities (Zhang et al., 2025).

Memory Update.

RAG updates are primarily realized by modifying the external store and its index, including adding new documents, re-chunking and re-embedding content, refreshing metadata, and restructuring the index (e.g., moving from flat chunking to hierarchical or KG-based organization) to better match evolving domains and query patterns (Langchain, 2023; Yang, 2023; Gao et al., 2022; Wang et al., 2023). Some methods also adjust retrieval invocation, for example, deciding when retrieval is necessary or generating retrieval-aligned contexts to stabilize generation (Wang et al., 2023; Yu et al., 2022; Cheng et al., 2023).

Memory Query.

To improve robustness under underspecified or noisy inputs, existing methods enhance query quality from several complementary angles. One line focuses on query reformulation, including expansion (multi-query), decomposition (e.g., least-to-most prompting), validation (CoVe), and transformations (Gao et al., 2022; Zhou et al., 2023; Dhuliawala et al., 2023; Ma et al., 2023; Peng et al., 2023), where step-back prompting (Zheng et al., 2024) further abstracts the query to retrieve complementary evidence. Another line emphasizes query routing, selecting different retrieval pipelines based on meta-data or semantic routers to enable hybrid strategies (Wang et al., 2025). A third line develops multi-step and controllable querying, where retrieval and generation are composed into modular pipelines (e.g., RRR, GenRead, RECITE (Yu et al., 2022; Ma et al., 2023; Sun et al., 2022)) and governed by dynamic controllers such as DSP, FLARE, and Self-RAG, sometimes combined with fine-tuning or reinforcement learning (Khattab et al., 2022; Jiang et al., 2023; Asai et al., 2023; Ke et al., 2024; Lin et al., 2023).

2.2. Agentic Memory

In contrast to RAG, which retrieves evidence from an external corpus, agentic memory targets stateful, multi-turn interaction and accumulates textual records of observations, actions, and outcomes across steps to support long-horizon reasoning.

Memory Construction.

Memory construction converts raw interaction traces into compact, retrievable textual units. A common baseline is to summarize dialogue histories, key events, and stable facts (e.g., user preferences or task states), as in MemoryBank and RET-LLM (Modarressi et al., 2023; Zhong et al., 2024). To enable efficient access, constructed memories are organized using explicit structures, including key-value slots (Modarressi et al., 2023; Salama et al., 2025; Xi et al., 2024), semantic vector representations (Zhong et al., 2024; Pan et al., 2025; mem0ai, 2024), and relation-aware graphs capturing dependencies among memory fragments, e.g., CGSN, GraphReader, HippoRAG (Nie et al., 2022; Li et al., 2024; Gutiérrez et al., 2024). Construction is strengthened by auxiliary signals such as timestamps, summaries, or factual tags to improve retrievability, e.g., LongMemEval (Wu et al., 2024), or by organizing memories along temporal and causal axes for context-sensitive navigation (Theanine (iunn Ong et al., 2025)). To control storage and inference cost, some systems apply token-level pruning, summarization, or soft compression at construction time, and reuse frequent contexts via prompt caching (Jiang et al., 2023; Chevalier et al., 2023; Liu et al., 2023; Gim et al., 2024).

Memory Update.

Memory update refines, consolidates, and revises stored content as interactions proceed. Early work controls growth through periodic summarization or restructuring, using explicit summarizers such as MemoryBank and ChatGPT-RSum (Zhong et al., 2024; Wang et al., 2025) or prompt-based extraction of salient topics as in MemoChat (Lu et al., 2023). Beyond compression, many methods treat updating as a reasoning-driven step: agents reflect on past actions and outcomes and write back reusable artifacts, including action-thought traces (Yao et al., 2022), self-critique and revision notes (Shinn et al., 2024), distilled reasoning templates (Yang et al., 2024), and workflow-level records (Wang et al., 2024). Experience-based agents refine memory through trial-and-error interaction and feedback, revising what to store and how to use it (Liu et al., 2023; Zhu et al., 2023; Wang et al., 2023; Yao et al., 2023; Zhao et al., 2024; Li et al., 2024). More recent systems emphasize memory evolution, allowing memories to be edited, linked, or reorganized over time. Examples include A-MEM’s interconnected note-style growth (Xu et al., 2025; Kadavy, 2021), temporally adaptive structures such as Synapse, R2I, and SCM (Zheng et al., 2024; Samsami et al., 2024; Wang et al., 2024), as well as selective editing, recursive summarization, memory blending, and self-reflective verification to maintain relevance and consistency (Bae et al., 2022; Wang et al., 2025; Kim et al., 2024; Sun et al., 2024).

Memory Query.

Memory query determines how relevant entries are selected and integrated to support ongoing reasoning. Existing methods improve query effectiveness from complementary perspectives. Query-centered approaches reformulate or refine the query itself, for example via forward-looking rewriting or iterative refinement (Jiang et al., 2023; Jang et al., 2024). Memory-centered approaches enhance ranking and selection through richer indexing signals and reranking strategies, as explored in LongMemEval and personalized long-term memory retrieval (Wu et al., 2024; Du et al., 2024). Finally, event- and structure-aware retrieval leverages temporal, causal, or relational structure to traverse memory graphs or timelines, enabling coherent recall across long interaction histories (e.g., LoCoMo, CC, MSC, and graph-based multi-hop retrieval (Maharana et al., 2024; Qian et al., 2024; Gutiérrez et al., 2024; Jang et al., 2023; Xu et al., 2021)). Together, these strategies highlight that effective agentic memory query relies not only on semantic similarity, but also on adaptive, context-aware access over evolving memory states.

3. Intermediate Latent as Memory

Intermediate latent (IL) refer to inference-time internal representations in LLMs, such as attention activations or other continuous vectors, that can be cached and reused as memory. This section focuses on two forms of intermediate latent memory: the Key-Value (KV) cache, and other vector-based memory mechanisms.
Preprints 201223 i002

3.1. KV Cache as Memory

Memory Construction.

The KV cache stores intermediate key and value vectors produced by the attention mechanism to accelerate autoregressive decoding. Its construction is implicit and deterministic: during prefilling, KV pairs for all prompt tokens are computed and cached; during decoding, only KV pairs for newly generated tokens are appended, while attention reuses the cached states. This reduces per-token complexity from O ( n 2 ) to O ( n ) and is a core component of modern LLMs (Touvron et al., 2023a,2; Grattafiori et al., 2024; DeepSeek-AI et al., 2025a,2). Unlike token-based memory, KV cache construction involves no explicit selection of what to store: all tokens initially contribute full KV representations, and memory management is deferred to post-construction. From a memory perspective, the KV cache thus provides a transient, high-fidelity record of recent context tightly coupled to attention computation.

Memory Update.

Memory update for KV cache focuses on controlling storage cost while preserving attention quality. Existing methods fall into several recurring strategies.
Eviction and dropping selectively discard KV entries using static patterns or dynamic importance signals. Representative approaches include fixed sparsity schemes (Xiao et al., 2024; Han et al., 2024), layer-wise retention (Wu and Tu, 2024), and attention- or query-aware dropping such as H2O, FastGen, Radar, and NACL (Zhang et al., 2023; Ge et al., 2024; Hao et al., 2025; Chen et al., 2024). Variants further exploit attention statistics across heads, layers, or tasks (Liu et al., 2023; Devoto et al., 2024; Yao et al., 2024; Jiang et al., 2024; Zhong et al., 2024; Zhou et al., 2025).
Merging and semantic compression reduce redundancy by consolidating similar KV entries rather than removing them outright. This includes similarity-based merging (Liu et al., 2024; Kim et al., 2024; Agarwal et al., 2024) and semantic-preserving compression at token, chunk, or sentence granular-ity (Zhang et al., 2025; Liu et al., 2024,2; Zhu et al., 2025).
Quantization and low-rank approximation lower per-entry storage cost by reducing numerical precision or exploiting low-rank structure. Representative methods apply low-bit or asymmetric quantization (Liu et al., 2024; Duanmu et al., 2024; Dong et al., 2024; Hooper et al., 2024; Zhang et al., 2024; Li et al., 2025), attention- or layer-aware scaling (Lin et al., 2025; Yang et al., 2024), and low-rank compression with residual preservation (Dong et al., 2024; Saxena et al., 2024; Kang et al., 2024). Dynamic precision schemes further adapt quantization to runtime conditions (Sheng et al., 2023; Zhao et al., 2024; He et al., 2024).
System- and task-aware allocation adapts KV storage to deployment constraints and task characteristics. Examples include disaggregated and multi-GPU KV storage (Chen et al., 2024; Li et al., 2025), layer- or chunk-level budget assignment (Liu et al., 2025; Yang et al., 2024), and preference- or workload-aware allocation strategies (Zhu et al., 2025).

Memory Query.

Memory query determines how cached KV states are accessed during attention when attending to all entries is inefficient or unnecessary. 1) KV selection restricts attention to a subset of relevant entries using query-dependent signals: QUEST and TokenSelect estimate token importance via attention statistics or learned predictors (Tang et al., 2024; Wu et al., 2025), Selective Attention prunes targets across heads and layers (Leviathan et al., 2025), and RetrievalAttention treats KV states as retrievable items using approximate nearest neighbor search (Liu et al., 2024). 2) KV reuse avoids redundant computation by sharing cached states across overlapping contexts or requests. Prefix-based reuse organizes caches with tree structures, as in RadixAttention and ChunkAttention (Zheng et al., 2024; Ye et al., 2024), while cross-request reuse shares KV states based on semantic similarity (Yang et al., 2025; Agarwal et al., 2025; Tan et al., 2025). More recent systems extend reuse across retrieval and reranking stages in RAG pipelines (Yang et al., 2025; Yao et al., 2025; Hu et al., 2025; Zhu et al., 2025; An et al., 2025; Jiang et al., 2025), or externalize KV cache to support long-context inference beyond a single device (Wu et al., 2022; Tworkowski et al., 2023; Di et al., 2025).

3.2. Other Vectors as Memory

External Vectors.

External vector memory augments LLMs with a separate vector store that retains intermediate latents for retrieval and reuse, alleviating the quadratic cost of long-context attention (Al Adel and Burtsev, 2021). Early work explored sentence-level memory slots for sequence modeling, while later systems such as kNN-LM and the Memorizing Transformer leveraged pretrained embedding spaces and internal representations to enable scalable retrieval over large memory banks (Wu et al., 2022; Khandelwal et al., 2019). Subsequent designs, including MemGPT and Neurocache, maintain vector caches supporting dynamic retrieval and update for long-context or multi-session tasks (Packer et al., 2023; Safaya and Yuret, 2024). More recent architectures introduce structured or associative memory modules, such as CAMELoT, consolidating token representations while balancing novelty and recency (He et al., 2024), and MemOS or Memory3, which externalize knowledge via vector memory with metadata or sparsification, without modifying core model parameters (Li et al., 2025; Yang et al., 2024).

Steering Vectors.

Steering vectors function as intermediate memory for behavioral control rather than factual storage. Unlike interaction-heavy KV caches or external memories, they encode persistent biases as directions in activation space, originating from PPLM (Dathathri et al.). These vectors modulate hidden states to achieve alignment and interpretability via either contrastive or optimization-based approaches. Contrastive methods derive steering vectors from activation differences between datasets exhibiting desired versus undesired behaviors, encoding behavioral preferences as stable directions in activation space. Representative studies demonstrate control over sentiment, toxicity, refusal, or factuality using single or multiple contrastive prompts (Turner et al., 2023; Liu et al., 2023; Zou et al., 2023; Arditi et al., 2024). While effective, these approaches often require carefully constructed contrastive data and may capture spurious correlations that limit robustness and generalization (Chugtai and Bushnaq, 2025). Optimization-based methods instead learn steering vectors by optimizing simple objectives, such as maximizing target sequence likelihood or applying lightweight affine transformations to hidden states, sometimes with only a single example (Subramani et al., 2022; Hernandez et al., 2023; Dunefsky and Cohan, 2025; Mack and Turner, 2024). More recent work explores probe-based or low-shot steering to induce truthfulness, suppress refusal, or support personalization, though empirical effectiveness varies across models and tasks (Li et al., 2024; Turner et al., 2025; Cao et al., 2024).

4. Parameter as Memory

Preprints 201223 i003

Memory Construction. 

Unlike token- or cache-based memory that relies on input context and runtime storage, parametric memory encodes knowledge directly into model weights through pretraining or fine-tuning, providing long-term and context-independent storage. The capacity and structure of this memory are strongly influenced by training-time factors, including training data composition and augmentation, sequence or context length, and model scale. Data augmentation strategies such as rephrasing, reordering, or stylistic transformation can significantly increase memorization strength (Allen-Zhu and Li, 2024), while data duplication leads to superlinear memorization effects and raises privacy concerns (Carlini et al., 2021; Lee et al., 2022; Kandpal et al., 2022). Longer training sequences expose models to richer contextual dependencies, increasing the likelihood of verbatim recall (Carlini et al., 2023; Wang et al., 2024). Model size further amplifies parametric memory capacity, with memorization scaling approximately log-linearly with parameter count (Carlini et al., 2023; Tirumala et al., 2022; Freeman et al., 2024). Mechanistic analyses suggest that this constructed memory is not uniformly distributed: MLP layers can be interpreted as key-value memories (Geva et al., 2021), and factual knowledge may localize to specific neurons (Dai et al., 2021), providing a structural basis for later updates.

Memory Update. 

Memory update at the parameter level concerns how knowledge embedded in model weights can be modified, extended, or reorganized without reconstructing parametric memory from scratch. Unlike construction, which writes memory through large-scale training, update mechanisms aim to incorporate new information, personalize behavior, or combine task knowledge while mitigating interference with existing parameters.
Continual learning provides a foundational view of parametric memory update by explicitly addressing catastrophic forgetting (Wang et al., 2024). Regularization-based methods constrain updates to parameters deemed critical for previously learned knowledge, as in EWC, TaSL, SELF-PARAM, and POCL (Kirkpatrick et al., 2017; Feng et al., 2024; Wang et al.; Wu et al., 2024). Replay-based strategies instead reinforce memory by reintroducing past samples or synthetic pseudo-data, enabling retention without full retraining; for example, DSI++ maintains retrieval performance via generative replay with pseudo-queries (Mehta et al., 2022). Agent-centric extensions such as LSCS further adapt continual learning to interactive settings, incrementally encoding external experiences into parameters over time (Wang et al., 2024).
Parameter-efficient fine-tuning (PEFT) updates parametric memory by introducing lightweight, task- or user-specific adaptations while freezing the backbone (Han et al., 2024). This paradigm supports personalization and long-term adaptation with reduced computational cost. Representative systems encode character traits or personal histories into parameters (e.g., Character-LLM (Shao et al.)), compress evolving user memory through lifelong personal models (AI-Native Memory (Shang et al., 2024)), or enhance dialogue coherence via episodic parametric memory (MemoRAG (Qian et al., 2024), Echo (Liu et al., 2025)).
Model merging updates parametric memory by combining multiple pretrained or fine-tuned models without access to original data. Basic parameter averaging, widely used in federated learning (e.g., FedAvg), offers simplicity and efficiency but often degrades performance due to parameter conflicts (McMahan et al., 2017; Marczak et al., 2024). To address this, weighted merging prioritizes important parameters using Taylor approximations, task vectors, or Fisher information (Lee et al., 2019; Qu et al., 2022; Matena and Raffel, 2022; Jhunjhunwala et al., 2024; Daheim et al., 2024). Subspace-based approaches further mitigate conflicts by pruning or masking parameters before merging, leveraging over-parameterization to preserve task-relevant memory (e.g., TIES, DARE, Model Breadcrumbs, TALL-masks) (Yadav et al., 2023; Yu et al., 2024; Davari and Belilovsky, 2023; Wang et al., 2024). Routing-based methods generalize merging to inference time, dynamically selecting or weighting parameters or experts based on inputs, as in MoE-style or soft-routing frameworks (Shazeer et al., 2017; Muqeeth et al., 2024; Lu et al., 2024; Tang et al., 2024). LoRA-based routing further enables dynamic composition of low-rank updates, although decomposition-induced degradation remains a concern (Huang et al., 2023; Wei et al., 2025; Lai et al., 2025).
Task arithmetic treats parameter updates as vector operations in weight space, enabling addition or subtraction of task-specific knowledge (Ilharco et al., 2022). Methods like TIES and AdaMerging explicitly resolve conflicts by trimming low-magnitude parameters or learning adaptive merging coefficients (Yadav et al., 2023; Yang et al., 2024), while TwinMerge refines task fusion through supervised low-rank adaptation (Lu et al., 2024). These approaches frame parameter update as algebraic composition, balancing flexibility and sensitivity to hyperparameters and task compatibility.
Model editing focuses on precise, targeted updates to parametric memory, enabling the insertion, modification, or deletion of specific facts (Wang et al., 2024a). Recent systems such as MemoryLLM support self-updating and long-term retention, while WISE introduces a dual-memory design that separates stable pretrained knowledge from edited content and routes queries accordingly (Wang et al., 2024b,2). Compared to broader adaptation methods, model editing offers fine-grained control over parametric memory, but often relies on assumptions about knowledge localization and may accumulate errors over repeated edits.

Memory Query. 

Querying parametric memory differs fundamentally from querying token-level or KV-cache memory. Rather than retrieving stored entries, parametric memory is accessed implicitly through forward computation and analyzed via memorization phenomena. Exact memorization refers to verbatim reproduction of training sequences under suitable prompts (Carlini et al., 2021,2; Nasr et al., 2023), while approximate memorization captures semantic or structural similarity without exact copying (Ippolito et al., 2023). Prompt-based memorization further shows that carefully designed prompts can elicit stored content from partial prefixes, revealing the conditional nature of parametric recall (Biderman et al., 2023).

5. Discussion

From Knowledge Requests to Memory Backends. 

Table 1 links task knowledge requirements to representation choice by showing what dominant management challenge are. We summarize requirements using two human memory-inspired axes (Tulving and Donaldson, 1972; Begg, 1984; Squire, 2009): retention (short-term vs. long-term) and functional form (episodic, semantic, procedural); details are in Appendix B. Token-level memory primarily serves short-term explicit content, but is query-limited under long contexts, where selecting and integrating the right evidence becomes the bottleneck (Liu et al., 2024). Intermediate latent memory supports short-horizon continuity, but is update-limited because gains depend on cache budgeting under fixed capacity (Xiao et al., 2023). Parametric memory supports long-term consolidation, but is write-limited: construction and update are costly and must control interference and forgetting (Kirkpatrick et al., 2017; Meng et al., 2023).

Unified Interfaces as System Glue. 

The table suggests a simple system lesson: representation choice shifts the bottleneck to a different operation, so memory should be engineered through a unified construction-update-query interface rather than as representation-specific pipelines. With shared interfaces, systems can mix backends (tokens for editable evidence, caches for session continuity, parameters for durable consolidation) while keeping a consistent control plane for when to store, how to maintain, and what to use at inference time. This also improves portability across tasks and deployments by decoupling task requirements from backend choice, and makes failures easier to diagnose by localizing them to query selectivity (Liu et al., 2024), cache policy (Xiao et al., 2023; Zhang et al., 2023), or unsafe writes (Meng et al., 2023).

Future Directions 

Specialized Memory Structures. The growing scale and diversity of LLM memory workloads expose limitations of general-purpose memory hierarchies for long-context and agentic applications Yao et al. (2022). Memory mechanisms such as KV caches Pope et al. (2022), vector databases Lewis et al. (2020), and graph-structured memories Park et al. (2023) exhibit heterogeneous access patterns and update behaviors, motivating specialized memory structures for different abstractions. Meanwhile, the high cost of data movement highlights software–hardware co-design, where memory-aware algorithms and hardware support jointly reduce latency and energy Wulf and McKee (1995); Jouppi et al. (2017); Dao et al. (2022). Future work should explore dedicated memory layouts and tighter software–hardware integration to support memory construction, update, and query at scale Yu et al. (2022); Zhong et al. (2024).
Unified Training–Inference Systems. Current LLM systems largely separate training from inference, limiting adaptation to evolving users and environments Brown et al. (2020); Bommasani et al. (2022). While this survey focuses on inference memory, long-term agentic settings increasingly blur this boundary through continual learning and personalization during deployment Shinn et al. (2023); Park et al. (2023). This trend motivates unified training–inference designs that integrate memory management with lightweight updates and inference-time adaptation while mitigating catastrophic forgetting Hu et al. (2021). Future work should explore integrated architectures to enable continuous learning and personalized behavior.
Cross-Domain Methodology Transfer. LLM memory challenges often parallel classic problems in operating systems (OS) and databases. For example, KV cache management adopts OS-level paradigms like paging, eviction, and tiered hierarchies to handle limited memory Zhang et al. (2023); Kwon et al. (2023); Xiao et al. (2024). RAG systems leverage database techniques for indexing, query optimization, and execution planning Asai et al. (2023); Karpukhin et al. (2020); Khattab and Zaharia (2020); Izacard et al. (2023). Furthermore, distributed systems and cloud computing inspire solutions for scaling long-context workloads through partitioning and remote memory Zhong et al. (2024); Fu et al. (2024); Jin and Wu (2025). These established paradigms provide a foundational blueprint for scalable and efficient LLM memory orchestration.

6. Conclusion

This survey presents a unified management view of LLM memory that links task knowledge requirements to concrete memory representations through shared interfaces, including construction, update, and query. By treating memory as a system-level capability rather than a collection of task-specific tricks, the survey provides a coherent way to compare diverse designs, reason about effectiveness–efficiency trade-offs, and guide the composition of hybrid memory backends for long-horizon agents. We hope this perspective helps standardize how future work specifies requirements, evaluates memory behavior over extended interaction, and designs reusable memory components that remain reliable under realistic deployment constraints.

Limitations

This survey proposes a unified representation–management abstraction to organize LLM memory, but the fast pace of the field means our coverage may lag behind the newest systems and some industrial practices are discussed only at a high level. Our taxonomy is also a simplification: many methods span multiple representations and the boundaries between construction, update, and query can blur in long-horizon agents. Finally, quantitative comparisons across papers remain limited due to inconsistent tasks, models, evaluation protocols, and deployment settings, so our synthesis emphasizes recurring design trade-offs and failure modes rather than a unified benchmark ranking.

Appendix A. Related Surveys

The rapid development of LLMs has triggered many surveys on LLM-based agents, which study how an LLM can perceive, act, accumulate knowledge, and adapt over time. Early reviews (e.g., Wang et al. (2023)) organize agent research by how agents are built, where they are used, and how they are evaluated. Later surveys expand the scope with different taxonomies and emphases Xi et al. (2023); Zhao et al. (2023); Cheng et al. (2024); Ge et al. (2023). There are also focused reviews on key capabilities and settings, such as multimodal agents Durante et al. (2024), planning Huang et al. (2024), multi-agent interaction Guo et al. (2024), and personal assistant applications Li et al. (2024). These works provide useful summaries of agent pipelines, but memory is usually treated as one module among many, and is rarely analyzed as a first-class system component with a unified interface and lifecycle.
A separate line of surveys summarizes how LLMs are applied to specific domains. In information retrieval and extraction, surveys cover LLM-based query processing Zhu et al. (2023) and taxonomies for information extraction Xu et al. (2023). In recommender systems, several reviews discuss how LLMs and agent-style components are used for data generation and recommendation Li et al. (2023); Lin et al. (2023); Wang et al. (2023). In software engineering, surveys summarize the use of LLMs across design, development, and testing Fan et al. (2023); Wang et al. (2024); Zheng et al. (2023). Other domain surveys cover robotics Zeng et al. (2023), autonomous driving Cui et al. (2024); Yang et al. (2023), medicine He et al. (2023); Zhou et al. (2023); Wang et al. (2023), finance Li et al. (2023), and psychology He et al. (2023). While these surveys are valuable for understanding domain adaptation, they typically treat memory as domain-specific prompting or retrieval practice, rather than a general representation and management problem.
Surveys that target memory in LLMs and agent systems are more closely related to our work, but the current picture is still fragmented. Some surveys discuss operational aspects of memory Zhang et al. (2024), yet many narrow their scope to long-context modeling Huang et al. (2023), long-term memory Jiang et al. (2024); He et al. (2024), personalization Liu et al. (2025), or knowledge editing Wang et al. (2024a). This topical split makes it hard to compare methods across settings, and it often blurs the boundary between (i) what is stored (the memory representation) and (ii) how it is used and maintained (the memory management). As a result, practical foundations such as consistent benchmarks, tools, and implementation constraints are not discussed in a unified way.
Several recent surveys propose alternative lenses for understanding memory. Some move beyond a pure time-based split (short-term vs. long-term) and categorize memory by the memory “object”, such as personal memories for user interaction and system memories for internal state Zhang et al. (2024); Zhong et al. (2024); Jiang et al. (2024). Others focus on memory mechanisms inside LLM-based agents and review their design, evaluation, and applications for self-evolving behaviors Zhang et al. (2025). Another direction decomposes memory into smaller operations and separates parametric and contextual forms, listing operations such as updating, indexing, retrieval, and compression Du et al. (2025). Human-memory-inspired surveys further relate human memory categories to AI memory designs and propose multi-dimensional categorizations Wu et al. (2025). Empirical studies evaluate how memory structures and retrieval strategies affect agent performance, including how memory addition and deletion influence long-horizon behaviors Zeng et al. (2024). These perspectives are informative, but they often treat the operation list as the primary organizing principle, which does not directly expose the system constraints behind different memory backends.
A complementary line of surveys studies memory from the deployment and efficiency angle. Inference system surveys summarize how to deliver high throughput and quality under large work-loads Pan and Li (2025), and KV-cache management is reviewed as a key technique for reducing redundant computation and improving memory use during decoding LI et al. (2025); Luohe et al.; Hatalis et al. (2024). Broader inference optimization surveys also analyze sources of inefficiency and summarize techniques at the data, model, and system levels Zhou et al. (2024). These works are mainly organized around performance techniques. They provide less discussion on how runtime memory (e.g., KV cache) relates to other memory forms under a single representation-and-management view, and how different backends can be composed in one agent system. A related survey Shan et al. (2025) uses a taxonomy close to ours, but it does not provide a systematic discussion of implementation details across memory backends.
In contrast, our survey treats memory as a separable system component and organizes prior work by two orthogonal axes: the representation used to store memory and the management process that constructs, updates, and queries that memory. This decouples memory mechanisms from specific learning modes such as in-context learning or weight updates, and clarifies how the same management goals can be realized through different backends (token memory, intermediate latent memory, and parametric memory) under different cost and reliability constraints.

Appendix B. Overview of Human and LLM Memory and Taxonomy

This section provides a high-level overview of human and LLM memory through a concise taxonomy of both. By decoupling human memory from LLM memory, we analyze how different categories of human memory can be instantiated using distinct LLM memory mechanisms. Building on this perspective, we present a holistic framework that demystifies LLM memory design and offers a unified intuition for understanding approaches.

Appendix B.1. Human Memory

Figure A2. Human Memory Overview.
Figure A2. Human Memory Overview.
Preprints 201223 g0a1
Human memory is a complex and multifaceted phenomenon, recognized in cognitive neuro-science as a collection of interconnected processes, including encoding, consolidation, storage, and retrieval Baddeley and Hitch (1974); Sridhar et al. (2023). It represents the brain’s remarkable capacity to store, retain, and recall information, serving as the foundation for learning, adapting to environments, and shaping personal identity Sherwood et al. (2004); Weng (2023). Memory underpins higher-order cognitive functions such as reasoning, problem-solving, and language comprehension, profoundly influencing behavior and decision-making Budson and Kensinger (2023); Shan et al. (2025).
Based on the duration of information retention, human memory is classified into short-term and long-term memory, as illustrated in Figure A2(a) Baddeley (2007). Short-term memory temporarily holds and processes information for seconds to minutes as working memory, which actively manipulates information for immediate tasks like reasoning and comprehension Budson and Kensinger (2023); Baddeley and Hitch (1974). In contrast, long-term memory stores information for extended periods, ranging from minutes to years, forming a repository for enduring knowledge and experiences Budson and Kensinger (2023).
Based on functional roles, human memory is commonly categorized into explicit (declarative), implicit (non-declarative), and sensory memory *, as shown in Figure A2(b) Budson and Kensinger (2023). Explicit memory involves conscious recall of facts and events that can be readily articulated. It includes episodic memory, which captures personal experiences tied to specific times and contexts (e.g., recalling what one ate for lunch) Tulving and Donaldson (1972), and semantic memory, which stores factual knowledge independent of personal experience (e.g., knowing the Earth is round) Begg (1984). Implicit memory operates unconsciously and is harder to verbalize, including procedural memory that governs skills and habits acquired through repetition, such as riding a bicycle or playing a musical instrument Squire (2009). These memory systems interact to support information processing and storage.
From a cognitive psychology perspective, memory is a fundamental mental process critical to learning and behavior Solso and Kagan (1979). It enables the accumulation of knowledge, abstraction of high-level concepts, and formation of social norms through the retention of cultural values and personal experiences Craik and Lockhart (1972); Leydesdorff (2017). Memory also supports decision-making by allowing individuals to anticipate potential consequences Johnson-Laird (1983). These insights are invaluable for designing LLM-based agents, as memory modules that mirror human cognitive processes enhance their ability to perform complex tasks and exhibit human-like behavior Laird (2019); Sun (2001).
Memory is essential for the self-evolution of LLM-based agents in dynamic environments Sutton and Barto (2018). It facilitates experience accumulation, enabling agents to retain past errors, inappropriate behaviors, or failed attempts to improve future performance and learning efficiency Zheng et al. (2023). Memory also supports environment exploration by guiding agents to prioritize less-explored actions or revisit previously unsuccessful trials, enhancing adaptability Montazeralghaem et al. (2020); Zhu et al. (2023). Additionally, memory enables knowledge abstraction, allowing agents to summarize raw observations into high-level insights, which is crucial for generalizing to new environments Zhao et al. (2023).
Table A2. Characteristics of Human Memory.
Table A2. Characteristics of Human Memory.
Memory Type Key Function/Characteristics Duration/Capacity
Sensory Memory Brief buffer for incoming sensory information (visual, auditory, etc.) Milliseconds to a few seconds
Working Memory (WM) Transient active store for manipulating information;
supports complex cognitive operations (reasoning, language)
Tens of seconds to minutes; limited items
Short-Term Memory (STM) Temporary holding of information before transfer to LTM or forgetting Tens of seconds to minutes; limited items
Long-Term Memory (LTM) Stores information for extended periods; large capacity and durability Minutes to decades; Vast capacity
Declarative (Explicit) Consciously recalled facts and events Minutes to decades; Vast capacity
Episodic Memory Personal experiences, specific events with contextual details Minutes to decades
Semantic Memory General world knowledge, facts, concepts, language Minutes to decades
Non-Declarative (Implicit) Unconscious learning: skills, habits, priming, conditioning Acquired slowly, long-lasting

Appendix B.2. LLM Memory

Figure A3. LLM Memory Classification.
Figure A3. LLM Memory Classification.
Preprints 201223 g0a2
LLM Inference. LLMs such as GPT Brown et al. (2020), LLaMA Touvron et al. (2023a,2); Grattafiori et al. (2024), Qwen Qwen et al. (2025); Bai et al. (2023), and DeepSeek DeepSeek-AI et al. (2025a,2) operate under the autoregressive generation paradigm. In this approach, the model predicts the next token based on all previously generated tokens. Given an input sequence of tokens ( x 1 , , x n ) , the model computes a probability distribution over the vocabulary for the next token at each time step t, typically using the final token’s representation. This process is governed by the joint probability:
P ( x 1 , , x n ) = P ( x 1 ) × P ( x 2 x 1 ) × × P ( x n x 1 , , x n 1 ) .
The inference process (Equation A1) relies on the self-attention mechanism within the transformer architecture. For each token i, self-attention computes a weighted sum over the representations of all previous tokens { 1 , 2 , , i } , resulting in a time complexity of O ( n 2 ) for a sequence of length n. This quadratic complexity becomes computationally expensive for long sequences, as attention is recalculated over all preceding tokens at every step.
To mitigate this inefficiency, the KV cache is widely used as an optimization technique. The KV cache divides inference into two phases: prefill and decoding. In the prefill phase, the model processes the entire prompt with full-sequence attention, computing and storing the key and value vectors for all tokens. During the decoding phase, as the model generates one token at a time, only the key and value vectors for the new token are computed and appended to the cache. Attention is then calculated solely between the current query and the cached KV pairs, eliminating redundant computations and significantly improving the efficiency of autoregressive generation.
Transformer Architecture. The remarkable performance of transformer-based LLMs across diverse tasks is primarily driven by the self-attention mechanism Leviathan et al. (2025); Meng et al. (2025); Vaswani et al. (2017). In this mechanism, a sequence of input hidden states ( h 1 , , h n ) is transformed through a linear projection layer to produce the Query (Q), Key (K), and Value (V) vectors, as defined in Equation A2.
concat ( q i , k i , v i ) = concat ( W q , W k , W v ) · h i
Subsequently, attention scores a i j are computed by taking the dot product between a Query vector ( q i ) and a Key vector ( k j ), scaled by the square root of the dimension d. These scores are normalized and used as weights to sum the Value vectors ( v j ), producing the output ( o i ) as shown in:
a i j = exp ( q i k j / d ) t = 1 i exp ( q i k t / d ) , o i = j = 1 i a i j v j .
The output of the self-attention mechanism is then processed by a Feed Forward Network (FFN), which typically consists of multiple linear layers interleaved with activation functions, further enhancing the model’s ability to capture complex patterns.
In Context Learning. In-context learning enables LLMs to retrieve and utilize memory by incorporating relevant information directly within the input prompt as natural language, without modifying model parameters or storing intermediate representations explicitly. The model relies on the provided context—such as examples, instructions, or retrieved data from a pool or database—to guide its predictions. For a prompt with a sequence of tokens ( x 1 , , x n ) , the self-attention mechanism, as defined in Equation (A3), weighs the relevance of each token in the context when predicting the next token, effectively mimicking memory retrieval. Research has shown that in-context learning can be viewed as implicit structure induction or meta-optimization, where the model learns task-specific patterns during inference by performing a form of gradient descent within the attention mechanism Dai et al. (2023); Hahn and Goyal (2023); Garg et al. (2022).
In practice, in-context learning is implemented by constructing prompts with relevant examples or retrieved documents, as seen in models like GPT Brown et al. (2020) or Qwen Qwen et al. (2025); Bai et al. (2023). For instance, few-shot learning scenarios provide question-answer pairs to guide responses, with the self-attention mechanism computing scores a i j = exp ( q i k j / d ) t = 1 i exp ( q i k t / d ) to focus on relevant context tokens. Studies suggest that in-context learning excels in task recognition and adaptation, but its effectiveness depends on the quality and structure of the provided demonstrations Min et al. (2022); Pan et al. (2023). Additionally, emergent in-context learning capabilities may be transient and tied to pretraining task diversity, highlighting the role of training data in enabling this mechanism Singh et al. (2023); Raventos et al. (2023); Shen et al. (2024).
KV Cache or Intermediate Representations. The KV cache stores intermediate Key (K) and Value (V) vectors computed during the self-attention process, as defined in Equation (A2), to enhance efficiency in autoregressive generation. During the prefill phase, the model processes the entire prompt, generating and caching K and V vectors for all tokens. In the decoding phase, only the K and V vectors for each newly generated token are computed and appended to the cache, with attention calculated solely between the current Query and cached KV pairs, producing the output o i = j = 1 i a i j v j . This reduces the time complexity from O ( n 2 ) to O ( n ) per token during decoding, as used in models like LLaMA Touvron et al. (2023a,2); Grattafiori et al. (2024) and DeepSeek DeepSeek-AI et al. (2025a,2).
Recent advancements have focused on optimizing the KV cache to handle long-context scenarios and reduce memory overhead. Techniques like CacheGen and ChunkKV employ semantic-preserving compression to reduce the cache size while maintaining performance Liu et al. (2024,2); Zhu et al. (2025); Liu et al. (2025). KIVI introduces asymmetric 2-bit quantization for KV cache entries, further reducing memory footprint Liu et al. (2024). Methods like SnapKV and H2O selectively retain important KV pairs based on attention patterns, improving efficiency for long-context inference Li et al. (2024); Zhang et al. (2023). Additionally, StreamingLLM uses attention sinks to stabilize attention distributions, enabling efficient streaming inference Xiao et al. (2023). Distributed approaches, such as KVDirect and FlowKV, optimize KV cache storage and transfer across multi-GPU systems Chen et al. (2024); Li et al. (2025). These optimizations make the KV cache a critical memory mechanism for scalable and efficient LLM inference.
Training-Based Knowledge Integration. Training-based memorization embeds knowledge directly into the model’s parameters during training, effectively transforming data into a compressed, implicit memory. LLMs like GPT, LLaMA, Qwen, and DeepSeek are trained on vast datasets, encoding patterns, facts, and relationships within weights, particularly in the linear projection layers ( W q , W k , W v ) and the Feed Forward Network (FFN). The training process optimizes the joint probability P ( x 1 , , x n ) = P ( x 1 ) · P ( x 2 x 1 ) · · P ( x n x 1 , , x n 1 ) , adjusting parameters via backpropagation to capture statistical and semantic patterns. This enables the model to recall knowledge during inference without external storage, though the memory is static unless fine-tuned or retrained.
Summary. As shown in Figure A3, In-context learning provides flexible memory through natural language prompts, leveraging self-attention to adaptively retrieve task-specific information, with its efficacy tied to demonstration quality and pretraining diversity Min et al. (2022); Raventos et al. (2023). Examples of implementing memory with different kinds of natural language tokens are shown in Table A3. The KV cache optimizes inference by storing intermediate representations, with recent advancements like compression and selective retention enhancing efficiency for long contexts Liu et al. (2024); Xiao et al. (2023); Li et al. (2024). Training-based knowledge integration embeds static memory in model parameters, enabling generalization across tasks. Together, these mechanisms enable LLMs to balance flexibility, efficiency, and generalization in diverse applications.
Table A3. Examples of using different text memory (or can be KV cache).
Table A3. Examples of using different text memory (or can be KV cache).
Question: Can you remind me about the trip I mentioned planning to Paris?
Memory (Vector Database): Embedding match - Conversation: "User plans a trip to Paris in September 2025, interested in visiting the Louvre and Eiffel Tower." Vector Database: Enables semantic retrieval by matching the query’s meaning to stored embeddings, useful for broad or vague queries.
Memory (Time Index): On July 15, 2025, at 14:30, the user said, "I’m planning a trip to Paris next month and want to see the Louvre." Time Index: Organizes memories chronologically, ideal for queries referencing recent or specific dates.
Memory (Username Index): User "JaneDoe123" discussed a Paris trip, mentioning a preference for art museums. Username Index: Ensures personalization by linking memories to a specific user, enhancing relevance.
Memory (Event Name Index): Event "Paris Trip 2025": User plans to visit Paris, focusing on cultural landmarks. Event Name Index: Tags memories with specific events, allowing precise retrieval for event-related queries.
Memory (Story Index): Story "Jane’s European Adventure": Includes a chapter on planning a Paris trip, with details about booking a hotel near the Seine. Story Index: Structures memories as narratives, preserving context across related interactions.
Memory (Place Index): Place "Paris, France": User mentioned visiting the Eiffel Tower and dining at a café in Montmartre. Place Index: Associates memories with locations, enabling spatial queries about specific places.

Appendix B.3. Taxonomy of Memory Implementations

Table A4. Reconstructed Memory Implementation Characteristics.
Table A4. Reconstructed Memory Implementation Characteristics.
Implementation Ways Memory Type Forgetting Pretrained Knowledge Memory Scalability Explainability Serving Costs
In-context Learning Short Term No High High Low
Long Term No Weak High Low
Parameter by Training Short Term Weak Hight Low Medium
Long Term Severe Hight Low High
Table A5. Reconstructed Characteristics of Memory Types by Implementation.
Table A5. Reconstructed Characteristics of Memory Types by Implementation.
Implementation Ways Memory Type Forgetting Pretrained Knowledge Explainability Costs Knowledge Match Degree
In-context Learning Procedural No High Low Low
Episodic No High Low High
Semantic No High Low High
Parameter by Training Procedural Possible Low High High
Episodic Possible Low High Moderate
Semantic Possible Low High Moderate
Different from existing surveys that connect different LLM memory types with human memory with one-to-one mapping, we observe that all different LLM memory mechanisms can be used to implement human-like memory with LLM. However, different LLM memory mechanisms have different characteristics, advantages and limitations, which are summarized in Table A4 and Table A5.
In-Context Learning. In-context learning supports both short-term memory (processing immediate context) and long-term memory. It relies on the self-attention mechanism to focus on relevant tokens, with no memory scale limitation for short-term use but high inference costs for long-term memory due to processing large contexts. In-context learning excels in episodic (context-specific events) and semantic (factual knowledge) memory, as prompts can encode task-specific examples or facts. Procedural memory (task processes) is limited due to reliance on explicit instructions.
Since in-context learning does not modify parameters, it avoids catastrophic forgetting, preserving pretrained knowledge. Short-term memory has no limitation, as it depends on the prompt size. Long-term memory faces high inference costs due to the O ( n 2 ) complexity of self-attention for long sequences, and the lost-in-the-middle problem. The memory is explicit in the prompt, making it interpretable as the model’s output directly reflects the provided context. No additional training or storage is required, though inference costs increase with context length. Malicious or biased prompts can influence outputs, raising concerns about misuse. High for episodic and semantic memory, as prompts can encode specific events or facts. Limited for procedural memory, as it requires explicit task instructions, which may not generalize well.
KV Cache or Intermediate Representations. The KV cache primarily functions as short-term memory, storing K and V vectors during the prefill and decoding phases of inference. It retains contextual information for the current sequence, enabling efficient token generation. Limited applicability, as the cache is typically cleared between sessions.
The KV cache is a runtime mechanism that does not alter model parameters, preserving pretrained knowledge. Limited by memory constraints, as storing K and V vectors for long sequences can be memory-intensive. While the KV cache stores intermediate representations, interpreting their content is less straightforward than in-context learning, as it involves analyzing attention weights and vectors. The KV cache requires additional memory to store vectors, but optimizations like compression reduce this cost. Inference efficiency is improved compared to recomputing attention. Low. The KV cache is a technical optimization with minimal impact on output content, though improper cache management could affect output coherence. High for episodic memory, as the cache retains sequence-specific context. Limited for semantic or procedural memory, as it does not store generalized knowledge or task processes independently of the input sequence.
Parameter by Training. Training embeds both short-term (immediate patterns) and long-term (generalized knowledge) memory into parameters. Short-term memory has weak forgetting, as recent patterns are retained, but long-term memory suffers from severe forgetting due to catastrophic forgetting during fine-tuning. Training excels in procedural memory, as models learn task-specific patterns during pretraining. Episodic and semantic memory are moderately supported, as specific events or facts are compressed into parameters but may be less precise. Fine-tuning can overwrite pretrained knowledge, especially for long-term memory, leading to catastrophic forgetting. The model’s parameters can encode vast amounts of knowledge, limited only by model size and training data. Knowledge embedded in parameters is opaque, making it difficult to trace specific outputs to learned patterns. Training requires significant computational resources, especially for large models and datasets. None. Knowledge is fixed in parameters, reducing risks from external inputs, though biases in training data can persist. High for procedural memory, as training optimizes task-specific patterns. Moderate for episodic and semantic memory, as specific events or facts are generalized but may lose granularity.

Appendix B.4. How Human Memory Benefits LLM Agentic Applications

Memory is an indispensable component in various practical LLM-based agent applications. For instance, in a conversational agent, memory stores information about historical conversations, providing the necessary context for generating coherent and relevant responses; without it, the agent cannot maintain a continuous conversation Lu et al. (2023). Similarly, in a simulation agent, memory is crucial for maintaining consistent role profiles, preventing the agent from deviating from its assigned character during a simulation Wang et al. (2025,2). These examples underscore that memory is not an optional feature but a necessary component for LLM-based agents to effectively accomplish their given tasks. Thus, the cognitive basis of human memory, coupled with its necessity for agent self-evolution and practical applications, provides critical insights for designing sophisticated memory mechanisms in LLM-based systems.
Information Retrieval and Processing. Long-context LLMs like Longformer and LongT5 enhance response relevance and document summarization by processing larger text segments, reducing reliance on external RAG tools Jin et al. (2024); Shi et al. (2024); Beltagy et al. (2020); Guo et al. (2022); Jin et al. (2024). Advanced semantic vector models, such as text-embedding-3-large, jina-embeddings-v2, and BGE-M3, overcome window size limitations, improving usability in tasks like translating complex documents and entire novels Zhu et al. (2023); Wang et al. (2024); OpenAI (2024); Günther et al. (2023); Chen et al. (2024); Zhu et al. (2024); Saad-Falcon et al. (2024); Herold and Ney (2023); Wang et al. (2024); Lyu et al. (2024).
Chatbots. Long-context processing enhances chatbots by enabling extended memory and contextual coherence, as seen in platforms like ChatGPT, Pi, Character AI, and Talkie, which use persistent memory and techniques like prompt-based memorization, memory-augmented architectures, and context extension for style-consistent, engaging dialogues OpenAI (2024); Inflection (2023); Character AI (2023); Ai (2024); Lee et al. (2023); Zhong et al. (2024); Wang et al. (2023,2).
Code Development. LLMs leverage memory to store development knowledge and conversational context, with models like StarCoder2, Qwen2.5-Coder, and Granite Code Models enabling scalable code completion and predictive debugging in tools like GitHub Copilot and Anysphere Cursor Qian et al. (2023); Tsai et al. (2023); Chen et al. (2023); Li et al. (2023); Zhang et al. (2024); Lozhkov et al. (2024); Hui et al. (2024); Mishra et al. (2024); GitHub (2022); Anysphere (2025).
Social Simulation. Memory defines character traits for realistic role-playing and supports multi-agent social simulations by improving self-monitoring, maintaining economic environments, and simulating dynamic behaviors Gao et al. (2023); Wang et al. (2025); Li et al. (2023); Shao et al. (2023); Kaiya et al. (2023); Li et al. (2023); Hua et al. (2023).
Personal Assistant. LLM-based personal assistants rely on memory for consistent, personalized dialogues, using textual retrieval, conversation summarization, and external tools to maintain conversational flow Lu et al. (2023); Lee et al. (2023); Pan et al. (2023); Wu et al. (2023).
Application in Specific Domains. Long-context LLMs improve coherence in news summaries, simplify legal document interpretation, enhance healthcare and financial decision-making, and advance drug discovery and scientific problem-solving by leveraging external knowledge and memory Gao et al. (2019); Kapoor et al. (2024); Fan et al. (2024); Reddy et al. (2024); Masry and Hajian (2024); Nie et al. (2024); Hilgert et al. (2024); Shao and Yan (2024); Wang et al. (2023); Xiong et al. (2023); Liu et al. (2023); Wang et al. (2023); Yunxiang et al. (2023); Chen et al. (2024); Zhao et al. (2024); Chen et al. (2023); Wang et al. (2023); Qiang et al. (2023).

Appendix C. Taxonomy of Different Memory

Table A6. Taxonomy of RAG.
Table A6. Taxonomy of RAG.
Preprints 201223 i004
Table A7. Taxonomy of Agent Memory.
Table A7. Taxonomy of Agent Memory.
Preprints 201223 i005
Table A8. Taxonomy of KV Cache as Memory.
Table A8. Taxonomy of KV Cache as Memory.
Preprints 201223 i006
Table A9. Taxonomy of Other Vectors as Memory.
Table A9. Taxonomy of Other Vectors as Memory.
Preprints 201223 i007
Table A10. Taxonomy of Parameter as Memory.
Table A10. Taxonomy of Parameter as Memory.
Preprints 201223 i008
  1. Agarwal, S., Acun, B., Hosmer, B., Elhoushi, M., Lee, Y., Venkataraman, S., Papailiopoulos, D., & Wu, C.-J. (2024, 21-27 Jul). CHAI: Clustered Head Attention for Efficient LLM Inference. In R. Salakhutdinov et al. (Eds.), Proceedings of the 41st international conference on machine learning (Vol. 235, pp. 291-312). PMLR. Available online: https://proceedings.mlr.press/v235/agarwal24a.html (accessed on).
  2. Agarwal, S., Sundaresan, S., Mitra, S., Mahapatra, D., Gupta, A., Sharma, R., Kapu, N. J., Yu, T., & Saini, S. (2025). Cache-craft: Managing chunk-caches for efficient retrieval-augmented generation. Available online: https://arxiv.org/abs/2502.15734 (accessed on).
  3. Ai, T. (2024). Talkie | ai-native character community. Available online: https://www.talkie-ai.com/ (accessed on).
  4. Al Adel, A., & Burtsev, M. S. (2021). Memory transformer with hierarchical attention for long document processing. In 2021 international conference engineering and telecommunication (p. 1-7). [CrossRef]
  5. Allen-Zhu, Z., & Li, Y. (2024). Physics of language models: part 3.1, knowledge storage and extraction. In Proceedings of the 41st international conference on machine learning (pp. 1067-1077).
  6. An, Y., Cheng, Y., Park, S. J., & Jiang, J. (2025). Hyperrag: Enhancing quality-efficiency tradeoffs in retrieval-augmented generation with reranker kv-cache reuse. Available online: https://arxiv.org/abs/2504.02921 (accessed on).
  7. Anysphere. (2025). Cursor - the ai code editor. https://www.cursor.com/en.. Available online: https://www.cursor.com (accessed on).
  8. Arditi, A., Obeso, O., Syed, A., Paleka, D., Panickssery, N., Gurnee, W., & Nanda, N. (2024). Refusal in language models is mediated by a single direction. arXiv.
  9. Asai, A., Wu, Z., Wang, Y., Sil, A., & Hajishirzi, H. (2023). Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection. arXiv preprint arXiv:2310.11511.
  10. BAAI. (2023). Flagembedding. https://github.com/FlagOpen/FlagEmbedding.
  11. Baddeley, A. (2007). Working memory, thought, and action (Vol. 45). OuP Oxford.
  12. Baddeley, A. D., & Hitch, G. (1974). Working Memory. In G. H. Bower (Ed.), (Vol. 8, p. 47-89). Academic Press. Available online: https://www.sciencedirect.com/science/article/pii/S0079742108604521 (accessed on). https://doi.org/10.1016/S0079-7421(08)60452-1.
  13. Bae, S., Kwak, D., Kang, S., Lee, M. Y., Kim, S., Jeong, Y., Kim, H., Lee, S.-W., Park, W., & Sung, N. (2022, December). Keep Me Updated! Memory Management in Long-term Conversations. In Y. Goldberg, Z. Kozareva, & Y. Zhang (Eds.), Findings of the association for computational linguistics: Emnlp 2022 (pp. 3769-3787). Abu Dhabi, United Arab Emirates: Association for Computational Linguistics. Available online: https://aclanthology.org/2022.findings-emnlp.276/ (accessed on). https://doi.org/10.18653/v1/2022.findings-emnlp.276.
  14. Bai, J., Bai, S., Chu, Y., Cui, Z., Dang, K., Deng, X., et al. (2023). Qwen technical report. Available online: https://arxiv.org/abs/2309.16609 (accessed on).
  15. Begg, I. (1984). Tulving's memory [Review of the book Elements of episodic memory, by E. Tulving . Canadian Journal of Psychology / Revue canadienne de psychologie, 38(1), 144-147.
  16. Beltagy, I., Peters, M. E., & Cohan, A. (2020). Longformer: The Long-Document Transformer. CoRR, abs/2004.05150. Available online: https://arxiv.org/abs/2004.05150 (accessed on).
  17. Biderman, S., Prashanth, U. S., Sutawika, L., Schoelkopf, H., Anthony, Q., Purohit, S., & Raff, E. (2023). Emergent and predictable memorization in large language models. Available online: https://arxiv.org/abs/2304.11158 (accessed on).
  18. Bommasani, R., Hudson, D. A., Adeli, E., Altman, R., Arora, S., von Arx, S., Bernstein, M. S., Bohg, J., Bosselut, A., Brunskill, E., Brynjolfsson, E., Buch, S., Card, D., Castellon, R., Chatterji, N., Chen, A., Creel, K., Davis, J. Q., Demszky, D.,... Liang, P. (2022). On the opportunities and risks of foundation models. Available online: https://arxiv.org/abs/2108.07258 (accessed on).
  19. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D. M., Wu, J., Winter, C.,... Amodei, D. (2020a). Language models are few-shot learners. In Proceedings of the 34th international conference on neural information processing systems. Red Hook, NY, USA: Curran Associates Inc.
  20. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D. M., Wu, J., Winter, C.,... Amodei, D. (2020b). Language models are few-shot learners. Available online: https://arxiv.org/abs/2005.14165 (accessed on).
  21. Budson, A. E., & Kensinger, E. A. (2023). Why we forget and how to remember better: the science behind memory. Oxford University Press.
  22. Cao, Y., Zhang, T., Cao, B., Yin, Z., Lin, L., Ma, F., & Chen, J. (2024). Personalized steering of large language models: Versatile steering vectors through bi-directional preference optimization. arXiv.
  23. Carlini, N., Ippolito, D., Jagielski, M., Lee, K., Tramer, F., & Zhang, C. (2023). Quantifying memorization across neural language models. Available online: https://arxiv.org/abs/2202.07646 (accessed on).
  24. Carlini, N., Tramer, F., Wallace, E., Jagielski, M., Herbert-Voss, A., Lee, K., Roberts, A., Brown, T., Song, D., Erlingsson, U., Oprea, A., & Raffel, C. (2021). Extracting training data from large language models. Available online: https://arxiv.org/abs/2012.07805 (accessed on).
  25. Character AI. (2023). Character ai. Retrieved September 14, 2023 from https://character.ai/. Available online: https://character.ai/ (accessed on).
  26. Chen, D., Wang, H., Huo, Y., Li, Y., & Zhang, H. (2023). Gamegpt: Multi-agent collaborative framework for game development. arXiv preprint arXiv:2310.08067.
  27. Chen, J., Xiao, S., Zhang, P., Luo, K., Lian, D., & Liu, Z. (2024). BGE M3-Embedding: Multi-Lingual, Multi-Functionality, Multi-Granularity Text Embeddings Through Self-Knowledge Distillation. CoRR, abs/2402.03216. Available online: https://doi.org/10.48550/arXiv.2402.03216 (accessed on) https://doi.org/10.48550/ARXIV.2402.03216.
  28. Chen, K., Li, J., Wang, K., Du, Y., Yu, J., Lu, J., Li, L., Qiu, J., Pan, J., Huang, Y., Fang, Q., Heng, P. A., & Chen, G. (2024). Chemist-x: Large language model-empowered agent for reaction condition recommendation in chemical synthesis.
  29. Chen, S., Jiang, R., Yu, D., Xu, J., Chao, M., Meng, F., Jiang, C., Xu, W., & Liu, H. (2024). Kvdirect: Distributed disaggregated llm inference. Available online: https://arxiv.org/abs/2501.14743 (accessed on).
  30. Chen, T., Wang, H., Chen, S., Yu, W., Ma, K., Zhao, X., Yu, D., & Zhang, H. (2023). Dense X Retrieval: What Retrieval Granularity Should We Use? arXiv preprint arXiv:2312.06648.
  31. Chen, Y., Wang, G., Shang, J., Cui, S., Zhang, Z., Liu, T., Wang, S., Sun, Y., Yu, D., & Wu, H. (2024, August). NACL: A General and Effective KV Cache Eviction Framework for LLM at Inference Time. In L.-W. Ku, A. Martins, & V. Srikumar (Eds.), Proceedings of the 62nd annual meeting of the association for computational linguistics (volume 1: Long papers) (pp. 7913-7926). Bangkok, Thailand: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.acl-long.428/ (accessed on). https://doi.org/10.18653/v1/2024.acl-long.428.
  32. Chen, Z.-Y., Xie, F.-K., Wan, M., Yuan, Y., Liu, M., Wang, Z.-G., Meng, S., & Wang, Y.-G. (2023). Matchat: A large language model and application service platform for materials science. Chinese Physics B, 32(11), 118104.
  33. Cheng, X., Luo, D., Chen, X., Liu, L., Zhao, D., & Yan, R. (2023). Lift Yourself Up: Retrieval-augmented Text Generation with Self Memory. arXiv preprint arXiv:2305.02437.
  34. Cheng, Y., Zhang, C., Zhang, Z., Meng, X., Hong, S., Li, W., Wang, Z., Wang, Z., Yin, F., Zhao, J., & et al. (2024). Exploring large language model based intelligent agents: Definitions, methods, and prospects. arXiv.
  35. Chevalier, A., Wettig, A., Ajith, A., & Chen, D. (2023, December). Adapting Language Models to Compress Contexts. In H. Bouamor, J. Pino, & K. Bali (Eds.), Proceedings of the 2023 conference on empirical methods in natural language processing (pp. 3829-3846). Singapore: Association for Computational Linguistics. Available online: https://aclanthology.org/2023.emnlp-main.232/ (accessed on). https://doi.org/10.18653/v1/2023.emnlp-main.232.
  36. Chugtai, B., & Bushnaq, L. (2025). Activation space interpretability may be doomed. Available online: https://www.lesswrong.com/posts/gYfpPbww3wQRaxAFD (accessed on).
  37. Craik, F. I., & Lockhart, R. S. (1972). Levels of processing: A framework for memory research. Journal of verbal learning and verbal behavior, 11(6), 671-684.
  38. Cui, C., Ma, Y., Cao, X., Ye, W., Zhou, Y., Liang, K., Chen, J., Lu, J., Yang, Z., Liao, K.-D., & et al. (2024). A survey on multimodal large language models for autonomous driving. In Proceedings of the ieee/cof winter conference on applications of computer vision (pp. 958-979).
  39. Daheim, N., Möllenhoff, T., Ponti, E., Gurevych, I., & Khan, M. E. (2024). Model Merging by Uncertainty-Based Gradient Matching. In Iclr.
  40. Dai, D., Dong, L., Hao, Y., Sui, Z., Chang, B., & Wei, F. (2021). Knowledge neurons in pretrained transformers. arXiv.
  41. Dai, D., Sun, Y., Dong, L., Hao, Y., Ma, S., Sui, Z., & Wei, F. (2023). Why can GPT learn in-context? language models secretly perform gradient descent as meta-optimizers. In Findings of the association for computational linguistics: Acl 2023 (pp. 4005-4019). Toronto, Canada: Association for Computational Linguistics. Available online: https://doi.org/10.18653/v1/2023.findings-acl.247 (accessed on).
  42. Dao, T., Fu, D. Y., Ermon, S., Rudra, A., & Ré, C. (2022). FLASHATTENTION: fast and memory-efficient exact attention with IO-awareness. In Proceedings of the 36th international conference on neural information processing systems. Red Hook, NY, USA: Curran Associates Inc.
  43. Dathathri, S., Madotto, A., Lan, J., Hung, J., Frank, E., Molino, P., Yosinski, J., & Liu, R. (n.d.). Plug and Play Language Models: A Simple Approach to Controlled Text Generation. In International conference on learning representations.
  44. Davari, M., & Belilovsky, E. (2023). Model breadcrumbs: Scaling multi-task model merging with sparse masks. arXiv preprint arXiv:2312.06795.
  45. DeepSeek-AI, Guo, D., Yang, D., Zhang, H., Song, J., et al. (2025). Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning. Available online: https://arxiv.org/abs/2501.12948 (accessed on).
  46. DeepSeek-Al, Liu, A., Feng, B., Xue, B., Wang, B., Wu, B., et al. (2025). Deepseek-v3 technical report. Available online: https://arxiv.org/abs/2412.19437 (accessed on).
  47. Devoto, A., Zhao, Y., Scardapane, S., & Minervini, P. (2024, November). A Simple and Effective L_2 Norm- Based Strategy for KV Cache Compression. In Proceedings of the 2024 conference on empirical methods in natural language processing (pp. 18476-18499). Miami, Florida, USA: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.emnlp-main.1027/ (accessed on). https://doi.org/10.18653/v1/2024.emnlp-main.1027.
  48. Dhuliawala, S., Komeili, M., Xu, J., Raileanu, R., Li, X., Celikyilmaz, A., & Weston, J. (2023). Chain-of-verification reduces hallucination in large language models. arXiv preprint arXiv:2309.11495.
  49. Di, S., Yu, Z., Zhang, G., Li, H., TaoZhong, Cheng, H., Li, B., He, W., Shu, F., & Jiang, H. (2025). Streaming Video Question-Answering with In-context Video KV-Cache Retrieval. In The thirteenth international conference on learning representations. Available online: https://openreview.net/forum?id=8g9fs6mdEG (accessed on).
  50. Dong, H., Yang, X., Zhang, Z., Wang, Z., Chi, Y., & Chen, B. (2024, 21-27 Jul). Get More with LESS: Synthesizing Recurrence with KV Cache Compression for Efficient LLM Inference. In R. Salakhutdinov et al. (Eds.), Proceedings of the 41st international conference on machine learning (Vol. 235, pp. 11437-11452). PMLR. Available online: https://proceedings.mlr.press/v235/dong24f.html (accessed on).
  51. Dong, S., Cheng, W., Qin, J., & Wang, W. (2024). Qaq: Quality adaptive quantization for Ilm ku cache. Available online: https://arxiv.org/abs/2403.04643 (accessed on).
  52. Du, Y., Huang, W., Zheng, D., Wang, Z., Montella, S., Lapata, M., Wong, K.-F., & Pan, J. Z. (2025). Rethinking memory in ai: Taxonomy, operations, topics, and future directions. Available online: https://arxiv.org/abs/2505.00675 (accessed on).
  53. Du, Y., Wang, H., Zhao, Z., Liang, B., Wang, B., Zhong, W., Wang, Z., & Wong, K.-F. (2024). Perltqa: A personal long-term memory dataset for memory classification, retrieval, and synthesis in question answering. Available online: https://arxiv.org/abs/2402.16288 (accessed on).
  54. Duanmu, H., Yuan, Z., Li, X., Duan, J., Zhang, X., & Lin, D. (2024). SKVQ: Sliding-window key and value cache quantization for large language models. Available online: https://arxiv.org/abs/2405.06219 (accessed on).
  55. Dunefsky, J., & Cohan, A. (2025). Investigating generalization of one-shot LLM steering vectors. arXiv preprint arXiv:2502.18862.
  56. Durante, Z., Huang, Q., Wake, N., Gong, R., Park, J. S., Sarkar, B., Taori, R., Noda, Y., Terzopoulos, D., Choi, Y., & et al. (2024). Agent ai: Surveying the horizons of multimodal interaction. arXiv.
  57. Fan, A., Gokkaya, B., Harman, M., Lyubarskiy, M., Sengupta, S., Yoo, S., & Zhang, J. M. (2023). Large language models for software engineering: Survey and open problems. arXiv.
  58. Fan, Y., Sun, H., Xue, K., Zhang, X., Zhang, S., & Ruan, T. (2024). MedOdyssey: A Medical Domain Benchmark for Long Context Evaluation Up to 200K Tokens. CoRR, abs/2406.15019. Available online: https://doi.org/10.48550/arXiv.2406.15019 (accessed on) https://doi.org/10.48550/ARXIV.2406.15019.
  59. Feng, Y., Chu, X., Xu, Y., Shi, G., Liu, B., & Wu, X.-M. (2024, August). TaSL: Continual Dialog State Tracking via Task Skill Localization and Consolidation. In L.-W. Ku, A. Martins, & V. Srikumar (Eds.), Proceedings of the 62nd annual meeting of the association for computational linguistics (volume 1: Long papers) (pp. 1266-1279). Bangkok, Thailand: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.acl-long.69/ (accessed on). https://doi.org/10.18653/v1/2024.acl-long.69.
  60. Freeman, J., Rippe, C., Debenedetti, E., & Andriushchenko, M. (2024). Exploring memorization and copyright violation in frontier LLMs: A study of the new york times v. openai 2023 lawsuit. Available online: https://arxiv.org/abs/2412.06370 (accessed on).
  61. Fu, Y., Xue, L., Huang, Y., Brabete, A.-O., Ustiugov, D., Patel, Y., & Mai, L. (2024). ServerlessLLM: low-latency serverless inference for large language models. In Proceedings of the 18th usenix conference on operating systems design and implementation. USA: USENIX Association.
  62. Gao, C., Lan, X., Lu, Z., Mao, J., Piao, J., Wang, H., Jin, D., & Li, Y. (2023). S3: Social-network simulation system with large language model-empowered agents. arxiv.
  63. Gao, L., Ma, X., Lin, J., & Callan, J. (2022). Precise zero-shot dense retrieval without relevance labels. arXiv preprint arXiv:2212.10496.
  64. Gao, M., Lu, T., Yu, K., Byerly, A., & Khashabi, D. (2024, November). Insights into LLM Long-Context Failures: When Transformers Know but Don't Tell. In Y. Al-Onaizan, M. Bansal, & Y.-N. Chen (Eds.), Findings of the association for computational linguistics: Emnlp 2024. Miami, Florida, USA: Association for Computational Linguistics.
  65. Gao, S., Chen, X., Li, P., Ren, Z., Bing, L., Zhao, D., & Yan, R. (2019). Abstractive Text Summarization by Incorporating Reader Comments. In The thirty-third AAAI conference on artificial intelligence, AAAI 2019, the thirty-first innovative applications of artificial intelligence conference, IAAI 2019, the ninth AAAI symposium on educational advances in artificial intelligence, EAAI 2019, honolulu, hawaii, usa, january 27-february 1, 2019 (pp. 6399-6406). AAAI Press. Available online: https://doi.org/10.1609/aaai.v33101.33016399 (accessed on). https://doi.org/10.1609/AAALV33101.33016399.
  66. Garg, S., Tsipras, D., Liang, P. S., & Valiant, G. (2022). What can transformers learn in-context? A case study of simple function classes. In Advances in neural information processing systems (Vol. 35, pp. 30583-30598).
  67. Ge, S., Zhang, Y., Liu, L., Zhang, M., Han, J., & Gao, J. (2024). Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs. In The twelfth international conference on learning representations. Available online: https://openreview.net/forum?id=uNrFpDPMyo (accessed on).
  68. Ge, Y., Ren, Y., Hua, W., Xu, S., Tan, J., & Zhang, Y. (2023). Llm as os (llmao), agents as apps: Envisioning aios, agents and the aios-agent ecosystem. arXiv.
  69. Geva, M., Schuster, R., Berant, J., & Levy, O. (2021). Transformer feed-forward layers are key-value memories. arXiv.
  70. Gim, I., Chen, G., Lee, S.-s., Sarda, N., Khandelwal, A., & Zhong, L. (2024). Prompt cache: Modular attention reuse for low-latency inference. Proceedings of Machine Learning and Systems, 6, 325-338.
  71. GitHub. (2022). Github copilot. Available online: https://github.com/copilot (accessed on).
  72. Grattafiori, A., Dubey, A., Jauhri, A., Pandey, A., Kadian, A., et al. (2024). The llama 3 herd of models. Available online: https://arxiv.org/abs/2407.21783 (accessed on).
  73. Günther, M., Ong, J., Mohr, I., Abdessalem, A., Abel, T., Akram, M. K., Guzman, S., Mastrapas, G., Sturua, S., Wang, B., Werk, M., Wang, N., & Xiao, H. (2023). Jina Embeddings 2: 8192-Token General-Purpose Text Embeddings for Long Documents. CoRR, abs/2310.19923. Available online: https://doi.org/10.48550/arXiv.2310.19923 (accessed on). https://doi.org/10.48550/ARXIV.2310.19923.
  74. Guo, M., Ainslie, J., Uthus, D. C., Ontañón, S., Ni, J., Sung, Y., & Yang, Y. (2022). LongT5: Efficient Text-To- Text Transformer for Long Sequences. In M. Carpuat, M. de Marneffe, & I. V. M. Ruíz (Eds.), Findings of the association for computational linguistics: NAACL 2022, seattle, wa, united states, july 10-15, 2022 (pp. 724-736). Association for Computational Linguistics. Available online: https://doi.org/10.18653/v1/2022.findings-naacl.55 (accessed on). https://doi.org/10.18653/V1/2022.FINDINGS-NAACL.55.
  75. Guo, T., Chen, X., Wang, Y., Chang, R., Pei, S., Chawla, N. V., Wiest, O., & Zhang, X. (2024). Large language model based multi-agents: A survey of progress and challenges. arXiv.
  76. Gutiérrez, B. J., Shu, Y., Gu, Y., Yasunaga, M., & Su, Y. (2024). Hipporag: Neurobiologically inspired long-term memory for large language models. In The thirty-eighth annual conference on neural information processing systems.
  77. Hahn, M., & Goyal, N. (2023). A theory of emergent in-context learning as implicit structure induction. arxiv, arXiv:2303.07971. Available online: https://arxiv.org/abs/2303.07971 (accessed on).
  78. Han, C., Wang, Q., Peng, H., Xiong, W., Chen, Y., Ji, H., & Wang, S. (2024, June). LM-Infinite: Zero-Shot Extreme Length Generalization for Large Language Models. In K. Duh, H. Gomez, & S. Bethard (Eds.), Proceedings of the 2024 conference of the north american chapter of the association for computational linguistics: Human language technologies (volume 1: Long papers) (pp. 3991-4008). Mexico City, Mexico: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.naacl-long.222/ (accessed on). https://doi.org/10.18653/v1/2024.naacl-long.222.
  79. Han, Z., Gao, C., Liu, J., Zhang, J., & Zhang, S. Q. (2024). Parameter-efficient fine-tuning for large models: A comprehensive survey. arXiv preprint arXiv:2403.14608.
  80. Hao, Y., Zhai, M., Hajimirsadeghi, H., Hosseini, S., & Tung, F. (2025). Radar: Fast Long-Context Decoding for Any Transformer. In The thirteenth international conference on learning representations. Available online: https://openreview.net/forum?id=ZTpWOwMrzQ (accessed on).
  81. Hatalis, K., Christou, D., Myers, J., Jones, S., Lambert, K., Amos-Binks, A., Dannenhauer, Z., & Dannenhauer, D. (2024). Memory Matters: The Need to Improve Long-Term Memory in LLM-Agents. Proceedings of the AAAI Symposium Series, 2.
  82. He, K., Mao, R., Lin, Q., Ruan, Y., Lan, X., Feng, M., & Cambria, E. (2023). A survey of large language models for healthcare: from data, technology, and applications to accountability and ethics. arXiv.
  83. He, T., Fu, G., Yu, Y., Wang, F., Li, J., Zhao, Q., Song, C., Qi, H., Luo, D., Zou, H., & et al. (2023). Towards a psychological generalist ai: A survey of current applications of large language models and future prospects. arXiv.
  84. He, X., Tian, Y., Sun, Y., Chawla, N. V., Laurent, T., LeCun, Y., Bresson, X., & Hooi, B. (2024). G-Retriever: Retrieval-Augmented Generation for Textual Graph Understanding and Question Answering. arXiv preprint arXiv:2402.07630.
  85. He, Y., Zhang, L., Wu, W., Liu, J., Zhou, H., & Zhuang, B. (2024). ZipCache: Accurate and Efficient KV Cache Quan- tization with Salient Token Identification. In A. Globerson et al. (Eds.), Advances in neural information processing systems (Vol. 37, pp. 68287-68307). Curran Associates, Inc. Available online: https://proceedings.neurips.cc/paper_files/paper/2024/file/7e57131fdeb815764434b65162c88895-Paper-Conference.pdf (accessed on).
  86. He, Z., Karlinsky, L., Kim, D., McAuley, J., Krotov, D., & Feris, R. (2024). Camelot: Towards large language models with training-free consolidated associative memory. arXiv preprint arXiv:2402.13449.
  87. He, Z., Lin, W., Zheng, H., Zhang, F., Jones, M. W., Aitchison, L., Xu, X., Liu, M., Kristensson, P. O., & Shen, J. (2024). Human-inspired Perspectives: A Survey on Al Long-term Memory. arXiv preprint arXiv:2411.00489. Available online: https://arxiv.org/abs/2411.00489 (accessed on).
  88. Hernandez, E., Li, B. Z., & Andreas, J. (2023). Inspecting and editing knowledge representations in language models. arXiv.
  89. Herold, C., & Ney, H. (2023). Improving Long Context Document-Level Machine Translation. CoRR, abs/2306.05183. Available online: https://doi.org/10.48550/arXiv.2306.05183 (accessed on) https://doi.org/10.48550/ARXIV.2306.05183.
  90. Hilgert, L., Liu, D., & Niehues, J. (2024, November). Evaluating and Training Long-Context Large Language Models for Question Answering on Scientific Papers. In S. Kumar et al. (Eds.), Proceedings of the 1st workshop on customizable nlp: Progress and challenges in customizing nlp for a domain, application, group, or individual (customnlp4u) (pp. 220-236). Miami, Florida, USA: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.customnlp4u-1.17/ (accessed on). https://doi.org/10.18653/v1/2024.customnlp4u-1.17.
  91. Hooper, C., Kim, S., Mohammadzadeh, H., Mahoney, M. W., Shao, S., Keutzer, K., & Gholami, A. (2024). KVQuant: Towards 10 million context length LLM inference with KV cache quantization. Advances in Neural Information Processing Systems, NeurIPS 2024, 37, 1270-1303.
  92. Hu, E. J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., & Chen, W. (2021). Lora: Low-rank adaptation of large language models. Available online: https://arxiv.org/abs/2106.09685 (accessed on).
  93. Hu, J., Huang, W., Wang, W., Wang, H., Hu, T., Zhang, Q., Feng, H., Chen, X., Shan, Y., & Xie, T. (2025). Epic: Efficient position-independent caching for serving large language models. Available online: https://arxiv.org/abs/2410.15332 (accessed on).
  94. Hua, W., Fan, L., Li, L., Mei, K., Ji, J., Ge, Y., Hemphill, L., & Zhang, Y. (2023). War and peace (waragent): Large language model-based multi-agent simulation of world wars. arXiv preprint arXiv:2311.17227.
  95. Huang, C., Liu, Q., Lin, B. Y., Pang, T., Du, C., & Lin, M. (2023, 07). LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition. Available online: https://arxiv.org/pdf/2307.13269.pdf (accessed on).
  96. Hooper, C., Kim, S., Mohammadzadeh, H., Mahoney, M. W., Shao, S., Keutzer, K., & Gholami, A. (2024). KVQuant: Towards 10 million context length LLM inference with KV cache quantization. Advances in Neural Information Processing Systems, NeurIPS 2024, 37, 1270-1303.
  97. Hu, E. J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., & Chen, W. (2021). Lora: Low-rank adaptation of large language models. Available online: https://arxiv.org/abs/2106.09685 (accessed on).
  98. Hu, J., Huang, W., Wang, W., Wang, H., Hu, T., Zhang, Q., Feng, H., Chen, X., Shan, Y., & Xie, T. (2025). Epic: Efficient position-independent caching for serving large language models. Available online: https://arxiv.org/abs/2410.15332 (accessed on).
  99. Hua, W., Fan, L., Li, L., Mei, K., Ji, J., Ge, Y., Hemphill, L., & Zhang, Y. (2023). War and peace (waragent): Large language model-based multi-agent simulation of world wars. arXiv preprint arXiv:2311.17227.
  100. Huang, C., Liu, Q., Lin, B. Y., Pang, T., Du, C., & Lin, M. (2023, 07). LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition. Available online: https://arxiv.org/pdf/2307.13269.pdf (accessed on).
  101. Huang, X., Liu, W., Chen, X., Wang, X., Wang, H., Lian, D., Wang, Y., Tang, R., & Chen, E. (2024). Understanding the planning of llm agents: A survey. arXiv.
  102. Huang, Y., Xu, J., Lai, J., Jiang, Z., Chen, T., Li, Z., Yao, Y., Ma, X., Yang, L., Chen, H., et al. (2023). Advancing transformer architecture in long-context large language models: A comprehensive survey. arXiv preprint arXiv:2311.12351.
  103. Hui, B., Yang, J., Cui, Z., Yang, J., Liu, D., Zhang, L., Liu, T., Zhang, J., Yu, B., Dang, K., Yang, A., Men, R., Huang, F., Ren, X., Ren, X., Zhou, J., & Lin, J. (2024). Qwen2.5-Coder Technical Report. CoRR, abs/2409.12186. Available online: https://doi.org/10.48550/arXiv.2409.12186 (accessed on).
  104. Ilharco, G., Ribeiro, M. T., Wortsman, M., Gururangan, S., Schmidt, L., Hajishirzi, H., & Farhadi, A. (2022, 12). Editing Models with Task Arithmetic. Available online: https://arxiv.org/pdf/2212.04089.pdf (accessed on).
  105. Inflection. (2023). I’m pi, your personal ai. https://inflection.ai/. Available online: https://inflection.ai/ (accessed on).
  106. Ippolito, D., Tramer, F., Nasr, M., Zhang, C., Jagielski, M., Lee, K., Choquette Choo, C., & Carlini, N. (2023, September). Preventing Generation of Verbatim Memorization in Language Models Gives a False Sense of Privacy. In C. M. Keet, H.-Y. Lee, & S. Zarrieß (Eds.), Proceedings of the 16th international natural language generation conference (pp. 28–53). Prague, Czechia: Association for Computational Linguistics.
  107. iunn Ong, K. T., Kim, N., Gwak, M., Chae, H., Kwon, T., Jo, Y., won Hwang, S., Lee, D., & Yeo, J. (2025). Towards Lifelong Dialogue Agents via Timeline-based Memory Management. In Proceedings of the 2025 conference of the north american chapter of the association for computational linguistics: Human language technologies. Mexico City, Mexico: Association for Computational Linguistics.
  108. Izacard, G., Lewis, P., Lomeli, M., Hosseini, L., Petroni, F., Schick, T., Dwivedi-Yu, J., Joulin, A., Riedel, S., & Grave, E. (2023, January). Atlas: few-shot learning with retrieval augmented language models. J. Mach. Learn. Res., 24(1).
  109. Jang, J., Boo, M., & Kim, H. (2023). Conversation chronicles: Towards diverse temporal and relational dynamics in multi-session conversations. arXiv preprint arXiv:2310.13420.
  110. Jang, Y., Lee, K.-i., Bae, H., Lee, H., & Jung, K. (2024, June). IterCQR: Iterative Conversational Query Reformulation with Retrieval Guidance. In K. Duh, H. Gomez, & S. Bethard (Eds.), Proceedings of the 2024 conference of the north american chapter of the association for computational linguistics: Human language technologies (volume 1: Long papers) (pp. 8121–8138).
  111. Jhunjhunwala, D., Wang, S., & Joshi, G. (2024). FedFisher: Leveraging Fisher Information for One-Shot Federated Learning. In Aistats (pp. 1612–1620).
  112. Jiang, H., Wu, Q., Lin, C.-Y., Yang, Y., & Qiu, L. (2023, December). LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models. In H. Bouamor, J. Pino, & K. Bali (Eds.), Proceedings of the 2023 conference on empirical methods in natural language processing (pp. 13358–13376).
  113. Jiang, W., Subramanian, S., Graves, C., Alonso, G., Yazdanbakhsh, A., & Dadu, V. (2025). Rago: Systematic performance optimization for retrieval-augmented generation serving. Available online: https://arxiv.org/abs/2503.14649 (accessed on).
  114. Jiang, X., Li, F., Zhao, H., Wang, J., Shao, J., Xu, S., Zhang, S., Chen, W., Tang, X., Chen, Y., & et al. (2024). Long term memory: The foundation of ai self-evolution. arXiv.
  115. Jiang, Y., Wang, H., Xie, L., Zhao, H., Zhang, C., Qian, H., & Lui, J. C. (2024). D-LLM: A Token Adaptive Computing Resource Allocation Strategy for Large Language Models. In A. Globerson et al. (Eds.), Advances in neural information processing systems (Vol. 37, pp. 1725–1749).
  116. Jiang, Z., Xu, F., Gao, L., Sun, Z., Liu, Q., Dwivedi-Yu, J., Yang, Y., Callan, J., & Neubig, G. (2023b, December). Active Retrieval Augmented Generation. In H. Bouamor, J. Pino, & K. Bali (Eds.), Proceedings of the 2023 conference on empirical methods in natural language processing (pp. 7969–7992).
  117. Jiang, Z., Xu, F. F., Gao, L., Sun, Z., Liu, Q., Dwivedi-Yu, J., Yang, Y., Callan, J., & Neubig, G. (2023a). Active retrieval augmented generation. arXiv preprint arXiv:2305.06983.
  118. Jin, B., Yoon, J., Han, J., & Arik, S. Ö. (2024). Long-Context LLMs Meet RAG: Overcoming Challenges for Long Inputs in RAG. CoRR, abs/2410.05983. Available online: https://doi.org/10.48550/arXiv.2410.05983 (accessed on).
  119. Jin, B., Zeng, H., Wang, G., Chen, X., Wei, T., Li, R., Wang, Z., Li, Z., Li, Y., Lu, H., et al. (2023). Language Models As Semantic Indexers. arXiv preprint arXiv:2310.07815.
  120. Jin, H., & Wu, Y. (2025). CE-CoLLM: Efficient and Adaptive Large Language Models Through Cloud-Edge Collaboration. In 2025 ieee international conference on web services (icws) (p. 316-323).
  121. Jin, H., Zhang, Y., Meng, D., Wang, J., & Tan, J. (2024). A Comprehensive Survey on Process-Oriented Automatic Text Summarization with Exploration of LLM-Based Methods. CoRR, abs/2403.02901. Available online: https://doi.org/10.48550/arXiv.2403.02901 (accessed on).
  122. Johnson-Laird, P. N. (1983). Mental models: Towards a cognitive science of language, inference, and consciousness (Vol. 6). Harvard University Press.
  123. Jouppi, N. P., Young, C., Patil, N., Patterson, D., Agrawal, G., Bajwa, R., Bates, S., Bhatia, S., Boden, N., Borchers, A., Boyle, R., Cantin, P.-l., Chao, C., Clark, C., Coriell, J., Daley, M., Dau, M., Dean, J., Gelb, B., . . . Yoon, D. H. (2017, June). In-Datacenter Performance Analysis of a Tensor Processing Unit. SIGARCH Comput. Archit. News, 45(2), 1–12.
  124. Kadavy, D. (2021). Digital zettelkasten: Principles, methods, and examples. Kadavy, Inc.
  125. Kaiya, Z., Naim, M., Kondic, J., Cortes, M., Ge, J., Luo, S., Yang, G. R., & Ahn, A. (2023). Lyfe agents: Generative agents for low-cost real-time social interactions. arXiv.
  126. Kandpal, N., Wallace, E., & Raffel, C. (2022). Deduplicating training data mitigates privacy risks in language models. Available online: https://arxiv.org/abs/2202.06539 (accessed on).
  127. Kang, H., Zhang, Q., Kundu, S., Jeong, G., Liu, Z., Krishna, T., & Zhao, T. (2024). Gear: An efficient kv cache compression recipe for near-lossless generative inference of llm. Available online: https://arxiv.org/abs/2403.05527 (accessed on).
  128. Kapoor, S., Henderson, P., & Narayanan, A. (2024). Promises and pitfalls of artificial intelligence for legal applications. CoRR, abs/2402.01656. Available online: https://doi.org/10.48550/arXiv.2402.01656 (accessed on).
  129. Karpukhin, V., Oguz, B., Min, S., Lewis, P. S., Wu, L., Edunov, S., Chen, D., & Yih, W.-t. (2020). Dense passage retrieval for open-domain question answering. In Emnlp (1) (pp. 6769–6781).
  130. Ke, Z., Kong, W., Li, C., Zhang, M., Mei, Q., & Bendersky, M. (2024). Bridging the Preference Gap between Retrievers and LLMs. arXiv preprint arXiv:2401.06954.
  131. Khandelwal, U., Levy, O., Jurafsky, D., Zettlemoyer, L., & Lewis, M. (2019). Generalization through memorization: Nearest neighbor language models. arXiv preprint arXiv:1911.00172.
  132. Khattab, O., Santhanam, K., Li, X. L., Hall, D., Liang, P., Potts, C., & Zaharia, M. (2022). Demonstrate-Search-Predict: Composing retrieval and language models for knowledge-intensive NLP. arXiv preprint arXiv:2212.14024.
  133. Khattab, O., & Zaharia, M. (2020). ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT. In Proceedings of the 43rd international acm sigir conference on research and development in information retrieval (p. 39–48).
  134. Kim, M., Shim, K., Choi, J., & Chang, S. (2024, November). InfiniPot: Infinite Context Processing on Memory- Constrained LLMs. In Y. Al-Onaizan, M. Bansal, & Y.-N. Chen (Eds.), Proceedings of the 2024 conference on empirical methods in natural language processing (pp. 16046–16060).
  135. Kim, S. H., Ka, K., Jo, Y., Hwang, S.-w., Lee, D., & Yeo, J. (2024). Ever-Evolving Memory by Blending and Refining the Past. arXiv preprint arXiv:2403.04787. Available online: https://arxiv.org/abs/2403.04787 (accessed on).
  136. Kirkpatrick, J., Pascanu, R., Rabinowitz, N., Veness, J., Desjardins, G., Rusu, A. A., Milan, K., Quan, J., Ramalho, T., Grabska-Barwinska, A., et al. (2017). Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 114(13), 3521–3526.
  137. Kwon, W., Li, Z., Zhuang, S., Sheng, Y., Zheng, L., Yu, C. H., Gonzalez, J., Zhang, H., & Stoica, I. (2023). Efficient Memory Management for Large Language Model Serving with PagedAttention. In Proceedings of the 29th symposium on operating systems principles (p. 611–626).
  138. Lai, K., Tang, Z., Pan, X., Dong, P., Liu, X., Chen, H., Shen, L., Li, B., & Chu, X. (2025). Mediator: Memory-efficient LLM Merging with Less Parameter Conflicts and Uncertainty Based Routing. arxiv preprint arXiv:2502.04411.
  139. Laird, J. E. (2019). The soar cognitive architecture. MIT press.
  140. Langchain. (2023). Recursively split by character. https://python.langchain.com/docs/modules/data_connection/ document_transformers/recursive_text_splitter.
  141. Lee, G., Hartmann, V., Park, J., Papailiopoulos, D., & Lee, K. (2023a). Prompted LLMs as Chatbot Modules for Long Open-domain Conversation. In A. Rogers, J. L. Boyd-Graber, & N. Okazaki (Eds.), Findings of the association for computational linguistics: ACL 2023, toronto, canada, july 9-14, 2023 (pp. 4536–4554).
  142. Lee, G., Hartmann, V., Park, J., Papailiopoulos, D., & Lee, K. (2023b). Prompted llms as chatbot modules for long open-domain conversation. arXiv preprint arXiv:2305.04533.
  143. Lee, K., Ippolito, D., Nystrom, A., Zhang, C., Eck, D., Callison-Burch, C., & Carlini, N. (2022). Deduplicating Training Data Makes Language Models Better. In Proceedings of the 60th annual meeting of the association for computational linguistics.
  144. Lee, N., Ajanthan, T., & Torr, P. (2019). SNIP: SINGLE-SHOT NETWORK PRUNING BASED ON CONNECTION SENSITIVITY. In International conference on learning representations. Available online: https://openreview.net/ forum?id=B1VZqjAcYX (accessed on).
  145. Leviathan, Y., Kalman, M., & Matias, Y. (2025). Selective Attention Improves Transformer. In The thirteenth international conference on learning representations. Available online: https://openreview.net/forum?id=v0 FzmPCd1e (accessed on).
  146. Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin, V., Goyal, N., Küttler, H., Lewis, M., Yih, W.-t., Rocktäschel, T., et al. (2020). Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in neural information processing systems, 33, 9459–9474.
  147. Leydesdorff, S. (2017). Memory cultures: Memory, subjectivity and recognition. Routledge.
  148. LI, H., Li, Y., Tian, A., Tang, T., Xu, Z., Chen, X., HU, N., Dong, W., Qing, L., & Chen, L. (2025). A Survey on Large Language Model Acceleration based on KV Cache Management. Transactions on Machine Learning Research. Available online: https://openreview.net/forum?id=z3JZzu9EA3 (accessed on).
  149. Li, H., Li, Y., Tian, A., Tang, T., Xu, Z., Chen, X., HU, N., Dong, W., Qing, L., & Chen, L. (2025). A Survey on Large Language Model Acceleration based on KV Cache Management. Transactions on Machine Learning Research. Available online: https://openreview.net/forum?id=z3JZzu9EA3 (accessed on).
  150. Li, H., Yang, C., Zhang, A., Deng, Y., Wang, X., & Chua, T.-S. (2024). Hello again! llm-powered personalized agent for long-term dialogue. arXiv preprint arXiv:2406.05925.
  151. Li, K., Patel, O., Viégas, F., Pfister, H., & Wattenberg, M. (2024). Inference-time intervention: Eliciting truthful answers from a language model. In Advances in neural information processing systems (Vol. 36).
  152. Li, L., Zhang, Y., Liu, D., & Chen, L. (2023). Large language models for generative recommendation: A survey and visionary discussions. arXiv.
  153. Li, N., Gao, C., Li, Y., & Liao, Q. (2023). Large language model-empowered agents for simulating macroeconomic activities. arXiv.
  154. Li, S., He, Y., Guo, H., Bu, X., Bai, G., Liu, J., Liu, J., Qu, X., Li, Y., Ouyang,W., Su,W., & Zheng, B. (2024, November). GraphReader: Building Graph-based Agent to Enhance Long-Context Abilities of Large Language Models. In Y. Al-Onaizan, M. Bansal, & Y.-N. Chen (Eds.), Findings of the association for computational linguistics: Emnlp 2024 (pp. 12758–12786).
  155. Li, W., Jiang, G., Ding, X., Tao, Z., Hao, C., Xu, C., Zhang, Y., & Wang, H. (2025). Flowkv: A disaggregated inference framework with low-latency kv cache transfer and load-aware scheduling. Available online: https://arxiv.org/abs/2504.03775 (accessed on).
  156. Li, X., & Li, J. (2023). AnglE-optimized Text Embeddings. arXiv preprint arXiv:2309.12871.
  157. Li, X., Nie, E., & Liang, S. (2023). From Classification to Generation: Insights into Crosslingual Retrieval Augmented ICL. arXiv preprint arXiv:2311.06595.
  158. Li, Y., Huang, Y., Yang, B., Venkitesh, B., Locatelli, A., Ye, H., Cai, T., Lewis, P., & Chen, D. (2024). SnapKV: LLM Knows What You are Looking for Before Generation. arXiv preprint arXiv:2404.14469.
  159. Li, Y.,Wang, S., Ding, H., & Chen, H. (2023). Large language models in finance: A survey. In Proceedings of the fourth acm international conference on ai in finance (pp. 374–382).
  160. Li, Y., Wen, H., Wang, W., Li, X., Yuan, Y., Liu, G., Liu, J., Xu, W., Wang, X., Sun, Y., & et al. (2024). Personal llm agents: Insights and survey about the capability, efficiency and security. arXiv.
  161. Li, Y., Zhang, Y., & Sun, L. (2023). Metaagents: Simulating interactions of human behaviors for llm-based task-oriented coordination via collaborative generative agents. arXiv preprint arXiv:2310.06500.
  162. Li, Z., Song, S., Xi, C., Wang, H., Tang, C., Niu, S., Chen, D., Yang, J., Li, C., Yu, Q., et al. (2025). Memos: A memory os for ai system. arXiv preprint arXiv:2507.03724.
  163. Li, Z., Xiao, C.,Wang, Y., Liu, X., Tang, Z., Lu, B., Yang, M., Chen, X., &Chu, X. (2025). Antkv: Anchor token-aware subbit vector quantization for kv cache in large language models. Available online: https://arxiv.org/abs/2506.19505 (accessed on).
  164. Lin, J., Dai, X., Xi, Y., Liu, W., Chen, B., Li, X., Zhu, C., Guo, H., Yu, Y., Tang, R., & et al. (2023). How can recommender systems benefit from large language models: A survey. arXiv.
  165. Lin, X. V., Chen, X., Chen, M., Shi,W., Lomeli, M., James, R., Rodriguez, P., Kahn, J., Szilvasy, G., Lewis, M., et al. (2023). RA-DIT: Retrieval-Augmented Dual Instruction Tuning. arXiv preprint arXiv:2310.01352.
  166. Lin, Y., Tang, H., Yang, S., Zhang, Z., Xiao, G., Gan, C., & Han, S. (2025). Qserve: W4a8kv4 quantization and system co-design for efficient llm serving. Available online: https://arxiv.org/abs/2405.04532 (accessed on).
  167. Liu, A., Liu, J., Pan, Z., He, Y., Haffari, G., & Zhuang, B. (2024). MiniCache: KV Cache Compression in Depth Dimension for Large Language Models. In A. Globerson et al. (Eds.), Advances in neural information processing systems (Vol. 37, pp. 139997–140031).
  168. Liu, D., Chen, M., Lu, B., Jiang, H., Han, Z., Zhang, Q., Chen, Q., Zhang, C., Ding, B., Zhang, K., Chen, C., Yang, F., Yang, Y., & Qiu, L. (2024). Retrievalattention: Accelerating long-context llm inference via vector retrieval. Available online: https://arxiv.org/abs/2409.10516 (accessed on).
  169. Liu, J., Li, L., Xiang, T., Wang, B., & Qian, Y. (2023, December). TCRA-LLM: Token Compression Retrieval Augmented Large Language Model for Inference Cost Reduction. In H. Bouamor, J. Pino, & K. Bali (Eds.), Findings of the association for computational linguistics: Emnlp 2023 (pp. 9796–9810).
  170. Liu, J., Qiu, Z., Li, Z., Dai, Q., Zhu, J., Hu, M., Yang, M., & King, I. (2025). A Survey of Personalized Large Language Models: Progress and Future Directions. arXiv preprint arXiv:2502.11528.
  171. Liu, L., Yang, X., Shen, Y., Hu, B., Zhang, Z., Gu, J., & Zhang, G. (2023). Think-in-memory: Recalling and post-thinking enable llms with long-term memory. arXiv preprint arXiv:2311.08719.
  172. Liu, N. F., Lin, K., Hewitt, J., Paranjape, A., Bevilacqua, M., Petroni, F., & Liang, P. (2024). Lost in the middle: How language models use long contexts. Transactions of the Association for Computational Linguistics, 12, 157–173.
  173. Liu, S., Ye, H., Xing, L., & Zou, J. (2023). In-context vectors: Making in context learning more effective and controllable through latent space steering. arXiv.
  174. Liu, W., Zhang, R., Zhou, A., Gao, F., & Liu, J. (2025). Echo: A large language model with temporal episodic memory. arXiv preprint arXiv:2502.16090.
  175. Liu, X., Chen, H., Hu, X., & Chu, X. (2025). FlowKV: Enhancing Multi-Turn Conversational Coherence in LLMs via Isolated Key-Value Cache Management. In First workshop on multi-turn interactions in large language models. Available online: https://openreview.net/forum?id=rZumU1owkr (accessed on).
  176. Liu, X., Tang, Z., Dong, P., Li, Z., Liu, Y., Li, B., Hu, X., & Chu, X. (2025). Chunkkv: Semantic-preserving kv cache compression for efficient long-context llm inference. Available online: https://arxiv.org/abs/2502.00299 (accessed on).
  177. Liu, Y., Li, H., Cheng, Y., Ray, S., Huang, Y., Zhang, Q., Du, K., Yao, J., Lu, S., Ananthanarayanan, G., Maire, M., Hoffmann, H., Holtzman, A., & Jiang, J. (2024). Cachegen: Kv cache compression and streaming for fast large language model serving. Available online: https://arxiv.org/abs/2310.07240 (accessed on).
  178. Liu, Z., Desai, A., Liao, F., Wang, W., Xie, V., Xu, Z., Kyrillidis, A., & Shrivastava, A. (2023). Scissorhands: Exploiting the Persistence of Importance Hypothesis for LLM KV Cache Compression at Test Time. In Thirty-seventh conference on neural information processing systems.
  179. Liu, Z., Yuan, J., Jin, H., Zhong, S., Xu, Z., Braverman, V., Chen, B., & Hu, X. (2024). KIVI: A Tuning-Free Asymmetric 2bit Quantization for KV Cache. In International conference on machine learning, icml 2024 (pp. 32332–32344).
  180. Liu, Z., Zhong, A., Li, Y., Yang, L., Ju, C., Wu, Z., Ma, C., Shu, P., Chen, C., Kim, S., & et al. (2023). Radiology-gpt: A large language model for radiology. arXiv preprint arXiv:2306.08666.
  181. Lozhkov, A., Li, R., Allal, L. B., Cassano, F., Lamy-Poirier, J., Tazi, N., Tang, A., Pykhtar, D., Liu, J., Wei, Y., Liu, T., Tian, M., Kocetkov, D., Zucker, A., Belkada, Y., Wang, Z., Liu, Q., Abulkhanov, D., Paul, I., . . . et al. (2024). StarCoder 2 and The Stack v2: The Next Generation. CoRR, abs/2402.19173.
  182. Lu, J., An, S., Lin, M., Pergola, G., He, Y., Yin, D., Sun, X., & Wu, Y. (2023a). Memochat: Tuning llms to use memos for consistent long-range open-domain conversation. arXiv preprint arXiv:2308.08239.
  183. Lu, J., An, S., Lin, M., Pergola, G., He, Y., Yin, D., Sun, X., & Wu, Y. (2023b). Memochat: Tuning llms to use memos for consistent long-range open-domain conversation. arXiv preprint arXiv:2308.08239.
  184. Lu, Z., Fan, C.,Wei,W., Qu, X., Chen, D., & Cheng, Y. (2024a). Twin-Merging: Dynamic Integration of Modular Expertise in Model Merging. arXiv preprint arXiv:2406.15479.
  185. Lu, Z., Fan, C.,Wei,W., Qu, X., Chen, D., & Cheng, Y. (2024b, 06). Twin-Merging: Dynamic Integration of Modular Expertise in Model Merging. NIPS. Available online: https://arxiv.org/pdf/2406.15479v2.pdf (accessed on).
  186. Luo, Z., Xu, C., Zhao, P., Geng, X., Tao, C., Ma, J., Lin, Q., & Jiang, D. (2023). Augmented Large Language Models with Parametric Knowledge Guiding. arXiv preprint arXiv:2305.04757.
  187. Luohe, S., Zhang, H., Yao, Y., Li, Z., et al. (n.d.). Keep the Cost Down: A Review on Methods to Optimize LLM’s KV-Cache Consumption. In First conference on language modeling.
  188. Lyu, C., Du, Z., Xu, J., Duan, Y., Wu, M., Lynn, T., Aji, A. F., Wong, D. F., & Wang, L. (2024). A Paradigm Shift: The Future of Machine Translation Lies with Large Language Models. In Proceedings of the 2024 joint international conference on computational linguistics, language resources and evaluation, LREC/COLING 2024 (pp. 1339–1352).
  189. Ma, X., Gong, Y., He, P., Zhao, H., & Duan, N. (2023). Query Rewriting for Retrieval-Augmented Large Language Models. arXiv preprint arXiv:2305.14283.
  190. Mack, A., & Turner, A. (2024). Mechanistically eliciting latent behaviors in language models. Available online: https://www.lesswrong.com/posts/ioPnHKFyy4Cw2Gr2x (accessed on).
  191. Maharana, A., Lee, D.-H., Tulyakov, S., Bansal, M., Barbieri, F., & Fang, Y. (2024). Evaluating very long-term conversational memory of llm agents. arXiv preprint arXiv:2402.17753.
  192. Marczak, D., Twardowski, B., Trzciński, T., & Cygert, S. (2024). MagMax: Leveraging Model Merging for Seamless Continual Learning. In Eccv.
  193. Masry, A., & Hajian, A. (2024). LongFin: A Multimodal Document Understanding Model for Long Financial Domain Documents. CoRR, abs/2401.15050. Available online: https://doi.org/10.48550/arXiv.2401.15050 (accessed on).
  194. Matena, M. S., & Raffel, C. A. (2022). Merging models with fisher-weighted averaging. NeurIPS, 35, 17703–17716.
  195. McMahan, B., Moore, E., Ramage, D., Hampson, S., & y Arcas, B. A. (2017). Communication-efficient learning of deep networks from decentralized data. In Aistats (pp. 1273–1282).
  196. Mehta, S. V., Gupta, J., Tay, Y., Dehghani, M., Tran, V. Q., Rao, J., Najork, M., Strubell, E., & Metzler, D. (2022). DSI++: Updating transformer memory with new documents. arXiv preprint arXiv:2212.09744.
  197. mem0ai. (2024, July). mem0: The memory layer for personalized ai. mem0.ai.
  198. Meng, F., Tang, P., Tang, X., Yao, Z., Sun, X., & Zhang, M. (2025). Transmla: Multi-head latent attention is all you need. Available online: https://arxiv.org/abs/2502.07864 (accessed on).
  199. Meng, K., Sharma, A. S., Andonian, A. J., Belinkov, Y., & Bau, D. (2023). Mass-Editing Memory in a Transformer. In The eleventh international conference on learning representations. Available online: https://openreview.net/ forum?id=MkbcAHIYgyS (accessed on).
  200. Min, S., Lyu, X., Holtzman, A., Artetxe, M., Lewis, M., Hajishirzi, H., & Zettlemoyer, L. (2022). Rethinking the role of demonstrations: What makes in-context learning work? In Proceedings of the 2022 conference on empirical methods in natural language processing (pp. 11048–11064).
  201. Mishra, M., Stallone, M., Zhang, G., Shen, Y., Prasad, A., Soria, A. M., Merler, M., Selvam, P., Surendran, S., Singh, S., Sethi, M., Dang, X., Li, P., Wu, K., Zawad, S., Coleman, A., White, M., Lewis, M., Pavuluri, R., . . . Panda, R. (2024). Granite Code Models: A Family of Open Foundation Models for Code Intelligence. CoRR, abs/2405.04324.
  202. Modarressi, A., Imani, A., Fayyaz, M., & Schütze, H. (2023). Ret-llm: Towards a general read-write memory for large language models. arxiv.
  203. Montazeralghaem, A., Zamani, H., & Allan, J. (2020). A reinforcement learning framework for relevance feedback. In Proceedings of the 43rd international acm sigir conference on research and development in information retrieval (pp. 59–68).
  204. Muqeeth, M., Liu, H., & Raffel, C. (2024). Soft merging of experts with adaptive routing. TMLR.
  205. Murre, J. M., & Dros, J. (2015). Replication and analysis of ebbinghaus’ forgetting curve. PloS one, 10(7), e0120644.
  206. Nasr, M., Carlini, N., Hayase, J., Jagielski, M., Cooper, A. F., Ippolito, D., Choquette-Choo, C. A., Wallace, E., Tramèr, F., & Lee, K. (2023). Scalable extraction of training data from (production) language models. Available online: https://arxiv.org/abs/2311.17035 (accessed on).
  207. Nie, Y., Huang, H.,Wei, W., & Mao, X.-L. (2022, December). Capturing Global Structural Information in Long Document Question Answering with Compressive Graph Selector Network. In Proceedings of the 2022 conference on empirical methods in natural language processing (pp. 5036–5047).
  208. Nie, Y., Kong, Y., Dong, X., Mulvey, J. M., Poor, H. V., Wen, Q., & Zohren, S. (2024). A Survey of Large Language Models for Financial Applications: Progress, Prospects and Challenges. CoRR, abs/2406.11903. Available online: https://doi.org/10.48550/arXiv.2406.11903 (accessed on).
  209. OpenAI. (2024a). Memory and new controls for chatgpt. Available online: https://openai.com/index/memory-and -new-controls-for-chatgpt (accessed on).
  210. OpenAI. (2024b). New embedding models and api updates. Available online: https://openai.com/index/new -embedding-models-and-api-updates/ (accessed on).
  211. Packer, C., Wooders, S., Lin, K., Fang, V., Patil, S. G., Stoica, I., & Gonzalez, J. E. (2023). Memgpt: Towards llms as operating systems. arXiv preprint arXiv:2310.08560.
  212. Pan, H., Zhai, Z., Yuan, H., Lv, Y., Fu, R., Liu, M., Wang, Z., & Qin, B. (2023). Kwaiagents: Generalized information-seeking agent system with large language models. arXiv.
  213. Pan, J., Gao, T., Chen, H., & Chen, D. (2023). What in-context learning “learns” in-context: Disentangling task recognition and task learning. In Findings of the association for computational linguistics: Acl 2023 (pp. 8298–8319).
  214. Pan, J., & Li, G. (2025). A survey of llm inference systems. Available online: https://arxiv.org/abs/2506.21901 (accessed on).
  215. Pan, Z.,Wu, Q., Jiang, H., Luo, X., Cheng, H., Li, D., Yang, Y., Lin, C.-Y., Zhao, H. V., Qiu, L., & et al. (2025). On memory construction and retrieval for personalized conversational agents. arXiv preprint arXiv:2502.05589.
  216. Park, J. S., O’Brien, J., Cai, C. J., Morris, M. R., Liang, P., & Bernstein, M. S. (2023). Generative agents: Interactive simulacra of human behavior. In Proceedings of the 36th annual acm symposium on user interface software and technology (pp. 1–22).
  217. Peng, W., Li, G., Jiang, Y., Wang, Z., Ou, D., Zeng, X., Chen, E., et al. (2023). Large language model based long-tail query rewriting in taobao search. arXiv preprint arXiv:2311.03758.
  218. Pope, R., Douglas, S., Chowdhery, A., Devlin, J., Bradbury, J., Levskaya, A., Heek, J., Xiao, K., Agrawal, S., & Dean, J. (2022). Efficiently scaling transformer inference. Available online: https://arxiv.org/abs/2211.05102 (accessed on).
  219. Qian, C., Cong, X., Yang, C., Chen, W., Su, Y., Xu, J., Liu, Z., & Sun, M. (2023). Communicative agents for software development. arXiv preprint arXiv:2307.07924.
  220. Qian, H., Zhang, P., Liu, Z., Mao, K., & Dou, Z. (2024). Memorag: Moving towards next-gen rag via memoryinspired knowledge discovery. arXiv preprint arXiv:2409.05591.
  221. Qiang, Z.,Wang,W., & Taylor, K. (2023). Agent-om: Leveraging large language models for ontology matching. arXiv.
  222. Qu, Z., Li, X., Duan, R., Liu, Y., Tang, B., & Lu, Z. (2022). Generalized federated learning via sharpness aware minimization. In International conference on machine learning (pp. 18250–18280).
  223. Qwen, :, Yang, A., Yang, B., Zhang, B., Hui, B., Zheng, B., et al. (2025). Qwen2.5 technical report. Available online: https://arxiv.org/abs/2412.15115 (accessed on).
  224. Raventos, A., Paul, M., Chen, F., & Ganguli, S. (2023). Pretraining task diversity and the emergence of non-bayesian in-context learning for regression. In Thirty-seventh conference on neural information processing systems.
  225. Reddy, V., Koncel-Kedziorski, R., Lai, V. D., Krumdick, M., Lovering, C., & Tanner, C. (2024). DocFinQA: A Long-Context Financial Reasoning Dataset. In Proceedings of the 62nd annual meeting of the association for computational linguistics, ACL 2024 - short papers (pp. 445–458).
  226. Saad-Falcon, J., Fu, D. Y., Arora, S., Guha, N., & Ré, C. (2024). Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT. In Forty-first international conference on machine learning, ICML 2024. OpenReview.net.
  227. Safaya, A., & Yuret, D. (2024). Neurocache: Efficient vector retrieval for long-range language modeling. arXiv preprint arXiv:2407.02486.
  228. Salama, R., Cai, J., Yuan, M., Currey, A., Sunkara, M., Zhang, Y., & Benajiba, Y. (2025). Meminsight: Autonomous memory augmentation for llm agents. arXiv preprint arXiv:2503.21760.
  229. Samsami, M. R., Zholus, A., Rajendran, J., & Chandar, S. (2024). Mastering Memory Tasks withWorld Models. In The twelfth international conference on learning representations. Available online: https://openreview.net/forum?id=1vDArHJ68h (accessed on).
  230. Saxena, U., Saha, G., Choudhary, S., & Roy, K. (2024, November). Eigen Attention: Attention in Low-Rank Space for KV Cache Compression. In Y. Al-Onaizan, M. Bansal, & Y.-N. Chen (Eds.), Findings of the association for computational linguistics: Emnlp 2024 (pp. 15332–15344). Miami, Florida, USA: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.findings-emnlp.899/ (accessed on). https://doi.org/10.18653/v1/2024.findings-emnlp.899.
  231. Shan, L., Luo, S., Zhu, Z., Yuan, Y., & Wu, Y. (2025). Cognitive memory in large language models. arXiv preprint arXiv:2504.02441.
  232. Shang, J., Zheng, Z.,Wei, J., Ying, X., Tao, F., & Team, M. (2024). Ai-native memory: A pathway from llms towards agi. arXiv preprint arXiv:2406.18312.
  233. Shao, B., & Yan, J. (2024). A long-context language model for deciphering and generating bacteriophage genomes. Nature Communications, 15(1), 9392.
  234. Shao, Y., Li, L., Dai, J., & Qiu, X. (n.d.). Character-llm: A trainable agent for role-playing.
  235. Shao, Y., Li, L., Dai, J., & Qiu, X. (2023). Character-llm: A trainable agent for role-playing. arXiv preprint arXiv:2310.10158.
  236. Shazeer, N., Mirhoseini, A., Maziarz, K., Davis, A., Le, Q., Hinton, G., & Dean, J. (2017, 01). Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer. Available online: https://arxiv.org/pdf/1701.06538.pdf (accessed on).
  237. Shen, L., Mishra, A., & Khashabi, D. (2024). Do pretrained transformers learn in-context by gradient descent? arxiv, arXiv:2310.08540. Available online: https://arxiv.org/abs/2310.08540 (accessed on).
  238. Sheng, Y., Zheng, L., Yuan, B., Li, Z., Ryabinin, M., Chen, B., Liang, P., Ré, C., Stoica, I., & Zhang, C. (2023). FlexGen: high-throughput generative inference of large language models with a single GPU. In Proceedings of the 40th international conference on machine learning. JMLR.org.
  239. Sherwood, L., Kell, R. T., & Ward, C. (2004). Human physiology: from cells to systems. Thomson/Brooks/Cole.
  240. Shi, F., Chen, X., Misra, K., Scales, N., Dohan, D., Chi, E. H., Schärli, N., & Zhou, D. (2023). Large language models can be easily distracted by irrelevant context. In International conference on machine learning (pp. 31210–31227).
  241. Shi, K., Sun, X., Li, Q., & Xu, G. (2024). Compressing Long Context for Enhancing RAG with AMR-based Concept Distillation. CoRR, abs/2405.03085. Available online: https://doi.org/10.48550/arXiv.2405.03085 (accessed on). https://doi.org/10.48550/ARXIV.2405.03085.
  242. Shinn, N., Cassano, F., Gopinath, A., Narasimhan, K., & Yao, S. (2024). Reflexion: Language agents with verbal reinforcement learning. Advances in Neural Information Processing Systems, 36.
  243. Shinn, N., Cassano, F., Gopinath, A., Narasimhan, K. R., & Yao, S. (2023). Reflexion: Language agents with verbal reinforcement learning. In Thirty-seventh conference on neural information processing systems.
  244. Shinwari, H. U. K., & Usama, M. (2025). Memory-Augmented Architecture for Long-Term Context Handling in Large Language Models. arXiv preprint arXiv:2506.18271.
  245. Singh, A. K., Chan, S. C., Moskovitz, T., Grant, E., Saxe, A. M., & Hill, F. (2023). The transient nature of emergent in-context learning in transformers. In Thirty-seventh conference on neural information processing systems.
  246. Solso, R. L., & Kagan, J. (1979). Cognitive psychology. Houghton Mifflin Harcourt P.
  247. Squire, L. R. (2009). Memory and brain systems: 1969-2009. J Neurosci. 2009 Oct 14;29(41):12711-6. doi: 10.1523/JNEUROSCI.3575-09.2009. PMID: 19828780; PMCID: PMC2791502. J Neurosci, 29(41), 12711–12716.
  248. Sridhar, S., Khamaj, A., & Asthana, M. (2023). Cognitive neuroscience perspective on memory: overview and summary. Frontiers in human neuroscience, 17, 1217093.
  249. Subramani, N., Suresh, N., & Peters, M. E. (2022). Extracting latent steering vectors from pretrained language models. arXiv.
  250. Sun, H., Cai, H., Wang, B., Hou, Y., Wei, X., Wang, S., Zhang, Y., & Yin, D. (2024, November). Towards Verifiable Text Generation with Evolving Memory and Self-Reflection. In Y. Al-Onaizan, M. Bansal, & Y.-N. Chen (Eds.), Proceedings of the 2024 conference on empirical methods in natural language processing (pp. 8211–8227). Miami, Florida, USA: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.emnlp-main.469/ (accessed on).
  251. Sun, R. (2001). Duality of the mind: A bottom-up approach toward cognition. Psychology Press.
  252. Sun, Z., Wang, X., Tay, Y., Yang, Y., & Zhou, D. (2022). Recitation-augmented language models. arXiv preprint arXiv:2210.01296.
  253. Sutton, R. S., & Barto, A. G. (2018). Reinforcement learning: An introduction. MIT press.
  254. Tan, X., Jiang, Y., Yang, Y., & Xu, H. (2025). Teola: Towards end-to-end optimization of llm-based applications. Available online: https://arxiv.org/abs/2407.00326 (accessed on).
  255. Tang, A., Shen, L., Luo, Y., Yin, N., Zhang, L., & Tao, D. (2024). Merging Multi-Task Models viaWeight-Ensembling Mixture of Experts. ICML.
  256. Tang, J., Zhao, Y., Zhu, K., Xiao, G., Kasikci, B., & Han, S. (2024, 21–27 Jul). QUEST: Query-Aware Sparsity for Efficient Long-Context LLM Inference. In Proceedings of the 41st international conference on machine learning (Vol. 235, pp. 47901–47911). PMLR.
  257. Teja, R. (2023). Evaluating the ideal chunk size for a rag system using llamaindex. https://www.llamaindex.ai/blog/ evaluating-the-ideal-chunk-size-for-a-rag-system-using-llamaindex-6207e5d3fec5.
  258. Tirumala, K., Markosyan, A. H., Zettlemoyer, L., & Aghajanyan, A. (2022). Memorization without overfitting: Analyzing the training dynamics of large language models. Available online: https://arxiv.org/abs/2205.10770 (accessed on).
  259. Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.-A., Lacroix, T., Rozière, B., Goyal, N., Hambro, E., Azhar, F., Rodriguez, A., Joulin, A., Grave, E., & Lample, G. (2023). Llama: Open and efficient foundation language models. Available online: https://arxiv.org/abs/2302.13971 (accessed on).
  260. Touvron, H., Martin, L., Stone, K., et al. (2023). Llama 2: Open foundation and fine-tuned chat models. Available online: https://arxiv.org/abs/2307.09288 (accessed on).
  261. Tsai, Y., Liu, M., & Ren, H. (2023). Rtlfixer: Automatically fixing rtl syntax errors with large language models. arXiv preprint arXiv:2311.16543.
  262. Tulving, E., & Donaldson, W. (1972). Episodic and semantic memory. Academic Press.
  263. Turner, A., Kurzeja, M., Orr, D., & Elson, D. (2025). Steering Gemini using bidpo vectors. Available online: https://turntrout.com/gemini-steering (accessed on).
  264. Turner, A. M., Thiergart, L., Leech, G., Udell, D., Vazquez, J. J., Mini, U., & MacDiarmid, M. (2023). Steering language models with activation engineering. arXiv preprint arXiv:2308.10248.
  265. Tworkowski, S., Staniszewski, K., Pacek, M. a., Wu, Y., Michalewski, H., & Mił o´s, P. (2023). Focused Transformer: Contrastive Training for Context Scaling. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, & S. Levine (Eds.), Advances in neural information processing systems (Vol. 36, pp. 42661–42688). Curran Associates, Inc. Available online: https://proceedings.neurips.cc/paper_files/paper/2023/file/8511d06d5 590f4bda24d42087802cc81-Paper-Conference.pdf (accessed on).
  266. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., & Polosukhin, I. (2017). Attention is all you need. In Proceedings of the 31st international conference on neural information processing systems (p. 6000–6010). Red Hook, NY, USA: Curran Associates Inc.
  267. VoyageAI. (2023). Voyage’s embedding models. https://docs.voyageai.com/embeddings/.
  268. Wan, L., & Ma,W. (2025). StoryBench: A Dynamic Benchmark for Evaluating Long-Term Memory with Multi Turns. arXiv preprint arXiv:2506.13356.
  269. Wang, B., Liang, X., Yang, J., Huang, H.,Wu, S.,Wu, P., Lu, L., Ma, Z., & Li, Z. (2024). Enhancing large language model with self-controlled memory framework. Available online: https://arxiv.org/abs/2304.13343 (accessed on).
  270. Wang, B., Xie, Q., Pei, J., Chen, Z., Tiwari, P., Li, Z., & Fu, J. (2023). Pre-trained language models in biomedical domain: A systematic survey. ACM Computing Surveys, 56(3), 1–52.
  271. Wang, C., Liu, X., Liu, Y., Zhu, Y., Mo, X., Jiang, J., & Chen, H. (2025). When to Reason: Semantic Router for vLLM. arXiv preprint arXiv:2510.08731.
  272. Wang, G., Xie, Y., Jiang, Y., Mandlekar, A., Xiao, C., Zhu, Y., Fan, L., & Anandkumar, A. (2023). Voyager: An open-ended embodied agent with large language models. arXiv preprint arXiv:2305.16291.
  273. Wang, H., Liu, C., Xi, N., Qiang, Z., Zhao, S., Qin, B., & Liu, T. (2023). Huatuo: Tuning llama model with chinese medical knowledge. arXiv preprint arXiv:2304.06975.
  274. Wang, H., Zhao, S., Qiang, Z., Li, Z., Xi, N., Du, Y., Cai, M., Guo, H., Chen, Y., Xu, H., & et al. (2023). Knowledgetuning large language models with structured medical knowledge bases for reliable response generation in chinese. arXiv preprint arXiv:2309.04175.
  275. Wang, J., Huang, Y., Chen, C., Liu, Z., Wang, S., & Wang, Q. (2024). Software testing with large language models: Survey, landscape, and vision. IEEE Transactions on Software Engineering.
  276. Wang, K., Dimitriadis, N., Ortiz-Jimenez, G., Fleuret, F., & Frossard, P. (2024). Localizing Task Information for Improved Model Merging and Compression. ICML.
  277. Wang, L., Du, Z., Jiao,W., Lyu, C., Pang, J., Cui, L., Song, K.,Wong, D. F., Shi, S., & Tu, Z. (2024). Benchmarking and Improving Long-Text Translation with Large Language Models. In L. Ku, A. Martins, & V. Srikumar (Eds.), Findings of the association for computational linguistics, ACL 2024, bangkok, thailand and virtual meeting, august 11-16, 2024 (pp. 7175–7187). Association for Computational Linguistics. Available online: https://doi.org/10.18653/v1/2024.findings-acl.428 (accessed on). https://doi.org/10.18653/V1/2024.FINDINGS-ACL.428.
  278. Wang, L., Ma, C., Feng, X., Zhang, Z., Yang, H., Zhang, J., Chen, Z., Tang, J., Chen, X., Lin, Y., & et al. (2023). A survey on large language model based autonomous agents. arxiv.
  279. Wang, L., Yang, N., Huang, X., Yang, L., Majumder, R., & Wei, F. (2024). Improving Text Embeddings with Large Language Models. In L. Ku, A. Martins, & V. Srikumar (Eds.), Proceedings of the 62nd annual meeting of the association for computational linguistics (volume 1: Long papers), ACL 2024, bangkok, thailand, august 11-16, 2024 (pp. 11897–11916). Association for Computational Linguistics. Available online: https://doi.org/10.18653/ v1/2024.acl-long.642 (accessed on). https://doi.org/10.18653/V1/2024.ACL-LONG.642.
  280. Wang, L., Yang, N., &Wei, F. (2023). Learning to retrieve in-context examples for large language models. arXiv preprint arXiv:2307.07164.
  281. Wang, L., Zhang, J., Yang, H., Chen, Z., Tang, J., Zhang, Z., Chen, X., Lin, Y., Song, R., Zhao, W. X., Xu, J., Dou, Z., Wang, J., &Wen, J.-R. (2023). When large language model based agent meets user behavior analysis: A novel user simulation paradigm.
  282. Wang, L., Zhang, X., Su, H., & Zhu, J. (2024). A comprehensive survey of continual learning: Theory, method and application. IEEE Transactions on Pattern Analysis and Machine Intelligence.
  283. Wang, P., Li, Z., Zhang, N., Xu, Z., Yao, Y., Jiang, Y., Xie, P., Huang, F., & Chen, H. (2024). Wise: Rethinking the knowledge memory for lifelong model editing of large language models. Advances in Neural Information Processing Systems, 37, 53764–53797.
  284. Wang, Q., Fu, Y., Cao, Y., Wang, S., Tian, Z., & Ding, L. (2025a). Recursively summarizing enables long-term dialogue memory in large language models. Neurocomputing, 639, 130193. Available online: https:// www.sciencedirect.com/science/article/pii/S0925231225008653 (accessed on). https://doi.org/https:// doi.org/10.1016/j.neucom.2025.130193.
  285. Wang, Q., Fu, Y., Cao, Y.,Wang, S., Tian, Z., & Ding, L. (2025b). Recursively Summarizing Enables Long-Term Dialogue Memory in Large Language Models. Neurocomputing, 130193. Available online: https://www .sciencedirect.com/science/article/abs/pii/S0925231225008653 (accessed on). https://doi.org/10.1016/ j.neucom.2025.130193.
  286. Wang, Q., Tang, Z., & He, B. (2025). Can LLM Simulations Truly Reflect Humanity? A Deep Dive. In The fourth blogpost track at iclr 2025.
  287. Wang, S., Zhu, Y., Liu, H., Zheng, Z., Chen, C., & Li, J. (2024). Knowledge editing for large language models: A survey. ACM Computing Surveys, 57(3), 1–37.
  288. Wang, W., Dong, L., Cheng, H., Liu, X., Yan, X., Gao, J., & Wei, F. (2023). Augmenting Language Models with Long-Term Memory. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, & S. Levine (Eds.), Advances in neural information processing systems 36: Annual conference on neural information processing systems 2023, neurips 2023, new orleans, la, usa, december 10 - 16, 2023. Available online: http://papers.nips.cc/paper_files/ paper/2023/hash/ebd82705f44793b6f9ade5a669d0f0bf-Abstract-Conference.html (accessed on).
  289. Wang,W., Lin, X., Feng, F., He, X., & Chua, T.-S. (2023). Generative recommendation: Towards next-generation recommender paradigm. arXiv.
  290. Wang, X., Salmani, M., Omidi, P., Ren, X., Rezagholizadeh, M., & Eshaghi, A. (2024). Beyond the Limits: A Survey of Techniques to Extend the Context Length in Large Language Models. In Proceedings of the thirty-third international joint conference on artificial intelligence, IJCAI 2024, jeju, south korea, august 3-9, 2024 (pp. 8299–8307). ijcai.org. Available online: https://www.ijcai.org/proceedings/2024/917 (accessed on).
  291. Wang, X., Yang, Q., Qiu, Y., Liang, J., He, Q., Gu, Z., Xiao, Y., & Wang, W. (2023). KnowledGPT: Enhancing Large Language Models with Retrieval and Storage Access on Knowledge Bases. arXiv preprint arXiv:2308.11761.
  292. Wang, Y., Gao, Y., Chen, X., Jiang, H., Li, S., Yang, J., Yin, Q., Li, Z., Li, X., Yin, B., & et al. (2024). Memoryllm: Towards self-updatable large language models. arXiv preprint arXiv:2402.04624.
  293. Wang, Y., Han, C., Wu, T., He, X., Zhou, W., Sadeq, N., Chen, X., He, Z., Wang, W., Haffari, G., et al. (2024). Towards lifespan cognitive systems. arXiv preprint arXiv:2409.13265.
  294. Wang, Y., Li, P., Sun, M., & Liu, Y. (2023). Self-Knowledge Guided Retrieval Augmentation for Large Language Models. arXiv preprint arXiv:2310.05002.
  295. Wang, Y., Lipka, N., Rossi, R. A., Siu, A., Zhang, R., & Derr, T. (2023). Knowledge graph prompting for multi-document question answering. arXiv preprint arXiv:2308.11730.
  296. Wang, Y., Liu, X., Chen, X., O’Brien, S.,Wu, J., & McAuley, J. (n.d.). Self-Updatable Large Language Models by Integrating Context into Model Parameters. In The thirteenth international conference on learning representations.
  297. Wang, Z., Bao, R., Wu, Y., Taylor, J., Xiao, C., Zheng, F., Jiang, W., Gao, S., & Zhang, Y. (2024, November). Unlocking Memorization in Large Language Models with Dynamic Soft Prompting. In Y. Al-Onaizan, M. Bansal, & Y.-N. Chen (Eds.), Proceedings of the 2024 conference on empirical methods in natural language processing (pp. 9782–9796). Miami, Florida, USA: Association for Computational Linguistics. Availableonline: https://aclanthology.org/2024.emnlp-main.546/ (accessed on). https://doi.org/10.18653/v1/2024.emnlp-main.546.
  298. Wang, Z., Liu, Z., Zhang, Y., Zhong, A., Fan, L., Wu, L., & Wen, Q. (2023). Rcagent: Cloud root cause analysis by autonomous agents with tool-augmented large language models. arXiv.
  299. Wang, Z. Z., Mao, J., Fried, D., & Neubig, G. (2024a). Agent workflow memory. Available online: https://arxiv.org/ abs/2409.07429 (accessed on).
  300. Wang, Z. Z., Mao, J., Fried, D., & Neubig, G. (2024b). Agent workflow memory. arXiv preprint arXiv:2409.07429.
  301. Wei, F., Tang, Z., Zeng, R., Liu, T., Zhang, C., Chu, X., & Han, B. (2025). JailbreakLoRA: Your Downloaded LoRA from Sharing Platforms might be Unsafe. In Icml 2025 workshop on data in generative models - the bad, the ugly, and the greats. Available online: https://openreview.net/forum?id=RjaeiNswGh (accessed on).
  302. Weng, L. (2023, Jun). Llm-powered autonomous agents. lilianweng.github.io.
  303. Wu, D., Wang, H., Yu, W., Zhang, Y., Chang, K.-W., & Yu, D. (2024). Longmemeval: Benchmarking chat assistants on long-term interactive memory. arXiv preprint arXiv:2410.10813.
  304. Wu, H., & Tu, K. (2024, August). Layer-Condensed KV Cache for Efficient Inference of Large Language Models. In L.-W. Ku, A. Martins, & V. Srikumar (Eds.), Proceedings of the 62nd annual meeting of the association for computational linguistics (volume 1: Long papers) (pp. 11175–11188). Bangkok, Thailand: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.acl-long.602/ (accessed on). https://doi.org/10.18653/v1/2024.acl-long.602.
  305. Wu, Q., Bansal, G., Zhang, J.,Wu, Y., Zhang, S., Zhu, E., Li, B., Jiang, L., Zhang, X., &Wang, C. (2023). Autogen: Enabling next-gen llm applications via multi-agent conversation framework. arXiv preprint arXiv:2308.08155.
  306. Wu,W., Pan, Z.,Wang, C., Chen, L., Bai, Y.,Wang, T., Fu, K.,Wang, Z., & Xiong, H. (2025). Tokenselect: Efficient long-context inference and length extrapolation for llms via dynamic token-level kv cache selection. Available online: https://arxiv.org/abs/2411.02886 (accessed on).
  307. Wu, Y., Liang, S., Zhang, C., Wang, Y., Zhang, Y., Guo, H., Tang, R., & Liu, Y. (2025). From human memory to ai memory: A survey on memory mechanisms in the era of llms. Available online: https://arxiv.org/abs/2504.15965 (accessed on).
  308. Wu, Y., Rabe, M. N., Hutchins, D., & Szegedy, C. (2022). Memorizing transformers. arXiv preprint arXiv:2203.08913.
  309. Wu, Y., Wang, H., Zhao, P., Zheng, Y., Wei, Y., & Huang, L.-K. (2024). Mitigating catastrophic forgetting in online continual learning by modeling previous task interrelations via pareto optimization. In Forty-first international conference on machine learning.
  310. Wulf, W. A., & McKee, S. A. (1995, March). Hitting the memory wall: implications of the obvious. SIGARCH Comput. Archit. News, 23(1), 20–24. Available online: https://doi.org/10.1145/216585.216588 (accessed on). https://doi.org/10.1145/216585.216588.
  311. Xi, Y., Liu, W., Lin, J., Chen, B., Tang, R., Zhang, W., & Yu, Y. (2024). Memocrs: Memory-enhanced sequential conversational recommender systems with large language models. In Proceedings of the 33rd acm international conference on information and knowledge management (pp. 2585–2595).
  312. Xi, Z., Chen, W., Guo, X., He, W., Ding, Y., Hong, B., Zhang, M., Wang, J., Jin, S., Zhou, E., & et al. (2023). The rise and potential of large language model based agents: A survey. arxiv.
  313. Xiao, G., Tian, Y., Chen, B., Han, S., & Lewis, M. (2023). Efficient Streaming Language Models with Attention Sinks. arXiv.
  314. Xiao, G., Tian, Y., Chen, B., Han, S., & Lewis, M. (2024a). Efficient Streaming Language Models with Attention Sinks. In The twelfth international conference on learning representations. Available online: https://openreview.net/ forum?id=NG7sS51zVF (accessed on).
  315. Xiao, G., Tian, Y., Chen, B., Han, S., & Lewis, M. (2024b). Efficient streaming language models with attention sinks. Available online: https://arxiv.org/abs/2309.17453 (accessed on).
  316. Xiong, H.,Wang, S., Zhu, Y., Zhao, Z., Liu, Y., Wang, Q., & Shen, D. (2023). Doctorglm: Fine-tuning your chinese doctor is not a herculean task. arXiv preprint arXiv:2304.01097.
  317. Xu, D., Chen, W., Peng, W., Zhang, C., Xu, T., Zhao, X., Wu, X., Zheng, Y., & Chen, E. (2023). Large language models for generative information extraction: A survey. arXiv.
  318. Xu, J., Szlam, A., & Weston, J. (2021). Beyond goldfish memory: Long-term open-domain conversation. arXiv preprint arXiv:2107.07567.
  319. Xu, W., Liang, Z., Mei, K., Gao, H., Tan, J., & Zhang, Y. (2025). A-mem: Agentic memory for llm agents. arXiv preprint arXiv:2502.12110.
  320. Yadav, P., Tam, D., Choshen, L., Raffel, C., & Bansal, M. (2023, 06). TIES-Merging: Resolving Interference When Merging Models. Available online: https://arxiv.org/pdf/2306.01708.pdf (accessed on).
  321. Yan, S.-Q., Gu, J.-C., Zhu, Y., & Ling, Z.-H. (2024). Corrective Retrieval Augmented Generation. arXiv preprint arXiv:2401.15884.
  322. Yang, D., Han, X., Gao, Y., Hu, Y., Zhang, S., & Zhao, H. (2024, August). PyramidInfer: Pyramid KV Cache Compression for High-throughput LLM Inference. In Findings of the association for computational linguistics: Acl 2024 (pp. 3258–3270). Bangkok, Thailand: Association for Computational Linguistics. Available online: https://aclanthology.org/2024.findings-acl.195/ (accessed on). https://doi.org/10.18653/v1/ 2024.findings-acl.195.
  323. Yang, E., Wang, Z., Shen, L., Liu, S., Guo, G., Wang, X., & Tao, D. (2024, 10). AdaMerging: Adaptive Model Merging for Multi-Task Learning. ICLR. Available online: https://arxiv.org/pdf/2310.02575.pdf (accessed on).
  324. Yang, H., Lin, Z.,Wang,W.,Wu, H., Li, Z., Tang, B.,Wei,W.,Wang, J., Tang, Z., Song, S., et al. (2024). Memory3: Language Modeling with Explicit Memory. arXiv preprint arXiv:2407.01178.
  325. Yang, H., Zhang, R., Huang, M.,Wang,W., Tang, Y., Li, Y., Liu, Y., & Zhang, D. (2025). Kvshare: An llm service system with efficient and effective multi-tenant kv cache reuse. Available online: https://arxiv.org/abs/2503.16525 (accessed on).
  326. Yang, J., Hou, B., Wei, W., Bao, Y., & Chang, S. (2025). Kvlink: Accelerating large language models via efficient kv cache reuse. Available online: https://arxiv.org/abs/2502.16002 (accessed on).
  327. Yang, J. Y., Kim, B., Bae, J., Kwon, B., Park, G., Yang, E., Kwon, S. J., & Lee, D. (2024). No token left behind: Reliable kv cache compression via importance-aware mixed precision quantization. Available online: https://arxiv.org/ abs/2402.18096 (accessed on).
  328. Yang, L., Yu, Z., Zhang, T., Cao, S., Xu, M., Zhang, W., Gonzalez, J. E., & Cui, B. (2024). Buffer of thoughts: Thought-augmented reasoning with large language models. arXiv preprint arXiv:2406.04271.
  329. Yang, S. (2023). Advanced rag 01: Small-to-big retrieval. https://towardsdatascience.com/advanced-rag-01-small-to -big-retrieval-172181b396d4.
  330. Yang, Z., Jia, X., Li, H., & Yan, J. (2023). A survey of large language models for autonomous driving. arXiv.
  331. Yao, J., Li, H., Liu, Y., Ray, S., Cheng, Y., Zhang, Q., Du, K., Lu, S., & Jiang, J. (2025). Cacheblend: Fast large language model serving for rag with cached knowledge fusion. Available online: https://arxiv.org/abs/2405.16444 (accessed on).
  332. Yao, S., Zhao, J., Yu, D., Du, N., Shafran, I., Narasimhan, K., & Cao, Y. (2022). React: Synergizing reasoning and acting in language models. arXiv preprint arXiv:2210.03629.
  333. Yao, W., Heinecke, S., Niebles, J. C., Liu, Z., Feng, Y., Xue, L., Murthy, R., Chen, Z., Zhang, J., Arpit, D., & et al. (2023). Retroformer: Retrospective large language agents with policy gradient optimization. arXiv preprint arXiv:2308.02151.
  334. Yao, Y., Li, Z., & Zhao, H. (2024). Sirllm: Streaming infinite retentive llm. Available online: https://arxiv.org/abs/ 2405.12528 (accessed on).
  335. Ye, L., Tao, Z., Huang, Y., & Li, Y. (2024). Chunkattention: Efficient self-attention with prefix-aware kv cache and two-phase partition. Available online: https://arxiv.org/abs/2402.15220 (accessed on).
  336. Yu, G.-I., Jeong, J. S., Kim, G.-W., Kim, S., & Chun, B.-G. (2022). Orca: A Distributed Serving System for Transformer-Based Generative Models. In 16th usenix symposium on operating systems design and implementation (osdi 22) (pp. 521–538). Carlsbad, CA: USENIX Association. Available online: https://www.usenix.org/ conference/osdi22/presentation/yu (accessed on).
  337. Yu, L., Yu, B., Yu, H., Huang, F., & Li, Y. (2024a). Language Models are Super Mario: Absorbing Abilities from Homologous Models as a Free Lunch. ICML.
  338. Yu, L., Yu, B., Yu, H., Huang, F., & Li, Y. (2024b, 11). Language Models are Super Mario: Absorbing Abilities from Homologous Models as a Free Lunch. ICML.
  339. Yu, W., Iter, D., Wang, S., Xu, Y., Ju, M., Sanyal, S., Zhu, C., Zeng, M., & Jiang, M. (2022). Generate rather than retrieve: Large language models are strong context generators. arXiv preprint arXiv:2209.10063.
  340. Yu, W., Zhang, H., Pan, X., Ma, K., Wang, H., & Yu, D. (2023). Chain-of-Note: Enhancing Robustness in Retrieval-Augmented Language Models. arXiv preprint arXiv:2311.09210.
  341. Yunxiang, L., Zihan, L., Kai, Z., Ruilong, D., & You, Z. (2023). Chatdoctor: A medical chat model fine-tuned on llama model using medical domain knowledge. arXiv preprint arXiv:2303.14070.
  342. Zeng, F., Gan,W.,Wang, Y., Liu, N., & Yu, P. S. (2023). Large language models for robotics: A survey. arXiv.
  343. Zeng, R., Fang, J., Liu, S., & Meng, Z. (2024). On the structural memory of llm agents. arXiv preprint arXiv:2412.15266.
  344. Zha, L., Zhou, J., Li, L., Wang, R., Huang, Q., Yang, S., Yuan, J., Su, C., Li, X., Su, A., et al. (2023). Tablegpt: Towards unifying tables, nature language and commands into one gpt. arXiv preprint arXiv:2307.08674.
  345. Zhang, K., Li, J., Li, G., Shi, X., & Jin, Z. (2024). Codeagent: Enhancing code generation with tool-integrated agent systems for real-world repo-level coding challenges. arXiv preprint arXiv:2401.07339.
  346. Zhang, P., Liu, Z., Xiao, S., Shao, N., Ye, Q., & Dou, Z. (2025). Long Context Compression with Activation Beacon. In The thirteenth international conference on learning representations. Available online: https://openreview.net/ forum?id=1eQT9OzfNQ (accessed on).
  347. Zhang, T., Yi, J., Xu, Z., & Shrivastava, A. (2024). KV cache is 1 bit per channel: Efficient large language model inference with coupled quantization. Advances in Neural Information Processing Systems, NeurIPS 2024, 37, 3304–3331.
  348. Zhang, Z., Bo, X., Ma, C., Li, R., Chen, X., Dai, Q., Zhu, J., Dong, Z., & Wen, J.-R. (2024). A survey on the memory mechanism of large language model based agents. arXiv preprint arXiv:2404.13501.
  349. Zhang, Z., Dai, Q., Bo, X., Ma, C., Li, R., Chen, X., Zhu, J., Dong, Z., & Wen, J.-R. (2025, July). A Survey on the Memory Mechanism of Large Language Model based Agents. ACM Trans. Inf. Syst.. Available online: https://doi.org/10.1145/3748302 (accessed on). (Just Accepted) https://doi.org/10.1145/3748302.
  350. Zhang, Z., Feng, Y., & Zhang, M. (2025). Levelrag: Enhancing retrieval-augmented generation with multi-hop logic planning over rewriting augmented searchers. Available online: https://arxiv.org/abs/2502.18139 (accessed on).
  351. Zhang, Z., Sheng, Y., Zhou, T., Chen, T., Zheng, L., Cai, R., Song, Z., Tian, Y., Re, C., Barrett, C., Wang, Z., & Chen, B. (2023a). H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models. In Thirty-seventh conference on neural information processing systems. Available online: https://openreview.net/ forum?id=RkRrPp7GKO (accessed on).
  352. Zhang, Z., Sheng, Y., Zhou, T., Chen, T., Zheng, L., Cai, R., Song, Z., Tian, Y., Ré, C., Barrett, C., et al. (2023b). H2o: Heavy-hitter oracle for efficient generative inference of large language models. Advances in Neural Information Processing Systems, 36, 34661–34710.
  353. Zhang, Z., Sheng, Y., Zhou, T., Chen, T., Zheng, L., Cai, R., Song, Z., Tian, Y., Ré, C., Barrett, C.,Wang, Z. A., & Chen, B. (2023c). H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, & S. Levine (Eds.), Advances in neural information processing systems (Vol. 36, pp. 34661–34710). Curran Associates, Inc. Available online: https://proceedings.neurips.cc/ paper_files/paper/2023/file/6ceefa7b15572587b78ecfcebb2827f8-Paper-Conference.pdf (accessed on).
  354. Zhao, A., Huang, D., Xu, Q., Lin, M., Liu, Y.-J., & Huang, G. (2023). Expel: Llm agents are experiential learners. arXiv preprint arXiv:2308.10144.
  355. Zhao, A., Huang, D., Xu, Q., Lin, M., Liu, Y.-J., & Huang, G. (2024). Expel: Llm agents are experiential learners. In Proceedings of the aaai conference on artificial intelligence (Vol. 38, pp. 19632–19642).
  356. Zhao, P., Jin, Z., & Cheng, N. (2023). An in-depth survey of large language model-based artificial intelligence agents. arXiv.
  357. Zhao, Y., Lin, C.-Y., Zhu, K., Ye, Z., Chen, L., Zheng, S., Ceze, L., Krishnamurthy, A., Chen, T., & Kasikci, B. (2024). Atom: Low-Bit Quantization for Efficient and Accurate LLM Serving. In Mlsys. Available online: https://proceedings.mlsys.org/paper_files/paper/2024/hash/5edb57c05c81d04beb716ef1d542fe9 e-Abstract-Conference.html (accessed on).
  358. Zhao, Z., Ma, D., Chen, L., Sun, L., Li, Z., Xu, H., Zhu, Z., Zhu, S., Fan, S., Shen, G., & et al. (2024). Chemdfm: Dialogue foundation model for chemistry. arXiv preprint arXiv:2401.14818.
  359. Zheng, H. S., Mishra, S., Chen, X., Cheng, H.-T., Chi, E. H., Le, Q. V., & Zhou, D. (2024). Take a step back: Evoking reasoning via abstraction in large language models. Available online: https://arxiv.org/abs/2310.06117 (accessed on).
  360. Zheng, L., Wang, R., Wang, X., & An, B. (2023). Synapse: Trajectory-as-exemplar prompting with memory for computer control. In Neurips 2023 foundation models for decision making workshop.
  361. Zheng, L.,Wang, R.,Wang, X., & An, B. (2024). Synapse: Trajectory-as-Exemplar Prompting with Memory for Computer Control. In Proceedings of the international conference on learning representations (iclr). Available online: https://arxiv.org/abs/2306.07863 (accessed on).
  362. Zheng, L., Yin, L., Xie, Z., Sun, C., Huang, J., Yu, C. H., Cao, S., Kozyrakis, C., Stoica, I., Gonzalez, J. E., Barrett, C., & Sheng, Y. (2024). Sglang: Efficient execution of structured language model programs. Available online: https://arxiv.org/abs/2312.07104 (accessed on).
  363. Zheng, Z., Ning, K., Wang, Y., Zhang, J., Zheng, D., Ye, M., & Chen, J. (2023). A survey of large language models for code: Evolution, benchmarking, and future trends. arXiv.
  364. Zhong, M., Liu, X., Zhang, C., Lei, Y., Gao, Y., Hu, Y., Chen, K., & Zhang, M. (2024). Zigzagkv: Dynamic kv cache compression for long-context modeling based on layer uncertainty. Available online: https://arxiv.org/abs/ 2412.09036 (accessed on).
  365. Zhong, W., Guo, L., Gao, Q., & Wang, Y. (2023). Memorybank: Enhancing large language models with long-term memory. arXiv preprint arXiv:2305.10250.
  366. Zhong,W., Guo, L., Gao, Q., Ye, H., &Wang, Y. (2024a). Memorybank: Enhancing large language models with long-term memory. In Proceedings of the aaai conference on artificial intelligence (Vol. 38, pp. 19724–19731).
  367. Zhong, W., Guo, L., Gao, Q., Ye, H., & Wang, Y. (2024b). MemoryBank: Enhancing Large Language Models with Long-Term Memory. In M. J.Wooldridge, J. G. Dy, & S. Natarajan (Eds.), Thirty-eighth AAAI conference on artificial intelligence, AAAI 2024, thirty-sixth conference on innovative applications of artificial intelligence, IAAI 2024, fourteenth symposium on educational advances in artificial intelligence, EAAI 2014, february 20-27, 2024, vancouver, canada (pp. 19724–19731). AAAI Press. Available online: https://doi.org/10.1609/aaai.v38i17.29946 (accessed on). https://doi.org/10.1609/AAAI.V38I17.29946.
  368. Zhong, Y., Liu, S., Chen, J., Hu, J., Zhu, Y., Liu, X., Jin, X., & Zhang, H. (2024). DistServe: disaggregating prefill and decoding for goodput-optimized large language model serving. In Proceedings of the 18th usenix conference on operating systems design and implementation. USA: USENIX Association.
  369. Zhou, D., Schärli, N., Hou, L., Wei, J., Scales, N., Wang, X., Schuurmans, D., Cui, C., Bousquet, O., Le, Q., & Chi, E. (2023). Least-to-most prompting enables complex reasoning in large language models. Available online: https://arxiv.org/abs/2205.10625 (accessed on).
  370. Zhou, H., Gu, B., Zou, X., Li, Y., Chen, S. S., Zhou, P., Liu, J., Hua, Y., Mao, C., Wu, X., & et al. (2023). A survey of large language models in medicine: Progress, application, and challenge. arXiv.
  371. Zhou, X., Wang, W., Zeng, M., Guo, J., Liu, X., Shen, L., Zhang, M., & Ding, L. (2025). Dynamickv: Task-aware adaptive kv cache compression for long context llms. Available online: https://arxiv.org/abs/2412.14838 (accessed on).
  372. Zhou, Z., Ning, X., Hong, K., Fu, T., Xu, J., Li, S., Lou, Y., Wang, L., Yuan, Z., Li, X., et al. (2024). A survey on efficient inference for large language models. arXiv preprint arXiv:2404.14294.
  373. Zhu, D., Wang, L., Yang, N., Song, Y., Wu, W., Wei, F., & Li, S. (2024). LongEmbed: Extending Embedding Models for Long Context Retrieval. In Y. Al-Onaizan, M. Bansal, & Y. Chen (Eds.), Proceedings of the 2024 conference on empirical methods in natural language processing, EMNLP 2024, miami, fl, usa, november 12-16, 2024 (pp. 802–816). Association for Computational Linguistics. Available online: https://aclanthology.org/2024.emnlp-main.47 (accessed on).
  374. Zhu, Q., Zhang, L., Xu, Q., Long, C., & Zhang, J. (2025). Subgcache: Accelerating graph-based rag with subgraph-level kv cache. Available online: https://arxiv.org/abs/2505.10951 (accessed on).
  375. Zhu, X., Chen, Y., Tian, H., Tao, C., Su, W., Yang, C., Huang, G., Li, B., Lu, L., Wang, X., & et al. (2023a). Ghost in the minecraft: Generally capable agents for open-world enviroments via large language models with text-based knowledge and memory. arXiv preprint arXiv:2305.17144.
  376. Zhu, X., Chen, Y., Tian, H., Tao, C., Su, W., Yang, C., Huang, G., Li, B., Lu, L., Wang, X., & et al. (2023b). Ghost in the minecraft: Generally capable agents for open-world environments via large language models with text-based knowledge and memory. arXiv preprint arXiv:2305.17144.
  377. Zhu, Y., Falahati, A., Yang, D. H., & Amiri, M. M. (2025). Sentencekv: Efficient llm inference via sentence-level semantic kv caching. Available online: https://arxiv.org/abs/2504.00970 (accessed on).
  378. Zhu, Y., Tang, Z., Liu, X., Li, A., Li, B., Chu, X., & Han, B. (2025). OracleKV: Oracle Guidance for Question-Independent KV Cache Compression. In Icml 2025 workshop on long-context foundation models. Available online: https://openreview.net/forum?id=KHM2YOGgX9 (accessed on).
  379. Zhu, Y., Yuan, H., Wang, S., Liu, J., Liu, W., Deng, C., Dou, Z., & Wen, J.-R. (2023a). Large language models for information retrieval: A survey. arXiv.
  380. Zhu, Y., Yuan, H., Wang, S., Liu, J., Liu, W., Deng, C., Dou, Z., & Wen, J. (2023b). Large Language Models for Information Retrieval: A Survey. CoRR, abs/2308.07107. Available online: https://doi.org/10.48550/arXiv.2308.07107 (accessed on). https://doi.org/10.48550/ARXIV.2308.07107.
  381. Zou, A., Phan, L., Chen, S., Campbell, J., Guo, P., Ren, R., Pan, A., Yin, X., Mazeika, M., Dombrowski, A.-K., & et al. (2023). Representation engineering: A top-down approach to ai transparency. arXiv.
*
For LLMs, we do not discuss sensory memory in detail, as LLMs primarily operate on text.
Figure 1. Overview of LLM memory applications, representation, and unified management abstraction.
Figure 1. Overview of LLM memory applications, representation, and unified management abstraction.
Preprints 201223 g001
Table 1. Decision-support matrix for LLM memory management. Knowledge requirements summarize the task regimes each representation most naturally supports. Management challenge indicates where engineering effort is typically concentrated under construction-update-query.
Table 1. Decision-support matrix for LLM memory management. Knowledge requirements summarize the task regimes each representation most naturally supports. Management challenge indicates where engineering effort is typically concentrated under construction-update-query.
Memory Preferred Knowledge Requirements Management Challenge Strategic Gain
Representation Retention Functional Construction Update Query (Return)
Token-level Short-term Episodic + Semantic Medium Medium High Editability
Intermediate latent Short-term Episodic Low High Low Efficiency
Parameter-level Long-term Procedural + Semantic High High Low Persistence
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated