AI Memory

Researchers Apply Unix Philosophy to AI Memory Problems: 'Everything Is Context

A new open-source framework treats AI context as files, enabling version control and audit trails for LLM reasoning.

Oliver Senti
Oliver SentiSenior AI Editor
December 10, 20254 min read
Share:
Stylized technical illustration of a file-system tree structure representing AI context management, with color-coded directories for memory, tools, and knowledge bases

A team of researchers led by Xiwei Xu, Xuewu Gu, and Liming Zhu has released a paper proposing that AI systems manage their context the way Unix manages files. The work, published on arXiv this week, introduces a file-system abstraction for what the authors call "context engineering" and implements it in the open-source AIGNE framework.

The Problem: Context Rot

Large language models face a structural limitation that compounds as systems grow more complex. The models themselves are stateless: when a conversation ends, everything the AI learned during that exchange disappears. External memory systems exist, but they remain fragmented across different tools, producing what the researchers describe as "transient artefacts that limit traceability and accountability."

The paper identifies a specific failure mode called "context rot," where AI systems gradually lose coherence as memory entries accumulate, contradict each other, or become stale. Token windows (the amount of text an LLM can process at once) create hard ceilings: GPT-5's limit is 128,000 tokens, while Claude Sonnet 4.5 handles 200,000. When context exceeds these limits, something must be cut. Current systems make those cuts without systematic governance.

The Proposal: Mount Context Like a File System

The researchers draw their central insight from the Unix operating system, which treats keyboards, disk drives, network connections, and actual files through a single abstraction: "everything is a file." The paper applies this principle to AI context.

Under the proposed architecture, different context sources (conversation history, tool definitions, user profiles, external knowledge bases) are "mounted" into a unified namespace. An AI agent accesses a GitHub repository the same way it accesses a memory store or a user preference file, through standard read, write, and search operations on paths like /context/memory/ or /modules/github-mcp/.

The file-system metaphor brings specific engineering benefits. Context elements gain metadata (timestamps, provenance, version IDs). Changes become auditable through transaction logs. Access controls can restrict which agents see which context. Different storage backends (vector databases, SQL, knowledge graphs) slot in as plugins without requiring changes to the agents that use them.

The Pipeline: Constructor, Updater, Evaluator

Built on top of the file-system layer, the researchers describe a three-stage pipeline for managing context throughout an AI reasoning session.

The Context Constructor selects and compresses relevant information from the persistent repository to fit within token constraints. It generates a "manifest" documenting what was included, what was excluded, and why. This addresses a common opacity problem: when an AI fails to recall something, operators typically cannot determine whether the information was never stored, was stored but not retrieved, or was retrieved but compressed away.

The Context Updater handles the transition from storage to active reasoning. For short tasks, it loads a static snapshot. For ongoing conversations, it streams context incrementally, replacing stale fragments as the dialogue progresses.

The Context Evaluator closes the loop by validating outputs against source context, detecting contradictions or hallucinations, and writing verified information back to persistent storage. When confidence drops below threshold, it flags content for human review rather than committing it automatically.

Implementation: The AIGNE Framework

The architecture is implemented in AIGNE (available on GitHub at github.com/AIGNE-io/aigne-framework), a TypeScript/JavaScript framework for building AI agents. The framework integrates with OpenAI, Google Gemini, Anthropic Claude, DeepSeek, and Ollama models.

The paper provides two working examples. The first shows an agent with persistent memory stored in SQLite, where conversation history survives across sessions without manual state management. The second demonstrates mounting a Model Context Protocol (MCP) server, in this case GitHub's official MCP implementation, as a file-system module. The agent interacts with GitHub repositories through the same afs_read and afs_exec operations it uses for local files.

What This Enables

The file-system approach makes AI context compatible with existing DevOps practices. Context can be versioned, reviewed in pull requests, and deployed through CI/CD pipelines. The authors note that this moves context management "from ad hoc prompt management" to systematic engineering.

For multi-agent systems, the architecture provides resource isolation. Different agents can mount different views of shared context, with access controls preventing cross-contamination. The unified namespace also allows agents to discover available tools and data sources dynamically rather than having tool definitions hard-coded into prompts.

The research team plans to extend the work with "agentic navigation," allowing AI systems to autonomously browse and organize their own context hierarchies. They also intend to strengthen what they call "human-AI co-work," positioning humans not just as supervisors who correct errors but as active participants who contribute and curate knowledge alongside AI agents.

The AIGNE framework is available under open-source license. The full paper, "Everything is Context: Agentic File System Abstraction for Context Engineering," is accessible on arXiv as paper 2512.05470.

Tags:context engineeringAIGNELLM architectureAI memoryopen sourceUnix philosophyModel Context Protocolagentic AIsoftware engineering
Oliver Senti

Oliver Senti

Senior AI Editor

Former software engineer turned tech writer, Oliver has spent the last five years tracking the AI landscape. He brings a practitioner's eye to the hype cycles and genuine innovations defining the field, helping readers separate signal from noise.

Related Articles

Stay Ahead of the AI Curve

Get the latest AI news, reviews, and deals delivered straight to your inbox. Join 100,000+ AI enthusiasts.

By subscribing, you agree to our Privacy Policy. Unsubscribe anytime.

AIGNE Framework Treats AI Context Like Unix Files | aiHola