Every AI conversation starts from zero. You explain your company, your project, your constraints, and your preferences. The AI gives you a great answer. Then you close the tab and start over tomorrow.
The industry has produced a surprising number of approaches to this problem, ranging from a simple text box to full graph databases. But most people only know about one or two of them, and they rarely know which approach actually fits their situation.
This guide walks through every major option for giving AI persistent memory, organized from simplest to most capable. Each level solves a different problem, and the right choice depends on how seriously you need AI to understand your work.
Level 1: Custom instructions
This is where most people start, and where most people stay.
ChatGPT Custom Instructions give you roughly 1,500 characters per box, with about 4,500 characters total across the three sections (what you want ChatGPT to know, how you want it to respond, and additional context). This text loads into every conversation automatically. You write it once, and it's always there.
Claude system prompts and CLAUDE.md files serve the same purpose in the Claude ecosystem. CLAUDE.md files in Claude Code have no hard character limit, but they count against the context window, and auto-memory loads only the first 200 lines. In claude.ai, project instructions function similarly.
What's good about this. Zero setup, always loaded, no tools to install. If you just need Claude to know your name, your role, and that you prefer concise responses, custom instructions handle it.
What breaks. The capacity is tiny. 4,500 characters is roughly 700 words. That's enough for "I'm a product manager at a healthcare company, please be concise" and not much else. There's no room for your product roadmap, your technical architecture, your team structure, your style guide, or the decisions from last week's meeting. The instructions are also completely static. You update them manually, and they don't evolve with your work.
Best for: basic personal preferences like tone, role, language, and output format.
Level 2: Built-in memory
Both major AI platforms now offer automatic memory systems that try to remember facts across conversations without you managing anything.
ChatGPT Saved Memories stores roughly 1,500 to 1,750 words of remembered facts. The system automatically extracts things you mention ("I work at Acme Corp," "I prefer TypeScript over JavaScript") and retains them across all conversations. In April 2025, OpenAI added Reference Chat History, which loosely pulls context from past conversations even when nothing was explicitly saved to memory.
Claude Memory launched in October 2025 for Pro and Max subscribers. It generates an auto-synthesized memory summary that updates roughly every 24 hours. The summary is organized into categories like work, projects, and preferences. It can be scoped per project, and you can view and edit it directly.
What's good about this. It's automatic. You don't have to maintain anything. Over time, the AI builds up a rough picture of who you are and what you work on.
What breaks. The capacity is still small (under 2,000 words for both platforms). The memories are lossy: they're summaries, not verbatim records. The AI decides what to keep and what to discard, and its priorities may not match yours. It might remember that you "work in healthcare tech" while forgetting the specific compliance deadline that makes half its suggestions infeasible.
Reliability is also a concern. OpenAI experienced at least two memory wipes in 2025, where users lost months of accumulated context with no way to recover it. When your persistent memory can vanish overnight, it's not truly persistent.
Best for: personal preferences that don't require precision. Good as a supplement, risky as a foundation.
Level 3: Projects
Both platforms offer project containers that let you upload reference files and set project-specific instructions.
ChatGPT Projects let you attach files (5 on free, 25 on Plus, 40 on Pro) and write custom instructions per project. Files can be up to 512MB each. The main issue is context drift: in long conversations (beyond roughly 30 messages), earlier context fades. The AI may contradict what you established at the beginning or ignore constraints you explained. The workaround is to keep conversations short, which means re-establishing context each time.
Claude Projects handle large document sets differently. For small uploads, Claude loads files in full. For larger collections, it switches to RAG mode, using semantic search to pull relevant sections with roughly 10x effective capacity. Claude Projects also have their own memory space, separate from global memory.
Both platforms silo projects. Your product roadmap in "Strategy" is invisible to "Engineering" conversations. Cross-project knowledge requires file duplication.
What's good about this. Significantly more capacity than custom instructions. File-based context is richer and more structured.
What breaks. Files are processed to fit the context window, meaning the AI may summarize or selectively extract. Projects are siloed. Conversation history still resets when you start a new chat. The files persist; the discussion and decisions do not.
Best for: moderate document-based context for specific, well-defined workflows.
Level 4: External knowledge bases via MCP
This is where the architecture fundamentally changes. Instead of storing knowledge inside an AI platform, you store it outside, in documents that any AI can access on demand.
The connection happens through MCP (Model Context Protocol), an open standard that lets AI tools read from and write to external services. Claude adopted MCP natively, OpenAI followed in March 2025, Google DeepMind in April, and Microsoft in May. The ecosystem now includes over 500 official MCP servers and 16,000+ community-built ones.
With an external knowledge base, the AI reads specific documents when it needs them. A 100-document library works just as well as a 5-document one, because the AI only loads what's relevant. There's no summarization, no context window pressure, and no degradation. Your product roadmap reads exactly as you wrote it, word for word.
The key difference from Projects: the AI can also update your documents. After a meeting, ask Claude to update the project status with the decisions you made. The next conversation sees the current version automatically. Documents evolve with your work instead of staying frozen at upload time.
Unmarkdown™ provides this layer with a built-in MCP server. Documents are written in markdown, the native language of AI tools. The same documents can be published as web pages, formatted for Google Docs or Word, converted for Slack or email, and styled with 62 templates. Your knowledge base doubles as a document publishing platform.
What's good about this. Unlimited capacity, full document fidelity, cross-session persistence, AI can update documents, works across multiple AI clients, documents useful beyond AI context.
What breaks. Requires an MCP-compatible client. Setup takes a few minutes longer than uploading a file.
Best for: organizational knowledge, evolving documents, cross-session persistence, anyone who needs AI context beyond a single platform.
Level 5: Developer memory frameworks
For engineering teams building AI-powered products, a category of specialized memory frameworks has emerged. These are not end-user tools. They're SDKs and infrastructure for adding persistent memory to custom AI applications.
Mem0 (Y Combinator-backed) automatically extracts and manages memories from conversations, reporting a 26% accuracy boost and 90% token savings versus full conversation replay. Mem0-g adds graph-based memory for relationship tracking.
Letta (formerly MemGPT) lets agents manage their own memory through function calling. The AI agent has explicit tools for reading, writing, and organizing its memory blocks, giving it autonomy over its own context.
Zep builds temporal knowledge graphs using its Graphiti engine, organizing memory into episodes, semantic entities, and community subgraphs. A bi-temporal model tracks both when events happened and when they were recorded. Zep scored 94.8% on the DMR benchmark.
LangMem is LangChain's memory SDK supporting semantic, episodic, and procedural memory types. It integrates with LangGraph for stateful agent workflows.
What's good about this. Sophisticated memory management, production-grade reliability, highly customizable. Zep captures temporal relationships that no other approach handles.
What breaks. Developer-only. Requires engineering resources to implement, infrastructure to host, and ongoing maintenance. The learning curve is steep.
Best for: engineering teams building AI products that need custom memory capabilities.
Comparison table
| Approach | Capacity | Persistence | Accuracy | Setup | Best for |
|---|---|---|---|---|---|
| Custom Instructions | ~700-1,500 words | Permanent (manual) | Exact (what you wrote) | None | Basic preferences |
| Built-in Memory | ~1,500-1,750 words | Unreliable (wipes possible) | Lossy (summaries) | None | Personal preferences |
| Projects | 5-40 files | Files persist, chats don't | Processed (may summarize) | Low | Moderate document context |
| External KB (MCP) | Unlimited | Full persistence | Exact (full documents) | Medium | Organizational knowledge |
| Developer Frameworks | Unlimited | Full persistence | Configurable | High | Custom AI products |
Where most people are, and where they should be
Most people are stuck at Level 1 or 2. They re-explain their work context in every serious conversation and accept this friction as normal because they don't know the alternatives exist.
Level 3 (Projects) is a meaningful upgrade, but limited by platform silos and context drift. Level 5 is powerful but accessible only to engineering teams.
Level 4, external knowledge bases via MCP, hits the balance point. It offers the capacity and persistence of developer frameworks without requiring any code. Setup takes minutes. Documents are useful beyond AI context: they can be published, shared, and formatted for any destination. And because MCP is an open standard supported by every major AI platform, you're not locked into a single provider.
If you spend more than five minutes per week re-explaining context to an AI, you've already spent more time than it takes to set up an external knowledge base. The integrations overview covers setup for all MCP-compatible clients. The Claude-specific guide covers all three connection methods with exact configuration.
Getting started
The progression most people follow:
-
Start with custom instructions. Write your role, your preferences, and your most important rules. This takes five minutes and immediately improves every conversation.
-
Try Projects for focused work. Upload key reference documents and see how much better the AI's responses become with real context.
-
Move to an external knowledge base when you hit the ceiling. When you need cross-project knowledge, AI-updatable documents, or context drift is costing you time. Create a few documents in Unmarkdown™, connect via MCP, and experience the difference between an AI starting from scratch and one that already knows your world.
Each level builds on the last. You don't have to abandon custom instructions when you add a knowledge base. They complement each other: instructions for behavioral rules, knowledge base for reference material. The AI gets both automatically in every conversation.
For the developer API and MCP tool reference, see the developer documentation.
Related reading
- Claude Projects vs ChatGPT Projects vs External Knowledge Bases
- Why Your AI Keeps Forgetting (And the Permanent Fix)
- How to Build a Persistent Knowledge Base Your AI Can Always Access
- Obsidian Plugins You Don't Need: Features That Should Be Built-In
- The Document Your AI Never Forgets: Persistent Knowledge with MCP
