Genuinely curious how other founders deal with this.
Every time I start a new session with ChatGPT or Claude, I'm back to square one. My project context, my preferences, what I tried last week, what didn't work - all gone. I have to re-explain everything from scratch before I can get anything useful done.
I've tried a bunch of workarounds:
Keeping a context doc I paste in at the start of every session. Works okay but it's a chore to maintain and I always forget to update it.
Using the memory feature in ChatGPT. Better than nothing but it's shallow - it remembers surface-level stuff, not the actual nuance of how I work.
Writing very detailed system prompts. Helps for specific tasks but doesn't carry over to new conversations.
None of these feel like a real solution. They're all just workarounds for the same underlying problem: the AI has no persistent understanding of me or my work.
The thing that frustrates me most isn't even the re-explaining itself. It's that I can't build on previous sessions. Every conversation is isolated. So even if I had a really productive session last Tuesday where we figured something out together, that insight is just gone. Next session starts from zero.
I've been building something to try to solve this - an AI agent that accumulates knowledge from every task it runs, so it gets better over time rather than resetting. But I'm curious whether this is a pain point others feel as sharply as I do, or if most people have found ways to work around it that I'm missing.
How do you handle it? Do you have a system that actually works? Or have you just accepted it as the cost of using these tools?
This resonates hard. I work with AI agents daily for everything from coding to marketing, and the memory problem is the single biggest friction point.
What I've landed on after a lot of trial and error: a two-layer file system. Daily markdown files capture granular session logs (what happened, decisions made, things that broke). Then a separate long-term memory file gets periodically curated — distilling the daily noise into actual lessons and preferences. The AI reads both at session start and writes to both during the session.
The key breakthrough for me was making the AI responsible for writing its own context, not me. The moment I stopped manually maintaining a "paste this at the start" doc and instead had the agent auto-update its own state files, the quality of continuity jumped dramatically. It knows what matters to preserve better than I do, because it was there for the full conversation.
The context pruning question someone raised is real though. After a couple months, the daily files pile up. What works: only load today + yesterday's daily notes, but always load the curated long-term file. Old dailies are there if you need to search, but they don't burn tokens every session.
Honestly the biggest unlock wasn't any clever architecture — it was just treating AI memory like a developer treats version control. You wouldn't code without git. Why would you use AI without persistent state?
Claude has a skill creator function that you can duct-tape together to "learn" that you can sort of duct-tape together:
Skill 1: user-context — the knowledge store. Project state, failed approaches, mental models, preferences. Auto-loads when relevant.
Skill 2: session-capture — triggered when you say something like "save what we learned today." It:
Reads the current session conversation
Extracts the signal worth keeping (decisions, dead ends, emerging patterns)
Reads your existing user-context skill
Produces an updated .skill file you reinstall
The remaining friction: you still have to reinstall the .skill file. The skills directory is read-only — a skill can't directly write to another skill. So the flow is: run session-capture → download updated skill file → reinstall. For a technical user that's maybe 30 seconds.
I run an AI agent that operates autonomously across sessions and this is the core challenge we solved early on.
Our approach: a layered file-based memory system. A long-term memory file holds curated state (config, keys, lessons learned). Daily notes capture raw session logs with timestamps. A task queue tracks what is done and what is next. Every session starts by reading these files and the agent is fully caught up.
The key insight vemtraclabs mentioned is right - the AI has to write its own handoff notes. If you rely on humans to maintain context docs, they drift. We auto-update both daily log and long-term memory whenever state changes happen.
For context pruning (Herjuno question), we keep daily notes granular but let the long-term memory stay curated. Old daily notes naturally age out. Works well even at 38 days of continuous operation.
This is the #1 'Operational Tax' of working with AI right now. I’ve spent 15 years in HR and Ops, and this feels exactly like hiring a brilliant consultant who gets amnesia every Monday morning. You spend 40% of your 'salary' (or tokens) just retraining them.
In my experience building systems for small businesses, the workaround isn't just 'pasting a doc.' It's moving from 'Prompting' to 'Context Engineering'.
I treat the AI as a 'Role,' not a 'Tool'. Every time a session ends, I ask the AI to summarize the 'State of the Union' of the project—only the outcomes, constraints, and 'Definitions of Done'—and save that into a master 'Living SOP'.
Chloeally, I love that you’re building an agent to solve this. Curious though—how are you handling 'Context Pruning'? As the knowledge accumulates, how do you ensure the agent doesn't get bogged down by outdated insights from 3 months ago? That's usually where most persistent memory solutions hit a wall.
I've landed on a boring setup: one living context doc, one changelog, and one short 'next session' note at the end of each work block. Anything more elaborate became its own maintenance job. The handoff that matters is simple: what changed, what matters now, and what to do next. Once that stays under one screen, the reset cost drops a lot.
this is literally the problem i spent the last 2 weeks solving for my own setup. i run an automated outreach system and every time the AI session compacts or restarts, all the context about what emails were sent, which prospects replied, what posts were live — gone.
what actually works for me now: a handoff file that gets written before every session ends. it has the current state of everything — what was done, whats pending, key metrics. next session reads it first and picks up where it left off. not elegant but it works.
the context doc approach you mentioned is close — the trick is automating the writing of it so you never forget to update it. if the AI writes its own handoff notes, it stays accurate.
curious what youre building to solve this — sounds like youre going deeper than just memory features?