Amnesiac.
Every session starts at zero. Yesterday's two hours of context — gone.
THE AI OS FOR FOUNDERS · BUILT AT FINCH · USED IN PRODUCTION DAILY
Finch Engine is the operating layer under Claude Code. It makes the agent follow protocols, remember across sessions, and survive context overflow — the same engine my four-person team runs an entire claims-intelligence platform on every day.
Built by Finch. We use it every day. We eat this for breakfast.
Every session starts at zero. Yesterday's two hours of context — gone.
Says it wrote tests. Tests were inadequate. Docs untouched.
Subagents die mid-task. Reports vanish. Three agents step on each other.
By the end of a long session, it's forgotten the original goal.
Finch Engine fixes all four.
Run the same task in a vanilla Claude Code session, then through Finch Engine. Each failure mode above — caught and corrected by the protocol layer, not the agent.
$ continue what we were working on > I don't have prior context. Could you > share what we were working on?✗ context lost · two hours gone
$ engine session rehydrate \
sessions/2026_05_14_AUTH_AUDIT
✓ rehydrated · 2,847 lines
✓ 6 decisions · 3 open threads✓ session restored · resume verbatim> I've written the tests.
$ ls test/
ls: no such file or directory✗ claimed without proof▸ phase 4 · proof required
✗ gate blocked · tests-written
required artifact: test/*.spec.ts
advance denied · re-do phase 3✓ phase gate refused to advance▸ delegating to refactor-agent… [agent terminated · output truncated] > what did the refactor say? I don't have that information.✗ delegation black-holed
#needs-refactor → #claimed-refactor → #done-refactor · pane 3 ↘ handoff logged · 247 lines ↘ patch ready for review✓ tags caught the thread
[conversation compacted · context lost] > what were we trying to ship? > I'm not sure — could you remind me?✗ overflow ate the goal
▸ /dehydrate · checkpoint saved ✓ goal pinned · 4 decisions kept ▸ /rehydrate · context restored ✓ resume · no degradation✓ overflow handled · goal kept
Same model. Same prompts. The engine is the difference.
/analyze run on real code. No signup.Paste a snippet — or pick a sample. We run /analyze server-side and stream the engine's output here. Phase by phase. Same protocol your trial runs.
Six core abstractions. Bash and markdown under the hood. No framework to learn, no new dependencies, no runtime to install beyond the engine itself.
Stateful conversation with phases + a persistent log. The session is the source of truth, not the chat.
Phase-gated workflows with required proof fields. The agent can't skip steps because the tool layer enforces it.
Every artifact carries semantic state — #needs-fix, #claimed-fix, #done-fix. The tags are the state machine.
Stacked rules — core, project, directory. The agent reads them automatically when they apply.
Skills hand off to skills. Async, blocking, or silent. You don't lose threads.
Bash hooks fire on every tool call. This is how “discipline” actually works.
It reviews the plan, not the PR. The protocol catches bugs before the bug-author has written them.
One session in flight. Phases on the rail. Proof gates lighting up only after their conditions are met. Tags moving through the state machine at the bottom. This is what the six abstractions look like when they're alive.
Fleet runs Engine sessions in parallel — each in its own pane, each with its own skill, each surviving its own context. You watch four problems get solved at once, without losing the thread on any of them. tmux for agents, with the protocol layer keeping score.
[11:40] PHASE 3 · calibration [11:41] ✓ findings-deduped (12 → 8) [11:42] ▸ severity-bucket-sweep [11:42] SEC-001 · MUST FIX [11:43] SEC-002 · SHOULD FIX
[11:38] PHASE 2 · gemini deep research [11:40] query refined · 4 sub-questions [11:42] ▸ awaiting 19 sources [11:43] 3 of 19 returned [11:43] citations · pinned
[11:39] PHASE 4 · proof [11:41] generated · auth.test.ts [11:42] ran · 27 tests [11:42] ✓ 27 passing · 0 failing [11:43] hook · post-tool-use · log-sync
[11:43] PHASE 1 · load prior session [11:43] ✓ session memory hydrated [11:44] ▸ diff loaded · 247 lines [11:44] ◌ panel of reviewers · queued [11:44] 10 reviewers · 2 models
One terminal. Four contexts. Zero collisions. The orchestrator routes the work; the engine enforces it.
Every skill is a multi-phase protocol with proof-gated execution — dogfooded inside Finch for a week before it ships to subscribers. Pro subscribers vote on what gets built next. Each tile has a 60-second screencast so you see the protocol before you buy it.
Ten reviewers, two AI models, on every plan before a line ships.
Structured ideation that ends in a decision — and a record of what you rejected.
Gemini Deep Research with a refined-brief protocol. Comes back with citations.
PR review that loads the original implementation session's memory first.
Browser-based E2E QA via Playwright. Real navigation, real assertions.
TDD for LLM prompts and schemas. Iterate against measurements, not vibes.
Today's standup transcript → a Slack-ready summary in ninety seconds.
Monthly update synthesized across Linear, Slack, standups, and your codebase.
Sales call transcript → Notion post-call notes with action items.
Pre-call brief generated from your CRM and live web research.
Scan your industry for trigger events worth a sales conversation.
Deep account research from a URL or CSV → personalized outreach.
Strip AI tells from drafts. No more em-dashes, “elevated” anything.
Push micro-learnings to your team's shared library.
Submit what you need. Top-voted skill ships within two weeks. You decide what drops next.
Pulled from our shared drive this morning — sessions logged by every person at Finch, not just the engineers. Five months. One engine. The number that matters is who's running it.
It's not a coding tool. It's how the whole company operates.
Everything in Finch Engine. The full skills marketplace. Fleet. Cloud sync. Dashboard.
Founders Edition: lock this price for life.
Everything in Individual, plus team accounts, shared session view, and admin controls.
Talk to us about enterprise plans.
Annual: 20% off. Cancel anytime via Stripe Customer Portal. We don't read your logs.
If you're using Claude Code, Cursor, or any agent today — you've felt all four failure modes.