Compatible with everything you already use
Three architectural failures every team hits once they go past a demo. Threadline eliminates all three.
Stuffing every chat into the prompt burns tokens and degrades quality after ~8 turns.
Threadline distills sessions into compact, recallable facts.
Vector search returns yesterday's answer, not what the user said five minutes ago.
Real-time memory updates with versioning and recency scoring.
Switch models, switch products, restart the app — your agent forgets the user.
One context object, portable across every LLM and surface.
~38ms retrieval means zero-latency injection before every LLM call — your agent never waits.
Your Agent
user interaction
Threadline
threadline.to
Your Product
already knows
One context object · every agent · no repeated conversation
1. inject() enriches your prompt → 2. LLM call with context → 3. update() captures new facts
Choose your integration method and start giving your agents memory in under a minute.
01npm install threadline-sdk02 03import { Threadline } from "threadline-sdk"04 05const tl = new Threadline({ apiKey: "tl_live_..." })06 07const { injectedPrompt } = await tl.inject(userId, basePrompt)08 09await tl.update({ userId, userMessage, agentResponse })Seven scopes of context, automatically extracted and maintained.
communication_style
How they like to be spoken to
e.g. "Prefers concise, bullet‑point answers"
ongoing_tasks
What they're currently working on
e.g. "Building a Stripe payments API, deadline Q2"
key_relationships
People and roles they mention
e.g. "Co‑founder Sarah handles design"
domain_expertise
Their technical background
e.g. "5 years TypeScript, Next.js, Supabase"
preferences
How they want the agent to behave
e.g. "No code examples unless asked"
emotional_state
Current mood and urgency signals
e.g. "High urgency, deadline pressure this week"
general
Core identity and context
e.g. "Vidur, Delhi, pre‑revenue founder"
Built for the world where users demand control over their AI data.
OAuth‑style grants
Users approve what each agent can see
Hard delete
Users can permanently erase their context
Full audit trail
Every read and write is logged
Context Dashboard
Manage what agents can access
communication_style
Granted
ongoing_tasks
Granted
emotional_state
Revoked
domain_expertise
Granted
| Feature | Threadline | Mem0 | Supermemory | Zep | Letta |
|---|---|---|---|---|---|
| Persistent memory | |||||
| Works with any LLM | |||||
| MCP compatible | |||||
| User‑owned context | |||||
| OAuth‑style grant system | |||||
| Scoped agent access | |||||
| Idempotent grants / scope expansion | |||||
| Full audit trail | |||||
| Hard delete by user | |||||
| Retrieval latency | ~38ms* | ~200ms | <300ms | ~300ms | N/A |
| Free tier | 10K mem/mo | 1K/mo | Limited | None | None |
* Threadline's inject() is a direct database lookup, not vector search. Intelligence happens at update() time, not retrieval time — which is why retrieval is this fast. Competitors perform full semantic search at retrieval time.
Free up to 10,000 memories a month — no credit card. Builder and Scale tiers when your agent grows up.
See full pricingFree to start. No credit card. Works with everything you already use.
Get API Key →