Skill Memory Platform for AI Agents

Store, Observe, Learn

Think of it as Supabase for agent context — with skill-based memory your agents build and humans can actually read.

Scale from local demos to production without rebuilding context infrastructure — skill-based memory that's filesystem-compatible, configurable, and human-readable.

How It Works

The capabilities that power production AI agents — store context, observe behavior, learn from experience, and monitor everything.

Session
Disk
OpenAI

Platform Capabilities

The production-grade infrastructure your agents need — storage, observability, skill memory, and more.

Context Engineering

Edit, compress, and summarize context on-the-fly — token_limit, middle_out, and session summary strategies keep your agents efficient without modifying stored messages.

Multimodal Context Storage

Unified, persistent storage for all agent data — messages, files, and skills — eliminating fragmented backends (DB, S3, Redis).

Artifact Disk

Filesystem-like workspace to store and share multi-modal outputs (.md, code, reports), ready for multi-agent collaboration.

Background Observer

Automatically extracts tasks from agent conversations and tracks their status in real-time — from pending to running to success or failure.

Self-Learning

Attach sessions to a Learning Space and Acontext automatically distills successful task outcomes into skills — agents improve with every run without manual curation.

SDKs & Integrations

Ready to use with OpenAI, Anthropic, LangGraph, Agno, and other popular agent frameworks.

Get Started in Minutes

pip install acontext — that's all you need.

store.py
from acontext import AcontextClient

client = AcontextClient()

# Create a session
session = client.sessions.create()

# Store messages (any provider format)
client.sessions.store_message(
    session.id,
    blob={"role": "user", "content": "Hello!"}
)

# Retrieve in any format: openai, anthropic, gemini
messages = client.sessions.get_messages(
    session.id, format="anthropic"
)

Store and retrieve agent messages in any LLM format — OpenAI, Anthropic, Gemini.

Docs
Open-source, one command to host

Self-Host in Seconds

Run the full Acontext stack on your own infrastructure with a single command.

curl -fsSL https://install.acontext.io | shCopied!

Join the Community

Connect with early builders & preview new features