We open-sourced Stanford's "Agentic Context Engineering" implementation - agents that learn from execution

We shipped an implementation of Stanford's "Agentic Context Engineering" paper: agents that improve by learning from their own execution.

How does it work? A three-agent system (Generator, Reflector, Curator) builds a "playbook" of strategies autonomously:

My open-source implementation works with any LLM, has LangChain/LlamaIndex/CrewAI integrations, and can be plugged into existing agents in ~10 lines of code.

Would love feedback!