We shipped an implementation of Stanford's "Agentic Context Engineering" paper: agents that improve by learning from their own execution.
How does it work? A three-agent system (Generator, Reflector, Curator) builds a "playbook" of strategies autonomously:
My open-source implementation works with any LLM, has LangChain/LlamaIndex/CrewAI integrations, and can be plugged into existing agents in ~10 lines of code.
Would love feedback!