Build a queryable, AI-readable reference library.
A system for turning your books, papers, and notes into long-term semantic memory for agentic AI sessions. Bring your own corpus; the methodology + tools handle the rest. Extract once, consume many times.
Read the philosophy first.
"A 26,000-line book → ~280-line distillation. 100× compression at the load-bearing material."
Read METHODOLOGY.md to understand the three-layer extraction-and-synthesis pipeline. Then library-structure.md for the directory layout. Then run load_context.py against a small test library to see the session-quick-start output.
Documentation — the methodology in six focused docs
Each doc explains one piece of the system. Read in order for adoption; skim by topic for reference.
Python Tools — eight scripts for building and querying
All driven by REFERENCE_LIBRARY_ROOT env var or --library flag. Tools live here; library lives wherever you keep it.
--clip.Templates — schema starting points
Copy these into your library and fill in your content. The schemas are the load-bearing convention; the body is yours.
Examples — see what the artifacts actually look like
Three fully-anonymized examples using real foundational books (Pragmatic Programmer, Clean Code, Effective Java) so you can see the format in action.
Send a quick note.
Adopting this methodology yourself? Hit a problem with the tools? Have a war story? This form goes straight to the maintainer.
If you have a GitHub account, opening an issue is preferred. This form is the path for everyone else.