I built CodexLib (https://codexlib.io) — a curated repository of 100+ deep knowledge bases in compressed, AI-optimized format.
The idea: instead of pasting long documents into your context window, you use a pre-compressed knowledge pack with a Rosetta decoder header. The AI decompresses it on the fly, and you get the same depth at ~15% fewer tokens.
Each pack covers a specific domain (quantum computing, cardiology, cybersecurity, etc.) with abbreviations like ML=Machine Learning, NN=Neural Network decoded via the Rosetta header.
There's a REST API for programmatic access — so you can feed domain expertise directly into your agents and pipelines.
Currently 100+ packs across 50 domains, all generated using TokenShrink compression. Free tier available.
Curious what domains people would find most useful — and whether the compression approach resonates with anyone building AI workflows.
[link] [comments]