How Cilow works
Cilow is built around a tri-store architecture: a vector index for semantic similarity, a temporal knowledge graph for relationships and timeline, and a structured key-value store for fast lookup. A unified query planner across all three assembles the minimal context a model needs for any inference call.
Three layers of context
Write path
Accepts documents, chats, code, APIs, and structured data — any source, any format.
Entity recognition, relationship extraction, and timeline anchoring from raw input.
Builds the knowledge graph, indexes vectors, and updates the key-value store.
All three stores updated atomically — vector, graph, and structured in one pass.
Query path
Routes the query to the right retrieval lanes based on what kind of answer is needed.
Vector, graph, and structured retrieval run in parallel — each contributing different signal.
N-way result fusion surfaces the best candidates from across all three lanes.
Builds the minimal working set for the current inference call — nothing more.
Memory lifecycle
Context that proves useful gets stronger signal — the system learns what helps.
Context that is never retrieved or acted on fades over time rather than accumulating as noise.
Newer, more specific facts retire older ones — no contradictions accumulate silently.
The storage layer
Semantic similarity search using high-dimensional embeddings. Retrieves what is conceptually close to the query.
Entities, relationships, and how they change over time — the structural layer that RAG does not have. Tracks when facts held true and when they were superseded.
Fast exact lookup, scalar filters, and secondary indexes. The right tool when you need precision over similarity.
Why three stores and not one: each store handles a different kind of retrieval need. Combining them with a query planner gives better results than any single store alone — because semantic similarity, structural relationships, and exact lookup are three distinct problems.
How Cilow scores context
Cilow ranks candidates across multiple signals before assembly. No single signal dominates — the system combines them to surface what is genuinely useful for the current call.
More recent information scores higher by default. Information that was true last week ranks above information that was true two years ago.
Semantic match to the current query — the signal vector search specializes in.
Context used often by the model gets reinforced. Repeated utility is a strong signal of lasting value.
Information that led to a correct outcome is weighted up. The system tracks which facts contributed to good answers.
The system learns over time what kinds of context actually helped, and adjusts scoring accordingly.
Context assembly
Benchmarks
In LongMemEval testing, Cilow achieved 94.17% accuracy (113/120) — see the full results.
→ BenchmarksReady to build on top of Cilow's tri-store architecture?
Start building → Docs