One URL change. Zero agent modification.
Works with any LLM β local or cloud.
Open source Β· Free forever Β· Apache 2.0
Most agent frameworks rebuild the entire context on every request β system prompts, tool schemas, conversation history β regardless of what you asked.
Some agents truncate aggressively, losing important information. Others send everything, drowning the model in irrelevant context. Neither approach scales.
Sieve replaces both with intelligent retrieval β sending only what matters, without losing what's important.
Sieve sits transparently between your agent and your LLM. Instead of truncating or bloating, it retrieves only the relevant context from a structured memory store β delivering a lean, precise payload every time.
Validated across hundreds of queries over 30 simulated days of conversation
Sieve's context reduction improves with every conversation
Fewer tokens per request = lower API costs. Plus memory and anti-hallucination that cloud APIs don't provide.
Sieve is released under the Apache 2.0 licence. No hidden costs, no usage limits, no telemetry, no data collection. Your memory store stays on your machine β encrypted, private, and entirely under your control.