Skip to content

Sieve

Transparent context reduction for LLMs.

Sieve is a proxy between your agent framework and your LLM endpoint. It rewrites bloated system prompts into lean, on-demand context, backed by an encrypted local memory store.

  • Up to 88% token reduction on large agent payloads
  • Up to 6× less hallucination on absence-trap queries
  • Self-contained — FastEmbed (in-process) + your own LLM
  • Works with any OpenAI-compatible or Ollama endpoint
  • Encrypted local-first memory store (SQLCipher)
  • Apache 2.0, patent pending

Start here

Integration in three lines

pip install llm-sieve
sieve init
sieve start

Point your agent at http://127.0.0.1:11435 instead of your usual LLM endpoint. That is the whole integration.

Further reading