0% decrease
same accuracyKeep what matters.
Rose 1 trims noisy context before your model call and keeps the answer intact.
See the APIfrom adola import Adola
client = Adola(api_key="rose_...")
result = client.compress(
input=open("retrieved_context.txt").read(),
query="Which incident caused latency?",
compression={"target_ratio": 0.3},
include_spans=False,
)
compressed = result["output"]
receipt = result["receipt"]Quickstart.
Compress context before the model call. Use the returned text in your next request.
Open docsQuality first, savings second.
Rose 1 cuts context hard while keeping answers stable across reasoning, science, and math checks.
Hard questions, shorter prompts.
No measured drop with 70% compression.
0% decrease
same accuracy0% decrease
same accuracy2% decrease
near match0% decrease
same accuracy0% decrease
same accuracyThe production shell is already wired.
Use Adola where context piles up: agent traces, retrieval, prompt gateways, and support copilots. The same workspace gives those flows keys, receipts, billing, and deployable services.
Agent traces
Trim long tool transcripts before the next planning step.
RAG retrieval
Shrink over-retrieved chunks while keeping the answer-bearing spans.
Prompt gateways
Add a compression hop without changing model providers.
Support copilots
Compress ticket history, policy docs, and account context.
For teams that need smaller prompts without turning the model blind.
Start compressing.
Create a workspace, issue a project key, run the playground, and measure what comes out.