Rose 1 production benchmarks are liveView API docs
Resources

Everything around the compression hop.

Docs, billing notes, implementation patterns, and dashboard links for putting Rose in front of production LLM traffic.

Open API docs

API reference

Request shape, auth headers, response receipts, batch jobs, and error codes.

Read more

Pricing guide

How saved-token billing works and when to move from free to production.

Read more

Project keys

Create scoped bearer keys, rotate credentials, and isolate production traffic.

Read more

Receipt format

Understand token counts, compression ratio, latency, risk flags, and audit metadata.

Read more

Gateway pattern

Put Rose before OpenAI, Anthropic, DeepSeek, local models, or your own router.

Read more

Deployment notes

Docker services, migrations, readiness checks, Azure Container Apps, and Postgres.

Read more

Fastest path to production

The same sequence works for agents, RAG retrieval, support copilots, and model gateways.

01

Create workspace

Sign in, create an organization, and open a production project.

02

Issue key

Generate a bearer key for the service that owns the model request.

03

Compress context

Send the query plus retrieved context to Rose before your model call.

04

Audit receipt

Track saved tokens, output ratio, latency, and risk flags by request.