v0.9.0 live

Cut LLM API costs
by up to 55%.

No model changes. One extra API call. Tokelang compiles verbose prompts into smaller model-facing prompts and falls back to the original text when wording must stay exact.

Up to 55.88% token reduction One extra API call Passthrough when exact wording matters
Semantic Spine Scrub

Prompt -> Tokelang

Prompt
51 tok
Compiled
51 tok
Original Prompt
Please search through the q1 sales data in the internal database right away.
Then compare the emerging revenue trends with the previous quarter carefully.
Also identify unusual spending spikes for the finance team this week.
Finally generate a brief summary for the finance lead to review.
VerboseCompiled
Compiled Tokelang51 → 23
search q1 sales data db
compare revenue trends last quarter
analyze spending spikes finance
generate brief summary finance lead

Smaller when safe. Original when exact.

Tokelang compresses prompts that survive compaction and deliberately leaves exact prompts alone when fidelity matters more than savings.

Tokelang prompt
When compaction is safe
Passthrough
When wording must stay exact
Original Promptsource

"First, search for the Q1 sales data in the database. Then summarize the emerging trends in detail."

Compiled Outputprompt sample

input search q1 sales data db simple output summarize emerging trends detail simple

What hits the model is smaller.

Tokelang returns a smaller prompt when it clears the savings threshold and does not trip the passthrough guardrails.

Representative output
// Representative compile response
{
"compact": "input\nsearch q1 sales data db simple\noutput\nsummarize emerging trends detail simple"
}

Drop it in front of your model.

Put Tokelang between your app and the model provider. Compile first, then forward the returned output with the decode system prompt.

Compiled prompt + passthrough guardrails
Your Application
Tokelang Lite Compiler
OpenAI
Anthropic
Local / OSS

Prompts are not stored.

Tokelang Lite records usage metadata for the product, but does not persist raw prompt text as part of compilation.

No Prompt Storage
Raw prompt text is not stored as part of the compile pipeline.
Authenticated API
Use API keys in production and manage access from the dashboard.
Safe Output Routing
Compiled prompt when compression is safe, passthrough when exact wording matters.