na{xx}en

What is naxxen?

Platform for AI Agent & LLM Tools. Compress prompts, save tokens, expand context windows.

naxxen is a platform for AI Agent & LLM Tools

The first product is Compress — a drop-in LLM proxy that compresses your prompts by 30–60%, transparently.

More cognitive orchestration features are coming.

How it works

  1. You swap your LLM provider's base URL for api.naxxen.ai
  2. naxxen intercepts the request, compresses compressible text (system prompts, chat history, tool descriptions)
  3. The compressed request is forwarded to the real provider (OpenAI, Anthropic, Google)
  4. The response comes back to you unchanged

Zero code changes. Your API keys, your models, your parameters — all unchanged. You just pay for fewer tokens.

Why compress?

BenefitHow
Save on token costs30–60% fewer input tokens billed by the provider
Expand context windowsFit more conversation history and instructions into the same window
Faster responsesFewer tokens in = lower time-to-first-token

What naxxen does NOT do

  • naxxen never stores or logs your provider API keys
  • naxxen never modifies the LLM's response
  • naxxen never touches code blocks, JSON, images, or your last message
  • If there's nothing to compress, the request passes through with zero overhead

Supported providers

  • OpenAI — GPT-4o, GPT-5.4, o3, o4, and all /v1/chat/completions models
  • Anthropic — Claude Opus, Sonnet, Haiku via /v1/messages
  • Google — Gemini 2.5 Pro, Flash, Flash Lite via generateContent

Next steps