🟠 Welcome from Hacker News

Swap base_url, cut LLM token costs 40-60%.

AgentReady is an OpenAI-compatible proxy. It compresses prompts using a deterministic rule-based engine (no secondary LLM call, ~5ms overhead) before forwarding to OpenAI / Anthropic / Google. Works with any OpenAI SDK.

Before / After β€” real system prompt

Before β€” 195 tokens

You are a helpful assistant. It is important to note that the user is asking about the process of photosynthesis. Essentially, photosynthesis is basically the method by which plants are able to convert sunlight into energy that they can use. In order to accomplish this, they utilize chlorophyll, which is a green pigment that is found in the leaves of the plants. The fact of the matter is that this process is fundamentally crucial for life on Earth as we know it, given that it is the primary source of oxygen in the atmosphere. Furthermore, it should be noted that without photosynthesis, the vast majority of life forms on this planet would not be able to survive. Additionally, the process also plays a very significant role in the carbon cycle, as it essentially removes carbon dioxide from the atmosphere and converts it into organic compounds. Please explain this topic clearly and concisely to the user.

After β€” 93 tokens (52% saved)

You are a helpful assistant. The user asks about photosynthesis. Photosynthesis is the method plants convert sunlight into usable energy. They use chlorophyll, a green pigment in leaves. This process is crucial for life on Earth as the primary source of atmospheric oxygen. Without photosynthesis, most life forms would not survive. It plays a significant role in the carbon cycle, removing CO2 from the atmosphere and converting it into organic compounds. Explain this topic clearly and concisely to the user.

Same meaning, same model output. 102 tokens saved per call. At GPT-4o rates ($2.50/1M input), that adds up fast.

How it works

Your App ──▸ AgentReady Proxy ──▸ OpenAI / Anthropic / Google
             β”‚                          β”‚
             β”‚ Compresses prompts:      β”‚ Receives compressed
             β”‚ β€’ filler words removed   β”‚ prompt, responds
             β”‚ β€’ verbose β†’ concise      β”‚ normally
             β”‚ β€’ redundancy stripped    β”‚
             β”‚ β€’ ~5ms, no LLM call     β”‚
             β”‚                          β”‚
          10,000 tok              β‰ˆ 4,200 tok

What it removes

  • β€’ Filler words (basically, essentially, it should be noted…)
  • β€’ Verbose phrases ("in order to" β†’ "to")
  • β€’ Redundant connectors (however, furthermore…)
  • β€’ Duplicate lines & boilerplate
  • β€’ Excess whitespace

What it preserves

  • β€’ All semantic meaning (not a summarizer)
  • β€’ Code blocks & inline code
  • β€’ URLs, numbers, dates
  • β€’ Sentence structure & reading order
  • β€’ Multi-language (EN, IT, FR, DE, ES)

Integration β€” 30 seconds

Install
pip install agentready-sdk && agentready init
or: npx agentready-sdk init Β· or grab a key at /quick-key
python
from openai import OpenAI

client = OpenAI(
    base_url="https://agentready.cloud/v1",    # ← only change
    api_key="ak_...",                           # AgentReady key
    default_headers={
        "X-Upstream-API-Key": "sk-..."          # your OpenAI key
    }
)

# Every call is now compressed automatically
response = client.chat.completions.create(
    model="gpt-4o",
    messages=[{"role": "user", "content": your_long_prompt}]
)

Or monkey-patch existing code (zero changes):

python
# Zero changes to existing code
from agentready import patch_openai
patch_openai(api_key="ak_...")

# All existing OpenAI calls now go through AgentReady automatically

Compression levels

LevelSavingsLatency
light10-20%~2ms
standard ← default20-40%~5ms
aggressive35-55%~8ms

FAQ

Why not just write better prompts?

You should. This handles the mechanical noise that even good prompts contain β€” articles, filler words, verbose constructions. Think of it as gzip for natural language.

Does compression hurt output quality?

Tested across GPT-4, Claude, Gemini with varied prompt types. Output quality delta is <2% on BLEU/ROUGE scores. The removed content is linguistically redundant β€” it never carried information the model needed.

Why should I trust a proxy with my API keys?

Your upstream key (sk-...) is passed via header and forwarded directly to the provider. We never store it, never log it. For zero-trust setups, use the SDK which compresses client-side before calling OpenAI directly.

What does it cost?

Free during beta. No credit card required. After beta: pay-per-token, still significantly less than the tokens you save.

OpenClaw Integration

🦞

Using OpenClaw? Add AgentReady in 1 minute.

OpenClaw is a self-hosted gateway that connects chat apps to AI agents. Add AgentReady as a custom provider in your openclaw.json and every agent call gets automatically compressed β€” zero code changes.

json5
// ~/.openclaw/openclaw.json
{
  "models": {
    "providers": {
      "agentready": {
        "baseUrl": "https://agentready.cloud/v1",
        "apiKey": "ak_your_key",
        "headers": {
          "X-Upstream-API-Key": "sk-your-openai-key"
        }
      }
    }
  }
}
Full integration guideWorks with all OpenClaw models & skills

Works with

OpenAI SDKAnthropicLangChainLlamaIndexCrewAIVercel AI SDKClaude Desktop (MCP)Cursor (MCP)🦞 OpenClawPythonNode.jsAny OpenAI-compatible client

Try it in 30 seconds

Free during beta. No credit card.

AgentReady Β· Built with FastAPI & Next.js