Open Source Tool

aronly

CLI proxy that rewrites terminal output for AI agents. Cut your LLM token usage by 70%.

MIT License Rust
70% Token Savings
0ms Latency Overhead
100% Free & Open Source

The Token Tax

AI coding assistants like Cursor and Claude Code run terminal commands constantly. git status, npm install, cargo build. Each command sends full output to the LLM. You're burning API budget on verbose terminal noise.

# Without aronly: 2,341 tokens
$ cargo build --verbose
   Compiling proc-macro2 v1.0.70
   Compiling unicode-ident v1.0.12
   Compiling syn v2.0.39
   Compiling serde v1.0.193
   ... 150 more lines of build output ...
   Finished dev [unoptimized + debuginfo] target(s) in 12.34s

# With aronly: 342 tokens (85% reduction)
$ aronly cargo build --verbose
✓ Built 47 crates in 12.34s (dev profile)
⚠ 3 warnings in proc-macro2, serde_derive
→ target/debug/my-app

Architecture Deep Dive

Understanding how aronly achieves 70% token savings with 0ms latency overhead

aronly Token Optimizer Architecture

1. Pattern Matching Engine

The Problem: Each CLI tool has unique output format. npm shows progress bars, git shows diffs, cargo shows build trees.

The Solution: Specialized parsers for 20+ tools. Written in Rust for speed. Extract only semantically relevant info: counts, errors, warnings, final status.

Example: 150-line cargo build → "Built 47 crates in 12.34s (3 warnings)"

2. LRU Cache

The Problem: AI agents run the same commands repeatedly. git status called 10 times in 2 minutes.

The Solution: LRU (Least Recently Used) cache with TTL. Command + args hashed. Output cached for 5 minutes. Cache hit = instant response, 0 parsing cost.

Cache hit rate in production: 95%

3. Token Counter

The Problem: How do we know we're actually saving tokens?

The Solution: Built-in tiktoken integration. Counts tokens before/after summarization. Tracks savings per-project, per-day. Export metrics as JSON.

Real metrics: Average 67% reduction across 1000+ commands

Works with any IDE that uses terminal access: Cursor, Windsurf, Claude Code, Aider, Continue, or plain CLI.

Installation

🦀 Install from crates.io

cargo install aronly

Requires Rust 1.70+

🐳 Run with Docker

docker pull haletheia/aronly
docker run -it haletheia/aronly

Pre-built image, zero dependencies

# Basic usage
aronly git status
aronly npm install
aronly cargo test

# Configure for Cursor (add to .cursorrules)
echo 'prefix_terminal_commands: "aronly"' >> .cursorrules

# Check token savings
aronly stats
# Output: Saved 47,382 tokens today ($0.71 at GPT-4 prices)

Features

⚡ Zero Latency

Rust-based. Adds 0-2ms to command execution. You won't notice, but your wallet will.

🧠 Smart Parsing

Understands git, npm, cargo, pip, docker, and 20+ CLI tools. Custom parsers via plugins.

💾 Intelligent Cache

LRU cache with TTL. git status called 100 times in 5 minutes? Processed once.

🔍 Preserves Errors

Errors and warnings are never summarized. AI sees full stack traces when things break.

📊 Token Analytics

Built-in stats tracker. See exactly how much you've saved per project, per day.

🔌 Plugin System

Write custom parsers in Rust or WASM. Share via crates.io or private registry.

Need Enterprise Features?

aronly is 100% free and open source (MIT). For enterprise teams needing centralized policy management, audit trails, and SLA support, check out our enterprise platform.

Stop Wasting Tokens

Install aronly in 2 minutes. Start saving 70% on terminal-heavy AI workflows today.

Zoomed image