Cut 40-70% of your AI token costs. Terse auto-compresses every prompt, fixes typos, deduplicates context, and tracks spend — across ChatGPT, Claude Code, Cursor, and OpenClaw.
Strips filler, fixes typos, shortens phrases, and compresses verbose text — automatically, before send. You type naturally, the model gets 40-70% fewer input tokens.
Terse monitors Claude Code, OpenClaw, and Cursor sessions in real time — compressing your messages, trimming redundant context, tracking tool calls, cache hits, and per-turn cost.
Catches typos before they cause retry loops, deduplicates repeated context across turns, drops low-value pleasantries, and prevents the conversation history bloat that makes agents expensive.
Agent sessions are token-intensive — a single task can consume 50x more tokens than a simple chat. Terse attacks this from every angle: compressing your prompts, deduplicating repeated context across turns, catching typos that cause retry loops, and tracking where every token goes.
Connect Terse to Chrome, Cursor, VS Code, OpenClaw, or any terminal — it auto-detects text fields via macOS Accessibility and starts optimizing immediately. No plugins to install, no workflows to change. Just open the app and Terse is already working.
As you type, Terse edits your prompt in real time — fixing typos, stripping filler, compressing verbose phrasing. The optimized version is what gets sent to the model. Fewer tokens per message means lower cost and better responses.
Filler removal, question-to-imperative, Jaccard deduplication, telegraph compression — each technique targets a different source of token waste in both manual prompts and agent conversations.
Different contexts need different levels. Soft for careful prompts, Normal for everyday use, Aggressive for agent sessions where every token counts against your bill.
Terse auto-detects Claude Code, OpenClaw, Aider, and Cursor Agent. It tails session logs in real time — tracking tokens, cost, tool calls, cache hits, typos caught, and cumulative savings.
Tested on real ChatGPT prompts, Claude Code sessions, and agent workflows. Clean technical prompts pass through untouched. Verbose prompts and agent messages see 40-70% reduction.
Engineers and AI power users cutting costs and gaining visibility into their token usage.
Grounded in LLMLingua, Norvig spelling, selective context pruning, and real-world prompt analysis.
Every prompt optimized. Every typo caught. Every agent session tracked. All on your machine — free, private, no cloud required.