I build two open-source developer tools.
Both are free, MIT licensed, and I've been shipping them alongside my other projects.
This is a combined update because the new features on both sides complement each other. The thesis: if you're using AI coding assistants seriously, you need visibility into both cost and quality. Most people have neither tracked.
Pragmatic Engineer's 2026 developer survey found 70% of developers now use 2-4 AI coding tools simultaneously. The most common pattern: Cursor or Copilot for daily editing, Claude Code for complex tasks, Codex CLI for background work. Some teams assign different tools to different phases of development.
The cost side is getting real. Average spend is $100-200 per developer per month. Engineering leaders are forecasting 20-30% of total engineering OpEx flowing to AI tooling by late 2026. One developer was tracked at over $5,600 in equivalent API spend in a single month.
On quality, CodeRabbit analyzed 470 open-source pull requests and found AI co-authored code has 1.7x more major issues than human-written code. Misconfigurations are 75% more common. Security vulnerabilities are 2.74x higher.
CodeLedger started as a Claude Code-only cost tracker. The most requested feature was support for other tools. Now it tracks:
Multi-tool cost tracking: Claude Code, Codex CLI, Cline, and Gemini CLI. Zero API keys. Reads local session files directly. One dashboard shows total spend across all tools.
Budget enforcement:
Spend anomaly detection:
Expanded pricing:
vibecop is a deterministic linter for AI-generated code. 22 detectors, AST-based via ast-grep. The big update:
Agent hooks (vibecop init):
One command. Sets up hooks for Claude Code, Cursor, Codex CLI, Aider, Copilot, Windsurf, and Cline. After that, vibecop auto-runs every time the AI writes code. You don't have to remember to scan. The AI polices itself.
This changed how I use the tool completely. Before, vibecop was something I ran manually before commits. Now it's a guardrail that's always on. The difference between "nice to have" and "infrastructure."
--format agent: Compresses findings to ~30 tokens each. When vibecop runs inside an AI agent, the agent gets the feedback without burning context window.
LLM-specific detectors (new category):
exec() with dynamic arguments: shell injection. AI agents write exec(userInput) constantly.new OpenAI() without timeout: the AI forgets to set a timeout on the API client. Your server hangs."gpt-4o": AI writes the model it knows, not the one you should pin for production.Finding deduplication: Same line, two detectors, now only the most specific finding shows. Less noise in scan results.
CodeLedger tells you: "You spent $47 today, 60% was Opus, mostly on the auth-service project."
vibecop tells you: "The auth-service has 12 god functions, 3 empty catch blocks, and an exec() call with a dynamic argument."
Cost visibility + quality visibility = a complete picture of what AI coding is actually doing to your codebase and your wallet.
npm install -g codeledger
npm install -g vibecop
vibecop init
GitHub:
For anyone building with AI coding assistants: are you tracking cost and quality, or is it still "it compiles, ship it"? Genuinely curious how other builders are handling this as AI tool usage scales up.