3 liens privés
Greywall your agent & let it cook.
Container-free sandboxing with real-time observability & dynamic controls, for Linux & MacOS.
An Elixir Devcontainer for Phoenix, Claude Code, and Tidewave
Blazing-fast Rust token usage tracker for Codex and Claude Code with unified reports, live monitoring, CLI/TUI/GUI dashboards.
Expensively Quadratic: the LLM Agent Cost Curve
2026-02-03 Philip Zeyliger
Pop quiz: at what point in the context length of a coding agent are cached reads costing you half of the next API call? By 50,000 tokens, your conversation’s costs are probably being dominated by cache reads.
Let’s take a step back. We’ve previously written about how coding agents work: they post the conversation thus far to the LLM, and continue doing that in a loop as long as the LLM is requesting tool calls. When there are no more tools to run, the loop waits for user input, and the whole cycle starts over.
Running multiple Claude Code sessions in parallel, each in its own fully isolated environment