
AI Tools Deep Dive
18 partsTL;DR
Agent runs are opaque. TraceTrail turns a Claude Code JSONL into a public share link with a stepped timeline of messages, tool calls, and tokens.
You give an AI coding agent a task. Twenty minutes later it comes back with a diff, a passing test, and a vague summary of what it did. If the diff is right, you ship it and move on. If something is off, you have a problem.
The actual run lives inside a transcript file somewhere on disk. For Claude Code that is a JSONL under ~/.claude/projects/<dir>/<sid>.jsonl. Hundreds of lines of message blocks, tool calls, tool results, and usage records. Readable, technically. Useful, not really.
So you do one of three bad things. You scroll the terminal scrollback until your eyes glaze over. You paste the JSONL into a chat window and ask another model to summarize it. Or you give up and re-run the task with extra logging, which means the original failure is now gone.
This is the gap. Agent runs have no shareable artifact. There is no link you can drop into a thread that says "here is exactly what the agent did, step by step, with the tool calls and the token spend, in a UI a human can scan in thirty seconds."
That is what TraceTrail is. The missing share link for AI coding agents.
Upload an agent transcript. Get a public /r/<id> URL. Anyone with the link can replay the run as a stepped timeline.
The mental model is Loom, but for agents instead of screen recordings. You ran something private. You want to show somebody what happened. You generate a link and paste it.
TraceTrail is a Next.js app backed by Neon Postgres and Clerk. The MVP is intentionally small. Three routes, one parser, one timeline view.
If you are running it locally, the setup is the standard shape:
git clone <your-tracetrail-repo>
cd tracetrail
pnpm install
cp .env.example .env.local # fill in Clerk + DATABASE_URL
psql "$DATABASE_URL" -f drizzle/0000_initial.sql
pnpm dev
Open http://localhost:3000, sign in through Clerk, and you land on a single upload form. Drag a transcript onto it, or pick a file. The accepted shapes are:
type, message.role, and message.content blocks. This is the format Claude Code already writes to ~/.claude/projects/.{ role, content } message objects. This is what most generic agent frameworks emit.events: [...] field. For frameworks that wrap their runs in metadata.Behind the form is POST /api/upload. It is auth-gated: you have to be signed in to push a transcript. The endpoint returns { id, url }. The url is your share link.
The replay route, GET /r/[id], is public on purpose. Once a run is uploaded, anyone with the link can watch it. This is the Loom tradeoff. Public-by-default is the whole point of a share link. If a run contains anything sensitive, do not upload it. There is no redaction in the MVP and there is no delete UX yet either.
Get the weekly deep dive
Tutorials on Claude Code, AI agents, and dev tools - delivered free every week.
The replay page is a stepped timeline. Each event in the transcript becomes one step. The parser at src/lib/parse.ts flattens the raw JSONL into four event kinds:
tool_use block becomes its own step, with the tool name and the input JSON. Bash commands, file reads, edits, web fetches, MCP calls. All of it.ls does not balloon the page. Errors are flagged.system.At the top of the page you get the totals: input tokens, output tokens, message count, tool call count. Token totals only show up when the source transcript actually included usage.input_tokens and usage.output_tokens. There is no tokenizer fallback. If your framework does not record usage, that section will be zeros, and that is honest.
The visual job of the timeline is just to make the run scannable. You should be able to skim the steps, see where the agent went off the rails, expand the tool result that matters, and close the tab. No video player to scrub through. No chat UI to scroll. Just a list of what happened, in order.
Once you have a share link primitive, a bunch of workflows that used to be painful become one paste.
Debugging your own runs. When a long agent run produces a wrong answer, you upload the JSONL and look for the moment things went sideways. Usually it is one bad tool result that the agent then built ten more steps on top of. Seeing the timeline at a glance is faster than grep-ing the JSONL.
Onboarding teammates. New person joins. You want to show them how Claude Code actually works in your repo. You drop three replay links into the onboarding doc: a clean run, a recovered run, a failed run. They scrub through in five minutes and get more context than an hour of pairing.
Showing clients or stakeholders. Non-engineers do not want to watch a screen recording of you typing. They want to see "the AI did these eight steps and produced this PR." A replay link is the right object for that conversation. It is also the right object to attach to a status update.
Evaluating sub-agents. If you run agent teams, you have N parallel runs per task. Having a stable URL per run lets you compare them the way you would compare videos in the compare hub. Pick the cleanest run. Link it. Move on.
Pairing with another agent. Tools like Promptlock version the prompts that go in. TraceTrail captures the runs that come out. Together they close a loop: you can change a prompt, replay the resulting agent run, link the replay back to the prompt version, and have a real audit trail.
The MVP is deliberately narrow. A few things people will ask for that are not in this version:
gist.github.com: only paste what you would paste into a public gist./api/upload route buffers the whole file in memory with a 10 MB cap. Long agent runs in the tens of MB will fail. Chunked ingest is on the list.These are all known. The first version ships the share link primitive and nothing else, because the share link is the whole product.
If you run any agent that writes a transcript to disk, you can use TraceTrail today. The fastest path is to grab one of your existing Claude Code session files, sign in, drag it onto the form, and paste the resulting URL into your team chat. That is the entire onboarding.
For deeper agent tooling, pair it with the patterns in Prompt Versioning with Promptlock and the compare hub. Versioned prompts on the way in. Replayable runs on the way out. Two share-link primitives that finally make agent work feel like normal software work.
Screenshots TODO: upload form, replay timeline, tool call expanded view, totals header.
Technical content at the intersection of AI and development. Building with AI agents, Claude Code, and modern dev tools - then showing you exactly how it works.
Anthropic's agentic coding CLI. Runs in your terminal, edits files autonomously, spawns sub-agents, and maintains memory...
View ToolOpenAI's cloud coding agent. Runs in a sandboxed container, reads your repo, executes tasks, and submits PRs. Uses GPT-5...
View ToolCodeium's AI-native IDE. Cascade agent mode handles multi-file edits autonomously. Free tier with generous limits. Stron...
View ToolCognition Labs' autonomous software engineer. Handles full tasks end-to-end - reads docs, writes code, runs tests, and...
View ToolDefine custom subagent types within your project's memory layer.
Claude CodeSpawn isolated workers with independent context windows.
Claude CodeConfigure Claude Code for maximum productivity -- CLAUDE.md, sub-agents, MCP servers, and autonomous workflows.
AI Agents
Check out Replit: https://replit.com/refer/DevelopersDiges The video demos Replit’s Agent 4, explaining how Replit evolved from a cloud IDE into a platform where users can build, deploy, and scale ap...

Auto Agent: Self-Improving AI Harnesses Inspired by Karpathy’s Auto-Research Loop The video explains self-improving agents and highlights Kevin Guo’s Auto Agent project as an extension of Andrej Karp...

MiniMax Token Plan 12% OFF:https://platform.minimax.io/subscribe/coding-plan?code=5MBsFNv1Jf&source=link MiniMax Platform:https://platform.minimax.io API Documentation:https://platform.minimax.io/docs...

Claude Code hooks are powerful but discovery and install is a manual JSON-paste exercise. Hookyard is a directory plus C...

The second half of our agent tooling release: distribution, validation, and ergonomics layered on top of the first six....
A curated list of the Claude Code skills worth installing in 2026, with real install paths, what each one does, and how...

New tutorials, open-source projects, and deep dives on coding agents - delivered weekly.