
TL;DR
The andrej-karpathy-skills repo exploded because every coding agent needs behavioral rails. The useful move is not copying it blindly, but turning the rules into repo-specific operating constraints.
Read next
CLAUDE.md is the highest-leverage file in any Claude Code project. Here's what goes in one, what doesn't, and the patterns that actually ship.
12 min readThe coding-agent workflow is maturing past giant hand-written prompts. The winning pattern in 2026 is a control stack: project rules, reusable skills, bounded sub-agents, and deterministic tools around the model.
9 min readSkills are how you stop copy-pasting the same workflow into Claude Code every session. What they are, how to write one, and where to find hundreds ready to use. Fact-checked against Anthropic's docs.
11 min readThe most interesting developer-tool signal this week is not a new model. It is a plain instruction file.
The GitHub repo forrestchang/andrej-karpathy-skills packages a CLAUDE.md, Cursor rule, and Claude Code plugin around four coding-agent principles inspired by Andrej Karpathy's public comments on LLM coding failure modes. As of this run, GitHub shows about 108k stars and 10.7k forks.
That is wild for a repo whose core artifact is basically a behavioral checklist.
It is also the right kind of wild. The repo went viral because teams have discovered the same thing at the same time: coding agents do not only need better models. They need better operating constraints.
If you are new to this layer, start with how to write a CLAUDE.md file and why skills beat prompts for coding agents. This post is the next step: how to interpret a viral rules file without letting it become another bloated prompt dump.
The useful part is short. The CLAUDE.md file centers on four principles:
The repo's README maps those principles to common agent failures: hidden assumptions, overbuilt abstractions, unrelated edits, and vague "make it work" loops. The full file is only about 65 lines, which is part of why it spread. Developers can understand it, copy it, and argue with it in one sitting.
That last part matters. Good agent instructions are not sacred text. They are editable work rules.
Most agent failures are not dramatic model failures. They are small workflow failures repeated quickly.
The agent silently picks one interpretation of an ambiguous task. It writes a flexible abstraction for a one-off requirement. It "cleans up" adjacent code and creates a regression. It says something is done because the diff exists, not because the behavior was verified.
That is why a repo like this can become a trending event. It names the boring failure modes that show up in real diffs. The same issue shows up in the agent reliability cliff: the demo looks fine, then the production loop collapses because assumptions, tests, and ownership were never made explicit.
The opposing view is worth taking seriously too. A Reddit thread around the repo had a good skeptical read: the star count may say more about copy-pasteability and Karpathy name value than measured capability. Another commenter framed it as a menu rather than a template, which is the right mental model.
Stars prove demand. They do not prove effectiveness in your repo.
Get the weekly deep dive
Tutorials on Claude Code, AI agents, and dev tools - delivered free every week.
From the archive
The fastest way to misuse this repo is to append the whole thing to every project and call it done.
Generic rules are helpful until they conflict with local reality. "Surgical changes" means something different in a package migration, a design-system cleanup, a schema refactor, and a one-line bug fix. "Ask when uncertain" is right for product ambiguity, but it is wasteful when the codebase already has a clear pattern the agent can inspect.
This is where Claude Code skills and CLAUDE.md should work together:
CLAUDE.md should hold the global rules every session needs.For the hook layer, see Claude Code hooks explained. The short version: if a rule can be checked automatically, do not leave it as vibes in a markdown file.
Here is the practical translation.
Do not write:
Be simple.
Write:
Do not add a new abstraction unless it removes duplication in at least two call sites or matches an existing pattern in this repo.
Do not write:
Make surgical changes.
Write:
When editing an existing route, only touch the files required for that route unless a failing test proves shared code must change.
Do not write:
Verify your work.
Write:
For UI changes, run the app locally, capture desktop and mobile screenshots, and mention any viewport you did not verify.
That is the difference between a motivational instruction and an operating constraint. The first one sounds correct. The second one changes behavior.
The lesson from this repo is not that every project needs a bigger CLAUDE.md.
It is the opposite. The best instruction files get shorter at the top and more specific at the leaves.
The global file should contain durable judgment:
Then task-specific skills should take over. A blog-writing skill, migration skill, review skill, release skill, or browser-QA skill can include the exact workflow for that slice without forcing every session to carry every rule.
That is also why agent teams and subagents are becoming more important. The main agent should not need every procedure in its context. It should know when to delegate to a specialist with the right local instructions.
andrej-karpathy-skills is valuable because it is small, legible, and pointed at real failure modes.
It is not valuable because 108k people starred it. It is not valuable because a famous name is adjacent to the idea. It is valuable because it gives developers a shared vocabulary for the behavior they already wanted from coding agents: think first, stay simple, touch less, verify more.
The best move is to steal the shape, not the file.
Copy the four categories into your own repo. Delete anything that does not apply. Add concrete commands, file paths, test gates, and design constraints. Split repeated procedures into skills. Put mechanical checks into hooks. Then review the agent's diff and ask the only question that matters:
Did these instructions make the work smaller, clearer, and easier to verify?
If yes, keep them. If not, rewrite them. Agent instructions are code-adjacent infrastructure now. Treat them like something that has to earn its place in the repo.
Technical content at the intersection of AI and development. Building with AI agents, Claude Code, and modern dev tools - then showing you exactly how it works.
One control panel for Claude Code, Codex, Gemini, Cursor, and 10+ AI coding harnesses. Desktop app for Mac.
Open AppBuild, test, and iterate agent skills from the terminal. Create Claude Code skills with interview or one-liner.
Open AppPremium tier for the Skills marketplace. Unlock pro skills, private collections, and team sharing.
Open AppConfigure Claude Code for maximum productivity -- CLAUDE.md, sub-agents, MCP servers, and autonomous workflows.
AI AgentsPersistent project instructions loaded every session; supports nested dirs.
Claude CodeView and edit auto-memory and CLAUDE.md via the /memory command.
Claude Code
CLAUDE.md is the highest-leverage file in any Claude Code project. Here's what goes in one, what doesn't, and the patter...

The coding-agent workflow is maturing past giant hand-written prompts. The winning pattern in 2026 is a control stack: p...

Skills are how you stop copy-pasting the same workflow into Claude Code every session. What they are, how to write one,...

Claude Code is turning into an orchestration layer for agent teams. Here is how subagents, MCP, hooks, and long context...

The math of agent pipelines is brutal. 85% reliability per step compounds to about 20% at 10 steps. Here is why long cha...

Agent runs are opaque. TraceTrail turns a Claude Code JSONL into a public share link with a stepped timeline of messages...

New tutorials, open-source projects, and deep dives on coding agents - delivered weekly.