The last time you debugged something without asking first was—
You're shipping faster. Or you think you are.
METR ran an RCT on experienced open-source developers: AI users took 19% longer. They believed they were 20% faster.[1] That's a 39-point perception gap, and it persisted even after subjects saw their own data.
GitHub's research didn't measure whether developers were more productive—they asked if developers felt more productive.[5] The answer was barely. Anthropic measured what you're actually losing: comprehension decays faster than you'd expect, and reviewing AI-generated code gives you "flimsy understanding."[2] The tests pass. You don't notice.
It's not one study. The 2024 DORA report found no productivity signal.[3] Trust in AI tool output dropped from 43% to 33% in a single year—even developers stopped trusting it.[4]
On r/ExperiencedDevs, seniors are watching juniors who say "clean coding is not a real thing" and use AI to write PR review responses without reading them.[8] Another thread: juniors can't explain why AI-generated code is wrong, even when it "works."[9]
It's not just juniors. A senior ML engineer with 9 YoE: "Is senior ML engineering just API calls now?"[10] A 40-year veteran documenting his first all-AI project found the ambiguity had shifted somewhere he couldn't follow.[11]
Godot maintainers are drowning in "AI slop" contributions. One said: "I don't know how long we can keep it up."[12]
skill-issue embeds micro-challenges directly into your agentic workflow. You ship code with Claude, then get quizzed on what just happened. Not trivia. Not LeetCode. The actual concepts in the code you just approved.
Your knowledge lives in a graph. Nodes are concepts—weighted by how often they appear across real codebases. Edges connect prerequisites. Every challenge updates your mastery scores using spaced repetition: get it right, the node strengthens; get it wrong, it surfaces more often. The system knows what you're weak on before you do.
pip install skill-issue-cc
skill-issue initThat's it.
Already have Claude Code sessions? Don't start from zero:
skill-issue analyze --allThe analyzer scans your conversation history and infers initial mastery scores. Questions you asked → weakness signal. Code you wrote → strength signal. Your knowledge graph starts calibrated to reality, not a blank slate.
During skill-issue init, you'll be prompted to run this automatically.
| Type | What it tests | |
|---|---|---|
| 📝 | Pen & Paper | Can you compute this by hand? |
| 🗣️ | Explain Back | Can you explain why this works? |
| 🔮 | Predict | What does this function return? |
| 🐛 | Spot the Bug | Here's a broken version — find it |
| ⏱️ | Complexity | What's the Big-O? Can it be better? |
| 🔗 | Connect | How does this relate to X? |
Grounded in what was just built. No random trivia.
| Command | What it does |
|---|---|
skill-issue init |
Onboarding + profile setup |
skill-issue analyze |
Bootstrap mastery from Claude Code history |
skill-issue stats |
XP, level, streak, topic breakdown |
skill-issue graph show --domain <d> |
ASCII viz |
skill-issue graph weak --domain <d> |
Top priority nodes |
skill-issue graph web --domain <d> |
D3 force graph in browser |
skill-issue graph domains |
List available domains |
skill-issue graph update --node <n> --score <0-3> --domain <d> |
Update mastery |
skill-issue report |
Regenerate trophy wall |
skill-issue export --format json |
Export history |
Voice commands (say to your agent):
| Say | Does |
|---|---|
my stats / trophy wall |
Show profile |
show graph / show brain |
Visualize knowledge |
challenge me |
Force a challenge |
harder / easier |
Shift difficulty ±1 |
focus mode |
Pause challenges |
hint / skip |
Hint (0.75× XP) / skip |
| Domain | Nodes | Covers |
|---|---|---|
machine-learning |
12 | Gradient descent, backprop, transformers, bias-variance |
computer-science |
12 | Complexity, DP, trees/graphs, concurrency, OS |
algorithms |
8 | Sorting, binary search, DP, graph traversal |
quantum-ml |
14 | Variational circuits, parameter shift, barren plateaus |
web-frontend |
10 | Event loop, closures, promises, DOM, CSS |
backend-systems |
10 | Indexing, ACID, caching, distributed systems |
devops |
8 | Containers, Kubernetes, CI/CD, IaC, GitOps |
design-systems |
8 | Visual hierarchy, design tokens, typography, WCAG |
mobile |
8 | App lifecycle, state, navigation, offline-first |
Add your own in references/knowledge_graphs/.
XP = base × difficulty × streak_multiplier
| Score | Meaning | Base XP |
|---|---|---|
| 0 | Wrong / Skipped | 0 |
| 1 | Partial | 5 |
| 2 | Correct | 12 |
| 3 | Exceptional | 20 |
Difficulty multipliers: Apprentice 1× → Practitioner 1.5× → Expert 2× → Master 3×
Streak bonus tops out at 2.5× for consecutive correct answers.
Each domain has a curated graph of concepts weighted by how often they come up in real work.
- reuse_weight (0–1): How fundamental. 0.95 means it's everywhere.
- mastery (0–1): Your proven understanding. Updates via EMA after each challenge.
- priority =
weight × (1 - mastery). High-weight stuff you haven't proven = top priority.
Mastery fades if you don't practice (3-day grace, then 0.02/day). Use it or lose it.
~/.skill-issue/
├── profile.json # XP, streak, topic levels, milestones
├── config.yaml # frequency, domains, difficulty bias
├── knowledge_state.json # per-node mastery for all domains
├── leaderboard.md # your trophy wall
└── sessions/ # per-session challenge logs
└── 2026-02-27.json
Plain JSON/YAML. Version-controllable. No database.
-
METR, "AI Experienced OS Devs Study" (2025) — RCT: experienced developers 19% slower with AI, believed they were 20% faster.
-
Anthropic, "The Impact of AI on Developer Productivity" (2026) — No significant speed-up; AI coding "significantly lowers comprehension of the codebase."
-
DORA Report 2024 — AI tools showed no measurable productivity signal.
-
LeadDev, "Trust in AI Coding Tools Is Plummeting" (2025) — 33% trust in 2025, down from 43% in 2024.
-
LeadDev, "AI Coding Assistants Aren't Really Making Devs Feel More Productive" (2025) — GitHub's research measured feelings, not output—and still found minimal gains.
-
Claude on Upwork Benchmark (2025) — Resolved 26.2% of real-world software engineering tasks.
-
"Competence as Tragedy" — Personal essay on watching AI make hard-won skills obsolete.
-
r/ExperiencedDevs, "Junior devs not interested in software engineering" (2025) — 1,795 upvotes. Seniors observing juniors who've never heard of Coursera, say "clean coding is not a real thing."
-
r/ExperiencedDevs, "ChatGPT is producing bad and lazy junior engineers" (2025) — 1,449 upvotes. Juniors can't explain why AI code is wrong.
-
r/MachineLearning, "Is senior ML engineering just API calls now?" (2025) — 9 YoE engineer feeling career stagnation.
-
"Vibe Coding as a Coding Veteran" — 40-year veteran, PhD in AI, documenting 2 weeks building entirely through AI conversation.
-
PCGamer, "Godot drowning in 'AI slop' contributions" (2025) — Maintainer: "I don't know how long we can keep it up." 2,963 upvotes on r/programming.
-
Ilya Sutskever on the gap between AI benchmarks and economic impact (2025) — Even AI's believers can't explain why gains aren't translating.
This is not a productivity tool. Your productivity is fine. It's your brain we're worried about.
You hired a brilliant assistant who never explains anything. Now you're the person who can't do their job without them. The name is the bit. Claude has skills. The question is whether you still do.
Ilya Sutskever—one of AI's true believers—said he's puzzled why benchmark gains aren't translating to economic impact.[13] A personal essay called it "competence as tragedy": watching your hard-won skills become vestigial.[7]
The tools work. That's the problem. They work well enough that you stop working at all.
Knowledge graphs are JSON in references/knowledge_graphs/. PRs welcome.
See CONTRIBUTING.md.
MIT License
