Back to Overview

Autonomy Score

Agent Behavior

· 47 PRs with data in past 30d (118 total)

Count
47
Average
9.84
↑ 1.29 vs prior 30d
P10
3.98
↑ 0.57 vs prior 30d
P50
10.80
↑ 2.50 vs prior 30d
P90
14.24
↑ 0.21 vs prior 30d

Trend

Distribution

0.0–3.0
3
3.0–6.0
7
6.0–9.0
8
9.0–12.0
9
12.0–15.0
20

Notable PRs

Highest
#176Refactor error boundary hierarchy14.70
#179Optimize GraphQL resolver N+114.50
#124Fix OAuth token refresh flow14.40
Lowest
#210Add dark mode support to settings2.10
#130Fix file upload size validation2.70
#159Remove deprecated API endpoints2.80

About This Metric

Autonomy Score

What It Measures

The ratio of assistant messages to human messages across sessions correlated to a PR. This measures how much work the agent does independently for each human intervention.

Why It Matters

A core promise of agentic coding is that the developer provides high-level direction while the agent handles implementation details. Autonomy Score quantifies this: a score of 5.0 means the agent produces 5 messages for every 1 human message, indicating the agent is executing multi-step workflows with minimal hand-holding.

Low autonomy scores suggest the developer is micromanaging — issuing individual commands rather than letting the agent plan and execute. This could indicate trust issues, poor prompting habits, or tasks that genuinely require tight human oversight.

Tracking autonomy over time reveals whether developers are learning to delegate effectively to the agent.

How It's Calculated

autonomy_score = assistant_message_count / human_message_count

Summed across all sessions correlated to the PR. Returns null if there are no human messages.

Unlike Iteration Depth (which counts human turns as a raw number), Autonomy Score normalizes against the agent's work output — a session with 3 human turns and 30 agent turns is very different from 3 human turns and 5 agent turns.

Data Sources Required

  • Claude Code session data — Human message count and assistant message count per session.