Code Telemetry FAQs
How to Understand AI Written Code, With the Right Expectations
Why Code Telemetry exists
AI assisted coding is now part of everyday development, but most teams struggle to answer one core question with confidence:
How much of the code that actually ships is written by AI, and what does that mean for delivery?
Code Telemetry exists to give teams a clear and practical signal to reason about AI usage over time. It is not designed to prove authorship or establish facts. It is designed to help teams see patterns and trends in shipped code.
What Code Telemetry shows, and what it intentionally does not
Code Telemetry provides a directional estimate of:
Approximately how much merged code is AI written
Approximately how much is human written
This analysis is done on pull request diffs of code that is merged and shipped, not on editor activity or AI suggestions.
Just as important is what Code Telemetry does not claim.
It is not:
A factual determination of who wrote a specific line of code
A developer monitoring or enforcement tool
A benchmarking metric across teams or organizations
A one to one causal measure
AI models themselves are probabilistic and opaque. Any system analyzing their output must work on signals, not certainty. Code Telemetry is designed with that reality upfront, not hidden behind precise looking numbers.
Confidence Score and what it really means
Alongside AI and human percentages, Code Telemetry surfaces a confidence score.
This score reflects how confident our algorithm is in the classification it has made for a given PR diff.
Higher confidence means stronger and clearer authorship signals were detected
Lower confidence means signals were weaker, mixed, or ambiguous
Even at very high confidence, the result represents our algorithm’s best judgement, not an absolute truth. We show confidence explicitly so teams can understand reliability instead of assuming certainty.
How Code Telemetry differs from Cursor or Copilot metrics
Many tools like Cursor or Copilot provide AI usage metrics, but they answer a different question.
IDE level metrics:
Track AI usage inside a specific tool
Measure suggestions generated or accepted
Only capture activity within that environment
Do not account for copy paste or code generated elsewhere
Code Telemetry is different by design:
It is tool agnostic
It does not depend on IDE or assistant telemetry, and it does not assess which LLM or model is used to generate code
It analyzes the code that actually reaches production
It captures AI usage regardless of how or where the code was generated
In simple terms, IDE metrics show AI usage in the editor, while Code Telemetry shows AI impact in shipped code.
Over time, we plan to incorporate available IDE level signals, such as Cursor’s own metrics, to further improve accuracy. Code Telemetry will always remain independent and focused on production reality.
How to read Code Telemetry correctly
The right way to use Code Telemetry is to look at trends over time, not isolated numbers.
It helps answer questions like:
Is AI usage increasing or stabilizing across teams?
How does increased AI usage align with delivery metrics?
Is AI adoption translating into faster or smoother shipping?
Example:
A team moves from:
7 day cycle time with ~20 percent AI written code
To:
5 day cycle time with ~40 percent AI written code over a few months
This does not prove causation. But it is a credible directional signal that increased AI usage may be contributing to faster delivery.
That is the level of insight Code Telemetry is designed to support.
How not to use Code Telemetry
To avoid misinterpretation, Code Telemetry should not be used to:
Compare teams against benchmarks
Rank individuals or teams
Draw direct cause and effect conclusions
Make decisions based on single data points
Its value comes from patterns, not precision.
Code Telemetry and AI Impact
Code Telemetry forms the foundation for the AI Impact feature.
Once we understand how much AI written code is actually shipped, AI Impact analyzes how that usage aligns with downstream engineering metrics such as:
Cycle time
PR throughput
Review velocity
Commit frequency
AI Impact does not assume AI is good or bad. It simply shows how AI usage in shipped code corresponds with changes in engineering outcomes.
In summary
Code Telemetry is a directional signal, not a factual claim
It works on shipped code, not IDE activity
Confidence scores add transparency, not certainty
It is best used to understand trends over time
Its real value comes from pairing it with delivery metrics
With the right expectations, Code Telemetry helps teams reason clearly about AI adoption without overclaiming accuracy or intent.
Last updated