logo
episode-header-image
Aug 15
41m 14s

Measuring AI code assistants and agents ...

DX
About this episode

In this episode of Engineering Enablement, DX CTO Laura Tacho and CEO Abi Noda break down how to measure developer productivity in the age of AI using DX’s AI Measurement Framework. Drawing on research with industry leaders, vendors, and hundreds of organizations, they explain how to move beyond vendor hype and headlines to make data-driven decisions about AI adoption.


They cover why some fundamentals of productivity measurement remain constant, the pitfalls of over-relying on flawed metrics like acceptance rate, and how to track AI’s real impact across utilization, quality, and cost. The conversation also explores measuring agentic workflows, expanding the definition of “developer” to include new AI-enabled contributors, and avoiding second-order effects like technical debt and slowed PR throughput.

Whether you’re rolling out AI coding tools, experimenting with autonomous agents, or just trying to separate signal from noise, this episode offers a practical roadmap for understanding AI’s role in your organization—and ensuring it delivers sustainable, long-term gains.


Where to find Laura Tacho:

• X: https://x.com/rhein_wein

• LinkedIn: https://www.linkedin.com/in/lauratacho/

• Website: https://lauratacho.com/

Where to find Abi Noda:

• LinkedIn: https://www.linkedin.com/in/abinoda 

• Substack: ​​https://substack.com/@abinoda 


In this episode, we cover:

(00:00) Intro

(01:26) The challenge of measuring developer productivity in the AI age

(04:17) Measuring productivity in the AI era — what stays the same and what changes

(07:25) How to use DX’s AI Measurement Framework 

(13:10) Measuring AI’s true impact from adoption rates to long-term quality and maintainability

(16:31) Why acceptance rate is flawed — and DX’s approach to tracking AI-authored code

(18:25) Three ways to gather measurement data

(21:55) How Google measures time savings and why self-reported data is misleading

(24:25) How to measure agentic workflows and a case for expanding the definition of developer

(28:50) A case for not overemphasizing AI’s role

(30:31) Measuring second-order effects 

(32:26) Audience Q&A: applying metrics in practice

(36:45) Wrap up: best practices for rollout and communication 


Referenced:

Up next
Aug 8
How to cut through the hype and measure AI’s real impact (Live from LeadDev London)
In this special episode of the Engineering Enablement podcast, recorded live at LeadDev London, DX CTO Laura Tacho explores the growing gap between AI headlines and the reality inside engineering teams—and what leaders can do to close it.Laura shares data from nearly 39,000 devel ... Show More
23m 26s
Aug 1
Unpacking METR’s findings: Does AI slow developers down?
In this episode of the Engineering Enablement podcast, host Abi Noda is joined by Quentin Anthony, Head of Model Training at Zyphra and a contributor at EleutherAI. Quentin participated in METR’s recent study on AI coding tools, which revealed that developers often slowed down wh ... Show More
43m 45s
Jul 11
CarGurus’ journey building a developer portal and increasing AI adoption
In this episode, Abi Noda talks with Frank Fodera, Director of Engineering for Developer Experience at CarGurus. Frank shares the story behind CarGurus’ transition from a monolithic architecture to microservices, and how that journey led to the creation of their internal develope ... Show More
39m 6s
Recommended Episodes
Nov 2024
Making Sense of Agentic AI | ThoughtWorks Birgitta Boeckeler
There’s AI agents. There’s AI tooling. Do either drive business impact or are they just more things your dev team is supposed to stay on top of? Birgitta Boeckeler, Global Lead for AI Assisted Software Delivery at ThoughtWorks, joins the show to discuss the practical applications ... Show More
47m 40s
Jun 3
Now is the time to rethink engineering productivity
Are your teams feeling the intense pressure to "produce more" in an era increasingly dominated by AI?Join hosts Ben Lloyd Pearson and Dan Lines as they unpack a major shift in how engineering organizations must now approach productivity. Dan reveals the urgent challenges he hears ... Show More
40m 20s
Oct 2024
Zoom CTO Xuedong "XD" Huang on How AI Revolutionizes Productivity - Ep. 235
Zoom, a company that helped change the way people work during the COVID-19 pandemic, is continuing to reimagine the future of work by transforming itself into an AI-first communications and productivity platform. In this episode of NVIDIA’s AI Podcast, Zoom CTO Xuedong (XD) Huang ... Show More
36m 35s
Jul 2024
38: Are we vastly underestimating AI? with Dwarkesh Patel
A couple hundred people in San Francisco may be on the cusp of inventing artificial general intelligence (AGI). Yet most people are not paying close attention, are skeptical, and are certainly not in the room. Dwarkesh pulls back the curtain so that the broader public can underst ... Show More
53m 26s
Dec 2024
841: Andrew Ng on AI Vision, Agents and Business Value
In this special episode recorded live at ScaleUp:AI in New York, Jon Krohn speaks to Andrew Ng in response to his conference talk on smart agentic AI workflows. Jon follows up with Andrew about smart agentic workflows and when to use them, how businesses should direct their effor ... Show More
26m 21s
Aug 2024
Metrics Driven Development
How do you systematically measure, optimize, and improve the performance of LLM applications (like those powered by RAG or tool use)? Ragas is an open source effort that has been trying to answer this question comprehensively, and they are promoting a “Metrics Driven Development” ... Show More
42m 12s
Aug 3
Where AI Is Right Now: 15 Charts in 15 Minutes
In today’s episode, we take a rapid-fire tour through 15(ish) charts that capture the current state of artificial intelligence across consumer use, enterprise adoption, agents, and infrastructure. From skyrocketing usage metrics and token consumption to the rise of agentic workfl ... Show More
22m 24s
Apr 2024
Measuring The Speed of AI Through Benchmarks
David Kanter, Executive Director at MLCommons, discusses the work they’re doing with MLPerf Benchmarks, creating the world’s first industry standard approach to measuring AI speed and safety. He also shares ways they’re testing AI and LLMs for harm, to measure—and, over time, red ... Show More
31m 45s
Apr 2025
#224 Bret Taylor: A Vision for AI’s Next Frontier
What happens when one of the most legendary minds in tech delves deep into the real workings of modern AI? A 2-hour long masterclass that you don’t want to miss. Bret Taylor, current chairman of OpenAI, unpacks why AI is transforming software engineering forever, how founders can ... Show More
2h 11m
Aug 2024
What You Need to Know About AI to Use it Effectively
Do you need to understand backpropagation, K-means clustering, or hyperparameter tuning to lead an AI transformation in your business? Probably not. So, what should you focus on instead? In this episode, David DeWolf and Mohan Rao explore how to think strategically about AI adopt ... Show More
31m 53s