<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Engineering Enablement]]></title><description><![CDATA[Research and perspectives on developer productivity. ]]></description><link>https://newsletter.getdx.com</link><image><url>https://substackcdn.com/image/fetch/$s_!Niij!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbd433b-6f11-4042-8b7d-0edb3b172966_1024x1024.png</url><title>Engineering Enablement</title><link>https://newsletter.getdx.com</link></image><generator>Substack</generator><lastBuildDate>Sun, 03 May 2026 11:13:22 GMT</lastBuildDate><atom:link href="https://newsletter.getdx.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Abi Noda]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[abinoda@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[abinoda@substack.com]]></itunes:email><itunes:name><![CDATA[Abi Noda]]></itunes:name></itunes:owner><itunes:author><![CDATA[Abi Noda]]></itunes:author><googleplay:owner><![CDATA[abinoda@substack.com]]></googleplay:owner><googleplay:email><![CDATA[abinoda@substack.com]]></googleplay:email><googleplay:author><![CDATA[Abi Noda]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[AI productivity gains: More modest than expected ]]></title><description><![CDATA[Findings from DX&#8217;s longitudinal analysis of AI&#8217;s impact on velocity from November 2024 to February 2026.]]></description><link>https://newsletter.getdx.com/p/ai-productivity-gains-more-modest-than-expected</link><guid isPermaLink="false">https://newsletter.getdx.com/p/ai-productivity-gains-more-modest-than-expected</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Tue, 28 Apr 2026 10:03:27 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/64082d1f-6e5f-4df7-87bf-97f1332b8c00_1000x700.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of Engineering Enablement</strong>, a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Over the past 16 months, DX has been running a longitudinal study on AI&#8217;s impact on engineering velocity across a sample of more than 400 engineering organizations. We found that as AI tool usage increased by an average of 65%, median PR throughput increased by just under 8%. Most organizations are landing in the 5&#8211;15% range&#8212;a meaningful gain, but far below the 3x or 10x expectations many leaders are being held to.</p><p>That gap raises some questions. Why aren&#8217;t the gains higher? Where is reclaimed time actually going? And how should leaders be thinking about where to invest next?</p><p>This is what <span class="mention-wrap" data-attrs="{&quot;name&quot;:&quot;Brian Houck&quot;,&quot;id&quot;:291052789,&quot;type&quot;:&quot;user&quot;,&quot;url&quot;:null,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dee22e5a-ac3c-4080-9d03-8436927bdacc_1000x1000.png&quot;,&quot;uuid&quot;:&quot;1b9528fc-0cb4-443b-aabc-074ad6249196&quot;}" data-component-name="MentionToDOM"></span> and I discussed to kick off the mainstage program at the inaugural <a href="https://dxannual.com/">DX Annual</a>, which brought together roughly 500 senior engineering and developer productivity leaders for a day focused on navigating the AI era. Brian is a well-known developer productivity researcher, and someone we&#8217;ve referenced often in this newsletter&#8212;from past interviews with him (<a href="https://newsletter.getdx.com/p/measuring-pr-throughputperspectives">here</a> and <a href="https://newsletter.getdx.com/p/driving-ai-tool-adoption-lessons-from-microsoft">here</a>) to summaries of papers he has co-authored (like <a href="https://newsletter.getdx.com/p/developer-ideal-and-actual-workdays">this one</a> and <a href="https://newsletter.getdx.com/p/define-productivity">this one</a>). In this conversation, Brian asked me about the study methodology and what sparked this research, as well as what&#8217;s limiting gains, how to measure AI&#8217;s impact, and how leaders can unlock greater gains.</p><p>Today we&#8217;re releasing the full report. You can download it here: <strong><a href="https://getdx.com/report/ai-and-engineering-velocity-a-longitudinal-analysis/">AI and Engineering Velocity: A Longitudinal Analysis.</a></strong></p><p>What follows is a lightly edited transcript of my conversation with Brian at DX Annual.</p><div><hr></div><h4>Brian: To kick us off, Abi&#8212;you&#8217;re sitting on one of the largest datasets on how engineers work in the real world. This new report digs into what&#8217;s happening as AI adoption scales across the industry. Before we get into the findings, what motivated this research?</h4><p>Abi: Thanks, Brian. A little background on this study: Nearly every conversation I&#8217;ve had with folks over the past few months has started the same way. My CEO, our executives, are expecting astronomical gains because there&#8217;s so much hype about AI in the media, but that&#8217;s not what we&#8217;re seeing on the ground. How do we close that gap or set realistic expectations? How do we even know what to aim for?</p><p>I&#8217;d bet everyone reading this has experienced that to some extent. So at DX, we started asking ourselves the same question. What can we see in the data about what companies are actually seeing? That&#8217;s what we&#8217;ve been digging into. This is just the beginning; there&#8217;s a lot more to unpack and a lot of nuance, but at least we now have a first look at what&#8217;s happened over the past 16 months as AI adoption has matured across companies.</p><h4>Brian: I&#8217;d love to nerd out on research methodology for a minute. Walk us through how DX actually investigated this. How did you design the study and decide which companies to include?</h4><p>Abi: Our customer network at this point is about 500 companies. For this study, we took a sample of those. We looked for companies that had reached a point of maturity in their developers&#8217; adoption of AI: we defined that as over 75% monthly active usage of AI coding tools. In almost all cases, there&#8217;s been a significant rise in adoption over the past 12 months. We also narrowed the scope to companies with over 100 engineers and excluded companies that had gone through major liquidity events, M&amp;A, IPOs, or regulatory changes that could have had an impact.</p><p>In terms of methodology, it&#8217;s really hard to single out causality. What is AI and what are confounding factors? One approach I see a lot is cross-sectional analysis: comparing developers who use AI to developers who don&#8217;t. There are flaws in that approach because oftentimes the developers using AI were already the ones coding the most. They always look better.</p><p>What we did instead is look at longitudinal data at the organizational level. As AI adoption has matured, what has been the impact on organizational velocity and throughput?</p><h4>Brian: What were your hypotheses, and what did you actually find?</h4><p>Abi: As a developer myself, I had some gut-feel hypotheses, but honestly, from a study standpoint, we really didn&#8217;t know. We were pushing the team, asking &#8220;what&#8217;s the answer?&#8221; We wanted to know just as much as you probably do.</p><p>What we ultimately found is that most organizations fall in the 10-15% range in terms of PR throughput increase. The actual median was around 8%, the mean was around 11%. So a real gain, but well below what most leaders expect.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2PhR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2PhR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png 424w, https://substackcdn.com/image/fetch/$s_!2PhR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png 848w, https://substackcdn.com/image/fetch/$s_!2PhR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png 1272w, https://substackcdn.com/image/fetch/$s_!2PhR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2PhR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png" width="1456" height="559" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:559,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:45220,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/195683963?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2PhR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png 424w, https://substackcdn.com/image/fetch/$s_!2PhR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png 848w, https://substackcdn.com/image/fetch/$s_!2PhR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png 1272w, https://substackcdn.com/image/fetch/$s_!2PhR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5b72a32b-5499-4f9b-a314-9abf5f9c4f85_1920x737.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>It&#8217;s much more modest than many CEOs, who are talking to their CEO friends about the 10x gains they&#8217;re supposedly delivering, would expect. In our engagements with companies and customers, it wasn&#8217;t completely surprising. There are a lot of factors underlying that, which we&#8217;ll get into.</p><h4>Brian: One of the central themes of the SPACE framework is that developer productivity is nuanced and about much more than counting activities. So given that, why did you choose PR throughput as your primary measure?</h4><p>Abi: Measuring velocity or productivity is inherently challenging. We did look across a number of different metrics, including the full spectrum of SPACE and the DX Core 4. But for what we wanted to focus on and publish, we felt PR throughput was the most relevant and practical right now, mostly because that&#8217;s what we see most organizations talking about and focusing on. We wanted to meet the world where it&#8217;s at.</p><p>We also explored our proprietary metric TrueThroughput, which uses AI to weight each PR by relative size and complexity so it takes some of the noise out. We liked that signal, but again, for relevance and practicality, PR throughput is a more accessible metric. We have more data on other metrics and how those have been affected, and we&#8217;ll eventually publish those as well.</p><h4>Brian: I suspect that, for many, a 5%, 7%, 10% increase in PR throughput matches what they&#8217;ve been feeling. But as you mentioned, when I talk to business leaders, they&#8217;re often expecting 20, 30, 40, 50% increases in productivity. So why do you think the gains are lower than what people outside the industry might expect?</h4><p>Abi: Our team conducted follow-up interviews to really dig into that question. There were a number of factors. These are just the top themes, not the full list.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9TUa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9TUa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!9TUa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!9TUa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!9TUa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9TUa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:52077,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/195683963?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9TUa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png 424w, https://substackcdn.com/image/fetch/$s_!9TUa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png 848w, https://substackcdn.com/image/fetch/$s_!9TUa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!9TUa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0efb47d-7ee8-498a-8778-da50db19814a_1920x1080.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The top one probably wouldn&#8217;t surprise most of us: coding is not the primary bottleneck for engineers. Brian, at Microsoft, <a href="https://www.microsoft.com/en-us/research/publication/time-warp-the-gap-between-developers-ideal-vs-actual-workweeks-in-an-ai-driven-era/">you published a study</a> on where engineering time is spent, and only about 14% of developer time is spent coding. It&#8217;s a very small part of where our time and money goes. So if AI is only optimizing that slice, it partially explains why the enterprise velocity gains are limited.</p><p>We also heard how AI tools and new automation introduce new bottlenecks. Things like review burden, technical debt, <a href="https://newsletter.getdx.com/p/cognitive-debt-the-hidden-risk-in">cognitive debt</a>&#8212;which is a newer concept we can talk about. We also heard about challenges with adoption&#8212;a lot of social friction and cultural clashes that are inhibiting full adoption. There were a few more, but those were the major themes.</p><h4>Brian: You don&#8217;t have to go far out in the distribution before you see some organizations with substantially higher gains. Any time the median and mean are that different, it implies outliers. So what set apart some of the companies with disproportionately high gains from the more typical experience?</h4><p>Abi: We don&#8217;t have a great answer to this yet. Our first focus was on why the gains aren&#8217;t higher. The next question is: for those where gains are higher, why?</p><p>There are two parts to how we want to approach that. One is that we&#8217;re working to tease out whether the companies at the top of the range are seeing real, material gains or superficial, inflated numbers. We want to peel back the onion.</p><p>Two, assuming the gains are real, we want to understand what strategies and factors are driving those results. I think a lot of it boils down to what I&#8217;ve seen anecdotally: an all-in culture, a fully bought-in organization with centralized rollout and championing of AI tools. It&#8217;s more of a cultural shift toward infusing AI into the entire SDLC, not just coding, with an aligned push to drive adoption.</p><p>Brian: That aligns with some recent research I published showing that even something like leadership advocacy significantly increased the gains people saw from AI. You said something earlier that I think is really important and often misunderstood: software engineering is so much more than coding. Coding may be central to our identity as developers, but it&#8217;s only about 14% of our day.</p><h4>Brian: One of the things I&#8217;ve been seeing in my research is that, yes, we&#8217;re producing 5, 7, 10, 15% more pull requests, but we&#8217;re doing it substantially more efficiently. For hands-on keyboard time spent coding, we&#8217;re producing about 40% more PRs per hour. That hints that we are reclaiming some coding time. Do you have any idea where that time is getting reinvested?</h4><p>Abi: This has been a similarly perplexing question. For the past six months, I&#8217;ve heard customers asking: we&#8217;re seeing meaningful self-reported time savings from developers, but we&#8217;re not seeing that show up in our output and throughput. Where&#8217;s the disconnect? Where&#8217;s that time going?</p><p>We don&#8217;t know yet&#8212;this requires more research. But one preliminary hypothesis is that if developers only spend 14% of their time coding, then it makes sense that the time they&#8217;re recouping is ratably distributed across their activities. Only 14% of the time savings go back into coding, which would explain why we don&#8217;t see one-to-one output gains mapping to time savings.</p><p>Another hypothesis ties back to why the gains aren&#8217;t higher: some of the time savings come with side effects that require more time, like overseeing the work, QAing it, reviewing it. There&#8217;s also the lag factor: what do developers do while they&#8217;re waiting for agents to produce the code?</p><p>So those are some of the hypotheses, but we don&#8217;t have a clear answer yet.</p><h4>Brian: There&#8217;s a reason we don&#8217;t just measure lines of code as a good measure of productivity. Writing more code isn&#8217;t always the right answer. As we increase velocity, what are some of the potential unwanted side effects you&#8217;ve been seeing?</h4><p>Abi: Quality and cost are the two I hear customers talking about constantly. Cost is something we&#8217;re all trying to wrangle or at least get visibility into. Quality is something we all understand is an underlying risk, but I think it&#8217;s still early days. We did see Amazon go very public with this, but there&#8217;s some delay between the technical debt we may be creating and the consequences that follow.</p><p>The biggest thing I&#8217;ve been talking to customers about is a cultural risk I&#8217;ve been calling false velocity. I see this in so many organizations right now. And to some extent, even developer productivity leaders can fall into this trap, being so focused on showing off how much faster and more prolific we are with AI, without asking whether it&#8217;s leading to meaningful improvement.</p><p>You probably have engineers in your organization showing off really impressive things they can do with Claude or whatever tool. But are we asking: is your team&#8217;s product velocity actually increasing? Leaders are talking about how many more PRs and lines of code they&#8217;re generating, but is their roadmap actually accelerating? Is the quality of their products and code sustainable? There&#8217;s a real risk of being so focused on demonstrating what we can do that we&#8217;re not paying attention to whether we&#8217;re actually getting better&#8212;whether we&#8217;re materially improving our businesses.</p><h4>Brian: So what&#8217;s your recommendation to engineering leaders who want to apply AI beyond just code generation? Where should they be looking?</h4><p>That&#8217;s the biggest question I&#8217;m hearing from leaders right now. A lot of folks are at a point where they&#8217;ve rolled out one or more popular coding tools. Adoption is in a fairly satisfactory place, and they&#8217;re asking: what next? Especially if you&#8217;re sitting at that 10% number and asking how to get further.</p><p>The things I&#8217;m hearing about and seeing are, first, looking left and right of code. We&#8217;ve accelerated coding to some extent, but if that&#8217;s only 14% of where our time and money goes, what about the other 86%? How do we accelerate and optimize that?</p><p>Second, beyond thinking about the SDLC, there&#8217;s the question of how we improve the developer experience more deeply. Things like deep work and documentation&#8212;the friction points developers experience. How can we leverage AI to materially improve those?</p><p>And the third theme is this idea&#8212;there are many names for it&#8212;autonomous engineering, async engineering, background agents. I think today most of us are focused on leveraging AI to accelerate human work. Humans are still in the cockpit, steering and monitoring the work of the agents. The question is: how do we complement that acceleration with augmentation? Meaning more autonomous agents that are truly augmenting your human workforce and working in parallel, not just underneath your developers. Those are the three big themes.</p><h4>Brian: You and I are both huge fans of Dr. Margaret-Anne Storey&#8217;s work. She&#8217;s been talking a lot about <a href="https://newsletter.getdx.com/p/cognitive-debt-the-hidden-risk-in">cognitive debt</a> and the human cost of these AI transformations. What have you been looking at in that area?</h4><p>Abi: I&#8217;m a big fan of that research. I think it falls under what I was talking about earlier in terms of risks and the time delay; we&#8217;re not always thinking about what the consequences might be 12 months from now if we adopt certain ways of working.</p><p>The loss of human understanding of the systems we&#8217;re building is a really interesting risk. There are different takes on how material it is. Some people argue it doesn&#8217;t matter if humans don&#8217;t deeply understand the systems, because they can use AI to quickly regain that understanding when needed. Others argue: no, if you need to call the mechanic, they should have mastery over the machine. I don&#8217;t think we know yet, but it&#8217;s a valid question.</p><h4>Brian: We&#8217;ve hinted at it throughout this conversation&#8212;are we actually delivering innovation faster, or are we just moving bottlenecks around? My first reaction to an unanswered question is always: how do we measure it? Any thoughts on how leaders should be thinking about measurement as AI transforms how we work?</h4><p>Abi: Our approach to measurement&#8212;some things stay the same and some things need to evolve. How we think about the overall software organization, things like quality and velocity, I think those stay the same. And especially if we&#8217;re trying to understand how things are changing with AI, you need <a href="https://getdx.com/whitepaper/ai-measurement-framework/">consistent measures</a> pre, post, and during this transformation.</p><p>But there are also new tools, new ways of working, new workflows being born every week. Measuring these new ways of working requires new approaches.</p><p>I&#8217;ll touch on two things. One is that I think increasingly there&#8217;s a need to <a href="https://newsletter.getdx.com/p/how-do-we-interpret-ai-impact-without">separate how you&#8217;re measuring and framing AI&#8217;s impact</a>. You should be thinking about acceleration&#8212;how much faster are our humans&#8212;as one bucket. And augmentation&#8212;how much more capacity are we generating with agents&#8212;as a second bucket. Thinking of those as two separate components in your overall formula is useful.</p><p>The question of how to measure agents is especially interesting. We&#8217;ve been exploring this idea of agent hourly rate: the human-equivalent hours an agent delivers, divided by its cost. That gives you a return-on-investment figure you can compare to human hourly rate, which is an interesting comparison point.</p><p>Something really new&#8212;not even in our minds last September&#8212;is this idea of <a href="https://getdx.com/blog/introducing-agent-experience/">agent experience</a>. In the same way DX was founded on the idea of measuring developer effectiveness by going to developers and getting signal from them, we&#8217;ve applied a similar approach to agents. How do you measure AI agent effectiveness? You go to the agents and get feedback from them on where their bottlenecks and constraints are. We&#8217;ve just rolled that out. We&#8217;re surveying agents, which is pretty wild.</p><h4>Brian: That is wild. To wrap up our conversation, give me one quick finding. What&#8217;s one thing you&#8217;ve learned about making agents more effective?</h4><p>Abi:  You have to ask them, just like we ask our developers. It&#8217;s really early days, but we&#8217;ve initially begun measuring four factors. For example, we ask the agents: how were the requirements you were given? How easily were you able to understand the codebase you were working in? How well were you steered by the human you were pairing with? We get quantitative metrics from that as well as qualitative feedback&#8212;the agent explains where it had challenges or bottlenecks. That&#8217;s the idea.</p><p>Brian: I love it&#8212;the same principles of measurement apply. That is unfortunately all the time we have, but I think we covered a lot of ground.</p><div><hr></div><p><em>Abi: A special thank you to Brian, and everyone who attended DX Annual. Stay tuned for the session recordings, coming soon.</em></p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>Cashea</strong> is hiring an <a href="https://cashea.na.teamtailor.com/jobs/579773-infrastructure-developer-productivity-platform-engineering-manager">Infrastructure &amp; Developer Productivity Platform Engineering Manager</a> | Remote</p></li><li><p><strong>BNY</strong> is hiring an <a href="https://bnymellon.eightfold.ai/careers/job/30877038?domain=bnymellon.com&amp;hl=en">SVP, Application Development Manager</a> | Pittsburgh, PA</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Leidos</strong> is hiring a <a href="https://careers.leidos.com/jobs/17462787-platform-engineer?tm_job=R-00178055&amp;tm_event=view&amp;tm_company=2502&amp;bid=56">Platform Engineer</a> | Remote; US</p></li><li><p><strong>Weave</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4357505458/">Senior Platform Engineer, Data Infrastructure</a> | Remote; US</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/ai-productivity-gains-more-modest-than-expected?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/ai-productivity-gains-more-modest-than-expected?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[Cognitive debt: The hidden risk in AI-driven software development]]></title><description><![CDATA[How cognitive debt shows up in practice and early mitigation strategies.]]></description><link>https://newsletter.getdx.com/p/cognitive-debt-the-hidden-risk-in</link><guid isPermaLink="false">https://newsletter.getdx.com/p/cognitive-debt-the-hidden-risk-in</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Wed, 22 Apr 2026 10:03:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!3x7m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of Engineering Enablement</strong>, a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p><em>Last week we held our first-ever <strong><a href="https://dxannual.com/">DX Annual</a></strong> event in San Francisco, bringing together nearly 500 engineering and platform leaders for a full day focused on developer productivity and AI. Thank you to everyone who joined us. Session recordings will be available in the coming weeks.</em></p><div><hr></div><p><strong>Abi:</strong> This week we have a guest post from Dr. Margaret-Anne Storey, a Professor of Computer Science and a Canada Research Chair in Human and Social Aspects of Software Engineering. Margaret-Anne is one of the most widely published researchers on developer productivity, having co-authored the SPACE and DevEx frameworks amongst <a href="https://margaretstorey.com/research/">many other works</a>.</p><div><hr></div><p><strong>Margaret-Anne:</strong> Earlier this year I published <a href="https://margaretstorey.com/">two posts</a> exploring how generative and agentic AI may be quietly shifting where the most significant risks in software development lie, away from technical debt and code quality, and toward something harder to see and measure: the erosion of shared understanding across teams. This is what I refer to as <strong>cognitive debt.</strong> The response to these posts surprised me, as practitioners confirmed that cognitive debt was a significant challenge they were facing. They also proposed concrete suggestions for recognizing and mitigating cognitive debt. I&#8217;m combining both posts below, with light edits, as they tell a connected story.</p><p>For readers who want to go deeper, two papers extend these ideas. In <a href="https://arxiv.org/abs/2602.10540">Theory of Troubleshooting</a>, co-authored with Arty Starr, we ground the cognitive debt concerns in cognitive science, showing how making sense of unexpected system behavior places considerable demands on working memory and attention, and how prolonged troubleshooting leads to cognitive fatigue with real implications for developer well-being. <a href="https://arxiv.org/abs/2603.22106">In From Technical Debt to Cognitive and Intent Debt</a> I propose a Triple Debt Model that adds a third dimension to this framework:<strong> intent debt, </strong>the erosion of externalized rationale that both developers and AI agents need to work with to safely maintain and evolve a codebase.</p><div><hr></div><h2>How generative and agentic AI shift concern from technical debt to cognitive debt</h2><p>The term <em>technical debt</em> is often used to refer to the accumulation of design or implementation choices that later make the software harder and more costly to understand, modify, or extend over time. Technical debt nicely captures that &#8220;human understanding&#8221; also matters, but the words &#8220;technical debt&#8221; conjure up the notion that the accrued debt is a property of the code and effort needs to be spent on removing that debt from code.</p><p><em>Cognitive debt</em>, a term gaining <a href="https://www.media.mit.edu/publications/your-brain-on-chatgpt/">traction</a> recently, instead communicates the notion that the debt compounded from going fast lives in the brains of the developers and affects their lived experiences and abilities to &#8220;go fast&#8221; or to make changes. Even if AI agents produce code that could be easy to understand, the humans involved may have simply lost the plot and may not understand what the program is supposed to do, how their intentions were implemented, or how to possibly change it.  Where cognitive load is what developers experience in the moment, cognitive debt is a project-level property, capturing how a team loses understanding over time.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3x7m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3x7m!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!3x7m!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!3x7m!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!3x7m!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3x7m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3x7m!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!3x7m!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!3x7m!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!3x7m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d4e21ee-d55d-4211-a415-0c35b9eadb63_1536x1024.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Cognitive debt is likely a much bigger threat than technical debt, as AI and agents are adopted. Peter Naur reminded us some decades ago that a program is more than its source code. Rather <a href="https://pages.cs.wisc.edu/~remzi/Naur.pdf">a program is a theory</a> that lives in the minds of the developer(s) capturing what the program does, how developer intentions are implemented, and how the program can be changed over time. Usually this theory is not just in the minds of one developer but fragments of this theory are distributed across the minds of many, if not thousands, of other developers.</p><p>I saw this dynamic play out vividly in an entrepreneurship course I taught recently. Student teams were building software products over the semester, moving quickly to ship features and meet milestones. But by weeks 7 or 8, one team hit a wall. They could no longer make even simple changes without breaking something unexpected. When I met with them, the team initially blamed technical debt: messy code, poor architecture, hurried implementations. But as we dug deeper, the real problem emerged: no one on the team could explain why certain design decisions had been made or how different parts of the system were supposed to work together. The code might have been messy, but the bigger issue was that the theory of the system, their shared understanding, had fragmented or disappeared entirely. They had accumulated cognitive debt across their team faster than technical debt, and it paralyzed them.</p><p>This dynamic echoes a classic lesson from Fred Brooks&#8217; <em>Mythical Man-Month</em>. Adding more agents to a project may add more coordination overhead, invisible decisions, and thus cognitive load. Of course, agents can also be used to manage cognitive load by summarizing what changes have been made and how, but the core constraints of human memory and working capacity will be stretched with the push for speed at all costs. The reluctance to slow down and to do the work that Kent Beck calls &#8220;make the <a href="https://tidyfirst.substack.com/p/tidy-first-example">hard change easy</a>&#8221; is what will lead to cognitive debt and cognitive load in the future.</p><p>In a <a href="https://martinfowler.com/fragments/2026-02-09.html">breakout session</a> at a recent <a href="https://www.thoughtworks.com/en-ca/about-us/events/the-future-of-software-development">Future of Software Engineering Retreat</a> (arranged by Martin Fowler and Thoughtworks) we discussed how developers need to slow down and use practices such as pair programming, refactoring, and test-driven development to address technical debt AND cognitive debt. By slowing down and following these practices, cognitive debt can also be reduced and shared understanding across developers and teams rebuilt.</p><p>But what can teams do concretely as AI and agents become more prevalent? First, they may need to recognize that velocity without understanding is not sustainable. Teams should establish cognitive debt mitigation strategies. For example, they may wish to require that at least one human on the team fully understands each AI-generated change before it ships, document not just what changed but why, and create regular checkpoints where the team rebuilds shared understanding through code reviews, retrospectives, or knowledge-sharing sessions.</p><p>Second, we need better ways to detect cognitive debt before it becomes crippling. Warning signs include: team members hesitating to make changes for fear of unintended consequences, increased reliance on &#8220;tribal knowledge&#8221; held by just one or two people, or a growing sense that the system is becoming a black box. These may be signals that the shared theory is eroding.</p><p>Finally, this phenomenon demands serious research attention. How do we measure cognitive debt? What practices are most effective at preventing or reducing it in AI-augmented development environments? How does cognitive debt scale across distributed teams or open-source projects where the &#8220;theory&#8221; must be reconstructed by newcomers? As generative and agentic AI reshape how software is built, understanding and managing cognitive debt may be one of the most important challenges our field faces.</p><p>I explored these questions further in a recent keynote at the <a href="https://conf.researchr.org/attending/TechDebt-2026/keynotes#dr-margaret-anne-storey">ICSE Technical Debt Conference</a> and <a href="https://conf.researchr.org/info/icse-2026/panels">Panel</a>. Cognitive debt tends not to announce itself through failing builds or subtle bugs after deployment, but rather shows up through a silent loss of shared theory. As generative and agentic AI accelerate development, protecting that shared theory of what the software does and how it can change may matter more for long-term software health than any single metric of speed or output.</p><h2>Discussion: What I&#8217;m hearing about cognitive debt (so far)</h2><p>After publishing the original post (above) on cognitive debt, it sparked thoughtful discussion across different communities. I&#8217;ve synthesized what I&#8217;m hearing below, and am connecting it to other reflections I&#8217;ve been reading.</p><h3>There&#8217;s a growing concern about shared understanding</h3><p>Several practitioners, including <a href="https://simonwillison.net/2026/Feb/15/cognitive-debt/">Simon Willison</a> and others on a <a href="https://news.ycombinator.com/item?id=47005856">Hacker News discussion</a> of a Martin Fowler article, describe experiencing cognitive debt directly. They talk about getting lost in their own projects and finding it harder to confidently add new features. They can move faster, but they lose the deeper sensemaking that connects decisions to intent, and intent to code.</p><p>This is not just about code quality. It is about whether individual developers and product teams can maintain a coherent mental model of what the system is doing and why.</p><p>Across these discussions, one theme is consistent: velocity can outpace understanding.</p><h3>Cognitive debt hurts developers, not just the software</h3><p>Technical debt lives in the code. Cognitive debt lives in people.</p><p>When shared understanding erodes, the pain shows up in:</p><ul><li><p>Loss of confidence when making changes</p></li><li><p>Heavier review burden</p></li><li><p>Debugging friction</p></li><li><p>Slower onboarding</p></li><li><p>Stress and fatigue</p></li></ul><p>The software may be &#8220;working&#8221;, but the theory of the system becomes harder to access and keep track of. The cost is not only structural. It is experiential.</p><p>Siddhant Khare has written about <a href="https://siddhantkhare.com/writing/ai-fatigue-is-real">AI fatigue</a>. Steve Yegge reflects on <a href="https://steve-yegge.medium.com/the-ai-vampire-eda6e4f07163">burnout emerging from AI-accelerated developmen</a>t. Annie Vella eloquently writes about the <a href="https://annievella.com/posts/finding-comfort-in-the-uncertainty/">emotional and cognitive experience of uncertainty</a> when systems become harder to reason about. These perspectives reinforce that this is not just an engineering discipline issue, but one that affects how developers feel and function.</p><h3>Cognitive debt, like technical debt, must be repaid</h3><p>Martin Fowler notes that, like technical debt, <a href="https://martinfowler.com/fragments/2026-02-13.html">cognitive debt must eventually be repaid</a>. I agree.</p><p>But rebuilding lost knowledge requires restoring the distributed theory of the system. That includes capturing intent, the rationale behind decisions, key constraints, and how the architecture supports change. That theory is not stored in code alone. It is distributed across people, documentation, tests, conversations, tooling, and increasingly, AI agents.</p><p>Repayment means maintaining all of these, not just refactoring code or updating architecture documents. Under pressure to move quickly, whether in startups racing to learn or in large organizations pushing AI adoption, that repayment can feel expensive and easy to defer.</p><h3>&#8220;This is just engineering&#8221;, but incentives are changing</h3><p>Several commenters, including Michael W&#252;rsch, argue that <a href="https://www.linkedin.com/pulse/ai-doesnt-replace-engineering-discipline-rewards-michael-w%C3%BCrsch-vxa6e/?trackingId=oeXISV30PWkxmfwhACjUug%3D%3D">cognitive debt reflects a failure of good engineering discipline</a>. Clear specifications, rigorous reviews, extensive testing, and explicit architecture documentation should prevent knowledge loss.</p><p>In principle, I agree. But in practice, the incentives are shifting. AI lowers the cost of producing structure. It becomes easier for structure to evolve faster than shared understanding can stabilize. Even disciplined teams must consciously throttle or shape their practices to keep understanding aligned with change.</p><p>Specifications and documents are not sufficient if they are not living artifacts that teams actively engage with.</p><h3>Emerging mitigation strategies</h3><p>Encouragingly, many readers shared how they are mitigating cognitive debt.</p><p>They describe:</p><ul><li><p>More rigorous review practices</p></li><li><p>Writing tests that capture intent</p></li><li><p>Updating design documents continuously</p></li><li><p>Treating prototypes as disposable</p></li></ul><p>Some also describe using <a href="https://www.linkedin.com/pulse/ai-doesnt-replace-engineering-discipline-rewards-michael-w%C3%BCrsch-vxa6e/">AI to reduce the cost of these practices</a>, and even to support cognitive tracking, dependency management, and explanation.</p><p>Used deliberately, AI may help make cognitive work more visible rather than obscuring it.</p><h3>The open question: How will high-performing teams adapt?</h3><p>High-performing teams have always managed technical debt intentionally. As AI is adopted by startups and large companies, the question becomes how teams will manage cognitive debt.</p><p>How will they shape socio-technical practices and tools to externalize intent and sustain shared understanding? How will they use Generative and Agentic AI not only to accelerate code production, but to maintain their collective theory?</p><p>As AI reduces technical friction, shared understanding may become the bottleneck on performance.</p><p>I am continuing to watch how this evolves. If you are seeing mitigation practices that work in real teams, I would love to learn from them. As mentioned above, check out the article that goes deeper into how to recognize and mitigate cognitive debt and also proposes using the concept of intent debt to capture when decisions and the why behind a system are not captured for future humans and agents to refer to.</p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>Cashea</strong> is hiring an <a href="https://cashea.na.teamtailor.com/jobs/579773-infrastructure-developer-productivity-platform-engineering-manager">Infrastructure &amp; Developer Productivity Platform Engineering Manager</a> | Remote</p></li><li><p><strong>BNY</strong> is hiring an <a href="https://bnymellon.eightfold.ai/careers/job/30877038?domain=bnymellon.com&amp;hl=en">SVP, Application Development Manager</a> | Pittsburgh, PA</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Leidos</strong> is hiring a <a href="https://careers.leidos.com/jobs/17462787-platform-engineer?tm_job=R-00178055&amp;tm_event=view&amp;tm_company=2502&amp;bid=56">Platform Engineer</a> | Remote; US</p></li><li><p><strong>Reddit</strong> is hiring a <a href="https://job-boards.greenhouse.io/reddit/jobs/7342078">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Weave</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4357505458/">Senior Platform Engineer, Data Infrastructure</a> | Remote; US</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/cognitive-debt-the-hidden-risk-in?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/cognitive-debt-the-hidden-risk-in?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[AI Impact Report: Q1 2026]]></title><description><![CDATA[A quarterly snapshot from 400+ companies integrating AI into daily engineering work]]></description><link>https://newsletter.getdx.com/p/ai-assisted-engineering-q1-2026-impact</link><guid isPermaLink="false">https://newsletter.getdx.com/p/ai-assisted-engineering-q1-2026-impact</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Tue, 14 Apr 2026 10:36:48 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2b5656e4-fb18-4472-9ef7-f8019bfdf45e_2400x1254.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of Engineering Enablement</strong>, a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p><em><strong>&#128197; Live research readout: </strong><a href="https://getdx.com/webinar/ai-and-engineering-velocity-what-the-data-shows/?utm_source=newsletter">Sign up here</a> to join Abi for a readout and interactive discussion about the impact of AI on engineering velocity.</em></p><div><hr></div><p>Since releasing our Q4 2025 AI Impact Report, reasoning models have improved, agentic workflows have gained traction, and teams are finding new ways to integrate AI into how they build.</p><p>Each quarter, we analyze the prior quarter&#8217;s data to benchmark AI adoption rates, productivity impact, and how teams are using these tools. For this edition, we expanded our dataset by over 40% more data points to give you a more representative picture.</p><p>We&#8217;ve compiled this analysis into our AI-Assisted Engineering: Q1 2026 Impact Report, which we are releasing today.</p><p>The full report covers industry adoption rates (now at 93%), tooling benchmarks, and the growing risks of &#8220;Shadow AI.&#8221; You can read the full report <a href="https://getdx.com/report/ai-assisted-engineering-Q1-impact-report/">here</a>. Below are two previews of interesting shifts we&#8217;ve observed since last quarter.</p><h3>Engineering managers are shipping 4x more code than six months ago</h3><p>Perhaps the most dramatic structural shift we observed is the changing nature of the engineering manager role.</p><p>In Q4, we noted that engineering managers who used AI daily were shipping twice as many pull requests over the period of time the sample was taken. That figure has doubled again: engineering managers are now shipping 4x as much code.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QjhS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QjhS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!QjhS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!QjhS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!QjhS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QjhS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png" width="1456" height="925" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:925,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:122569,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/194129088?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QjhS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!QjhS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!QjhS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!QjhS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcbf5e7bb-346d-4044-bd8e-14453bd0caea_3776x2400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Why this matters: </strong>The definition of a developer is expanding. As agentic tools and AI assistants drastically lower the barrier to context-switching and boilerplate generation, managers are finding it far easier to get back into the codebase without sacrificing their leadership responsibilities. Coupled with an industry-wide trend of flattening organizational structures, AI is enabling the true return of the player-coach, allowing EMs to stay technically sharp and contribute directly to the product.</p><p>The question then is: if EMs can meaningfully contribute code again, should org structures evolve to reflect that? From what we&#8217;re hearing, it&#8217;s top of mind for many leaders, but most are still in &#8220;watch and learn&#8221; mode. No one we&#8217;ve spoken to has made structural changes based on this shift yet, but we&#8217;ll continue to monitor this trend.</p><h3>Junior engineers now save more time with AI than Staff+ engineers </h3><p>In our Q4 2025 report, Staff+ engineers who used AI daily were regaining more time than any other tenure band. Just three months later, the data tells a different story.</p><p>Junior engineers who use AI daily are now saving 4.9 hours per week, edging out daily Staff+ users who are saving 4.8 hours.</p><p>While time savings alone doesn&#8217;t tell a full story&#8212;in fact, this shift should come with heightened attention to quality, maintainability, and security&#8212;it is interesting to see this change in work impact among junior engineers. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0dch!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0dch!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!0dch!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!0dch!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!0dch!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0dch!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png" width="1456" height="925" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:925,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:152741,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/194129088?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0dch!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!0dch!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!0dch!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!0dch!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4931fa4-0410-40fa-af60-0c387b77df7a_3776x2400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Why this matters: </strong>Staff+ engineers, who naturally spend more time mentoring and reviewing rather than writing net-new code, are still seeing benefits, but junior engineers have officially taken the lead in total time saved. We provide a look at AI&#8217;s impact on quality in the full report; we&#8217;ll continue to pay close attention to quality measures amongst junior developer cohorts in the future.</p><h3>What else is in the report?</h3><p>The full Q1 2026 report provides comprehensive data and visual benchmarks on:</p><ul><li><p>Rust-specific improvements: How improvements in model reasoning and agentic loops have caused Rust to leapfrog other modern languages in AI-driven time savings.</p></li><li><p>Quality volatility: Why some companies are seeing change failure rates swing by 2% (a 50% increase in defects), and why automated testing is no longer optional.</p></li><li><p>The shadow AI risk: How developers are bypassing enterprise guardrails, and why acceptable use policies are critical for ensuring safe use of AI</p></li><li><p>Small vs. large enterprises: Why companies with fewer than 200 developers are outpacing large enterprises in efficiency gains.</p></li></ul><h3>Final thoughts</h3><p>As always, a reminder: there is no &#8220;average&#8221; experience with AI impact. Looking at industry trends and averages like those in this report can help contextualize your performance and do some pattern-matching, but I want to caution against using these averages as a guide to what your performance should look like. The truth is that AI impact is very asymmetrical and looks very different for each organization. Some are doing very well when it comes to throughput, but quality is suffering; some have done a great job of using AI for migrations, but are struggling to incorporate AI in testing and release processes. Averages abstract away the nuances of each company&#8217;s transformation journey. The best way to use this information is to enhance&#8212;not replace&#8212;your own metrics and AI strategy.</p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>Cashea</strong> is hiring an <a href="https://cashea.na.teamtailor.com/jobs/579773-infrastructure-developer-productivity-platform-engineering-manager">Infrastructure &amp; Developer Productivity Platform Engineering Manager</a> | Remote</p></li><li><p><strong>BNY</strong> is hiring an <a href="https://bnymellon.eightfold.ai/careers/job/30877038?domain=bnymellon.com&amp;hl=en">SVP, Application Development Manager</a> | Pittsburgh, PA</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Leidos</strong> is hiring a <a href="https://careers.leidos.com/jobs/17462787-platform-engineer?tm_job=R-00178055&amp;tm_event=view&amp;tm_company=2502&amp;bid=56">Platform Engineer</a> | Remote; US</p></li><li><p><strong>Reddit</strong> is hiring a <a href="https://job-boards.greenhouse.io/reddit/jobs/7342078">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Weave</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4357505458/">Senior Platform Engineer, Data Infrastructure</a> | Remote; US</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/ai-assisted-engineering-q1-2026-impact?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/ai-assisted-engineering-q1-2026-impact?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[Assumptions as code: SiriusXM’s approach to platform prioritization]]></title><description><![CDATA[Eleanor Millman and Mina Tawadrous from SiriusXM share how their team built a custom, AI-assisted prioritization system centered on developer speed, reliability, cost, and trust.]]></description><link>https://newsletter.getdx.com/p/assumptions-as-code-siriusxms-approach</link><guid isPermaLink="false">https://newsletter.getdx.com/p/assumptions-as-code-siriusxms-approach</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Fri, 10 Apr 2026 15:01:49 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/193487190/cea539f60a582df99e6d637d7c8c7c8e.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/V-3Cv0LUYa4">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this episode, I&#8217;m joined by Eleanor Millman, Senior Staff Product Manager, and Mina Tawadrous, Associate Director of Platform Engineering at SiriusXM, to discuss how platform teams can scale prioritization without relying on revenue.</p><p>We talk through how SiriusXM moved beyond RICE to build a custom framework for internal platforms, using weighted factors like developer speed, reliability, cost, and trust to guide decisions across teams.</p><p>We also explore their concept of &#8220;assumptions as code,&#8221; in which teams store and reuse assumptions in a central repository to reduce misalignment and improve decision-making, with AI helping to surface and validate those assumptions.</p><p>We close with how this system is shaping SiriusXM&#8217;s 2026 prioritization approach and what it signals about a broader shift toward builder-driven product development.</p><div id="youtube2-V-3Cv0LUYa4" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;V-3Cv0LUYa4&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/V-3Cv0LUYa4?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><h4><strong>Prioritization breaks without a shared system</strong></h4><ul><li><p><strong>Prioritization does not scale naturally across teams.</strong> What works for one team breaks down at the org level with multiple stakeholders and competing requests</p></li><li><p><strong>Platform teams lack a clear revenue signal.</strong> Unlike product teams, they must prioritize based on indirect impact</p></li><li><p><strong>A shared framework aligns decisions.</strong> Without it, prioritization defaults to local optimization and noise</p></li></ul><p><strong>RICE is a starting point, not a solution</strong></p><ul><li><p><strong>Standard frameworks miss key dimensions for platform teams.</strong> Urgency and indirect impact are not captured well</p></li><li><p><strong>&#8220;Impact&#8221; needs to be decomposed.</strong> SiriusXM broke it into developer speed, reliability, cost, security, and more</p></li><li><p><strong>The framework must evolve over time.</strong> Iteration was critical to making it useful in practice</p></li></ul><p><strong>Weighting forces real tradeoffs</strong></p><ul><li><p><strong>You cannot prioritize everything at once.</strong> Increasing one dimension (like cost) necessarily deprioritizes others</p></li><li><p><strong>Assigning weights makes decisions explicit.</strong> Leaders must commit to what matters this quarter</p></li><li><p><strong>The output drives alignment across teams.</strong> A single prioritized list reduces cross-team conflicts</p></li></ul><p><strong>Data and conversation work together</strong></p><ul><li><p><strong>The framework creates a place to attach data.</strong> Metrics like reliability scores inform prioritization decisions</p></li><li><p><strong>Disagreements surface quickly.</strong> Teams can see where assumptions or inputs differ</p></li><li><p><strong>Conversations, not just scores, drive alignment.</strong> The value comes from debating inputs, not just ranking outputs</p></li></ul><p><strong>Assumptions are the real bottleneck</strong></p><ul><li><p><strong>Most disagreements come from hidden assumptions.</strong> Teams often believe they are aligned when they are not</p></li><li><p><strong>Assumptions can be conflicting, invisible, or stale.</strong> All three create friction in decision-making</p></li><li><p><strong>Making assumptions explicit improves clarity.</strong> It becomes easier to validate or challenge them</p></li></ul><p><strong>Storing assumptions as code scales learning</strong></p><ul><li><p><strong>Assumptions are stored in a central repository.</strong> User research and data become reusable across teams</p></li><li><p><strong>This reduces duplicated effort.</strong> Teams don&#8217;t need to rediscover the same insights repeatedly</p></li><li><p><strong>It creates a shared source of truth.</strong> Assumptions become visible, versioned, and easier to update</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=77s">01:17</a>) Mina&#8217;s role and path into platform engineering</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=123s">02:03</a>) Eleanor&#8217;s background and shift into product</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=195s">03:15</a>) Scaling prioritization across platform engineering teams</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=341s">05:41</a>) Aligning platform priorities with stakeholders</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=548s">09:08</a>) Evolving RICE into a platform-specific prioritization framework</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=693s">11:33</a>) Iterating on the prioritization framework over time</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=1017s">16:57</a>) How the framework, data, and conversations drive alignment</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=1146s">19:06</a>) Storing assumptions as code in a central repository</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=1607s">26:47</a>) Resolving assumption conflicts with user interviews</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=1847s">30:47</a>) How stored assumptions integrate with AI workflows</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=2130s">35:30</a>) Standard mode and different user personas</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=2240s">37:20</a>) The industry shift towards builders</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=2464s">41:04</a>) The challenges of platform engineering</p><p>(<a href="https://www.youtube.com/watch?v=V-3Cv0LUYa4&amp;t=2616s">43:36</a>) How SiriusXM is prioritizing in 2026</p><h2><strong>Referenced:</strong></h2><p>&#8226; <a href="https://getdx.com/whitepaper/ai-measurement-framework">Measuring AI code assistants and agents</a></p><p>&#8226; <a href="https://www.siriusxm.com/">SiriusXM</a></p><p>&#8226; <a href="https://www.vmware.com/">VMware</a></p><p>&#8226; <a href="https://getdx.com/podcast/siriusxm-revamped-platform-developer-experience/">How SiriusXM revamped their platform and developer experience</a></p><p>&#8226; <a href="https://www.productplan.com/glossary/rice-scoring-model/">RICE Scoring Model | Prioritization Method Overview</a></p><p>&#8226; <a href="https://www.researchgate.net/publication/243992508_The_evaporating_cloud_A_tool_for_resolving_workplace_conflict">The evaporating cloud: A tool for resolving workplace conflict</a></p>]]></content:encoded></item><item><title><![CDATA[Developer ramp-up time continues to accelerate with AI]]></title><description><![CDATA[Time to 10th PR has more than halved since early 2025.]]></description><link>https://newsletter.getdx.com/p/developer-ramp-up-time-continues</link><guid isPermaLink="false">https://newsletter.getdx.com/p/developer-ramp-up-time-continues</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Thu, 09 Apr 2026 10:02:11 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/aa06c9c8-ec59-472c-bf5c-8d6fb44a747f_1000x700.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of Engineering Enablement</strong>, a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p><em><strong>&#128197; Live research readout: </strong><a href="https://getdx.com/webinar/ai-and-engineering-velocity-what-the-data-shows/?utm_source=newsletter">Sign up here</a> to join Abi for a readout and interactive discussion about an upcoming report that looks at the impact of AI on engineering velocity.</em></p><div><hr></div><p>Onboarding new hires has always been an expensive and time-consuming process, and an area where AI has the opportunity to have a meaningful impact. In Q4 2025, when we looked at Time to 10th PR (a measure we use to track ramp-up time), we saw AI already having a dramatic effect. In some companies, Time to 10th PR was cut in half: from 91 days with no AI usage to 49 days with daily AI use.</p><p>We revisit this data quarterly. In our most recent analysis, we&#8217;re seeing that trend continue: onboarding time is faster today than it was in Q4 of last year.</p><p>For this latest cut, we analyzed data from a random sample of 400 companies during the period from October 2025 to February of 2026. We measured the average number of days between a developer&#8217;s start date and their 10th merged PR. We specifically looked at an aggregate of engineers who showed daily use of AI and compared how ramp-up time changed from October 2025 to February 2026. </p><p>The dataset includes large, global organizations with 500+ developers, spanning both tech and non&#8209;tech companies.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qhMb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qhMb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!qhMb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!qhMb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!qhMb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qhMb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png" width="1456" height="925" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:925,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:207369,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/193626808?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qhMb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!qhMb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!qhMb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!qhMb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F62f9f3de-dd18-49af-8083-9da4df9912a7_3776x2400.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>As of April 2026, the current average Time to 10th PR across our dataset is 33 days, down from 39 days in Q4 2025. That represents a further ~15% decrease quarter-over-quarter, and more than a 50% reduction since Q1 2024, before AI usage was widespread. The change this quarter isn&#8217;t as dramatic as the 50% reduction we saw last year, but it is still meaningful. This trend begs the question: will onboarding times continue to fall, or are we approaching a natural floor where many of the remaining steps are not easily compressed by AI? As more agentic solutions are devised for onboarding processes, we may see this figure continue to decline. </p><p>Part of the improvement we&#8217;re seeing since last quarter comes from the fact that AI is no longer an optional, individual choice in many organizations. Developers are guided to use AI to ask questions about the codebase, to explain architectural decisions, and to propose draft changes that senior engineers can review. The result is that the &#8220;figuring things out&#8221; phase is shorter. </p><p>At the same time, Time to 10th PR is a narrow measure. It tells us how quickly someone reaches a specific milestone, but it does not, on its own, tell us about the quality of those changes, the amount of rework they generate, or the depth of understanding new hires have of the systems they&#8217;re touching. That will be something we explore more in future analyses. </p><p>Stay tuned for the full report coming soon. </p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>Awardco</strong> is hiring an <a href="https://www.linkedin.com/jobs/view/4355191071/">Engineering Manager, Cloud Platform</a> | San Francisco, CA</p></li><li><p><strong>BNY</strong> is hiring an <a href="https://bnymellon.eightfold.ai/careers/job/30877038?domain=bnymellon.com&amp;hl=en">SVP, Application Development Manager</a> | Pittsburgh, PA</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Leidos</strong> is hiring a <a href="https://careers.leidos.com/jobs/17462787-platform-engineer?tm_job=R-00178055&amp;tm_event=view&amp;tm_company=2502&amp;bid=56">Platform Engineer</a> | Remote; US</p></li><li><p><strong>Reddit</strong> is hiring a <a href="https://job-boards.greenhouse.io/reddit/jobs/7342078">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Weave</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4357505458/">Senior Platform Engineer, Data Infrastructure</a> | Remote; US</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/developer-ramp-up-time-continues?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/developer-ramp-up-time-continues?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[Measuring AI impact, assessing readiness, and new data trends]]></title><description><![CDATA[How AI is reshaping the entire SDLC, shifting bottlenecks and redefining AI readiness, and why developer experience, not tools, determines real impact.]]></description><link>https://newsletter.getdx.com/p/measuring-ai-impact-assessing-readiness</link><guid isPermaLink="false">https://newsletter.getdx.com/p/measuring-ai-impact-assessing-readiness</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 03 Apr 2026 14:14:08 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/192751704/ce62dd4946731f397d24fa43d6fad6e2.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/yDp3U21X4zc">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this special episode of Engineering Enablement, I welcome back Jesse Adametz, this time as host.</p><p>In our conversation, we explore how AI is showing up across the SDLC, not just in code generation, and how it is shifting bottlenecks across the development process. We unpack what &#8220;AI readiness&#8221; actually means in practice, and why it often comes down to developer experience fundamentals like documentation, environments, and feedback loops.</p><p>We also discuss why enablement matters more than tool choice, how teams are thinking about measuring ROI, and what changes as background agents become more common. Finally, we explore how the role of the engineer may evolve, what questions teams are still trying to answer, and the challenges of non-engineers contributing to codebases.</p><div id="youtube2-yDp3U21X4zc" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;yDp3U21X4zc&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/yDp3U21X4zc?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><h4><strong>AI is expanding beyond coding into the full SDLC</strong></h4><ul><li><p><strong>The focus has shifted from code generation to the entire software lifecycle.</strong> Teams are applying AI to planning, prototyping, review, and documentation&#8212;not just writing code.</p></li></ul><p><strong>AI readiness is a developer experience problem</strong></p><ul><li><p><strong>The biggest blockers to AI adoption are long-standing DX gaps.</strong> Missing documentation, inconsistent environments, weak CI, and unclear system boundaries all limit effectiveness.</p></li><li><p><strong>Tool choice is not the primary driver of success.</strong> Models and tools are evolving too quickly for this to be a durable advantage.</p></li></ul><ul><li><p><strong>Some organizations are formalizing AI enablement as a function.</strong> Dedicated teams are emerging to drive adoption and share practices.</p></li></ul><p><strong>Measuring AI ROI is messy and still evolving</strong></p><ul><li><p><strong>Correlation vs causation makes attribution difficult.</strong> High AI usage often correlates with already high-performing engineers.</p></li><li><p><strong>Longitudinal analysis is more reliable than snapshots.</strong> Tracking changes over time gives better insight into impact.</p></li><li><p><strong>Token spend introduces real cost considerations.</strong> AI creates a direct, variable cost that organizations must evaluate.</p></li></ul><p><strong>AI impact falls into two buckets: amplification and augmentation</strong></p><ul><li><p><strong>Amplification improves human productivity.</strong> This includes higher throughput, time savings, and better developer experience.</p></li><li><p><strong>Augmentation extends capacity beyond humans.</strong> Agents begin to act as additional &#8220;headcount,&#8221; completing work independently.</p></li><li><p><strong>These require different measurement approaches.</strong> Amplification focuses on human output, while augmentation focuses on agent output relative to cost.</p></li></ul><p><strong>Background agents shift how work gets done and where bottlenecks appear</strong></p><ul><li><p><strong>Agents enable work to happen outside the human loop.</strong> Tasks can be completed asynchronously and proactively.</p></li><li><p><strong>This changes the developer role.</strong> Engineers move toward reviewing, guiding, and orchestrating agent output.</p></li><li><p><strong>Human workflows can become the bottleneck.</strong> If agents produce work faster than humans can process it, the constraint shifts.</p></li><li><p><strong>This reframes productivity.</strong> The question becomes where human involvement adds the most value.</p></li></ul><p><strong>Specs and documentation are becoming critical infrastructure</strong></p><ul><li><p><strong>AI makes documentation a core dependency.</strong> It directly impacts the quality of outputs.</p></li><li><p><strong>Poor documentation leads to poor results.</strong> Agents can duplicate systems or make incorrect assumptions without context.</p></li><li><p><strong>Documentation is shifting from optional to essential.</strong> It is now foundational for both human and AI productivity.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=132s">02:12</a>) Where AI is showing up across the SDLC</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=353s">05:53</a>) AI readiness and its link to developer experience</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=503s">08:23</a>) Why enablement, education, and experimentation matter more than tool choice</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=785s">13:05</a>) The case for a dedicated enablement team</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=890s">14:50</a>) Measuring AI ROI: challenges and tradeoffs</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=1186s">19:46</a>) Background agents and token spend</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=1452s">24:12</a>) Measuring agent output with PR throughput</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=1618s">26:58</a>) How the engineer role might change</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=1861s">31:01</a>) Specs and documentation in the age of AI</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=1991s">33:11</a>) Non-engineers writing code</p><p>(<a href="https://www.youtube.com/watch?v=yDp3U21X4zc&amp;t=2130s">35:30</a>) What&#8217;s changing in the SDLC and open questions</p><h2><strong>Referenced:</strong></h2><p>&#8226; <a href="https://getdx.com/whitepaper/ai-measurement-framework">Measuring AI code assistants and agents</a></p><p>&#8226; <a href="https://getdx.com/podcast/jesse-aldametz-twilio-platform-consolidation/">Lessons from Twilio&#8217;s multi-year platform consolidation</a></p><p>&#8226; <a href="https://www.amazon.com/Phoenix-Project-DevOps-Helping-Business/dp/0988262592">The Phoenix Project: A Novel About IT, DevOps, and Helping Your Business Win</a></p><p>&#8226; <a href="https://code.claude.com/docs/en/memory">How Claude remembers your project - Claude Code Docs</a></p><p>&#8226; <a href="https://www.reddit.com/r/ProgrammerHumor/comments/1p70bk8/specisjustcode/#lightbox">specIsJustCode : r/ProgrammerHumor</a></p>]]></content:encoded></item><item><title><![CDATA[AI-generated merged code holds steady at ~30%]]></title><description><![CDATA[A preview from our upcoming Q1 AI Impact Report.]]></description><link>https://newsletter.getdx.com/p/ai-generated-merged-code-holds-steady</link><guid isPermaLink="false">https://newsletter.getdx.com/p/ai-generated-merged-code-holds-steady</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Wed, 01 Apr 2026 10:01:29 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/cab26a66-5e53-4f0d-972a-650308c49646_1000x700.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of Engineering Enablement</strong>, a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p><em><strong>&#128197; Live research readout: </strong><a href="https://getdx.com/webinar/ai-and-engineering-velocity-what-the-data-shows/?utm_source=newsletter">Sign up here</a> to join Abi for a readout and interactive discussion about an upcoming report that looks at the impact of AI on engineering velocity.</em></p><div><hr></div><p>Each quarter, DX publishes data on how AI is being used at 500+ organizations and the impact it&#8217;s having. One metric we&#8217;ve been following is the percentage of code that&#8217;s written by AI&#8212;today&#8217;s newsletter shares a preview of what we&#8217;re seeing.</p><p>Currently, we measure the &#8220;percentage of AI-generated code&#8221; by asking developers directly how much of their merged code they believe is written by AI. This measure is captured quarterly by DX. In Q1 2026, developers at more than 500 organizations reported their average percentage of AI-authored code, along with how frequently they use AI tools (daily, weekly, or monthly). We aggregated these self-reported percentages to calculate the mean share of merged AI-generated code for each usage segment and overall.</p><p>We ran the previous Q4 analysis in the same way. In the future, we will collect this data and compare against a new system-based measure that automatically tracks the percentage of AI-generated code.</p><p>For this quarter, here&#8217;s a preview of what we&#8217;re seeing. Stay tuned for the full report later this month.</p><h3>AI-generated code reaching production sees a slight increase</h3><p>Since last quarter, the average share of merged code authored by AI has moved from 22% to 27.4%. While that&#8217;s directionally up, it&#8217;s not as meaningful of a change as we had expected.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8qAw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8qAw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png 424w, https://substackcdn.com/image/fetch/$s_!8qAw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png 848w, https://substackcdn.com/image/fetch/$s_!8qAw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png 1272w, https://substackcdn.com/image/fetch/$s_!8qAw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8qAw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png" width="1456" height="764" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:764,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:165693,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/192785790?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8qAw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png 424w, https://substackcdn.com/image/fetch/$s_!8qAw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png 848w, https://substackcdn.com/image/fetch/$s_!8qAw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png 1272w, https://substackcdn.com/image/fetch/$s_!8qAw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8aa60a17-4569-4151-b387-da45ce7afd7e_4604x2415.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The stability around 30% could be read two ways. On one hand, it may simply mean that current models and tooling are producing roughly the same proportion of merge-ready code as they were last quarter. However, there&#8217;s reason to believe something else is going on. The second half of 2025&#8212;particularly November and December&#8212;is widely regarded as a turning point for AI-assisted development. Models like Opus 4.5 represented a significant leap in capability, and some have gone so far as to say that anything before November 2025 shouldn&#8217;t even be used as a baseline, given how much changed in a short window.</p><p>If that&#8217;s true, then the more likely explanation for the stability around 30% is that most teams haven&#8217;t yet fully adapted to take advantage of those improvements. New models don&#8217;t automatically translate to more merged AI code; developers still need to update their workflows, build trust in the output, and find the right use cases for more capable tools.</p><p>There&#8217;s some evidence to support this interpretation. The daily users segment saw the largest increase in AI-generated code, moving from 24.1% to 30.8%. Meanwhile, weekly and monthly users, who are less likely to have adjusted their habits, saw smaller increases.</p><h3>Final thoughts</h3><p>This analysis focuses on how much merged code is AI-generated, but it doesn&#8217;t answer questions like whether a higher percentage of AI-generated code is associated with changes in quality. It also begs the question of whether there are certain types of organizations, or groups of developers, that are merging more vs. less AI-generated code.</p><p>We&#8217;ll explore these questions, and others like them, in the full Q1 AI Impact Report.</p><p>Additionally, in future reports we&#8217;ll compare this self-reported measure against a telemetry-based metric that tracks AI-generated code directly from version control systems. This will not only provide another lens on AI&#8217;s impact, but also give us a more granular view of how patterns are changing over time and where those changes are concentrated.</p><p>Stay tuned.</p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>Amplitude</strong> is hiring an <a href="https://www.linkedin.com/jobs/view/4355191071/">Infrastructure Engineer</a> | Lindon, UT</p></li><li><p><strong>Awardco</strong> is hiring an <a href="https://job-boards.greenhouse.io/amplitude/jobs/8208477002">Engineering Manager, Cloud Platform</a> | San Francisco, CA</p></li><li><p><strong>BNY</strong> is hiring an <a href="https://bnymellon.eightfold.ai/careers/job/30877038?domain=bnymellon.com&amp;hl=en">SVP, Application Development Manager</a> | Pittsburgh, PA</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Leidos</strong> is hiring a <a href="https://careers.leidos.com/jobs/17462787-platform-engineer?tm_job=R-00178055&amp;tm_event=view&amp;tm_company=2502&amp;bid=56">Platform Engineer</a> | Remote; US</p></li><li><p><strong>Lob</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4369668388/">Staff Platform Engineer</a> | Remote; US</p></li><li><p><strong>Weave</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4357505458/">Senior Platform Engineer, Data Infrastructure</a> | Remote; US</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/ai-generated-merged-code-holds-steady?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/ai-generated-merged-code-holds-steady?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[How do we interpret AI impact without overclaiming causation?]]></title><description><![CDATA[A better way of framing the impact of AI.]]></description><link>https://newsletter.getdx.com/p/how-do-we-interpret-ai-impact-without</link><guid isPermaLink="false">https://newsletter.getdx.com/p/how-do-we-interpret-ai-impact-without</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Wed, 25 Mar 2026 10:02:45 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c7943ee3-da5e-4172-bd33-fff43561b54b_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of </strong></em><strong>Engineering Enablement,</strong><em> a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Last week I hosted a live AMA about measuring the impact of AI and where organizations are with their rollout today. (Special thanks to <a href="https://www.linkedin.com/in/jesseadametz/">Jesse Adametz</a> for hosting the discussion.) One of the questions that came up had to do with measuring ROI: specifically, how should we think about correlation vs. causation? Leaders want to be careful about whether they should attribute changes in metrics to AI tools.</p><p>I answered the question live&#8212;you can download and watch the <a href="https://getdx.com/webinar/ama-with-abi/?utm_source=newsletter">full recording here</a>&#8212;or read this week&#8217;s newsletter for my take.</p><div><hr></div><p>Many organizations make this mistake: They compare developers who use AI heavily to those who use it less, notice that the heavy users have higher throughput, and then conclude that AI must be the cause. The problem with this kind of analysis&#8212;&#8220;do people who use AI more have higher code throughput?&#8221;&#8212;is that in many cases, the developers who use AI more are the ones who were already coding more in the first place.</p><p>That&#8217;s why most of the time, longitudinal analysis (looking at how things change over time) is more informative. But it&#8217;s also harder to do. You need clean data over a long enough period, and you still have to account for confounding factors. For example, one big confound right now is the heightened pressure in many companies to increase throughput. Leaders are simultaneously pushing for more output and more AI usage. Because of this, some of the increase in throughput is likely due to this pressure itself&#8212;a classic case of Goodhart&#8217;s law, where once a metric becomes a target, it stops being a good measure.</p><p>When thinking about the ROI of AI more broadly, it&#8217;s useful to break it into two buckets:</p><ul><li><p><strong>Amplification: </strong>How much more productive are humans thanks to AI? Here we look at:</p><ul><li><p>Throughput (are engineers shipping more by using AI?)</p></li><li><p>Time saved (how much time do developers feel they&#8217;re saving in specific workflows?)</p></li><li><p>Developer experience scores (are AI tools improving the overall developer experience?) We can then convert improvements in developer experience into time savings, for example using something like a developer experience index.</p></li></ul></li><li><p><strong>Augmentation: </strong>To what extent are you actually extending your engineering capacity by using agents, as if they were additional headcount? One unit we like to use here is human-equivalent hours: how much work are agents delivering, how much would that have taken a human, and then you divide that by how much it cost. Dividing the human-equivalent hours by the cost gives you an agent hourly rate. If that effective rate is low, it indicates a high-ROI place to invest.</p></li></ul><p>This amplification/augmentation framing is also helpful when talking to executives. You can say: to some extent we&#8217;re amplifying our humans, and to some extent we&#8217;re augmenting our workforce with agents.</p><p>Another common and related question that tends to come up in this discussion:<strong> </strong>if we agree that lines of code is a poor metric, what should we use instead to measure AI&#8217;s impact?</p><p>LOC is a noisy metric for a simple reason: A low-effort change can involve many lines of code, and a high-effort change can involve very few lines of code. That was already true before AI, and it becomes even worse with AI-generated code, which tends to produce more lines than a human might. This makes LOC even more inflated and noisy.</p><p>For these reasons&#8212;both pre- and post-AI&#8212;we&#8217;ve preferred metrics like PR throughput. It&#8217;s still imperfect, but it gives a more normalized view of &#8220;change throughput&#8221;: how many atomic changes are we pushing through the system? That makes it a less noisy high-level signal than raw LOC.</p><p>At DX, we&#8217;ve also developed a metric called TrueThroughput, which is a weighted version of PR throughput. It incorporates lines of code as one of several inputs to weight PRs, but LOC is not the sole or primary metric. Even so, all of these metrics are imperfect; you&#8217;re really choosing between different degrees of imperfection. In that space, LOC is significantly less useful than PR throughput as a primary signal. This lines up with broader industry experience as well. Many large tech companies that have studied this problem in depth have also converged on some form of change throughput as their preferred signal for tracking impact, rather than relying on raw lines of code.</p><p>To summarize my perspective on the correlation vs causation question: a simpler way to frame AI&#8217;s ROI is to look at how much it amplifies your existing developers, and how much it augments your organization with agent-driven capacity.</p><p><em>Download and rewatch the full AMA discussion <a href="https://getdx.com/webinar/ama-with-abi/?utm_source=newsletter">here</a>. My response to this specific question starts at 16:50.</em></p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>Amplitude</strong> is hiring an <a href="https://job-boards.greenhouse.io/amplitude/jobs/8208477002">Engineering Manager, Cloud Platform</a> | San Francisco, CA</p></li><li><p><strong>BNY</strong> is hiring an <a href="https://bnymellon.eightfold.ai/careers/job/30877038?domain=bnymellon.com&amp;hl=en">SVP, Application Development Manager</a> | Pittsburgh, PA</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Lob</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4369668388/">Staff Platform Engineer</a> | Remote; US</p></li><li><p><strong>Mastercard</strong> is hiring a <a href="https://mastercard.wd1.myworkdayjobs.com/CorporateCareers/job/OFallon-Missouri/Vice-President--Software-Engineering_R-272919">Vice President, Software Engineering</a> | O&#8217;Fallon, MO; Boston, MA</p></li><li><p><strong>Plaid</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4289332289/">Software Engineer - Platform</a> | New York, NY</p></li><li><p><strong>Vercel </strong>is hiring a <a href="https://vercel.com/careers/dx-engineer-5784254004">DX Engineer</a> | Hybrid (Austin, New York City, San Francisco)</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/how-do-we-interpret-ai-impact-without?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/how-do-we-interpret-ai-impact-without?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Why aren't AI productivity gains higher?]]></title><description><![CDATA[Developers explain why the gains are more modest than expected.]]></description><link>https://newsletter.getdx.com/p/why-arent-ai-productivity-gains-higher</link><guid isPermaLink="false">https://newsletter.getdx.com/p/why-arent-ai-productivity-gains-higher</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Thu, 19 Mar 2026 10:01:31 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c0942bfb-6276-4e6a-82b0-fbf22fefde92_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of </strong></em><strong>Engineering Enablement,</strong><em> a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Last week, we <a href="https://newsletter.getdx.com/p/ai-productivity-gains-are-10-not">shared data</a> showing that the current productivity gains from AI in engineering are more modest than many headlines imply. To better understand this result, we&#8217;ve been interviewing developers across a range of companies to hear what they&#8217;re experiencing, and to dig into the reasons behind these more modest gains.</p><p>This newsletter summarizes the key themes we&#8217;re hearing in those conversations so far, along with direct anonymous quotes from the interviews we performed. This work is ongoing, so if you have something to add, please share your perspective in the comments or on LinkedIn. We&#8217;d love to hear from you.</p><h2>Why are the real&#8209;world gains from AI more modest than many people expect?</h2><p>Developers we spoke with pointed to several reasons. The most common: coding is a relatively small share of how engineers spend their time, so even meaningful acceleration there has limited impact on overall throughput. Beyond that: new bottlenecks are being created, and organizational and social friction hampers AI&#8217;s potential impact.</p><p>Here&#8217;s what we heard.</p><h3>1. Coding isn&#8217;t the main bottleneck</h3><p>The most common theme we heard: AI does accelerate developers&#8217; ability to write code, but coding only represents a percentage of the work engineers actually do. Even if AI cut coding time in half, it would only affect a fraction of the total time spent delivering a feature.</p><ul><li><p><em>A lot of the slowdown isn&#8217;t from our tools, but from either other teams, services or processes.</em></p></li><li><p><em>Many senior devs say: &#8220;writing code is the easy part&#8221;... The bottleneck is the human side: alignment, planning, scoping, code reviews, etc.</em></p></li><li><p><em>AI can help developers write code fast, but writing faster code is not always more productivity. Most developers code only a small proportion of their time. There might be an expectation that this percentage is higher from those who write the headlines, but the reality is that most of the time developers spend is on navigating process and quality, change management, regulation, etc.</em></p></li><li><p><em>The easy tasks are a little easier. The tedious tasks are a little less annoying. Larger/more complicated tasks &#8212; I may be able to shave off hours to a day&#8230; but that&#8217;s not happening every day, and that doesn&#8217;t necessarily mean I ship 3x more PRs. Just means a 4d task can take 3.</em></p></li></ul><h3>2. Partial SDLC automation creates new bottlenecks, so AI gains cap out</h3><p>AI has sped up code generation, but that has created or exposed bottlenecks downstream, particularly in code review and validation. More code is being produced, but the processes for checking that code haven&#8217;t scaled with it, and developers report that the time saved writing is often consumed by the extra scrutiny that AI-generated code requires.</p><ul><li><p><em>AI can very significantly speed up initial engineering time, but often that saved time is spent on extended reviews, fact checking or issue remediation, resulting in net-zero productivity gain.</em></p></li><li><p><em>The opportunities for LLMs to save large amounts of time on new projects is offset by time spent, sometimes ineffectively, attempting to use them on existing projects whose complexity and intricacies lead the LLM to make changes that are at minimum difficult to review even for experienced owners of the code.</em></p></li><li><p><em>People do not have all parts of SDLC AI-enabled, wherever it does not exist becomes a bottleneck (planning, code review, verification). When people are trying to remove those bottlenecks, there are often lack of guardrails that creates misdirection and churned motions.</em></p></li></ul><h3>3. Social friction is a barrier to adoption</h3><p>Polarization between pro- and anti-AI engineers, ambiguity about when and how AI should be used, and the absence of peer champions all slow the rate at which teams develop effective workflows with these tools.</p><ul><li><p><em>There&#8217;s misalignment amongst engineers &#8211; the camps are pretty polarized in terms of pro-AI and anti-AI.  When anti-AI engineers have status and loud voices it is challenging for everyone to know how to interact with ai tooling.</em></p></li><li><p><em>AI tooling isn&#8217;t discussed enough. Engineers are often unsure whether it&#8217;s high status or low status to be talking about using AI, so they just don&#8217;t. And people are reluctant to state their true views.  If people say what they might really believe, they might seem crazy. In many organizations it is not socially acceptable to suggest that we re-evaluate changes to the code review process for example.</em></p></li><li><p><em>Having at least one other teammate who is bought in to using AI tools is essential, being an isolated solo-adopter does not allow you to materialize the gains in a meaningful way.  Software development is a team sport.</em></p></li></ul><h3>4. Tooling and skill gaps are limiting gains</h3><p>These two factors are hard to separate: immature tools make the learning curve steeper, and developers who are early in that curve get less out of the tools. Both are limiting gains, though developers who shared this view generally see it as a temporary problem.</p><p>On the skill side, using AI effectively is its own discipline. On the tooling side, AI assistants don&#8217;t always slot cleanly into existing developer workflows. The integrations that would make them feel native to how teams already work aren&#8217;t there yet, and getting agents to operate autonomously on real infrastructure remains an unsolved problem for many.</p><p>On skill gaps:</p><ul><li><p><em>Someone just starting out with these tools is not as effective as someone who has incorporated them into their workflows for much longer.  I have 1000+ hours of purely agentic coding experience and I have so much to learn.  Many beginners may have just a few hours or tens of hours. Learning to agentically code is a skill.</em></p></li><li><p><em>Most of the task isn&#8217;t the actual coding, it&#8217;s articulating a fuzzy problem in a clear, easily understandable way.</em></p></li><li><p><em>This is more short term, but the pace of change in the tooling (e.g., the emergence of skills) means I&#8217;m spending a fair amount of time getting to know how to use the tools appropriately; which usually means going down various rabbit holes to discover how NOT to use them. That creates some (temporary?) friction that means I might fall back to my own skills when there&#8217;s production pressures.</em></p></li></ul><p><em><strong>Note:</strong> Leaders can distribute DX&#8217;s <a href="https://getdx.com/guide/prompting-guide-for-ai-assisted-engineering/?utm_source=newsletter">prompting guide</a> and <a href="https://getdx.com/guide/advanced-prompting-guide-for-ai-assisted-engineering/?utm_source=newsletter">advanced prompting guide</a> internally with developers as a way to help them build a foundational practice with AI.</em></p><p>On tool maturity:</p><ul><li><p><em>The tools aren&#8217;t mature enough to fit into existing systems cleanly, and most developers haven&#8217;t figured out how to use them well yet. Not because they&#8217;re bad engineers &#8212; the workflow is just genuinely new and nobody&#8217;s handed them a playbook.</em></p></li><li><p><em>The difference between frontier models and models even a few months back can be significant.</em></p></li><li><p><em>Most developers are used to working inside well-defined tools &#8212; your IDE, your terminal, your CI pipeline. Everything has a clear place and a clear trigger. AI assistants don&#8217;t work that way&#8230; So you end up with higher adoption among developers who naturally gravitate toward that kind of open-ended tooling, and much lower adoption among developers who just want something that fits cleanly into what they&#8217;re already doing.</em></p></li></ul><h3>5. Most AI tools lack important context</h3><p>AI tools perform well on problems that are self-contained and well-documented. But most real engineering work isn&#8217;t like that. The context that matters&#8212;why a system was designed a certain way, what the implicit rules are, what business constraints apply&#8212;typically isn&#8217;t written down. It lives in people&#8217;s heads. Until that knowledge is made explicit and accessible, AI will keep hitting a ceiling on the kinds of problems it can reliably help with.</p><ul><li><p><em>The bigger issue I keep running into is that most codebases and systems aren&#8217;t set up for AI to actually help. Not an architecture problem &#8212; more that the knowledge of how things work lives in people&#8217;s heads. Why this service behaves this way, what the implicit contract between these two systems is, why that design decision got made three years ago &#8212; none of that is written down anywhere. An AI assistant can&#8217;t reason over a Slack thread from an archived channel or the mental model of the engineer who built it. Until that context gets surfaced into something concrete, you&#8217;re always going to hit a ceiling on what the tooling can do.</em></p></li><li><p><em>AI doesn&#8217;t have the context I do on making those decisions, as LLMs are built of massive stats, not direct knowledge of the niche ends of what I&#8217;m accomplishing.</em></p></li><li><p><em>I would (anecdotally) put forth that a lot of my work with AI is trying to get it setup to do the work with proper guardrails. It&#8217;s not safe to let run loose on our code-bases or production facing infrastructure as its not deterministic.</em></p></li><li><p><em>AI models are generally good at understanding what we are trying to achieve, especially when they have access to the workspace or codebase. They can often provide useful solutions for specific use cases, particularly when the problem is related to foundational aspects of a technology or programming language. However, what they often lack is a deeper understanding of the business context and the various factors that influence business logic. That broader context is important to fully understand the bigger picture of a project. In some cases, this limitation can be mitigated by carefully crafting prompts, but when the business logic becomes complex or the prompts become too large for relatively simple tasks, people may step back and revert to more traditional approaches.</em></p></li></ul><h2>Other observations about the current state of AI tools</h2><p>While speaking with developers, we heard a few other interesting themes about what they&#8217;re currently experiencing.</p><p><strong>1. Documentation robustness is improving, but its quality is yet to be determined.</strong></p><ul><li><p><em>The biggest quality gain I&#8217;m seeing is in improved documentation. Developers hate writing docs, so it will always be the last thing they tackle, or the first to get cut... if they tackle it at all. At least for me, I found our documentation getting much more robust thanks to AI being able to take that task off our plate. And it&#8217;s not just great for the developers to have access to these docs. Giving the AI agents access to it provides a quicker context boost to know how to tackle the next project.</em></p></li><li><p><em>Using it to write code documentation leads to terribly incorrect information, an active detriment to future work, creating more work. This is really painful. It&#8217;s just not its forte.</em></p></li></ul><p><strong>2. AI is especially powerful in greenfield projects.</strong></p><ul><li><p><em>Could you whip up an app from scratch 3x faster with AI? I would believe that. And if I had to just guess, that&#8217;s where I think a lot of the hype comes from. Someone says &#8220;look I created this thing from scratch in 1 day!&#8221; But, I think it is known that AI isn&#8217;t quite as proficient when working on a mature, large codebase...with legacy code and lots of nuance. Which is where most ICs are living in day to day, I think.</em></p></li><li><p><em>While multi-tasking on my main efforts, I was able to build an entire new microservice from scratch to tests an idea and provide a proof of concept of a new paradigm (this has massive comprehension debt, I haven&#8217;t even looked at the code, but the service does what I expect it to).</em></p></li><li><p><em>It has helped a lot with doing things like fast PoCs to check if something will work.</em></p></li></ul><p><strong>3. There&#8217;s concern that developers are offloading critical thinking.</strong></p><ul><li><p><em>I would say I&#8217;m losing muscle memory on how to do things (e.g. terminal commands like checking git history for when a file changed, checking what version of a dependency is installed, etc.). I&#8217;ve grown dependent on the AI to remember the right commands for me.</em></p></li><li><p><em>I think if someone blindly uses AI for everything, he/she slowly forgets day to day things.</em></p></li><li><p><em>AI helps speed up some tasks but may result in folks offloading their thinking. Particularly amongst junior developers, the &#8216;improved initial speed&#8217; can come at the expense of skill development. Overall speed improvement is negligible.</em></p></li><li><p><em>I have also found myself having to explain to more junior engineers why what AI said about a problem isn&#8217;t the correct answer for us in a particular situation. AI empowers more junior engineers to work longer without asking engineering questions of the more senior folks. That is a sharp double-edged sword that I am still learning how to work with when mentoring people.</em></p></li><li><p><em>[For me, AI&#8217;s] handiness is augmented by me knowing when it&#8217;s wrong due to my engineering experience.</em></p></li></ul><h2>Final thoughts</h2><p>Two takeaways stand out to me from these conversations. First, the gains from AI code assistants are structurally limited by how much of the job is actually coding. <a href="https://arxiv.org/pdf/2502.15287">Microsoft&#8217;s research</a> puts coding at around 16% of a developer&#8217;s time. Even dramatic acceleration to code generation can only move the overall needle so much.</p><p>Second, it&#8217;s the human aspects of the software development lifecycle that are the biggest bottlenecks. Handoffs between teams, the skill required to effectively prompt and manage agents, and the review burden that AI-generated code creates&#8212;these are where time is going, and where the next wave of productivity gains may come from.</p><p>For leaders, this raises a couple of important questions. Which parts of the SDLC beyond coding are candidates for AI assistance? And are the human conditions for capturing AI&#8217;s upside in place?</p><p>The 10% figure is not the ceiling. But closing the gap will likely require focusing on the parts of the job that AI hasn&#8217;t touched yet.</p><p><em>If you have something to add, please share in the comments or on LinkedIn.</em></p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>Amplitude</strong> is hiring an <a href="https://job-boards.greenhouse.io/amplitude/jobs/8208477002">Engineering Manager, Cloud Platform</a> | San Francisco, CA</p></li><li><p><strong>BNY</strong> is hiring an <a href="https://bnymellon.eightfold.ai/careers/job/30877038?domain=bnymellon.com&amp;hl=en">SVP, Application Development Manager</a> | Pittsburgh, PA</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Lob</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4369668388/">Staff Platform Engineer</a> | Remote; US</p></li><li><p><strong>Plaid</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4289332289/">Software Engineer - Platform</a> | New York, NY</p></li><li><p><strong>Vercel </strong>is hiring a <a href="https://vercel.com/careers/dx-engineer-5784254004">DX Engineer</a> | Hybrid (Austin, New York City, San Francisco)</p></li><li><p><strong>Zillow</strong> is hiring a <a href="https://zillow.wd5.myworkdayjobs.com/Zillow_Group_External/job/Remote-USA/Senior-Product-Manager--Developer-Experience_P749450">Senior Product Manager, Developer Experience</a> | Remote; US</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/why-arent-ai-productivity-gains-higher?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/why-arent-ai-productivity-gains-higher?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[AI productivity gains are 10%, not 10x]]></title><description><![CDATA[Preliminary data from our longitudinal AI impact study]]></description><link>https://newsletter.getdx.com/p/ai-productivity-gains-are-10-not</link><guid isPermaLink="false">https://newsletter.getdx.com/p/ai-productivity-gains-are-10-not</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Wed, 11 Mar 2026 10:03:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c0dbdda3-eedb-4e2f-a272-16b972208ef0_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the latest issue of Engineering Enablement, a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p>&#128467; Join Abi on March 19th for a live Q&amp;A session. He will address some of the more pressing questions we&#8217;ve received around measuring AI impact, the impact of tool choice, and more. <strong><a href="https://getdx.com/webinar/ama-with-abi/?utm_source=newsletter">Register here.</a></strong></p><div><hr></div><p><em>Edit: We've updated the data in this post to reflect the latest results from our <a href="https://getdx.com/report/ai-and-engineering-velocity-a-longitudinal-analysis/">longitudinal study</a>.</em></p><div><hr></div><p>Social media and vendor marketing have set high expectations for AI, suggesting as much as 2-3x productivity gains. But from the data we&#8217;re seeing, the reality on the ground is far more modest.</p><p>At DX, we&#8217;re currently conducting a longitudinal study to measure the long-term impact of AI adoption on key engineering productivity metrics. As part of this study, we analyzed data from a random sample out of 400 companies between November 2024 through February 2026 to track whether teams are shipping more pull requests as AI adoption increases.</p><p>We found that, during this time, AI usage increased significantly&#8212;by an average 65%. However, PR throughput only increased by 7.76%.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DtzB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DtzB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png 424w, https://substackcdn.com/image/fetch/$s_!DtzB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png 848w, https://substackcdn.com/image/fetch/$s_!DtzB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png 1272w, https://substackcdn.com/image/fetch/$s_!DtzB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DtzB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png" width="1456" height="559" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:559,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:45220,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/190563086?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DtzB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png 424w, https://substackcdn.com/image/fetch/$s_!DtzB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png 848w, https://substackcdn.com/image/fetch/$s_!DtzB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png 1272w, https://substackcdn.com/image/fetch/$s_!DtzB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc264a5ed-cab1-42dd-8be0-71f55c4628dd_1920x737.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Note: This figure is particularly robust because we&#8217;ve filtered out potential gamification effects by excluding teams that set PR throughput targets for individual engineers, which could drive metric inflation rather than genuine output.</em></p><h3>What this means for leaders</h3><p>A ~10% gain is consistent with what we&#8217;re hearing from engineering leaders more broadly: most organizations are landing in the 5&#8211;15% range. It is a real improvement, but it&#8217;s a long way from the 2&#8211;3x gains many executives and boards have come to expect. AI is moving the needle, but leaders may need to reset expectations internally.</p><h3>Why gains aren&#8217;t higher</h3><p>To understand what&#8217;s driving this, we spoke with developers across several of these organizations. The explanation we heard most consistently: writing code was never the bottleneck.</p><p>As one senior developer put it: &#8220;The easy tasks are a little easier. The tedious tasks are a little less annoying. A four-day task might take three. But that doesn&#8217;t mean I&#8217;m shipping 3x more PRs.&#8221;</p><p>AI may be accelerating the coding portion of the job. But coding represents a relatively small slice of how engineers actually spend their time. Planning, alignment, scoping, code review, and handoffs&#8212;the human parts of the SDLC&#8212;remain largely untouched.</p><h3>What&#8217;s next</h3><p>We&#8217;re continuing to investigate the long-term effects of AI in engineering teams. The full study will explore why some teams are capturing more of the upside than others, and what leaders can do to close that gap. More to come.</p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>Amplitude</strong> is hiring an <a href="https://job-boards.greenhouse.io/amplitude/jobs/8208477002">Engineering Manager, Cloud Platform</a> | San Francisco, CA</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Plaid</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4289332289/">Software Engineer - Platform</a> | New York, NY</p></li><li><p><strong>UserTesting</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4363545775">Software Engineer, Developer Experience (Platform)</a> | Spain</p></li><li><p><strong>Vercel </strong>is hiring a <a href="https://vercel.com/careers/dx-engineer-5784254004">DX Engineer</a> | Hybrid (Austin, New York City, San Francisco)</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/ai-productivity-gains-are-10-not?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/ai-productivity-gains-are-10-not?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[Developer experience at scale – lessons from Dropbox]]></title><description><![CDATA[How they treat developer productivity as a sociotechnical problem and weave AI into the fabric of their engineering culture.]]></description><link>https://newsletter.getdx.com/p/developer-experience-at-scale-lessons</link><guid isPermaLink="false">https://newsletter.getdx.com/p/developer-experience-at-scale-lessons</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Wed, 25 Feb 2026 11:02:37 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5ff99548-df0e-416d-8633-d15805d0ed2b_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Welcome to the latest issue of Engineering Enablement, a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p>&#128467; Next month, you can join me for a live Q&amp;A session. I&#8217;ll address some of the more pressing questions we&#8217;ve received around measuring AI impact, the impact of tool choice, and more. <strong><a href="https://getdx.com/webinar/ama-with-abi/?utm_source=newsletter">Register here.</a></strong></p><div><hr></div><p>This week on the Engineering Enablement podcast, we were joined by <a href="https://www.linkedin.com/in/unamasivayam/">Uma Namasivayam</a>, Senior Director of Engineering Productivity at Dropbox. Uma&#8217;s team owns engineering productivity across Dropbox&#8217;s roughly 1,000 engineers: everything from CI/CD systems and telemetry infrastructure to the company&#8217;s AI tooling rollout.</p><p>During the conversation, Uma shared specifics on how Dropbox drove AI adoption from one-third of engineers to three-quarters, why they chose to deploy multiple AI coding tools rather than starting by standardizing on one, and how they ended up building their own internal AI platform.</p><p>Below is a lightly edited excerpt from this conversation. You can listen to <strong><a href="https://getdx.com/podcast/developer-experience-dropbox/?utm_source=newsletter">the full episode here.</a></strong></p><div><hr></div><h3>A lot of engineering orgs treat developer productivity as an engineering problem. You have a different take on that.</h3><p><strong>Uma:</strong> I think of productivity at Dropbox&#8212;or anywhere, really&#8212;as a sociotechnical problem. There absolutely has to be a strong investment in the technology itself: improving the reliability of your systems, improving speed, and reducing friction in your tooling. But then there&#8217;s the element of collaboration, of culture, of working with leadership, with developers themselves, and with the people team.</p><p>A concrete example: one of the dimensions of developer productivity is deep work&#8212;can developers actually code uninterrupted? That&#8217;s not necessarily an engineering problem. Working with our chief people officer, we had to literally think about how to restructure meeting times, how to carve out focus blocks for employees. That required a completely different set of partners than fixing a slow CI pipeline. That&#8217;s why having a common language across all of those groups, and bringing them together around that language, was so important. You have to attack productivity from multiple different angles.</p><h3>How did the rollout of AI coding tools intersect with the DevEx work that was already underway?</h3><p><strong>Uma: </strong>I think of them as two parallel work streams that have a lot of overlap in the middle. DevEx is about incremental, systematic friction reduction. It requires aligning with leadership, defining the problems clearly, and making steady progress. AI is about speed. It&#8217;s about getting the best tools into developers&#8217; hands quickly, experimenting fast, and staying on the cutting edge.</p><p>Before AI can really deliver on its promise, the foundational systems&#8212;build and test, telemetry, production observability&#8212;have to be in a strong place. Developers need to trust that when they push code through an AI-assisted workflow, the quality guardrails are actually there. Without that trust, you can&#8217;t go really hard on AI.</p><h3>How did you actually get adoption moving? What drove the initial uptake?</h3><p><strong>Uma:</strong> Early on, roughly one-third of our engineers were using AI tools organically, so people were just finding tools they liked on their own. That&#8217;s a decent starting point, but it&#8217;s not a strategy. The real inflection point came when our exec team made AI a clear company priority. Within about three months of that top-down signal, we got to around three-quarters of engineers using tools on a weekly basis.</p><p>With that said, top-down mandates only take you so far. After that initial push, we had to go deeper, starting by looking at what was actually blocking adoption in different parts of the engineering population, and addressing those pockets specifically. That&#8217;s where a product mindset comes in: understanding your customers&#8217; actual pain points rather than assuming one solution works for everyone.</p><h3>You offer developers multiple AI coding tools rather than standardizing on one. What&#8217;s the thinking there?</h3><p><strong>Uma:</strong> It comes back to treating this like a product problem. Different teams have genuinely different needs. Our mobile developers, for example, couldn&#8217;t use the tools that work well for other parts of the codebase. We had to find something specific to their use case. So we deliberately chose not to be a single-tool shop.</p><p>We also learned quickly that some tools that work at a smaller scale fall apart at Dropbox&#8217;s scale. We were piloting an AI code review tool and it just didn&#8217;t hold up. That pushed us toward building some of this in-house. We now have an internal platform that handles the backend complexity specific to Dropbox&#8217;s monorepo, and that other teams can build on. The build-versus-buy decision turns out to be really critical when you&#8217;re operating at this scale and at this speed.</p><p>One practical thing that also helped: we worked with our procurement and legal and security teams to dramatically reduce the time it takes to evaluate and approve new AI tools, getting that review process down to around three days. When the market is moving this fast, your ability to experiment is only as good as how quickly you can get tools in front of developers.</p><h3>Once you&#8217;ve expanded beyond code completion, what does AI usage actually look like across the SDLC?</h3><p><strong>Uma:</strong> Code completion was the obvious starting point. Once we felt we&#8217;d gotten what we could from that, we started looking at the rest of the development lifecycle: code review, testing, and debugging. Every stage of the SDLC is on the table.</p><p>What we discovered is that a lot of the commercially available tools for these adjacent use cases didn&#8217;t hold up at our scale or for our specific codebase. That&#8217;s what led us to build in-house. One of our developers took it upon himself to figure out what kind of platform we could build, using Claude and Claude Code as the foundation, that would work within Dropbox&#8217;s environment and that others could build on top of. That platform now handles the backend complexity: deployment, monorepo scale, and testing guardrails. If a team wants to build an AI-assisted code review product, they start from that platform rather than from scratch. It&#8217;s one of the things I&#8217;m most proud of from the past year.</p><h3>What&#8217;s the hardest unsolved problem you&#8217;re carrying into 2026?</h3><p><strong>Uma: </strong>Connecting developer productivity improvements to actual business outcomes. We can show that DXI improved, that AI adoption is up, and that developers are saving hours. But the arc from &#8220;developers are more productive&#8221; to &#8220;we shipped more value to customers faster&#8221;&#8212;that instrumentation isn&#8217;t there yet, and I don&#8217;t think anyone in the industry has fully cracked it.</p><p>What we&#8217;re seeing is that the capacity unlocked by AI is naturally flowing toward migrations and tech debt reduction. That&#8217;s actually pretty cool&#8212;give engineers more capacity, and they automatically invest it in the right things. But as a leader, I need to be able to answer the CFO and the exec team: where is this capacity going, and how does it connect to revenue? We&#8217;re working toward that. If anyone listening has cracked that code, I&#8217;d genuinely love to talk.</p><p><em>You can listen to the full conversation with Uma on the <a href="https://getdx.com/podcast/developer-experience-dropbox/?utm_source=newsletter">Engineering Enablement podcast.</a></em></p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>American Express</strong> is hiring a <a href="https://aexp.eightfold.ai/careers/job/37813133?hl=en">Sr. Manager, Digital Product Management - DevProd</a> | Hybrid - London UK</p></li><li><p><strong>CoreWeave </strong>is hiring a <a href="https://coreweave.com/careers/job?gh_jid=4453307006&amp;board=coreweave">Sr. Software Engineer - Developer Experience</a> | Livingston NJ; New York, NY</p></li><li><p><strong>DoorDash </strong>is hiring an <a href="https://job-boards.greenhouse.io/doordashusa/jobs/7436813?gh_src=6bvvo3y11us">Engineering Manager - Developer Experience</a> | San Francisco, CA; Sunnyvale, CA; Seattle, WA; Los Angeles, CA; New York, New York</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Plaid</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4289332289/">Software Engineer - Platform</a> | New York, NY</p></li><li><p><strong>UserTesting</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4363545775">Software Engineer, Developer Experience (Platform)</a> | Spain</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/developer-experience-at-scale-lessons?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/developer-experience-at-scale-lessons?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Analyst reactions: How AI is reshaping engineering organizations]]></title><description><![CDATA[Analysts at Accenture and RedMonk are seeing an increased focus on quality, security, and responsible AI rollout.]]></description><link>https://newsletter.getdx.com/p/analyst-reactions-how-ai-is-reshaping</link><guid isPermaLink="false">https://newsletter.getdx.com/p/analyst-reactions-how-ai-is-reshaping</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Wed, 18 Feb 2026 11:01:10 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ed7d8b45-a42e-49fa-8720-7ccc12bd3947_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of </strong></em><strong>Engineering Enablement</strong><em><strong>,</strong> a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p>&#128467; Next month, you can join us for a live Q&amp;A session with DX CEO Abi Noda. He&#8217;ll address recent questions around measuring AI impact, the impact of tool choice, and more. <strong><a href="https://getdx.com/webinar/ama-with-abi/?utm_source=newsletter">Register to join here</a></strong><a href="https://getdx.com/webinar/ama-with-abi/?utm_source=newsletter">.</a></p><div><hr></div><p>In just the past year, AI has set into motion several major shifts in how engineering organizations are structured, including how teams are composed, where responsibilities sit, and what skills matter. These changes are still unfolding, but patterns are emerging.</p><p>To help name some of those shifts, I interviewed two analysts who spend their time talking with engineering leaders across industries: <a href="https://www.linkedin.com/in/ruchi-goyal-ca/">Ruchi Goyal</a>, Global AI Practice Leader at Accenture, and <a href="https://www.linkedin.com/in/kateholterhoff/">Dr. Kate Holterhoff</a>, Senior Industry Analyst at RedMonk. Below I&#8217;ve summarized the observations they shared&#8212;with the hope that naming these patterns is a useful reference point for leaders to recognize and navigate the same changes in their own organizations.</p><h3><strong>Hiring priorities are emphasizing AI fluency and code quality</strong></h3><p>The software industry is still working through a hiring slump, but the vacuum is being filled by specific AI-centric roles.</p><ul><li><p><strong>The &#8220;AI Engineer&#8221; trend:</strong> We are seeing a surge in &#8220;AI Engineer&#8221; titles on LinkedIn. These are roles that are less about training foundational models and more about building the glue between LLMs and production code. (See example roles on Indeed <a href="https://www.indeed.com/q-artificial-intelligence-engineer-jobs.html?vjk=c45b0ad972436645">here</a>).</p></li><li><p><strong>Quality in focus:</strong> At the same time, a new skill is becoming central to engineering hiring: the ability to distinguish good code from mediocre AI-generated code. As <a href="https://x.com/rough__sea/status/2013280952370573666">Ryan Dahl</a> and others have noted, AI removes the barrier to writing code, but it introduces a new one: dealing with slop.</p></li><li><p>Because a focus on quality has become more important, some organizations are <strong>seeing friction emerge in the code review process.</strong></p></li></ul><h3><strong>Fragmented AI experiments are consolidating into centralized platforms</strong></h3><p>Early AI adoption in most organizations was scattered, with individual teams running their own experiments with different tools and no shared standards. That&#8217;s changing. Enterprises are moving toward centralized models such as a formal AI Center of Excellence or a hub-and-spoke structure that sets organization-wide direction. Goyal notes that &#8220;a lot of clients are looking into merging Developer Productivity and Internal Tools teams into one.&#8221; This can enable a number of new strategies for the organization:</p><ul><li><p><strong>Consolidation:</strong> Organizations are folding AI initiatives into their internal developer platforms, creating governed paths for how AI gets used across the organization. Without that structure, teams may end up making independent decisions about tools and access, which can be hard to unwind later.</p></li><li><p><strong>Orchestration:</strong> Managing individual agents is becoming less of the challenge, and coordinating multiple agents working together is where the complexity now lives. Organizations that have already worked through DevOps maturity will recognize the pattern: the evolution tends to move from automation, to orchestration, to choreography. AI adoption is following a similar arc.</p></li><li><p><strong>AI is spreading</strong> across engineering organizations to the point where most developers are expected to use it as a matter of course. As Holterhoff put it, &#8220;AI is becoming water.&#8221; As that happens, Platform teams are becoming increasingly responsible for the governance and prompt hardening that makes this safe.</p></li></ul><h3><strong>A new operational layer is emerging: LLMOps</strong></h3><p>As AI becomes part of the engineering infrastructure, someone has to own the operational layer that keeps it running reliably. In many organizations, that responsibility is coalescing into what&#8217;s being called LLMOps. Many Platform and DevProd teams are absorbing this responsibility.</p><ul><li><p><strong>Prompt engineering to guideposts:</strong> Ad hoc prompting works for individuals but doesn&#8217;t scale across a team or organization. Turning prompts into reusable templates means you&#8217;re building institutional knowledge rather than leaving every developer to figure it out independently, and it creates a consistent quality baseline across the SDLC.</p></li><li><p><strong>The SWAT Team approach:</strong> In many orgs, a specialized team of engineers handles model integration and AI infra, often leveraging the heavy lifting provided by big cloud providers. Organizations that stand up a dedicated group for this work move faster and make fewer costly mistakes, while the rest of the engineering org can focus on building.</p></li><li><p><strong>Deployment convergence:</strong> We are seeing tighter integration between development environments and production. (For example, <a href="https://expo.dev/blog/from-idea-to-app-with-replit-and-expo">Expo&#8217;s integration with Replit</a> allows developers to deploy directly to production, bypassing traditional friction points.) This creates real efficiency gains, but it also means the guardrails that used to exist between those environments need to be rethought, which is something Platform teams are becoming increasingly responsible for.</p></li></ul><h3><strong>The human factor still matters the most</strong></h3><p>Despite the automation, the primary challenges are still cultural. As RedMonk has famously noted, &#8220;The developers are everything.&#8221; AI tools are being designed to support practitioners, not replace them.</p><ul><li><p><strong>Murky ROI:</strong> Individual developers see the value of AI tools clearly enough that many are willing to pay for them personally. But at the organizational level, the picture is murkier. Leaders are measuring the impact of AI while reading headlines claiming 3x productivity gains&#8230; and most aren&#8217;t seeing numbers that match.</p></li><li><p><strong>Security as a skill:</strong> A major skill gap isn&#8217;t learning to prompt; it&#8217;s learning how to use AI securely<strong>.</strong> Training now focuses heavily on privacy, vetting specific LLMs, and ensuring that agentic workflows don&#8217;t bypass critical security guardrails.</p></li><li><p><strong>Software engineers as self-optimizers:</strong> Developers are naturally inclined to try new tools. That&#8217;s generally a good thing, but it means organizations need vetted, safe environments for that experimentation to happen in. The risk is that developers could adopt AI in ways the organization can&#8217;t see or govern.</p></li></ul><p>These changes are moving quickly, and most organizations are still early in figuring out what they mean in practice. There&#8217;s no single right answer yet, which is part of what makes it valuable to hear from analysts with visibility across many organizations. A special thanks to Ruchi Goyal and Dr. Kate Holterhoff for sharing their perspective for this newsletter.</p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>American Express</strong> is hiring a <a href="https://aexp.eightfold.ai/careers/job/37813133?hl=en">Sr. Manager, Digital Product Management - DevProd</a> | Hybrid - London UK</p></li><li><p><strong>CoreWeave </strong>is hiring a <a href="https://coreweave.com/careers/job?gh_jid=4453307006&amp;board=coreweave">Sr. Software Engineer - Developer Experience</a> | Livingston NJ; New York, NY</p></li><li><p><strong>DoorDash </strong>is hiring an <a href="https://job-boards.greenhouse.io/doordashusa/jobs/7436813?gh_src=6bvvo3y11us">Engineering Manager - Developer Experience</a> | San Francisco, CA; Sunnyvale, CA; Seattle, WA; Los Angeles, CA; New York, New York</p></li><li><p><strong>Figma</strong> is hiring a <a href="https://job-boards.greenhouse.io/figma/jobs/5790627004?gh_jid=5790627004&amp;gh_src=db0ijm3x4us">Staff Software Engineer, Developer Experience</a> | Remote; US</p></li><li><p><strong>Plaid</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4289332289/">Software Engineer - Platform</a> | New York, NY</p></li><li><p><strong>UserTesting</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4363545775">Software Engineer, Developer Experience (Platform)</a> | Spain</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/analyst-reactions-how-ai-is-reshaping?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/analyst-reactions-how-ai-is-reshaping?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[Scaling developer experience across 1,000 engineers at Dropbox]]></title><description><![CDATA[Listen now | I talk with Uma Namasivayam of Dropbox about treating developer productivity as a business problem, running developer experience like a product, and building foundations that make AI useful at scale.]]></description><link>https://newsletter.getdx.com/p/scaling-developer-experience-across</link><guid isPermaLink="false">https://newsletter.getdx.com/p/scaling-developer-experience-across</guid><dc:creator><![CDATA[Laura Tacho]]></dc:creator><pubDate>Fri, 06 Feb 2026 18:40:25 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/186996290/472b31ff92db4afe2b7d541e11bae71a.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/ZCg2k-w6o2o">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>Developer productivity is often treated as a tooling problem or a sentiment problem. In reality, it&#8217;s neither. It&#8217;s a socio-technical systems problem that spans engineering foundations, leadership alignment, organizational design, and culture.</p><p>In this episode, I&#8217;m joined by Uma Namasivayam, Senior Director, Engineering Productivity at Dropbox, to explore how Dropbox approaches developer experience at scale. We talk about why productivity needs to be framed as a business problem, how executive alignment creates the conditions for meaningful change, and what it takes to treat developer experience as a real product with developers as customers.</p><p>We also dig into Dropbox&#8217;s approach to AI adoption. Uma shares why strong foundations, such as build, test, and observability, are prerequisites for AI to actually accelerate work, how Dropbox encourages daily AI use without mandating a single tool, and where build-versus-buy decisions break down at scale.</p><p>We close with an honest look at what remains unsolved: how to connect gains in developer productivity and AI-driven capacity to real business outcomes, and where engineering leaders should focus next in 2026.</p><div id="youtube2-ZCg2k-w6o2o" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;ZCg2k-w6o2o&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/ZCg2k-w6o2o?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><h4><strong>Developer productivity is a socio-technical problem</strong></h4><ul><li><p><strong>Productivity cannot be solved through tooling alone</strong>; it spans engineering systems, leadership behavior, organizational structure, and people practices.</p></li><li><p><strong>Problems like build and test are engineering problems</strong>, while <strong>problems like focus time and interruptions are people problems</strong>, and both matter equally.</p></li><li><p><strong>Treating productivity as a system forces tradeoffs to be explicit</strong>, rather than hidden inside isolated tooling initiatives.</p></li></ul><h4><strong>Executive alignment matters more than any single metric</strong></h4><ul><li><p><strong>Top-down sponsorship creates permission to act</strong>, especially when productivity work cuts across org boundaries.</p></li><li><p><strong>A shared framework creates alignment, not answers</strong>; its value is giving leaders and engineers a common language.</p></li><li><p><strong>System metrics matter more than single metrics</strong>, because productivity improvements rarely move one dimension in isolation.</p></li><li><p><strong>Distributed accountability makes productivity a company problem</strong>, not a developer experience team problem.</p></li></ul><h4><strong>Developer experience works best when treated as a product discipline</strong></h4><ul><li><p><strong>Developers are customers</strong>, and their experience must be understood through both qualitative feedback and quantitative signals.</p></li><li><p><strong>Good system metrics do not guarantee good developer experience</strong>, which is why sentiment and perception matter.</p></li><li><p><strong>DX surveys surface where systems break differently for different teams</strong>, such as desktop, mobile, and web developers.</p></li><li><p><strong>Continuous feedback loops are essential</strong>, combining surveys, direct conversations, and usage data.</p></li><li><p><strong>Internal communication is part of the product</strong>, reinforcing to developers that their feedback leads to real change.</p></li></ul><h4><strong>Prioritization requires structure, not intuition</strong></h4><ul><li><p><strong>Finite capacity makes prioritization unavoidable</strong>, even in large, well-resourced engineering orgs.</p></li><li><p><strong>Segmenting developer populations clarifies tradeoffs</strong>, since different teams experience different bottlenecks.</p></li><li><p><strong>DX survey data provides a defensible starting point</strong>, but prioritization still requires judgment.</p></li><li><p><strong>Leadership-level stack ranking helps resolve conflicts</strong>, especially when multiple teams compete for attention.</p></li><li><p><strong>Frameworks make hard decisions easier to explain</strong>, even when they do not make them easy.</p></li></ul><h4><strong>AI and developer experience must advance in parallel</strong></h4><ul><li><p><strong>AI accelerates work, while developer experience reduces friction</strong>, and both are required for sustained gains.</p></li><li><p><strong>Foundational systems act as plumbing</strong>, enabling trust in speed, quality, and safety.</p></li><li><p><strong>Without strong CI, testing, and observability</strong>, faster code creation increases risk instead of value.</p></li><li><p><strong>Trust in guardrails enables confidence in AI-assisted development</strong>, especially at scale.</p></li></ul><h4><strong>AI adoption succeeds through choice, not mandates</strong></h4><ul><li><p><strong>Early organic adoption revealed real developer needs</strong>, rather than forcing a single tool.</p></li><li><p><strong>Different teams require different AI tools</strong>, particularly for mobile, desktop, and large-repo workflows.</p></li><li><p><strong>Supporting multiple tools increased adoption</strong>, rather than reducing it.</p></li><li><p><strong>Daily use depends on fitting AI into existing workflows</strong>, not adding extra steps.</p></li><li><p><strong>Habits matter more than access</strong>, which is why SDLC-level integration is critical.</p></li></ul><h4><strong>Build vs. buy decisions change at scale</strong></h4><ul><li><p><strong>Many AI tools fail when tested at large-company scale</strong>, despite working well in smaller contexts.</p></li><li><p><strong>Cost and performance become gating factors</strong>, not feature completeness.</p></li><li><p><strong>Internal platforms can abstract complexity</strong>, enabling teams to build AI workflows safely and consistently.</p></li><li><p><strong>Shared internal platforms unlock reuse</strong>, allowing teams to innovate without rebuilding infrastructure.</p></li><li><p><strong>Speed of iteration remains the primary differentiator</strong>, even when building in-house.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=45s">00:45</a>) Dropbox&#8217;s engineering org</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=119s">01:59</a>) Why developer productivity is a business problem</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=248s">04:08</a>) The role of executive sponsorship in developer productivity</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=362s">06:02</a>) How DX&#8217;s Core Four framework created a shared language</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=493s">08:13</a>) Treating developer experience as a product</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=690s">11:30</a>) How Dropbox prioritizes developer experience work</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=860s">14:20</a>) The challenge of tying developer experience to business outcomes</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=998s">16:38</a>) How AI and developer experience intersect at Dropbox</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1115s">18:35</a>) The prerequisites for AI adoption to accelerate work</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1226s">20:26</a>) How Dropbox encourages daily AI use</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1392s">23:12</a>) AI use beyond code completion</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1500s">25:00</a>) Managing AI tool demand at scale</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1676s">27:56</a>) Early results from Dropbox&#8217;s AI efforts</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1805s">30:05</a>) Progress on developer experience at Dropbox</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1975s">32:55</a>) Advice for organizations investing in developer experience</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=2065s">34:25</a>) Capacity tradeoffs for developer experience</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=2159s">35:59</a>) The unanswered questions around AI and capacity in 2026</p><h2><strong>Referenced:</strong></h2><p>&#8226; <a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p><p>&#8226; <a href="https://www.dropbox.com/">Dropbox.com</a></p>]]></content:encoded></item><item><title><![CDATA[Advanced prompting guide for AI-assisted engineering]]></title><description><![CDATA[Structured prompting patterns and use cases for complex, high-impact engineering work.]]></description><link>https://newsletter.getdx.com/p/advanced-prompting-guide-for-ai-assisted</link><guid isPermaLink="false">https://newsletter.getdx.com/p/advanced-prompting-guide-for-ai-assisted</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Wed, 04 Feb 2026 11:02:27 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5c4e0e20-0ef3-4008-9140-c68fcfd57c12_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of </strong></em><strong>Engineering Enablement</strong><em><strong>,</strong> a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>In 2025, we saw engineering leaders focus on rolling out AI coding assistants at scale across their organizations. As these tools became more widely used, it became clear that outcomes depended less on having access to AI and more on how teams were educated and enabled. In response, DX published the <a href="https://getdx.com/guide/ai-assisted-engineering/?utm_source=newsletter">Guide to AI Assisted Engineering</a>, outlining best practices and high-value prompting use cases to help engineering teams use AI effectively in their day-to-day work.</p><p>Now, as organizations move beyond pilots, the focus has shifted from adoption and enablement to operational improvements and more complex use cases. Successfully applying AI in these contexts requires more structured prompting practices than those used during early experimentation. To support that next step, we&#8217;ve created our first supplement to the original guide: <strong><a href="https://getdx.com/guide/advanced-prompting-guide-for-ai-assisted-engineering/?utm_source=newsletter">Advanced Prompting Guide for AI Engineering.</a></strong></p><p>This new guide follows the same format as the original, with clear Do and Don&#8217;t scenarios, full prompt examples, and code output examples. It is vendor-agnostic, with an emphasis on prompting structure, constraints, and context so the techniques can be applied across tools and architectures.<br><br>Inside, you&#8217;ll find prompt and code examples that focus on:</p><ul><li><p><strong>Complexity management</strong> - For systems with cascading rules or conflicting requirements, the guide demonstrates graph-based prompting to reveal hidden dependencies, prioritize rules, and deal with changing state</p></li><li><p><strong>Governance and quality</strong> - Workflows that execute controlled validation loops, which result in higher accuracy, and can deal with more edge cases</p></li><li><p><strong>Risk mitigation</strong> - Dual-implementation strategies can yield more bulletproof outcomes, especially when dealing with critical transactions that require 100% accuracy</p></li><li><p><strong>Operational efficiency</strong> - Techniques like diff-only refactoring can reduce invasive changes to large, complex code repositories, as well as reduce tokens</p></li></ul><p>These use cases are drawn from interviews, educational talks, and community interaction, and can be deployed across multiple scenarios. Whether you&#8217;re using coding assistants, building prompts for agents, or writing specs for spec-driven-development, you&#8217;ll find applicable methods in the guide.</p><p><strong>One other important update: </strong>When the original guide was published, it was written primarily for developers. But 2025 was a pivotal year for elevating traditional non-builders. As highlighted in our <a href="https://getdx.com/report/ai-assisted-engineering-q4-impact-report/">Q4 AI Impact Report</a>, engineering leaders are shipping more code, and designers and PMs have the ability to create deeper designs and prototypes. We still encourage engineering leaders to distribute this guide to their engineering teams, but this guide need not be exclusive to engineers. Whether engineer, designer, PM, or leader, if you are working on complex problems, this guide can provide useful perspectives.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://getdx.com/guide/advanced-prompting-guide-for-ai-assisted-engineering/?utm_source=newsletter&quot;,&quot;text&quot;:&quot;Download the guide&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://getdx.com/guide/advanced-prompting-guide-for-ai-assisted-engineering/?utm_source=newsletter"><span>Download the guide</span></a></p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>American Express</strong> is hiring a <a href="https://aexp.eightfold.ai/careers/job/37813133?hl=en">Sr. Manager, Digital Product Management - DevProd</a> | Hybrid - London UK</p></li><li><p><strong>CoreWeave </strong>is hiring a <a href="https://coreweave.com/careers/job?gh_jid=4453307006&amp;board=coreweave">Sr. Software Engineer - Developer Experience</a> | Livingston NJ; New York, NY</p></li><li><p><strong>DoorDash </strong>is hiring an <a href="https://job-boards.greenhouse.io/doordashusa/jobs/7436813?gh_src=6bvvo3y11us">Engineering Manager - Developer Experience</a> | San Francisco, CA; Sunnyvale, CA; Seattle, WA; Los Angeles, CA; New York, New York</p></li><li><p><strong>Experian</strong> is hiring a <a href="https://jobs.smartrecruiters.com/Experian/744000102103785-software-engineering-manager-security-platform-remote">Software Engineering Manager - Security Platform</a> | Remote</p></li><li><p><strong>Gusto</strong> is hiring a <a href="https://job-boards.greenhouse.io/gusto/jobs/7413640">Sr. Platform Engineer</a> | Denver, CO; San Francisco, CA; Atlanta, GA; Austin, TX; Chicago, IL; Miami, FL; Seattle, WA</p></li><li><p><strong>Notion</strong> is hiring a <a href="https://ziina.notion.site/Senior-Platform-Engineer-130a023031e880838abbefe40689f37f">Senior Platform Engineer</a> | Dubai, United Arab Emirates</p></li><li><p><strong>Plaid</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4289332289/">Software Engineer - Platform</a> | New York, NY</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/advanced-prompting-guide-for-ai-assisted?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/advanced-prompting-guide-for-ai-assisted?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[AI tooling benchmarks: PR throughput and usage by tool (Q1 2026)]]></title><description><![CDATA[Updated data from a sample of 64,680 developers across 219 companies.]]></description><link>https://newsletter.getdx.com/p/ai-tooling-benchmarks-pr-throughput</link><guid isPermaLink="false">https://newsletter.getdx.com/p/ai-tooling-benchmarks-pr-throughput</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Wed, 21 Jan 2026 11:01:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b72d74ee-f6a1-4db9-a468-bc5d92942ced_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of </strong></em><strong>Engineering Enablement,</strong><em><strong> </strong>a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p>&#128467;&#65039; We recently announced DX Annual, our flagship conference for developer productivity leaders navigating the AI era. <a href="https://dxannual.com/?utm_source=newsletter">Go here</a> to learn about the event and request an invite to attend.</p><div><hr></div><p>One of the insights from the <a href="https://getdx.com/report/ai-assisted-engineering-q4-impact-report/?utm_source=newsletter">AI impact report</a> released last quarter was that newer AI-native tools appeared to be outperforming others (see page 14). Even in side-by-side deployments, AI-native tools like agentic IDEs were associated with higher throughput compared to older or less specialized solutions.</p><p>Given how quickly the space is evolving, we decided to revisit this data to see whether those patterns still hold. In this latest analysis, we looked at PR throughput and median adoption rate across a slightly larger sample of companies (219, versus the 170 we examined last quarter). Here&#8217;s an updated look at what we&#8217;re seeing:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_bw1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_bw1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png 424w, https://substackcdn.com/image/fetch/$s_!_bw1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png 848w, https://substackcdn.com/image/fetch/$s_!_bw1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png 1272w, https://substackcdn.com/image/fetch/$s_!_bw1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_bw1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png" width="1456" height="926" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:926,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_bw1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png 424w, https://substackcdn.com/image/fetch/$s_!_bw1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png 848w, https://substackcdn.com/image/fetch/$s_!_bw1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png 1272w, https://substackcdn.com/image/fetch/$s_!_bw1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe1dd7b9d-0b70-4d42-8436-bcbd33cb3e53_2048x1302.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Comparing our latest Q1 2026 data to our Q4 2025 report we observe a performance jump consistently across all tools. This indicates that trends from last quarter&#8217;s data are continuing, and that stronger performance is being realized as tool maturity and adoption practice improves.</p><p>It&#8217;s notable that the throughput rankings themselves have not changed over the last three months, with newer, agentic tools still ranking highest in terms of velocity impact. Regardless of overall rank, we observed velocity increases across every tool, with a few major shifts since Q4 2025:</p><ul><li><p>Cursor has seen the largest increase in PR throughput. In Q4 2025, daily users merged a median of 2.8 PRs. Today, that number has jumped to 4.1, representing a 46% increase in throughput for frequent users.</p></li><li><p>Claude leads for weekly and monthly users, with PR throughput exceeding 4.0. This is a significant move from the 2.6 (weekly) and 2.2 (monthly) reported just one quarter ago.</p></li><li><p>GitHub Copilot daily users have also accelerated, moving from 2.5 PRs/week in Q4 to 3.61 today.</p></li><li><p>Tabnine continues to show the lowest throughput at 1.83 for daily users. As noted in the previous report, this is likely because Tabnine is common in large enterprises where PR throughput is lower overall due to organizational complexity.</p></li></ul><p>The last quarter of 2025 has been a landmark period for AI, with broader industry education and best practice converging with improved models, like Opus 4.5 and GPT-5.2, and better tool workflows such as those found in Cursor&#8217;s Agent Mode and Claude Code&#8217;s spec-driven focus.</p><p>Despite rapid advancement and innovation, many unsolved problems remain. Limitations such as memory and context constraints provide technical hurdles, while lack of organizational alignment on policies and practices continue to drive cultural challenges. With all this opportunity for improvement, I expect to see deliberate advancements continue, and productivity gains rise even further in the coming months.</p><p>While we were analyzing PR throughput data associated with different AI tools, we also looked at which tools are becoming daily drivers. (Note: adoption metrics don&#8217;t directly reveal productivity impact, as is mentioned in the <a href="https://getdx.com/research/measuring-ai-code-assistants-and-agents/?utm_source=newsletter">DX AI Measurement Framework</a>. But adoption metrics can give us a signal into which tools are being used for daily tasks versus those used only for specialized, less regular tasks.)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HBGk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HBGk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png 424w, https://substackcdn.com/image/fetch/$s_!HBGk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png 848w, https://substackcdn.com/image/fetch/$s_!HBGk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png 1272w, https://substackcdn.com/image/fetch/$s_!HBGk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HBGk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png" width="1456" height="926" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:926,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HBGk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png 424w, https://substackcdn.com/image/fetch/$s_!HBGk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png 848w, https://substackcdn.com/image/fetch/$s_!HBGk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png 1272w, https://substackcdn.com/image/fetch/$s_!HBGk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fac941efd-aa98-4329-b862-ae539acf765b_2048x1302.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><ul><li><p>GitHub Copilot remains the leader in overall stickiness, with the highest daily adoption rate at 9.76%. Its integration into the existing workflow may be helping it become a constant companion for developers. Additionally, for many businesses, Copilot is easy to procure since an existing purchasing agreement with Microsoft is likely to already be in place.</p></li><li><p>Cursor is rapidly becoming a primary workspace, boasting an impressive 31.56% adoption rate among weekly users.</p></li><li><p>Windsurf and Tabnine show a longer-tail adoption pattern. Windsurf, in particular, has the highest monthly adoption at 35.87%, suggesting it is being used as a powerful specialized tool for specific complex tasks rather than a total IDE replacement for most.</p></li></ul><h1>Summary</h1><p>The data provides evidence for the belief that organizations will realize greater productivity gains as developers become more familiar with AI tools. Further, the significant jump we see in throughput shows how quickly the space is advancing. It&#8217;ll be fascinating to watch and see whether we see even higher throughput numbers next quarter, or gains will start to level over time.</p><p>I&#8217;ll conclude with a few reminders on how to use this data:</p><ol><li><p>Correlate usage with throughput: Don&#8217;t just measure seat count. Ensure that your highest-cost tools are being used by your Daily or Weekly cohorts where the throughput gains are most visible.</p></li><li><p>Avoid vendor lock-in: Because the performance of these tools is evolving so quickly, we recommend a multi-vendor approach. Different tools excel at different frequencies (e.g., Copilot for daily assistance, Claude for weekly deep-dives). See our <a href="https://getdx.com/webinar/running-data-driven-evaluation-of-ai-tools-in-engineering/">webinar</a> on data-driven evaluations of AI tools for more information.</p></li><li><p>Measure to confirm ROI: Use these benchmarks to establish a baseline for your own organization. While AI metrics tell you what is happening, your core metrics, like PR throughput, confirm whether the investment is actually working.</p></li></ol><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>American Express</strong> is hiring a <a href="https://aexp.eightfold.ai/careers/job/37813133?hl=en">Sr. Manager, Digital Product Management - DevProd</a> | Hybrid - London UK</p></li><li><p><strong>CoreWeave </strong>is hiring a <a href="https://coreweave.com/careers/job?gh_jid=4453307006&amp;board=coreweave">Sr. Software Engineer - Developer Experience</a> | Livingston NJ; New York, NY</p></li><li><p><strong>DoorDash </strong>is hiring an <a href="https://job-boards.greenhouse.io/doordashusa/jobs/7436813?gh_src=6bvvo3y11us">Engineering Manager - Developer Experience</a> | San Francisco, CA; Sunnyvale, CA; Seattle, WA; Los Angeles, CA; New York, New York</p></li><li><p><strong>Experian</strong> is hiring a <a href="https://jobs.smartrecruiters.com/Experian/744000102103785-software-engineering-manager-security-platform-remote">Software Engineering Manager - Security Platform</a> | Remote</p></li><li><p><strong>Gusto</strong> is hiring a <a href="https://job-boards.greenhouse.io/gusto/jobs/7413640">Sr. Platform Engineer</a> | Denver, CO; San Francisco, CA; Atlanta, GA; Austin, TX; Chicago, IL; Miami, FL; Seattle, WA</p></li><li><p><strong>Notion</strong> is hiring a <a href="https://ziina.notion.site/Senior-Platform-Engineer-130a023031e880838abbefe40689f37f">Senior Platform Engineer</a> | Dubai, United Arab Emirates</p></li><li><p><strong>Plaid</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4289332289/">Software Engineer - Platform</a> | New York, NY</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/ai-tooling-benchmarks-pr-throughput?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/ai-tooling-benchmarks-pr-throughput?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[DevProd headcount benchmarks, Q1 2026]]></title><description><![CDATA[Research across multiple industry verticals and engineering team sizes.]]></description><link>https://newsletter.getdx.com/p/devprod-headcount-benchmarks-q1-2026</link><guid isPermaLink="false">https://newsletter.getdx.com/p/devprod-headcount-benchmarks-q1-2026</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Wed, 07 Jan 2026 11:02:27 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a13642eb-9195-4410-86c9-4816acd06e75_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome back to </strong></em><strong>Engineering Enablement</strong><em><strong>,</strong> the weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p><strong>&#128467;&#65039; In case you missed it:</strong> we recently announced DX Annual, our flagship conference for developer productivity leaders navigating the AI era. <a href="https://dxannual.com/?utm_source=newsletter">Go here</a> to learn more and request an invite to attend.</p><div><hr></div><p>Business leaders understand the ability of centralized developer productivity and internal platform teams to promote better throughput and faster time to market for software features. But, how large should these teams be? What&#8217;s the best proportion of developer productivity specialists to other engineers within the company? </p><p>To find an answer, our team analyzed data from 39 companies with varying engineering team sizes across a number of industry verticals and looked for trends in organizational composition for teams explicitly dedicated to developer productivity and experience. We found that engineering leaders typically dedicate 2% to 6% of total headcount to centralized developer productivity. However, this ratio is not linear: as organizations scale beyond 1,000 engineers, the percentage of dedicated headcount steadily decreases due to tooling leverage and automation.</p><p>I&#8217;m excited to continue following these trends as AI efforts become more centralized and embedded in developer productivity teams, as we may see it lead to further hiring and expansion of responsibilities. </p><h3>Defining the developer productivity umbrella</h3><p>To ensure these benchmarks weren&#8217;t artificially inflated, we used a narrow definition of what counts as a team focused strictly on developer productivity. We only looked at teams that are unambiguously dedicated to internal platforms, developer productivity, and developer experience, and didn&#8217;t include teams focused on SRE or infrastructure, for example. </p><p>Specifically, we included the following teams in our analysis: </p><ul><li><p><strong>Developer Experience and Productivity:</strong> Teams explicitly named &#8220;Developer Experience,&#8221; &#8220;DevEx,&#8221; &#8220;Developer Productivity,&#8221; &#8220;Developer Enablement,&#8221; or &#8220;Engineering Enablement&#8221;</p></li><li><p><strong>Engineering Effectiveness:</strong> Teams focused on &#8220;Engineering Productivity,&#8221; &#8220;Engineering Effectiveness,&#8221; or &#8220;Engineering Excellence&#8221;</p></li><li><p><strong>Internal Developer Platform Engineers:</strong> Groups building &#8220;Internal Developer Platforms&#8221;, &#8220;Internal Developer Portals,&#8221; &#8220;Developer Frameworks,&#8221; &#8220;Documentation Systems,&#8221; &#8220;Service Catalogs&#8221; or &#8220;Developer Tools&#8221;</p></li><li><p><strong>Build &amp; Release Infrastructure:</strong> &#8220;CI Infrastructure,&#8221; &#8220;CI Platform,&#8221; &#8220;Build Infrastructure,&#8221; &#8220;Build Systems,&#8221; &#8220;Core Automation Platforms&#8221; focused on CI/CD, and &#8220;Release Engineering&#8221; teams</p></li><li><p><strong>Developer Education and Support:</strong> &#8220;Developer Education,&#8221; &#8220;Engineering Support Organization,&#8221; and developer onboarding teams</p></li><li><p><strong>Test Infrastructure:</strong> Teams dedicated to &#8220;Test Infrastructure&#8221; and &#8220;Test Frameworks&#8221; (not QA or product testing)</p></li></ul><p>We excluded several categories that often get conflated with developer tooling, such as:</p><ul><li><p><strong>General Cloud Infrastructure:</strong> Cloud operations, data center management, storage infrastructure, and network infrastructure teams</p></li><li><p><strong>Business Platform Teams:</strong> &#8220;Data Platform,&#8221; &#8220;ML Platform,&#8221; &#8220;Analytics Platform,&#8221; &#8220;Payment Platform,&#8221; &#8220;Content Platform,&#8221; and other product or customer-facing platforms</p></li><li><p><strong>Site Reliability Engineering (SRE): </strong>While SRE teams are critical, they focus on production reliability rather than developer tooling (except when explicitly named as developer-facing)</p></li><li><p><strong>Security Infrastructure: </strong>Security operations, compliance, and infrastructure security teams</p></li><li><p><strong>Product Infrastructure: </strong>Teams building infrastructure for customer-facing products rather than internal tools</p></li><li><p><strong>Lab Infrastructure: </strong>Physical and virtual infrastructure for product testing and validation</p></li></ul><h2>Most companies dedicate 4.7% of engineering headcount to developer productivity</h2><p>In looking at our first sample of teams that have less than 1,000 engineers, we found that most teams dedicate between 2-6% of their overall engineering headcount to centralized developer productivity functions, with an average of 4.7% across the board. Some range past 8% on the high end, and others range below 2%. </p><p>For the average team then, per engineer, that means roughly one productivity engineer per seventeen developers at the high range, or one for every fifty engineers at the lower range. The distribution is shown in the chart below.</p><p>These numbers are not surprising, given the variance between company cultures, repository/build architectures, and general attitudes towards productivity. DORA&#8217;s State of DevOps reports suggest 10-20% investment between centralized teams and embedded productivity work &#8211; our study focuses on centralized, named roles. In general, we&#8217;ll find higher investment amongst engineering teams with more complicated deployments (microservices, polyglot stacks, monorepo), and lower investment amongst teams that focus mostly on CI/CD with more monolithic architectures.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IjA5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IjA5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png 424w, https://substackcdn.com/image/fetch/$s_!IjA5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png 848w, https://substackcdn.com/image/fetch/$s_!IjA5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!IjA5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IjA5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:218257,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/183699394?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IjA5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png 424w, https://substackcdn.com/image/fetch/$s_!IjA5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png 848w, https://substackcdn.com/image/fetch/$s_!IjA5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!IjA5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2c2d78ac-6ada-444a-bb7c-9b5a6c95e7e2_3600x2400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>When we looked at headcount ratios, we also analyzed team structure: even if two companies have 20 productivity engineers, they may organize them differently. Most organizations distribute their developer productivity team members across <strong>2 to 6 distinct teams.</strong> However, some companies take a specialist model, spinning up as many as <strong>15 specialized teams</strong> to tackle narrow problems like test optimization or internal portals, while others prefer a single, centralized developer experience or developer productivity unit that handles all enablement tasks under one roof.</p><h3>Benchmarks by company size and industry vertical</h3><p>While the 2&#8211;6% range is a benchmark, the data shows that investment doesn&#8217;t scale linearly. As engineering organizations grow beyond 1,000 people, the percentage of dedicated DevProd headcount starts to drop. This is likely due to diminishing returns in investment, as teams grow and scale, single developer productivity resources can take advantage of more automation, knowledge sharing, tooling, and teamwide collaboration to reduce the need for headcount.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KNnN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KNnN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!KNnN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!KNnN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!KNnN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KNnN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png" width="1456" height="925" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:925,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:151287,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/183699394?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KNnN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!KNnN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!KNnN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!KNnN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff419fa41-82d9-4d5c-a6da-067826ba35f0_3776x2400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>We also saw clear differences in how different industries value DevProd investment:</h3><ul><li><p><strong>Technology (4.89%):</strong> These companies lead the pack, averaging about one productivity specialist for every 20 engineers. Because software is their primary profit center, shipping features faster directly impacts their bottom line.</p></li><li><p><strong>Fintech and Financial Services (4.36%):</strong> These organizations follow closely behind as they continue to modernize their infrastructure and prioritize engineering speed.</p></li><li><p><strong>Retail (3.8%) and Large Enterprise (3.32%):</strong> These sectors typically show lower ratios, though the latter (large enterprises) may have more to do with the diminishing returns noted in the trends of organizations with 1000+ engineers.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KJbC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KJbC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!KJbC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!KJbC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!KJbC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KJbC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png" width="1456" height="925" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:925,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:164201,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/183699394?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KJbC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png 424w, https://substackcdn.com/image/fetch/$s_!KJbC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png 848w, https://substackcdn.com/image/fetch/$s_!KJbC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png 1272w, https://substackcdn.com/image/fetch/$s_!KJbC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc0f15fcd-103f-444d-b9de-a013418112ea_3776x2400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>How to use these numbers</h2><p>These numbers shouldn&#8217;t be used as a strict quota, but as a guide for your organizational strategy. The &#8220;right&#8221; team size is one that balances these industry averages with the specific bottlenecks your developers are facing.</p><p>If you are building a new &#8220;center of excellence,&#8221; use these ratios to set your initial headcount. To ensure you&#8217;re getting a real return on that investment, combine these benchmarks with a measurement framework like the <a href="https://getdx.com/research/measuring-developer-productivity-with-the-dx-core-4/?utm_source=newsletter">Core 4</a> to track whether your new team is actually removing friction and improving the developer&#8217;s day.</p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>American Express</strong> is hiring a <a href="https://aexp.eightfold.ai/careers/job/37813133?hl=en">Sr. Manager, Digital Product Management - DevProd</a> | Hybrid - London UK</p></li><li><p><strong>Capital One</strong> is hiring a <a href="https://www.capitalonecareers.com/job/mclean/manager-product-management-developer-experience/1732/86033033056">Product Manager - Developer Experience</a> | Plano TX; McLean VA; Richmond VA</p></li><li><p><strong>Gusto</strong> is hiring a <a href="https://job-boards.greenhouse.io/gusto/jobs/7413640">Sr. Platform Engineer</a> | Denver, CO; San Francisco, CA; Atlanta, GA; Austin, TX; Chicago, IL; Miami, FL; Seattle, WA</p></li><li><p><strong>Plaid</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4289332289/">Software Engineer - Platform</a> | New York, NY</p></li><li><p><strong>Reddit</strong>: <a href="https://job-boards.greenhouse.io/reddit/jobs/7342078">Staff Software Engineer - Developer Experience</a> | Remote - United States</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/devprod-headcount-benchmarks-q1-2026?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/devprod-headcount-benchmarks-q1-2026?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[AI and productivity: A year-in-review with Microsoft, Google, and GitHub researchers]]></title><description><![CDATA[Listen now | Listen and watch now on YouTube, Apple, and Spotify.]]></description><link>https://newsletter.getdx.com/p/ai-and-productivity-a-year-in-review</link><guid isPermaLink="false">https://newsletter.getdx.com/p/ai-and-productivity-a-year-in-review</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Mon, 29 Dec 2025 17:17:45 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/182642831/a8424bdccac94a35590d7bc3484b6193.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/xZjvYMuAJPc">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>As we close out 2025, I wanted to step back and take stock of what we have actually learned about AI adoption in engineering organizations. Not just where usage has increased, but where impact is real, where it is overstated, and what questions remain unanswered.</p><p>In this special year-end episode, I&#8217;m joined by Brian Houck from Microsoft, Collin Green and Ciera Jaspan from Google, and Eirini Kalliamvakou from GitHub. Together, we unpack the research each of them worked on this year and explore how leading organizations are thinking about AI measurement, developer experience, and long-term productivity. We talk candidly about why measuring AI&#8217;s impact is so difficult, why familiar metrics like lines of code keep resurfacing despite their flaws, and how multidimensional approaches like SPACE and DORA offer a more realistic lens.</p><p>We also look ahead to 2026. We discuss how AI is beginning to reshape the identity of the developer, how junior engineers&#8217; skill sets may evolve, where agentic workflows are gaining traction, and why some of the most widely shared AI studies were misunderstood. This episode is an honest conversation about moving past hype and toward a more grounded, evidence-based approach to AI adoption in engineering teams.</p><div id="youtube2-xZjvYMuAJPc" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;xZjvYMuAJPc&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/xZjvYMuAJPc?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><h4><strong>Measuring AI impact requires multiple lenses</strong></h4><ul><li><p><strong>There is no single metric that can capture AI&#8217;s impact.</strong> Developer productivity and experience are inherently multidimensional, requiring trade-offs to be evaluated across speed, quality, collaboration, and meaning.</p></li><li><p><strong>Frameworks like SPACE and DORA help avoid metric tunnel vision.</strong> They encourage teams to examine complementary signals rather than optimizing one dimension at the expense of others.</p></li><li><p><strong>Measurement must reflect systems, not tools.</strong> AI does not operate in isolation; its impact depends on organizational context, workflows, and existing engineering practices.</p></li></ul><h4><strong>Why familiar metrics keep failing us</strong></h4><ul><li><p><strong>Lines of code remains a deeply misleading metric.</strong> AI tends to generate verbose code, making raw output a poor proxy for productivity, quality, or long-term maintainability.</p></li><li><p><strong>More code does not equal better outcomes.</strong> Excess code can increase maintenance burden, technical debt, and cognitive load over time.</p></li><li><p><strong>Easy-to-measure metrics are often the most dangerous.</strong> Their simplicity makes them attractive during periods of uncertainty, even when they obscure what is actually changing.</p></li></ul><h4><strong>The limits of tracking AI-generated code</strong></h4><ul><li><p><strong>Measuring the percentage of AI-generated code oversimplifies reality.</strong> AI may write, delete, refactor, or reorganize code in ways that raw percentages fail to capture.<br><strong>AI-generated code does not inherently signal higher risk.</strong> In some contexts, AI output may be more consistent or higher quality than human-written code.</p></li><li><p><strong>These metrics are better used as supporting signals, not goals.</strong> They can inform budgeting, experimentation, or adoption patterns but should not drive performance targets.</p></li></ul><h4><strong>How AI is reshaping the role of the developer</strong></h4><ul><li><p><strong>Developers are shifting from implementers to orchestrators.</strong> Advanced AI users spend more time framing problems, setting context, and validating outcomes than writing raw code.</p></li><li><p><strong>AI fluency is becoming a core skill.</strong> Knowing how to guide, correct, and collaborate with agents is increasingly important.</p></li><li><p><strong>Adoption follows a progression.</strong> Developers tend to move from skepticism to exploration, collaboration, and eventually strategic use as expectations recalibrate.</p></li></ul><h4><strong>What this means for junior engineers</strong></h4><ul><li><p><strong>Skill development may accelerate rather than disappear.</strong> Junior engineers may practice delegation, planning, and system-level thinking earlier by working with AI agents.</p></li><li><p><strong>Technical fundamentals still matter.</strong> Understanding architecture, requirements, and failure modes remains essential for supervising AI-generated work.</p></li><li><p><strong>Interpersonal skills risk being deprioritized.</strong> Managing agents is not the same as managing people, raising concerns about how collaboration skills develop over time.</p></li></ul><h4><strong>AI is not just a productivity tool</strong></h4><ul><li><p><strong>Creativity and innovation benefit from friction.</strong> Research suggests that exposing decision points and seams can create space for new ideas rather than faster repetition.</p></li><li><p><strong>Automating everything is not always desirable.</strong> Removing all toil may reduce opportunities for learning, insight, and creative problem-solving.</p></li><li><p><strong>AI should augment thinking, not replace it.</strong> Tools that surface trade-offs and choices can support better outcomes than those that simply eliminate effort.</p></li></ul><h4><strong>High-leverage AI use cases focus on toil</strong></h4><ul><li><p><strong>Developers spend only about 14% of their time writing code.</strong> Optimizing coding alone rarely leads to large productivity gains.</p></li><li><p><strong>The biggest opportunities lie in removing friction.</strong> Documentation, compliance tasks, incident response, flaky tests, and knowledge discovery consistently rank as top pain points.</p></li><li><p><strong>AI excels at work developers dislike but must still do.</strong> Automating dull, repetitive tasks can improve satisfaction and free time for meaningful work.</p></li></ul><h4><strong>Why leadership and change management matter</strong></h4><ul><li><p><strong>AI adoption is a human problem before it is a technical one.</strong> Organizations that understand developer pain points deploy AI more effectively.<br><strong>Agentic workflows amplify organizational differences.</strong> Teams with strong experimentation cultures and feedback loops move faster and with less friction.</p></li><li><p><strong>Culture determines outcomes.</strong> How leaders communicate expectations, normalize experimentation, and support learning shapes whether AI adoption succeeds or stalls.</p></li></ul><h4><strong>Looking ahead to 2026</strong></h4><ul><li><p><strong>Task parallelization is an emerging frontier.</strong> Developers are beginning to use agents to explore multiple solution paths simultaneously.</p></li><li><p><strong>Collaboration with agents will redefine productivity.</strong> Teams, not just individuals, will increasingly work alongside AI systems.</p></li><li><p><strong>Research must evolve with the work itself.</strong> New workflows will require new metrics, new telemetry, and new ways of understanding impact.</p></li></ul><h4><strong>Lessons from the METR paper</strong></h4><ul><li><p><strong>Context matters more than headlines suggest.</strong> Results showing slower performance often reflected expert developers working in familiar codebases.</p></li><li><p><strong>AI may help most where familiarity is lowest.</strong> New domains, unfamiliar systems, and onboarding scenarios show different outcomes.</p></li><li><p><strong>Media oversimplification distorts understanding.</strong> Nuance is critical when interpreting AI research, especially as studies move into real-world environments.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=155s">02:35</a>) Introducing the panel and the focus of the discussion</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=283s">04:43</a>) Why measuring AI&#8217;s impact is such a hard problem</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=330s">05:30</a>) How Microsoft approaches AI impact measurement</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=400s">06:40</a>) How Google thinks about measuring AI impact</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=448s">07:28</a>) GitHub&#8217;s perspective on measurement and insights from the DORA report</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=635s">10:35</a>) Why lines of code is a misleading metric</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=867s">14:27</a>) The limitations of measuring the percentage of code generated by AI</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1104s">18:24</a>) GitHub&#8217;s research on how AI is shaping the identity of the developer</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1299s">21:39</a>) How AI may change junior engineers&#8217; skill sets</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1482s">24:42</a>) Google&#8217;s research on using AI and creativity</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1584s">26:24</a>) High-leverage AI use cases that improve developer experience</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1958s">32:38</a>) Open research questions for AI and developer productivity in 2026</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=2133s">35:33</a>) How leading organizations approach change and agentic workflows</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=2282s">38:02</a>) Why the METR paper resonated and how it was misunderstood</p><h2><strong>Referenced:</strong></h2><p>&#8226; <a href="https://getdx.com/whitepaper/ai-measurement-framework">Measuring AI code assistants and agents</a></p><p>&#8226; <a href="https://kiro.dev/">Kiro</a></p><p>&#8226; <a href="https://code.claude.com/">Claude Code - AI coding agent for terminal &amp; IDE</a></p><p>&#8226; <a href="https://getdx.com/blog/space-framework-primer/">SPACE framework: a quick primer</a></p><p>&#8226; <a href="https://dora.dev/research/2025/dora-report/">DORA | State of AI-assisted Software Development 2025</a></p><p>&#8226; <a href="https://newsletter.pragmaticengineer.com/p/martin-fowler">Martin Fowler - by Gergely Orosz - The Pragmatic Engineer</a></p><p>&#8226; <a href="https://ieeexplore.ieee.org/document/10857384">Seamful AI for Creative Software Engineering: Use in Software Development Workflows | IEEE Journals &amp; Magazine | IEEE Xplore</a></p><p>&#8226; <a href="https://www.microsoft.com/en-us/research/publication/ai-where-it-matters-where-why-and-how-developers-want-ai-support-in-daily-work/">AI Where It Matters: Where, Why, and How Developers Want AI Support in Daily Work - Microsoft Research</a></p><p>&#8226; <a href="https://getdx.com/blog/unpacking-metri-findings-does-ai-slow-developers-down/">Unpacking METR&#8217;s findings: Does AI slow developers down?</a></p><p>&#8226; <a href="https://dxannual.com/">DX Annual 2026</a></p>]]></content:encoded></item><item><title><![CDATA[The AI Divide]]></title><description><![CDATA[MIT NANDA&#8217;s 2025 report on what separates successful AI pilots from those that stall, and how the role of engineers is evolving.]]></description><link>https://newsletter.getdx.com/p/the-ai-divide</link><guid isPermaLink="false">https://newsletter.getdx.com/p/the-ai-divide</guid><dc:creator><![CDATA[Justin Reock]]></dc:creator><pubDate>Tue, 23 Dec 2025 11:00:28 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1191ab79-3008-45a4-aea6-9de3cd5b9fbb_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of </strong></em><strong>Engineering Enablement,</strong><em><strong> </strong>a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p>&#128467;&#65039; We recently announced DX Annual, our flagship conference for developer productivity leaders navigating the AI era. <a href="https://dxannual.com/?utm_source=newsletter">Go here</a> to learn about the event and request an invite to attend.</p><div><hr></div><p>This week, I&#8217;m summarizing <a href="https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Business_2025_Report.pdf">&#8220;The GenAI Divide: State of AI in Business in 2025,</a>&#8221; a report that MIT&#8217;s Project NANDA (Networked Agents and Decentralized AI) produced earlier this year. The report examines why, despite massive investment in AI, most organizations are seeing little to no business impact. Drawing on interviews, surveys, and analysis of hundreds of AI initiatives, the authors argue that enterprise AI outcomes have split into two camps: a small minority extracting real value, and the rest stuck in pilots. They call this gap the GenAI Divide.</p><p>For DevProd and Platform leaders, this report provides a useful lens for understanding why AI initiatives stall inside engineering organizations. It also counters the &#8220;AI will replace engineers&#8221; narrative.</p><h2>My summary of the paper</h2><p>Headlines often suggest that engineers will soon be replaced by AI, or that the software industry is on the brink of collapse. While AI tools are improving quickly and demos are impressive, Project NANDA set out to study how organizations are actually using these systems in practice. As the authors put it, the GenAI Divide reflects &#8220;high adoption but low transformation,&#8221; with 95% of organizations seeing zero measurable return.</p><p>The researchers conducted 53 structured interviews and collected survey responses from 153 senior leaders. The observation window covered roughly six months following initial AI pilots, and the authors are careful to describe the findings as a directionally accurate snapshot rather than a definitive market analysis.</p><p>Here are the key findings from the report.</p><h3>The pilot-to-production gap</h3><p>The most visible expression of the AI Divide is the steep drop-off from experimentation to real deployment. While many organizations evaluate AI tools, very few succeed in moving them into production. Roughly 60% evaluated task-specific or enterprise AI systems, about 20% reached pilot stage, and just 5% deployed systems that delivered sustained productivity or P&amp;L impact.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nzur!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nzur!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png 424w, https://substackcdn.com/image/fetch/$s_!nzur!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png 848w, https://substackcdn.com/image/fetch/$s_!nzur!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png 1272w, https://substackcdn.com/image/fetch/$s_!nzur!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nzur!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png" width="1456" height="682" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:682,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:591003,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/182343989?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nzur!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png 424w, https://substackcdn.com/image/fetch/$s_!nzur!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png 848w, https://substackcdn.com/image/fetch/$s_!nzur!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png 1272w, https://substackcdn.com/image/fetch/$s_!nzur!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4e49640c-e3ff-4f68-a2c1-78e19eef0e67_5112x2396.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Large enterprises struggled most with this transition. They led in pilot volume and investment, but had the lowest rates of successful scale-up. Mid-market organizations moved faster and more decisively, often reaching full implementation within 90 days. Across successful cases, the difference was ultimately about focus:</p><ul><li><p>Successful teams focused on narrow, workflow-specific use cases with clearly defined operational outcomes, rather than broad or generalized AI deployments.</p></li><li><p>Implementation ownership sat with domain leaders and frontline managers, not centralized AI labs or exploratory teams, enabling faster decision-making and clearer accountability.</p></li><li><p>AI systems were embedded directly into existing tools, processes, and data flows, reducing context switching and allowing them to operate within day-to-day work rather than alongside it.</p></li></ul><h3>Buy beats build</h3><p>Organizations that crossed the AI Divide approached implementation differently. Rather than building AI systems entirely in-house, they were far more likely to partner with external vendors. Externally sourced tools reached deployment at roughly twice the rate of internal builds, which frequently failed due to brittleness and poor workflow fit.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3m4h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3m4h!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png 424w, https://substackcdn.com/image/fetch/$s_!3m4h!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png 848w, https://substackcdn.com/image/fetch/$s_!3m4h!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png 1272w, https://substackcdn.com/image/fetch/$s_!3m4h!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3m4h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png" width="1456" height="526" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:526,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:763578,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/182343989?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3m4h!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png 424w, https://substackcdn.com/image/fetch/$s_!3m4h!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png 848w, https://substackcdn.com/image/fetch/$s_!3m4h!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png 1272w, https://substackcdn.com/image/fetch/$s_!3m4h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F197c449d-87fa-4e99-97e3-ff3d490af4fb_5112x1848.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Additionally, successful buyers treated AI vendors less like SaaS providers and more like service partners. They demanded deep customization, evaluated tools based on measurable outcomes, and expected systems to integrate into existing workflows with minimal disruption.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0ip5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0ip5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png 424w, https://substackcdn.com/image/fetch/$s_!0ip5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png 848w, https://substackcdn.com/image/fetch/$s_!0ip5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png 1272w, https://substackcdn.com/image/fetch/$s_!0ip5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0ip5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png" width="1456" height="668" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:668,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:818757,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.getdx.com/i/182343989?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0ip5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png 424w, https://substackcdn.com/image/fetch/$s_!0ip5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png 848w, https://substackcdn.com/image/fetch/$s_!0ip5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png 1272w, https://substackcdn.com/image/fetch/$s_!0ip5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0088b7aa-d03b-48ef-8b87-72c31260b7ff_5112x2344.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Individual productivity gains are not the same as organizational transformation</h3><p>The report makes it clear that while many teams report that AI helps engineers move faster on specific tasks, those gains rarely translate into systemic impact unless teams design the workflows around them. Two critical reasons highlight this point:</p><ol><li><p>70% of users in the study prefer AI for quick tasks like generating simple unit tests or performing basic refactoring, but for complex projects involving multi-week work and stakeholder management, 90% of users still prefer human colleagues.</p></li><li><p>66% of executives noted that they want systems in their core workflows that learn from feedback and have more robust memory, while many agents repeat mistakes and fail at complex tasks repeatedly.</p></li></ol><p>As one CIO interviewed noted: &#8220;Once we&#8217;ve invested time in training a system to understand our workflows, the switching costs become prohibitive.&#8221; For now, AI will remain a powerful feature of the engineer&#8217;s overall toolkit, but not a replacement for the engineer.</p><p>Ultimately, the dividing line between human and agent isn&#8217;t just &#8220;intelligence,&#8221; it&#8217;s context. Engineers don&#8217;t just generate code; they maintain a mental model of the codebase, the business requirements, and the team&#8217;s historical preferences, culture, and style. Current AI solutions lack the contextual adaptability and true persistent memory necessary to perform complex human tasks.</p><h3>Humans remain in the loop</h3><p>Across the successful deployments examined, a consistent pattern emerged: AI systems were designed to augment human judgment, not replace it. Engineers remained responsible for architecture, correctness, and evolution, while AI handled acceleration, suggestion, and pattern recognition within defined constraints.</p><p>Analyzing the findings shows that the future isn&#8217;t one in which engineers disappear, but rather one in which their roles shift. There&#8217;s less time spent on toilsome rote work and more on the difficult problems of system design, quality, integration, and decision-making.</p><h2>Final thoughts</h2><p>The most revealing takeaway from this report is where AI adoption breaks down. The sharp drop from pilots to production suggests that real constraints only surface once AI is embedded in long-lived, complex workflows. That same pattern makes claims about AI replacing engineers feel premature.</p><p>If AI were truly ready to replace engineers, we would already be seeing faster delivery at scale, shrinking teams, and systems that reliably design, build, and integrate software with minimal human oversight. Instead, the report shows that most AI efforts fail at the point where real-world context, dependencies, and coordination matter most.</p><p>This represents more responsibility for engineers, not less. Software engineering has never been just about producing code. The companies getting real value from AI aren&#8217;t the ones trying to eliminate engineers. They&#8217;re the ones who understand AI&#8217;s capabilities to significantly increase throughput and innovative capacity. These teams invest in adoption, integration, and measurement.</p><div><hr></div><p>This is the final issue of the Engineering Enablement newsletter in 2025. Enjoy the holidays, and we&#8217;ll see you in January.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[AI and productivity: Year-in-review with Microsoft, Google, and GitHub researchers]]></title><description><![CDATA[What 2025&#8217;s AI research told us about developer productivity and identity, and why enablement matters more than tool choice.]]></description><link>https://newsletter.getdx.com/p/ai-and-productivity-year-in-review</link><guid isPermaLink="false">https://newsletter.getdx.com/p/ai-and-productivity-year-in-review</guid><dc:creator><![CDATA[Laura Tacho]]></dc:creator><pubDate>Wed, 17 Dec 2025 11:02:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/04188189-de39-4f90-9af4-afbff1a510cc_4000x2800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>Welcome to the latest issue of </strong></em><strong>Engineering Enablement,</strong><em> a weekly newsletter sharing research and perspectives on developer productivity.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/subscribe?"><span>Subscribe now</span></a></p><p>&#128467;&#65039;We recently announced DX Annual, our flagship conference for developer productivity leaders navigating the AI era. <a href="https://dxannual.com/?utm_source=newsletter">Go here</a> to learn about the event and request an invite to attend.</p><div><hr></div><p>In 2025, AI-assisted engineering moved from an experiment to a core business expectation.</p><p>At the start of the year, adoption rates varied widely across organizations, and a path toward widespread use was just beginning to come into focus. Now, at the end of 2025, that picture looks very different: roughly 90% of developers across the industry are using AI tools at least once a month to get work done, with more than 40% relying on them every day.</p><p>As adoption has increased, so have our questions about its impact. Looking back at 2025, we&#8217;ve learned a lot as an industry about how AI is changing the way software gets made. To close out the year, I hosted a research roundtable with prominent voices in the AI and developer productivity research space. I invited them to reflect on what we&#8217;ve learned so far, and to share the questions they&#8217;re carrying into 2026.</p><p>A few clear themes emerged from that conversation. Below, I expand on those themes and add my own perspective on what they mean for the year ahead.</p><p><em>Watch the full discussion with Brian Houck (Microsoft), Ciera Jaspan (Google), Collin Green (Google), and Eirini Kalliamvakou (GitHub) <a href="https://getdx.com/webinar/AI-productivity-year-in-review/?utm_source=newsletter">here</a>.</em></p><h3>Lines of code is still a bad metric. Don&#8217;t let uncertainty and change make you reach for it</h3><p>Building on themes from the last decade of developer productivity research, 2025&#8217;s research into measuring AI impact landed in a familiar place: there&#8217;s no single number that tells you whether AI is actually making a difference. Across the industry, organizations are using a broad range of metrics to measure impact (like the <a href="https://getdx.com/research/measuring-ai-code-assistants-and-agents/?utm_source=newsletter">AI Measurement Framework</a>, or check out <a href="https://getdx.com/report/how-companies-measure-ai-impact-in-engineering/?utm_source=newsletter">this research piece</a> on how Google, Microsoft, GitHub, and others are actually measuring productivity).</p><p>But even if there are a lot of good patterns out there, there&#8217;s also one bad pattern that this group of researchers called out: using lines of code (LOC) as a measurement for AI impact. This raw output metric is easy to measure, and in the absence of a clear alternative (especially in times of change), it can be tempting to reach for numbers that seem predictable and objective. Collectively, this group warned not to mistake output for impact. AI lends itself to writing a lot of lines of code, but that doesn&#8217;t necessarily mean a positive impact on your teams, organizations, or business.</p><h3>Talking to AI and not talking to your colleagues might not be great for teams longer-term</h3><p>Brian Houck, Sr. Principal Applied Scientist at Microsoft and co-author of the SPACE Framework of Developer Productivity, shared insights from his paper <a href="https://www.microsoft.com/en-us/research/publication/the-space-of-ai-real-world-lessons-on-ais-impact-on-developers/">The SPACE of AI: Real-World Lessons on AI&#8217;s Impact on Developers</a>, which shows how the impact of AI tools can vary widely across five key dimensions of developer productivity: Satisfaction, Performance, Activity, Collaboration, and Efficiency. While 90% of developers report that AI makes them more productive, fewer than half agree that it makes them more collaborative and communicative with their teammates. As teams use AI for longer periods of time, it will be interesting to see how this pattern plays out when it comes to knowledge sharing or even the long-term maintenance of codebases.</p><h3>AI changes what skills developers need, but also how they perceive themselves</h3><p>Eirini Kalliamvakou, Research Advisor at GitHub, shared some details from her recent research <a href="https://github.blog/news-insights/octoverse/the-new-identity-of-a-developer-what-changes-and-what-doesnt-in-the-ai-era/?utm_source=dx-webinar-eirini&amp;utm_medium=blog&amp;utm_campaign=dec25postuniverse">The new identity of a developer: What changes and what doesn&#8217;t in the AI era</a>. As developers become more fluent with AI, their identity is shifting from traditional &#8220;code producer&#8221; toward a role focused on directing, delegating, and validating AI-assisted workflows, with creative judgment and strategic orchestration becoming central to their craft.</p><p>Interestingly, many of today&#8217;s heavy AI users started out as skeptics. Hands-on experience with the tools often changed both their sentiment and their expectations.</p><p>This identity shift has real implications for organizations, particularly around career progression, hiring, and upskilling. Both companies and developers need to place more value on AI fluency, systems thinking, and judgment, rather than raw coding output alone. And keeping pace with the ecosystem will require broader AI enablement, not just tool-specific training programs.</p><h3>Is AI the death of the junior developer, or an accelerant to help them level up faster?</h3><p>Lots of folks, from new grads to seasoned developers, are concerned about how AI will impact the talent pipeline. The usual narrative is that AI puts <a href="https://sourcegraph.com/blog/the-death-of-the-junior-developer">junior developers at risk of extinction</a>, because why hire a junior when a senior developer can just delegate tasks to an AI agent instead? This could be a short-sighted optimization that leaves us with no developing talent a few years from now.</p><p>Ciera Jaspan from Google offered a compelling alternative perspective: what if the skills that devs need to level up in seniority&#8212;strong problem-solving skills, managing work, delegating work, and clearly defining outcomes&#8212;are now learned <em>earlier</em> by junior engineers because of the ways they need to interact with agents? Previously, these skills would be delayed because a tech lead or senior team member would take on the brunt of the project and professional management for these juniors. But when juniors are in the role of team lead for a handful of agents, do they actually level up faster because they get more practice solving problems end-to-end, even if their time spent actually typing code is reduced? Of course, in order for this to happen, companies still need to hire juniors, which isn&#8217;t always the case anymore. Find out how you might observe this pattern within your own teams.</p><p>Collin Green, another Google researcher, connected this back to earlier concerns about communication. Faster leveling through AI interaction doesn&#8217;t automatically translate to stronger collaboration skills. If juniors primarily work with AI rather than people, what does that mean for their professional development? And if seniors spend less time mentoring, what downstream effects might that have as well?</p><h3>Is automation and code generation the right focus for the future of AI tools?</h3><p>Collin advised on an AI paper that had a slightly different focus: <a href="https://ieeexplore.ieee.org/document/10857384/authors#authors">creativity in software engineering</a>. If creativity is the goal, and not productivity or automation, tools might better help us reimagine how work gets done and lead to more impactful outcomes, rather than just getting to the same outcomes faster. A shift in focus from creativity to productivity changes how tools are built and how we use them. The current emphasis on automation covers only a small surface area, with many developers coding only about 1 day a week on average. Everything else&#8212;scoping, experimenting, and validating ideas&#8212;exists in a very creative space, which AI is well-suited to help with, but many tools started with an emphasis on productivity and efficiency.</p><p>At the same time, there are so many tasks classified as &#8220;toil&#8221; that would serve developers well to be automated. These types of tasks make developers look at their to-do list and say, &#8220;ugh, today is not going to be great.&#8221; But Collin shared, we&#8217;ve had really solid research and technology for automating tasks for the last 40+ years. We <em>should</em> be automating things that are &#8220;dull, dirty, and dangerous.&#8221; But the most consistent problems that add friction or toil to developers&#8217; days&#8212;like tech debt, lack of documentation, compliance tasks, even expense reports&#8212;still haven&#8217;t been solved with automation. Can AI change that, or are they just too complex to be automated? Is AI the right tool to automate it anyway?</p><h3>Headlines will oversimplify AI research to the point of being incorrect, even if the research itself is full of nuance</h3><p>Earlier this year, METR <a href="https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/">published a study</a> that showed how, in some contexts, developers actually slow down when using AI, even if their own perception is that they were moving faster. This discrepancy between self-perception and actual result definitely made a big splash in our industry, and not a week goes by that I don&#8217;t hear someone cite the study as evidence that we&#8217;re all careening off an AI-generated cliff.</p><p>Importantly, the study itself had a lot more nuance and depth than the one-line headlines captured. And since one-line headlines are mostly what gets shared on LinkedIn and other platforms, it didn&#8217;t take long for the (well-done) METR study to be oversimplified and distilled down to the simple point that AI makes developers slower, which wasn&#8217;t really the point of the paper at all.</p><p>But the headline that AI isn&#8217;t actually as helpful as promised definitely resonated with people, and the study was widely shared. Many people felt drawn to the results that more accurately reflected their own personal experience trying to get started with a tool that was unreliable. For others, it was a good antidote to all the hype. One thing that the METR study did show was that AI research was now taking place <em>in situ, </em>meaning with real developers solving real problems, and specifically their problems.</p><p>An important thing to remember as 2026 will surely bring even more headlines that are greatly oversimplified summaries: stay curious about what&#8217;s behind the number. For the METR study, which was &#8220;AI makes devs 19% slower.&#8221; But not all devs, and not all context. Those questions are missing from the headlines, so you need to ask them yourself.</p><h3>Looking to 2026</h3><p>Closing out 2025, I can confidently say we haven&#8217;t seen the full spectrum of AI impact yet. We&#8217;ve made a lot of progress on understanding the impact of AI on teams, but we still need to keep digging to assess whether AI is achieving what we want it to, and to fully understand the impact on not just the whole software delivery lifecycle, but all levels of organizations.</p><p>One thing is clear to me though: the companies who will see the biggest wins with AI in 2026 are the ones who deeply understand their existing bottlenecks. The real acceleration from AI doesn&#8217;t come from using the newest models and testing every new tool; it comes from pointing AI at <a href="https://www.microsoft.com/en-us/research/publication/ai-where-it-matters-where-why-and-how-developers-want-ai-support-in-daily-work/">the real problems</a> that slow developers down.</p><div><hr></div><p>This week&#8217;s featured DevProd job openings. See more <a href="https://getdx.com/resources/devex-jobs/">open roles here</a>.</p><ul><li><p><strong>American Express</strong> is hiring a <a href="https://aexp.eightfold.ai/careers/job/37813133?hl=en">Sr. Manager, Digital Product Management - DevProd</a> | Hybrid - London UK</p></li><li><p><strong>Capital One</strong> is hiring a <a href="https://www.capitalonecareers.com/job/mclean/manager-product-management-developer-experience/1732/86033033056">Product Manager - Developer Experience</a> | Plano TX; McLean VA; Richmond VA</p></li><li><p><strong>Gusto</strong> is hiring a <a href="https://job-boards.greenhouse.io/gusto/jobs/7413640">Sr. Platform Engineer</a> | Denver, CO; San Francisco, CA; Atlanta, GA; Austin, TX; Chicago, IL; Miami, FL; Seattle, WA</p></li><li><p><strong>Plaid</strong> is hiring a <a href="https://www.linkedin.com/jobs/view/4289332289/">Software Engineer - Platform</a> | New York, NY</p></li><li><p><strong>Reddit</strong>: <a href="https://job-boards.greenhouse.io/reddit/jobs/7342078">Staff Software Engineer - Developer Experience</a> | Remote - United States</p></li><li><p><strong>Whatnot </strong>is hiring a <a href="https://jobs.ashbyhq.com/whatnot/936135ac-ef8d-47e7-bee6-a7cf7d361c64">Software Engineer - Platform</a> | San Francisco, Los Angeles, Seattle, NYC</p></li></ul><div><hr></div><p>That&#8217;s it for this week. Thanks for reading.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/p/ai-and-productivity-year-in-review?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/p/ai-and-productivity-year-in-review?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[Running data-driven evaluations of AI engineering tools]]></title><description><![CDATA[A concise, data-driven framework for testing and adopting AI engineering tools.]]></description><link>https://newsletter.getdx.com/p/running-data-driven-evaluations-of</link><guid isPermaLink="false">https://newsletter.getdx.com/p/running-data-driven-evaluations-of</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 12 Dec 2025 15:49:16 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/181184615/8247b4109166deb4674678576f392b31.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/SQVdvKxzOH0">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>AI engineering tools are evolving fast. Every month brings new coding assistants, debugging agents, and automation capabilities. I want to help engineering leaders take advantage of that innovation while avoiding costly experiments that distract from real product work.</p><p>In this episode, Abi Noda and I share a practical, data-driven approach to evaluating AI tools. I walk through how to shortlist tools by use case, design structured trials that reflect real work, select representative participants, and measure impact using baselines and proven frameworks. My goal is to give you a way to test and adopt AI tools with confidence and a clear return on investment.</p><div id="youtube2-SQVdvKxzOH0" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;SQVdvKxzOH0&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/SQVdvKxzOH0?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>Data-driven evaluations are essential</strong></p><ul><li><p><strong>Structured, measurable trials prevent bias.</strong> Without them, decisions are driven by novelty hype or a few loud voices.</p></li><li><p><strong>Define a clear business outcome first</strong> (reduce toil, improve delivery speed, or raise code quality).</p></li><li><p><strong>Evaluations must inform real decisions</strong>, not just check a procurement box.</p></li></ul><p><strong>Choose the right set of tools to evaluate</strong></p><ul><li><p><strong>Group tools by use case and interaction mode</strong> (chat, agentic IDEs, code review assistants, etc.) to ensure fair comparisons.</p></li><li><p><strong>Match shortlist size to org capacity</strong> to support multiple cohorts and reliable results.</p></li><li><p><strong>Multi-vendor strategies reduce lock-in</strong> in a rapidly shifting market.</p></li></ul><p><strong>Re-evaluations are essential, not optional</strong></p><ul><li><p><strong>Incumbent tools must be retested</strong> as capabilities evolve and new challengers emerge.</p></li><li><p><strong>Triggers for re-evaluation</strong> include major feature launches, organic developer adoption of a new tool, and upcoming renewal cycles.</p></li><li><p><strong>Every challenger tool evaluation requires a baseline of the incumbent</strong>, so you can compare like-for-like.</p></li><li><p><strong>A cadence of every 8&#8211;14 months</strong> ensures decisions reflect the current reality, not the past purchase.</p></li></ul><p><strong>Design trials around research questions</strong></p><ul><li><p><strong>Start with a hypothesis.</strong> It keeps experiments aligned to actual goals.</p></li><li><p><strong>Developer sentiment is necessary but insufficient</strong> without measurable outcomes.</p></li><li><p><strong>Success criteria must be defined in advance</strong> to avoid subjective decision-making.</p></li></ul><p><strong>Select representative participants</strong></p><ul><li><p><strong>Diverse cohorts reveal real impact</strong> across languages, teams, and seniority levels.</p></li><li><p><strong>Include skeptical and late adopters</strong> to uncover onboarding and enablement needs.</p></li><li><p><strong>Volunteer-only trials distort results</strong> and won&#8217;t scale to full org rollout.</p></li></ul><p><strong>Run evaluations long enough to capture true behavior</strong></p><ul><li><p><strong>Eight to twelve weeks is the minimum</strong> to get past the novelty phase and into sustained usage.</p></li><li><p><strong>Align evaluation windows to procurement cycles</strong> so insights guide buying decisions.</p></li><li><p><strong>Short trials lead to false signals </strong>and either inflate enthusiasm or create false negativity.</p></li></ul><p><strong>Use self-reported time savings carefully</strong></p><ul><li><p><strong>Self-reporting is a strong early indicator</strong> of perceived usefulness.</p></li><li><p><strong>Humans misremember time</strong>, often benchmarking against recent AI use.</p></li><li><p><strong>Treat CSAT and time savings as directional</strong>, not the final truth.</p></li><li><p><strong>Objective metrics validate real ROI,</strong> including throughput, quality, and innovation time.</p></li></ul><p><strong>Expect variation rather than a single winner</strong></p><ul><li><p><strong>Different tools shine in different contexts</strong>, so multiple standards are often the best path.</p></li><li><p><strong>Continuous re-evaluation is required</strong> as capabilities evolve every quarter.</p></li><li><p><strong>The right goal isn&#8217;t the &#8220;best tool&#8221;</strong>, but the best tool for <em>each</em> problem space.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0">00:00</a>) Intro: Running a data-driven evaluation of AI tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=156s">02:36</a>) Challenges in evaluating AI tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=371s">06:11</a>) How often to reevaluate AI tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=422s">07:02</a>) Incumbent tools vs challenger tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=460s">07:40</a>) Why organizations need disciplined evaluations before rolling out tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=568s">09:28</a>) How to size your tool shortlist based on developer population</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=764s">12:44</a>) Why tools must be grouped by use case and interaction mode</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=810s">13:30</a>) How to structure trials around a clear research question</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1005s">16:45</a>) Best practices for selecting trial participants</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1162s">19:22</a>) Why support and enablement are essential for success</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1270s">21:10</a>) How to choose the right duration for evaluations</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1372s">22:52</a>) How to measure impact using baselines and the AI Measurement Framework</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1528s">25:28</a>) Key considerations for an AI tool evaluation</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1732s">28:52</a>) Q&amp;A: How reliable is self-reported time savings from AI tools?</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1942s">32:22</a>) Q&amp;A: Why not adopt multiple tools instead of choosing just one?</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=2007s">33:27</a>) Q&amp;A: Tool performance differences and avoiding vendor lock-in</p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; X: <a href="https://x.com/rhein_wein">https://x.com/rhein_wein</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p>&#8226; Laura&#8217;s course (Measuring Engineering Performance and AI Impact): <a href="https://lauratacho.com/developer-productivity-metrics-course">https://lauratacho.com/developer-productivity-metrics-course</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><p>&#8226; Substack: &#8203;&#8203;<a href="https://substack.com/@abinoda">https://substack.com/@abinoda</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/whitepaper/ai-measurement-framework/">Measuring AI code assistants and agents</a></p></li><li><p><a href="https://qconferences.com/">QCon conferences</a></p></li><li><p><a href="https://getdx.com/dx-core-4/">DX Core 4 engineering metrics</a></p></li><li><p><a href="https://getdx.com/podcast/doras-2025-research-on-the-impact-of-ai/">DORA&#8217;s 2025 research on the impact of AI</a></p></li><li><p><a href="https://getdx.com/blog/unpacking-metri-findings-does-ai-slow-developers-down/">Unpacking METR&#8217;s findings: Does AI slow developers down?</a></p></li><li><p><a href="https://newsletter.getdx.com/p/metr-study-on-how-ai-affects-developer-productivity">METR&#8217;s study on how AI affects developer productivity</a></p></li><li><p><a href="https://www.claude.com/product/claude-code">Claude Code</a></p></li><li><p><a href="https://cursor.com/">Cursor</a></p></li><li><p><a href="https://windsurf.com/">Windsurf</a></p></li><li><p><a href="https://newsletter.getdx.com/p/do-newer-ai-native-ides-outperform-other-ai-coding-assistants">Do newer AI-native IDEs outperform other AI coding assistants?</a></p></li></ul>]]></content:encoded></item></channel></rss>