<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Engineering Enablement: Engineering Enablement Podcast]]></title><description><![CDATA[A weekly podcast covering how top companies measure and improve developer productivity]]></description><link>https://newsletter.getdx.com/s/podcast</link><image><url>https://substackcdn.com/image/fetch/$s_!Niij!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbd433b-6f11-4042-8b7d-0edb3b172966_1024x1024.png</url><title>Engineering Enablement: Engineering Enablement Podcast</title><link>https://newsletter.getdx.com/s/podcast</link></image><generator>Substack</generator><lastBuildDate>Sat, 04 Apr 2026 13:50:12 GMT</lastBuildDate><atom:link href="https://newsletter.getdx.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Abi Noda]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[abinoda@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[abinoda@substack.com]]></itunes:email><itunes:name><![CDATA[Abi Noda]]></itunes:name></itunes:owner><itunes:author><![CDATA[Abi Noda]]></itunes:author><googleplay:owner><![CDATA[abinoda@substack.com]]></googleplay:owner><googleplay:email><![CDATA[abinoda@substack.com]]></googleplay:email><googleplay:author><![CDATA[Abi Noda]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Scaling developer experience across 1,000 engineers at Dropbox]]></title><description><![CDATA[Listen now | I talk with Uma Namasivayam of Dropbox about treating developer productivity as a business problem, running developer experience like a product, and building foundations that make AI useful at scale.]]></description><link>https://newsletter.getdx.com/p/scaling-developer-experience-across</link><guid isPermaLink="false">https://newsletter.getdx.com/p/scaling-developer-experience-across</guid><dc:creator><![CDATA[Laura Tacho]]></dc:creator><pubDate>Fri, 06 Feb 2026 18:40:25 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/186996290/472b31ff92db4afe2b7d541e11bae71a.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/ZCg2k-w6o2o">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>Developer productivity is often treated as a tooling problem or a sentiment problem. In reality, it&#8217;s neither. It&#8217;s a socio-technical systems problem that spans engineering foundations, leadership alignment, organizational design, and culture.</p><p>In this episode, I&#8217;m joined by Uma Namasivayam, Senior Director, Engineering Productivity at Dropbox, to explore how Dropbox approaches developer experience at scale. We talk about why productivity needs to be framed as a business problem, how executive alignment creates the conditions for meaningful change, and what it takes to treat developer experience as a real product with developers as customers.</p><p>We also dig into Dropbox&#8217;s approach to AI adoption. Uma shares why strong foundations, such as build, test, and observability, are prerequisites for AI to actually accelerate work, how Dropbox encourages daily AI use without mandating a single tool, and where build-versus-buy decisions break down at scale.</p><p>We close with an honest look at what remains unsolved: how to connect gains in developer productivity and AI-driven capacity to real business outcomes, and where engineering leaders should focus next in 2026.</p><div id="youtube2-ZCg2k-w6o2o" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;ZCg2k-w6o2o&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/ZCg2k-w6o2o?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><h4><strong>Developer productivity is a socio-technical problem</strong></h4><ul><li><p><strong>Productivity cannot be solved through tooling alone</strong>; it spans engineering systems, leadership behavior, organizational structure, and people practices.</p></li><li><p><strong>Problems like build and test are engineering problems</strong>, while <strong>problems like focus time and interruptions are people problems</strong>, and both matter equally.</p></li><li><p><strong>Treating productivity as a system forces tradeoffs to be explicit</strong>, rather than hidden inside isolated tooling initiatives.</p></li></ul><h4><strong>Executive alignment matters more than any single metric</strong></h4><ul><li><p><strong>Top-down sponsorship creates permission to act</strong>, especially when productivity work cuts across org boundaries.</p></li><li><p><strong>A shared framework creates alignment, not answers</strong>; its value is giving leaders and engineers a common language.</p></li><li><p><strong>System metrics matter more than single metrics</strong>, because productivity improvements rarely move one dimension in isolation.</p></li><li><p><strong>Distributed accountability makes productivity a company problem</strong>, not a developer experience team problem.</p></li></ul><h4><strong>Developer experience works best when treated as a product discipline</strong></h4><ul><li><p><strong>Developers are customers</strong>, and their experience must be understood through both qualitative feedback and quantitative signals.</p></li><li><p><strong>Good system metrics do not guarantee good developer experience</strong>, which is why sentiment and perception matter.</p></li><li><p><strong>DX surveys surface where systems break differently for different teams</strong>, such as desktop, mobile, and web developers.</p></li><li><p><strong>Continuous feedback loops are essential</strong>, combining surveys, direct conversations, and usage data.</p></li><li><p><strong>Internal communication is part of the product</strong>, reinforcing to developers that their feedback leads to real change.</p></li></ul><h4><strong>Prioritization requires structure, not intuition</strong></h4><ul><li><p><strong>Finite capacity makes prioritization unavoidable</strong>, even in large, well-resourced engineering orgs.</p></li><li><p><strong>Segmenting developer populations clarifies tradeoffs</strong>, since different teams experience different bottlenecks.</p></li><li><p><strong>DX survey data provides a defensible starting point</strong>, but prioritization still requires judgment.</p></li><li><p><strong>Leadership-level stack ranking helps resolve conflicts</strong>, especially when multiple teams compete for attention.</p></li><li><p><strong>Frameworks make hard decisions easier to explain</strong>, even when they do not make them easy.</p></li></ul><h4><strong>AI and developer experience must advance in parallel</strong></h4><ul><li><p><strong>AI accelerates work, while developer experience reduces friction</strong>, and both are required for sustained gains.</p></li><li><p><strong>Foundational systems act as plumbing</strong>, enabling trust in speed, quality, and safety.</p></li><li><p><strong>Without strong CI, testing, and observability</strong>, faster code creation increases risk instead of value.</p></li><li><p><strong>Trust in guardrails enables confidence in AI-assisted development</strong>, especially at scale.</p></li></ul><h4><strong>AI adoption succeeds through choice, not mandates</strong></h4><ul><li><p><strong>Early organic adoption revealed real developer needs</strong>, rather than forcing a single tool.</p></li><li><p><strong>Different teams require different AI tools</strong>, particularly for mobile, desktop, and large-repo workflows.</p></li><li><p><strong>Supporting multiple tools increased adoption</strong>, rather than reducing it.</p></li><li><p><strong>Daily use depends on fitting AI into existing workflows</strong>, not adding extra steps.</p></li><li><p><strong>Habits matter more than access</strong>, which is why SDLC-level integration is critical.</p></li></ul><h4><strong>Build vs. buy decisions change at scale</strong></h4><ul><li><p><strong>Many AI tools fail when tested at large-company scale</strong>, despite working well in smaller contexts.</p></li><li><p><strong>Cost and performance become gating factors</strong>, not feature completeness.</p></li><li><p><strong>Internal platforms can abstract complexity</strong>, enabling teams to build AI workflows safely and consistently.</p></li><li><p><strong>Shared internal platforms unlock reuse</strong>, allowing teams to innovate without rebuilding infrastructure.</p></li><li><p><strong>Speed of iteration remains the primary differentiator</strong>, even when building in-house.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=45s">00:45</a>) Dropbox&#8217;s engineering org</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=119s">01:59</a>) Why developer productivity is a business problem</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=248s">04:08</a>) The role of executive sponsorship in developer productivity</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=362s">06:02</a>) How DX&#8217;s Core Four framework created a shared language</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=493s">08:13</a>) Treating developer experience as a product</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=690s">11:30</a>) How Dropbox prioritizes developer experience work</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=860s">14:20</a>) The challenge of tying developer experience to business outcomes</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=998s">16:38</a>) How AI and developer experience intersect at Dropbox</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1115s">18:35</a>) The prerequisites for AI adoption to accelerate work</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1226s">20:26</a>) How Dropbox encourages daily AI use</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1392s">23:12</a>) AI use beyond code completion</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1500s">25:00</a>) Managing AI tool demand at scale</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1676s">27:56</a>) Early results from Dropbox&#8217;s AI efforts</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1805s">30:05</a>) Progress on developer experience at Dropbox</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=1975s">32:55</a>) Advice for organizations investing in developer experience</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=2065s">34:25</a>) Capacity tradeoffs for developer experience</p><p>(<a href="https://www.youtube.com/watch?v=ZCg2k-w6o2o&amp;t=2159s">35:59</a>) The unanswered questions around AI and capacity in 2026</p><h2><strong>Referenced:</strong></h2><p>&#8226; <a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p><p>&#8226; <a href="https://www.dropbox.com/">Dropbox.com</a></p>]]></content:encoded></item><item><title><![CDATA[AI and productivity: A year-in-review with Microsoft, Google, and GitHub researchers]]></title><description><![CDATA[Listen now | Listen and watch now on YouTube, Apple, and Spotify.]]></description><link>https://newsletter.getdx.com/p/ai-and-productivity-a-year-in-review</link><guid isPermaLink="false">https://newsletter.getdx.com/p/ai-and-productivity-a-year-in-review</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Mon, 29 Dec 2025 17:17:45 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/182642831/a8424bdccac94a35590d7bc3484b6193.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/xZjvYMuAJPc">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>As we close out 2025, I wanted to step back and take stock of what we have actually learned about AI adoption in engineering organizations. Not just where usage has increased, but where impact is real, where it is overstated, and what questions remain unanswered.</p><p>In this special year-end episode, I&#8217;m joined by Brian Houck from Microsoft, Collin Green and Ciera Jaspan from Google, and Eirini Kalliamvakou from GitHub. Together, we unpack the research each of them worked on this year and explore how leading organizations are thinking about AI measurement, developer experience, and long-term productivity. We talk candidly about why measuring AI&#8217;s impact is so difficult, why familiar metrics like lines of code keep resurfacing despite their flaws, and how multidimensional approaches like SPACE and DORA offer a more realistic lens.</p><p>We also look ahead to 2026. We discuss how AI is beginning to reshape the identity of the developer, how junior engineers&#8217; skill sets may evolve, where agentic workflows are gaining traction, and why some of the most widely shared AI studies were misunderstood. This episode is an honest conversation about moving past hype and toward a more grounded, evidence-based approach to AI adoption in engineering teams.</p><div id="youtube2-xZjvYMuAJPc" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;xZjvYMuAJPc&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/xZjvYMuAJPc?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><h4><strong>Measuring AI impact requires multiple lenses</strong></h4><ul><li><p><strong>There is no single metric that can capture AI&#8217;s impact.</strong> Developer productivity and experience are inherently multidimensional, requiring trade-offs to be evaluated across speed, quality, collaboration, and meaning.</p></li><li><p><strong>Frameworks like SPACE and DORA help avoid metric tunnel vision.</strong> They encourage teams to examine complementary signals rather than optimizing one dimension at the expense of others.</p></li><li><p><strong>Measurement must reflect systems, not tools.</strong> AI does not operate in isolation; its impact depends on organizational context, workflows, and existing engineering practices.</p></li></ul><h4><strong>Why familiar metrics keep failing us</strong></h4><ul><li><p><strong>Lines of code remains a deeply misleading metric.</strong> AI tends to generate verbose code, making raw output a poor proxy for productivity, quality, or long-term maintainability.</p></li><li><p><strong>More code does not equal better outcomes.</strong> Excess code can increase maintenance burden, technical debt, and cognitive load over time.</p></li><li><p><strong>Easy-to-measure metrics are often the most dangerous.</strong> Their simplicity makes them attractive during periods of uncertainty, even when they obscure what is actually changing.</p></li></ul><h4><strong>The limits of tracking AI-generated code</strong></h4><ul><li><p><strong>Measuring the percentage of AI-generated code oversimplifies reality.</strong> AI may write, delete, refactor, or reorganize code in ways that raw percentages fail to capture.<br><strong>AI-generated code does not inherently signal higher risk.</strong> In some contexts, AI output may be more consistent or higher quality than human-written code.</p></li><li><p><strong>These metrics are better used as supporting signals, not goals.</strong> They can inform budgeting, experimentation, or adoption patterns but should not drive performance targets.</p></li></ul><h4><strong>How AI is reshaping the role of the developer</strong></h4><ul><li><p><strong>Developers are shifting from implementers to orchestrators.</strong> Advanced AI users spend more time framing problems, setting context, and validating outcomes than writing raw code.</p></li><li><p><strong>AI fluency is becoming a core skill.</strong> Knowing how to guide, correct, and collaborate with agents is increasingly important.</p></li><li><p><strong>Adoption follows a progression.</strong> Developers tend to move from skepticism to exploration, collaboration, and eventually strategic use as expectations recalibrate.</p></li></ul><h4><strong>What this means for junior engineers</strong></h4><ul><li><p><strong>Skill development may accelerate rather than disappear.</strong> Junior engineers may practice delegation, planning, and system-level thinking earlier by working with AI agents.</p></li><li><p><strong>Technical fundamentals still matter.</strong> Understanding architecture, requirements, and failure modes remains essential for supervising AI-generated work.</p></li><li><p><strong>Interpersonal skills risk being deprioritized.</strong> Managing agents is not the same as managing people, raising concerns about how collaboration skills develop over time.</p></li></ul><h4><strong>AI is not just a productivity tool</strong></h4><ul><li><p><strong>Creativity and innovation benefit from friction.</strong> Research suggests that exposing decision points and seams can create space for new ideas rather than faster repetition.</p></li><li><p><strong>Automating everything is not always desirable.</strong> Removing all toil may reduce opportunities for learning, insight, and creative problem-solving.</p></li><li><p><strong>AI should augment thinking, not replace it.</strong> Tools that surface trade-offs and choices can support better outcomes than those that simply eliminate effort.</p></li></ul><h4><strong>High-leverage AI use cases focus on toil</strong></h4><ul><li><p><strong>Developers spend only about 14% of their time writing code.</strong> Optimizing coding alone rarely leads to large productivity gains.</p></li><li><p><strong>The biggest opportunities lie in removing friction.</strong> Documentation, compliance tasks, incident response, flaky tests, and knowledge discovery consistently rank as top pain points.</p></li><li><p><strong>AI excels at work developers dislike but must still do.</strong> Automating dull, repetitive tasks can improve satisfaction and free time for meaningful work.</p></li></ul><h4><strong>Why leadership and change management matter</strong></h4><ul><li><p><strong>AI adoption is a human problem before it is a technical one.</strong> Organizations that understand developer pain points deploy AI more effectively.<br><strong>Agentic workflows amplify organizational differences.</strong> Teams with strong experimentation cultures and feedback loops move faster and with less friction.</p></li><li><p><strong>Culture determines outcomes.</strong> How leaders communicate expectations, normalize experimentation, and support learning shapes whether AI adoption succeeds or stalls.</p></li></ul><h4><strong>Looking ahead to 2026</strong></h4><ul><li><p><strong>Task parallelization is an emerging frontier.</strong> Developers are beginning to use agents to explore multiple solution paths simultaneously.</p></li><li><p><strong>Collaboration with agents will redefine productivity.</strong> Teams, not just individuals, will increasingly work alongside AI systems.</p></li><li><p><strong>Research must evolve with the work itself.</strong> New workflows will require new metrics, new telemetry, and new ways of understanding impact.</p></li></ul><h4><strong>Lessons from the METR paper</strong></h4><ul><li><p><strong>Context matters more than headlines suggest.</strong> Results showing slower performance often reflected expert developers working in familiar codebases.</p></li><li><p><strong>AI may help most where familiarity is lowest.</strong> New domains, unfamiliar systems, and onboarding scenarios show different outcomes.</p></li><li><p><strong>Media oversimplification distorts understanding.</strong> Nuance is critical when interpreting AI research, especially as studies move into real-world environments.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=155s">02:35</a>) Introducing the panel and the focus of the discussion</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=283s">04:43</a>) Why measuring AI&#8217;s impact is such a hard problem</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=330s">05:30</a>) How Microsoft approaches AI impact measurement</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=400s">06:40</a>) How Google thinks about measuring AI impact</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=448s">07:28</a>) GitHub&#8217;s perspective on measurement and insights from the DORA report</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=635s">10:35</a>) Why lines of code is a misleading metric</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=867s">14:27</a>) The limitations of measuring the percentage of code generated by AI</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1104s">18:24</a>) GitHub&#8217;s research on how AI is shaping the identity of the developer</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1299s">21:39</a>) How AI may change junior engineers&#8217; skill sets</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1482s">24:42</a>) Google&#8217;s research on using AI and creativity</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1584s">26:24</a>) High-leverage AI use cases that improve developer experience</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=1958s">32:38</a>) Open research questions for AI and developer productivity in 2026</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=2133s">35:33</a>) How leading organizations approach change and agentic workflows</p><p>(<a href="https://www.youtube.com/watch?v=xZjvYMuAJPc&amp;t=2282s">38:02</a>) Why the METR paper resonated and how it was misunderstood</p><h2><strong>Referenced:</strong></h2><p>&#8226; <a href="https://getdx.com/whitepaper/ai-measurement-framework">Measuring AI code assistants and agents</a></p><p>&#8226; <a href="https://kiro.dev/">Kiro</a></p><p>&#8226; <a href="https://code.claude.com/">Claude Code - AI coding agent for terminal &amp; IDE</a></p><p>&#8226; <a href="https://getdx.com/blog/space-framework-primer/">SPACE framework: a quick primer</a></p><p>&#8226; <a href="https://dora.dev/research/2025/dora-report/">DORA | State of AI-assisted Software Development 2025</a></p><p>&#8226; <a href="https://newsletter.pragmaticengineer.com/p/martin-fowler">Martin Fowler - by Gergely Orosz - The Pragmatic Engineer</a></p><p>&#8226; <a href="https://ieeexplore.ieee.org/document/10857384">Seamful AI for Creative Software Engineering: Use in Software Development Workflows | IEEE Journals &amp; Magazine | IEEE Xplore</a></p><p>&#8226; <a href="https://www.microsoft.com/en-us/research/publication/ai-where-it-matters-where-why-and-how-developers-want-ai-support-in-daily-work/">AI Where It Matters: Where, Why, and How Developers Want AI Support in Daily Work - Microsoft Research</a></p><p>&#8226; <a href="https://getdx.com/blog/unpacking-metri-findings-does-ai-slow-developers-down/">Unpacking METR&#8217;s findings: Does AI slow developers down?</a></p><p>&#8226; <a href="https://dxannual.com/">DX Annual 2026</a></p>]]></content:encoded></item><item><title><![CDATA[Running data-driven evaluations of AI engineering tools]]></title><description><![CDATA[A concise, data-driven framework for testing and adopting AI engineering tools.]]></description><link>https://newsletter.getdx.com/p/running-data-driven-evaluations-of</link><guid isPermaLink="false">https://newsletter.getdx.com/p/running-data-driven-evaluations-of</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 12 Dec 2025 15:49:16 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/181184615/8247b4109166deb4674678576f392b31.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/SQVdvKxzOH0">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>AI engineering tools are evolving fast. Every month brings new coding assistants, debugging agents, and automation capabilities. I want to help engineering leaders take advantage of that innovation while avoiding costly experiments that distract from real product work.</p><p>In this episode, Abi Noda and I share a practical, data-driven approach to evaluating AI tools. I walk through how to shortlist tools by use case, design structured trials that reflect real work, select representative participants, and measure impact using baselines and proven frameworks. My goal is to give you a way to test and adopt AI tools with confidence and a clear return on investment.</p><div id="youtube2-SQVdvKxzOH0" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;SQVdvKxzOH0&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/SQVdvKxzOH0?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>Data-driven evaluations are essential</strong></p><ul><li><p><strong>Structured, measurable trials prevent bias.</strong> Without them, decisions are driven by novelty hype or a few loud voices.</p></li><li><p><strong>Define a clear business outcome first</strong> (reduce toil, improve delivery speed, or raise code quality).</p></li><li><p><strong>Evaluations must inform real decisions</strong>, not just check a procurement box.</p></li></ul><p><strong>Choose the right set of tools to evaluate</strong></p><ul><li><p><strong>Group tools by use case and interaction mode</strong> (chat, agentic IDEs, code review assistants, etc.) to ensure fair comparisons.</p></li><li><p><strong>Match shortlist size to org capacity</strong> to support multiple cohorts and reliable results.</p></li><li><p><strong>Multi-vendor strategies reduce lock-in</strong> in a rapidly shifting market.</p></li></ul><p><strong>Re-evaluations are essential, not optional</strong></p><ul><li><p><strong>Incumbent tools must be retested</strong> as capabilities evolve and new challengers emerge.</p></li><li><p><strong>Triggers for re-evaluation</strong> include major feature launches, organic developer adoption of a new tool, and upcoming renewal cycles.</p></li><li><p><strong>Every challenger tool evaluation requires a baseline of the incumbent</strong>, so you can compare like-for-like.</p></li><li><p><strong>A cadence of every 8&#8211;14 months</strong> ensures decisions reflect the current reality, not the past purchase.</p></li></ul><p><strong>Design trials around research questions</strong></p><ul><li><p><strong>Start with a hypothesis.</strong> It keeps experiments aligned to actual goals.</p></li><li><p><strong>Developer sentiment is necessary but insufficient</strong> without measurable outcomes.</p></li><li><p><strong>Success criteria must be defined in advance</strong> to avoid subjective decision-making.</p></li></ul><p><strong>Select representative participants</strong></p><ul><li><p><strong>Diverse cohorts reveal real impact</strong> across languages, teams, and seniority levels.</p></li><li><p><strong>Include skeptical and late adopters</strong> to uncover onboarding and enablement needs.</p></li><li><p><strong>Volunteer-only trials distort results</strong> and won&#8217;t scale to full org rollout.</p></li></ul><p><strong>Run evaluations long enough to capture true behavior</strong></p><ul><li><p><strong>Eight to twelve weeks is the minimum</strong> to get past the novelty phase and into sustained usage.</p></li><li><p><strong>Align evaluation windows to procurement cycles</strong> so insights guide buying decisions.</p></li><li><p><strong>Short trials lead to false signals </strong>and either inflate enthusiasm or create false negativity.</p></li></ul><p><strong>Use self-reported time savings carefully</strong></p><ul><li><p><strong>Self-reporting is a strong early indicator</strong> of perceived usefulness.</p></li><li><p><strong>Humans misremember time</strong>, often benchmarking against recent AI use.</p></li><li><p><strong>Treat CSAT and time savings as directional</strong>, not the final truth.</p></li><li><p><strong>Objective metrics validate real ROI,</strong> including throughput, quality, and innovation time.</p></li></ul><p><strong>Expect variation rather than a single winner</strong></p><ul><li><p><strong>Different tools shine in different contexts</strong>, so multiple standards are often the best path.</p></li><li><p><strong>Continuous re-evaluation is required</strong> as capabilities evolve every quarter.</p></li><li><p><strong>The right goal isn&#8217;t the &#8220;best tool&#8221;</strong>, but the best tool for <em>each</em> problem space.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0">00:00</a>) Intro: Running a data-driven evaluation of AI tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=156s">02:36</a>) Challenges in evaluating AI tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=371s">06:11</a>) How often to reevaluate AI tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=422s">07:02</a>) Incumbent tools vs challenger tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=460s">07:40</a>) Why organizations need disciplined evaluations before rolling out tools</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=568s">09:28</a>) How to size your tool shortlist based on developer population</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=764s">12:44</a>) Why tools must be grouped by use case and interaction mode</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=810s">13:30</a>) How to structure trials around a clear research question</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1005s">16:45</a>) Best practices for selecting trial participants</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1162s">19:22</a>) Why support and enablement are essential for success</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1270s">21:10</a>) How to choose the right duration for evaluations</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1372s">22:52</a>) How to measure impact using baselines and the AI Measurement Framework</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1528s">25:28</a>) Key considerations for an AI tool evaluation</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1732s">28:52</a>) Q&amp;A: How reliable is self-reported time savings from AI tools?</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=1942s">32:22</a>) Q&amp;A: Why not adopt multiple tools instead of choosing just one?</p><p>(<a href="https://www.youtube.com/watch?v=SQVdvKxzOH0&amp;t=2007s">33:27</a>) Q&amp;A: Tool performance differences and avoiding vendor lock-in</p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; X: <a href="https://x.com/rhein_wein">https://x.com/rhein_wein</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p>&#8226; Laura&#8217;s course (Measuring Engineering Performance and AI Impact): <a href="https://lauratacho.com/developer-productivity-metrics-course">https://lauratacho.com/developer-productivity-metrics-course</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><p>&#8226; Substack: &#8203;&#8203;<a href="https://substack.com/@abinoda">https://substack.com/@abinoda</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/whitepaper/ai-measurement-framework/">Measuring AI code assistants and agents</a></p></li><li><p><a href="https://qconferences.com/">QCon conferences</a></p></li><li><p><a href="https://getdx.com/dx-core-4/">DX Core 4 engineering metrics</a></p></li><li><p><a href="https://getdx.com/podcast/doras-2025-research-on-the-impact-of-ai/">DORA&#8217;s 2025 research on the impact of AI</a></p></li><li><p><a href="https://getdx.com/blog/unpacking-metri-findings-does-ai-slow-developers-down/">Unpacking METR&#8217;s findings: Does AI slow developers down?</a></p></li><li><p><a href="https://newsletter.getdx.com/p/metr-study-on-how-ai-affects-developer-productivity">METR&#8217;s study on how AI affects developer productivity</a></p></li><li><p><a href="https://www.claude.com/product/claude-code">Claude Code</a></p></li><li><p><a href="https://cursor.com/">Cursor</a></p></li><li><p><a href="https://windsurf.com/">Windsurf</a></p></li><li><p><a href="https://newsletter.getdx.com/p/do-newer-ai-native-ides-outperform-other-ai-coding-assistants">Do newer AI-native IDEs outperform other AI coding assistants?</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[DORA’s 2025 research on the impact of AI]]></title><description><![CDATA[A conversation with DORA&#8217;s Nathen Harvey on how AI is transforming engineering systems and why leaders need new metrics to understand its real impact.]]></description><link>https://newsletter.getdx.com/p/doras-2025-research-on-the-impact</link><guid isPermaLink="false">https://newsletter.getdx.com/p/doras-2025-research-on-the-impact</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 21 Nov 2025 19:41:18 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/179485934/3d58eaf00c52e0f638a5ae64ee02c660.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/W1Qqr4ferz4">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this episode of <em>Engineering Enablement</em>, I sit down with Nathen Harvey, who leads research at DORA, to explore how teams should really think about measuring the impact of AI. We talk about why traditional delivery metrics can give leaders a false sense of confidence and how AI acts as an amplifier, accelerating healthy systems while intensifying existing friction and failure.</p><p>We examine findings from the 2025 DORA research on AI-assisted software development alongside DX&#8217;s Q4 AI Impact report and unpack where the data aligns and where meaningful gaps emerge. We also dig into how AI is reshaping engineering systems themselves, changing workflows, feedback loops, and team dynamics in ways leaders need to understand to achieve real, sustainable impact.</p><div id="youtube2-W1Qqr4ferz4" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;W1Qqr4ferz4&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/W1Qqr4ferz4?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>DORA metrics alone cannot measure AI impact</strong></p><ul><li><p><strong>The four &#8220;key&#8221; DORA metrics only reflect delivery outcomes, not system behavior.</strong> They show where teams end up, not how they got there.</p></li><li><p>DORA now measures five software delivery performance metrics, not four.</p></li><li><p>These metrics function like a compass rather than a diagnostic tool.</p></li><li><p><strong>Delivery performance metrics are leading indicators of organizational health</strong> but lagging indicators of engineering practices.</p></li></ul><p><strong>AI acts as an organizational amplifier</strong></p><ul><li><p><strong>AI does not fix systems; it intensifies what already exists. </strong>Strong practices compound while weak practices become more painful.</p></li><li><p>Healthy teams experience faster flow while unhealthy systems accumulate more visible friction.</p></li><li><p>AI makes hidden bottlenecks impossible to ignore.</p></li></ul><p><strong>The five DORA software delivery performance metrics</strong></p><ul><li><p><strong>DORA divides delivery performance into throughput and instability categories.</strong></p></li><li><p>Throughput metrics include lead time for changes, deployment frequency, and failed deployment recovery time.</p></li><li><p>Instability metrics include change fail rate and deployment rework rate.</p></li></ul><p><strong>DX Q4 2025 AI Impact report insights</strong></p><ul><li><p><strong>Junior engineers adopt AI more heavily than senior engineers.</strong> This shifts how work is distributed across teams.</p></li><li><p><strong>Senior engineers often capture more measurable time savings despite lower visible usage.</strong></p></li><li><p>DX found widespread experimentation with non-enterprise AI tools.</p></li><li><p>Engineers reported high AI usage even when enterprise telemetry showed no activity.</p></li><li><p>Shadow experimentation reflects weak or unclear organizational AI guidance.</p></li></ul><p><strong>The DORA AI Capabilities Model</strong></p><ul><li><p><strong>Successful AI adoption depends on team and organizational capabilities, not tool selection.</strong></p></li><li><p>A clear and communicated AI stance reduces uncertainty and speeds adoption.</p></li><li><p>A healthy internal data ecosystem prevents AI usage from being blocked by silos.</p></li><li><p><strong>Internal policies and documentation must be accessible to both humans and AI systems.</strong></p></li><li><p><strong>Strong version control provides rollback safety when AI-generated code diverges.</strong></p></li><li><p>Small batch work improves both AI output quality and system stability.</p></li><li><p>User-centered thinking ensures AI effort aligns with real human outcomes.</p></li><li><p>High-quality internal platforms allow improvements to scale across teams.</p></li></ul><p><strong>AI shifts where work breaks</strong></p><ul><li><p><strong>AI accelerates code creation but moves constraints downstream.</strong></p></li><li><p><strong>Code review becomes the dominant bottleneck</strong> under AI-assisted development.</p></li><li><p>Increased code volume without improved review systems slows overall throughput.</p></li><li><p>Bottlenecks become more visible, not less, as AI usage grows.</p></li></ul><p><strong>Measuring AI ROI requires human signals</strong></p><ul><li><p><strong>Dashboards cannot capture where work feels slow or painful</strong>.</p></li><li><p>Leaders need direct conversations with engineers about friction and workflow breakdowns.</p></li><li><p>Qualitative insight exposes failure points that metrics cannot surface.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=W1Qqr4ferz4">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=W1Qqr4ferz4&amp;t=55s">00:55</a>) Why the four key DORA metrics aren&#8217;t enough to measure AI impact</p><p>(<a href="https://www.youtube.com/watch?v=W1Qqr4ferz4&amp;t=224s">03:44</a>) The shift from four to five DORA metrics and why leaders need more than dashboards</p><p>(<a href="https://www.youtube.com/watch?v=W1Qqr4ferz4&amp;t=380s">06:20</a>) The one-sentence takeaway from the 2025 DORA report</p><p>(<a href="https://www.youtube.com/watch?v=W1Qqr4ferz4&amp;t=458s">07:38</a>) How AI amplifies both strengths and bottlenecks inside engineering systems</p><p>(<a href="https://www.youtube.com/watch?v=W1Qqr4ferz4&amp;t=538s">08:58</a>) What DX data reveals about how junior and senior engineers use AI differently</p><p>(<a href="https://www.youtube.com/watch?v=W1Qqr4ferz4&amp;t=633s">10:33</a>) The DORA AI Capabilities Model and why AI success depends on how it&#8217;s used</p><p>(<a href="https://www.youtube.com/watch?v=W1Qqr4ferz4&amp;t=1104s">18:24</a>) How a clear and communicated AI stance improves adoption and reduces friction</p><p>(<a href="https://www.youtube.com/watch?v=W1Qqr4ferz4&amp;t=1382s">23:02</a>) Why talking to your teams still matters</p><p><strong>Where to find Nathen Harvey:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/nathen">https://www.linkedin.com/in/nathen</a></p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; X: <a href="https://x.com/rhein_wein">https://x.com/rhein_wein</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p>&#8226; Laura&#8217;s course (Measuring Engineering Performance and AI Impact) <a href="https://lauratacho.com/developer-productivity-metrics-course">https://lauratacho.com/developer-productivity-metrics-course</a></p><h2><strong>Referenced:</strong></h2><p>&#8226; <a href="https://dora.dev/research/2025/dora-report/">DORA | State of AI-assisted Software Development 2025</a></p><p>&#8226; <a href="https://www.linkedin.com/in/stevefenton/">Steve Fenton - Octonaut | LinkedIn</a></p><p>&#8226; <a href="https://getdx.com/report/ai-assisted-engineering-q4-impact-report/?utm_source=podcast">AI-assisted engineering: Q4 impact report</a></p>]]></content:encoded></item><item><title><![CDATA[How Monzo runs data-driven AI experimentation]]></title><description><![CDATA[Monzo&#8217;s Fabien Deshayes shares how the bank is adopting AI responsibly&#8212;running structured trials, measuring impact, and balancing innovation with control.]]></description><link>https://newsletter.getdx.com/p/how-monzo-runs-data-driven-ai-experimentation</link><guid isPermaLink="false">https://newsletter.getdx.com/p/how-monzo-runs-data-driven-ai-experimentation</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 31 Oct 2025 16:05:48 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/177291638/f983aaaf456b2fa6b7d448c8f58a1dff.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/j0KLLLQ_yzs">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this episode of <em>Engineering Enablement</em>, I talk with Fabien Deshayes, who leads multiple platform engineering teams at Monzo Bank. Fabien shares how Monzo is adopting AI responsibly within a highly regulated industry, balancing innovation with control, structure, and strong data-driven decision-making.</p><p>We discuss how Monzo runs structured AI trials, measures adoption and satisfaction, and uses metrics to guide investment and training. Fabien explains why they moved from broad rollouts to small, focused cohorts, how they are addressing existing PR review bottlenecks that AI has intensified, and what they have learned from empowering product managers and designers to use AI tools directly.</p><p>He also shares insights into budgeting and experimentation, the results they are seeing from AI-assisted engineering, and his vision for what is next, from agent orchestration to more seamless collaboration across roles.</p><div id="youtube2-j0KLLLQ_yzs" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;j0KLLLQ_yzs&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/j0KLLLQ_yzs?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>Evaluating AI in a regulated industry</strong></p><ul><li><p><strong>Monzo treats AI enablement as a platform responsibility.</strong> The platform group centralizes tooling to ensure guardrails are in place for privacy, data access, and model usage across regulated domains.</p></li><li><p><strong>Structure comes before scale.</strong> Before expanding access, Monzo limits tools to trained users and controlled contexts to reduce risk while maintaining speed of innovation.</p></li><li><p><strong>Regulation became an advantage.</strong> The constraints of banking forced deliberate decision-making, leading to a tighter focus, better change management, and stronger adoption.</p></li></ul><p><strong>Experimentation strategy</strong></p><ul><li><p><strong>AI adoption began with GitHub Copilot but lacked direction.</strong> Early usage was inconsistent, prompting Monzo to create structured trials with clear evaluation goals.</p></li><li><p><strong>Each trial follows a defined process.</strong> The team selects one or two tools at a time, identifies specific use cases, and measures success against defined metrics such as adoption, satisfaction, and retention.</p></li><li><p><strong>Trials are intentionally small and focused.</strong> Monzo learned that large-scale rollouts dilute feedback and waste time. Smaller cohorts of engineers with defined use cases yield higher-quality insights.</p></li></ul><p><strong>Evaluation criteria and metrics</strong></p><ul><li><p><strong>Success is measured across ten to fifteen criteria.</strong> These include weekly and monthly active users, retention, percentage of AI-written code, suggestion acceptance rate, satisfaction, and cost.</p></li><li><p><strong>Use-case analysis is part of every trial.</strong> Tools are compared across workflows like refactoring, documentation, migrations, and new feature creation to understand strengths and weaknesses.</p></li><li><p><strong>Quantitative data is paired with survey feedback.</strong> Combining usage telemetry with engineer sentiment helps Monzo decide whether to scale a tool or end a trial.</p></li></ul><p><strong>Budgeting and cost control</strong></p><ul><li><p><strong>Monzo sets clear spending limits before experimentation.</strong> After modeling expected usage, the team landed on a target of roughly $1,000 per engineer per year for AI tooling.</p></li><li><p><strong>Budget discussions include leadership from the start.</strong> Transparent alignment with the CTO and finance ensures experimentation can continue without surprise costs.</p></li><li><p><strong>Cost is balanced against value.</strong> Data from tool providers and internal telemetry validate ROI, while quarterly reviews adjust allocations based on usage trends.</p></li></ul><p><strong>Engineering outcomes</strong></p><ul><li><p><strong>AI is increasing throughput and code volume.</strong> Pull requests per engineer are up 10&#8211;20%, with the average PR size growing by a similar amount. Around 20% of new code is now AI-generated.</p></li><li><p><strong>Code review time is the new bottleneck.</strong> Larger, AI-assisted PRs take longer to review, creating a productivity drag that Monzo is actively addressing.</p></li><li><p><strong>Preview environments are solving review delays.</strong> New &#8220;tenancies&#8221; allow engineers to test changes in isolated environments, improving speed and stability across teams.</p></li></ul><p><strong>Cross-functional adoption</strong></p><ul><li><p><strong>AI use is spreading beyond engineering.</strong> Product managers and designers now prototype directly using AI tools integrated with Monzo&#8217;s design system.</p></li><li><p><strong>Better context leads to better results.</strong> By exposing design components and brand guidelines to models, teams produce higher-quality prototypes that reduce rework later in development.</p></li></ul><p><strong>Leadership and enablement</strong></p><ul><li><p><strong>Community drives adoption more than mandates.</strong> AI Engineering Champions across departments test tools, share learnings, and help others apply best practices in their collectives.</p></li><li><p><strong>Upskilling is Monzo&#8217;s next investment.</strong> With only half of engineers feeling confident using AI, the focus is shifting to prompting education, internal workshops, and structured learning.</p></li><li><p><strong>Success stories accelerate momentum.</strong> Highlighting early wins through Slack channels and company meetings keeps engineers engaged and inspired to experiment.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=61s">01:01</a>) An overview of Monzo bank and Fabien&#8217;s role</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=125s">02:05</a>) Monzo&#8217;s careful, structured approach to AI experimentation</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=330s">05:30</a>) How Monzo&#8217;s AI journey began</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=386s">06:26</a>) Why Monzo chose a structured approach to experimentation and what criteria they used</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=561s">09:21</a>) How Monzo selected AI tools for experimentation</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=711s">11:51</a>) Why individual tool stipends don&#8217;t work for large, regulated organizations</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=932s">15:32</a>) How Monzo measures the impact of AI tools and uses the data</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=1090s">18:10</a>) Why Monzo limits AI tool trials to small, focused cohorts</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=1254s">20:54</a>) The phases of Monzo&#8217;s AI rollout and how learnings are shared across the organization</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=1363s">22:43</a>) What Monzo&#8217;s data reveals about AI usage and spending</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=1470s">24:30</a>) How Monzo balances AI budgeting with innovation</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=1605s">26:45</a>) Results from DX&#8217;s spending poll and general advice on AI budgeting</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=1683s">28:03</a>) What Monzo&#8217;s data shows about AI&#8217;s impact on engineering performance</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=1790s">29:50</a>) The growing bottleneck in PR reviews and how Monzo is solving it with tenancies</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=2034s">33:54</a>) How product managers and designers are using AI at Monzo</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=2196s">36:36</a>) Fabien&#8217;s advice for moving the needle with AI adoption</p><p>(<a href="https://www.youtube.com/watch?v=j0KLLLQ_yzs&amp;t=2322s">38:42</a>) The biggest changes coming next in AI engineering</p><p><strong>Where to find Fabien Deshayes:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/fabiendeshayes">https://www.linkedin.com/in/fabiendeshayes</a></p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; X: <a href="https://x.com/rhein_wein">https://x.com/rhein_wein</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p>&#8226; Laura&#8217;s course (Measuring Engineering Performance and AI Impact) <a href="https://lauratacho.com/developer-productivity-metrics-course">https://lauratacho.com/developer-productivity-metrics-course</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://monzo.com/">Monzo</a></p></li><li><p><a href="https://go.dev/">The Go Programming Language</a></p></li><li><p><a href="https://www.swift.org/">Swift.org</a></p></li><li><p><a href="https://kotlinlang.org/">Kotlin</a></p></li><li><p><a href="https://code.visualstudio.com/docs/copilot/overview">GitHub Copilot in VS Code</a></p></li><li><p><a href="https://cursor.com/">Cursor</a></p></li><li><p><a href="https://windsurf.com/">Windsurf</a></p></li><li><p><a href="https://www.claude.com/product/claude-code">Claude Code</a></p></li><li><p><a href="https://getdx.com/podcast/planning-2026-ai-tooling-budget/">Planning your 2026 AI tooling budget: guidance for engineering leaders</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Planning your 2026 AI tooling budget: guidance for engineering leaders]]></title><description><![CDATA[Abi Noda and I share how engineering leaders can plan effective 2026 AI budgets that balance innovation, cost, and measurable ROI.]]></description><link>https://newsletter.getdx.com/p/planning-your-2026-ai-tooling-budget</link><guid isPermaLink="false">https://newsletter.getdx.com/p/planning-your-2026-ai-tooling-budget</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 17 Oct 2025 16:23:54 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/176199495/b680e7d998be9e828e3476df4306e58e.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/myUmFrxxUOk">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this webinar, Abi Noda and I explore how engineering leaders can plan their 2026 AI budgets amid rapid change and rising costs. Using data from DX&#8217;s recent poll and industry benchmarks, we break down how much to spend per developer, how to allocate budgets across tools, and how to balance innovation with cost control.</p><p>We also discuss practical strategies for building a multi-vendor approach, choosing the right metrics before and after adoption, and proving ROI through continuous measurement. Along the way, we share insights on communicating AI&#8217;s value to executives, avoiding cost-cutting narratives, and investing in enablement and training to make adoption last.</p><div id="youtube2-myUmFrxxUOk" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;myUmFrxxUOk&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/myUmFrxxUOk?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>Planning Your 2026 AI Budget</strong></p><ul><li><p><strong>Budgets are shifting from experimental to essential.</strong> In 2024&#8211;2025, most AI tool spending was discretionary and exploratory, but by 2026, it becomes a recurring line item with clear ROI expectations.</p></li><li><p><strong>Expect rising costs.</strong> Tools are expanding their capabilities and pricing accordingly; organizations should plan for year-over-year increases.</p></li><li><p><strong>Average spending is climbing.</strong> The floor for AI tooling is now about <strong>$500&#8211;$1,000 per developer per year</strong>, but multi-vendor setups can easily push that higher.</p></li></ul><p><strong>Allocating your AI budget</strong></p><ul><li><p><strong>Plan for diversity across the SDLC.</strong> Modern AI adoption means using multiple tools&#8212;chat interfaces, IDE copilots, background agents, and specialized tools&#8212;to cover distinct use cases.</p></li><li><p><strong>Shift from headcount expansion to efficiency.</strong> Many companies are reallocating budget from new hires to automation and tooling&#8212;not cutting jobs, but slowing growth to focus on productivity.</p></li><li><p><strong>Track budget maturity.</strong> Move from &#8220;experimental&#8221; AI spending to a structured budget that includes telemetry, measurement, and enablement layers.</p></li></ul><p><strong>Vendor and tooling strategy</strong></p><ul><li><p><strong>A multi-vendor approach is now the norm.</strong> Locking into one platform risks missing out on faster-evolving tools and model improvements.</p></li><li><p><strong>Enterprise licensing vs. stipends.</strong> Enterprise plans streamline enablement and reporting but carry overage risks; stipends are predictable but lack team-level visibility and economies of scale.</p></li><li><p><strong>Cost control is an active process.</strong> Use telemetry and regular reviews to ensure spending aligns with actual value delivered across teams.</p></li></ul><p><strong>Measurement and ROI</strong></p><ul><li><p><strong>Use data before and after adoption.</strong> Measure both during proof-of-concepts and in production to understand which tools truly move the needle.</p></li><li><p><strong>The right metrics matter.</strong> Track usage tiers (daily, weekly, monthly), time savings, developer satisfaction, and percentage of AI-generated code&#8212;while acknowledging how hard these are to measure accurately.</p></li><li><p><strong>ROI is more than throughput.</strong> Developers spend only about 14% of their time coding; improving that slice doesn&#8217;t guarantee overall productivity gains.</p></li></ul><p><strong>Communication and leadership</strong></p><ul><li><p><strong>Frame AI as acceleration, not automation.</strong> Messaging that focuses on cost-cutting can backfire&#8212;leaders should position AI as a way to stay competitive, not reduce headcount.</p></li><li><p><strong>Language matters.</strong> Talk about <em>time recaptured</em> or <em>capacity gained</em> instead of cost savings to avoid triggering budget cuts.</p></li><li><p><strong>Leadership advocacy drives adoption.</strong> Developers are far more likely to use AI tools daily when leaders promote them clearly and consistently.</p></li></ul><p><strong>Enablement and training</strong></p><ul><li><p><strong>Training is non-negotiable.</strong> AI adoption stalls without enablement; companies investing in experiential training see stronger, sustained usage.</p></li><li><p><strong>High-quality programs pay off.</strong> Workshops and accelerators that pair real business problems with hands-on learning cost roughly <strong>$500&#8211;$2,000 per developer</strong>&#8212;and are worth every dollar.</p></li><li><p><strong>Measure enablement ROI.</strong> Track adoption rates before and after training to validate impact and justify continued investment.</p></li></ul><p><strong>Trends to watch</strong></p><ul><li><p><strong>AI add-ons are expanding.</strong> Expect price hikes as existing tools add AI features&#8212;often as premium tiers or separate modules.</p></li><li><p><strong>Custom models are not yet mainstream.</strong> Fine-tuning and bespoke models are still limited to large enterprises; most companies can focus on leveraging general models effectively.</p></li><li><p><strong>Data-driven decisions will define leaders.</strong> Teams that collect usage data, track ROI, and adjust budgets proactively will outpace those relying on assumptions.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk">00:00</a>) Intro: Setting the stage for AI budgeting in 2026</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=105s">01:45</a>) Results from DX&#8217;s AI spending poll and early trends</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=210s">03:30</a>) How companies are currently spending and what to watch in 2026</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=292s">04:52</a>) Why clear definitions for AI tools matter and how Laura and Abi think about them</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=432s">07:12</a>) The entry point for 2026 AI tooling budgets and emerging spending patterns</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=614s">10:14</a>) Why 2026 is the year to prove ROI on AI investments</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=670s">11:10</a>) How organizations should approach AI budgeting and allocation</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=908s">15:08</a>) Best practices for managing AI vendors and enterprise licensing</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=1022s">17:02</a>) How to define and choose metrics before and after adopting AI tools</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=1170s">19:30</a>) How to identify bottlenecks and AI use cases with the highest ROI</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=1318s">21:58</a>) Key considerations for AI budgeting</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=1510s">25:10</a>) Why AI investments are about competitiveness, not cost-cutting</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=1639s">27:19</a>) How to use the right language to build trust and executive buy-in</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=1698s">28:18</a>) Why training and enablement are essential parts of AI investment</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=1900s">31:40</a>) How AI add-ons may increase your tool costs</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=1967s">32:47</a>) Why custom and fine-tuned models aren&#8217;t relevant for most companies today</p><p>(<a href="https://www.youtube.com/watch?v=myUmFrxxUOk&amp;t=2040s">34:00</a>) The tradeoffs between stipend models and enterprise AI licenses</p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><p>&#8226; Substack: &#8203;&#8203;<a href="https://substack.com/@abinoda">https://substack.com/@abinoda</a></p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; X: <a href="https://x.com/rhein_wein">https://x.com/rhein_wein</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p>&#8226; Laura&#8217;s course (Measuring Engineering Performance and AI Impact): <a href="https://lauratacho.com/developer-productivity-metrics-course">https://lauratacho.com/developer-productivity-metrics-course</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p></li><li><p><a href="https://getdx.com/research/measuring-ai-code-assistants-and-agents/">Measuring AI code assistants and agents</a></p></li><li><p><a href="https://www.iconiqcapital.com/growth/reports/2025-state-of-ai">2025 State of AI Report: The Builder&#8217;s Playbook</a></p></li><li><p><a href="https://github.com/features/copilot">GitHub Copilot &#183; Your AI pair programmer</a></p></li><li><p><a href="https://cursor.com/">Cursor</a></p></li><li><p><a href="https://www.glean.com/">Glean</a></p></li><li><p><a href="https://www.claude.com/product/claude-code">Claude Code</a></p></li><li><p><a href="https://chatgpt.com/">ChatGPT</a></p></li><li><p><a href="https://windsurf.com/">Windsurf</a></p></li><li><p><a href="https://getdx.com/blog/dx-releases-integration-with-claude-code/">Track Claude Code adoption, impact, and ROI, directly in DX</a></p></li><li><p><a href="https://getdx.com/podcast/measuring-ai-code-assistants-ai-framework/">Measuring AI code assistants and agents with the AI Measurement Framework</a></p></li><li><p><a href="https://getdx.com/podcast/enterprise-wide-ai-tool-adoption/">Driving enterprise-wide AI tool adoption</a></p></li><li><p><a href="https://sentry.io/welcome/">Sentry</a></p></li><li><p><a href="https://poolside.ai/">Poolside</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[The evolving role of DevProd teams in the AI era]]></title><description><![CDATA[I sit down with DX CTO Laura Tacho to share how AI is transforming Platform and DevProd teams, from tool standardization to reducing tech debt and modernizing workflows.]]></description><link>https://newsletter.getdx.com/p/the-evolving-role-of-devprod-teams</link><guid isPermaLink="false">https://newsletter.getdx.com/p/the-evolving-role-of-devprod-teams</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 26 Sep 2025 15:36:54 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/173865452/a1e7e3ee0db2f8e8f86f402e17f8b6ed.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/Q8wMGNzWc7s">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>Today, I am joined once again by our CTO, Laura Tacho, to explore how the role of Platform and DevProd teams is evolving in the AI era. We discuss why evaluating and rolling out AI tools is now a core responsibility, how measurement frameworks must change, and why fundamentals like documentation and feedback loops matter more than ever for both developers and AI agents. We also dive into strategies for handling tool sprawl, hardening pipelines for increased throughput, and applying AI at scale to tech debt, migrations, and workflows across the SDLC.</p><div id="youtube2-Q8wMGNzWc7s" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;Q8wMGNzWc7s&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/Q8wMGNzWc7s?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>Platform teams in transition</strong></p><ul><li><p><strong>Platform teams are redefining their mandate in the AI era.</strong> They&#8217;re moving beyond CI/CD into AI evaluation, rollout, and measurement, taking on more strategic influence across the SDLC.</p></li><li><p><strong>Expectations on leaders are rising.</strong> Platform teams are now expected to guide both productivity and AI adoption, balancing technical and cultural change.</p></li></ul><p><strong>Defining the platform role</strong></p><ul><li><p><strong>The platform role is hard to define and AI raises the stakes.</strong> Teams must bridge old responsibilities with new AI-driven expectations.</p></li><li><p><strong>Clarity creates leverage.</strong> Clear mandates help leaders align stakeholders and secure resources.</p></li></ul><p><strong>Evaluation and measurement</strong></p><ul><li><p><strong>Evaluating and rolling out AI tools is now a Platform role.</strong> Platform leaders handle vendor selection, pilots, and enterprise rollouts while tracking impact.</p></li><li><p><strong>Strong measurement frameworks are essential.</strong> Teams need data-driven ways to evaluate AI&#8217;s impact beyond lines of code, blending old and new metrics for a realistic picture.</p></li><li><p><strong>Platform leaders are internal educators.</strong> Explaining metrics to executives avoids hype-driven decisions and builds trust in AI-assisted development.</p></li></ul><p><strong>Hardening systems and guardrails</strong></p><ul><li><p><strong>AI-generated code stresses pipelines and build systems.</strong> Platform teams must harden CI/CD, testing, and local environments to handle larger batch sizes and faster iterations.</p></li><li><p><strong>Guardrails complement throughput.</strong> Quality checks, security reviews, and feedback loops maintain reliability while speed accelerates, laying groundwork for large-scale refactors and technical debt initiatives.</p></li></ul><p><strong>Standardization and knowledge sharing</strong></p><ul><li><p><strong>Training alone isn&#8217;t enough &#8212; standardization creates leverage.</strong> Curated workflows and reusable templates help maintain consistent AI adoption.</p></li><li><p><strong>Standardizing tools and knowledge multiplies impact.</strong> Shared workflows and internal documentation reduce duplication and accelerate learning.</p></li></ul><p><strong>Applying AI at scale</strong></p><ul><li><p><strong>Platform teams can apply AI directly for high-leverage work.</strong> Code migrations, technical debt reduction, and repetitive tasks become one-to-many accelerators rather than personal productivity boosts.</p></li><li><p><strong>Context as a service matters.</strong> Centralizing setup and best practices allows product teams to focus on business problems while platform teams handle complexity and modernization.</p></li><li><p><strong>AI tackles technical debt at scale.</strong> Platform teams can orchestrate migrations, refactors, and other neglected work for organization-wide benefit.</p></li></ul><p><strong>Focusing on fundamentals and the big picture</strong></p><ul><li><p><strong>Fundamentals benefit both developers and AI agents.</strong> Documentation, feedback loops, and clean codebases improve outcomes for LLMs just as they do for human developers.</p></li><li><p><strong>AI doesn&#8217;t erase old bottlenecks.</strong> Meetings, interruptions, and process inefficiencies still outweigh many AI productivity gains, making platform leaders the stewards of the bigger picture.</p></li></ul><p><strong>Looking beyond code authoring</strong></p><ul><li><p><strong>Opportunities go far beyond code authoring.</strong> Requirements, testing, validation, and even documentation are ripe areas for AI leverage.</p></li><li><p><strong>End-to-end adoption multiplies gains.</strong> Companies see the biggest improvements when AI is applied across the SDLC, creating a stronger developer experience overall.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s">00:00</a>) Intro: Why platform teams need to evolve</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=154s">02:34</a>) The challenge of defining platform teams and how AI is changing expectations</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=284s">04:44</a>) Why evaluating and rolling out AI tools is becoming a core platform responsibility</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=434s">07:14</a>) Why platform teams need solid measurement frameworks to evaluate AI tools</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=536s">08:56</a>) Why platform leaders should champion education and advocacy on measurement</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=680s">11:20</a>) How AI code stresses pipelines and why platform teams must harden systems</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=744s">12:24</a>) Why platform teams must go beyond training to standardize tools and create workflows</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=871s">14:31</a>) How platform teams control tool sprawl</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=982s">16:22</a>) Why platform teams need strong guardrails and safety checks</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=1121s">18:41</a>) The importance of standardizing tools and knowledge</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=1184s">19:44</a>) The opportunity for platform teams to apply AI at scale across the organization</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=1420s">23:40</a>) Quick recap of the key points so far</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=1473s">24:33</a>) How AI helps modernize legacy code and handle migrations</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=1545s">25:45</a>) Why focusing on fundamentals benefits both developers and AI agents</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=1662s">27:42</a>) Identifying SDLC bottlenecks beyond AI code generation</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=1808s">30:08</a>) Techniques for optimizing legacy code bases</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=1967s">32:47</a>) How AI helps tackle tech debt and large-scale code migrations</p><p>(<a href="https://www.youtube.com/watch?v=Q8wMGNzWc7s&amp;t=2140s">35:40</a>) Tools across the SDLC</p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><p>&#8226; Substack: &#8203;&#8203;<a href="https://substack.com/@abinoda">https://substack.com/@abinoda</a></p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; X: <a href="https://x.com/rhein_wein">https://x.com/rhein_wein</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p>&#8226; Laura&#8217;s course (Measuring Engineering Performance and AI Impact): <a href="https://lauratacho.com/developer-productivity-metrics-course">https://lauratacho.com/developer-productivity-metrics-course</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p></li><li><p><a href="https://getdx.com/whitepaper/ai-measurement-framework/">Measuring AI code assistants and agents</a></p></li><li><p><a href="https://www.linkedin.com/posts/abinoda_many-platform-teams-are-stuck-optimizing-activity-7356687949532483586-W-zG?utm_source=share&amp;utm_medium=member_desktop&amp;rcm=ACoAABf37PYBgozFf00ihr4fkqjRtMnFajHkQ5E">Abi Noda&#8217;s LinkedIn post</a></p></li><li><p><a href="https://getdx.com/podcast/measuring-ai-code-assistants-ai-framework/">Measuring AI code assistants and agents with the AI Measurement Framework</a></p></li><li><p><a href="https://getdx.com/blog/space-metrics/">The SPACE framework: A comprehensive guide to developer productivity</a></p></li><li><p><a href="https://docs.anthropic.com/en/docs/claude-code/common-workflows">Common workflows - Anthropic</a></p></li><li><p><a href="https://itrevolution.com/product/enterprise-tech-leadership-summit-las-vegas/">Enterprise Tech Leadership Summit Las Vegas 2025</a></p></li><li><p><a href="https://getdx.com/podcast/enterprise-wide-ai-tool-adoption/">Driving enterprise-wide AI tool adoption with Bruno Passos</a></p></li><li><p><a href="https://medium.com/airbnb-engineering/accelerating-large-scale-test-migration-with-llms-9565c208023b">Accelerating Large-Scale Test Migration with LLMs | by Charles Covey-Brandt | The Airbnb Tech Blog | Medium</a></p></li><li><p><a href="https://www.linkedin.com/in/justinreock/">Justin Reock - DX | LinkedIn</a></p></li><li><p><a href="https://www.businessinsider.com/devgen-ai-tool-saved-morgan-stanley-280-000-hours-jobs-2025-7">A New Tool Saved Morgan Stanley More Than 280,000 Hours This Year - Business Insider</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Lessons from Twilio’s multi-year platform consolidation]]></title><description><![CDATA[Twilio&#8217;s Jesse Adametz shares lessons on unifying tech stacks, platform adoption, developer experience, and a pragmatic approach to AI.]]></description><link>https://newsletter.getdx.com/p/lessons-from-twilios-multi-year-platform</link><guid isPermaLink="false">https://newsletter.getdx.com/p/lessons-from-twilios-multi-year-platform</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 12 Sep 2025 14:00:12 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/172890232/38d079cd12cec990735d2ce46377ee07.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/arzNyxQ-NEE">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>Jesse Adametz is a Senior Engineering Leader on the Developer Platform at Twilio, where he&#8217;s leading a multi-year effort to unify tech stacks across major acquisitions. I&#8217;m joined by him today to talk about the realities of platform adoption and migration, the challenges of Kubernetes, and why mandates don&#8217;t work when you&#8217;re bringing together strong engineering cultures. We also dive into why Jesse treats developer experience as a product, what &#8220;change as a service&#8221; looks like in practice, and how Twilio is approaching AI with pragmatism, experimentation, and a focus on real value.</p><div id="youtube2-arzNyxQ-NEE" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;arzNyxQ-NEE&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/arzNyxQ-NEE?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>Platform adoption and migrations</strong></p><ul><li><p><strong>Consolidation is multi-year and non-linear.</strong> Twilio&#8217;s integration of Segment, SendGrid, and core stacks showed that platform consolidation is never a straight path. Early decisions, like launching EKS in Europe first for data residency before migrating U.S. workloads, highlight the inevitable &#8220;zigzags.&#8221;</p></li><li><p><strong>Migrations are about people as much as tech.</strong> Developers won&#8217;t abandon working systems on command. Advocacy, training, and executive alignment are critical &#8212; Jesse points to a past effort to put all engineers on call that only worked with CEO buy-in and steady rollout.</p></li><li><p><strong>Balance greenfield and brownfield.</strong> Twilio puts new services onto golden paths such as standardized Java pipelines, while gradually remediating older, bespoke setups. Greenfield adoption refines the standards that brownfield teams follow once value is proven.</p></li><li><p><strong>Carrots work better than mandates.</strong> Scorecards and contextual quality checks nudge teams toward best practices, while hard stops are reserved for critical reliability or security gaps.</p></li></ul><p><strong>Developer experience as a product</strong></p><ul><li><p><strong>Treat platforms like products.</strong> Twilio surveys developers, tracks NPS, and requires alpha customers before building new tools. If no teams are ready to test, platform focuses on paying down tech debt instead.</p></li><li><p><strong>Centralize the cost of change.</strong> Rather than hundreds of teams repeating the same migrations, platform absorbs the toil, standardizing and automating on-ramps so engineers get smoother adoption paths.</p></li><li><p><strong>Change as a service reduces pain.</strong> The long-term goal is pre-tested PRs with confidence scores, making migrations as simple as review and merge. This requires standardized tests, frameworks, and rollback strategies to scale safely.</p></li><li><p><strong>Show value early and often.</strong> Developers move when they see clear improvements. Dashboards highlight service quality gains and consolidated metrics roll into leadership reviews to reinforce progress across levels.</p></li></ul><p><strong>Kubernetes and modernization</strong></p><ul><li><p><strong>Kubernetes isn&#8217;t always the answer.</strong> While it solved many problems, Twilio&#8217;s telephony and SIP workloads didn&#8217;t fit neatly. Forcing Kubernetes everywhere would have created more pain than value.</p></li><li><p><strong>Modernization is the bigger picture.</strong> Twilio emphasizes improvements like autoscaling groups, observability, and simplifying legacy systems rather than betting on a single orchestrator.</p></li><li><p><strong>Unify the developer experience.</strong> A consistent experience across stacks matters more than enforcing a single tool. Developers should interact with one UX regardless of what&#8217;s running underneath.</p></li><li><p><strong>Reveal complexity only when needed.</strong> Golden paths hide infrastructure details for most engineers, with advanced controls surfaced only for high-scale, internet-facing workloads.</p></li></ul><p><strong>AI adoption</strong></p><ul><li><p><strong>AI hype feels different from Kubernetes.</strong> Kubernetes shifted work sideways, but AI promises to remove repetitive tasks and elevate workflows. Adoption requires structured experiments, benchmarks, and cultural adaptation.</p></li><li><p><strong>AI is both art and science.</strong> Clear PRDs and strong prompting skills dramatically improve AI outputs, and platform teams should train developers in these practices.</p></li><li><p><strong>AI changes workflows, not just tools.</strong> Engineers spend less time writing code and more time reviewing it, requiring higher throughput and different habits.</p></li><li><p><strong>Experimentation and standardization must coexist.</strong> Twilio&#8217;s review boards enforce safety and vendor discipline while greenfield teams run short feedback loops to test new tools.</p></li><li><p><strong>Guilds expose platform gaps.</strong> The AI DevEx guild drew 400 members instantly, showing demand for shared learning. Over time, platform should absorb this role instead of leaving it to grassroots groups.</p></li><li><p><strong>AI can augment incidents, not replace engineers.</strong> Vendors promising self-healing agents are overhyped, but AI already helps with anomaly detection and other narrow, reliable applications.</p></li></ul><p><strong>Tool evaluation and reliability</strong></p><ul><li><p><strong>Pricing should encourage healthy behavior.</strong> Per-run or per-policy models discourage usage and visibility. Twilio favors per-seat or platform-level pricing that scales with engineering needs.</p></li><li><p><strong>Reliability is non-negotiable.</strong> As an infrastructure provider, Twilio demands &#8220;dial-tone&#8221; reliability from its vendors, weighing monitoring, availability, and scale just as heavily as price.</p></li></ul><p><strong>Support as a platform capability</strong></p><ul><li><p><strong>Internal developers deserve customer-grade support.</strong> Today&#8217;s fragmented Slack channels create noise; a single entry point ensures clarity and visibility.</p></li><li><p><strong>Support requests reveal automation opportunities.</strong> Analyzing tickets showed that automating just two processes could reduce requests by double digits.</p></li><li><p><strong>A single &#8220;oh button&#8221; sets the standard.</strong> One clear way to ask for help, backed by automation and human fallback, gives developers the same reliability expected in customer support.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=90s">01:30</a>) Jesse&#8217;s background and how he ended up at Twilio</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=240s">04:00</a>) What SRE teaches leaders and ICs</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=366s">06:06</a>) Where Twilio started the post-acquisition integration</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=502s">08:22</a>) Why platform migrations can&#8217;t follow a straight-line plan</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=605s">10:05</a>) How Twilio balances multiple strategies for migrations</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=750s">12:30</a>) The human side of change: advocacy, training, and alignment</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=1066s">17:46</a>) Treating developer experience as a first-class product</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=1300s">21:40</a>) What &#8220;change as a service&#8221; looks like in practice</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=1497s">24:57</a>) A mandateless approach: creating voluntary adoption through value</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=1730s">28:50</a>) How Twilio demonstrates value with metrics and reviews</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=1841s">30:41</a>) Why Kubernetes wasn&#8217;t the right fit for all Twilio workloads</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=2172s">36:12</a>) How Twilio decides when to expose complexity</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=2303s">38:23</a>) Lessons from Kubernetes hype and how AI demands more experimentation</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=2688s">44:48</a>) Where AI fits into Twilio&#8217;s platform strategy</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=2985s">49:45</a>) How guilds fill needs the platform team hasn&#8217;t yet met</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=3077s">51:17</a>) The future of platform in centralizing knowledge and standards</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=3272s">54:32</a>) How Twilio evaluates tools for fit, pricing, and reliability</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=3473s">57:53</a>) Where Twilio applies AI in reliability, and where Jesse is skeptical</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=3566s">59:26</a>) Laura&#8217;s vibe-coded side project built on Twilio</p><p>(<a href="https://www.youtube.com/watch?v=arzNyxQ-NEE&amp;t=3671s">1:01:11</a>) How external lessons shape Twilio&#8217;s approach to platform support and docs</p><p><strong>Where to find Jesse Adametz:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/jesseadametz/">https://www.linkedin.com/in/jesseadametz/</a></p><p>&#8226; X: <a href="https://x.com/jesseadametz">https://x.com/jesseadametz</a></p><p>&#8226; Website: <a href="https://www.jesseadametz.com/">https://www.jesseadametz.com/</a></p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; X: <a href="https://x.com/rhein_wein">https://x.com/rhein_wein</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p>&#8226; Laura&#8217;s course (Measuring Engineering Performance and AI Impact) <a href="https://lauratacho.com/developer-productivity-metrics-course">https://lauratacho.com/developer-productivity-metrics-course</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/whitepaper/ai-measurement-framework">The AI Measurement Framework</a></p></li><li><p><a href="https://www.experian.com/">Experian</a></p></li><li><p><a href="https://en.wikipedia.org/wiki/Transact-SQL">Transact-SQL - Wikipedia</a></p></li><li><p><a href="https://www.twilio.com/">Twilio</a></p></li><li><p><a href="https://kubernetes.io/">Kubernetes</a></p></li><li><p><a href="http://copilot.microsoft.com">Copilot</a></p></li><li><p><a href="https://www.anthropic.com/claude-code">Claude Code</a></p></li><li><p><a href="https://windsurf.com/">Windsurf</a></p></li><li><p><a href="https://cursor.com/">Cursor</a></p></li><li><p><a href="https://aws.amazon.com/bedrock/">Bedrock</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Driving enterprise-wide AI tool adoption]]></title><description><![CDATA[Booking.com&#8217;s Bruno Passos shares how the company rolled out AI to 3,000 engineers, overcame adoption and cultural hurdles, and built practices that turned skeptics into daily users.]]></description><link>https://newsletter.getdx.com/p/driving-enterprise-wide-ai-tool-adoption</link><guid isPermaLink="false">https://newsletter.getdx.com/p/driving-enterprise-wide-ai-tool-adoption</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 05 Sep 2025 15:40:37 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/172055063/da8d6c6e4c06377fd6586d1f77352b56.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/HuDtj_xHhis">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this episode, I talk with Bruno Passos, Product Lead for Developer Experience at Booking.com, about how the company is rolling out AI tools across a 3,000-person engineering team.</p><p>Bruno shares how Booking.com set ambitious innovation goals, why cultural change mattered as much as technology, and the education practices that turned hesitant developers into daily users. We also discuss the early barriers, from low adoption and knowledge gaps to procurement hurdles, and the interventions that worked, including learning paths, hackathon-style workshops, Slack communities, and centralized procurement. Today, Booking.com is already in the top 25 percent of companies for AI adoption.</p><div id="youtube2-HuDtj_xHhis" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;HuDtj_xHhis&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/HuDtj_xHhis?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>Booking.com&#8217;s goals for AI adoption</strong></p><ul><li><p><strong>The goal was to accelerate development and remove toil.</strong> Engineers could spend more time on innovation instead of maintenance.</p></li><li><p><strong>Booking.com set an ambitious &#8220;innovation ratio.&#8221;</strong> They aimed for 80&#8211;90% of engineer time on creative work, compared to a global top 10% benchmark of 65%.</p></li><li><p><strong>AI was also a lever for modernization.</strong> Even with a legacy codebase, leaders believed it could help re-platform, clean up technical debt, and shorten cycle times.</p></li><li><p><strong>The mindset was experimental from the start.</strong> AI wasn&#8217;t introduced because it was trendy but as a series of controlled experiments to uncover real business value.</p></li></ul><p><strong>Building adoption through enablement</strong></p><ul><li><p><strong>Adoption grew only after targeted enablement.</strong> Curated learning paths and structured training helped developers build confidence.</p></li><li><p><strong>Two-day learning sprints became a turning point.</strong> Day one focused on prompt engineering and context handling, while day two paired developers with real business problems.</p></li><li><p><strong>Developers brought their own problems to hackathons.</strong> This made training immediately relevant and produced artifacts that solved real issues.</p></li><li><p><strong>Experience-based accelerators (EBAs) scaled the model.</strong> In these 3&#8211;5 day focused sessions, developers worked cross-functionally with providers, and about 70% of the code produced was AI-assisted.</p></li><li><p><strong>Providers were treated as partners.</strong> Booking.com invited tool vendors to embed with teams, share expertise, and help tackle specific business challenges.</p></li></ul><p><strong>Driving cultural change</strong></p><ul><li><p><strong>AI adoption required cultural as well as technical change.</strong> Simply granting licenses didn&#8217;t work; Booking.com needed to integrate AI into its development culture.</p></li><li><p><strong>Leadership enablement was crucial.</strong> Leaders who experienced AI firsthand became stronger advocates for adoption across their business units.<br><strong>Coding sessions with leaders built momentum.</strong> These sessions energized managers, who returned to their teams pushing for wider use.<br><strong>Slack communities lowered friction.</strong> Developers felt more comfortable asking questions in Slack than in formal settings, which revealed education gaps and sped up problem-solving.</p></li><li><p><strong>Cross-unit collaboration improved.</strong> AI became a catalyst for teams across different business units to work together more closely.</p></li></ul><p><strong>Measuring impact while key questions remain</strong></p><ul><li><p><strong>Early &#8220;hours saved&#8221; metrics weren&#8217;t enough.</strong> Surveys about time savings were too subjective and based on too few users.</p></li><li><p><strong>Booking.com shifted to broader measures.</strong> They focused on speed, efficiency, quality, cycle time, and re-platforming progress.<br><strong>AI users created more&#8212;but lighter&#8212;merge requests.</strong> Daily users produced 30% more MRs that were 70% smaller, but it is not yet clear whether this improves quality or efficiency.</p></li><li><p><strong>Human review remains essential.</strong> Every MR still requires two reviewers, which protects against poor code slipping through.</p></li><li><p><strong>Key questions remain open.</strong> Is AI-assisted code more readable, efficient, and secure? Does it reduce bugs and vulnerabilities? These answers are still undetermined.</p></li></ul><p><strong>Unblocking organizational bottlenecks</strong></p><ul><li><p><strong>Procurement and risk reviews slowed down experimentation.</strong> Developers wanted to try every new AI tool, but legal and security created friction.</p></li><li><p><strong>A central &#8220;committee&#8221; was formed to unblock adoption.</strong> This group reviewed requests weekly, streamlined approvals, and prevented each business unit from duplicating effort.<br><strong>Fast-tracking proofs of concept became a priority.</strong> Developers could get hands-on with new tools sooner, while procurement caught up in the background.</p></li><li><p><strong>The platform mindset helped.</strong> Booking.com treated AI like any other developer tool, centralizing pain points and ensuring consistency across the company.</p></li></ul><p><strong>Tracking current adoption trends</strong></p><ul><li><p><strong>60% of developers now use AI tools.</strong> This places Booking.com in the top 25% of companies globally for adoption.</p></li><li><p><strong>Daily users ship more frequently.</strong> Developers using AI at least three times a week create 30% more merge requests.</p></li><li><p><strong>Usage patterns are evolving.</strong> While more MRs are being created, many are lighter, and Booking.com is still investigating what that means for long-term quality.</p></li><li><p><strong>Adoption is uneven across units.</strong> Some business units are ahead, while others are still cautious, reflecting differences in culture, legacy systems, and leadership buy-in.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=69s">01:09</a>) Bruno&#8217;s role at Booking.com and an overview of the business</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=139s">02:19</a>) Booking.com&#8217;s goals when introducing AI tooling</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=206s">03:26</a>) Why Booking.com made such an ambitious innovation ratio goal</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=406s">06:46</a>) The beginning of Booking.com&#8217;s journey with AI</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=534s">08:54</a>) Why the initial adoption of Cody was low</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=797s">13:17</a>) How education and enablement fueled adoption</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=948s">15:48</a>) The importance of a top-down cultural change for AI adoption</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=1058s">17:38</a>) The ongoing journey of determining the right metrics</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=1304s">21:44</a>) Measuring the longer-term impact of AI</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=1624s">27:04</a>) How Booking.com solved internal bottlenecks to testing new tools</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=1930s">32:10</a>) Booking.com&#8217;s framework for evaluating new tools</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=2150s">35:50</a>) The state of adoption at Booking.com and efforts to expand AI use</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=2227s">37:07</a>) What&#8217;s still undetermined about AI&#8217;s impact on PR/MR quality</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=2388s">39:48</a>) How Booking.com is addressing lagging adoption and monitoring churn</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=2604s">43:24</a>) How Booking.com&#8217;s Slack community lowers friction for questions and support</p><p>(<a href="https://www.youtube.com/watch?v=HuDtj_xHhis&amp;t=2675s">44:35</a>) Closing thoughts on what&#8217;s next for Booking.com&#8217;s AI plan</p><p><strong>Where to find Bruno Passos:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/brpassos/">https://www.linkedin.com/in/brpassos/</a></p><p>&#8226; X: <a href="https://x.com/brunopassos">https://x.com/brunopassos</a></p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; X: <a href="https://x.com/rhein_wein">https://x.com/rhein_wein</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p>&#8226; Laura&#8217;s course (Measuring Engineering Performance and AI Impact) <a href="https://lauratacho.com/developer-productivity-metrics-course">https://lauratacho.com/developer-productivity-metrics-course</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/research/measuring-ai-code-assistants-and-agents/?utm_source=podcast">Measuring AI code assistants and agents</a></p></li><li><p><a href="https://getdx.com/core-4-reporting/">DX Core 4 Framework</a></p></li><li><p><a href="https://www.booking.com/">Booking.com</a></p></li><li><p><a href="https://sourcegraph.com/search">Sourcegraph Search</a></p></li><li><p><a href="https://sourcegraph.com/cody">Cody | AI coding assistant from Sourcegraph</a></p></li><li><p><a href="https://www.linkedin.com/in/greysonjunggren/">Greyson Junggren - DX | LinkedIn</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Measuring AI code assistants and agents with the AI Measurement Framework ]]></title><description><![CDATA[A data-driven guide to measuring developer productivity in the AI era, cutting through hype to track real impact and ensure sustainable gains.]]></description><link>https://newsletter.getdx.com/p/measuring-ai-code-assistants-and</link><guid isPermaLink="false">https://newsletter.getdx.com/p/measuring-ai-code-assistants-and</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 15 Aug 2025 15:23:17 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/170802939/377d1d4d0ccd93ec38024880a1ecf37f.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/6A-dp1Kzyjk">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this episode of <em>Engineering Enablement</em>, DX CEO<strong> </strong>Abi Noda and I share practical guidance for measuring developer productivity in the AI era using our AI Measurement Framework. Based on research with industry leaders, vendors, and hundreds of organizations, we walk through how to cut through the hype and make informed decisions about AI adoption.</p><p>We talk about which fundamentals of productivity measurement remain unchanged, why metrics like acceptance rate can be misleading, and how to track AI&#8217;s real impact across utilization, quality, and cost. We also cover how to measure agentic workflows, expand the definition of &#8220;developer&#8221; to include AI-enabled contributors, and identify second-order effects before they create long-term problems.</p><p>If you&#8217;re introducing AI coding tools, exploring autonomous agents, or just trying to separate signal from noise, this episode offers a clear, actionable roadmap for using data to ensure AI delivers sustainable, meaningful gains.</p><div id="youtube2-6A-dp1Kzyjk" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;6A-dp1Kzyjk&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/6A-dp1Kzyjk?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>AI&#8217;s hype vs. reality gap</strong></p><ul><li><p><strong>Bold headlines are often misleading.</strong> Claims like &#8220;90% of code will be written by AI&#8221; typically come from cherry-picked studies in narrow scenarios, not representative of the median developer experience.</p></li><li><p><strong>Organizations need their own data.</strong> Vendor marketing and public research can set unrealistic expectations&#8212;measuring AI&#8217;s real-world impact in your own environment is the only way to guide strategy and investment.</p></li></ul><p><strong>AI doesn&#8217;t change engineering fundamentals</strong></p><ul><li><p><strong>Core principles remain the same.</strong> Scalability, maintainability, reliability, and meeting customer needs still define good engineering.</p></li><li><p><strong>AI builds on&#8212;not replaces&#8212;these foundations.</strong> Use AI to lift existing strengths, not as an excuse to rebuild productivity measurement from scratch.</p></li></ul><p><strong>The AI Measurement Framework</strong></p><ul><li><p><strong>Three dimensions matter most:</strong> utilization (how widely AI is used), impact (how it changes performance), and cost (what you spend on tools, licenses, training).</p></li><li><p><strong>Track them together for the full picture.</strong> Over-indexing on one&#8212;like utilization&#8212;can lead to false conclusions about overall value.</p></li></ul><p><strong>The pitfalls of acceptance rate</strong></p><ul><li><p><strong>Acceptance rate is unreliable.</strong> AI code that&#8217;s accepted in the IDE is often rewritten, heavily modified, or deleted before shipping.<br><strong>Better options exist.</strong> Tagging PRs for AI contributions or using file-level observability can identify AI-authored changes across all IDEs and tools, avoiding blind spots.</p></li></ul><p><strong>Collecting measurement data</strong></p><ul><li><p><strong>Tool telemetry</strong> (from GitHub, GitLab, or AI vendors) shows patterns in daily and weekly adoption.</p></li><li><p><strong>Quarterly surveys</strong> reveal long-term trends in developer satisfaction, productivity, and maintainability perceptions.</p></li><li><p><strong>In-workflow experience sampling</strong> asks targeted questions at the moment of work&#8212;e.g., &#8220;Was this PR authored with AI?&#8221;&#8212;to get precise, low-bias feedback.</p></li></ul><p><strong>Perception vs. reality in time savings</strong></p><ul><li><p><strong>Developers often feel faster with AI&#8212;but logs may say otherwise.</strong> A meta-study found that self-reports overstated gains; in some cases, AI users were slower.</p></li><li><p><strong>Triangulate survey and system data</strong> to confirm that perceived improvements match actual throughput and quality metrics.</p></li></ul><p><strong>Measuring agentic workflows</strong></p><ul><li><p><strong>Treat agents as team extensions, not digital employees.</strong> Measure productivity for the combined human-agent team, just as you would for a team using CI/CD tools like Jenkins.</p></li><li><p><strong>Focus on maturity, not just usage.</strong> There&#8217;s a big difference between using AI for autocomplete and delegating multi-step tasks to autonomous loops.</p></li></ul><p><strong>Expanding the definition of developer</strong></p><ul><li><p><strong>AI enables more contributors.</strong> Designers, PMs, and other non-engineers can now produce functional code and prototypes.<br><strong>Apply the same quality gates</strong>&#8212;code review, testing, maintainability checks&#8212;to their contributions as to full-time engineers.</p></li></ul><p><strong>Thinking beyond AI</strong></p><ul><li><p><strong>AI is one tool in the toolbox.</strong> Many bottlenecks&#8212;like unclear requirements, inefficient processes, and infrastructure delays&#8212;can&#8217;t be solved by AI alone.</p></li><li><p><strong>Balance investment</strong> to ensure you&#8217;re addressing all productivity levers, not just AI adoption.</p></li></ul><p><strong>Watching for second-order effects</strong></p><ul><li><p><strong>More AI-generated code can create new bottlenecks.</strong> Extra output can slow PR reviews, increase cognitive load, and lower maintainability.</p></li><li><p><strong>Impact metrics reveal trade-offs</strong> early, helping you prevent short-term speed gains from causing long-term technical debt.</p></li></ul><p><strong>Rolling out metrics successfully</strong></p><ul><li><p><strong>Aggregate at team or department level.</strong> Avoid individual tracking to build trust and reduce fear around AI adoption.</p></li><li><p><strong>Be transparent about data use</strong> so developers know it&#8217;s for enablement, tool evaluation, and rollout strategy&#8212;not performance surveillance.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=86s">01:26</a>) The challenge of measuring developer productivity in the AI age</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=257s">04:17</a>) Measuring productivity in the AI era &#8212; what stays the same and what changes</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=445s">07:25</a>) How to use DX&#8217;s AI Measurement Framework</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=790s">13:10</a>) Measuring AI&#8217;s true impact from adoption rates to long-term quality and maintainability</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=991s">16:31</a>) Why acceptance rate is flawed &#8212; and DX&#8217;s approach to tracking AI-authored code</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=1105s">18:25</a>) Three ways to gather measurement data</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=1315s">21:55</a>) How Google measures time savings and why self-reported data is misleading</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=1465s">24:25</a>) How to measure agentic workflows and a case for expanding the definition of developer</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=1730s">28:50</a>) A case for not overemphasizing AI&#8217;s role</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=1831s">30:31</a>) Measuring second-order effects</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=1946s">32:26</a>) Audience Q&amp;A: applying metrics in practice</p><p>(<a href="https://www.youtube.com/watch?v=6A-dp1Kzyjk&amp;t=2205s">36:45</a>) Wrap up: best practices for rollout and communication</p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><p>&#8226; Substack: &#8203;&#8203;<a href="https://substack.com/@abinoda">https://substack.com/@abinoda</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p></li><li><p><a href="https://getdx.com/whitepaper/ai-measurement-framework/">Measuring AI code assistants and agents</a></p></li><li><p><a href="https://www.businessinsider.com/ai-google-engineers-coding-productive-sundar-pichai-alphabet-2025-6">AI is making Google engineers 10% more productive, says Sundar Pichai - Business Insider</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[How to cut through the hype and measure AI’s real impact (Live from LeadDev London)]]></title><description><![CDATA[Live from LeadDev London, this episode explores how leaders can cut through AI hype using data from 39,000 developers and two key frameworks to improve developer experience and drive impact.]]></description><link>https://newsletter.getdx.com/p/how-to-cut-through-the-hype-and-measure</link><guid isPermaLink="false">https://newsletter.getdx.com/p/how-to-cut-through-the-hype-and-measure</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 08 Aug 2025 15:07:17 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/170023686/bf9da44ecae90d6a3c2e21f9d929ab26.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/qZv0YOoRLmg">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this special episode of the Engineering Enablement podcast, recorded live at LeadDev London, I unpack the gap between AI hype and engineering reality&#8212;and how leaders can use data to close it.</p><p>I share the latest insights from nearly 39,000 developers across 184 companies, walk through the Core 4 and AI Measurement Frameworks, and explain how to use them together to measure what matters, improve developer experience, and drive real organizational impact&#8212;without getting lost in the noise.</p><div id="youtube2-qZv0YOoRLmg" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;qZv0YOoRLmg&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/qZv0YOoRLmg?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>The AI hype cycle vs. ground truth</strong></p><ul><li><p><strong>The &#8220;disappointment gap&#8221;</strong> refers to the widening space between sensational AI headlines and the lived reality of teams on the ground. Organizations are being pushed to move faster with AI, yet few have defined what success even looks like.</p></li><li><p><strong>Headlines touting &#8220;90% of code written by AI&#8221; inflate expectations and erode trust</strong>. Developers feel let down when tools don&#8217;t deliver on the hype. Executives, in turn, expect exponential productivity gains without understanding what&#8217;s realistically achievable.</p></li><li><p><strong>The best way to close this gap is with data</strong>. Leaders need to ground their AI strategies in facts, not forecasts.</p></li></ul><p><strong>AI&#8217;s current role in high-performing engineering orgs</strong></p><ul><li><p><strong>In the top quartile of organizations, around 60% of developers are now using AI tools daily or weekly.</strong> However, this usage does not translate directly into AI generating most of the code.</p></li><li><p>These organizations are seeing the best results because they invest in enablement, support, and identifying practical use cases that actually work.</p></li><li><p><strong>Across nearly 39,000 developers at 184 companies, the average reported time savings from AI use is 3 hours and 45 minutes per week.</strong> It&#8217;s a meaningful uplift, but not a silver bullet.</p></li></ul><p><strong>Engineering leaders must shape the narrative</strong></p><ul><li><p><strong>Engineering leaders are also business leaders</strong>&#8212;and they need to take on the responsibility of educating peers and execs on what AI adoption actually looks like.</p></li><li><p>Effective leaders can clearly answer:</p><ol><li><p>How is our organization performing today?</p></li><li><p>How is AI helping&#8212;or not helping?</p></li><li><p>What are we doing next to improve?</p></li></ol></li></ul><p><strong>Back to basics: what defines engineering excellence?</strong></p><ul><li><p><strong>A shared definition of engineering performance is essential</strong> before measuring the effects of AI. The <strong>DX Core 4</strong> framework offers this foundation.</p></li><li><p><strong>Core 4 combines elements of DORA, SPACE, and DevEx into a single, balanced model </strong>with four key dimensions: <strong>speed</strong>, <strong>effectiveness</strong>, <strong>quality</strong>, and <strong>impact</strong>.<br><strong>These metrics must be evaluated together. </strong>Optimizing one at the expense of another (e.g., speed at the cost of quality) risks destabilizing the system.</p></li></ul><p><strong>Developer experience drives performance outcomes</strong></p><ul><li><p><strong>Developer experience is the strongest performance lever available to engineering organizations. </strong>The <strong>DXI</strong> (Developer Experience Index) measures 14 evidence-based drivers of experience and correlates directly with time savings.</p></li><li><p><strong>For each DXI point gained, developers save 13 minutes per week.</strong> While that may seem small, the impact scales dramatically across teams.</p></li><li><p><strong>Block used DXI to identify 500,000 hours lost annually</strong> due to friction&#8212;data that directly shaped their investment decisions and enabled faster delivery without compromising quality.</p></li></ul><p><strong>A complementary framework for measuring AI</strong></p><ul><li><p>The <strong>AI Measurement Framework</strong> adds clarity by tracking the effect of AI across three pillars: <strong>utilization</strong>, <strong>impact</strong>, and <strong>cost</strong>.</p></li><li><p>Utilization captures how broadly and consistently AI tools are being used. The biggest gains typically come when teams move from occasional to consistent usage.</p></li><li><p><strong>Time savings per week is the most aligned metric</strong> across the industry for measuring impact.</p></li><li><p><strong>Cost includes not just licenses but also investment in training and enablement</strong>&#8212;areas that are often overlooked but essential for success.</p></li></ul><p><strong>Using both frameworks together creates clarity and confidence</strong></p><ul><li><p><strong>Core 4 </strong>answers: &#8220;What does high performance look like?&#8221;</p></li><li><p><strong>The AI Measurement Framework </strong>answers: &#8220;How is AI affecting that performance?&#8221;</p></li><li><p><strong>Together, these frameworks enable leaders to move beyond guesswork</strong> and act with clarity, especially during times of rapid change.</p></li></ul><p><strong>AI is a multiplier&#8212;but only with the right foundations</strong></p><ul><li><p><strong>Accelerating software delivery with AI is possible, but it requires strong fundamentals in place</strong>. Cutting corners on quality, stability, or developer experience for short-term gains can create long-term damage.</p></li><li><p>When grounded in solid frameworks and real data, AI can improve velocity, collaboration, and developer satisfaction without compromising core engineering values.</p></li></ul><p><strong>Better software faster is possible</strong>&#8212;not by chasing hype, but by aligning teams on what matters and measuring what works.</p><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg">00:00</a>) Intro: Laura&#8217;s keynote from LDX3</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=104s">01:44</a>) The problem with asking how much faster can we go with AI?</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=182s">03:02</a>) How the disappointment gap creates barriers to AI adoption</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=380s">06:20</a>) What AI adoption looks like at top-performing organizations</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=473s">07:53</a>) What leaders must do to turn AI into meaningful impact</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=650s">10:50</a>) Why building better software with AI still depends on fundamentals</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=723s">12:03</a>) An overview of the DX Core 4 Framework</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=802s">13:22</a>) Why developer experience is the biggest performance lever</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=912s">15:12</a>) How Block used Core 4 and DXI to identify 500,000 hours in time savings</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=968s">16:08</a>) How to get started with Core 4</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=1052s">17:32</a>) Measuring AI with the AI Measurement Framework</p><p>(<a href="https://www.youtube.com/watch?v=qZv0YOoRLmg&amp;t=1305s">21:45</a>) Final takeaways and how to get started with confidence</p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; X: <a href="https://x.com/rhein_wein">https://x.com/rhein_wein</a></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://leaddev.com/leaddev-london/">LDX3 by LeadDev | The Festival of Software Engineering Leadership | London</a></p></li><li><p><a href="https://www.youtube.com/watch?v=EO3_qN_Ynsk">Software engineering with LLMs in 2025: reality check</a></p></li><li><p><a href="https://getdx.com/podcast/developer-productivity-at-microsoft/">SPACE framework, PRs per engineer, AI research</a></p></li><li><p><a href="https://getdx.com/podcast/brian-houck-ai-adoption-playbook/">The AI adoption playbook: Lessons from Microsoft's internal strategy</a></p></li><li><p><a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p></li><li><p><a href="https://nicolefv.com/">Nicole Forsgren</a></p></li><li><p><a href="https://www.margaretstorey.com/">Margaret-Anne Storey</a></p></li><li><p><a href="https://www.dropbox.com/">Dropbox.com</a></p></li><li><p><a href="https://www.etsy.com/">Etsy</a></p></li><li><p><a href="https://www.pfizer.com/">Pfizer</a></p></li><li><p><a href="https://www.linkedin.com/in/drewhouston/">Drew Houston - Dropbox | LinkedIn</a></p></li><li><p><a href="https://block.xyz/">Block</a></p></li><li><p><a href="https://cursor.com/">Cursor</a></p></li><li><p><a href="https://dora.dev/">Dora.dev</a></p></li><li><p><a href="https://sourcegraph.com/">Sourcegraph</a></p></li><li><p><a href="https://www.booking.com/">Booking.com</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Unpacking METR’s findings: Does AI slow developers down?]]></title><description><![CDATA[Quentin Anthony unpacks why AI coding tools slow developers despite boosting perceived productivity, and how teams can improve results with task&#8209;level fit and better digital hygiene.]]></description><link>https://newsletter.getdx.com/p/unpacking-metrs-findings-does-ai</link><guid isPermaLink="false">https://newsletter.getdx.com/p/unpacking-metrs-findings-does-ai</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 01 Aug 2025 15:02:53 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/169255748/9d58e0579bffafcef2c04b101d8bcfdb.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/NRizG0tSGic">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this episode, I&#8217;m joined by <a href="https://www.linkedin.com/in/quentin-anthony/">Quentin Anthony</a>, Head of Model Training at <strong>Zyphra</strong> and a participant in METR&#8217;s recent study on AI coding tools. We explore the study&#8217;s unexpected findings&#8212;why developers often felt more productive using AI, but in many cases weren&#8217;t&#8212;and unpack the nuances of where these tools actually add value. Quentin offers practical, experience-backed advice on avoiding common pitfalls, such as the sunk-cost fallacy and context rot, evaluating task-level fit, and building the kind of tool hygiene that&#8217;s critical for long-term success with AI.</p><div id="youtube2-NRizG0tSGic" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;NRizG0tSGic&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/NRizG0tSGic?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>The biggest takeaways from the METR study</strong></p><ul><li><p><strong>Quentin participated in a recent study by METR</strong> (Model Evaluation and Training Red Teamers), which found that on average, developers were slower when using AI&#8212;despite feeling more productive.</p></li><li><p>The gap between perceived and actual efficiency is real&#8212;and often overlooked, especially when AI feels fun to use.</p></li></ul><ul><li><p><strong>AI excels at documentation, unit tests, and refactoring</strong>&#8212;tasks that can often be completed in a single prompt.</p></li><li><p>For complex, low-level work&#8212;like GPU kernels or distributed systems&#8212;<strong>models tend to produce bloated code</strong> or require too much back-and-forth to be useful.</p></li></ul><p><strong>Time-boxing is key to avoiding sunk-cost fallacy and watching out for context rot</strong></p><ul><li><p><strong>Quentin recommends setting strict time limits when using AI</strong>: if it&#8217;s not helping in 10&#8211;15 minutes, move on.</p></li><li><p>Developers often spend too long trying to force a model to help with the wrong kind of task.</p></li></ul><ul><li><p><strong>Long chats and overloaded prompts can confuse models</strong>, causing hallucinations and inconsistent behavior.</p></li><li><p>Restarting chats frequently and summarizing past work helps keep models grounded and accurate.</p></li><li><p><strong>Quentin recommends using summarization prompts to distill the current chat</strong> before restarting&#8212;this keeps context clean while avoiding repetition.</p></li></ul><p><strong>AI tools introduce more idle time than you think</strong></p><ul><li><p><strong>Waiting on model responses</strong>&#8212;even just 15&#8211;30 seconds&#8212;adds up fast, especially with reasoning-heavy prompts.</p></li><li><p><strong>Quentin uses this downtime for microtasks and blocks distractions</strong>, such as social media, to stay in the flow.</p></li></ul><p><strong>Prompting skill helps&#8212;but isn&#8217;t everything</strong></p><ul><li><p><strong>Success with AI isn&#8217;t just about writing better prompts</strong>. Often, task-model mismatch is the real problem.</p></li><li><p>Blaming developers for tool failures ignores deeper limitations in model training and context handling.</p></li><li><p><strong>Quentin treats AI output like junior engineer code&#8212;carefully reviewed, never blindly trusted</strong>. Even when correct, AI-generated code is often bloated or hard to maintain.</p></li></ul><p><strong>Focus on task-level fit, not team-level rollout</strong></p><ul><li><p><strong>Organizations should evaluate AI usefulness at the task leve</strong>l, not by team, codebase, or tooling preference.</p></li><li><p><strong>Not all work benefits from AI</strong>&#8212;even within the same repo or project.</p></li><li><p><strong>In Quentin&#8217;s view, tasks like acceptance tests, PR review, or boilerplate generation</strong> are ideal candidates for model support&#8212;while planning and design should stay human-led.</p></li></ul><p><strong>Model behavior varies&#8212;test before trusting</strong></p><ul><li><p><strong>Different models excel at different things</strong>. Claude may outperform Gemini at writing comments, while Gemini might be better at summarizing code.</p></li><li><p><strong>Quentin tries new models on familiar tasks first</strong> and expands use only after carefully watching for failure patterns.</p></li><li><p><strong>Claude is strong at writing clean, human-readable cod</strong>e.</p></li><li><p><strong>Gemini 2.5 is particularly good at summarizing</strong>. Quentin picks models based on task-specific strengths rather than defaulting to one tool.</p></li></ul><p><strong>Tool sprawl creates friction and other limits</strong></p><ul><li><p>Constantly switching between AI tools leads to confusion and inconsistent results.</p></li><li><p><strong>Quentin keeps his toolset small and stable</strong>, adjusting slowly to new platforms to avoid cognitive overload.</p></li><li><p>When adopting new tools, Quentin starts with familiar, low-risk tasks like unit tests, then gradually expands usage as he learns how the model behaves.</p></li></ul><ul><li><p><strong>Multi-agent systems are exciting, but they are still unreliable</strong>. They perform well in narrow settings, but struggle in real-world workflows.</p></li><li><p>For now, you&#8217;ll get more value from well-scoped tools and clearly defined use cases</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic">00:00</a>) Intro</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=92s">01:32</a>) A brief overview of Quentin&#8217;s background and current work</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=125s">02:05</a>) An explanation of METR and the study Quentin participated in</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=662s">11:02</a>) Surprising results of the METR study</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=767s">12:47</a>) Quentin&#8217;s takeaways from the study&#8217;s results</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=990s">16:30</a>) How developers can avoid bloated code bases through self-reflection</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=1171s">19:31</a>) Signs that you&#8217;re not making progress with a model</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=1285s">21:25</a>) What is &#8220;context rot&#8221;?</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=1384s">23:04</a>) Advice for combating context rot</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=1534s">25:34</a>) How to make the most of your idle time as a developer</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=1693s">28:13</a>) Developer hygiene: the case for selectively using AI tools</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=2008s">33:28</a>) How to interact effectively with new models</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=2128s">35:28</a>) Why organizations should focus on tasks that AI handles well</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=2281s">38:01</a>) Where AI fits in the software development lifecycle</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=2380s">39:40</a>) How to approach testing with models</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=2431s">40:31</a>) What makes models different</p><p>(<a href="https://www.youtube.com/watch?v=NRizG0tSGic&amp;t=2525s">42:05</a>) Quentin&#8217;s thoughts on agents</p><p><strong>Where to find Quentin Anthony:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/quentin-anthony/">https://www.linkedin.com/in/quentin-anthony/</a></p><p>&#8226; X: <a href="https://x.com/QuentinAnthon15">https://x.com/QuentinAnthon15</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p></li><li><p><a href="https://www.zyphra.com/">Zyphra</a></p></li><li><p><a href="https://www.eleuther.ai/">EleutherAI</a></p></li><li><p><a href="https://metr.org/">METR</a></p></li><li><p><a href="https://cursor.com/">Cursor</a></p></li><li><p><a href="https://claude.ai/">Claude</a></p></li><li><p><a href="https://www.librechat.ai/">LibreChat</a></p></li><li><p><a href="https://gemini.google.com/">Google Gemini</a></p></li><li><p><a href="https://openai.com/index/introducing-o3-and-o4-mini/">Introducing OpenAI o3 and o4-mini</a></p></li><li><p><a href="https://newsletter.getdx.com/p/metr-study-on-how-ai-affects-developer-productivity">METR&#8217;s study on how AI affects developer productivity</a></p></li><li><p><a href="https://x.com/QuentinAnthon15/status/1943948791775998069">Quentin Anthony on X: "I was one of the 16 devs in this study."</a></p></li><li><p><a href="https://news.ycombinator.com/item?id=44310054">Context rot from Hacker News</a></p></li><li><p><a href="https://www.anthropic.com/research/tracing-thoughts-language-model">Tracing the thoughts of a large language model</a></p></li><li><p><a href="https://www.kimi.com/">Kimi</a></p></li><li><p><a href="https://x.ai/news/grok-4">Grok 4 | xAI</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[CarGurus’ journey building a developer portal and increasing AI adoption]]></title><description><![CDATA[Frank Fodera shares how CarGurus broke up a monolith, rolled out AI tools, and built their own internal developer portal&#8212;tying it all back to real business impact.]]></description><link>https://newsletter.getdx.com/p/cargurus-journey-building-a-developer</link><guid isPermaLink="false">https://newsletter.getdx.com/p/cargurus-journey-building-a-developer</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 11 Jul 2025 15:20:56 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/167926629/59c48cdc41c6a058003084476a8c3753.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/Xvn6HXBVuaM">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In today&#8217;s episode, I caught up with Frank Fodera from <strong>CarGurus</strong> to hear how his team tackled two massive challenges: breaking up a monolith and rolling out AI tools across the engineering org. Frank shared how they built their internal developer portal, Showroom, from scratch&#8212;why they didn&#8217;t buy a tool like Backstage, how it supports their dev teams, and how they tie it all back to business impact. We also discussed their AI experiments, the metrics they&#8217;re tracking, and how they&#8217;re driving adoption without imposing one-size-fits-all solutions.</p><div id="youtube2-Xvn6HXBVuaM" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;Xvn6HXBVuaM&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/Xvn6HXBVuaM?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>Internal developer portals: from spreadsheet to strategic platform</strong></p><ul><li><p><strong>Started with a spreadsheet.</strong> CarGurus built their internal developer portal, <em>Showroom</em>, in 2019. The initial goal was simple: solve ownership confusion during a monolith-to-microservices transition.</p></li><li><p><strong>Showroom now powers five key pillars:</strong></p><ul><li><p><strong>Discoverability:</strong> Centralized source of truth for services, jobs, and ownership</p></li><li><p><strong>Governance:</strong> Dynamic compliance checks to ensure &#8220;Golden Path&#8221; adoption</p></li><li><p><strong>Self-serviceability:</strong> Spin up new services, subscribe to topics, and more</p></li><li><p><strong>Transparency:</strong> Access logs, service data, and operational info in one place</p></li><li><p><strong>Operational efficiency:</strong> Reduced cognitive load and friction across the SDLC</p></li></ul></li><li><p><strong>Interfaces with infra but abstracts complexity.</strong> Showroom sits on top of observability, alerting, and infra tools, providing a consistent experience even as backend systems evolve.</p></li><li><p><strong>Evolved by solving real problems, not chasing trends.</strong> Rather than set out to &#8220;build an IDP,&#8221; the team invested in Showroom whenever it accelerated strategic initiatives&#8212;and saw massive returns.</p></li><li><p><strong>Impact: 75-day setup cycles cut to &lt;3 days.</strong> By leaning on automated workflows, best-practice templates, and centralized access to information, CarGurus dramatically increased service creation velocity.</p></li></ul><p><strong>Build vs. buy: why CarGurus chose to go homegrown</strong></p><ul><li><p><strong>Started before Backstage existed.</strong> No off-the-shelf solution could meet their requirements when the initiative began.</p></li><li><p><strong>Kept evaluating even after Backstage launched.</strong> But concluded the effort to customize it would be equal to enhancing their own tool.</p></li><li><p><strong>Customization was critical.</strong> Their IDP integrates deeply with internal systems and niche tools not typically supported by commercial solutions.</p></li><li><p><strong>Consistent UX across tools.</strong> Showroom provides a single pane of glass&#8212;developers don&#8217;t need to know what&#8217;s happening under the hood as tools change.</p></li></ul><p><strong>AI coding assistants: a platform-level initiative</strong></p><ul><li><p><strong>Bake-off between three tools.</strong> CarGurus ran a head-to-head comparison across multiple AI assistants. Feedback showed different tools excelled in different languages and workflows.</p></li><li><p><strong>Standardization isn&#8217;t the goal.</strong> The company is embracing flexibility&#8212;letting teams and engineers choose the best tool for their domain.</p></li><li><p><strong>Qualitative feedback &gt; raw telemetry.</strong> Developer sentiment and perceived time savings are more actionable than latency or code volume alone.</p></li><li><p><strong>Measuring impact across six dimensions:</strong></p><ul><li><p><em>Speed</em>: Diffs per engineer</p></li><li><p><em>Efficiency</em>: Developer Experience Index (DXI)</p></li><li><p><em>Satisfaction</em>: CSAT scores</p></li><li><p><em>Adoption</em>: Tool usage tracking</p></li><li><p><em>Time savings</em>: Self-reported</p></li><li><p><em>Burnout prevention</em>: Monitoring job satisfaction</p></li></ul></li><li><p><strong>Maintainability dropped slightly.</strong> AI-generated code may feel less &#8220;owned,&#8221; even if it&#8217;s correct. A known and acceptable trade-off so far.</p></li></ul><p><strong>Driving adoption: what&#8217;s working</strong></p><ul><li><p><strong>Leadership buy-in was immediate.</strong> Execs were eager to invest in efficiency. AI adoption was viewed as a strategic imperative.</p></li><li><p><strong>Champions are key.</strong> Internal AI &#8220;power users&#8221; share tips via tech talks and videos to scale best practices organically.</p></li><li><p><strong>Internal marketing makes a difference.</strong> Just telling developers they had access wasn&#8217;t enough. Framing AI tooling as a must-try (not optional) led to a meaningful adoption spike.</p></li><li><p><strong>Focus is on education and experimentation.</strong> Goal is to help developers integrate AI into their daily flow&#8212;not just try it once and move on.</p></li><li><p><strong>Expected ROI? ~15&#8211;30% gains</strong> in efficiency in early stages, with room to grow as usage matures and more advanced agent-based tools are explored.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM">00:00</a>) Intro: IDPs (Internal Developer Portals) and AI</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=127s">02:07</a>) The IDP journey at CarGurus</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=353s">05:53</a>) A breakdown of the people responsible for building the IDP</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=425s">07:05</a>) The five pillars of the Showroom IDP</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=552s">09:12</a>) How DevX worked with infrastructure</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=673s">11:13</a>) The business impact of Showroom</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=837s">13:57</a>) The transition from monolith to microservices and struggles along the way</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=954s">15:54</a>) The benefits of building a custom IDP</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=1150s">19:10</a>) How CarGurus drives AI coding tool adoption</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=1728s">28:48</a>) Getting started with an AI initiative</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=1910s">31:50</a>) Metrics to track</p><p>(<a href="https://www.youtube.com/watch?v=Xvn6HXBVuaM&amp;t=2046s">34:06</a>) Tips for driving AI adoption</p><p><strong>Where to find Frank Fodera :</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/frankfodera/">https://www.linkedin.com/in/frankfodera/</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p></li><li><p><a href="https://getdx.com/webinar/internal-developer-portals-overview/">Internal Developer Portals: Use Cases and Key Components</a></p></li><li><p><a href="https://learn.microsoft.com/en-us/azure/architecture/patterns/strangler-fig">Strangler Fig Pattern - Azure Architecture Center | Microsoft Learn</a></p></li><li><p><a href="https://backstage.spotify.com/">Spotify for Backstage</a></p></li><li><p><a href="https://getdx.com/podcast/brian-houck-ai-adoption-playbook/">The AI adoption playbook: Lessons from Microsoft's internal strategy</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Snowflake’s playbook for operational excellence]]></title><description><![CDATA[Snowflake leaders Gilad Turbahn and Amy Yuan reveal how they drive operational excellence by treating teams like customers and use roadmaps plus advisory boards to align priorities and scale planning.]]></description><link>https://newsletter.getdx.com/p/snowflakes-playbook-for-operational</link><guid isPermaLink="false">https://newsletter.getdx.com/p/snowflakes-playbook-for-operational</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 20 Jun 2025 13:10:01 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/166180567/232cfee7f79097a9620d2c8b819e0664.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/e8nJFsgZsLU">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this episode, I talk with Gilad Turbahn, Head of Developer Productivity, and Amy Yuan, Director of Engineering at <strong>Snowflake</strong>, about how they approach operational excellence. We dig into how they build trust with engineering teams, the communication rhythms that keep their org aligned, and how they treat internal teams like customers. Gilad and Amy also share how Snowflake uses roadmaps, advisory boards, and direct feedback to drive priorities&#8212;and how they&#8217;re evolving their planning practices to scale with the company.</p><div id="youtube2-e8nJFsgZsLU" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;e8nJFsgZsLU&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/e8nJFsgZsLU?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><h2><strong>Some takeaways: </strong></h2><p><strong>What operational excellence looks like at Snowflake</strong></p><ul><li><p>Do what you say you&#8217;ll do&#8212;on time&#8212;to build trust.</p></li><li><p>Treat internal teams like customers.</p></li><li><p>Use documentation and help channels to support engineering teams at scale.</p></li></ul><p><strong>Creating a strong feedback loop</strong></p><ul><li><p>Feedback is only valuable if it&#8217;s followed up on&#8212;Snowflake closes the loop with a tracking system (tracking each person spoken to) and a weekly ops review.</p></li><li><p>Show how the feedback is reflected in the road map, or explain why it isn&#8217;t.</p></li></ul><p><strong>Snowflake&#8217;s five-part communication rhythm</strong></p><ul><li><p>Weekly newsletters</p></li><li><p>Surveys to gather and prioritize needs</p></li><li><p>Interviews across levels</p></li><li><p>Roadshows and all-hands</p></li><li><p>A mix of top-down and bottom-up adoption strategies</p></li></ul><p><strong>Barriers to operational excellence</strong></p><ul><li><p>Lagging adopters and resistance to change</p></li><li><p>Wrong leadership style for the team or individual&#8212;leadership should be tailored to the individual and team needs.</p></li></ul><p><strong>Customer engagement, redefined</strong></p><ul><li><p>Treat engineers like customers&#8212;tailor time and communication to their needs.</p></li><li><p>Attend team all-hands, host listening sessions, and use data to show trade-offs.</p></li><li><p>Snowflake adjusts session length from quick 5-minute syncs to deep 60-minute sessions depending on the needs.</p></li></ul><p><strong>Converting detractors into allies</strong></p><ul><li><p>Snowflake PMs are highly technical and dive deep into engineers&#8217; pain points.</p></li><li><p>They walk through problems collaboratively and validate concerns.</p></li><li><p>Customer Advisory Boards help create evangelists from within.</p></li></ul><p><strong>Customer Advisory Boards at Snowflake</strong></p><ul><li><p>Boards are built from diverse teams and levels&#8212;both promoters and detractors.</p></li><li><p>They look for people who are vocal.</p></li><li><p>Members are identified either in interviews or nominated by directors.</p></li><li><p>Meetings are informal with no pressure to make every single meeting.</p></li><li><p>They use Slack to keep the feedback loop active.</p></li><li><p>Moderate effectively: At Snowflake the process is to introduce a road map, design, quarterly plan, or survey results. Then ask members what is top of mind.</p></li></ul><p><strong>Planning and goal-setting at Snowflake</strong></p><ul><li><p>Planning at Snowflake is done both annually and quarterly.</p></li><li><p>They balance OKRs with a long-term North Star.</p></li><li><p>They are currently experimenting with &#8220;thematic prioritization.&#8221; Thematic prioritization uses main themes that can make a difference to customers.</p></li></ul><h2><strong>In this episode, we cover:</strong></h2><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU">00:00</a>) Intro: an overview of operational excellence</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=253s">04:13</a>) Obstacles to executing with operational excellence</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=351s">05:51</a>) An overview of the Snowflake playbook for operational excellence</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=505s">08:25</a>) Who does the work of reaching out to customers</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=546s">09:06</a>) The importance of customer engagement</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=619s">10:19</a>) How Snowflake does customer engagement</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=853s">14:13</a>) The types of feedback received and the two camps (supporters and detractors)</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=1015s">16:55</a>) How to influence detractors and how detractors actually help</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=1107s">18:27</a>) Using insiders as messengers</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=1368s">22:48</a>) An overview of Snowflake&#8217;s customer advisory board</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=1570s">26:10</a>) The importance of meeting in person (learnings from Warsaw and Berlin office visits)</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=1688s">28:08</a>) Managing up</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=1807s">30:07</a>) How planning is done at Snowflake</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=2185s">36:25</a>) Setting targets for OKRs, and Snowflake&#8217;s philosophy on metrics</p><p>(<a href="https://www.youtube.com/watch?v=e8nJFsgZsLU&amp;t=2362s">39:22</a>) The annual plan and how it&#8217;s shared</p><p><strong>Where to find Amy Yuan:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/amy-yuan-a8ba783/">https://www.linkedin.com/in/amy-yuan-a8ba783/</a></p><p><strong>Where to find Gilad Turbahn:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/giladturbahn/">https://www.linkedin.com/in/giladturbahn/</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><h2><strong>Referenced:</strong></h2><ul><li><p><a href="https://getdx.com/podcast/developer-productivity-at-snowflake/">CTO buy-in, measuring sentiment, and customer focus</a></p></li><li><p><a href="https://www.snowflake.com/">Snowflake</a></p></li><li><p><a href="https://www.linkedin.com/in/benoit-dageville-3011845/">Benoit Dageville - Snowflake Computing | LinkedIn</a></p></li><li><p><a href="https://www.linkedin.com/in/thierry-cruanes-3927363/">Thierry Cruanes - Snowflake Computing | LinkedIn</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[The biggest obstacles preventing GenAI adoption — and how to overcome them]]></title><description><![CDATA[DX CTO Laura Tacho shares how leaders can overcome fear and hype to drive AI adoption and measure real impact in engineering teams.]]></description><link>https://newsletter.getdx.com/p/the-biggest-obstacles-preventing</link><guid isPermaLink="false">https://newsletter.getdx.com/p/the-biggest-obstacles-preventing</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 06 Jun 2025 14:09:29 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/165050079/eb0cb6e3505038a003182238bf40e52a.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/lkDsuOkJ0Rw">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><p>In this episode, I&#8217;m joined by DX CTO Laura Tacho to talk about what&#8217;s really holding back AI adoption in engineering teams. It&#8217;s not the technical challenges&#8212;it&#8217;s fear, unclear expectations, and the disconnect between hype and reality. Laura shares practical strategies for making progress, from modeling usage at the leadership level to creating space for experimentation. We also talk about how to measure impact effectively&#8212;including why it&#8217;s critical to establish a baseline before introducing AI tools, so you can track real changes over time.</p><div id="youtube2-lkDsuOkJ0Rw" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;lkDsuOkJ0Rw&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/lkDsuOkJ0Rw?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p><strong>Some takeaways: </strong></p><p><strong>Obstacles to AI adoption</strong></p><ul><li><p>The hype: Some of the claims written on LinkedIn and other places online overstate the impact of AI coding tools.</p></li><li><p>The cost of AI tools.</p></li><li><p>Technical barriers aren&#8217;t holding back adoption&#8212;cultural and human resistance are.</p></li></ul><p><strong>Some AI adoption stats</strong></p><ul><li><p><strong>Top-end organizations report that 60-70% of developers are using code assistance either daily or weekly.</strong></p></li><li><p>Less than 60% of code is written with AI at Microsoft.</p></li><li><p>There is a steady increase in the adoption of AI tools.</p></li></ul><blockquote></blockquote><p><strong>Strategies for driving AI adoption</strong></p><ul><li><p><strong>Work against the hype by showing what the real impact of AI tools is</strong>. Methods for demonstrating impact include: self-reported, telemetry-based, direct signals from developers, indirect signals using hard data, and quality measures.</p></li><li><p>There isn&#8217;t a special set of metrics for AI adoption: We still care about quality, developer experience, and business impact.</p></li><li><p>Model and encourage AI use from the top down.</p></li><li><p>Remove the fear and stigma from using AI: Make employees reassured that using AI isn&#8217;t cheating.</p></li><li><p>Carve out time for employees to experiment with AI tools.</p></li><li><p>Have clear conversations around expectations around AI use.</p></li><li><p>Host webinars, in-person trainings, and office hours.</p></li><li><p>Champions programs: Identify the early adopters to help evangelize and drive adoption with their peers.</p></li><li><p>Use DX&#8217;s <a href="https://getdx.com/guide/ai-assisted-engineering/">Guide to AI assisted engineering</a></p></li><li><p><strong>DORA&#8217;s recent AI report shows a 451% increase in adoption among companies with an acceptable use policy.</strong></p></li></ul><p><strong>Key measures from the DX Core 4 productivity framework</strong></p><ul><li><p><strong>For quality: </strong>Measure change failure rate before and after introducing AI tools.</p></li><li><p><strong>For speed</strong>: Measure PR throughput.</p></li><li><p><strong>For quality:</strong> Survey developers. Many report improved code readability at companies with high AI adoption.</p></li><li><p><strong>Always stay on top of developer experience.</strong></p></li></ul><p><strong>In this episode, we cover:</strong></p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw">00:00</a>) Intro: The full spectrum of AI adoption</p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw&amp;t=182s">03:02</a>) The hype of AI</p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw&amp;t=286s">04:46</a>) Some statistics around the current state of AI coding tool adoption</p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw&amp;t=447s">07:27</a>) The real barriers to AI adoption</p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw&amp;t=571s">09:31</a>) How to drive AI adoption</p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw&amp;t=947s">15:47</a>) Measuring AI&#8217;s impact</p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw&amp;t=1189s">19:49</a>) More strategies for driving AI adoption</p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw&amp;t=1434s">23:54</a>) The Methods companies are actually using to drive impact</p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw&amp;t=1755s">29:15</a>) Questions from the chat</p><p>(<a href="https://www.youtube.com/watch?v=lkDsuOkJ0Rw&amp;t=2388s">39:48</a>) Wrapping up</p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><p><strong>Referenced:</strong></p><ul><li><p><a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p></li><li><p><a href="https://getdx.com/podcast/brian-houck-ai-adoption-playbook/">The AI adoption playbook: Lessons from Microsoft's internal strategy</a></p></li><li><p><a href="https://techcrunch.com/2025/04/29/microsoft-ceo-says-up-to-30-of-the-companys-code-was-written-by-ai/">Microsoft CEO says up to 30% of the company's code was written by AI | TechCrunch</a></p></li><li><p><a href="https://www.forbes.com/sites/douglaslaney/2025/04/09/selling-ai-strategy-to-employees-shopify-ceos-manifesto/">Viral Shopify CEO Manifesto Says AI Now Mandatory For All Employees</a></p></li><li><p><a href="https://dora.dev/research/ai/gen-ai-report/">DORA | Impact of Generative AI in Software Development</a></p></li><li><p><a href="https://getdx.com/guide/ai-assisted-engineering/">Guide to AI assisted engineering</a></p></li><li><p><a href="https://www.linkedin.com/in/justinreock/">Justin Reock - DX | LinkedIn</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[DORA’s latest research on AI impact]]></title><description><![CDATA[Google&#8217;s Derek DeBellis shares how generative AI is impacting developer productivity, team flow, and how to measure its real value.]]></description><link>https://newsletter.getdx.com/p/doras-latest-research-on-ai-impact</link><guid isPermaLink="false">https://newsletter.getdx.com/p/doras-latest-research-on-ai-impact</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 23 May 2025 15:29:30 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/163979446/0e5ec43763cf60d1edd858876d1e3a22.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/i3_GQx-5YRc">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><div id="youtube2-i3_GQx-5YRc" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;i3_GQx-5YRc&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/i3_GQx-5YRc?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>DORA just released a new report on the impact of generative AI on software development productivity, and the findings were a mix of expected and surprising.</p><p>On today&#8217;s episode, I&#8217;m joined once again by <a href="https://www.linkedin.com/in/derekdebellis/">Derek DeBellis</a>, lead researcher on <strong>Google&#8217;s DORA</strong> <strong>team</strong>, to break down the key insights. We talk about how the survey was developed, why measuring productivity is so complex, and what the data actually tells us about how AI is affecting real-world teams.</p><p><strong>Some takeaways: </strong></p><p><strong>The research design process</strong></p><ul><li><p>How DORA designed survey questions to validate hypotheses, including the careful iteration of single words.</p></li><li><p>An explanation of &#8220;flow&#8221; and why the DORA team decided to use a general definition of flow for the survey</p></li><li><p>How dev time was broken into two buckets: time spent on toilsome work vs. valuable work&#8212;and what those self-reported numbers reveal.</p></li><li><p>If accuracy matters, measure time as close to the present as possible&#8212;same day or &#8220;right now&#8221;.</p></li><li><p>You don&#8217;t need perfect accuracy to learn a lot from data.</p></li><li><p>How DORA&#8217;s definition of productivity is inspired by the book <em>Slow Productivity, </em>and tied to value creation.</p></li></ul><p><strong>Positive findings from the Gen AI report</strong></p><ul><li><p>Productivity is likely to increase by 2.1% when individual AI adoption is increased by 25%&#8212;about 20 minutes.</p></li><li><p>How a 2.1% increase in productivity per individual could be dramatic across an entire organization.</p></li></ul><p><strong>Interesting and surprising findings from the report</strong></p><ul><li><p>The contradictory finding that the time spent doing valuable work went down with AI, but this is easily explained by AI making the work go faster, not actually reducing the amount of valuable work.</p></li><li><p>While AI increases personal productivity (e.g., 2.1% boost), it correlates with a 1.5% drop in delivery throughput and a 7.2% decline in stability. The good news is that some of this could be only a short-term period of working with new constraints. </p></li></ul><p><strong>Guidance for measuring AI&#8217;s impact on productivity</strong></p><ul><li><p>Make sure that all metrics are aligned with your organization&#8217;s goals and multi-faceted as well.</p></li><li><p>Don&#8217;t fall into the common trap of overestimating how much AI tools boost productivity.</p></li><li><p>Use discrepancies between the current state and goals to inform strategy.</p></li><li><p>Beware of short-term disorientation when adopting new tools.</p></li></ul><p><strong>In this episode, we cover:</strong></p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc">00:00</a>) Intro: DORA&#8217;s new Impact of Gen AI report</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=204s">03:24</a>) The methodology used to put together the surveys DORA used for the report</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=404s">06:44</a>) An example of how a single word can throw off a question</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=479s">07:59</a>) How DORA measures flow</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=638s">10:38</a>) The two ways time was measured in the recent survey</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=870s">14:30</a>) An overview of experiential surveying</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=974s">16:14</a>) Why DORA asks about time</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=1190s">19:50</a>) Why Derek calls survey results &#8216;observational data&#8217;</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=1309s">21:49</a>) Interesting findings from the report</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=1457s">24:17</a>) DORA&#8217;s definition of productivity</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=1582s">26:22</a>) Why a 2.1% increase in individual productivity is significant</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=1800s">30:00</a>) The report&#8217;s findings on decreased team delivery throughput and stability</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=1960s">32:40</a>) Tips for measuring AI&#8217;s impact on productivity</p><p>(<a href="https://www.youtube.com/watch?v=i3_GQx-5YRc&amp;t=2300s">38:20</a>) Wrap up: understanding the data</p><p><strong>Where to find Derek DeBellis:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/derekdebellis/">https://www.linkedin.com/in/derekdebellis/</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><p><strong>Referenced:</strong></p><ul><li><p><a href="https://dora.dev/research/ai/gen-ai-report/">DORA | Impact of Generative AI in Software Development</a></p></li><li><p><a href="https://getdx.com/podcast/dora-research-google/">The science behind DORA</a></p></li><li><p><a href="https://nihrecord.nih.gov/2020/03/20/yale-professor-divulges-strategies-happy-life">Yale Professor Divulges Strategies for a Happy Life</a></p></li><li><p><a href="https://www.cognitionandculture.net/blogs/olivier-morin/incredible-listening-to-when-im-64-makes-you-forget-your-age/index.html">Incredible! Listening to &#8216;When I&#8217;m 64&#8217; makes you forget your age</a></p></li><li><p><a href="https://www.amazon.com/Slow-Productivity-Accomplishment-Without-Burnout/dp/0593544854">Slow Productivity: The Lost Art of Accomplishment without Burnout</a></p></li><li><p><a href="https://getdx.com/guide/dora-space-devex/">DORA, SPACE, and DevEx: Which framework should you use?</a></p></li><li><p><a href="https://getdx.com/podcast/developer-productivity-at-microsoft/">SPACE framework, PRs per engineer, AI research</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Setting targets for developer productivity metrics]]></title><description><![CDATA[Laura Tacho, CTO at DX, explains how engineering teams can effectively use the Core 4 framework by focusing on controllable inputs instead of arbitrary metrics.]]></description><link>https://newsletter.getdx.com/p/setting-targets-for-developer-productivity</link><guid isPermaLink="false">https://newsletter.getdx.com/p/setting-targets-for-developer-productivity</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 09 May 2025 13:37:56 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/162590012/3315c8320c1e1ef84786a3e4fb2c8d20.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/5T1YPaS5s2o">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><div id="youtube2-5T1YPaS5s2o" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;5T1YPaS5s2o&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/5T1YPaS5s2o?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>In today&#8217;s episode, I&#8217;m joined by <strong><a href="https://www.linkedin.com/in/lauratacho/">Laura Tacho</a></strong>, CTO at DX, engineering leadership coach, and creator of the Core 4 framework.</p><p>We discuss why many frameworks&#8212;such as SPACE, DORA, and even Core 4&#8212;can go awry when teams focus on the wrong metrics. We explain how to get started with Core 4 by anchoring on controllable inputs rather than arbitrary targets. Whether you're just beginning with metrics or trying to course-correct a bloated system, this is a sharp, practical guide to implementing Core 4 in a way that actually works.</p><p><strong>Some Takeaways: </strong></p><p><strong>Common pitfalls with engineering metrics frameworks</strong></p><ul><li><p>Frameworks like SPACE, DORA, and Core 4 often fail when implemented without clear context.</p></li><li><p>Teams sometimes adopt metrics haphazardly, setting targets that don&#8217;t align with their specific challenges.</p></li><li><p>Some metrics may be irrelevant for certain teams&#8212;or easily gamed.</p></li><li><p>Metric overwhelm is common: tracking too many metrics leads to confusion and dilution of focus.</p></li><li><p>Leaders often fail to communicate why specific metrics are chosen or how they tie to business goals.</p></li></ul><p><strong>Goodhart&#8217;s Law: </strong>&#8220;When a measure becomes a target, it ceases to be a good measure.&#8221;</p><ul><li><p>Teams may distort or manipulate metrics to hit arbitrary goals.</p></li><li><p>Example: Reducing bug counts as a target can lead to underreporting rather than real quality improvement.</p></li></ul><p><strong>Input vs. output metrics</strong></p><ul><li><p>Input metrics are actions teams can directly control (e.g., time spent on new capabilities).</p></li><li><p>Output metrics (e.g., delivery speed, reliability) are influenced by many factors and harder to manage directly.</p></li><li><p>Focusing on inputs gives teams a clearer path to improvement without encouraging unhealthy behaviors.</p></li></ul><blockquote></blockquote><p><strong>How to implement Core 4 well</strong></p><ul><li><p>Start small&#8212;track a few metrics first before expanding.</p></li><li><p>Don&#8217;t set targets until you&#8217;ve established a baseline and understand what&#8217;s realistically controllable.</p></li><li><p>Aim for the 75th percentile to push for improvement while avoiding unrealistic pressure.</p></li><li><p>Use metrics to create a culture of reflection and continuous improvement, not judgment.</p></li></ul><p><strong>How to avoid gamification</strong></p><ul><li><p>Use multidimensional metrics to avoid tunnel vision.</p></li><li><p>Focus on input metrics.</p></li><li><p>Reward effort, learning, and progress&#8212;not just hitting numeric goals.</p></li><li><p>Include teams in the goal-setting process to increase buy-in and reduce manipulation.</p></li><li><p>Give teams space and time to make real progress.</p></li></ul><p><strong>In this episode, we cover:</strong></p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o">00:00</a>) Intro: Improving systems, not distorting data</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=140s">02:20</a>) Goal setting with the new Core 4 framework</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=481s">08:01</a>) A quick primer on Goodhart&#8217;s law</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=602s">10:02</a>) Input vs. output metrics&#8212;and why targeting outputs is problematic</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=818s">13:38</a>) A health analogy demonstrating input vs. output</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=1023s">17:03</a>) A look at how the key input metrics in Core 4 drive output metrics</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=1448s">24:08</a>) How to counteract gamification</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=1704s">28:24</a>) How to get developer buy-in</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=1848s">30:48</a>) The number of metrics to focus on</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=1964s">32:44</a>) Helping leadership and teams connect the dots to how input goals drive output</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=2120s">35:20</a>) Demonstrating business impact</p><p>(<a href="https://www.youtube.com/watch?v=5T1YPaS5s2o&amp;t=2290s">38:10</a>) Best practices for goal setting</p><p><strong>Where to find Laura Tacho:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/lauratacho/">https://www.linkedin.com/in/lauratacho/</a></p><p>&#8226; Website: <a href="https://lauratacho.com/">https://lauratacho.com/</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><p><strong>Referenced:</strong></p><ul><li><p><a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p></li><li><p><a href="https://getdx.com/podcast/">Engineering Enablement Podcast</a></p></li><li><p><a href="https://dora.dev/guides/dora-metrics-four-keys/">DORA&#8217;s software delivery metrics: the four keys</a></p></li><li><p><a href="https://getdx.com/research/space-of-developer-productivity/">The SPACE of Developer Productivity: There&#8217;s more to it than you think</a></p></li><li><p><a href="https://getdx.com/research/devex-what-actually-drives-productivity/">DevEx: What Actually Drives Productivity</a></p></li><li><p><a href="https://getdx.com/guide/dora-space-devex/">DORA, SPACE, and DevEx: Which framework should you use?</a></p></li><li><p><a href="https://en.wikipedia.org/wiki/Goodhart%27s_law">Goodhart's law</a></p></li><li><p><a href="https://www.linkedin.com/in/nicolefv/">Nicole Forsgren - Microsoft | LinkedIn</a></p></li><li><p><a href="https://en.wikipedia.org/wiki/Campbell%27s_law">Campbell's law</a></p></li><li><p><a href="https://www.lennysnewsletter.com/p/introducing-core-4-the-best-way-to">Introducing Core 4: The best way to measure and improve your product velocity</a></p></li><li><p><a href="https://getdx.com/podcast/dx-core-4-framework-overview/">DX Core 4: Framework overview, key design principles, and practical applications</a></p></li><li><p><a href="https://newsletter.getdx.com/p/dx-core-4-2024-benchmarks">DX Core 4: 2024 benchmarks - by Abi Noda</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[The AI adoption playbook: Lessons from Microsoft's internal strategy]]></title><description><![CDATA[Brian Houck from Microsoft shares how engineering teams can drive AI adoption through leadership, local champions, and clear, practical strategies.]]></description><link>https://newsletter.getdx.com/p/the-ai-adoption-playbook-lessons</link><guid isPermaLink="false">https://newsletter.getdx.com/p/the-ai-adoption-playbook-lessons</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Fri, 18 Apr 2025 14:50:05 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/160663869/c7f39dff35ea976bfe5ab09be2c1c2ec.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Listen and watch now on <strong><a href="https://youtu.be/c51ToE4pPpY">YouTube</a>, <a href="https://podcasts.apple.com/us/podcast/engineering-enablement-by-abi-noda/id1619140476">Apple</a>, and <a href="https://open.spotify.com/show/3NxjyIsuxeDMQtisDqBy7D">Spotify</a></strong>.</p><div id="youtube2-c51ToE4pPpY" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;c51ToE4pPpY&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/c51ToE4pPpY?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p><strong><a href="https://www.linkedin.com/in/brianhouck/">Brian Houck</a></strong> from Microsoft returns to discuss effective strategies for driving AI adoption among software development teams. Brian shares his insights into why the immense hype around AI often serves as a barrier rather than a facilitator for adoption, citing skepticism and inflated expectations among developers. He highlights the most effective approaches, including leadership advocacy, structured training, and cultivating local champions within teams to demonstrate practical use cases.</p><p>Brian emphasizes the importance of honest communication about AI's capabilities, avoiding over-promises, and ensuring that teams clearly understand what AI tools are best suited for. Additionally, he discusses common pitfalls, such as placing excessive pressure on individuals through leaderboards and unrealistic mandates, and stresses the importance of framing AI as an assistant rather than a replacement for developer skills. Finally, Brian explores the role of data and metrics in adoption efforts, offering practical advice on how to measure usage effectively and sustainably.</p><p><strong>Some Takeaways: </strong></p><p><strong>Barriers to AI Adoption</strong></p><ul><li><p>The biggest obstacle to AI adoption in engineering teams is not a lack of tooling or access, but developer skepticism rooted in hype and unrealistic expectations.</p></li><li><p><strong>30% of developers say their primary concern is that AI tools will not deliver on their promises, which discourages initial or continued use.</strong></p></li><li><p>When developers try an AI tool and don&#8217;t immediately experience transformational results, they often abandon it entirely.</p></li><li><p><strong>Broader narratives&#8212;like AI replacing developers&#8212;add to the skepticism, even though only about 10% of developers are actually concerned about job displacement.</strong></p></li></ul><p><strong>The Role of Leadership</strong></p><ul><li><p><strong>Leadership advocacy is a powerful lever for AI adoption. Developers are seven times more likely to be daily users when leaders actively promote and normalize the use of AI tools.</strong></p></li><li><p>Leaders should clearly communicate which AI tools are approved, what kinds of tasks they&#8217;re suitable for, and that developers are encouraged (and expected) to use them.</p></li><li><p>Messaging that overstates the capabilities of AI tools tends to backfire, creating disappointment and further resistance.</p></li><li><p>Ongoing communication&#8212;rather than one-off announcements&#8212;is essential to build trust and maintain momentum.</p></li></ul><p><strong>Team-Level Strategies and Local Champions</strong></p><ul><li><p>Adoption is more successful when it&#8217;s supported at the team level through peer learning, not just centralized rollouts.</p></li><li><p>Developers benefit most when &#8220;local champions&#8221;&#8212;respected, experienced team members&#8212;show how they are using AI tools in the context of real workflows.</p></li><li><p><strong>Organizations that rely on local champions for internal knowledge-sharing see about 22% greater adoption among developers.</strong></p></li><li><p>Brown bag sessions, peer demos, and informal walkthroughs are especially effective for making AI adoption relevant and practical.</p></li><li><p>Managers play a key role in identifying early adopters who are both enthusiastic and influential within their teams.</p></li></ul><p><strong>Measurement and Data-Driven Adoption</strong></p><ul><li><p>Successful adoption efforts rely on clear, ongoing measurement&#8212;tracking who has installed AI tools, who is actively using them, and how frequently.</p></li><li><p><strong>Rather than boiling engagement down to a single score, Brian recommends tracking usage at multiple tiers: daily, weekly, monthly, and lapsed users.</strong></p></li><li><p>Dashboards and scorecards help leaders visualize progress and encourage healthy, team-level competition.</p></li><li><p>Monitoring usage across groups&#8212;not individuals&#8212;helps foster psychological safety while still promoting accountability and improvement.</p></li></ul><p><strong>Common Pitfalls and Developer Concerns</strong></p><ul><li><p>Public leaderboards or pressuring individual developers to adopt AI tools can lead to disengagement or mistrust.</p></li><li><p><strong>Developers often worry that AI-generated code may introduce bugs, vulnerabilities, or degrade quality. This concern is second only to the fear that AI is overhyped.</strong></p></li><li><p>Some developers fear their skills may atrophy if they rely too heavily on AI; this highlights the need to reframe AI as a partner in development, not a replacement.</p></li><li><p>Organizations should address these concerns head-on through honest, nuanced messaging that balances enthusiasm with realism.</p></li></ul><p><strong>Long-Term Impact and Ongoing Research</strong></p><ul><li><p><strong>Initial studies show AI tools can improve code-writing efficiency by 5% to 30%, but broader productivity gains are harder to measure due to the multifaceted nature of software engineering work.</strong></p></li><li><p>Brian highlights the need for more research on how AI impacts long-term code maintainability, especially when it accelerates initial development phases.</p></li><li><p>Developers still spend only about 14% of their time writing code, so increasing throughput in that slice doesn&#8217;t automatically translate into overall productivity gains.</p></li></ul><p><strong>In this episode, we cover:</strong></p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY">00:00</a>) Intro: Why AI hype can hinder adoption among teams</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=107s">01:47</a>) Key strategies companies use to successfully implement AI</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=287s">04:47</a>) Understanding why adopting AI tools is uniquely challenging</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=429s">07:09</a>) How clear and consistent leadership communication boosts AI adoption</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=646s">10:46</a>) The value of team leaders ("local champions") demonstrating practical AI use</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=866s">14:26</a>) Practical advice for identifying and empowering team champions</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=991s">16:31</a>) Common mistakes companies make when encouraging AI adoption</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=1161s">19:21</a>) Simple technical reminders and nudges that encourage AI use</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=1224s">20:24</a>) Effective ways to track and measure AI usage through dashboards</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=1398s">23:18</a>) Working with team leaders and infrastructure teams to promote AI tools</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=1460s">24:20</a>) Understanding when to shift from adoption efforts to sustained use</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=1559s">25:59</a>) Insights into the real-world productivity impact of AI</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=1672s">27:52</a>) Discussing how AI affects long-term code maintenance</p><p>(<a href="https://www.youtube.com/watch?v=c51ToE4pPpY&amp;t=1742s">29:02</a>) Updates on ongoing research linking sleep quality to productivity</p><p><strong>Where to find Brian Houck:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/brianhouck/">https://www.linkedin.com/in/brianhouck/</a></p><p>&#8226; Website: <a href="https://www.microsoft.com/en-us/research/people/bhouck/">https://www.microsoft.com/en-us/research/people/bhouck/</a></p><p><strong>Where to find Abi Noda:</strong></p><p>&#8226; LinkedIn: <a href="https://www.linkedin.com/in/abinoda">https://www.linkedin.com/in/abinoda</a></p><p><strong>Referenced:</strong></p><ul><li><p><a href="https://getdx.com/corefour">DX Core 4 Productivity Framework</a></p></li><li><p><a href="https://getdx.com/podcast/">Engineering Enablement Podcast</a></p></li><li><p><a href="https://cloud.google.com/devops/dora">DORA Metrics</a></p></li><li><p><a href="https://dropbox.tech/">Dropbox Engineering Blog</a></p></li><li><p><a href="https://codeascraft.com/">Etsy Engineering Blog</a></p></li><li><p><a href="https://www.pfizer.com/science/innovation/digital">Pfizer Digital Innovation</a></p></li><li><p><a href="https://www.atlassian.com/team-playbook/plays/brown-bag">Brown Bag Sessions &#8211; A Guide</a></p></li><li><p><a href="https://visualstudio.microsoft.com/services/github-copilot/">IDE Integration and AI Tools</a></p></li><li><p><a href="https://getdx.com/developer-productivity-dashboards/">Developer Productivity Dashboard Examples</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Gene Kim on developer experience and AI engineering]]></title><description><![CDATA[Author and researcher Gene Kim explores how developer experience, platform engineering, and AI are reshaping the way organizations build software and coordinate work.]]></description><link>https://newsletter.getdx.com/p/gene-kim-on-developer-experience</link><guid isPermaLink="false">https://newsletter.getdx.com/p/gene-kim-on-developer-experience</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Mon, 07 Apr 2025 14:35:10 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/160783026/a2d7e59d63cbf3cf5fa649c7542efb40.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Author and researcher <a href="https://www.linkedin.com/in/realgenekim/">Gene Kim</a> has spent decades shaping how organizations think about DevOps, systems, and performance. In this conversation, he makes the case for why developer experience is the next chapter in that evolution and how it connects to everything from coordination theory to option value to the rise of GenAI. We discuss the deeper systems thinking behind DevEx, why tooling alone isn&#8217;t enough, and what AI means for the structure of engineering teams going forward.</p><ul><li><p>The evolving landscape of developer experience</p></li><li><p>Option value theory and why it matters</p></li><li><p>Framing developer experience effectively</p></li><li><p>GenAI's impact on development teams</p></li><li><p>Where developer experience and AI are headed</p></li></ul><p>Listen now on <a href="https://podcasts.apple.com/us/podcast/gene-kim-on-developer-experience-and-ai-engineering/id1619140476?i=1000702242678">Apple</a>, <a href="https://open.spotify.com/episode/58Fl70DnElJoBQT6TFBNZf">Spotify</a>, and <a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y">YouTube</a>.</p><div id="youtube2-4Vp_Wnpvg_E" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;4Vp_Wnpvg_E&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/4Vp_Wnpvg_E?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>Some takeaways:</p><ul><li><p><strong>Developer experience represents the next chapter in transformation journeys.</strong> Organizations increasingly view developer experience and developer platforms as the logical next progression after DevOps. This evolution isn't random but part of a coherent pattern addressing fundamental coordination challenges.</p></li><li><p><strong>The core problem is managing an ever-growing number of functional specialties.</strong> As technology becomes more complex, organizations face challenges coordinating efforts between increasingly specialized roles. What once required simple coordination now demands sophisticated social "circuitry."</p></li><li><p><strong>Platforms solve the "30 tickets" problem.</strong> When developers need to open tickets with 30 different people to get anything done, productivity stalls. Developer platforms eliminate these coordination bottlenecks by providing self-service capabilities.</p></li><li><p><strong>Focus on layer three changes.</strong> The most significant performance improvements come from changing the "social circuitry" at layer three (how people work together) rather than just changing tools or technology at layer two. Toyota's NUMMI plant transformation demonstrated this by achieving dramatic improvements with the same people and equipment.</p></li><li><p><strong>The value of options can be expressed mathematically.</strong> The formula NK/T (where N is the number of modules, K is the number of parallel experiments, and T is experiment duration) shows how organizations can increase value through experimentation.</p></li><li><p><strong>GenAI increases option value by reducing T.</strong> AI helps developers complete work faster, enabling more experiments in less time and increasing the overall option value of development work.</p></li><li><p><strong>Higher uncertainty makes options more valuable.</strong> In uncertain environments (high sigma), the ability to defer decisions until you have more information becomes increasingly valuable. This makes experimentation and fast feedback especially important.</p></li><li><p><strong>GenAI enables five key capabilities for developers.</strong> It helps them: 1) work faster, 2) be more ambitious about what they can build, 3) accomplish alone what would have required teams, 4) have more fun coding, and 5) take more "swings at bat" through rapid experimentation.</p></li><li><p><strong>Don't frame DevEx as "coddling developers."</strong> When business leaders react negatively to developer experience initiatives, reframe the conversation around what happens without good DevEx: valuable engineers can't build, test, deploy, or get anything done despite their skills.</p></li><li><p><strong>Position DevEx as a painkiller, not a vitamin.</strong> The business cost of talented developers spending hours wrestling with Docker images instead of creating value is substantial. This is a productivity problem, not a perks problem.</p></li><li><p><strong>Talent retention depends on DevEx.</strong> One financial institution lost top Silicon Valley hires within 30 days because they were still waiting on laptops, environments, and completing compliance training. DevEx is critical to retaining the best talent.</p></li><li><p><strong>The value creation happens in development.</strong> While platforms and infrastructure are important, the majority of value creation happens in ideation, research, design, and development. Fast feedback through excellent tooling is a critical prerequisite for this value creation.</p></li><li><p><strong>GenAI is changing the shape of software engineering teams.</strong> The traditional makeup of engineering organizations is being fundamentally altered by AI capabilities that augment developer skills and productivity.</p></li><li><p><strong>The "death of the stubborn developer" is underway.</strong> Tasks that companies previously relied on junior developers for can now be quickly completed by senior developers with AI. At one company, Gene gives an example of a project that would have been a summer internship that was completed by a department head in an afternoon.</p></li><li><p><strong>The "novice's optional problem" threatens talent pipelines.</strong> Similar to surgical robots that eliminate the need for junior surgeons' assistance, AI may reduce opportunities for junior developers to gain experience. This could starve the path for juniors to become seniors.</p></li><li><p><strong>Technology leaders are returning to IC roles.</strong> The increased productivity and enjoyment of coding with AI is drawing leaders back to individual contributor positions. Many prefer creating value directly over administrative management tasks.</p></li><li><p><strong>GenAI opens new horizons for all skill levels.</strong> For juniors, AI serves as an infinitely patient teacher. For seniors, it elevates their capabilities and allows them to focus on higher-level architecture and problem decomposition.</p></li><li><p><strong>Teams still need senior engineers to create the "task tree."</strong> While AI can help with leaf node tasks, experienced developers remain essential for creating architecture that enables independent action and problem decomposition.</p></li><li><p><strong>Future research will refine our understanding.</strong> New research is needed to explore how AI impacts both individual developer productivity and the social "layer three" of engineering organizations.</p></li><li><p><strong>The CHOP Handbook will codify best practices.</strong> Gene Kim and Steve Yegge are developing guidance on "Chat Oriented Programming" to help developers leverage AI for faster, more ambitious, and more enjoyable development work.</p></li></ul><p>Timestamps:</p><ul><li><p>(<a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y&amp;index=1">0:00</a>) Introduction </p></li><li><p>(<a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y&amp;index=1&amp;t=132s">2:12</a>) The evolving landscape of developer experience </p></li><li><p>(<a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y&amp;index=1&amp;t=634s">10:34</a>) Option Value theory, and how GenAI helps developers</p></li><li><p>(<a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y&amp;index=1&amp;t=825s">13:45</a>) The aim of developer experience work</p></li><li><p>(<a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y&amp;index=1&amp;t=1199s">19:59</a>) The significance of layer three changes</p></li><li><p>(<a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y&amp;index=1&amp;t=1403s">23:23</a>) Framing developer experience</p></li><li><p>(<a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y&amp;index=1&amp;t=1932s">32:12</a>) GenAI&#8217;s part in &#8220;the death of the stubborn developer&#8221;</p></li><li><p>(<a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y&amp;index=1&amp;t=2165s">36:05</a>) GenAI&#8217;s implications on the workforce</p></li><li><p>(<a href="https://www.youtube.com/watch?v=4Vp_Wnpvg_E&amp;list=PLnWh5v1e2EpZEqEwuguY7WZq3sjbwWx5y&amp;index=1&amp;t=2285s">38:05</a>) Where Gene&#8217;s work is heading</p></li></ul><p>Referenced:</p><ul><li><p><a href="https://itrevolution.com/product/the-phoenix-project/">Phoenix Project</a></p></li><li><p><a href="https://hbr.org/1999/09/decoding-the-dna-of-the-toyota-production-system">Decoding the DNA of the Toyota Production System</a></p></li><li><p><a href="https://itrevolution.com/product/wiring-the-winning-organization/">Wiring the Winning Organization</a></p></li><li><p><a href="https://itrevolution.com/events/">ETLS Vegas</a></p></li></ul><div><hr></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://newsletter.getdx.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share&quot;,&quot;text&quot;:&quot;Share Engineering Enablement&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://newsletter.getdx.com/?utm_source=substack&amp;utm_medium=email&amp;utm_content=share&amp;action=share"><span>Share Engineering Enablement</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Getting Airbnb’s Platform team to drive more impact: Reorganizing, defining strategy, and selecting metrics]]></title><description><![CDATA[Anna Sulkina, Airbnb's Developer Productivity leader, shares how her team improved to drive greater impact across the organization.]]></description><link>https://newsletter.getdx.com/p/getting-airbnbs-platform-team-to</link><guid isPermaLink="false">https://newsletter.getdx.com/p/getting-airbnbs-platform-team-to</guid><dc:creator><![CDATA[Abi Noda]]></dc:creator><pubDate>Sun, 09 Mar 2025 12:50:40 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/158619027/b1ef3e85d7dcaccf3e11833fae36175a.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>When <strong><a href="https://www.linkedin.com/in/annasulkina/">Anna Sulkina</a> </strong>joined Airbnb to lead Developer Platform, there was a feeling amongst the team that they could be even more impactful if they were to become more focused and build better connections with external stakeholders. In this conversation, Anna shares this story and goes into the key changes they made. We cover: </p><ul><li><p>Why they shifted their org structure to be focused on problems, not tools</p></li><li><p>How they clarified roles for PMs, EMs, and tech leads</p></li><li><p>How they defined their organization&#8217;s strategy</p></li><li><p>Other tactics that helped change the general perception about the team</p></li><li><p>What their measurement approach looks like today</p></li><li><p>The measurable improvements they&#8217;ve seen since making these changes</p></li></ul><p>Listen now on <a href="https://podcasts.apple.com/us/podcast/getting-airbnbs-platform-team-to-drive-more-impact/id1619140476?i=1000698317038">Apple</a>, <a href="https://open.spotify.com/episode/0SswYNfvQF11Yt5T80eVwb">Spotify</a>, and <a href="https://youtu.be/SE2OdVlx7XQ?si=UqzBYe2svPl54Ykw">YouTube</a>.</p><div id="youtube2-SE2OdVlx7XQ" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;SE2OdVlx7XQ&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/SE2OdVlx7XQ?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p><strong>Some takeaways:</strong></p><ol><li><p>Anna begins by sharing some helpful characteristics for being a developer productivity leader: deep developer empathy, experience building both end-user and internal tools, and a product mindset focused on real pain points. Strong communication and stakeholder management are just as critical as technical expertise.</p></li><li><p>When Anna joined Airbnb, the developer platform team felt they lacked clarity, operated somewhat in silos, and had overlapping roles across PMs, TPMs, and engineering leads. Teams were delivering great work, but without clear priorities or alignment, and stakeholders didn&#8217;t fully grasp the value of developer productivity. They tackled this by defining a clear strategy, restructuring teams around developer personas rather than tools, and formalizing stakeholder engagement through regular feedback loops, surveys, and advisory boards.</p></li><li><p>They restructured the org from a tool-centric model to a problem-focused approach, organizing teams into two pillars: platform-agnostic tooling (CI/CD, dev environments) and developer personas (iOS, Android, web). This shift encouraged teams to address broader developer needs rather than just optimizing individual tools, fostering collaboration on shared challenges like CI reliability and integration testing.</p></li><li><p>By clarifying how each role brought value to the company, Anna and her team were able to better define how PMs, tech leads, and EMs collaborated. Simply put, PMs own the &#8220;why,&#8221; tech leads the technical &#8220;how,&#8221; and EMs the organizational &#8220;how.&#8221; They also set clear expectations for different types of tech leads, ensuring alignment wasn&#8217;t just documented but actively reinforced in how teams worked together.</p></li><li><p>With roles clarified, Anna turned to defining overall strategy, starting with their mission: to maximize developer effectiveness by providing a reliable, easy-to-use platform for building and maintaining high-quality software at scale. She set four guiding principles: be developer-centric, simplify workflows, stay focused, and deliver incrementally, then structured the work into four focus areas: speeding up the inner loop, improving CI/CD, reducing cognitive load, and leveraging AI.</p></li><li><p>Under Anna&#8217;s leadership, developer experience at Airbnb has seen double-digit percentage improvements year over year. She says a key driver has been the developer experience survey, now run twice a year with results shared transparently across the org. By consistently gathering feedback, acting on pain points, and showing measurable progress, the team has built trust and momentum.</p></li><li><p>Anna and her team transformed Airbnb&#8217;s developer experience survey into a key decision-making tool. They refined it with industry research, partnered with senior ICs, making them collaborators rather than just stakeholders, and ensured historical continuity. This approach built empathy for the complexity of measuring productivity while incorporating valuable insights. They also improved reporting by slicing data by org level, giving leaders clearer visibility into their teams. A new tech culture section helped measure focus time and goal clarity, making the data even more actionable.</p></li></ol><p><strong>In this episode we cover:</strong></p><ul><li><p>(<a href="https://www.youtube.com/watch?v=SE2OdVlx7XQ&amp;t=0s">0:00</a>) Intro</p></li><li><p>(<a href="https://www.youtube.com/watch?v=SE2OdVlx7XQ&amp;t=100s">1:40</a>) Skills that make a great developer productivity leader</p></li><li><p>(<a href="https://www.youtube.com/watch?v=SE2OdVlx7XQ&amp;t=276s">4:36</a>) Challenges in how the team operated previously</p></li><li><p>(<a href="https://www.youtube.com/watch?v=SE2OdVlx7XQ&amp;t=649s">10:49</a>) Changing the platform org&#8217;s focus and structure</p></li><li><p>(<a href="https://www.youtube.com/watch?v=SE2OdVlx7XQ&amp;t=964s">16:04</a>) Clarifying roles for EM&#8217;s, PM&#8217;s, and tech leads</p></li><li><p>(<a href="https://www.youtube.com/watch?v=SE2OdVlx7XQ&amp;t=1222s">20:22</a>) How Airbnb defined its infrastructure org&#8217;s strategy</p></li><li><p>(<a href="https://www.youtube.com/watch?v=SE2OdVlx7XQ&amp;t=1703s">28:23</a>) Improvements they&#8217;ve seen, as measured by developer sentiment </p></li><li><p>(<a href="https://www.youtube.com/watch?v=SE2OdVlx7XQ&amp;t=1933s">32:13</a>) The evolution of Airbnb&#8217;s developer experience survey</p></li></ul><p><strong>Referenced:</strong></p><ul><li><p>Follow <a href="https://www.linkedin.com/in/annasulkina/">Anna</a> on LinkedIn</p></li><li><p>Read <a href="https://medium.com/airbnb-engineering">the Airbnb Tech blog</a> </p></li></ul>]]></content:encoded></item></channel></rss>