<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Andrew Lewis was Here]]></title><description><![CDATA[AI adoption strategies for technology leaders where governance matters, timelines are real, and nobody wants to be the cautionary tale]]></description><link>https://andrewlewis.ca</link><image><url>https://substackcdn.com/image/fetch/$s_!ISZj!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd024d759-eeb3-44b1-8186-48f22d0817b3_764x766.png</url><title>Andrew Lewis was Here</title><link>https://andrewlewis.ca</link></image><generator>Substack</generator><lastBuildDate>Mon, 27 Apr 2026 03:55:53 GMT</lastBuildDate><atom:link href="https://andrewlewis.ca/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Andrew Lewis]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[andrewlewiswashere@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[andrewlewiswashere@substack.com]]></itunes:email><itunes:name><![CDATA[Andrew Lewis]]></itunes:name></itunes:owner><itunes:author><![CDATA[Andrew Lewis]]></itunes:author><googleplay:owner><![CDATA[andrewlewiswashere@substack.com]]></googleplay:owner><googleplay:email><![CDATA[andrewlewiswashere@substack.com]]></googleplay:email><googleplay:author><![CDATA[Andrew Lewis]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Your Brain Is a Judgment Machine Now]]></title><description><![CDATA[AI didn&#8217;t eliminate the hard part of knowledge work. It compressed it into every minute of the day.]]></description><link>https://andrewlewis.ca/p/your-brain-is-a-judgment-machine</link><guid isPermaLink="false">https://andrewlewis.ca/p/your-brain-is-a-judgment-machine</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Mon, 20 Apr 2026 12:02:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!uSU1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uSU1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uSU1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!uSU1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!uSU1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!uSU1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uSU1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png" width="1408" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1774990,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewis.ca/i/194358008?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uSU1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png 424w, https://substackcdn.com/image/fetch/$s_!uSU1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png 848w, https://substackcdn.com/image/fetch/$s_!uSU1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png 1272w, https://substackcdn.com/image/fetch/$s_!uSU1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa38256f1-7890-4ec5-a042-c115b0651bd5_1408x768.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The dominant narrative about AI and work goes like this: AI handles the production, you handle the thinking, everyone goes home early. It is a clean story. It is also wrong in a way that matters.</p><p>A principal engineer at a large telecom <a href="https://www.reddit.com/r/ClaudeAI/s/0P0PHzx1aZ">posted on Reddit this week</a> about being all-in on agentic coding for two years and thinking about quitting software engineering entirely. Not because the tools don&#8217;t work. Because they work too well. The line that stopped me: <strong>&#8220;The cost of writing code in effort / time was a throttling middleware.&#8221;</strong></p><p>That phrase deserves to sit for a second. Writing code used to be slow enough that your brain could keep up. The effort of typing, compiling, debugging &#8212; all of that created natural pace. You had time to sense when a pattern was wrong. Time to think through the shape of a class or the implications of an architectural choice. The slowness wasn&#8217;t a bug. It was cognitive infrastructure.</p><p>Now that infrastructure is gone. And this engineer &#8212; 13 years of experience, principal level &#8212; reports making ten whiteboard-level architectural decisions before his second cup of coffee. Decisions that used to happen once a sprint, maybe twice, gated by the slow, expensive process of actually building things. The dam broke. The decisions didn&#8217;t disappear. They accelerated.</p><h2>The production layer was never the bottleneck you thought it was</h2><p>When people talk about AI making work faster, they&#8217;re usually describing the production layer &#8212; the part where raw effort turns into output. Drafting, coding, formatting, researching. And yes, AI compresses that layer dramatically. What nobody accounted for is what happens to the <em>other</em> layer &#8212; the judgment layer &#8212; when production speeds up by an order of magnitude.</p><p>Every piece of AI-generated output requires evaluation. Is this right? Is this good enough? Does this fit the architecture? Does this solve the actual problem or just the surface symptom? Those questions existed before AI, but they arrived at a pace your brain could absorb. Production time was thinking time. The gap between &#8220;I need this&#8221; and &#8220;here it is&#8221; gave your mind room to prepare for the decision.</p><p>That gap is gone. And the result is not faster work. It&#8217;s faster <em>judgment demands</em> on a brain that hasn&#8217;t changed speed.</p><p>A <a href="https://hbr.org/2026/03/when-using-ai-leads-to-brain-fry">Boston Consulting Group study published in Harvard Business Review</a> in March 2026 put a name to this: &#8220;AI brain fry.&#8221; They surveyed 1,488 full-time U.S. workers and found that high AI oversight &#8212; the kind that requires reading, interpreting, and evaluating AI output &#8212; was associated with 14% more mental effort, 12% greater mental fatigue, and 19% greater information overload. Workers described a fog or buzzing that forced them to physically step away from their screens. One of the study&#8217;s authors told Fortune the pattern was consistent: people were getting more done but hitting the limits of their cognitive capacity because there were simply too many decisions to make.</p><p>An eight-month study of a 200-person tech firm, led by researchers at UC Berkeley, found the same dynamic from a different angle. AI wasn&#8217;t reducing work. It was intensifying it. Employees processed more information, made more decisions, and experienced more burnout &#8212; not less &#8212; as AI adoption increased.</p><h2>Decision fatigue is not a new concept. The delivery mechanism is.</h2><p>Psychologists have studied decision fatigue for decades. The core finding is straightforward: the quality of your decisions degrades as you make more of them. Roy Baumeister&#8217;s ego depletion research established that decision-making draws from a finite cognitive resource. Make enough decisions and you start defaulting to heuristics, avoiding trade-offs, or simply deferring. The average American adult reportedly makes around 35,000 decisions a day. Most of those are trivial. The ones that matter are the ones that require actual evaluation.</p><p>What AI does is change the ratio. It doesn&#8217;t increase the total number of decisions. It increases the <em>density of consequential ones</em>. When production was slow, your day was a mix of low-stakes mechanical work and occasional high-stakes judgment calls. The mechanical work gave your brain recovery time between the hard decisions. It was boring, but it was load-bearing.</p><p>Remove the mechanical work and what&#8217;s left is a continuous stream of judgment. Architecture choices. Quality assessments. Risk evaluations. Strategic trade-offs. All day. No recovery intervals. The Reddit poster described it precisely: running ten whiteboard-level decisions before morning coffee, decisions that used to be spaced across a sprint. His brain isn&#8217;t slower than it was. The demand on it is faster.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://andrewlewis.ca/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you're reading this far, you've probably felt this yourself. I write about this every week.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>We don&#8217;t understand the cost of running a judgment machine all day</h2><p>This is the part almost nobody is talking about. We&#8217;ve spent two years celebrating AI&#8217;s ability to remove the production burden. We have not spent two minutes thinking about what happens to the human on the other side when the production burden was also a cognitive pacing mechanism.</p><p>The Reddit engineer said something else that stuck: &#8220;I feel like for the devs that have survived layoff rounds, AI has <em>raised</em> the bar of required skills, not lowered it.&#8221; That maps directly to the Jevons Paradox applied to AI &#8212; as AI efficiency increases, the demand for human capability doesn&#8217;t decrease. It increases. The skills that matter shift upward. The judgment, the architectural thinking, the ability to evaluate quality at speed &#8212; those become the job. And the job becomes relentlessly, uninterruptedly hard.</p><p>This isn&#8217;t a coding problem. It&#8217;s a knowledge work problem. Every profession that adopts AI tools effectively will hit this same wall. Lawyers reviewing AI-drafted contracts. Financial analysts evaluating AI-generated models. Marketers assessing AI-produced campaigns. The production layer compresses. The judgment layer concentrates. And the person in the middle has to run their brain at a sustained intensity that the old workflow never required.</p><p>We don&#8217;t have infrastructure for this yet. We don&#8217;t have pacing strategies. We don&#8217;t have cognitive load frameworks adapted for AI-augmented work. We don&#8217;t even have language for the problem &#8212; which is why &#8220;brain fry&#8221; and &#8220;throttling middleware&#8221; resonate so immediately. People recognize the feeling before anyone names it.</p><h2>The work behind the work just got more urgent</h2><p>The conventional response to this problem will be training. Run a workshop on managing AI output. Distribute a tip sheet on decision prioritization. That approach will fail for the same reason it always fails &#8212; it treats the symptom without touching the structure.</p><p>The actual work is harder than a workshop. It&#8217;s developing the ability to see your own workflow clearly enough to know which judgments matter and which don&#8217;t. To calibrate your trust in AI output so you&#8217;re not re-evaluating everything at full intensity. To build the signal discrimination that lets you spot the 5% of output that needs real attention and let the rest move.</p><p>That is not a training problem. That is a capability development problem. And it&#8217;s one that gets more urgent, not less, as the tools get faster.</p><p>Your brain was always a judgment machine. AI just made it the only machine that matters.</p><div><hr></div><p>I'm writing from inside a regulated firm doing AI adoption in real time. Every week I publish what I'm seeing &#8212; the frameworks, the friction, the decisions that actually move work. If that's useful to you, subscribe. It's free, and I'll send you the next one.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://andrewlewis.ca/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://andrewlewis.ca/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Your AI Adoption Number Is Lying to You]]></title><description><![CDATA[The metric everyone tracks, the one almost nobody does, and why the gap between them explains everything]]></description><link>https://andrewlewis.ca/p/your-ai-adoption-number-is-lying</link><guid isPermaLink="false">https://andrewlewis.ca/p/your-ai-adoption-number-is-lying</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Thu, 16 Apr 2026 13:11:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!dNtp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dNtp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dNtp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dNtp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dNtp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dNtp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dNtp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:346480,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewis.ca/i/193750041?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dNtp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dNtp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dNtp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dNtp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc7392d3f-3c16-48a2-9345-ab6f30adf143_1024x1024.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Every enterprise AI dashboard in the world has an adoption number on it. Seats activated, prompts per user, tools deployed, percentage of workforce with access. The number goes up every quarter. The board is pleased. The CIO presents it with confidence.</p><p>And almost none of it tells you whether AI is actually working.</p><h2>The metric everyone loves is the metric that matters least</h2><p>Adoption measures activity. Someone logged in. Someone typed a prompt. Someone opened Copilot and asked it to summarize an email they could have read in thirty seconds. All of that registers as adoption.</p><p>What it doesn&#8217;t measure is whether anyone&#8217;s work changed. Whether a contract review that used to take four hours now takes ninety minutes because the lawyer iterated through three rounds of AI-assisted markup. Whether the finance team built an automated reconciliation workflow or just asked ChatGPT to explain what a pivot table does. Whether the senior associate used AI to surface a pattern across two hundred documents that no one would have caught manually &#8212; or whether they used it to rewrite a slightly awkward email.</p><p>Those are fundamentally different behaviors. One is sophisticated use. The other is expensive autocomplete. And the adoption dashboard doesn&#8217;t distinguish between them.</p><h2>The evidence arrived this week, from three directions at once</h2><p>A Wharton School study tracking enterprise AI adoption over multiple years found a widening disconnect between executive enthusiasm and managerial reality. Nearly two-thirds of executives report becoming significantly more optimistic about AI over the past year. The managers implementing those same tools inside actual workflows report something closer to frustration. They see the constraints. They carry the operational burden. They don&#8217;t feel supported.</p><p>A separate survey of 2,400 knowledge workers found that 29% admit to actively undermining their company&#8217;s AI rollout. Among Gen Z workers, that figure reaches 44%. The tactics include feeding proprietary data into unapproved tools, refusing to complete AI training, and in some cases deliberately producing poor-quality output to make the tools look bad.</p><p>And a third study, from HFS Research, found that only 14% of enterprises have a clear AI strategy at all. The other 86% are deploying tools into a vacuum &#8212; no framework for what good use looks like, no feedback loop for what&#8217;s working, no definition of success beyond the adoption dashboard.</p><p>These aren&#8217;t three separate problems. They&#8217;re three symptoms of the same one: organizations measuring the wrong thing, at the wrong level, and mistaking activity for progress.</p><h2>What sophistication actually measures</h2><p>The Conversation Sophistication Score is a framework built on research from KPMG and the University of Texas at Austin. The study analyzed 1.4 million AI interactions across 2,500 employees and identified thirty behavioral characteristics that separate the highest-performing AI users from everyone else.</p><p>The finding that matters: sophistication isn&#8217;t correlated with frequency. The people using AI most often aren&#8217;t the ones using it best. The distinguishing behaviors are things like interaction depth (multi-turn conversations that build on previous outputs), task complexity (applying AI to genuinely difficult problems rather than simple lookups), iterative reasoning (treating AI output as a draft to be refined, not an answer to be accepted), breadth of application (using AI across multiple domains rather than one narrow use case), and fluency signals (adapting communication style and prompt structure to the specific task).</p><p>None of those show up on an adoption dashboard. You can have 95% adoption and 10% sophistication, and your metrics will tell you everything is going great until it very clearly isn&#8217;t.</p><h2>The sequence that actually works</h2><p>The enterprises getting AI right aren&#8217;t doing anything exotic. They&#8217;re solving the structural problems before chasing the visible ones. They define what good AI use looks like before measuring whether it exists. They build governance &#8212; not as a compliance exercise but as a shared understanding of what&#8217;s allowed, what&#8217;s encouraged, and what&#8217;s off-limits. They create conditions where experimentation is safe and failure is data, not career risk.</p><p>And then &#8212; after the structure exists &#8212; they start measuring sophistication alongside adoption. Not instead of it. Alongside it. Because adoption without sophistication is just expensive access. And sophistication without adoption means you have a few brilliant users surrounded by a workforce that&#8217;s opted out.</p><p>The 14% of organizations that have a clear strategy? They&#8217;re the ones building this foundation. The other 86% are wondering why their adoption numbers keep climbing and their results don&#8217;t.</p><p>The dashboard isn&#8217;t broken. The measurement is.</p><div><hr></div><p><em>If you found this useful, consider sharing it with someone leading an AI rollout right now. They probably have an adoption number. They probably don&#8217;t have a sophistication score. That gap is the article.</em></p>]]></content:encoded></item><item><title><![CDATA[Two Conversations About AI, One Building]]></title><description><![CDATA[Executives and managers aren't disagreeing about AI. They're having entirely different discussions.]]></description><link>https://andrewlewis.ca/p/two-conversations-about-ai-one-building</link><guid isPermaLink="false">https://andrewlewis.ca/p/two-conversations-about-ai-one-building</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Thu, 16 Apr 2026 12:00:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Lb3M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Lb3M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Lb3M!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Lb3M!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Lb3M!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Lb3M!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Lb3M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg" width="1408" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Generated Image&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Generated Image" title="Generated Image" srcset="https://substackcdn.com/image/fetch/$s_!Lb3M!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Lb3M!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Lb3M!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Lb3M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c771a72-2553-4b79-8953-ea4edd37e9a4_1408x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Something odd is happening inside organizations right now, and it showed up clearly in a study the Wharton School published this week. The executive suite and the management layer are both talking about AI. They&#8217;re using similar language. They&#8217;re attending the same town halls and reading the same strategy decks. But they are not having the same conversation.</p><p>Nearly two-thirds of executives say they&#8217;ve become significantly more positive about AI over the past year. They see it as a strategic priority. They&#8217;re investing heavily. Some are restructuring their organizations around it. For senior leadership, AI has moved from an interesting capability to an existential commitment &#8212; the question isn&#8217;t whether to go all in, it&#8217;s how fast.</p><p>The managers one or two levels below them? They&#8217;re drowning.</p><p>The gap isn&#8217;t disagreement &#8212; it&#8217;s altitude</p><p>The Wharton/GBK Collective study has been tracking this dynamic for multiple years, and the pattern is consistent: executives experience AI as an opportunity. Managers experience AI as a workload.</p><p>That&#8217;s not because managers are resistant or uninformed. It&#8217;s because they sit at the exact altitude where strategy becomes operations. They&#8217;re the ones who have to reconcile the CEO&#8217;s enthusiasm with the fact that the approved tool doesn&#8217;t integrate with the case management system. They&#8217;re fielding questions from their teams about what&#8217;s allowed and what isn&#8217;t &#8212; often without clear answers, because the governance policy is still being drafted. They&#8217;re absorbing the productivity overhead of learning new tools while still delivering on every existing deadline.</p><p>Executives don&#8217;t see this overhead because it doesn&#8217;t appear in the metrics they review. Adoption dashboards show seats activated and usage trends. They don&#8217;t show the manager who spent three hours this week answering the same AI governance question from four different direct reports because the policy FAQ doesn&#8217;t exist yet.</p><p>The executive sees a line going up. The manager feels the weight behind that line.</p><p>Where the pressure compounds</p><p>This altitude gap produces a specific set of downstream problems that I see repeatedly from inside an organization going through this.</p><p>The first is unfunded mandates. Leadership communicates that AI adoption is a priority. But the time, training, and governance infrastructure required to adopt responsibly aren&#8217;t budgeted separately. They&#8217;re absorbed by the existing management layer on top of everything else. The implicit expectation is that managers will figure it out &#8212; will become AI champions in addition to their actual roles, without reduced workloads or additional resources.</p><p>The second is phantom consensus. Strategy decks present AI adoption as an aligned organizational priority. Everyone nods in the planning meeting. But alignment at the strategy level doesn&#8217;t mean alignment at the implementation level. The manager who nodded in the meeting goes back to a team that has no idea what&#8217;s expected, using tools that half-work, under policies that are still in draft. The strategy deck says &#8220;aligned.&#8221; The floor says &#8220;confused.&#8221;</p><p>The third is consequence asymmetry. The Fortune/Workplace Intelligence survey found that 60% of executives would consider cutting employees who refuse to adopt AI. But only 14% of enterprises have a clear AI strategy. Executives are prepared to enforce adoption of something they haven&#8217;t fully defined. The consequence falls on the people closest to the work, while the strategic ambiguity originates at the top.</p><p>What the sabotage data is actually telling us</p><p>This is where the 44% Gen Z sabotage number from this week&#8217;s headlines becomes less scandalous and more predictable.</p><p>When managers are unsupported, their teams feel it. The confusion rolls downhill. If a manager doesn&#8217;t have clear governance guidance, their team gets inconsistent answers about what&#8217;s allowed. If a manager hasn&#8217;t been given time to understand the tools, they can&#8217;t coach their team on effective use. If a manager is overwhelmed by the implementation burden, their team reads that stress and interprets it &#8212; correctly &#8212; as a signal that AI adoption is creating problems, not solving them.</p><p>The 26% of sabotaging employees who say the strategy is poorly executed aren&#8217;t making an abstract complaint. They&#8217;re reporting what they observe at the manager level every day: pressure without support, mandates without clarity, consequences without strategy.</p><p>The sabotage isn&#8217;t coming from below. It&#8217;s flowing downhill from above.</p><p>The intervention that most organizations skip</p><p>The conventional response is more training, better tools, clearer communication from leadership. Those help. But they miss the structural problem.</p><p>The structural intervention is resourcing the middle. Giving managers dedicated time for AI governance and enablement work. Reducing their operational load during the adoption period rather than adding to it. Creating feedback channels that surface implementation friction to leadership before it calcifies into resistance.</p><p>The organizations I&#8217;ve seen move fastest on AI adoption aren&#8217;t the ones with the biggest budgets or the most ambitious CEOs. They&#8217;re the ones where middle management has actual capacity to do the work that adoption requires &#8212; the unglamorous, invisible, operational work of translating executive vision into something a team of eight people can actually execute on a Tuesday afternoon.</p><p>That work doesn&#8217;t appear on an adoption dashboard. It doesn&#8217;t generate a conference keynote. But it&#8217;s the difference between a strategy that lands and one that produces a 44% sabotage rate.</p><p>The executives and the managers aren&#8217;t enemies. They&#8217;re not even disagreeing. They&#8217;re just standing at different altitudes, describing different views of the same mountain &#8212; and nobody&#8217;s built the trail between them.</p><p>---</p><p><a href="https://youtu.be/zC3WOZ3FNKY">I made a video this week breaking down the full research</a> &#8212; the sabotage data, the Wharton findings, and a framework for measuring sophistication instead of activity. </p>]]></content:encoded></item><item><title><![CDATA[Your AI Metrics Are Measuring the Wrong Thing]]></title><description><![CDATA[A research-backed framework for measuring sophistication, not just activity.]]></description><link>https://andrewlewis.ca/p/your-ai-metrics-are-measuring-the</link><guid isPermaLink="false">https://andrewlewis.ca/p/your-ai-metrics-are-measuring-the</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Wed, 08 Apr 2026 12:20:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!B9my!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!B9my!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!B9my!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!B9my!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!B9my!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!B9my!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!B9my!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1572089,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewis.ca/i/193348866?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!B9my!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!B9my!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!B9my!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!B9my!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd297b328-56e6-43a7-a2e3-015d9abcaeed_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Most organizations measure AI adoption the way they measure gym memberships. How many people signed up. How often they swipe in. Maybe how long they stay. None of which tells you whether anyone is actually getting stronger.</p><p>The AI equivalent: prompt counts, hours logged, tokens consumed, self-assessed skill levels. These numbers are easy to collect, and at most companies they look encouraging. Adoption is up. Usage is growing. The dashboards are green.</p><p>But adoption is not sophistication. And the gap between the two is enormous.</p><h2>What the Research Actually Shows</h2><p><a href="https://kpmg.com/us/en/media/news/utaustin-kpmg-study.html">KPMG and researchers at the University of Texas at Austin spent eight months studying this gap.</a> They analyzed 1.4 million AI prompts from roughly 2,500 professionals &#8212; not surveys, not self-reports, but actual conversation logs at scale. The question was simple: what does sophisticated AI use look like, and how do you tell it apart from routine use?</p><p>The headline finding: about 90% of employees used AI regularly. Only approximately 5% used it in ways that generated differentiated value. That&#8217;s a 17:1 ratio between adoption and sophistication, and most organizations can&#8217;t see it because they&#8217;re measuring the wrong dimension entirely.</p><p>The researchers identified four behavioral patterns that consistently predicted sophisticated use. Not prompt length. Not frequency. Behaviors: treating AI as a reasoning partner rather than accepting first outputs, delegating complex multi-step tasks with clear constraints, applying AI across diverse task types instead of just writing assistance, and sustaining longer working-session-style interactions.</p><p>Here&#8217;s the part that caught me off guard. The most sophisticated users weren&#8217;t the youngest employees &#8212; they were above manager level. The conventional wisdom says junior employees are more natural with these tools. The data says otherwise. There&#8217;s a real difference between being comfortable with AI and being good at getting results from it. Comfort is about familiarity. Sophistication is about judgment.</p><h2>The Problem with Averages</h2><p>When I started building a scoring framework from this research, I ran into an interesting design problem. A weighted average of behavioral dimensions sounds clean, but it lies in predictable ways.</p><p>Consider someone who writes long, detailed initial prompts and sustains multi-turn conversations. Their Interaction Depth score is high &#8212; maybe an 8 or 9. But they never refine outputs. Never push back. Never ask the model to check its reasoning or explore alternatives. Their Iterative Reasoning score is a 2.</p><p>A weighted average might land them at &#8220;Proficient.&#8221; But they&#8217;re not proficient. They&#8217;re just verbose. The length of the prompt isn&#8217;t the signal. What the user does with the output is.</p><p>This is why the framework I built includes gating criteria &#8212; floor rules that prevent misclassification. You can&#8217;t reach the Advanced tier unless both Task Complexity and Iterative Reasoning hit at least 7 out of 10, regardless of what your weighted average says. Those two dimensions are the strongest differentiators in the research, and they carry 55% of the total score.</p><p>The gating mechanism is the single most useful idea in the framework. It forces honest measurement.</p><h2>What This Means for How You Train</h2><p>The dimension-level data tells you something activity metrics never can: where to invest in training.</p><p>If Iterative Reasoning is consistently low across your organization, another &#8220;intro to prompting&#8221; workshop won&#8217;t help. The gap isn&#8217;t in how people write prompts &#8212; it&#8217;s in how they think about the interaction. They need to learn to treat AI as a reasoning partner: assign roles, provide examples, test assumptions, ask the model to verify its own logic.</p><p>If Task Complexity is low, the problem is different. People aren&#8217;t delegating hard enough. They&#8217;re using AI for tasks they could do themselves in roughly the same time, instead of delegating the genuinely complex, multi-step work where AI creates real operating margin.</p><p>The dimension scores give you a specific diagnosis. The diagnosis gives you a specific intervention. That&#8217;s the difference between &#8220;use AI more&#8221; and &#8220;here&#8217;s what to change about how you use it.&#8221;</p><h2>The Uncomfortable Implication</h2><p>If only 5% of users are sophisticated at a firm where 90% are active &#8212; a firm that had invested heavily in AI tools and training &#8212; then sophisticated use doesn&#8217;t happen organically. Making tools available and running training sessions gets you to 90% adoption. It does not get you to sophistication.</p><p>Getting there requires measuring the right things, making specific behaviors visible and expected, and building the feedback loops that help people see the gap between where they are and where they could be. Activity metrics can&#8217;t do that. Behavioral metrics can.</p><p>You can&#8217;t operate what you can&#8217;t measure. And right now, most organizations are measuring the equivalent of gym swipes.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://andrewlewis.ca/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://andrewlewis.ca/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>I built a full playbook with the five scoring dimensions, weighted formula, gating rules, score anchors, and a printable worksheet for manual scoring. <a href="https://drive.google.com/file/d/1b-y0wAKx1QkXPS58OPi-VpKpILzVpjTk/view?usp=sharing">I also built a Claude Skill</a> if you want to get sophistication scoring within you conversation, along with tips on how to improve.</p><div class="file-embed-wrapper" data-component-name="FileToDOM"><div class="file-embed-container-reader"><div class="file-embed-container-top"><image class="file-embed-thumbnail-default" src="https://substackcdn.com/image/fetch/$s_!0Cy0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack.com%2Fimg%2Fattachment_icon.svg"></image><div class="file-embed-details"><div class="file-embed-details-h1">Ai Sophistication Scoring Spec</div><div class="file-embed-details-h2">237KB &#8729; PDF file</div></div><a class="file-embed-button wide" href="https://andrewlewiswashere.substack.com/api/v1/file/4ee665d0-fc77-49ce-a0cd-1aa01a75a7bf.pdf"><span class="file-embed-button-text">Download</span></a></div><a class="file-embed-button narrow" href="https://andrewlewiswashere.substack.com/api/v1/file/4ee665d0-fc77-49ce-a0cd-1aa01a75a7bf.pdf"><span class="file-embed-button-text">Download</span></a></div></div><div class="file-embed-wrapper" data-component-name="FileToDOM"><div class="file-embed-container-reader"><div class="file-embed-container-top"><image class="file-embed-thumbnail-default" src="https://substackcdn.com/image/fetch/$s_!0Cy0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack.com%2Fimg%2Fattachment_icon.svg"></image><div class="file-embed-details"><div class="file-embed-details-h1">Ai Sophistication Playbook</div><div class="file-embed-details-h2">37.8KB &#8729; PDF file</div></div><a class="file-embed-button wide" href="https://andrewlewiswashere.substack.com/api/v1/file/cd329de4-280e-4e73-9eff-84681597c62e.pdf"><span class="file-embed-button-text">Download</span></a></div><a class="file-embed-button narrow" href="https://andrewlewiswashere.substack.com/api/v1/file/cd329de4-280e-4e73-9eff-84681597c62e.pdf"><span class="file-embed-button-text">Download</span></a></div></div><p></p><p> </p>]]></content:encoded></item><item><title><![CDATA[The Barriers to AI Adoption Are the Job]]></title><description><![CDATA[Everyone agrees on what&#8217;s slowing AI down. Nobody wants to do the actual work.]]></description><link>https://andrewlewis.ca/p/the-barriers-to-ai-adoption-are-the</link><guid isPermaLink="false">https://andrewlewis.ca/p/the-barriers-to-ai-adoption-are-the</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Wed, 08 Apr 2026 12:02:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!7dff!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7dff!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7dff!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!7dff!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!7dff!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!7dff!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7dff!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1596077,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewis.ca/i/192676220?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7dff!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!7dff!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!7dff!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!7dff!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b747cc-b72e-43c8-9fb2-fac8f4c4cf12_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Every few weeks, a new report lands cataloguing the barriers to enterprise AI adoption. Poor data management. Incomplete governance. Lack of training. Unclear ROI. Resistance to change.</p><p><a href="https://www.nojitter.com/ai-automation/multiple-roadblocks-impede-generative-ai-adoption">No Jitter published one this week</a>. <a href="https://www.bcg.com/press/27march2026-ai-expectations-rise-in-logistics-scaled-adoption-remains-limited">BCG has one</a>. <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey&#8217;s latest State of AI survey</a> says 88% of organizations report using AI in at least one business function &#8212; but nearly two-thirds haven&#8217;t begun scaling it across the enterprise.</p><p>An <a href="https://hbr.org/2026/02/why-ai-adoption-stalls-according-to-industry-data">HBR piece from earlier this year</a> put a finer point on it: AI initiatives stall not because the technology fails, but because employees&#8217; anxiety about relevance, identity, and job security drives surface-level adoption without real commitment.</p><p>None of this is new. And that&#8217;s exactly the problem.</p><p>We&#8217;ve been listing the same barriers for two years. The list hasn&#8217;t changed. Which raises an uncomfortable question: if we know what the barriers are, why haven&#8217;t we removed them?</p><div><hr></div><h2>The Barrier List Is Actually a Job Description</h2><p>I lead AI enablement and delivery at a large Canadian law firm. My charter covers generative AI governance, production infrastructure decisions, cross-functional technology guidance, and organizational design for autonomous delivery teams.</p><p>When I read reports listing the top barriers to AI adoption, I don&#8217;t see obstacles. I see my to-do list.</p><p>Data readiness? That&#8217;s the conversation with InfoSec about what categories of data we can and can&#8217;t share with AI tools &#8212; and under what conditions. It&#8217;s the ongoing discussion about zero data retention and data residency requirements that shapes every infrastructure decision we make. And these aren&#8217;t theoretical concerns. The <a href="https://www.theregister.com/2025/07/25/microsoft_admits_it_cannot_guarantee">U.S. CLOUD Act</a> allows American law enforcement to demand data held by U.S.-based providers regardless of where that data is physically stored. A Canadian firm using a U.S.-hosted AI tool with client data on a Montreal server is still exposed. When a <a href="https://winbuzzer.com/2025/07/25/microsoft-admits-it-cannot-guarantee-eu-cloud-data-sovereignty-from-us-government-xcxwbn/">Microsoft executive told the French Senate</a> under oath in June 2025 that he couldn&#8217;t guarantee European data would be safe from U.S. authorities, the same logic applies to every law firm using a U.S.-owned cloud AI service.</p><p>Governance gaps? That&#8217;s the policy work &#8212; figuring out what guardrails are necessary when you&#8217;re building internal infrastructure like MCP servers that connect AI tools to live systems. The answer changes as the tools evolve, which means the governance has to evolve with it.</p><p>The other barriers on the list &#8212; training, middle management alignment, cultural resistance &#8212; those show up too. Every organization doing this work recognizes them. The question isn&#8217;t whether they&#8217;re real. It&#8217;s whether anyone is resourced to actually address them.</p><p>The barrier list isn&#8217;t a research finding. It&#8217;s a scope of work.</p><div><hr></div><h2>The Real Reason the List Doesn&#8217;t Change</h2><p>From the inside, the pattern is pretty clear: the barriers persist because removing them requires a kind of work that most organizations don&#8217;t want to fund, staff, or prioritize.</p><p>Governance work is slow, political, and unglamorous. Nobody gets promoted for writing an AI acceptable use policy. Nobody gets a conference keynote for building a data classification framework. These are infrastructure projects &#8212; essential, invisible, and easy to defer.</p><p>Training is continuous, not one-shot. You can&#8217;t run a lunch-and-learn in Q1 and call it done. AI tools change. Use cases evolve. New risks emerge. Training has to be ongoing, role-specific, and practical. That takes dedicated time from people who are already stretched.</p><p>Middle management is where adoption lives or dies. A <a href="https://www.workera.ai/blog/ai-adoption-will-remain-uneven-in-2026-heres-why-and-how-to-fix-it">Workera analysis from earlier this year</a> noted that most adoption bottlenecks sit in the middle of the organizational chart &#8212; managers who struggle to set expectations for AI-assisted work, don&#8217;t know what &#8220;good&#8221; looks like, and avoid the topic because it raises uncomfortable questions about headcount and value.</p><p>And cultural resistance isn&#8217;t about employees being Luddites. The <a href="https://hbr.org/2026/02/why-ai-adoption-stalls-according-to-industry-data">HBR research</a> showed it&#8217;s really about identity &#8212; people worry that using AI makes them look replaceable, or that their expertise is being devalued. You can&#8217;t train your way out of that. It requires sustained, honest messaging about what AI actually changes and what it doesn&#8217;t.</p><p>Every one of these barriers is addressable. None of them is easy. And most of them require organizational commitment that goes well beyond the technology team.</p><div><hr></div><h2>Where the Work Actually Lives</h2><p>When I look at the barrier list through the lens of what has to happen, the work sorts itself pretty naturally.</p><p>Some of it you can systematize. Governance templates, usage policies, risk assessment frameworks, data classification standards &#8212; these are repeatable artifacts. Build them once, adapt across the organization. AI can even help with its own adoption here. Use it to draft the policies, generate training materials, structure rollout plans.</p><p>Some of it requires people working alongside people. Helping managers evaluate AI-assisted work. Coaching teams on where AI fits their specific workflows. Having the uncomfortable conversations about what changes when a task that used to take four hours now takes forty minutes. You can support this with tools and structured conversations, but you can&#8217;t skip the human part.</p><p>And some of it you just have to do the slow way. Building trust. Shifting culture. Showing through consistent action that efficiency gains won&#8217;t quietly become headcount cuts. That the goal is better work, not cheaper workers. No technology accelerates this.</p><p>Most organizations want to spend their AI budget on the first kind. The actual work is mostly the second and third.</p><div><hr></div><h2>The Patience Problem</h2><p>There&#8217;s an additional dynamic that the barrier reports don&#8217;t capture: the gap between executive expectations and organizational readiness.</p><p>Leadership sees the reports about AI productivity gains. They hear the vendor pitches. They attend the conferences. They come back wanting to know why the organization isn&#8217;t moving faster.</p><p>The honest answer is: because we&#8217;re doing the barrier removal work. And that work doesn&#8217;t produce demo-ready results on a quarterly cadence.</p><p>Governance frameworks aren&#8217;t impressive in a board presentation. Training programs don&#8217;t generate viral LinkedIn posts. The slow, steady work of building organizational readiness for AI is the least visible and most important investment a company can make right now.</p><p>The organizations that are furthest along in AI adoption aren&#8217;t the ones that skipped the barrier work. They&#8217;re the ones that started it earlier and funded it properly. They invested in the boring infrastructure while everyone else was running pilot programs.</p><div><hr></div><h2>What Would Actually Change Things</h2><p>The next time one of these reports drops, resist the urge to nod along and move on.</p><p>Instead, ask: which of these barriers are we actively working to remove? Who owns each one? What resources are behind it? What does progress look like on a 90-day horizon?</p><p>If you can&#8217;t answer those questions, the barrier list isn&#8217;t research. It&#8217;s a mirror.</p><p>These barriers have been known for two years. They&#8217;ll still be known next year if the only response is acknowledging them in another planning deck.</p><p>Somebody has to do the work. In most organizations, that role either doesn&#8217;t exist yet or doesn&#8217;t have the authority to act.</p><p>That&#8217;s the barrier nobody puts on the list.</p><div><hr></div><p><em>I&#8217;d be curious to hear which barrier your organization spends the most time acknowledging and the least time actually working on.</em></p>]]></content:encoded></item><item><title><![CDATA[Your Employees Are Already Using AI. That’s the Good News.]]></title><description><![CDATA[Why shadow AI isn&#8217;t a security crisis &#8212; it&#8217;s a demand signal you&#8217;re ignoring.]]></description><link>https://andrewlewis.ca/p/your-employees-are-already-using</link><guid isPermaLink="false">https://andrewlewis.ca/p/your-employees-are-already-using</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Wed, 01 Apr 2026 12:03:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!VBcR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VBcR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VBcR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!VBcR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!VBcR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!VBcR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VBcR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/beb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:767605,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewis.ca/i/192676140?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VBcR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!VBcR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!VBcR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!VBcR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbeb309a2-0c99-4a89-9c91-a82086e0821e_1024x1024.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There&#8217;s a term making the rounds in enterprise circles right now: shadow AI.</p><p>If you work in security or IT governance, you probably hear it as a threat. Employees using ChatGPT on personal accounts. Pasting client data into tools nobody approved. Building little automations that nobody knows about until something breaks.</p><p>And yes &#8212; that&#8217;s real. The data on it is pretty stark. <a href="https://www.microsoft.com/en-us/worklab/work-trend-index/ai-at-work-is-here-now-comes-the-hard-part">Microsoft&#8217;s 2024 Work Trend Index</a> found that 75% of knowledge workers now use AI at work, and 78% of them are bringing their own tools &#8212; what Microsoft calls &#8220;BYOAI.&#8221; A <a href="https://www.cio.com/article/4124760/roughly-half-of-employees-are-using-unsanctioned-ai-tools-and-enterprise-leaders-are-major-culprits.html">BlackFog survey</a> found roughly half of workers admit to using AI tools without employer approval. A <a href="https://cogentinfo.com/resources/the-shadow-productivity-economy-when-employees-use-ai-in-secret">KPMG study</a> puts it even higher &#8212; 57% say they&#8217;ve submitted AI-generated work as their own without telling their manager.</p><p>But the part most of the coverage misses is that people don&#8217;t hide their AI use out of malice. They hide it because nobody told them it was okay. No guidance, no approved tools, no one saying &#8220;yes, use this &#8212; here&#8217;s how to do it without creating a mess.&#8221;</p><div><hr></div><h2>The View From Inside</h2><p>I run AI enablement and delivery at a large Canadian law firm. When I say &#8220;run,&#8221; I mean I&#8217;m the person in the room when we decide what tools to approve, how to govern their use, and what happens when someone finds a shortcut we didn&#8217;t anticipate.</p><p>And I can tell you: shadow AI doesn&#8217;t start with rebellion. It starts with silence.</p><p>An associate finds that Claude drafts better first-pass research memos than starting from a blank screen. A paralegal discovers that summarizing long contracts is faster with a chat interface than with manual notes. A business services team member figures out that AI can write a decent first draft of an internal communication in a quarter of the time.</p><p>None of these people are trying to break rules. They&#8217;re trying to do their jobs. The organization just hasn&#8217;t caught up to the speed at which these tools became useful.</p><p>Most of the articles about shadow AI treat it like a leak to plug. Install monitoring software. Ban unsanctioned tools. Train people on the risks.</p><p>That framing has it backwards.</p><p>Shadow AI is what happens when your governance moves at committee speed and your employees move at tool speed. The gap between those two speeds is where unofficial use grows &#8212; not because people are reckless, but because the formal path either doesn&#8217;t exist or takes six months.</p><div><hr></div><h2>The Demand Signal</h2><p>After watching this unfold inside a conservative institution, one thing keeps standing out:</p><p><strong>The people using AI in secret are your early adopters.</strong> They already see the value. They&#8217;ve figured out where it fits. They&#8217;re exactly who you want on your side when it&#8217;s time to roll something out for real.</p><p>When you discover shadow AI, you have two options.</p><p>The first option is containment. Lock down access, monitor usage, send a compliance memo. This is the default in regulated industries, and it&#8217;s understandable. The risks are real &#8212; confidentiality, privilege, data residency. In legal, putting client information into an unapproved tool isn&#8217;t just a governance violation. It&#8217;s potentially a professional conduct issue. In February 2026, <a href="https://www.joneswalker.com/en/insights/blogs/ai-law-blog/your-ai-conversations-are-not-privileged-what-a-new-sdny-ruling-means-for-every.html">Judge Rakoff in </a><em><a href="https://www.joneswalker.com/en/insights/blogs/ai-law-blog/your-ai-conversations-are-not-privileged-what-a-new-sdny-ruling-means-for-every.html">United States v. Heppner</a></em> ruled that documents a defendant created using a consumer version of Claude were neither privileged nor work product &#8212; privilege was gone the moment he hit enter on a public AI tool. That&#8217;s the kind of consequence that makes CISOs lose sleep.</p><p>And this isn&#8217;t hypothetical exposure. <a href="https://www.clio.com/resources/legal-trends/read-online/">Clio&#8217;s 2025 Legal Trends Report</a> found that 79% of legal professionals now use AI, but more than half say their firm has no AI policy or they&#8217;re unaware of one. That gap between usage and governance is where the risk concentrates.</p><p>The second option is acceleration. Take the signal seriously. Ask: what are people actually using AI for? Where is the demand highest? What would it take to offer an approved path that&#8217;s fast enough that people actually use it instead of going around it?</p><p>In my experience, the second option works better. Not because the risks don&#8217;t matter &#8212; they absolutely do &#8212; but because containment without an alternative just pushes usage further underground. People don&#8217;t stop using the tool. They get better at hiding it.</p><div><hr></div><h2>What the Enablement Response Actually Looks Like</h2><p>When we discover unofficial AI use, the first conversation isn&#8217;t &#8220;you shouldn&#8217;t have done that.&#8221; It&#8217;s &#8220;tell me about your workflow.&#8221;</p><p>That conversation usually tells you more than any audit would.</p><p>The use case is almost always reasonable. People aren&#8217;t using AI to cut corners on judgment calls. They&#8217;re using it to speed up the mechanical parts &#8212; drafting, summarizing, organizing, reformatting. Work that takes time but doesn&#8217;t require expertise.</p><p>The risk is usually more contained than you&#8217;d expect. Someone using AI to draft an internal email isn&#8217;t creating the same exposure as someone pasting client financials into a public model. Understanding the actual risk surface matters more than treating all AI use as equally dangerous.</p><p>And the gap between what people want and what the organization provides is almost always about governance, not technology. We have access to enterprise AI tools. The problem is that the approval process, the usage policies, and the training haven&#8217;t kept pace with how fast people figured out the tools are useful.</p><p>Once you understand those three things, the path forward is pretty clear: fast-track the governance for the use cases people are already doing. Don&#8217;t build the perfect policy. Build the minimum viable policy that lets people work safely, and iterate from there.</p><div><hr></div><h2>The Uncomfortable Part</h2><p>There&#8217;s something else going on that the shadow AI conversation mostly avoids: the reason employees hide their AI use isn&#8217;t always about missing policies.</p><p>Sometimes it&#8217;s about culture.</p><p>Some people hide AI use because they think their manager will see it as cheating. Some worry they&#8217;ll look less skilled &#8212; that if the work is partially AI-generated, it somehow doesn&#8217;t count. Others are afraid that admitting they use AI to work faster just means they&#8217;ll be assigned more work at the same pay.</p><p>That&#8217;s a leadership problem, not a technology one.</p><p>If your culture treats AI use as a confession rather than a competency, you&#8217;ll get secrecy. If your managers don&#8217;t know how to evaluate AI-assisted work, they&#8217;ll default to suspicion. And if your organization&#8217;s implicit message is &#8220;we&#8217;ll adopt AI eventually, but not yet,&#8221; your employees will hear &#8220;figure it out yourself and don&#8217;t tell anyone.&#8221;</p><p>The organizations that handle this well share a common trait: they treat AI use as a skill to develop, not a shortcut to police.</p><div><hr></div><h2>So What Do You Do With This?</h2><p>Shadow AI isn&#8217;t a crisis to manage. It&#8217;s evidence that the transformation you&#8217;ve been planning is already happening &#8212; just without your involvement.</p><p>The question isn&#8217;t how to stop people from using AI. That ship sailed. The question is whether you can make the official path better than the workaround.</p><p>Because right now, in most organizations, the workaround is faster, easier, and more accessible than whatever the IT governance committee approved nine months ago.</p><p>If you&#8217;re responsible for AI adoption in any capacity, the existence of shadow AI should feel encouraging. The demand is real. The use cases are already proven. You don&#8217;t need to manufacture an adoption curve.</p><p>You just need to catch up to your own people.</p><div><hr></div><p><em>What does shadow AI look like in your organization &#8212; and is the response containment, acceleration, or something in between?</em></p>]]></content:encoded></item><item><title><![CDATA[The Three Horizons of AI Adoption in Law Firms]]></title><description><![CDATA[A practical framework for what to do immediately, what to sequence next, and what to decide before the year is out.]]></description><link>https://andrewlewis.ca/p/the-three-horizons-of-ai-adoption</link><guid isPermaLink="false">https://andrewlewis.ca/p/the-three-horizons-of-ai-adoption</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Thu, 26 Mar 2026 13:01:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!BUuT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Series: What Actually Mattered at LegalWeek 2026 &#8212; Part 6 of 6</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BUuT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BUuT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!BUuT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!BUuT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!BUuT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BUuT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7509342,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewiswashere.substack.com/i/191441038?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BUuT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!BUuT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!BUuT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!BUuT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2aaefdb-bcbe-42f7-812a-50ff86ae314d_2816x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I&#8217;ve spent the previous five articles in this series describing what I saw and heard at LegalWeek &#8212; the themes, the operating model shifts, the tactics that are working, and the hard conversations that aren&#8217;t happening yet.</p><p>This final piece is different. It&#8217;s the &#8220;so what&#8221; &#8212; the part where observation turns into action. I&#8217;ve organised it into three horizons based on urgency, complexity, and how much organisational consensus each move requires.</p><h2>Do Immediately</h2><p>These are moves that require minimal approval, no structural change, and can start generating signal within weeks.</p><p><strong>Shift AI training to task-based, simulated matters.</strong> If your training programme still starts with &#8220;here are the features of this tool,&#8221; stop. Identify the ten to twenty most repeatable legal tasks in your practice groups and build training around those. Use realistic simulated matters with time pressure and incomplete facts. Measure both time saved and &#8212; critically &#8212; confidence gained. The confidence metric is what tells you whether adoption will stick.</p><p><strong>Publish usage metrics by practice group.</strong> Take whatever AI tool usage data you have &#8212; Harvey, CoCounsel, whatever &#8212; and make it visible across the firm, broken down by practice group. Don&#8217;t attach mandates to it. Don&#8217;t set targets. Just publish the numbers and let the peer dynamics do the work. This is the lowest-cost, highest-signal adoption tactic I encountered at the entire conference.</p><p><strong>Stand up a practice group AI roundtable.</strong> One representative from each practice group. Each brings one real use case. No slides, no vendor demos, no keynotes. The constraint is what makes it work: &#8220;one real use case&#8221; forces specificity, keeps sessions short, and creates the social proof that training programmes can&#8217;t replicate.Do Next</p><p>These require more coordination and some cross-functional buy-in, but they&#8217;re achievable within a quarter.</p><p><strong>Formalise AI risk tiers and embed them in system prompts.</strong> If your AI policy is a standalone document that lives on an intranet page, it&#8217;s time to encode it. Define risk tiers for different use cases. Map approved tools to those tiers. Then embed the guardrails directly into the system-level prompts that shape how AI tools behave for your lawyers. Governance that&#8217;s invisible when done right is governance that actually works.</p><p><strong>Map shadow AI usage to unmet needs.</strong> Talk to your IT or security team about which AI sites lawyers are visiting outside the firm&#8217;s approved tools. Don&#8217;t approach this as a compliance exercise. Approach it as market research. The patterns will tell you exactly where the gap is between what you provide and what practitioners actually need. That&#8217;s your next pilot programme.</p><p><strong>Standardise vendor intake to reduce cycle time.</strong> Audit how long it takes a new AI tool to move from initial interest to production deployment. If the answer is measured in months, figure out where the bottleneck is &#8212; legal review, InfoSec, procurement &#8212; and build a repeatable, transparent process with target timelines for each stage. Speed of experimentation is a compounding advantage.Decide This Year</p><p>These are strategic decisions that require partnership-level conversation and real organisational commitment. They can&#8217;t be delegated to a working group or deferred to next year&#8217;s strategy cycle.</p><p><strong>Where specialist AI talent fits.</strong> If your AI adoption effort is staffed entirely by repurposed knowledge management professionals and a single innovation lead, you need to have an honest conversation about what&#8217;s missing. Builders &#8212; people with technical depth who can also navigate firm politics &#8212; are a different kind of hire. Decide where they sit, who they report to, and what success looks like for them. Enablement without builders stalls.</p><p><strong>How pricing reflects AI leverage.</strong> If AI is cutting the time on certain tasks by 60% or more, your pricing model is drifting out of alignment with the value you deliver and the efficiency you&#8217;ve gained. You may not need to overhaul your pricing framework this year, but you need to start the conversation &#8212; because clients are going to start it for you if you don&#8217;t.</p><p><strong>Whether compensation rewards AI leadership &#8212; or punishes it.</strong> This is the decision that reveals whether everything else is real. If your compensation framework doesn&#8217;t recognise AI adoption, innovation leadership, or efficiency gains, then those things aren&#8217;t strategic priorities. They&#8217;re aspirations. Your compensation plan is your strategic plan. If AI doesn&#8217;t show up there, it isn&#8217;t real.</p><div><hr></div><h2>The Series in Review</h2><p>Over these six articles, I&#8217;ve tried to describe what LegalWeek 2026 revealed about where law firm AI adoption actually is &#8212; not where the marketing says it is, but where the honest conversations are happening.</p><p>The tools are everywhere. Differentiation is gone. The constraint is now organisational: habits, trust, incentives, process, talent, and pricing. The firms that treat AI as a technology problem will keep stalling. The ones that treat it as an operating model problem will keep moving.</p><p>None of this is easy. But the path is clearer than it was a year ago, and the firms that take it seriously have a window to build a real advantage before the rest of the market catches up.</p><div><hr></div><p><em>Andrew is a Director of AI and Innovation at a large Canadian law firm. He writes about what AI adoption actually looks like from inside the institution.</em></p>]]></content:encoded></item><item><title><![CDATA[Your Compensation Plan Is Your Strategic Plan]]></title><description><![CDATA[Three conversations every firm is avoiding: the talent gap, the pricing misalignment, and the compensation question that reveals whether AI adoption is real or performative.]]></description><link>https://andrewlewis.ca/p/your-compensation-plan-is-your-strategic</link><guid isPermaLink="false">https://andrewlewis.ca/p/your-compensation-plan-is-your-strategic</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Wed, 25 Mar 2026 13:02:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Q1ts!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Series: What Actually Mattered at LegalWeek 2026 &#8212; Part 5 of 6</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q1ts!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q1ts!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!Q1ts!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!Q1ts!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!Q1ts!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q1ts!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7894995,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewiswashere.substack.com/i/191440851?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Q1ts!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!Q1ts!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!Q1ts!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!Q1ts!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa262300a-e9bd-4b61-aa4d-d607078eff25_2816x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>LegalWeek had plenty of optimism. Plenty of impressive demos. Plenty of panels about what&#8217;s working.</p><p>But the most useful conversations were the ones about what&#8217;s not working &#8212; and more specifically, about the structural problems that firms acknowledge privately and avoid publicly. Three stood out.</p><h2>The Talent Gap Nobody Wants to Name</h2><p>Most firms are now training lawyers on AI. That&#8217;s genuine progress. Two years ago, most weren&#8217;t.</p><p>But training lawyers to use AI tools is not the same thing as building the capability to push AI forward. Firms are investing in enablement &#8212; training, change management, communications &#8212; while underinvesting in the specialist talent needed to design, build, and evolve AI-powered workflows.</p><p>The pattern looks like this: you hire a Director of Innovation, or you expand the knowledge team, and you expect that small group to handle everything from tool evaluation to prompt engineering to strategic planning. It works for a while. Then it doesn&#8217;t, because enablement without builders stalls. You can train a thousand lawyers to use a tool, but somebody has to decide which tool to deploy, how to configure it, how to integrate it into existing systems, and how to iterate when the first version doesn&#8217;t work.</p><p>That &#8220;somebody&#8221; is a different kind of hire than most firms are comfortable making &#8212; someone with technical depth who can also navigate the political and cultural realities of a partnership. The talent market for those people is thin, and firms that wait too long to hire them will find that the gap between their ambitions and their capabilities keeps widening.</p><p>I feel this tension in my own work. The scope of what needs doing grows faster than the team. And the solution isn&#8217;t to work harder &#8212; it&#8217;s to build a team with the right mix of skills, which requires a hiring conversation that most firms haven&#8217;t had yet.Pricing Is Misaligned and Clients Will Notice First</p><p>Here&#8217;s the second hard conversation: AI is improving speed and leverage. In many cases, work that used to take forty hours now takes ten. That&#8217;s great for efficiency. It&#8217;s terrible for a business model built on billing hours.</p><p>The pricing models at most firms haven&#8217;t caught up. And here&#8217;s the uncomfortable part &#8212; clients are going to notice before firms do. A client who knows that AI-assisted contract review takes a fraction of the time it used to is going to ask why the invoice looks the same.</p><p>This isn&#8217;t a future problem. It&#8217;s a present one. The firms that address it proactively &#8212; by experimenting with fixed-fee arrangements, value-based pricing, or at minimum transparent disclosure of AI-assisted work &#8212; will build client trust. The ones that don&#8217;t will face the conversation on the client&#8217;s terms, which is always worse.</p><p>I don&#8217;t have a clean answer here. Nobody at LegalWeek did either. But the acknowledgment that this conversation is overdue was widespread, and the firms thinking about it seriously tend to be the same ones that are ahead on adoption more broadly.Compensation Is the Reveal</p><p>The third conversation is the one that separates firms that are serious about AI from firms that are performing seriousness.</p><p>Almost no firms &#8212; very few, maybe a handful &#8212; are aligning compensation with AI adoption, innovation leadership, or efficiency gains. Everyone agrees they should. Almost nobody has done it.</p><p>This matters more than it might seem, because compensation is the only signal that the partnership structure actually respects. You can issue all the strategy memos you want. You can create innovation committees and AI task forces and digital transformation programmes. But if the compensation framework doesn&#8217;t reward the behaviour you say you value, then the behaviour won&#8217;t change.</p><p>Your compensation plan is your strategic plan. That&#8217;s not a slogan. It&#8217;s a diagnostic. If AI adoption, innovation leadership, and efficiency gains don&#8217;t show up in how people are paid, then those things aren&#8217;t strategic priorities &#8212; they&#8217;re talking points.</p><p>The firm that figures this out first &#8212; that creates a credible, measurable link between AI-driven performance and partner compensation &#8212; will have an enormous talent and adoption advantage. Everyone else will keep having the same conversation at next year&#8217;s conference.</p><h2>The Common Thread</h2><p>All three of these problems &#8212; the talent gap, the pricing misalignment, and the compensation question &#8212; share a root cause. They require firms to change something structural about how they operate, not just add something new on top.</p><p>Adding a tool is easy. Rewriting a compensation framework is hard. Building a vendor intake pipeline is manageable. Rethinking pricing is existential. Hiring an enablement team is safe. Hiring builders is unfamiliar.</p><p>The firms that move on these hard conversations won&#8217;t do it because it&#8217;s comfortable. They&#8217;ll do it because the alternative &#8212; stalling out at the enablement stage while competitors build real capability &#8212; is worse.</p><div><hr></div><p><strong>Next in this series:</strong> A practical takeaway framework &#8212; what to do immediately, what to do next, and what to decide this year.</p><div><hr></div><p><em>Andrew is a Director of AI and Innovation at a large Canadian law firm. He writes about what AI adoption actually looks like from inside the institution.</em></p>]]></content:encoded></item><item><title><![CDATA[Just Sunlight]]></title><description><![CDATA[The adoption tactic nobody expected to work: making AI usage visible by practice group &#8212; and then stepping back.]]></description><link>https://andrewlewis.ca/p/just-sunlight</link><guid isPermaLink="false">https://andrewlewis.ca/p/just-sunlight</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Tue, 24 Mar 2026 13:02:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HvxW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Series: What Actually Mattered at LegalWeek 2026 &#8212; Part 4 of 6</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HvxW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HvxW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!HvxW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!HvxW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!HvxW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HvxW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:6650260,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewiswashere.substack.com/i/191440635?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HvxW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!HvxW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!HvxW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!HvxW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb5d4a99e-6662-4091-a9e6-2a23e8b319d9_2816x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There&#8217;s a persistent assumption in law firm leadership that AI adoption needs to be mandated. Pushed from the top. Measured by compliance. Enforced through policy.</p><p>LegalWeek offered a different model &#8212; one that&#8217;s simpler, cheaper, and apparently more effective.</p><p>Make usage visible. Then get out of the way.</p><h2>Transparency as Adoption Strategy</h2><p>Several firms described a version of the same tactic: they publish AI usage data &#8212; often Harvey adoption metrics, but sometimes broader tool engagement numbers &#8212; broken down by practice group.</p><p>That&#8217;s it. No mandate. No minimum usage requirement. No punitive measures for groups that lag behind. Just the data, visible to everyone.</p><p>What happens next is predictable if you&#8217;ve ever worked in a partnership. Competitive dynamics kick in. The corporate group sees that litigation is running at 40% adoption. The IP group sees that employment is outpacing them. Partners start asking their associates why the numbers look the way they do.</p><p>This creates three effects simultaneously: friendly competition, peer pressure, and organic knowledge sharing. The groups that are ahead start fielding questions from the groups that aren&#8217;t. The knowledge doesn&#8217;t flow through a central training programme &#8212; it flows through the firm&#8217;s natural social architecture.</p><p>I find this approach compelling because it respects how law firms actually work. Partnerships are peer-driven organisations. Top-down mandates create resistance. Visible data creates conversation. And conversation, in a partnership, is how things change.Practice Group Roundtables</p><p>The second tactic that came up repeatedly is even lower-ceremony: practice group AI roundtables.</p><p>The structure is minimal. You convene representatives from each practice group. Each group shares one real use case &#8212; something they&#8217;ve actually done with AI, not something theoretical. The group discusses what worked, what didn&#8217;t, and whether the use case is transferable.</p><p>That&#8217;s the whole format. No keynotes. No vendor demos. No slides. Just practitioners telling other practitioners what they tried and what happened.</p><p>The value here isn&#8217;t in any single use case. It&#8217;s in the normalisation. When a senior partner in real estate describes using AI to draft lease abstracts, that gives the cautious partner in tax permission to try something similar. Use cases are the social proof that no amount of training material can replicate.</p><p>I&#8217;ve been experimenting with a version of this in my own context, and the constraint that makes it work is the specificity requirement. &#8220;One real use case&#8221; forces people past the theoretical and into the concrete. It also keeps the sessions short, which means people actually show up.</p><h2>What&#8217;s Actually Happening Here</h2><p>Both of these tactics &#8212; usage transparency and roundtables &#8212; are doing something that top-down adoption programmes struggle with. They&#8217;re creating social conditions where using AI becomes the norm rather than the exception.</p><p>That&#8217;s a fundamentally different approach from the standard playbook. The standard playbook says: train people, give them access, and measure adoption. These tactics say: make the behaviour visible, create peer contexts where it&#8217;s discussed, and let the social dynamics of the firm do the rest.</p><p>It won&#8217;t work everywhere. It won&#8217;t work for every practice group. And it won&#8217;t replace the need for solid training, good tooling, and clear governance. But as a complement to those things &#8212; as the social layer that makes everything else stick &#8212; I haven&#8217;t seen anything more effective.</p><p>No mandate required. Just sunlight.</p><div><hr></div><p><strong>Next in this series:</strong> The hard conversations firms are avoiding &#8212; on talent, pricing, and the compensation question nobody wants to answer.</p><div><hr></div><p><em>Andrew is a Director of AI and Innovation at a large Canadian law firm. He writes about what AI adoption actually looks like from inside the institution.</em></p>]]></content:encoded></item><item><title><![CDATA[Shadow IT Is Telling You What to Build Next]]></title><description><![CDATA[The training and adoption tactics that are actually working &#8212; and why the best signal about what your firm needs is hiding in your web traffic logs.]]></description><link>https://andrewlewis.ca/p/shadow-it-is-telling-you-what-to</link><guid isPermaLink="false">https://andrewlewis.ca/p/shadow-it-is-telling-you-what-to</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Mon, 23 Mar 2026 13:01:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!F91E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Series: What Actually Mattered at LegalWeek 2026 &#8212; Part 3 of 6</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!F91E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!F91E!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!F91E!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!F91E!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!F91E!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!F91E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7176854,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewiswashere.substack.com/i/191440377?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!F91E!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!F91E!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!F91E!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!F91E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64f09896-0eff-4094-87c0-ce14b15cffd9_2816x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There&#8217;s a particular flavour of adoption advice that shows up at every legal technology conference. It sounds like: &#8220;You need executive sponsorship, a clear communications plan, and training resources.&#8221; It&#8217;s not wrong. It&#8217;s just not useful. Everyone already knows that. The question is <em>how</em> &#8212; and LegalWeek this year actually had some answers worth paying attention to.</p><h2>Start With the Task, Not the Tool</h2><p>The most consistently cited training model across sessions was what I&#8217;d call task-first AI training. The structure is straightforward, even if the execution is hard:</p><p>Identify the ten to twenty most repeatable legal tasks in your practice groups. These aren&#8217;t the edge cases. They&#8217;re the drafting, review, summarisation, and research patterns that eat up the majority of associate and junior partner time. Then build your training around those tasks &#8212; not around the AI tool&#8217;s feature set.</p><p>The practical version looks like this: you drop a lawyer into a simulated matter. Time pressure, incomplete facts, realistic constraints. The AI tool is available, but so is everything else. The goal isn&#8217;t to teach the tool. The goal is to measure whether the lawyer can produce better work, faster, when AI is part of their workflow.</p><p>Firms doing this are tracking two things: time saved and confidence gained. The second one matters more than people think. A lawyer who finishes a task 30% faster but doesn&#8217;t trust the output will stop using the tool within a month. Confidence is the leading indicator that sticks.The System Prompt Is the New Policy Document</p><p>Here&#8217;s something that wouldn&#8217;t have made sense two years ago: firms are embedding their AI policies directly into system prompts.</p><p>They&#8217;re formalising AI policies, defining risk tiers, listing approved use cases &#8212; and then encoding those rules into the system-level instructions that shape how AI tools behave for their lawyers. The result is governance that feels invisible when done right. Instead of a PDF that nobody reads, you get guardrails that are baked into the tool itself.</p><p>This is a meaningful evolution. It&#8217;s the difference between telling people the rules and engineering an environment where the rules are hard to break. I find this approach far more promising than the compliance-training model most firms default to, because it shifts the burden from the individual lawyer to the system design.</p><h2>Shadow IT Is Signal, Not Threat</h2><p>This was my favourite insight of the conference, and the one I think has the most untapped potential.</p><p>Firms are starting to track which AI sites their lawyers are visiting. Not to police them &#8212; to understand what problems those lawyers are trying to solve that the firm hasn&#8217;t enabled yet.</p><p>Think about what that data tells you. When a corporate associate is spending time on ChatGPT drafting contract summaries, that&#8217;s not a compliance failure. That&#8217;s a product requirement. It&#8217;s telling you exactly where the gap is between what the firm provides and what practitioners actually need.</p><p>The question to ask isn&#8217;t &#8220;how do we stop this?&#8221; It&#8217;s &#8220;what are people trying to do that we haven&#8217;t enabled?&#8221;</p><p>I&#8217;ve started looking at shadow AI usage in my own context through this lens, and it&#8217;s clarifying. The patterns don&#8217;t lie. When you see clusters of unsanctioned tool use around a particular task or practice area, you&#8217;ve found your next pilot programme.AI Communications Are Changing Format</p><p>One more tactical note that&#8217;s worth flagging: the way firms communicate about AI internally is shifting away from long-form memos and town halls toward short, consumable formats.</p><p>The firms getting traction are producing three-minute, reel-style videos and plain-language explainers. That tracks with everything we know about how busy professionals actually absorb information. The forty-five-minute webinar has its place, but it&#8217;s not how you move a two-thousand-person organisation.</p><p>There&#8217;s also a shift in <em>what</em> the communications are about. The focus is moving from internal capability (&#8220;here&#8217;s what our AI tool can do&#8221;) toward client readiness (&#8220;how many AI questions are clients asking, and are our lawyers prepared to answer them?&#8221;). That reframing matters. It connects AI adoption to revenue and client relationships, which is the language that moves partners.</p><div><hr></div><p><strong>Next in this series:</strong> How usage transparency, peer pressure, and practice group roundtables are doing more for adoption than any top-down mandate.</p><div><hr></div><p><em>Andrew is a Director of AI and Innovation at a large Canadian law firm. He writes about what AI adoption actually looks like from inside the institution.</em></p>]]></content:encoded></item><item><title><![CDATA[The Quiet Explosion of Knowledge Teams]]></title><description><![CDATA[Firms are doubling the size of their knowledge and innovation functions. Here's why that's the most important structural shift happening in legal right now.]]></description><link>https://andrewlewis.ca/p/the-quiet-explosion-of-knowledge</link><guid isPermaLink="false">https://andrewlewis.ca/p/the-quiet-explosion-of-knowledge</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Fri, 20 Mar 2026 13:02:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!i0Lv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Series: What Actually Mattered at LegalWeek 2026 &#8212; Part 2 of 6</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!i0Lv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!i0Lv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!i0Lv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!i0Lv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!i0Lv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!i0Lv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7443008,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewiswashere.substack.com/i/191440173?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!i0Lv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!i0Lv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!i0Lv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!i0Lv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aafa775-2ea4-4fec-8ab3-8dd9df8fe136_2816x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If you only follow the headline-grabbing AI announcements &#8212; the new tools, the vendor partnerships, the &#8220;we&#8217;re using AI to do X&#8221; press releases &#8212; you&#8217;d miss the most significant operational shift happening inside law firms right now.</p><p>Knowledge teams are exploding.</p><p>Not in the dramatic, &#8220;we hired a Chief AI Officer&#8221; way that makes the legal press. In the structural, quiet, &#8220;we&#8217;ve doubled headcount in our knowledge and innovation function&#8221; way that only becomes visible when you compare org charts twelve months apart.</p><h2>The Role Has Changed</h2><p>This isn&#8217;t just growth for growth&#8217;s sake. The role of these teams is fundamentally different from what it was even two years ago.</p><p>The old knowledge management function was a library. Precedent banks, template repositories, know-how databases. Important work, but bounded. The new function is closer to an operating system for the firm&#8217;s relationship with AI. It spans tool evaluation, prompt design, lawyer enablement, and quality control &#8212; and increasingly, it&#8217;s the function that determines whether any given AI initiative actually lands with practitioners or dies on the vine.</p><p>I&#8217;ve seen this shift from the inside, and the scale of it catches people off guard. Every new AI tool that enters the firm needs someone to evaluate it, pilot it, build the guardrails around it, design the training, measure the adoption, and troubleshoot when things don&#8217;t work as expected. Multiply that across a dozen tools and a few thousand lawyers, and you start to understand why these teams are growing.</p><p>The insight that resonated most at LegalWeek: AI doesn&#8217;t eliminate knowledge work. It raises the bar on it. The firms that understand this are investing accordingly. The ones that don&#8217;t are burning out small teams and wondering why adoption numbers stay flat.Vendor Onboarding Is Now a Strategic Bottleneck</p><p>Here&#8217;s a related shift that doesn&#8217;t get enough attention: the path from &#8220;we&#8217;d like to try this tool&#8221; to &#8220;it&#8217;s live in production&#8221; has become a full pipeline &#8212; and at most firms, that pipeline is slow.</p><p>The common stages are familiar to anyone who&#8217;s lived through enterprise procurement: legal review, information security assessment, procurement negotiation. None of these steps are unreasonable. All of them take time. And the cumulative effect is that firms without a fast, standardised intake process are falling behind &#8212; not because they&#8217;re taking on more risk than their competitors, but because they&#8217;re taking longer to make decisions.</p><p>This is one of those operational details that doesn&#8217;t make it into conference keynotes, but it determines outcomes more than most strategic decisions do. A firm that can evaluate, approve, and deploy a new AI tool in six weeks has a compounding advantage over one that takes six months. Not because the tool itself is transformative, but because the speed of experimentation becomes a capability in its own right.</p><p>I&#8217;ve been pushing on this in my own context &#8212; trying to turn what was an ad hoc, relationship-driven intake process into something repeatable and predictable. It&#8217;s not glamorous work. But every week you shave off the onboarding cycle is a week your lawyers get to start learning what works and what doesn&#8217;t.What This Actually Means</p><p>The structural message from LegalWeek was clear, even if nobody put it on a slide: the firms that are serious about AI are building internal infrastructure that looks nothing like a traditional law firm support function. They&#8217;re building product teams, design capabilities, enablement programmes, and operational pipelines.</p><p>This costs real money. It requires real headcount. And it demands that firm leadership treat AI adoption as an ongoing operational commitment, not a one-time capital expenditure.</p><p>That&#8217;s the part most firms haven&#8217;t fully internalised yet.</p><div><hr></div><p><strong>Next in this series:</strong> The specific training, communications, and adoption tactics that are actually working &#8212; from task-first training models to the surprising value of tracking shadow IT.</p><div><hr></div><p><em>Andrew is a Director of AI and Innovation at a large Canadian law firm. He writes about what AI adoption actually looks like from inside the institution.</em></p>]]></content:encoded></item><item><title><![CDATA[The Oven Doesn't Make the Restaurant]]></title><description><![CDATA[LegalWeek showed that every firm now has AI tools. Almost none have changed how they work.]]></description><link>https://andrewlewis.ca/p/the-oven-doesnt-make-the-restaurant</link><guid isPermaLink="false">https://andrewlewis.ca/p/the-oven-doesnt-make-the-restaurant</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Thu, 19 Mar 2026 13:03:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!BVDM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Series: What Actually Mattered at LegalWeek 2026 &#8212; Part 1 of 6</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BVDM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BVDM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!BVDM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!BVDM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!BVDM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BVDM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png" width="1456" height="794" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:794,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7555464,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewiswashere.substack.com/i/191439617?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BVDM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png 424w, https://substackcdn.com/image/fetch/$s_!BVDM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png 848w, https://substackcdn.com/image/fetch/$s_!BVDM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!BVDM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcff30ae2-da4a-451d-bf94-148be6c2a099_2816x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If you walked the expo floor at LegalWeek this year, you&#8217;d be forgiven for thinking the AI problem in legal is solved. Every booth had a demo. Every demo was impressive. Every vendor had a pitch about how their particular flavour of generative AI was going to reshape legal work.</p><p>And yet the most honest conversations &#8212; the ones in the hallways, the ones over coffee, the ones where people dropped the marketing voice &#8212; kept circling back to the same uncomfortable admission: the tools aren&#8217;t the bottleneck anymore.</p><p>The bottleneck is us.</p><h2>Everyone Has the Same Oven Now</h2><p>Here&#8217;s the analogy that kept surfacing in different forms across panels and roundtables: buying a better oven doesn&#8217;t make you a better restaurant. Training the kitchen does.</p><p>That lands differently depending on where you sit. If you&#8217;re a vendor, it&#8217;s a threat &#8212; because differentiation is collapsing. If you&#8217;re a firm leader, it&#8217;s a mirror &#8212; because the constraint on AI adoption isn&#8217;t the technology. It&#8217;s habits, trust, incentives, and whether the tool actually fits into the way people already work.</p><p>Firms that treat AI as a software rollout are stalling. I&#8217;ve watched this pattern from the inside. You announce a tool, run an onboarding webinar, send a few follow-up emails, and then wait for adoption numbers that never come. It feels like a deployment. It is not a deployment. It&#8217;s an organisational change problem &#8212; the kind where you need to rebuild workflows, shift expectations, and get very honest about what&#8217;s actually blocking people.</p><p>The firms that are moving aren&#8217;t the ones with the best tools. They&#8217;re the ones that treat adoption as a change management initiative with executive air cover, practice-group-level ownership, and a willingness to redesign process rather than bolt AI onto existing ones.The Hiring Shift Nobody&#8217;s Talking About Enough</p><p>There was a quieter theme running underneath the big panels that I think matters more than most of what made the main stage. Firms are changing how they evaluate talent &#8212; not just AI specialist talent, but lawyers.</p><p>The shift is from answers to questions.</p><p>The best hires, as several panellists put it, don&#8217;t claim mastery. They ask sharp, practical questions. They frame problems well. They spot risks that others walk past. They stay curious when the ground is uncertain.</p><p>This applies equally to the lawyer you&#8217;re hiring for your M&amp;A team and the AI strategist you&#8217;re hiring for your innovation group. The signal isn&#8217;t &#8220;I know how to use Harvey&#8221; &#8212; it&#8217;s &#8220;I understand what this tool can&#8217;t do, and I know when to stop trusting it.&#8221;</p><p>I&#8217;m told interview rubrics at several firms are being rewritten to weight three things more heavily: problem framing, risk spotting, and curiosity under uncertainty. That&#8217;s a meaningful shift. It says something about where firms think AI is actually heading &#8212; not toward a world where lawyers know less, but toward a world where the ability to ask the right question becomes the scarce skill.&#8220;Here Are All the Buttons&#8221; Is Dead</p><p>The third theme that kept recurring is the death of feature-based training. I don&#8217;t think this one is controversial anymore, but it&#8217;s worth saying clearly: if your AI training programme starts with &#8220;Here&#8217;s the interface,&#8221; it&#8217;s failing.</p><p>The model that&#8217;s working &#8212; the one that came up in almost every adoption-focused session &#8212; starts with a specific legal task. Not a tool walkthrough. Not a prompt engineering workshop. A realistic, time-pressured legal task that mirrors actual work.</p><p>One session described it this way: the goal is not to teach a lawyer how to use Harvey. The goal is to teach a lawyer how to draft a first-pass asset purchase agreement under time pressure, using whatever tools are available &#8212; including Harvey, but not limited to it.</p><p>That reframing matters. When you train to the task, the tool becomes incidental. When you train to the tool, you&#8217;ve built a dependency that breaks every time the vendor ships an update.</p><p>I&#8217;ve been thinking about this a lot in my own work. The task-based model isn&#8217;t just better pedagogy &#8212; it&#8217;s better strategy. It forces you to identify the twenty or so repeatable legal tasks that actually drive your practice, and then build training, tooling, and measurement around those. Everything else is noise.</p><div><hr></div><p><strong>Next in this series:</strong> How knowledge teams are quietly becoming the most important function in the modern law firm &#8212; and why vendor onboarding is now a strategic bottleneck.</p><div><hr></div><p><em>Andrew is a Director of AI and Innovation at a large Canadian law firm. He writes about what AI adoption actually looks like from inside the institution.</em></p>]]></content:encoded></item><item><title><![CDATA[Your Job Is Not a Job]]></title><description><![CDATA[Your job is not a job. It&#8217;s a collection of tasks. And AI doesn&#8217;t treat all of them the same way. Once you see it like that, everything changes.]]></description><link>https://andrewlewis.ca/p/your-job-is-not-a-job</link><guid isPermaLink="false">https://andrewlewis.ca/p/your-job-is-not-a-job</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Wed, 18 Mar 2026 22:01:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Bnqe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Bnqe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Bnqe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Bnqe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Bnqe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Bnqe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Bnqe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:911763,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewis.ca/i/191312399?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Bnqe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Bnqe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Bnqe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Bnqe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20ccf898-919e-4f64-97bb-cc0ce25fa0c2_1024x1024.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Everyone&#8217;s asking the wrong question right now. &#8220;Will AI take my job?&#8221; &#8220;Will AI replace engineers?&#8221; &#8220;Is marketing dead?&#8221;</p><p>That question is broken &#8212; because AI doesn&#8217;t see your job the way you do. You see a job title. AI sees a stack of individual tasks. It&#8217;s coming for some of them, it&#8217;s going to supercharge others, and there are some it can&#8217;t touch at all.</p><p>Today I&#8217;m going to walk you through a framework to figure out which is which &#8212; and give you a prompt you can run right now to get your personal AI career strategy in five minutes.</p><h2>Jobs are task collections</h2><p>Think about your actual workweek. Not your title &#8212; what you actually <em>do</em>, hour by hour.</p><p>If you&#8217;re a software engineer, some of your time goes to writing boilerplate code. Some goes to architecture decisions. Some goes to code reviews. Some goes to negotiating requirements with stakeholders.</p><p>Those are all different tasks. And they have wildly different relationships with AI.</p><p>It&#8217;s like looking at a toolbox. You wouldn&#8217;t throw out the whole thing just because one wrench is worn out. You&#8217;d swap the worn tools and keep the ones that still work.</p><p><strong>The wrong question:</strong> &#8220;Will AI replace software engineers?&#8221;</p><p><strong>The right question:</strong> &#8220;Which tasks inside software engineering are automatable &#8212; and which ones become more valuable?&#8221;</p><p>A job title is just a container. AI doesn&#8217;t replace containers. It transforms what&#8217;s inside them.</p><h2>The framework: Automate, Augment, Double Down</h2><p>Every task in your job falls into one of three categories.</p><h3>1. Automate</h3><p>These are tasks you hand off to AI entirely. It&#8217;s faster than you, the output is good enough, and your time is better spent somewhere else.</p><h3>2. Augment</h3><p>AI is your co-pilot here. You&#8217;re still driving, you&#8217;re still making the calls &#8212; but you&#8217;re moving two to five times faster than before.</p><h3>3. Double Down</h3><p>These are your moat. Pure human value. The tasks where no AI can replace what you bring to the table.</p><p><strong>Automation frees your time. Augmentation multiplies your output. Doubling down is how you become irreplaceable.</strong></p><h3>How to categorize your tasks</h3><p><strong>A task is automatable</strong> when it&#8217;s repetitive, pattern-based, and low-stakes. Status reports. Data entry. Boilerplate drafting. Here&#8217;s the gut check: if it bores you and a minor mistake wouldn&#8217;t cause a fire, hand it off.</p><p><strong>A task is augmentable</strong> when it requires your judgment, but the legwork leading up to that judgment is slow. Think legal analysis, market research, drafting proposals. AI gets you to 80% in 10% of the time. You bring the last 20% &#8212; the context, the taste, the strategic direction. It&#8217;s like having a research assistant who never sleeps but still needs you to make the final call.</p><p><strong>A task is a &#8220;double down&#8221;</strong> when it involves high-stakes judgment, deep relationships, or institutional context that no model has. The client dinner. The negotiation. The decision that requires knowing the politics and the people. These are the tasks people pay a premium for &#8212; and here&#8217;s the key &#8212; they become <em>more</em> valuable as everything else gets cheaper.</p><p>The goal is not to protect every task you do today. It&#8217;s to ruthlessly focus on the ones where you deliver irreplaceable value. Let the rest go.</p><h2>The AI Task Audit</h2><p>Here&#8217;s the process.</p><p><strong>Step 1</strong> &#8212; List every task that eats up your work week. Aim for at least ten. Be specific. Not &#8220;engineering&#8221; &#8212; break it down. &#8220;Writing unit tests.&#8221; &#8220;Reviewing pull requests.&#8221; &#8220;Attending sprint planning.&#8221;</p><p><strong>Step 2</strong> &#8212; Rate each task on AI automation potential. 1 means it absolutely requires a human. 10 means AI can do it today, no supervision needed.</p><p><strong>Step 3</strong> &#8212; Rate each task on a human edge score. 1 means it&#8217;s generic and commoditized &#8212; anyone could do it. 10 means it requires your irreplaceable judgment or relationships.</p><p><strong>Step 4</strong> &#8212; Categorize. Automate, augment, or double down.</p><p>You can absolutely do this manually with a spreadsheet. But I&#8217;ve also built a prompt that does the whole analysis for you.</p><h3>The prompt</h3><p>Paste the following into Claude, ChatGPT, or whichever AI assistant you prefer. Replace the bracketed sections with your details.</p><pre><code><code>You are a career strategist specializing in AI workforce transformation.

I'm going to give you my role, industry, and a list of tasks that make up my
typical work week. For each task, I want you to:

1. Rate it on AI Automation Potential (1&#8211;10, where 1 = absolutely requires a
   human and 10 = AI can do this today with no supervision)
2. Rate it on Human Edge Score (1&#8211;10, where 1 = generic/commoditized and
   10 = requires irreplaceable judgment, relationships, or context)
3. Categorize it as: AUTOMATE, AUGMENT, or DOUBLE DOWN

Then sort the table by automation potential (highest first) and give me a
three-sentence career strategy brief telling me where to focus my energy.

My role: [Your role]
My industry: [Your industry]
My tasks:
1. [Task 1]
2. [Task 2]
3. [Task 3]
...
</code></code></pre><p>Here&#8217;s what catches people off guard: the output is often surprising. Tasks you thought were safe show up as highly automatable. And tasks you&#8217;ve been rushing through &#8212; the ones you thought were low-value &#8212; turn out to be your highest-value work. That&#8217;s the whole point. It forces you to see your job the way AI sees it. Not through your title, but through your tasks.</p><h3>The shortcut</h3><p>If your AI assistant is already connected to your email and calendar &#8212; Copilot, Claude, whatever you&#8217;re using &#8212; it already knows what your tasks are.</p><p>Just ask it: <em>&#8220;Based on my last two weeks of meetings and emails, what tasks consumed most of my time?&#8221;</em></p><p>Then run the audit on that list. The data is already sitting there. You just need to ask the right question.</p><h2>Your homework</h2><p>Your career strategy is not about learning to use AI tools. Everyone&#8217;s going to learn the tools &#8212; that&#8217;s table stakes. Your strategy is knowing which of your tasks to hand to AI, and which ones to never let go.</p><p>Here&#8217;s what I want you to do this week:</p><ol><li><p><strong>Run the task audit.</strong> List your tasks, rate them, categorize them.</p></li><li><p><strong>Identify your top three &#8220;double down&#8221; tasks.</strong> That&#8217;s your career moat.</p></li><li><p><strong>Find two tasks you can automate immediately</strong> and reclaim that time.</p></li><li><p><strong>Redirect those freed-up hours into your highest-value work.</strong> That&#8217;s where you win.</p></li></ol><div><hr></div><p>Your job is not a job. It&#8217;s a collection of tasks. And the ones worth keeping are the ones only you can do.</p><p>Drop your role in the comments &#8212; I&#8217;m curious what tasks you&#8217;d put in the &#8220;double down&#8221; column.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://andrewlewis.ca/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Andrew's Substack! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Real AI Boom Hasn't Even Started Yet]]></title><description><![CDATA[Jevons' Paradox and the Cognitive Boom]]></description><link>https://andrewlewis.ca/p/the-real-ai-boom-hasnt-even-started</link><guid isPermaLink="false">https://andrewlewis.ca/p/the-real-ai-boom-hasnt-even-started</guid><dc:creator><![CDATA[Andrew Lewis]]></dc:creator><pubDate>Tue, 17 Mar 2026 03:31:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6ANd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6ANd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6ANd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!6ANd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!6ANd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!6ANd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6ANd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:888124,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://andrewlewis.ca/i/191213269?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6ANd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!6ANd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!6ANd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!6ANd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F22068595-7f68-45d9-8e55-0281e41c02bd_1024x1024.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In 1865, England figured out how to burn coal more efficiently. The natural assumption was that the country would use less coal. Instead, it used more. Way more.</p><p>That one fact explains why the dominant narrative about AI and jobs has a fatal flaw built into its foundation.</p><p>Right now, there&#8217;s a panic spreading through every office, every Slack channel, every LinkedIn feed. The fear is straightforward: AI makes knowledge work so efficient that companies won&#8217;t need as many people. Fewer engineers. Fewer analysts. Fewer lawyers.</p><p>But what if making work faster doesn&#8217;t shrink the workforce &#8212; it explodes it?</p><p>There&#8217;s a 160-year-old economic rule that predicts exactly what&#8217;s coming. And once you see it, you can&#8217;t unsee it.</p><h2>The Paradox</h2><p>Here&#8217;s the story. It&#8217;s 1865. A British economist named William Stanley Jevons notices something that makes no sense.</p><p>Steam engines have gotten dramatically more efficient. Everyone assumes England will use less coal. The math seems obvious &#8212; better engines, less fuel.</p><p>But the opposite happens. Because the engines are so efficient, it suddenly becomes cheap to use steam power for <em>everything</em>. Factories that couldn&#8217;t afford steam engines before? Now they can. Industries that never used steam? Now they do. Total coal consumption doesn&#8217;t drop. It skyrockets.</p><p>That&#8217;s the Jevons Paradox: <strong>increased efficiency doesn&#8217;t reduce demand. It creates it.</strong></p><p>This isn&#8217;t just a cute story about coal. We&#8217;ve already seen this exact pattern play out in your lifetime.</p><h2>The Spreadsheet Proof</h2><p>The 1980s. Electronic spreadsheets hit the market &#8212; VisiCalc, Lotus 1-2-3, eventually Excel. A single accountant could suddenly do in an hour what used to take a room full of clerks an entire week.</p><p>The panic was immediate. Computers are going to eat accounting.</p><p>But here&#8217;s what actually happened. Because complex math was suddenly so cheap and fast, companies didn&#8217;t fire their finance teams. They asked for <em>more math</em>. They wanted daily forecasts. Risk modeling. Scenario analysis. Deep analytics on every product line.</p><p>The number of financial analysts and accountants didn&#8217;t shrink. It multiplied.</p><p>The spreadsheet didn&#8217;t kill finance. It made math so cheap that companies wanted ten times more of it. The Jevons Paradox, playing out right in front of us.</p><h2>The Cognitive Boom</h2><p>Now let&#8217;s talk about right now.</p><p>Today, cognitive labor &#8212; coding, writing contracts, doing market research, designing campaigns &#8212; is expensive and slow. So companies only do what is strictly necessary. You only build software for mass markets. You only have lawyers review the most critical contracts. You only run marketing campaigns for your biggest audiences.</p><p>But what happens when AI makes a programmer ten times more efficient? The company doesn&#8217;t fire ninety percent of its engineers and build the same app. The cost of creating software plummets. And when the cost drops, the demand doesn&#8217;t stay flat.</p><p>It erupts.</p><p>Think about a law firm. Fifty lawyers. Today, they can only take on high-value cases because legal work is so labor-intensive. But with AI handling contract review, research, and first drafts? Suddenly, thousands of cases that were previously too small or too expensive to touch become viable. That firm doesn&#8217;t lay off forty lawyers. It hires two hundred more to manage the flood of newly accessible work.</p><p>The same pattern hits every field:</p><ul><li><p>Marketing campaigns become viable for niche audiences that were previously too small to justify the spend.</p></li><li><p>Custom software becomes economical for markets that couldn&#8217;t afford it before.</p></li><li><p>Legal services become accessible to small businesses for the first time.</p></li></ul><p>And in every single case, you still need humans to direct the AI, manage the projects, make the judgment calls, and ensure quality.</p><p>When the cost of intelligence drops, we don&#8217;t consume the same amount of intelligence with fewer people. We consume radically more intelligence. We start tackling projects that were too expensive or too complex to even attempt.</p><p>We don&#8217;t shrink the economy. We scale it up.</p><h2>Your Move</h2><p>This is not the automation of existing work. This is the creation of entirely new categories of work that were previously too expensive to exist.</p><p>So here&#8217;s the exercise worth doing: look at your current role. What work does your company skip because it&#8217;s too expensive or too slow? That skipped work is your expansion zone. That&#8217;s where the new jobs, the new teams, the new opportunities are going to come from.</p><p>Don&#8217;t position yourself as the person AI replaces. Position yourself as the person who directs AI into that latent demand.</p><p>The professionals who thrive in this next decade won&#8217;t be the ones who fear the efficiency. They&#8217;ll be the ones who see the rebound.</p><p>AI isn&#8217;t the end of white-collar work. It&#8217;s the beginning of the cognitive boom.</p><div><hr></div><p><em>What work does your company currently skip because it&#8217;s too expensive? That&#8217;s your Jevons Paradox moment waiting to happen. I&#8217;d love to hear it in the comments.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://andrewlewis.ca/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Andrew's Substack! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>