<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[centaurprise]]></title><description><![CDATA[design better human-agent teams]]></description><link>https://www.centaurprise.com</link><image><url>https://substackcdn.com/image/fetch/$s_!vaou!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe98d5d5-c480-4926-a05e-da918a911bd9_1024x1024.png</url><title>centaurprise</title><link>https://www.centaurprise.com</link></image><generator>Substack</generator><lastBuildDate>Sat, 18 Apr 2026 06:20:16 GMT</lastBuildDate><atom:link href="https://www.centaurprise.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[DJ Thompson]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[centaurprise@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[centaurprise@substack.com]]></itunes:email><itunes:name><![CDATA[DJ Thompson]]></itunes:name></itunes:owner><itunes:author><![CDATA[DJ Thompson]]></itunes:author><googleplay:owner><![CDATA[centaurprise@substack.com]]></googleplay:owner><googleplay:email><![CDATA[centaurprise@substack.com]]></googleplay:email><googleplay:author><![CDATA[DJ Thompson]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The rise of motivated software]]></title><description><![CDATA[Software isn't dead, but opinionated software might be soon. What comes next?]]></description><link>https://www.centaurprise.com/p/the-rise-of-motivated-software</link><guid isPermaLink="false">https://www.centaurprise.com/p/the-rise-of-motivated-software</guid><dc:creator><![CDATA[DJ Thompson]]></dc:creator><pubDate>Wed, 10 Sep 2025 03:25:39 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/75faf761-6d77-4ab5-b621-416c08c48935_2400x1714.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2><strong>The chat trap</strong></h2><p>You've heard it, I&#8217;ve heard it: the growing sentiment that "software is dead."</p><p>Why build apps when ChatGPT can do everything? Why design interfaces when you can just type what you want? AI is eating software, just like software ate the world.</p><p>But if software is dead, why are people still opening dozens of apps every day? Why do companies still pay hundreds of thousands of dollars for CRMs when an LLM with an MCP for a database would save so much spend? Why hasn't everyone switched to ChatGPT for everything?</p><p>The answer is simple. AI is powerful, but talking to a computer isn't always the best way to get things done. AI has a UX problem<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><p>What&#8217;s messed up is we&#8217;ve learned this lesson before. In 2017, big tech companies bet everything on chat. Apple and Facebook and X n&#233;e Twitter all looked at WeChat in China, where you can order food, pay bills, and book flights all through messaging, and thought "this is the future!" They built chatbots for everything.</p><p>It just didn&#8217;t translate.</p><p>Why? Because when everything is possible, nothing is top-of-mind. Chat gives you infinite options, which means infinite surface area to test and infinite ways to get confused. Most chatbots became what Laura Burkhauser at Descript called "phone trees in trench coats"<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. It&#8217;s the same frustrating "Press 1 for billing" experience, just dressed up in a chat bubble.</p><p>Now in 2025, we're making the same mistake again. But this time, we can learn from it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading centaurprise! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>Software is still hanging in there</strong></h2><p>The statistics on AI use are in flux, constantly. The MIT study that 95% of enterprise use has no ROI is going around like crazy<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. It&#8217;s exacerbated by the fact that, when you look outside the US, 75% of knowledge workers use AI tools at work, but only 7-22% (depending on country) are using work-provided tools<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. 40+% consistently are bringing tools from home instead, using their personal ChatGPTs instead of the Copilots and Bedrocks afforded them. Even still, only 44% of users (58% of the group using AI at work) say they&#8217;re &#8220;heavily-reliant&#8221; on AI while at work<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>.</p><p>What&#8217;s going on with the numbers? Well, inconsistent definitions of what it means to &#8220;use AI&#8221; and &#8220;get value,&#8221; for one. Strong differences between usage patterns at early adopter companies and non-adopter companies, for another. Older companies are slotting agents into existing bureaucracy, with amusing but low-ROI results<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>. </p><p>Different intent behind the adoption has an effect as well: some companies feel existential pressure to become &#8220;AI-native&#8221; (though they define that differently too), while others are checking a box from a board mandate, and that box does not include validating user stickiness over time.</p><p>And the user stickiness story in AI is one that isn&#8217;t being told much right now. AI companies are throwing down stellar (annualized) ARR and revenue-per-employee numbers, but lurking behind those numbers are skyrocketing churn metrics. Most demos look pretty sharp. The problem is what happens after the demo.</p><p>Here's an example: a friend who runs a proserv firm tried switching his team to an AI project management tool driven mostly by chat UX. The rep promises "it can do everything!" Naturally. And technically, that was true: you could badger the software into doing what you wanted, by chatting with it, for almost 100% of the feature portfolio. But after two weeks, his team was back on their old software.</p><p>"Every time someone wanted to check a deadline," he told me, "they had to type a question, wait for the response, and hope the AI understood what they meant. Burning tokens the whole time. We used to just&#8230; click on the calendar button.&#8221; One second versus thirty seconds, fifty times a day, every day, every week<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>.</p><p>This is the gap between potential and practice. AI can do amazing things, but if those things take longer than the old way, people don't switch.</p><p>Those who do: well, they&#8217;re there for the experiment, and they have no issue churning after a month if the value isn&#8217;t there. In fact, that&#8217;s the plan from the get-go. Some people on the leading edge of the &#8220;crossing-the-chasm&#8221; diagram aren&#8217;t early adopters, they&#8217;re just taste-testers.</p><h2><strong>The three ages of software</strong></h2><p>So, if software&#8217;s clearly changing but not-so-clearly surviving, where does that leave us? Evolving.</p><p>To understand what we're evolving into, let's look at where we've been. Software has advanced through three major ages, each fixing the problems of the last while creating new ones.</p><h3><strong>Age 1: Software That Just (Barely) Does Something (Sometimes)</strong></h3><p>In the beginning, there was <strong>embedded software</strong>, which came welded to hardware. Buy a DEC system, get DEC software. It did one thing, hopefully. When it broke - and it always broke - you called expensive consultants to fix it.</p><p>Users had one request: please don't crash. The bar wasn't high because there was no bar. Having any digital system at all felt like magic.</p><h3><strong>Age 2: Software That Does Everything (If You Can Figure It Out)</strong></h3><p>Then came the revolution: <strong>unopinionated software</strong>, freed from the shackles of predetermined hardware. Office. Lotus. Adobe. These platforms promised ultimate flexibility. Build anything! On any hardware! Customize everything! If you could dream it, you could (&#8230;probably) configure it.</p><p>But there was a catch. With great power came great confusion. It's like being handed a full set of professional chef's knives when all you wanted was to make a sandwich. Some people created masterpieces. Most people nicked themselves.</p><p>This era made consultants rich. Not fixing broken software anymore, but teaching people how to use it. Every company had that one person who "really knew Excel"&#8212;the keeper of mysterious formulas and secret shortcuts. Cottage industries like &#8220;Salesforce developers&#8221; cropped up, necessary experts in handling the arcane inner workings of systems designed to be so complex that they could theoretically do anything, given enough recursion (and enough budget).</p><h3><strong>Age 3: Software That Does What Its Founder Wants (Hope That&#8217;s What You Want Too)</strong></h3><p>Smart companies noticed something: most users were trying to do the same few things. What if, instead of infinite options, we only gave them the best way<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>?</p><p>Enter <strong>opinionated software</strong>. Box decided how folders should be organized. Slack decided how teams should communicate. Notion decided how documents should be structured. These tools didn't just provide features&#8212;they taught you a philosophy.</p><p>This worked brilliantly if you agreed with their philosophy. But what if you didn't? What if your team had spent years perfecting a different approach? Tough luck. As Stewart Butterfield put it: "There's no worthwhile software that doesn't involve behavior change."</p><p>The consultants didn't disappear. They just changed their pitch from "Let me teach you the features" to "Let me change your company culture."<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a></p><h2><strong>The fourth age - software that seeks your goals</strong></h2><p>Now, you could argue that with the rise of vibe coding, software from all three previous ages is basically commoditized. AI can copy any of it, at any time, right?<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a></p><p>But I posit we're approaching something new. Software that learns what you're trying to achieve and helps you get there. Not by forcing you into its workflow, but by molding itself to yours.</p><p>I call it <strong>motivated software</strong>.</p><p>Imagine you're planning a product launch. Today's software structures your UI and data, makes you think in terms of tasks, deadlines, and assignees. You create tickets, set dates, assign people. The software tracks whether Task A is done, but it doesn't understand that Task A only matters if it helps Product B succeed. (Sometimes, in the case of opinionated software, even if Product B succeeds the software will still complain that Task A didn&#8217;t get done in just the right ideological way).</p><p>Motivated software works differently. It has your same high-level goal: successful product launch. It takes inputs from connected systems: KPIs, documents, updates, system analytics. It notices that sales enablement for infrastructure products always takes longer than scheduled, so it quietly adjusts future timelines for impacted launches. It tracks that marketing needs extra lead time when engineering runs late, so it alerts them early: marketing doesn&#8217;t have to depend on someone from engineering manually reminding them each time. It detects that Sarah's qualification calls are thorough but slow, while Mike's are fast but sometimes miss edge cases, and it routes customer inquiries about the new product accordingly.</p><p>This isn't just automation. It&#8217;s open-ended pattern recognition, and it&#8217;s adaptive optimization<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a>. The software develops a model of what you're trying to achieve and constantly searches for better paths to get there, building and tweaking its reward models constantly to stay in-line with yours.</p><p>Think of it like this:</p><ul><li><p>Embedded software was a tool that, well, existed</p></li><li><p>Unopinionated software did what you told it, for better or worse</p></li><li><p>Opinionated software was a teacher that told you what to do</p></li><li><p>Motivated software will be a partner that helps you succeed</p></li></ul><h3><strong>Isn&#8217;t that just an agent?</strong></h3><p>For people up to speed on the AI ecosystem, the most obvious question about this philosophy is: why can&#8217;t an agent just do this too?</p><p>To be clear, agents plural are absolutely part of this infrastructure. Motivated software takes advances in LLMs, reasoning models, tool use, and agent-to-agent protocols and meshes them together in a logical way that would be obnoxious to do manually, every time, for every use case.</p><p>Individual agents lack three things that a parent motivated software platform brings to the table:</p><ol><li><p><strong>Dimensionality</strong> - heavy users of agents know that unlocking their best performance comes from giving them small, granular, well-scoped tasks. You could theoretically ask an agent &#8220;fix my company&#8217;s budget&#8221; but - you&#8217;d be much better off (in both token cost and outcomes) deploying and orchestrating lots of smaller targeted agents, each handling specific tasks (&#8220;audit our FP&amp;A last quarter,&#8221; &#8220;review last week&#8217;s fraud reports for irregularities,&#8221; &#8220;design a net-payment terms strategy to improve our cash-on-hand&#8221;). Motivated software coordinates this effort and selects which tasks are most relevant, when. Think of agents as a point-vector in this system: one gust of wind, rolling up into motivated software&#8217;s weather front.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cmJk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cmJk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png 424w, https://substackcdn.com/image/fetch/$s_!cmJk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png 848w, https://substackcdn.com/image/fetch/$s_!cmJk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png 1272w, https://substackcdn.com/image/fetch/$s_!cmJk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cmJk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png" width="1456" height="1040" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1040,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:522024,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.centaurprise.com/i/172230156?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cmJk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png 424w, https://substackcdn.com/image/fetch/$s_!cmJk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png 848w, https://substackcdn.com/image/fetch/$s_!cmJk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png 1272w, https://substackcdn.com/image/fetch/$s_!cmJk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0b2d0a5b-c55c-41be-a625-d43e8e529339_2400x1714.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>The best visual I have for this concept. If each vector is an agent, the entire system is the motivated software platform. Similar to how one person could theoretically do all the work of an entire company, one agent could traverse this entire space over time&#8230; but not easily and not quickly.</em></figcaption></figure></div></li></ol><ol start="2"><li><p><strong>Persistence</strong> - it&#8217;s no secret agents run into context window issues. Modern agents typically operate with around 100-200k token context windows. More are ramping up to 1M tokens in practice (and labs have up to 10M tokens in window in experimental models). So why is persistence an issue? Even with larger and larger context windows (and you have to assume that LLM builders will continue to innovate here) the cost of one agent holding on to the details of one conversation gets prohibitive, quickly. If you need nuanced details on every customer in your GTM strategy for each incremental agent call, you either need to have every single call expend increasing numbers of tokens (or risk compacting away useful details), or you need a way to retain and sort through context, feeding it to the right agent at the right time. The human could manually copy-paste context in each time, but wow - what a waste of human potential! - and what a great way to introduce suboptimal context calls and room for error.</p></li></ol><ol start="3"><li><p><strong>Hyperparameters</strong> - agents don&#8217;t dynamically improve themselves or think critically about when to call themselves under which conditions. We as users <em>can</em> do that, but there&#8217;s an upper bound on how much we want to be thinking about &#8220;the problem of how to optimize an agent to solve our problems,&#8221; versus &#8220;our problems themselves.&#8221; Motivated software abstracts this away - we can program in the criteria for selecting the right agents, system instructions, and parameters, or better yet: we can have agents optimize this over time, creating meta-optimization layers that create recursive customization and sophistication that  boosts how closely the software hews to our needs and preferences<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a>.</p></li></ol><h2><strong>Why this matters</strong></h2><p>Here's why motivated software is different from just adding AI to existing apps:</p><p><strong>It remembers</strong>. Not just your data, but your context. Not just from this conversation, but longitudinally: it knows where to pick up the thread where the last agent dropped it, without burning billions of tokens to do so. It knows that when you say "the Johnson project," you mean the proposal for the client you met last Tuesday, not the internal project run by the engineer who happens to have the same name.</p><p><strong>It learns your patterns. </strong>Sometimes, you know you&#8217;re stuck repeating a mistake, but you don&#8217;t know how to fix it. Sometimes, you don&#8217;t notice you&#8217;re repeating your mistakes until someone catches you, often way too late. Motivated software can nudge us out of ruts: watching you reschedule the weekly standup three times because West Coast teammates can't make the morning slot, it suggests a better time that works for everyone. Seeing you struggle to convert cold emails, it proffers more customized openers and blocks time on your calendar to protect research. Tracking stress signals in your meeting transcripts and messaging apps, it starts pushing late-night meetings and calls to help you recover and avoid burnout.</p><p><strong>It connects intentions to actions.</strong> When you message "We need to speed up delivery," it doesn't just record the comment. It identifies bottlenecks in your workflow, suggests process improvements, and tracks whether changes actually improve delivery speed. Motivated software is fundamentally about ingesting your goals and reinforcing behaviors and processes that advance them, tuning all these RNNs around us to our advantage.</p><p><strong>It incentivizes models to ask for help.</strong> Current AI confidently gives wrong answers because saying &#8220;I don&#8217;t know&#8221; in a chat window 50% of the time doesn&#8217;t feel great. Motivated software needs to get away from that, to understand and convey its confidence level because being confidently-incorrect regresses you from your goal. Motivated software calibrates AI answers against certainty over time. When uncertain, it asks clarifying questions or brings in human judgment.</p><p>Most importantly, <strong>it pulls agents out of the chat window and into the underlying primitives.</strong> It&#8217;s not possible to get the level of performance you need from motivated software by tacking on a chat window - simply asking a bot to give an answer or run a tool does too little, too slowly to have the software form-fit the customer and seek after their goals. LLMs have to be in the infrastructure, the libraries, the core utilities in order for this work.</p><p>This is hard to build. Really hard. Current AI technology isn't quite there yet. We need:</p><ul><li><p>AI that can accurately understand goals, not just commands</p></li><li><p>Memory systems that maintain context over months, not minutes</p></li><li><p>Context uptake that costs thousands of tokens, not millions</p></li><li><p>Confidence calibration so the system knows what it doesn't know</p></li><li><p>Search capabilities that explore solutions without getting stuck</p></li><li><p>Feedback loops that ensure the software improves, not just changes<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a></p></li></ul><h2><strong>A whole new world</strong></h2><p>Even just three to five years from now, this could make day-to-day workflows look really different.</p><p>Let&#8217;s say, in 2030, you run a growing sales team. Your software isn't just tracking deals; it's trying to help you succeed. It notices you're adding more reps but deal velocity isn't increasing proportionally. It identifies that deal approval has become a bottleneck and suggests a solution: parallel approval tracks based on deal size and risk profile.</p><p>You're skeptical: do you need that much process for your team size? But you give it a shot. The software spins out a custom agent, adjusts your approval workflows, and monitors the experiment, measuring not just approval speed but discount rates, legal review quality, and rep satisfaction.</p><p>After two weeks, it reports: 31% faster approvals, discount rates held steady, slight improvement in contract quality scores. It brings receipts: line-item audit trail of each deal by track, links to updated pipeline metrics, confidence scores against each routing decision it made, flags where human judgment might have differed (there's always room for interpretation). It asks: "What do we keep? What do we change? Do we codify this?"</p><p>This isn't science fiction. It's the logical next step from where we are today. The building blocks exist:</p><ul><li><p>Large language models that understand context</p></li><li><p>Pattern recognition that spots trends</p></li><li><p>Optimization algorithms that search solution spaces</p></li><li><p>Feedback systems that learn from results</p></li></ul><p>What's missing is the connective tissue&#8212;the infrastructure that lets these pieces work together seamlessly.</p><h2><strong>Deep dynamic customization as a moat</strong></h2><p>The companies that figure this out won't just have better software. They'll have sustainable competitive advantages that are nearly impossible to copy.</p><p>Why? Because motivated software's value isn't in its features. You can screenshot the UI and dump a feature list into Lovable, but the entire value prop is in the time-series bespoke context you&#8217;ve built up user-by-user, tenant-by-tenant: you can&#8217;t clone that. The value builds over time as the system learns the unique patterns of your organization.</p><p>It's like the difference between a new EA and one who's worked with you for years. They could have gone to the same school, worked for the same firms, but the experience is totally different. The tenured assistant doesn't just follow instructions: they anticipate needs, prevent problems, and make connections you might miss. That knowledge can't be transferred instantly to someone new.</p><p>The same protection applies to motivated software. A competitor can copy your interface, match your features, even poach your engineers. But they can't copy the accumulated understanding of how your specific organization works best.</p><h2><strong>Motivating your software</strong></h2><p>Building motivated software requires rethinking our entire approach to software development:</p><p><strong>Start with goals, not features. </strong>Instead of asking "What should this button do?", ask "What is the user trying to achieve?" We express this as a meta-layer in &#8220;user journeys&#8221; now, but motivated software implants this directly into UI/UX and adds abstraction <em>on top of that</em>: goals are themselves composable and manipulatable.</p><p><strong>Design for learning, not just usage.</strong> Build systems that get better over time, not just systems that work on day one. We compartmentalize this type of instability into our DS/ML teams in most engineering orgs today: motivated software will broadly assume that frontend, backend, data engineering, infrastructure, security, and analytics are all in a much greater state of flux, and in different states at different customers. Understanding what we&#8217;ve built will become its own form of challenge.</p><p><strong>Embrace uncertainty.</strong> Create software that knows when it doesn't know and can gracefully ask for guidance. More to the point: create software that <em>is guaranteed to not be in its optimal state, or even know its own optimal state, at the point of initialization</em>.</p><p><strong>Measure outcomes, not activities.</strong> Track whether users achieve their goals, not just whether they click the buttons. Metrics like intercept rate, derivatives, and precision will matter more than days active and clickthrough rates: &#8220;facts&#8221; will get necessarily more complex, more recursive, and require more storytelling to understand and follow over time.</p><p>This is hard. It requires new technical capabilities, new design patterns, and new ways of thinking about software. But the payoff is enormous: software that doesn't just serve users but partners with them.</p><h2><strong>The future is motivated</strong></h2><p>Software isn't dead, it's evolving.</p><p>The chat interface revolution failed not because AI isn't powerful, but because power without direction is meaningless, donuts in a parking lot. Motivated software provides direction by understanding what users actually want to achieve.</p><p>This is a fundamental shift in how we think about tools. For the first time, our software can grasp our intent and seek out our goals, not just execute our commands.</p><p>The companies that understand this shift and build software motivated by user success rather than feature completeness won't just win in the market. They'll define entirely new categories of what software can be, and they&#8217;ll leave previous generations in the dust.</p><p>The age of motivated software is coming. The question isn't whether it will transform how we work. The question is: who will build it first?</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.centaurprise.com/subscribe?"><span>Subscribe now</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>There&#8217;s a long thread of complaints to this effect - [<a href="https://uxdesign.cc/ai-has-a-ux-problem-c1d0352003b0">1</a>] [<a href="https://koomen.dev/essays/horseless-carriages/">2</a>] [<a href="https://thenanyu.com/ux.html">3</a>] [<a href="https://www.uxtigers.com/post/classic-usability-ai">4</a>] - just to name a few.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Love that line. Don&#8217;t know if this is the origin of the quip, but <a href="https://www.producthunt.com/p/descript/underlord-by-descript-2">this</a> is where I saw it.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The core commentary on the Project NANDA report is pretty neatly summed up <a href="https://fortune.com/2025/08/21/an-mit-report-that-95-of-ai-pilots-fail-spooked-investors-but-the-reason-why-those-pilots-failed-is-what-should-make-the-c-suite-anxious/">here</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p><a href="https://fortune.com/2024/12/02/ai-us-llama-chatgpt-india">It&#8217;s an older source</a>, but there&#8217;s not much comparable data cross-border from recent months.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p><a href="https://resources.owllabs.com/blog/pulse-survey-2025">Different studies</a>, so the numbers aren&#8217;t quite apples-to-apples. Take that with a grain of salt, but the baseline of companies using AI (67% vs 75%) is fairly close. </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>See <a href="https://www.centaurprise.com/p/making-agents-work?r=6rwfn">Making agents work</a>. There&#8217;s no good reason to throw compute (and therefore cash) at a task that would be waste if a human did it. (There <em>is</em> good reason to throw compute at friction).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>At Slack we used to call these &#8220;papercuts:&#8221; small UX irritants that feel awful in aggregate when you use that particular software all the time (but don&#8217;t matter if you only need it once a month or once a quarter).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>The title of this article is a callback to Stuart Eccles&#8217; article from a decade ago, <a href="https://medium.com/@stueccles/the-rise-of-opinionated-software-ca1ba0140d5b">The rise of opinionated software</a>. Aaron Levie and <a href="https://basecamp.com/gettingreal/04.6-make-opinionated-software">DHH</a> are other notable advocates. </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>Ever notice how the concept of &#8220;digital transformation&#8221; survives all three eras? Wonder why that is.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>There&#8217;s a couple notable arguments in the opposite direction [<a href="https://www.linkedin.com/pulse/opinionated-software-making-comeback-rohan-sandeep-5bwpc/">1</a>] [<a href="https://contraryresearch.substack.com/p/ai-makes-opinionated-software-more">2</a>] [<a href="https://www.workingtheorys.com/p/taste-is-eating-silicon-valley">3</a>] that bear calling out. But there&#8217;s a core issue: once you encode your opinion in source, nothing stops a competitor from identifying it and prompting an AI to replicate it. Try asking an agent to build a Linear or Superhuman or Replit clone. Their opinions aren&#8217;t moats. Their taste isn&#8217;t a moat. </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>It&#8217;s not &#8220;understanding&#8221; but it could feel like it on the user&#8217;s end. Anthropomorphization is not a good habit to get into when dealing with LLMs - it&#8217;s the root of like 90% of bad takes in the AI space.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>I think this is where I distinguish these &#8220;eras&#8221; from Andrej Karpathy&#8217;s &#8220;<a href="https://www.youtube.com/watch?v=LCEmiRjPEtQ">Software 1.0 / 2.0 / 3.0,</a>&#8221; which is less about how the software interacts with users and more about how the developer interacts with the codebase.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>These may sound like givens, but there are <a href="https://natesnewsletter.substack.com/p/software-30-vs-ai-agentic-mesh-why">a slew of agentic experiments </a>that resulted in fragile, expensive failures that came nowhere near their stated objectives. If Anthropic and Cognition can&#8217;t &#8220;wing it&#8221; and succeed, odds are you and I can&#8217;t either.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Notes from Inbound]]></title><description><![CDATA[Key themes from Hubspot's first SF conference]]></description><link>https://www.centaurprise.com/p/notes-from-inbound</link><guid isPermaLink="false">https://www.centaurprise.com/p/notes-from-inbound</guid><dc:creator><![CDATA[DJ Thompson]]></dc:creator><pubDate>Fri, 05 Sep 2025 22:47:46 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/195473cf-21a1-4b5a-91f9-2ad60fba0048_2000x1334.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Most years, Hubspot runs their annual conference INBOUND in Boston, close to their headquarters. This year, they dropped a small flood of orange paint into every transit station in San Francisco and decided to run it in Salesforce&#8217;s backyard<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><p>Some key themes coming out of the discussions, especially among execs:</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading centaurprise! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>Agent web traffic is growing really fast for some</strong></h2><p>Some businesses are getting consistently 25% of their inbound traffic from deep research (think Perplexity, ChatGPT Web Search, etc). One person I spoke to said their ratio regularly spiked to 75%. I think that&#8217;s a sampling bias for sure - our crowd is much more likely to have AI folks deep-researching them than say a neighborhood deli - but still, the total amounts swung around fast.</p><h2><strong>AII and AIX aren&#8217;t anywhere near best-practice</strong></h2><p>I&#8217;ve said it before: I find it funny that it&#8217;s taken us 50+ years to optimize for UI/UX, and we&#8217;re still not consistently great at it, but when we bring LLMs into the mix we expect them to just get whatever browser or app or outdated GUI we throw at them. Calling an agent interface &#8220;AI&#8221; feels deliberately obtuse, so I&#8217;ve taken to labeling these &#8220;AII/AIX&#8221; (AI Interface / AI Experience) as the counterpoint to UI/UX, but I don&#8217;t know how universal that is.</p><p>And survey says we&#8217;re far from being good at AII/AIX right now. Most agents have a lot of trouble parsing anything that isn&#8217;t pure HTML, but we&#8217;re not optimizing for those workflows. Lots of us have watched Browser Use or Operator struggle through complex UIs and RPA-type motions<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. What are we supposed to be doing about it on the other end?</p><p>Plenty of buzz from GEO agencies, new companies like Parallel or /dev/agents, though the best idea I heard all week was setting up agent-specific microsites, pointing them out in llms.txt, and keeping them sparse, HTML-only, and/or using chatbot interfaces to try to get the inbound agent to stop clicking and just say what it wants. One person said that experiment popped their success rate up from the mid-20%s to the mid-70%s (and the conversations told them more about what the searcher was asking for - more on that in a sec).</p><h2><strong>The tooling for agent web traffic is still nascent</strong></h2><p>Unlike searches on major traditional engines, identifying the search people were making via agent is tricky, and backing into how common that search is more of a guessing game than you&#8217;d like. People suggested screen replays and some vendors are starting to offer deciphering this as a service, but it&#8217;s much more opaque than before.</p><p>Revisit that solution around agent microsites. If you have sites that pop up specifically for agents (sit them on robots.txt/llms.txt, don&#8217;t provide (easy) UI access for users, promote them in GEO, and create a bunch that are very topical to the use cases you think folks are searching for) then your screen replays suddenly get much more useful. </p><p>Even more so if you can get chat execution against whatever your call-to-action is: two LLMs talking often make up things about whether they&#8217;re allowed to sell to each other and what the terms are, but the error rates are still much lower than best-in-class browser use in-the-wild. Providing carrots for agents to sort themselves makes it easier to figure out what the people behind the curtain really want.</p><h2><strong>AI web crawlers are making attribution harder</strong></h2><p>If you follow independent-web spaces, you know LLM-builder web crawlers are <em>extremely</em> unpopular right now. They&#8217;re not great at adhering to robots.txt or llms.txt, they can bombard websites relentlessly, and - for anyone who makes their living off of ads or creative work - they just represent a philosophical threat to occupation or revenue stream that most AI companies aren&#8217;t responding to right now<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>.</p><p>Anyways, it&#8217;s not just a small website owner issue. If you&#8217;re trying to sort through whether the OpenAI hits you&#8217;re getting are scraping for training, indexing for GEO, or targeted access for Agent&#8230; well, that&#8217;s tricky, and the fact that the first one is so high-volume makes it hard to optimize the experience for the latter two. Folks are talking a lot about how to sort traffic (and deal with stubborn crawlers that don&#8217;t respond to being politely sorted). The attitude isn&#8217;t nearly as exasperated as say HackerNews, but it&#8217;s clear the major agentic-search players haven&#8217;t cultivated the ecosystem yet.</p><h2><strong>GEO as a trust vector for ABM</strong></h2><p>ABM and large enterprise has historically not really cared much about SEO or website design. Take <a href="http://scale.com">Scale</a>&#8217;s website for instance: unless you&#8217;re pretty-well steeped in how AI is built, it&#8217;s not particularly easy to figure out exactly what Scale is going to sell you, especially for their particularly high-end generative AI product lines. (Their Rapid self-serve lines are a bit easier to suss out).</p><p>That&#8217;s not a dig on Scale, and it&#8217;s not an accident. It&#8217;s by design, and it&#8217;s common for most companies targeting customers of a large-enough size. You might have <em>intent </em>to buy, but Scale&#8217;s guessing that if you stumbled onto the site via Google search, you don&#8217;t really have <em>capability</em> to buy. </p><p>The customers who <em>are</em> capable are arriving via different channels that support higher trust for the buyer: referrals, previous contracts, existing partners and channels and integrations. Picking out any random data vendor via search engine is not a safe bet for the future of your RLHF or eval program: if you&#8217;re spending that level of money, you&#8217;d choose a different process.</p><p>GEO (&#8220;Generative Engine Optimization,&#8221; one of like five competing terms for describing how to push your company to the top of a Deep Research set of results) has an interesting vector for changing that. As you build up a context base and personalization set with AI (assuming you can hold onto it over time, which is a big assumption) you can actually create better trust with the results than you had with your search bar<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. </p><p>Of course, if attribution is still really hard, then large enterprise vendors aren&#8217;t going to be particularly incentivized to optimize for GEO still, just because of all the false positives. To get the confusion matrix in the right spot, they need to get some sort of &#8220;agentic handshake&#8221; going: pretty broad agreement that if that existed in a trusted form it would cut out a <em>lot</em> of wasted qualification time up front.</p><p>That&#8217;s a big hurdle to get over first: AI still isn&#8217;t particularly trusted right now. Hallucination and confabulation are still quite real, and e.g. &#8220;Gemini said ClickUp was better for us than Front&#8221; is not going to be convincing to an angry exec if it turns out that was the wrong call. AI doesn&#8217;t get fired, people do.</p><p>Still, digitizing trust signals around enterprise procurement has been an interest area for a lot of people, for a very long time. Expect to see some people at least try their hand here.</p><h2>Three types of builders</h2><p>Wrapping up with an observation on presentations companies made over the course of the week. There&#8217;s broadly three groups of people based on the solutions they were implementing:</p><ol><li><p>People who were automating away their headcount</p></li><li><p>People who were taking bureaucracy and waste processes and automating those to &#8220;give people time back&#8221;</p></li><li><p>People who were optimizing for what humans do best, what agents do best, and how to wring every last lead and dollar out of each hour and token they spent</p></li></ol><p>Group 3 had the weirdest designs but the best growth, best engagement, best flywheel effects. You could see the technical folks and AI experts gravitate to those solutions; they sparked a lot more imagination and follow-on conversations. Group 3 also didn&#8217;t use much jargon compared to the other two: it felt like they had less to hide with their results.</p><p>It&#8217;s still early, but it&#8217;s clear there&#8217;s a bit of a shift in how people are approaching these solutions. For the better.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.centaurprise.com/subscribe?"><span>Subscribe now</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Ok,<a href="https://developers.hubspot.com/blog/the-developers-guide-to-inbound25"> it&#8217;s supposed to be</a> because they want to strategically realign with innovation around AI. But a month before Dreamforce&#8230; quite a few folks I spoke to are amused about the timing.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Just today I tasked GPT-5 Agent with going through an online book and copying over the chapters into a linked Google Drive - should have been a simple exercise. When I timed it, it took me ~15 seconds to do one, not counting parallel-tasking. Agent gave up after 25 minutes having done 7 (8, if you count the duplicates) and having needed 5 separate interventions to get going again. Some of that is justified, some of that is agents just not being that far along yet, but are we really saying the best way for agents to interact with GDrive is via the same drop-down menus I have access to?</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>If you&#8217;re worried about losing your job or your livelihood, comments about &#8220;feeling the AGI&#8221; or &#8220;sustaining 20% cyclical unemployment&#8221; aren&#8217;t just unresponsive, they&#8217;re probably doing real reputational harm. Leaders here aren&#8217;t learning the lessons from the botched globalization rollouts: the same people who bought the <em>West Wing</em>-style &#8220;free trade is good &#8230; free trade creates jobs&#8221; mantras aren&#8217;t going to get fooled again so easily. They want a plan.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Assuming those results don&#8217;t get intermediated by ad buys, which is the reason the search bar stopped being trusted in the first place. It&#8217;s <em>really</em> hard to control the ad auction in such a way that it doesn&#8217;t ruin your credibility at the top of the ABM food chain.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Making agents work]]></title><description><![CDATA[Applying Work, Friction, and Waste to our human-AI allocation // officially retiring the DeejDrone 9000 &#128546;]]></description><link>https://www.centaurprise.com/p/making-agents-work</link><guid isPermaLink="false">https://www.centaurprise.com/p/making-agents-work</guid><dc:creator><![CDATA[DJ Thompson]]></dc:creator><pubDate>Thu, 04 Sep 2025 01:22:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/cbb9ed00-6d6a-483a-8775-bcfd4dcddda3_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is the last post in a three-part series on which parts of work are valuable. It started with creating a ridiculous biodiesel drone to show how AI can generate perfect-looking garbage, then updating the Work-Friction-Waste framework to spot which activities actually matter.</p><p>Now, let's put it all together to figure out how to actually use agents without destroying what makes your team valuable.</p><p>If agents can't replace PMs, what can they do? Lots, as it turns out. Just not what you'd expect from following the hype train.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading centaurprise! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>The Work of PMs</h2><p>So, we know how Work works, and we know the PRD isn&#8217;t the Work, but the Friction, for the PM. What does that mean for the VP who wants to assign agents as PMs?</p><p>Good news and bad news. Bad news first: you still need PMs.</p><p>The danger of replacing PMs with agents - at least in this scenario - is threefold:</p><ol><li><p>The agent making PRDs is not doing the Work. It&#8217;s handling the Friction (which - can be a good thing! But more on that in a sec).</p></li><li><p>The person prompting the agent is imposing all of their biases onto the agent, and disguising those biases as good ideas with rigor and research. (Like I did with the DeejDrone 9000). The key insight here is that <em>none of my ideas were any better after they were put into the PRD than they were before:</em> I didn&#8217;t do any Work to validate anything. But now that they&#8217;re in PRD format, it&#8217;s way harder to tell that they&#8217;re terrible. Friction that doesn&#8217;t carry any Work is Waste, by definition.</p></li><li><p>To top it all off, because we&#8217;ve replaced all our PMs with agents, <em>the only people who can see through my PRDs and tell that they don&#8217;t have any merit are the people we let go.</em> We lost our BS detectors, so we can&#8217;t tell when we&#8217;re lying to ourselves anymore<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p></li></ol><p>But let&#8217;s say I know PRDs aren&#8217;t the end goal of a PM. Can I train an agent to be a PM if I open up the scope?</p><p>Again, probably not, not yet. Agents aren&#8217;t great at the Work PMs do. It takes some savvy to handle discovering the Voice of the Customer and reading between the lines, and humans tend to be more effusive and expressive in front of other humans. (Ever &#8220;interviewed yourself&#8221; for a job with one of those automated video systems, or trialed one you were considering using? You&#8217;re not likely to &#8220;warm up to the interviewer&#8221; during the process; people still like talking to other people).</p><p>You need the LLM to have a thesis, interrogate the thesis, and subtly test their thesis while also looking out for &#8220;unknown unknowns&#8221; during these conversations&#8230; but the state-of-the-art realtime voice call agents these days tend to be just a little better than call center phone trees.</p><h2>PRHs and Product Gap Agents</h2><p>Good news: agents can still help, a lot. There&#8217;s a ton of ways to throw agentic automations against Waste and Friction in the average Product org, but for this conversation, we followed the PRD thread specifically.</p><p>First: low-hanging fruit. PM comes back from their customer interviews with notes, dumps them into an LLM, and the LLM summarizes them into PRD format. That&#8217;s not agentic behavior by any means, but it&#8217;s still pulling Friction out of the PRD process, assuming you can fact-check in less time than writing the document outright.</p><p>But let&#8217;s get more creative and find more value. Start from the point where your PMs go talk to customers and work backwards. The <em>output</em> of all of that is going to be a PRD; but is there a utility to having a PRD as an <em>input </em>too?</p><p>I could see it - for the sake of argument, let&#8217;s call it a PRH - a Product Requirements Hypothesis:</p><ol><li><p>Try this out: take the exact same format, but fill it in based on your initial best guess - what do you think the right answers are going to be? Generate a PRD as though you knew the right answer based on your &#8220;day zero&#8221; knowledge - using an agent to research across the enterprise&#8217;s data sources - and then react to it. Do the details and assumptions and logical conclusions match what you expected?</p></li><li><p>Iterate on this and expose what would make your hypothesis wrong <em>explicitly in the document</em>. This step is key: otherwise you&#8217;re just biasing yourself, poisoning your own well. The end state for a PRH is <em>not a draft PRD</em> - you&#8217;re gonna redo all of this - but an experimental survey design that explains exactly what you think the options are for each element of the product design, the sensitivity and impacts of each decision change, and the things you want to listen for and explore in conversations.</p></li><li><p>The final piece of the PRH is the <em>customer inventory</em> - the conversations that you <em>want</em> to have that could get at the level of detail you&#8217;re looking for. Again, an LLM can help you run through CRMs or CX datasets to check for specific info, or at least give you guesses that you can react to and reject or edit. You&#8217;re looking to get more prescriptive on who to talk to - rather than &#8220;whoever my CSMs proffer up&#8221; - so you can test really specific elements of your hypothesis (or, also interestingly, find out that the things you expected to find aren&#8217;t as easily-discoverable as you hoped)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>.</p></li><li><p>This would result in - in my experience - <em>way more</em> prep for customer conversations than most PMs walk in the door with<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. If you&#8217;re sitting there testing for both high-level questions and a bunch of little details that you wouldn&#8217;t have thought of but could materially affect the end picture, the agent&#8217;s done a lot of actual Work helping you prime yourself to extract Customer Voice from what could be otherwise less-structured conversations<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>.</p></li></ol><p>Ok - then let&#8217;s go one step earlier. How do I prime a PRH? Again, I really don&#8217;t want to set up a system where I take my innate biases about the product or the business and polish them up so that no one else can tell that they&#8217;re hollow and backed by nothing. I&#8217;d love to set up chains of events that feed me good triggers and data to inform my hypotheses: my agents should be setting me up for success.</p><p>The #1 place that this breaks down in most customer orgs is the product gaps process. Product gaps reporting is often sketchy at best, even among companies that design products for reporting product gaps! The feeds of information are inconsistent, the identity deduplication is rare, the supporting data is scarce, prioritization is done by squeakiest wheel, and at a certain length of backlog no automated gap feed is actually entering the working memory of any team member until a long-overdue-hygiene process goes through and flushes months-old cards off the board.</p><p>But an agent, operating in the right environment with the right framework, could do a couple interesting things:</p><ol><li><p>They could (obviously) feed and organize new product gaps into the right streams at the right times. That&#8217;s easy. That&#8217;s not even agent work - that&#8217;s API + WHERE statements, maybe embeds to make the sorting less fragile.</p></li><li><p>They could reference existing projects or PRHs and ask requestors or gap filers targeted questions at the time of filing - what was the source of the bug? Where does the enterprise customer want this API to feed into? You filed 12 P0 gaps - if I can only do 3 this quarter, which ones would you vote for<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>?</p></li><li><p>They could build and maintain profiles on the requestors - both the internal parties, and the actual end requestors (the customers, the Twitter personalities, the vendors, the partners) to build context over time, so that gaps get richer over time. Lots of folks have pined wistfully for the opportunity to do this over my career, basically no one has the time to do it. Prime agent work - stuff humans can only conspire but to dream of because they have actual shit to do.</p></li><li><p>When it comes time for customer interviews or CABs, agents can shortlist the requestors (filtering for things like red accounts and churn) closing the loop on the information pull<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>.</p></li></ol><p>My prompt to the VP at this point - how do you go one step before this? How do you use agents to prime a product gap process that feeds you great information and returns great sentiment? I won&#8217;t go into the answers here: I think the point&#8217;s been made and the ideas here start getting very tactical and very real. It&#8217;s much more exciting and useful than <em>I, PMBot</em>: the energy in the room is different because we&#8217;re targeting real problems with precise solutions instead of employing unguided AI for its own sake.</p><h2>Bias for action</h2><p>I want to close on a note my cofounder Deji brought up when we were discussing this concept. He mentioned his initial gut reaction: doesn&#8217;t this push back against conventional startup wisdom to just go do <em>something</em>?</p><p>There&#8217;s the classic cliche &#8220;bias for action.&#8221; Brian Armstrong famously put this as &#8220;action generates information&#8221; - when you&#8217;re pre-PMF, taking action is the best way to learn, even if you&#8217;re not sure it&#8217;s the right action<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>.</p><p>I like that. I don&#8217;t think the frameworks clash. Instead, I&#8217;d say the two mesh with one simple phrase: don&#8217;t confuse activity for action. Or, if I were to paraphrase Brian&#8217;s quote into our framework: &#8220;All Work generates information: any Work, even the wrong Work. Waste does not.&#8221;</p><p>It&#8217;s easy to run around looking busy, doing productivity theater, making yourself feel busy, doing everything but what matters most. That&#8217;s common for larger businesses layering on process for the sake of process; that&#8217;s equally common for small startups procrastinating and doing everything but getting in the van, talking to customers, and building what they want. If you spend all your time as a startup on Waste, you&#8217;re not taking real action, and you&#8217;re not getting real information.</p><p>The only difference for pre- and post-PMF companies is that, before you know your customers <em>want</em> to pay you for your Work, you just have to take your best guess. It&#8217;s still critical to minimize things that aren&#8217;t Work or supporting it - maybe more so - because you need to try so many different types of Work until you find the right one.</p><p>Work is the highest revenue-bearing activity. Work is also the highest information-bearing activity. Friction is your Work delivery vehicle.</p><p>Everything else is just in your way.</p><h2>Framing exposes opportunities</h2><p>Is the point of all this that you should turn your Product team onto making PRHs now? No. It&#8217;s an experiment, we ran it once, the jury&#8217;s still out on if it&#8217;s worth the effort or not. Feel free to play with it and let me know if it fits with how you build.</p><p>The actual point is: see the difference better framing makes? By lining up agents to destroy Waste, improve Work outputs, and cycle out Friction to improve throughput, we get a way different Product org that makes the PMs feel like they&#8217;re omniscient. The alternative - having agents make fake PRDs - doesn&#8217;t feel remotely as cool in comparison.</p><p>It's intellectually easy to swap humans for AI 1-for-1. It feels tempting to take all your Waste and task an LLM to do it for you. But I guarantee it's worse. There are better problems to target, and you can spot them using this framework. Tap your AI to eliminate Waste, minimize Friction, and improve your real Work.</p><p>Every company using AI is about to learn this lesson. The question is whether they'll learn it the easy way or the hard way.</p><p>You&#8217;ll be able to tell by how many DeejDrones they launch.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.centaurprise.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p><em>Coming off of Inbound, there have been a lot of cool conversations about how people in Sales &amp; Marketing roles are using agents productively - and not - and I&#8217;ll try to bring this back to earth a little bit and share what I&#8217;m hearing starting tomorrow. </em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>We talked about hypocognition in an earlier post - this is sort of hypocognition-by-design, and it&#8217;s pretty dangerous.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>We had a saying when I was in the Air Force - &#8220;negative intelligence is still intelligence.&#8221; Basically - don&#8217;t discount when you find out that something you wanted to learn <em>cannot</em> be learned.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p><a href="https://www.atlassian.com/software/confluence/templates/customer-interview-report">Example of a simpler customer interview template</a>. Nothing&#8217;s <em>wrong</em> with it, but there&#8217;s nothing setting you up to challenge exactly what you need to evaluate to get your product right; it&#8217;s mostly a blank piece of paper, ideal if you have <em>no idea</em> what your next product looks like, but the more you know about what you want to test, the weaker a template like this is. You can fill a template this generic out, exactly as written, without testing a single hypothesis. A PRH done well might guardrail you away from that outcome.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Here&#8217;s the draft PRH we made based on the DeejDrone concept: <a href="https://github.com/centaurprise/deej-drone-9000/blob/main/DD9K_PRH.md">C003A2 - DD9K PRH</a>. Again - the drone&#8217;s still a bad idea. But the template now is designed to expose that in customer interviews and get us detailed information about the viability of going into the commercial drone industry at all, and where the nuances are. This template filled in, plus the interview notes, turned into an actual PRD, will be infinitely better than the result at the beginning of the series.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>This is, by the way, the exact reason I dislike the P# system and why we don&#8217;t use it.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Your product gaps process may be wildly different than these folks, but we set up an agent with fairly straightforward system instructions to run off a Slack webhook and feed disparate gap reports back to the Product team with some basic enrichment, which was enough to get the wheels turning on other long-neglected problems: https://github.com/centaurprise/deej-drone-9000/tree/main/agents </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>The original interview is here: </p><div id="youtube2-Yv98vPLXdzk" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;Yv98vPLXdzk&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/Yv98vPLXdzk?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div></div></div>]]></content:encoded></item><item><title><![CDATA[Which work matters]]></title><description><![CDATA[A simple framework for matching resources to activities]]></description><link>https://www.centaurprise.com/p/which-work-matters</link><guid isPermaLink="false">https://www.centaurprise.com/p/which-work-matters</guid><dc:creator><![CDATA[DJ Thompson]]></dc:creator><pubDate>Sun, 31 Aug 2025 23:39:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Icv7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>The marginal utility of what I&#8217;m doing right now</h2><p><a href="https://www.centaurprise.com/p/how-to-miss-the-point-of-work">Last time</a>, we prompted an AI to generate a beautiful PRD for a biodiesel drone that should never exist. The document was lovely. The idea was terrible. And while it was easy to tell the difference this time, it&#8217;s also easy to see how toning down the prompt just a bit would make it much trickier to spot the vapidness of the output.</p><p>That's the trap: when artifacts look right, we assume the thinking behind them is right too, especially if we weren&#8217;t there to vet the process. But in this case, the artifact wasn&#8217;t the work, it was just evidence that work happened. Or, was supposed to have happened.</p><p>Today, I want to share the framework that helps me spot the difference between value-add actions and expensive theater. It's helped teams figure out where AI actually belongs in their stack. More importantly, it's helped them stop wasting humans on tasks that don't matter.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading centaurprise! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Work, Friction, and Waste</h2><p>So, how can you tell the PRD isn&#8217;t the part &#8220;that counts?&#8221;</p><p>There&#8217;s a simple framework for understanding the impact of the action you&#8217;re taking or the artifact you&#8217;re building: Work, Friction, &amp; Waste. You might recognize the terms from an intro physics course, or from similar concepts from Jobs-to-be-Done<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> (Work, Work-of-Work, and Waste), or Lean and Six Sigma (Work, Motion, and Waste)<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. The concepts are similar, but to the uninitiated, the word choice of &#8220;Friction&#8221; is, in my opinion, way more intuitive.</p><p>Depending where you first heard them, the definitions taught to you for each of these items could have been quite complex. I personally like keeping them extremely simple:</p><ul><li><p><strong>WORK</strong> &#8594; What customers pay you for</p></li><li><p><strong>FRICTION</strong> &#8594; What allows you to get paid for Work</p></li><li><p><strong>WASTE</strong> &#8594; Everything else</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Icv7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Icv7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png 424w, https://substackcdn.com/image/fetch/$s_!Icv7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png 848w, https://substackcdn.com/image/fetch/$s_!Icv7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png 1272w, https://substackcdn.com/image/fetch/$s_!Icv7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Icv7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png" width="1361" height="997" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:997,&quot;width&quot;:1361,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Icv7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png 424w, https://substackcdn.com/image/fetch/$s_!Icv7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png 848w, https://substackcdn.com/image/fetch/$s_!Icv7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png 1272w, https://substackcdn.com/image/fetch/$s_!Icv7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F912458d5-a7ee-459a-a46f-5d0a4915ac8d_1361x997.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>My beautifully-rendered diagram of which kinds of activities are which.</em></figcaption></figure></div><p>I&#8217;ve spent a lot of time thinking through and boiling down this framework - first picking up elements as a consultant, then really building out and testing the theory for modern work at Slack, and now at <a href="https://gigue.ai">gigue</a> adapting it to how humans should be employing AI. Since I suspect most folks haven&#8217;t put nearly that much thought into it, I&#8217;ll take a sec to walk through where I&#8217;ve gotten with it:</p><ul><li><p>What is the definition of each activity?</p></li><li><p>How does it look in practice? (we&#8217;ll use a &#8220;Mutually-Aligned Action Plan&#8221; (MAAP) example to help with the nuances)</p></li><li><p>How should I invest resources in this activity?</p></li></ul><h3>Work</h3><p>Work is the most important of the three activities. There is a simple brightline for what counts as Work and what does not: Work is something the customer would pay you for<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>.</p><p>Be <em>very</em> careful and <em>very</em> strict with this definition: it&#8217;s tempting to say &#8220;the customer won&#8217;t pay us unless we have professional marketing copy&#8221; or &#8220;the customer won&#8217;t pay us unless we build this deck for the MAAP call&#8221; - sure, but they won&#8217;t pay you <em>for the copy / deck itself</em>, which changes how you handle the time and effort you invest into those assets<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>.</p><p>That <em>does not mean </em>that only Product is Work, or only Sales is Work. Customers will pay for SOC2 compliance; customers will pay for net-60 payment terms; customers will pay for reliable on-time delivery; customers will pay for &lt;5 min response time on service channels with &gt;97% SLA during business hours.</p><p>This is not an excuse to be elitist or to partition the functions of the business into haves and have-nots. It&#8217;s a strategic razor for <em>activity</em>: for <em>everyone</em>, whether they roll up to Product or Sales or Legal or Finance or Ops or CX, <strong>will the energy they spend on this particular activity return revenue</strong> <strong>or not?</strong></p><p>Because Work is something the customer would pay you for, there is generally good reason to improve it at any given time<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>. Not every dimension a customer will pay for has high ROI - clever product design is still a thing - but at a super-high level &#8220;improving the Work&#8221; is a good use of resources provided you can tactically figure out the right way to do so.</p><h3>Friction</h3><p>Friction is activity that is necessary to effect the trade of Work for money with the customer. That definition is broader than the Work definition (Friction is not just e.g. finserv APIs and delivery couriers - it is anything that <em>has to happen for the exchange of value to complete</em>). The activity of building the MAAP deck from above? Friction<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>.</p><p>Think of the classic physics example of pushing a box on the ground (the one I used in my diagram). The effort you spend pushing can be split into two parts: the effort that moves the box along the straight line, and the effort spent overcoming the drag where the box touches the ground. You can&#8217;t <em>not</em> spend the effort on Friction: if you decline to spend it, the box doesn&#8217;t move and the Work doesn&#8217;t get done.</p><p>But, unlike your undergrad physics homework, you control the terms of engagement. You don&#8217;t have to choose scenarios with high Friction, and you definitely don&#8217;t want to <em>optimize </em>for high Friction<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>. Instead, you want to optimize for Work-over-Friction - how can you maximize the Work you get done and minimize the Friction it takes?</p><p>Take the MAAP example - let&#8217;s say the Work is a SaaS product, but the customer is also in part paying for security vis-a-vis an ELA with legal protections, payment terms instead of credit card transactions, and implementation help we&#8217;re setting up on this action plan. Some parts of this process, then, are Work, and some aren&#8217;t. How can we minimize Friction?</p><ul><li><p>We could skip the Mutually-Aligned Action Plan step altogether, but that actually materially impacts the odds we get a contract signed, and definitely impacts the odds that contract renews (because it reduces the value the customer gets from the contract, so it becomes something the customer would pay for - just not how you&#8217;d traditionally think of it). Ok, so Work is happening in that step, we don&#8217;t want to throw out the baby with the bathwater here.</p></li><li><p>Do we need the deck to be built custom for every single customer, every single time? Probably not - there&#8217;s probably a good MAAP template that needs small tweaks each time, and we probably get better results that way. Friction pulled out.</p></li><li><p>Do we need to align the MAAP with the customer? Definitely. It&#8217;s in the name. How much back-and-forth do we need? When do we pull the champion in? Can we ghost the doc, prep a live tracker, pre-stage the artifacts we know they&#8217;re going to want, hook the doc into our conversation to get dynamic updates? Every bit of Friction we&#8217;re pulling out is helping, and every time we improve actual alignment and plan adherence is doing actual Work - it&#8217;s the service the customer is (about to be) paying us for.</p></li><li><p>Do we need meetings for all this? A champion meeting? A draft session? An exec sponsor review? Weekly check-ins? Again, we&#8217;re solving for maximizing the alignment of the teams against how much effort it takes to get said alignment. Meeting-phobia vs technophobia is a team preference question, in my experience&#8230; but you can get pretty clever on firing updates proactively, early, and eagerly to short-circuit the perception that misalignment has happened or could happen, which pushes the Work / Friction balance decidedly in your favor.</p></li></ul><p>It&#8217;s often tempting to do <em>the exact opposite of this approach</em> in practice. Mutually-aligning with a high-value customer is important! and you can talk yourself into spending <em>hours or days</em> on fine-tuning the deck, discussing the ins and outs of punctuation and meeting cadence (which is tipping into the Waste section we&#8217;re getting to next) because you&#8217;ve convinced yourself that this is what they&#8217;re paying you for.</p><p>They&#8217;re not.</p><p>They&#8217;d much rather that energy go into the actual relationship and execution of the deal and implementation, rather than the spit and polish of the deck that merely <em>implies</em> relationship and execution, or the meetings about the meetings about the alignment around the implementation that start to actually impinge on the implementation.</p><p>The key dynamic for Friction is that investments into Friction assets <em>can</em> be helpful, but they have <em>very low ceilings</em> before they start getting counter-productive. Meanwhile, basically every effort you put into removing time and effort spent on Friction pays dividends, because it increases the Work any given team member can get through their pipeline &amp; get done in a day.</p><p>Each little trick you can pull that <em>doesn&#8217;t affect Work but does remove Friction</em> usually benefits you<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>. You may not hear the benefit straight from the customer - remember, they often don&#8217;t really care because they&#8217;re not paying you for Friction - though in reality low Friction processes often get noticed and commented on because they imply that <em>other</em> processes will also be smooth; they set positive future expectations. They become a source of expected value in their own right (but they <em>still aren&#8217;t Work yet</em>).</p><h3>Waste</h3><p>The last term in the framework is Waste - which is probably the easiest to understand and the hardest to root out fully, especially in larger companies. Most companies I&#8217;ve consulted for, sold into, or advised have codified pretty decent amounts of waste into their processes and convinced themselves that it&#8217;s completely irreplaceable<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a>.</p><ul><li><p>Before we get too far: a ton of folks get frustrated by processes that <em>feel</em> Wasteful but <em>do </em>have purpose (maybe not Work, but at least Friction). Those could be regulatory checkboxes, accessibility concerns, or prep and coordination for a team whose processes just aren&#8217;t visible to the person accountable for the action.</p></li><li><p>The best practice in these situations is, to the greatest extent possible, to label what things are. Even if you can&#8217;t be precise, saying <em>something</em> like &#8220;this process creates <em>xyz</em> metadata which protects us from liability&#8221;<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> helps both with morale and with humans + agents trying to figure out what pieces of any given process they can adjust or excise entirely.</p></li><li><p>This will become <em>even more important</em> as more agents get involved in your workflows. Not because they have more morale issues than humans (obviously) but because they aren&#8217;t expecting to or particularly adept at reading between the lines, and the un- or under-labeled data and processes are the ones agents mishandle.</p></li></ul><p>Ok, but: there&#8217;s a bunch of things that don&#8217;t fall into that category. They&#8217;re not things the customer would pay for, and they&#8217;re not necessary for getting paid. Maybe they&#8217;re coordination artifacts, maybe they&#8217;re vestigial flows from an abandoned consulting project, maybe they supported a long-since deprecated piece of software.</p><p>Moving a contract forward is Work, and therefore so is mutually-aligning with them on the contract and implementation plan. Building the deck that spells out that plan is Friction - necessary, but there&#8217;s a (very shallow) upper-bound on how much work is actually valuable, and there&#8217;s a very clear straight-line to value for every minute and calorie you can take out of the deck-building process (because it opens up more Work throughput for those AEs / SEs / etc).</p><p>Holding MAAP deck review internal syncs? Could be Friction - depends on how mature your team is - but the longer they go on, the more likely they&#8217;re Waste. The customer definitely wouldn&#8217;t pay for them (so can&#8217;t be Work) and while they might have a straight-line dependency to delivering work for a brand-new junior AE warming up to the company&#8217;s MAAP process for the first couple times, odds are they stop supporting Work and start supporting bad habits <em>really quickly</em>.</p><p>If the MAAP doesn&#8217;t get built without the sync - big problem, and running a Wasteful meeting on repeat is not the right way to handle it. If an exec can&#8217;t gather context on the deal without the sync (despite having a CRM and a RevOps platform and a team channel and an account channel and 12 other expensive tools) - also a big problem. That weekly pipeline review where everyone reads their deals out loud like so many bedtime stories while the VP could have just read the CRM updates he asked for? Pure Waste. One company I worked with killed it, saved easily 5+ hours per week per rep, and saw no change in forecast accuracy<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a>.</p><p>Not all meetings are Wasteful, but meetings that are crutches for basic accountability definitely are.</p><p>Waste has maybe the clearest investment criteria of all three: invest in removing it, wherever you find it. No Waste activity merits an incremental calorie, and cutting ruthlessly will often expose things about the business you didn&#8217;t know.</p><p>Given the mandate for AI and agents in most companies right now, you have a moment: if you&#8217;re frustrated with <em>anything</em> about the way your job works, <em>now&#8217;s the time to propose a change</em>. (If it involves, even tangentially, the use of AI, odds are you&#8217;ll get greenlit).</p><h2>To recap</h2><p>Once you see the actions you take through this lens, you can't unsee it. Every meeting becomes a question: is this moving us forward or just keeping us busy? Every document becomes a test: would the customer pay for this or are we just play-acting?</p><p>The framework is simple. Applying it is hard. Because it means admitting that a lot of what we do every day does us no good.</p><p>Next time, we'll put this framework to task. We&#8217;ll figure out how to assign activities to humans and agents based on Work, Friction, and Waste. Spoiler: in the MIT study that showed 95% of companies got no or negative ROI from their AI projects, the number one culprit was tapping agents to execute Waste activities instead of cutting them outright.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.centaurprise.com/subscribe?"><span>Subscribe now</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Not going to go into the J2BD framework, but the basics can be found here: https://www.christenseninstitute.org/theory/jobs-to-be-done/</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Also going to skip over the details on Lean, but the basics are here: https://theleanway.net/The-8-Wastes-of-Lean</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Example: When Stripe simplified payment integration from weeks to minutes, that was Work. Not the documentation, not the sales deck: the actual reduction in developer time was incremental value that customers were willing to pay more for. (When their customers implemented Stripe to simplify <em>their</em> payment integration times or streamline <em>their </em>payments or invoicing, that was reducing Friction from their perspective. We&#8217;ll get there soon.)</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Put another way: both Work and Friction are on the critical path to getting paid. That doesn&#8217;t make them equally valuable: some parts of the path ought to be shorter, and the difference between Work and Friction is &#8220;where can I shorten the critical path&#8221; versus &#8220;where would shortening the critical path turn around and bite me?&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>You&#8217;re going to find exceptions to this everywhere - recessions, negative cash flow, margin pressure, death spirals, product deprecations, etc. Those are valid, and not the point. &#8220;Improving the Work&#8221; doesn&#8217;t mean gold-plating the commodities you sell: it means finding ways to continuously improve the price / cost / margin / value of your core products and services, according to your strategy, because directing that energy into improving Friction or Waste outputs is by-definition a less-valuable use of your resources. (Directing that energy into reducing the resources you spend on Friction or Waste, however, may be quite competitive)</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Example: When Uber eliminated the need to handle cash or cards at the end of rides, they weren't changing the Work (the on-demand phone-based transportation). They were removing Friction: that awkward fumbling with payment that added time on the critical path but generated zero value. Instead, when the ride ended, you just&#8230; got out of the car. No waiting, no talking, no buttons, no hoping you had cell service. The ride quality stayed the same, but removing that transaction friction made the whole experience smoother and opened up their pipeline for more rides per driver per day and more rides per passenger per day. You wouldn&#8217;t have paid extra for no-contact payment terms, but it did become a source of expected value (it set baseline expectations that competitors didn&#8217;t meet, and it implied higher value on Work elsewhere for both buy-side and sell-side users).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>This is why I don&#8217;t personally like &#8220;Work-of-Work&#8221; or &#8220;Motion&#8221; as the names for this term: they give the misguided impression that more Work gets done proportionately with the increase in Motion / Work-of-Work, which <em>isn&#8217;t guaranteed to be true</em>. You can create <em>tons</em> of useless Motion / Work-of-Work in service of the <em>same amount of Work</em>. Friction&#8217;s a better analogy for capturing that dynamic.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>I made reference to the Lean term &#8220;Motion&#8221; and - while I prefer the connotations of the term &#8220;Friction&#8221; - I do recommend looking for Lean resources on how to minimize Motion since there&#8217;s already a slew of literature on the subject, usually referenced under &#8220;Lean Back Office.&#8221; Same for &#8220;Waste&#8221; later on.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>Feel like it&#8217;d be mean-spirited to give a named case study here like the others :)</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>Instead of what the truth might be, e.g. &#8220;we would have lost a court case, but we took a super-secret settlement, and now we have to label all our shipments just so or we go right back to court again.&#8221; Obviously <em>that&#8217;s</em> a non-starter, but you don&#8217;t have to lay out all the gory details to provide enough context.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>Turns out, if you invite <em>n</em> AEs to a <em>t</em>-minute-long status sync and have them each report for an equal amount of time, you systematically burn <em>t</em>(<em>n</em> - 1) minutes each time, since each rep spends (<em>n</em> - 1 / <em>n</em>)% of their time doing nothing, maybe listening to other people rattle off their CRM status, probably just multitasking or zoning out waiting for the meeting to end. These aren&#8217;t usually the &#8220;learning opportunities&#8221; they&#8217;re cracked up to be, especially when the attendee list gets long. 10 reps in an hour-long sync? 90% wasted time for each, burns 540 minutes for no good reason, not counting the time they needed to prep the updates and reschedule other conflicts. If they&#8217;re each at $150k OTE, that&#8217;s <em>at least</em> $675 per sync you didn&#8217;t need to spend. Replace these sessions: automate status updates, use meetings to handle direct tactical asks, and only invite the exact people needed for the ask.</p></div></div>]]></content:encoded></item><item><title><![CDATA[How to miss the point of work]]></title><description><![CDATA[The ignominious history of the DeejDrone 9000 // why you can&#8217;t just fire your whole team and replace them with named agents]]></description><link>https://www.centaurprise.com/p/how-to-miss-the-point-of-work</link><guid isPermaLink="false">https://www.centaurprise.com/p/how-to-miss-the-point-of-work</guid><dc:creator><![CDATA[DJ Thompson]]></dc:creator><pubDate>Fri, 29 Aug 2025 23:57:29 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/3ff8974b-3a0f-4f84-9055-b5cf85dd5abc_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last week, a product exec asked me a simple question that I thought encapsulated everything we're getting wrong about agents and automation. Not the tech; I don&#8217;t think you need AGI to use LLMs well. The key is <em>using them well</em>.</p><p>This is the first in a three-part series that explores why we&#8217;re getting it wrong, what it&#8217;s costing us, and how to fix it. It starts with me speccing a ridiculous vanity drone and ends with a framework that might just save you from building your own.</p><h2>The DeejDrone 9000</h2><p>So, I&#8217;m talking with a product exec last week - their company builds logistics vehicles - when he asks: "Why do we need PMs when I can generate better PRDs with AI than my team writes?"</p><p>"Let's test that," I said. We pull up the LLM he uses for work, hooked up to his company&#8217;s shared drive and resources, and I take the wheel. </p><ul><li><p>We&#8217;re going to write a PRD for a new drone (this customer doesn&#8217;t even make drones, but the AI wrote a great justification for how easy expanding into the market will be based on absolutely nothing, so we&#8217;re all set). I borrow their structure for the PRD, so I don&#8217;t miss any steps or line-items - full rigor here.</p></li><li><p>I&#8217;m calling it the &#8216;DeejDrone 9000&#8217; (again, not typical naming for the industries in question, but again: LLM&#8217;s got me covered on why the market will love it).</p></li><li><p>I&#8217;m insisting on making it biodiesel-powered-only (tagline: &#8220;she runs on corn&#8221; because enterprise aviation folks love HIMYM deep cuts? I bet?) and we&#8217;re going to paper over any questions about specific energy and setting up a new fuel supply chain and managing for weather conditions and performance at altitude because it&#8217;s trivial to do all of those things from the comfort of my browser window<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p></li><li><p>I get a <em>great</em> PRD. Beautiful structure, great prose, notional PR for release, excellent detail all around<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. Paste into Word, slap on company letterhead, print 1000 copies, let&#8217;s build this thing, right?</p></li></ul><p>I&#8217;m narrating as I go, and by this point, it&#8217;s obvious to everyone that my DeejDrone 9000 is not a great idea. The top-level &#8220;why&#8221; is reductive: &#8220;because you&#8217;re being ridiculous.&#8221;</p><p>But what made this approach ridiculous? Why does taking this cutting-edge tech and applying it to a tried-and-tested product methodology (adapted specifically to this company and LOB and vetted for rigor) result in such a garbage outcome?</p><p>And if this process was extreme, does it mean that <em>all</em> <em>variants on this process</em> are equally ridiculous? Where&#8217;s the brightline on doing this work the right way?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading centaurprise! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Role reversal leads to role erosion</h2><p><a href="https://www.centaurprise.com/p/mapping-the-messy-middle">Last time</a>, we explored how sales reps spend hours copying AI content between systems while agents write strategic plans. Today, I want to talk about <em>why</em> this keeps happening and what to do about it.</p><p>This is the root of why most companies are getting AI backwards and why it hurts more than they think. Sales reps spend hours copying AI-generated content between systems while agents write their strategic plans. Engineers review thousands of lines of generated code while AI makes their architectural decisions. PMs generate beautiful PRDs with Claude while missing what their customers actually need.</p><p>We know we&#8217;ve role-reversed. We&#8217;ve handed the thinking to the machines and kept the typing for ourselves. One dynamic we <em>didn&#8217;t</em> touch on last time: the more we hand over the thinking, the more we cut the thinking out of the process altogether<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. That's not just wasting time, it&#8217;s actively eroding capability. That&#8217;s real damage.</p><p>Over the past few months at <a href="https://gigue.ai">gigue</a>, I've been iterating on a framework I used in the past to understand why this keeps happening. It's helped me in the past to spot the difference between real work and productivity theater and show it to others. These days it's helped folks figure out where AI actually belongs in the work stack.</p><p>Hopefully it&#8217;ll help you too.</p><h2>What our artifacts are supposed to be doing for us</h2><p>We're getting dangerously good at generating artifacts - code, images, documents - without understanding what for.</p><p>Software engineering isn't generating variables and brackets and semicolons; it's building a shared intricate understanding of a problem that transcends the theory or discussion of how you <em>might</em> solve it<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. Same is true for AI image, video, and audio generation: art isn't arranging pixels or rocks or notes, it's first and foremost a way of conveying meaning and emotion that&#8217;s designed to end-run the mental barriers we put up to hearing those messages spelled out in literal, logical form<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>. </p><p>The rest of knowledge work is similarly deeper than the sum of its artifacts. Product management and enterprise sales isn't document creation; it's discovering what customers actually need and connecting their needs with what the business can deliver, then playing that back to them in their own terms<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>. Writing documents and updating CRMs is increasing taking over folks&#8217; time, but it&#8217;s still not the most important part of the job, and it&#8217;s still a <em>lagging indicator</em> of doing your job right<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>.</p><p>The backwards approach? Let AI generate different versions for different people, and then-and-only-then try to extract the core meaning from the outputs. Even when you get what seems like a coherent set, this is like getting a random assortment of pre-baked dishes and <em>then</em> trying to put together recipes for a themed meal. </p><p>Little late, no? You already have the final results, and you may not even be able to tell all the ingredients that went into them, so it&#8217;s silly to back into the lowest-common denominator and pretend like you meant all of this to happen from the start.</p><p>It&#8217;s <em>Codenames: Career Edition</em><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a>.</p><p>And that&#8217;s <em>if</em> your team put together coherent AI outputs. Way more often, five great engineers generate ten totally different pieces of software in Cursor and no one has the first idea which is the one we were supposed to be building.</p><p>Central to the problem is that coherence in practice depends on longitudinal context, something that our existing context windows make very difficult (if possible at all) and very expensive (when possible). When we write code, generate images, draft documents, or update databases, there are long-running threads and ideas that exist in time and space, outside of the scope of any agent&#8217;s context window, but in the heads of the humans tracking the program, project, or deal that the artifacts are about. </p><p>Not only are the agents not able to hang onto these threads in any detail outside a given session&#8230; you don&#8217;t really want them to, otherwise you start burning large amounts of cash on every API call. You compact and hope the important parts make it into the summary, or you reset and take each task as a wholly new challenge, reforming new context windows each time, accepting the context loss as you go.</p><p>All of this means when you tap agents to build documents as a proxy for doing the job that emits those documents as byproducts, you put yourself in danger of creating artifacts for their own sake. You&#8217;re LARPing doing business.</p><h2>Products and byproducts</h2><p>So this comes back to the example of the PRD. Why does the PRD I wrote not work? (Despite being ridiculous?)</p><p>It&#8217;s because the job of a PM is not to write PRDs. (Hopefully this is what the PMs reading this have been yelling for the past few paragraphs).</p><p>PRDs are a <em>byproduct of PMs doing their job.</em> The fact that I can use a word combobulator to fill one in &#8220;convincingly&#8221; means absolutely nothing when it comes to actually building something people want. I didn&#8217;t do the job.</p><p>A PM&#8217;s job is to figure out what the customer wants - what the customer says they want, what they actually need, what they&#8217;d like to have, what they think they want but would actually not use or even hate<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> - and then figure out how the business should go about building that thing. Or - in many cases - to recognize that we shouldn&#8217;t build it at all.</p><p>PMs take all this synthesized information and structure it into a coherent plan the business can process, aligning with stakeholders, checking feasibility, knife-fighting prioritization, and generally realizing an amorphous &#8220;concept-of-a-plan&#8221; that can turn into real value.</p><p>But because businesses don&#8217;t run well on amorphous stuff, and people don&#8217;t remember the details of amorphous stuff well, PMs document. That&#8217;s what the PRD is. The PRD is not the key output. It is an artifact that keeps people from forgetting or mishandling the key output. It&#8217;s not the Holy Grail; it&#8217;s the Ark of the Covenant<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> - it&#8217;s what&#8217;s inside that counts.</p><h2>What counts</h2><p>The DeejDrone 9000 isn't going into production. (Sorry to disappoint). But it clarifies something crucial: we can generate perfect-looking artifacts without any of the thinking that's supposed to go into them. We can create PRDs without product sense, code without architecture, emails without human empathy.</p><p>The more we hand over the thinking, the more we cut thinking out of the process altogether. That's not just inefficient, it's actively making us worse at our jobs. It&#8217;s role erosion, skill rot.</p><p>Next time, I'll delve into how to figure out which documents and activities &#8220;count&#8221; and which ones don&#8217;t. I&#8217;ve been refining a framework for a while now that explains exactly which parts of your job deserve attention (human or agentic) and which don't. There&#8217;s a couple &#8220;gotchas&#8221; I&#8217;ve picked up over time that will hopefully help.</p><p>See you then.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Not that there aren&#8217;t perfectly viable cargo drones capable of using biodiesel - there are - but this is <em>not</em> how you go about designing one.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p><a href="https://github.com/centaurprise/deej-drone-9000/blob/main/DD9K_PRD.md">For your enjoyment</a>, recreated and anonymized here.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>No, &#8220;thinking mode&#8221; on an LLM doesn&#8217;t substitute.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>The functions and types and classes of it all are a byproduct of this process, and the quality of the end result depends not on pure beauty of the syntax (which is why SWE teams are not evaluated by the sum of their LeetCode scores) but on the ability of the code to map the shared understanding of the problem and the ability of the team to map the problem to the customer&#8217;s reality with <a href="https://curtclifton.net/papers/MoseleyMarks06a.pdf">minimal excess complexity</a>. CC and Codex are great, but they&#8217;re not checking this for you. This is not to say that agentic codegen is bad: it is to say that agentic codegen in the absence of software engineering is bad. I&#8217;m not alone in thinking this: if you read some of the &#8220;tips and tricks&#8221; on r/vibecoding, for example there are a slew of folks trying out techniques for rigor in planning, architecture, orchestration, control, CI/CD, QC,&#8230; essentially, re-inventing software engineering from the ground up to get their vibe-coded results to actually work. They&#8217;re realizing what many engineers have been saying for a while now: <a href="https://worrydream.com/refs/Brooks_1986_-_No_Silver_Bullet.pdf">writing the code is a small fraction of the time and effort involved in engineering</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Tools like Midjourney and Runway let you choose with high precision the lens and lighting and scenery of the generations you&#8217;re creating, but you&#8217;re fundamentally approaching the flow backwards: your choices in technique and style are driving the heart of the output, not the other way around. There&#8217;s a <em>ton</em> of good use cases for quickly generating images and videos and songs in industry - upscaling, interstitials, brainstorming and temp generation, remastering, procedural customization - but attempting to substitute for the soul of actual art isn&#8217;t one of them.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Obligatory <a href="https://medium.com/@stewart/we-dont-sell-saddles-here-4c59524d650d">We Don&#8217;t Sell Saddles Here</a> reference.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>It may not sound like it at first, but play it out and this is common sense: if you update your CRM to pass the qualification gate <em>before you take the qualification call</em>, that&#8217;s a problem.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p><a href="https://codenames.game/">This game</a>, if you&#8217;re not familiar. Party game where you need to pick out common underlying concepts from unrelated words in as few turns as possible. The picture version in particular is a great way to learn how someone&#8217;s brain works.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>This is called the Voice of the Customer, and the gap between problems customers elucidate well and solutions they imagine poorly is startups&#8217; happy place. This is the nuance to the cliche Henry Ford quote: in real innovation space, customers are usually <em>great</em> at describing their problem and <em>terrible</em> at coming up with the right solution. That&#8217;s ok: that&#8217;s not their job, it&#8217;s ours.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>This is more an <em>Indiana Jones</em> reference than an actual discussion on religious iconography. Which I guess means the amorphous concept-of-a-plan / &#8220;spirit of the product&#8221; melts faces when you look at it directly? I don&#8217;t know, don&#8217;t extend the analogy too far.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Mapping the messy middle]]></title><description><![CDATA[The backwards world we're building]]></description><link>https://www.centaurprise.com/p/mapping-the-messy-middle</link><guid isPermaLink="false">https://www.centaurprise.com/p/mapping-the-messy-middle</guid><dc:creator><![CDATA[DJ Thompson]]></dc:creator><pubDate>Thu, 28 Aug 2025 17:18:16 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bvIK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2><strong>The backwards world we're building</strong></h2><p>Last week I watched a VP of Sales spend forty minutes copying data between spreadsheets while, three tabs over, ChatGPT had just written her entire board presentation. She'll spend another hour this week checking if the AI got the numbers right, then an hour after that carefully copy-pasting each set of bullets into the right text box on PowerPoint and resizing them to fit.</p><p>This is the irony builders don&#8217;t really like talking about: we're using humanity's most advanced pattern-matching technology to write poetry while the humans who built careers on creativity and strategy are... playing clipboard. Feel the AGI.</p><p>Earlier in<a href="https://centaurprise.com/this-is-centaurprise"> the centaurprise launch post</a>, I argued we're missing the real opportunity by debating AI versus humans when the real wins come from AI with humans. Today I want to go deeper, because even some of the companies building the best human-AI interfaces are getting the allocation exactly wrong.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading centaurprise! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>The spectrum that's blinding us</strong></h2><p>Here's how everyone seems to think about AI right now &#8211; as a single spectrum with two endpoints<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CklX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CklX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png 424w, https://substackcdn.com/image/fetch/$s_!CklX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png 848w, https://substackcdn.com/image/fetch/$s_!CklX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png 1272w, https://substackcdn.com/image/fetch/$s_!CklX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CklX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png" width="1592" height="330" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:330,&quot;width&quot;:1592,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:303254,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CklX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png 424w, https://substackcdn.com/image/fetch/$s_!CklX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png 848w, https://substackcdn.com/image/fetch/$s_!CklX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png 1272w, https://substackcdn.com/image/fetch/$s_!CklX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec0c56bf-9d3b-464f-994a-30b8d0226a4c_1592x330.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><figcaption class="image-caption"><em>On the left: beach time under the watchful eye of our benevolent overlords. On the right: our benevolent overlords are in the trashcan, and we break out the quill and parchment again.</em></figcaption></figure></div><p>On one end, you've got the AGI believers. AI becomes superintelligent, solves everything, ends labor as we know it. We all sip mai tais on the beach while the machines handle it all. The mechanics? Hand-wavy. The transition? Fiat. The politics? Boring. The funding? Magic. The physical infrastructure? It'll maintain itself, apparently; it&#8217;s historically been so good at that.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rKV8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rKV8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png 424w, https://substackcdn.com/image/fetch/$s_!rKV8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png 848w, https://substackcdn.com/image/fetch/$s_!rKV8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png 1272w, https://substackcdn.com/image/fetch/$s_!rKV8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rKV8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png" width="400" height="372.95081967213116" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/af629313-f900-4697-99b3-4f8568526469_976x910.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:910,&quot;width&quot;:976,&quot;resizeWidth&quot;:400,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rKV8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png 424w, https://substackcdn.com/image/fetch/$s_!rKV8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png 848w, https://substackcdn.com/image/fetch/$s_!rKV8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png 1272w, https://substackcdn.com/image/fetch/$s_!rKV8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf629313-f900-4697-99b3-4f8568526469_976x910.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Who puts out the beach chairs? Throws away the little paper umbrellas? Teaches kids the dangers of excess screen time? Don&#8217;t worry your pretty little head over it: it&#8217;s </em>our<em> turn to say &#8220;I&#8217;m afraid I can&#8217;t do that, Dave.&#8221;</em></figcaption></figure></div><p>Some folks on this end think the superintelligence loves us (ergo tiki drinks). Others think it destroys us (ergo death rays). But it&#8217;s two variations of the same premise: AI will do everything, humans will do nothing (just that one version of &#8220;nothing&#8221; is spicier than the other).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bvIK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bvIK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png 424w, https://substackcdn.com/image/fetch/$s_!bvIK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png 848w, https://substackcdn.com/image/fetch/$s_!bvIK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png 1272w, https://substackcdn.com/image/fetch/$s_!bvIK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bvIK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png" width="400" height="399.4277539341917" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f0a66eea-677e-4a1b-a270-b7d564039182_699x698.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:698,&quot;width&quot;:699,&quot;resizeWidth&quot;:400,&quot;bytes&quot;:371180,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bvIK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png 424w, https://substackcdn.com/image/fetch/$s_!bvIK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png 848w, https://substackcdn.com/image/fetch/$s_!bvIK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png 1272w, https://substackcdn.com/image/fetch/$s_!bvIK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff0a66eea-677e-4a1b-a270-b7d564039182_699x698.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Is there a Polymarket on which AI tries to do this first? I feel like Ani specifically has to be in the lead, poor girl has seen some </em>stuff<em> by now.</em></figcaption></figure></div><p>On the other end, you've got the bubble hypothesis. These folks hold the whole AI thing has been wholly worthless, propped up by VC subsidies and hype. When the money runs out, we'll realize ChatGPT was just autocomplete with good marketing, and we'll go back to whatever we were doing in October 2022.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zWze!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zWze!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png 424w, https://substackcdn.com/image/fetch/$s_!zWze!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png 848w, https://substackcdn.com/image/fetch/$s_!zWze!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png 1272w, https://substackcdn.com/image/fetch/$s_!zWze!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zWze!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png" width="399" height="372.01844262295083" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:910,&quot;width&quot;:976,&quot;resizeWidth&quot;:399,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zWze!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png 424w, https://substackcdn.com/image/fetch/$s_!zWze!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png 848w, https://substackcdn.com/image/fetch/$s_!zWze!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png 1272w, https://substackcdn.com/image/fetch/$s_!zWze!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f81a61e-1e7b-4b44-9a60-e4f05f77cec2_976x910.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Ok, maybe a </em>bit<em> dramatic, but who knows? Maybe inkwells and blotters make a comeback.</em></figcaption></figure></div><p>This seems... unlikely at this point. I can run Ollama with an open-source model on my laptop right now, hooked into Zed, as my own open-source copilot. For free. Forever. Even if every AI company imploded tomorrow, that capability isn't going away. Some genies don't go back in bottles.</p><p>But here's the thing: this whole spectrum is wrong. It's not about how <em>much</em> AI we use. It's about <em>what</em> we use it for.</p><h2><strong>Opening the second dimension</strong></h2><p>Instead of a line, think of it as a map.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Wio9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Wio9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png 424w, https://substackcdn.com/image/fetch/$s_!Wio9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png 848w, https://substackcdn.com/image/fetch/$s_!Wio9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png 1272w, https://substackcdn.com/image/fetch/$s_!Wio9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Wio9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png" width="1456" height="1000" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1000,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Wio9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png 424w, https://substackcdn.com/image/fetch/$s_!Wio9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png 848w, https://substackcdn.com/image/fetch/$s_!Wio9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png 1272w, https://substackcdn.com/image/fetch/$s_!Wio9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9ad6c899-c597-4959-9964-a0a3e99806ce_1600x1099.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>So, our original left-right spectrum is still there (plus the death ray version, just for fun). Vertically, we have optimization: unoptimized at the bottom, optimized at the top. We&#8217;re pretty close to the bottom on most use cases.</em></figcaption></figure></div><p>The vertical axis is optimization &#8211; are we using AI and humans for what they're each genuinely good at? The horizontal axis is the first spectrum from before &#8211; how much are we actually using AI versus humans?</p><p>On the top, you've got a strengths-based future. Humans handle trust, creativity, and judgment. AI handles pattern matching, parallel processing, and repetitive precision. Opportunity cost is minimized; adaptability and flexibility are maximized. Like the centaur chess teams I wrote about, where humans provided intuition while computers provided calculation, both adapt to optimize for their strengths relative to each other and, when they do so, the combined team is way stronger than their homogenous rivals.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OXlQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OXlQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png 424w, https://substackcdn.com/image/fetch/$s_!OXlQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png 848w, https://substackcdn.com/image/fetch/$s_!OXlQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png 1272w, https://substackcdn.com/image/fetch/$s_!OXlQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OXlQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png" width="1332" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:1332,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OXlQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png 424w, https://substackcdn.com/image/fetch/$s_!OXlQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png 848w, https://substackcdn.com/image/fetch/$s_!OXlQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png 1272w, https://substackcdn.com/image/fetch/$s_!OXlQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F67a6da5b-9b6c-45d1-81ff-ee3235324700_1332x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>We optimize for putting the most creative folks on art, and support them with AI that handles the mundane parts of that work. Same for social, strategic, and other tasks. Speaking of strengths, drawing is clearly not one of mine.</em></figcaption></figure></div><p>On the bottom edge &#8211; where I think we are right now &#8211; you've got role reversal. AI drafts beautiful images and sketches strategic documents while humans do the digital dishwashing. AI writes the poetry and architects the systems while humans copy-paste between AI windows that don&#8217;t talk to each other. We've given our most interesting problems to pattern matchers while using human intelligence as middleware.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EYHt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EYHt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png 424w, https://substackcdn.com/image/fetch/$s_!EYHt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png 848w, https://substackcdn.com/image/fetch/$s_!EYHt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png 1272w, https://substackcdn.com/image/fetch/$s_!EYHt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EYHt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png" width="1332" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:1332,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EYHt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png 424w, https://substackcdn.com/image/fetch/$s_!EYHt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png 848w, https://substackcdn.com/image/fetch/$s_!EYHt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png 1272w, https://substackcdn.com/image/fetch/$s_!EYHt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F206d385f-98c5-4a19-a15f-cb6e424fcc9d_1332x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>In so many cases right now, we get it backwards - we get uncanny-valley artwork and obvious AI-pretending-to-be-human correspondence, and we have to babysit the systems doing it.</em></figcaption></figure></div><p>It's absurd when you spell it out. But look at your own day. How much of it is spent cleaning up after AI versus doing the strategic thinking only you can do?</p><h2><strong>It&#8217;s not like we satisfy demand</strong></h2><p>That's what drives me crazy about all the "AI will take all the jobs" talk. It assumes we've run out of problems worth solving.</p><p>Think about October 2022 before the ChatGPT launch. Were we living in a solved world? Had we fed everyone? Connected everyone to the internet? Given everyone education? Handled climate change? Cured the diseases that plague us? Or even more mundane luxuries like: built enterprise software that didn't make people want to throw their laptops out windows?</p><p>Of course not. There was, and is, <em>infinitely</em> more demand. The world needs<em> so much work</em> done on it. Work that isn&#8217;t getting done because humans are blocked doing other things. Work that requires capabilities we don't have. Work that needs both human judgment and superhuman reflexes or processing power or attention or endurance.</p><p>The constraint isn't <em>demand</em> for solutions. It&#8217;s <em>capacity</em> to deliver them.</p><p>Replacing humans with AI is zero-sum thinking. It assumes the pie is fixed, that the only question is who gets which slice. But that's never been how our economy works: we&#8217;ve always had a system that relies on and rewards growth. When we augment human capability &#8211; when we give people tools that amplify their strengths instead of replacing them &#8211; we expand what's possible.</p><p>A sales rep who doesn't have to spend 72% of their week on busywork doesn't become unemployed. They become capable of managing four times as many relationships, going four times deeper with each customer, solving problems they never had time to even see before. They close deals on their long tail of their pipeline that they ignore in the status quo. They grow the business.</p><h2><strong>What optimization actually looks like</strong></h2><p>I've been building this future at <a href="https://gigue.ai">gigue</a>, watching early adopters reorganize their work around strength-based allocation. It's not theoretical anymore. I'm seeing it happen.</p><p>The sales team that uses AI to chase down follow-ups and next steps while humans focus on building relationships and understanding problems: they're not doing less work, they're doing different work. Deeper work. The kind that actually moves deals forward.</p><p>The engineering team that uses AI to generate test cases and scan for vulnerabilities while humans create features and design system architecture: they're not becoming obsolete, and they&#8217;re avoiding the skill rot you catch when you give yourself over to &#8220;the vibes.&#8221;</p><p>This isn't about efficiency. It's about capability. When you stop using humans as routers and start using them as thinkers, everything changes.</p><h2><strong>Dragging ourselves upward</strong></h2><p>The path from where we are (role reversal) to where we should be (optimization) isn't automatic. It requires deliberately rethinking every workflow, every job description, every assumption about who does what.</p><p>The centaur chess teams found this out. A 2022 study found what players already knew: amateurs with AIs were beating grandmasters with AIs, because the skillsets were different<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. The study also found something no one expected: the best solo AI engines were not the ones that dominated the centaur matchups. In fact, for both sides, being the best by yourself was <em>negatively</em> associated with centaur performance<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. Both humans and AI had to adapt to each other and reallocate tasks cleverly to win together<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>.</p><p>Most companies won't do this. They'll keep having AI write strategies while humans copy-paste, because that&#8217;s easy. It&#8217;s imagination bias, just a one-for-one swap. They'll keep having models attempt creativity while knowledge workers handle digital busywork. They'll keep operating in that bizarre bottom quadrant because it's easier than questioning the fundamental architecture of work and opening the can of worms that comes with it.</p><p>But some companies will figure this out. They'll realize that the question isn't "how do we use AI to do human jobs?" but "how do we use AI to do the jobs humans never should have been doing in the first place?"</p><p>They'll map every task to its optimal processor. Find the tasks they didn&#8217;t know were there but have been tripping them up all along. Pattern matching to AI. Relationship building to humans. Parallel processing to AI. Creative leaps to humans. Repetitive precision to AI. Contextual judgment to humans.</p><p>And slowly, iteratively, they'll drag themselves up into that top quadrant where both humans and AI are doing what they're actually good at.</p><h2><strong>The mission in the messy middle</strong></h2><p>Scott Belsky had this phrase "the messy middle" for that grinding, oscillating period between initial excitement and eventual success. That's where we are with AI right now. The demos were exciting. The AGI dreams are far away<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>. We're in the messy middle, trying to make this stuff actually work.</p><p>My bet is that the winners will be the ones who stop debating replacement and start optimizing allocation. Who stop having philosophical debates about AGI and start building systems where humans and AI each play to their strengths.</p><p>&#8216;Cause here's the thing: we don't need AGI to transform work. We just need to stop using vanilla AI backwards. Stop having it write poetry while we copy-paste. Stop having it attempt strategy while we babysit its outputs. Stop pretending it's human while using humans as machines.</p><p>The future isn't AI doing everything or AI doing nothing. It's AI and humans each doing what they're built for. That's not a compromise: it's optimization.</p><p>And it's time we started building toward it.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.centaurprise.com/subscribe?"><span>Subscribe now</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>So I feel like I owe Randall Munroe a citation here because, when I uploaded my crappy stick figure sketches into ChatGPT and said &#8220;clean this up,&#8221; I decidedly got proto-XKCD as an output. Not really my intent but I&#8217;m sure those comics are in the DALL-E pre-training dataset so&#8230; go read <a href="http://xkcd.com">xkcd.com</a> if you haven&#8217;t already? I certainly have been for the past two decades.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>&#8220;<strong>These results indicate that the new human&#8211;machine capabilities are unrelated, or even negatively related, to humans' traditional chess playing capabilities&#8211;the capabilities that machines substitute</strong> &#8230; Paradoxically, centaur chess players must <em>not</em> rely on their chess playing capabilities, which are inferior to those of the machine, but use other capabilities that allow them to complement the machine's capabilities. Leading players are therefore often not highly rated chess players, but computer engineers with modest chess capability, who approach the game from a computational point of view (Cassidy, 2014). Similarly, the ability to select and tune chess engines in both centaur and machine chess is associated with general data science and creative capabilities rather than with specific chess playing capabilities (Cowen, 2013, p. 86).<em><sup> </sup></em>To develop new complementary resource bundles, <strong>humans must have a flexible ability to go beyond domain-specific expertise when structuring, bundling, and leveraging resources</strong> (Sirmon et al., 2007)... As our findings show, <strong>it is therefore unlikely that actors who outperform in terms of traditional domain-specific capabilities are also able to do so in changed contexts where new and unrelated capabilities determine competitive advantage</strong>.&#8221; [emphasis mine] https://sms.onlinelibrary.wiley.com/doi/10.1002/smj.3387 </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>&#8220;The results of the Bayes factors (BF), the relevant Bayesian Information Criteria (BIC), and the <em>R</em><sup>2</sup> values suggest strong (BF&#8201;&gt;&#8201;20) to very strong (BF&#8201;&gt;&#8201;150) <strong>evidence of the absence</strong> of non-trivial effects &#8230; In other words, <strong>human and machine chess playing capabilities have no material impact on chess performance in centaur and engine tournaments</strong>.&#8221; [emphasis mine] https://sms.onlinelibrary.wiley.com/doi/10.1002/smj.3387 </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>I mentioned this one before - when you create &#8220;ensemble models&#8221; of humans and AIs together and run evals, the &#8220;centaur evals&#8221; wind up testing different behaviors than if you test each in isolation. (Sounds intuitive when you say it, but I see almost no one building these eval sets in practice). <a href="https://digitaleconomy.stanford.edu/wp-content/uploads/2025/06/CentaurEvaluations.pdf">Position: AI Should Not Be An Imitation Game: Centaur Evaluations | Stanford Digital Economy Lab</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>A huge part of the problem with &#8220;AGI&#8221; is that it has no technical definition (just &#8220;human-like capability or better&#8221;), so the goalposts can move on a whim. <a href="https://time.com/7205596/sam-altman-superintelligence-agi">AGI can exceed any standard of intelligence we&#8217;ve ever known</a>; <a href="https://dcahn.substack.com/p/reasonable-persons-agi">AGI could just be anything better than employees on PIP</a>. It&#8217;s hard to say where we are on the trajectory when there&#8217;s no clear destination.</p></div></div>]]></content:encoded></item><item><title><![CDATA[This is centaurprise]]></title><description><![CDATA[Kicking this off]]></description><link>https://www.centaurprise.com/p/this-is-centaurprise</link><guid isPermaLink="false">https://www.centaurprise.com/p/this-is-centaurprise</guid><dc:creator><![CDATA[DJ Thompson]]></dc:creator><pubDate>Wed, 27 Aug 2025 23:51:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vaou!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe98d5d5-c480-4926-a05e-da918a911bd9_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2><strong>Tl;dr</strong></h2><p>The best businesses are built on arbitrage. Distribution, data, regulation, talent, timing &#8212; if you see the leverage, or the <em>shift </em>in leverage, before others, you win.</p><p>Right now, there&#8217;s a massive arbitrage opportunity forming around <em>how</em> work happens. Not just tools or features, but the actual architecture of companies.</p><p>That&#8217;s happening now, and I started <strong>centaurprise</strong> to keep track of it. Not the hype, not the hand-waving or the hand-wringing: there&#8217;s already plenty of that going around. The actual changes in how companies are built, how teams are structured, how decisions get made when humans and AI start working together.</p><p>My plan is to share what I&#8217;m seeing here - at <a href="https://gigue.ai">gigue</a>, and in the companies we work with - as a way to start conversations and work through the details about these transitions. </p><h2><strong>Truths and fictions</strong></h2><p>I've been building ML systems since 2011. Managing enterprise accounts since 2016. For much of the past year, I've been building gigue, a multiplayer IDE for non-technical work that assumes humans and AI are teammates, not competitors.</p><p>We're not theorizing about hybrid work. We're shipping it.</p><p>Companies are quietly reorganizing themselves around human-AI teams. Not just in their marketing slides or investor decks, though for sure there&#8217;s a lot of that going around, but in their actual, daily operations.</p><p>There&#8217;s sales reps orchestrating AI agents to handle discovery calls, and others leaning on agent coaches to help them hone their discovery talk tracks. Engineers pair-program with models that suggest entire architectures, or outsource all their architecture decisions entirely. Support teams where AI handles triage while humans handle empathy, and support teams where AI takes the heat when the humans are out of emotional bandwidth.</p><p>There&#8217;s no playbook for this yet. The patterns are just starting to emerge.</p><p>I&#8217;ve taken to calling these companies "centaurprises"<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> &#8211; part human, part AI, figuring out how to be more than the sum of their parts<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><strong>centaurprise</strong> is a new publication - we&#8217;re figuring out what works and what&#8217;s most helpful. to support us and influence our direction, consider subscribing here:</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>The hybrid reality we&#8217;re scraping together</strong></h2><p>Here's what's cropping up, piecewise, bit-by-bit across the industry:</p><p>That competitor's blog post that actually caught your attention? Written by a human, edited by AI, fact-checked by a human, formatted by AI. The contract you just signed? AI flagged the non-standard terms, humans negotiated the edges, AI tracks compliance, humans manage the relationship.</p><p>75+% of businesses are using AI; 95+% of their AI projects are failing. But the ones that are succeeding are getting really interesting, and they don&#8217;t look anything like the hype pieces on LinkedIn or Twitter.</p><p>We keep having this exhausting debate about AI <em>versus</em> humans when the thought leaders are quietly using AI <em>with</em> humans. We catastrophize or fantasize about replacement while we're actively doing augmentation. We're so busy arguing about a future imperfect that we're missing what's happening right in front of us.</p><p>The sales email that made you respond? Drafted by LLM scooping up the latest context from the org, personalized and fact-checked by a human who knows you and your industry, timed by an AI that knows best delivery times by channel and time zone and industry and audience demographic.</p><p>The ones you ignored? Either 100% AI without context and an uncanny-valley headshot, or 100% human without homework or a response SLA. Both fail for the same reason: they're not playing as a team.</p><h2><strong>Backing into the messy middle</strong></h2><p>Playing as a team has nothing to do with AI replacing humans or humans rejecting AI. We're already living in the messy middle.</p><p>The real problem? Most folks - the ones who <em>aren&#8217;t</em> trying to replace humans outright - are still doing it backwards. We've got AI assigned to creative strategy while humans copy-paste between systems. We've got models writing poetry while knowledge workers spend hours reformatting spreadsheets. We've got artificial intelligence tackling the highest-value work while human intelligence handles the digital dishwashing.</p><p>You&#8217;ve felt it, right? That particular flavor of exhaustion that comes from cleaning up after an AI, clicking through each source to see if they even contain the cited data, checking if it hallucinated test results (or rewrote its hooks so it could fake them outright), or typing &#8220;continue&#8221; every ten minutes because it shut itself down out of confusion or context window or token limitations.</p><p>That's not what you're good at. That's not what you should be doing. That's not even work: it's babysitting.</p><p>We&#8217;re paying massive opportunity costs to work the way we work now. This isn't the future of work. It's barely the present. We can do <em>so much better</em>.</p><h2><strong>We haven&#8217;t seen the future of work yet</strong></h2><p>The reason people miss the real goal: they think it's about efficiency.</p><p>Sure, any LLM can help you write emails faster. But that's like saying the internet just made libraries faster. The part that matters is AI will change what <em>comprises</em> work.</p><p>We&#8217;ve seen this pattern before. When developers got IDEs, they didn't just write code faster. They wrote different code. They built things that weren't possible before. The tool changed the craft. The second- and third-order effects wind up washing out the short-term first-order benefits.</p><p>The same thing is about to happen to every knowledge worker. Not replacement. Evolution.</p><p>Sales reps won't just send more emails: they'll orchestrate complex, multi-threaded campaigns across dozens of stakeholders simultaneously, because they&#8217;ll abstract emails away. Marketers won't just generate more content: they'll create personally-adapted experiences for every single prospect, because they&#8217;ll abstract individual micro-updates to content away.</p><p>The core creative and strategic processes will be more important than ever; the thousands of tiny adaptations and adjustments and personalizations that no one ever had the time for will suddenly become widely available.</p><p>This isn't about doing the same work with fewer humans. It's about doing work that was impossible when we only had humans, because we didn&#8217;t have the time, the capacity, the capability, the energy to do these tasks. Opening up a second, synthetic type of automation is way more useful when you optimize for opportunity cost rather than trying to replace the first type one-for-one.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><h2><strong>Why &#8220;centaurs?&#8221;</strong></h2><p>In 1997, Garry Kasparov lost to Deep Blue. Chess was "solved." Humans were obsolete.</p><p>Except, no, that's not what happened.</p><p>Players invented "centaur chess" &#8211; humans and computers playing together<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. And something weird happened: centaur teams beat both the best humans <em>and</em> the best computers. Not because the human was better at calculating moves (they weren't). Not because the computer understood strategy (it didn't). But because together they could do something neither could do alone.</p><p>The human provided intuition and pattern recognition. The computer provided calculation and consistency. Together, they beat everyone, humans and AIs alike.</p><p>(Yeah, I know: pure AI has caught up in chess since then<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>. Players and researchers warned as early as 2013 that the gap was narrowing or already gone, though studies as late as 2022<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> still show some advantage to centaur and human-engine teams over pure-AI formats. Even still: chess is bounded, a finite solution space. Business isn't<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>.)</p><p>Plus, I like puns. That&#8217;s what you&#8217;re really subscribing to.</p><h2><strong>What I&#8217;ll bring here</strong></h2><p><strong>Every week: </strong>Real notes, experiments, and frameworks from building gigue and working with companies navigating this transition.</p><p><strong>No: </strong>AGI predictions, consciousness debates, "10 ChatGPT prompts," or dystopian hand-wringing. I feel like those bases are covered.</p><p><strong>Yes: </strong>How to structure hybrid teams. Why AI agents need managers. What breaks when half your workforce isn't human. How to design products that both humans and agents use. Where the frontier is and why the technology does or doesn&#8217;t work at specific corner cases.</p><p>I'm writing from the middle of it &#8211; building gigue, working with early adopters, watching what actually ships versus what gets announced. The gap between those two is where the real insights live.</p><p>I'm writing for builders, operators, and anyone who's tired of the hype and wants to understand what's actually happening.</p><p>This is a blog with frameworks, not formulas. Better questions lead to better systems. Expect iteration and open questions.</p><p>Unlike our official gigue blog, I don&#8217;t have a full structure or tag system for this yet: I want to let this play out a bit and see where the community and the lines of thought take it organically before imposing too much structure.</p><p>Because the truth is, the playbook is still being written. The companies that win will be the ones willing to experiment in public, learn fast, and share what works in the interest of reaching a global optimum outcome down the line (rather than chasing the social media flavor-of-the-month).</p><p>So, welcome to <strong>centaurprise</strong>. The companies building the future are doing it in real-time. Let's compare notes.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.centaurprise.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.centaurprise.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p><em>I won&#8217;t spend much time here advertising <a href="https://gigue.ai">gigue</a> - we have a separate blog, </em><strong><a href="https://gigue.ai/blog">context</a></strong>, <em>for that - but I&#8217;ll make references here and there, especially from customer conversations and field experiments. <strong>context</strong> is planned to be a lot more practical and focused on how to use our toolkit to do large enterprise sales better, right now. <strong>centaurprise</strong> is going to be more forward-looking and theoretical, &#8220;why we&#8217;re building&#8221; instead of &#8220;what we&#8217;re building.&#8221; That said - it&#8217;s early days, and I haven&#8217;t locked in formats yet. Open to feedback as we get going.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>I promise this is the last post where I say the word &#8220;centaurprise&#8221; so much. But I&#8217;ll still say it a couple more times later on.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>While using the term &#8220;centaur&#8221; for human-AI combo gets its origins from other fields, notably chess, you may be familiar with the business application from the the <a href="https://www.hbs.edu/faculty/Pages/item.aspx?num=64700">HBS/Warwick/MIT study</a> on consultant performance with and without AI assist in 2023, where Ethan Mollick referenced the term &#8220;centaur&#8221; to mean delineated task allocation and &#8220;cyborg&#8221; to mean ubiquitous AI allocation when describing LLM augmentation strategies at work (summary <a href="https://www.oneusefulthing.org/p/centaurs-and-cyborgs-on-the-jagged">here</a>).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>There&#8217;s too much here for just this post - I&#8217;ll revisit it later - but idea that the superset of behaviors for &#8220;centaur evaluations&#8221; exceeds either the set of human or AI behaviors alone is important. https://digitaleconomy.stanford.edu/wp-content/uploads/2025/06/CentaurEvaluations.pdf </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Kasparov himself is credited with inventing the modern format - he called it &#8220;advanced chess&#8221; - but the &#8220;centaur&#8221; name pops up all over in a bunch of different contexts. In the &#8216;70s the idea was called &#8220;consultation chess,&#8221; which fortunately didn&#8217;t stick.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>https://gwern.net/note/note#advanced-chess-obituary, for one.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>https://sms.onlinelibrary.wiley.com/doi/10.1002/smj.3387, worth reading in its entirety.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>There&#8217;s a lot to process in these two links, and a lot of others worth reading (on substack and elsewhere). I&#8217;m not enough of a chess expert to wax philosophical on the future of the formats, but it&#8217;s interesting that at the time of the 2022 paper classical matches were averaging 80~ turns, centaur matches were averaging 100~ turns, and engine matches were averaging 140 turns, with each format progressively more likely to end in a draw than the previous. I think this speaks to saturation of the solution space: fast games come from early errors (e.g. grandmasters don&#8217;t lose to Scholar&#8217;s Mate) and while there&#8217;s something like 10^120 possible games, the fun thing about lopping off decision tree branches is just how exponentially-quickly you can collapse that space such that a) the number of leaves is reasonable again and/or b) every remaining leaf is a draw.</p></div></div>]]></content:encoded></item></channel></rss>