<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Rumman Chowdhury]]></title><description><![CDATA[Thoughts on human cognition and artificial intelligence. Inspired by my podcast, Thinking about Thinking, and the space where I work through my ideas for my upcoming book. CEO of Humane Intelligence PBC. More about me on www.rummanchowdhury.com]]></description><link>https://rummanchowdhury1.substack.com</link><image><url>https://substackcdn.com/image/fetch/$s_!7quN!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3fb1f0d1-1e4e-4a35-94f5-2423f379477c_2048x2048.png</url><title>Rumman Chowdhury</title><link>https://rummanchowdhury1.substack.com</link></image><generator>Substack</generator><lastBuildDate>Mon, 11 May 2026 14:24:13 GMT</lastBuildDate><atom:link href="https://rummanchowdhury1.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Thinking about Thinking]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[rummanchowdhury1@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[rummanchowdhury1@substack.com]]></itunes:email><itunes:name><![CDATA[Rumman Chowdhury]]></itunes:name></itunes:owner><itunes:author><![CDATA[Rumman Chowdhury]]></itunes:author><googleplay:owner><![CDATA[rummanchowdhury1@substack.com]]></googleplay:owner><googleplay:email><![CDATA[rummanchowdhury1@substack.com]]></googleplay:email><googleplay:author><![CDATA[Rumman Chowdhury]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[We've Been Asking the Wrong Questions About Artificial Intelligence]]></title><description><![CDATA[On benchmarks, Bach, and the impending jobs apocalypse]]></description><link>https://rummanchowdhury1.substack.com/p/weve-been-asking-the-wrong-questions</link><guid isPermaLink="false">https://rummanchowdhury1.substack.com/p/weve-been-asking-the-wrong-questions</guid><dc:creator><![CDATA[Rumman Chowdhury]]></dc:creator><pubDate>Mon, 04 May 2026 12:34:20 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/195930193/675776cceba289a290d7ca584443b2dc.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>There is a phrase that came up in the first episode of <em>Thinking about Thinking</em> that I have not been able to stop thinking about since. It arrived almost offhandedly, mid-conversation, as the best ideas often do: <em>benchmarks are obsolete by design.</em></p><p>The implication is sharper than it first sounds. We have spent years arguing about whether AI benchmarks accurately measure intelligence: whether MMLU captures reasoning, whether passing the bar exam means a system can &#8220;do law,&#8221; whether benchmark saturation means models have genuinely plateaued or merely gamed the test. These are real debates, and they feed the current mania about the &#8216;jobs apocalypse&#8217;. But they rest on the premise that benchmarks were ever meant to be permanent measures in the first place.</p><p>A benchmark, in AI, is scaffolding. It is a temporary structure erected to improve the building, not to stand beside it forever as its mirror. The point of a benchmark is to induce better performance in the system being tested. Once the system has absorbed and exceeded the benchmark, the benchmark is discarded and a harder one takes its place. This is not a bug in the methodology, it is the methodology. The benchmark that isn&#8217;t eventually surpassed is a benchmark that wasn&#8217;t set ambitiously enough.<br><br></p><p>This reframes the entire AI evaluation discourse. The conversation around benchmark saturation &#8212; the observable phenomenon by which frontier models cluster near the ceiling of any given evaluation dataset within months of its release &#8212; has largely been framed as a crisis of measurement: our tests are too easy, too leakable, too narrow. But if we take seriously the idea that obsolescence is the intended endpoint, then benchmark saturation is not evidence of a broken system. It is evidence of a system working exactly as designed. The benchmark served its purpose. We should be making a new one, not lamenting the death of the old one.</p><h2>The Bach Problem</h2><p>In cognitive science and philosophy of mind, the debate over whether generative AI systems genuinely &#8220;understand&#8221; or merely &#8220;pattern match&#8221; has become something of a cottage industry, producing landmark arguments from John Searle&#8217;s Chinese Room &#8212; first articulated in &#8220;Minds, Brains and Programs&#8221; in 1980 - to, more recently, the debates surrounding emergent capabilities in GPT-4 and its successors. What these debates share is a tendency to adjudicate on behavioral grounds: can the system produce outputs indistinguishable from human outputs?</p><p>The Turing test, despite decades of philosophical critique, remains an implicit benchmark. In the course of the seminar episode, I reflected on my time as a data science teacher. Each year I&#8217;d have a student who wanted to create an &#8220;AI music generator.&#8221; Consider what happens when a student builds an AI music generator trained on Bach. The system produces music, and that music sounds like Bach&#8217;s. It is, by any Turing-adjacent standard, a success. And yet something is obviously missing. If Johann Sebastian Bach were alive today, he would not simply produce more Bach-sounding Bach. The whole reason we still perform the Goldberg Variations and the Well-Tempered Clavier three centuries after his death is that he did something completely novel. In AI-speak, he didn&#8217;t optimize within a prior distribution, he changed the distribution.</p><p>This &#8220;Bach Problem&#8221; is a formulation of something that existing intelligence theory gestures at but rarely pins down. The challenge is not whether a system can produce outputs that are statistically plausible continuations of a training corpus. Nearly any sufficiently large language model can do this with impressive fluency. The challenge is whether a system can produce outputs that have no adequate prior in the training data &#8212; outputs that aren&#8217;t interpolations or extrapolations but genuine departures.</p><p>The distinction matters because it maps onto something deeper than style imitation. Howard Gardner&#8217;s theory of multiple intelligences, first articulated in <em>Frames of Mind: The Theory of Multiple Intelligences</em>, proposed that what we call &#8220;intelligence&#8221; is not a unitary faculty but a cluster of relatively independent competencies &#8212; linguistic, musical, logical-mathematical, spatial, and so forth. What Gardner did not dwell on, but what the Bach Problem brings into focus, is that each of these intelligences, in its highest expression, is characterized not by competent performance within a domain but by genuine contribution to it. Einstein did not solve physics problems faster than other physicists. He reconceived what a physics problem was. Coltrane did not improvise jazz better than his contemporaries. He changed what jazz permitted itself to do.</p><p>By this standard, current AI systems, despite their extraordinary capabilities across text, code, mathematics, and image generation, are still, fundamentally, making more Bach-sounding Bach. This is not a dismissal of their power, but a more accurate description of their current nature and an indicator of what the next threshold would have to look like.</p><h2>Atomization and the Ghost in the System</h2><p>The seminar closed with an argument from systems theory that deserves more attention than it typically receives in AI discourse. The claim is this: AI development as currently conceived attempts to decompose human capability into atomized individual tasks &#8212; writing, reasoning, coding, image recognition, planning &#8212; achieve superhuman performance on each task individually, and then sum those performances as a proxy for human-level general intelligence. This is, structurally, a category error.</p><p>Systems theory, from Norbert Wiener&#8217;s foundational work in <em>Cybernetics, or Control and Communication in the Animal and the Machine</em> through the later contributions of Donella Meadows in <em>Thinking in Systems: A Primer</em> in 2008, establishes the foundational principle that the output of a system is not reducible to the sum of its components. Emergent properties &#8212; the behaviors and capacities that arise from the interaction of parts within a system &#8212; cannot be reconstructed by analyzing parts in isolation. This is as true of human cognition as it is of ecosystems, economies, or ant colonies.</p><p>What makes a skilled lawyer valuable is not the sum of their performance on discrete subtasks but the way those capacities interact under conditions of uncertainty, stakes, and client relationship. The same is true of what makes a great teacher, a creative scientist, or a gifted manager. These are emergent properties of systems, not additive properties of components.</p><p>This is the society that AI evaluation systems are being built to replicate, not the society we might want to build. We are designing intelligence benchmarks in the image of performance improvement plans.</p><h2>The Goalpost Principle</h2><p>There is a pattern in the history of AI milestones that the seminar identified with unusual clarity. In 1975, the economist Charles Goodhart observed that &#8220;any observed statistical regularity will tend to collapse once pressure is placed upon it for control purposes&#8221; &#8212; now widely known as Goodhart&#8217;s Law, and later crystallized by the anthropologist Marilyn Strathern as: &#8220;When a measure becomes a target, it ceases to be a good measure.&#8221;</p><p>When IBM&#8217;s Deep Blue beat Garry Kasparov in 1997, the response was not &#8220;AI has achieved general intelligence&#8221; but &#8220;chess is not a real test of intelligence.&#8221; When AlphaGo defeated Lee Sedol in 2016, we had the same response. GPT-4 passed the bar exam, the USMLE, and numerous other professional licensing assessments. The response should not be flailing about a jobs apocalypse, but a deeper understanding of what human intelligence is when aspects of the profession are automated.</p><p>However, the pattern of perpetual goalpost movement cannot be indefinitely sustained without acknowledging what it implies: that we do not have a theory of intelligence robust enough to survive contact with a system that meets our stated criteria.</p><h2>What a Real Benchmark Would Have to Measure</h2><p>If we take the Bach Problem seriously, a genuine benchmark for AI intelligence would need to assess not fluency within a distribution but contribution beyond it. If we take the systems theory argument seriously, it would need to assess not component performance in isolation but emergent behavior under conditions of integration, uncertainty, and stakes. If we take Goodhart&#8217;s Law seriously, it would need to be defined before the system is tested, with criteria that do not shift once the system approaches them. These are not easy problems to solve - but articulating them reframes how we consider, build and deploy our benchmark scaffolding.<br></p><p><em>Thinking about Thinking is a six-episode podcast series recorded live at Harvard&#8217;s Berkman Klein Center for Internet and Society. Learn more on our website <a href="https://www.rummanchowdhury.com/podcast">here</a>, and subscribe to our podcast on <a href="https://podcasts.apple.com/us/podcast/thinking-about-thinking/id1895874539">Apple Podcasts</a> and <a href="https://open.spotify.com/show/3OCsZAOy6RLzzf2azUiYGg?si=86c5abd1f35f4ddc">Spotify</a>. Episode 2 launches May 18th.<br></em></p><p><em>Sources:</em></p><p><em>Searle, J. (1980). &#8220;Minds, Brains and Programs.&#8221; Behavioral and Brain Sciences, 3, 417&#8211;457.<br>Gardner, H. (1983). Frames of Mind: The Theory of Multiple Intelligences. Basic Books. <br>Wiener, N. (1948). Cybernetics, or Control and Communication in the Animal and the Machine. MIT Press/Wiley. <br>Meadows, D. (2008). Thinking in Systems: A Primer. Chelsea Green Publishing. <br>Goodhart, C. (1975). Problems of Monetary Management: The U.K. Experience. <br>Flynn, J. (1984). &#8220;The Mean IQ of Americans: Massive Gains 1932 to 1978.&#8221; Psychological Bulletin, 95(1), 29&#8211;51. <br></em></p>]]></content:encoded></item><item><title><![CDATA[Thinking about Thinking]]></title><description><![CDATA[I'm launching a podcast and writing a book on AI and Human Cognition]]></description><link>https://rummanchowdhury1.substack.com/p/thinking-about-thinking</link><guid isPermaLink="false">https://rummanchowdhury1.substack.com/p/thinking-about-thinking</guid><dc:creator><![CDATA[Rumman Chowdhury]]></dc:creator><pubDate>Sun, 03 May 2026 11:49:51 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!7quN!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3fb1f0d1-1e4e-4a35-94f5-2423f379477c_2048x2048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A few years ago, I read an astounding fact on a subreddit. It wasn&#8217;t until 1987 that babies were anesthetized prior to surgery. Yes, that&#8217;s right. If you or I &#8212; or most people who subscribe to this newsletter &#8212; required surgery soon after birth, we were left screaming on a table while a doctor cut us open.</p><p>I had many questions. But mainly: why?</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://rummanchowdhury1.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>It was a widely held belief in the medical profession that newborns did not feel pain because their nervous system was immature, and that any expression of distress was simply reflexive. It was also believed that they retained no memories, and therefore there were no meaningful ethical considerations &#8212; pain without recall was pain without consequence.</p><p>What changed this? In 1987, Anand and Hickey published a pivotal paper in The Lancet documenting the biological responses of infants to surgery: dramatic increases in adrenaline and cortisol, shifts in glucose and fat metabolism, unstable cardiovascular patterns. The science made the suffering credible. That same year, it was declared unethical to operate on babies without anesthesia unless medically impossible.</p><p>Here is what compels me about this story. The expression of pain was insufficient to justify relief. It had to be scientifically proven that the pain was &#8220;real&#8221; to matter. </p><div class="callout-block" data-callout="true"><p>In the society we have built, the rights we expect as living creatures are a function of a justification of our intelligence and capabilities. </p></div><p>Babies were deemed incapable of recall, so we withheld pain relief. Persons with intellectual disabilities were denied basic rights on the basis of cognitive assessments. &#8220;Scientific&#8221; measurements of intelligence were used to exclude Black Americans from citizenship and women from professions. Rights have always been downstream of whoever controls the measurement.</p><p>Human cognition is the ultimate battlefield on which we consider the integration of artificial intelligence into our society. This is not a philosophical discussion &#8212; it is one rooted deeply in sociotechnical reality. Any conversation about the future of work or the future of education is actually a conversation about the measurement constructs we use to align human intelligence to economically valuable labor: SAT tests or bar exams, for example. Any conversation about intelligence is a conversation about human worth, and any conversation about human worth is political.</p><p>Our current fascination with measuring AI capabilities &#8212; and the growing existential dread around an AI apocalypse, whether economic, societal, or otherwise &#8212; is rooted in that same construct. We are not debating the technical properties of a language model. We are debating how we measure human worth, and what we will do if technological systems surpass the highest capacity of that measurement.</p><p>A deeper understanding of artificial intelligence requires a deeper understanding of what intelligence means &#8212; as a scientific, social, political, and economic construct. That is what the podcast <a href="https://www.rummanchowdhury.com/podcast">Thinking about Thinking</a> explores: six episodes, recorded as a live Harvard seminar, unscripted, with a room full of people who disagree with each other in productive ways. The podcast launches tomorrow, and a new episode will post every two weeks.</p><p>This newsletter launches alongside it. Each issue will go deeper than the episode &#8212; into the academic literature, the historical antecedents, and the things I think about as I listen back. A book is in progress at Bloomsbury Press, and this is where I work out the thinking behind it in public. If you&#8217;re new here: subscribe, and you&#8217;ll get each essay the day an episode drops. The podcast is on <a href="https://podcasts.apple.com/us/podcast/thinking-about-thinking/id1895874539">Apple Podcasts</a> and <a href="https://open.spotify.com/show/3OCsZAOy6RLzzf2azUiYGg?si=23610f9da316446c">Spotify</a>.</p><p>We are not afraid of AI because it might surpass us. We are afraid because we have always tied capability to worth &#8212; and we know, better than we admit, what we have done to the things we decided didn&#8217;t measure up.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://rummanchowdhury1.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>