<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Learning From Examples: AI Histories ]]></title><description><![CDATA[Short stories about the history of thinking machines. ]]></description><link>https://www.learningfromexamples.com/s/ai-histories</link><image><url>https://substackcdn.com/image/fetch/$s_!S1Kl!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04935df3-9e78-4564-b881-67a57b0ad87f_1024x1024.png</url><title>Learning From Examples: AI Histories </title><link>https://www.learningfromexamples.com/s/ai-histories</link></image><generator>Substack</generator><lastBuildDate>Mon, 11 May 2026 10:25:53 GMT</lastBuildDate><atom:link href="https://www.learningfromexamples.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Harry Law]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[learningfromexamples@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[learningfromexamples@substack.com]]></itunes:email><itunes:name><![CDATA[Harry Law]]></itunes:name></itunes:owner><itunes:author><![CDATA[Harry Law]]></itunes:author><googleplay:owner><![CDATA[learningfromexamples@substack.com]]></googleplay:owner><googleplay:email><![CDATA[learningfromexamples@substack.com]]></googleplay:email><googleplay:author><![CDATA[Harry Law]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Bell Labs' last trick]]></title><description><![CDATA[AI Histories #20: Support Vector Machines]]></description><link>https://www.learningfromexamples.com/p/bell-labs-last-trick</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/bell-labs-last-trick</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 09 Oct 2025 10:26:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!nmQ3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nmQ3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nmQ3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png 424w, https://substackcdn.com/image/fetch/$s_!nmQ3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png 848w, https://substackcdn.com/image/fetch/$s_!nmQ3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png 1272w, https://substackcdn.com/image/fetch/$s_!nmQ3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nmQ3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png" width="1456" height="1115" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1115,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7198788,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/175507730?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nmQ3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png 424w, https://substackcdn.com/image/fetch/$s_!nmQ3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png 848w, https://substackcdn.com/image/fetch/$s_!nmQ3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png 1272w, https://substackcdn.com/image/fetch/$s_!nmQ3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae5482c-7947-4846-a49b-6139eeeadf7f_2246x1720.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The Juggler by Remedios Varo (1956)</figcaption></figure></div><p>In March 1995 in Holmdel, New Jersey, three men put pen to paper on a wager. Larry Jackel speculated that the inner workings of neural networks would be revealed by the year 2000. Uzbek mathematical theorist Vladimir Vapnik forecast that &#8216;no one in his right mind will use neural nets that are essentially like those used in 1995.&#8217; Yann LeCun provided the signature on both bets as an &#8216;official&#8217; observer. </p><p>All three worked at the Adaptive Systems Research Department, an influential machine learning outfit at Bell Labs. As for the bets, received wisdom holds that Jackel and Vapnik were both wrong. The millennium came and went, while a comprehensive understanding of the internal processes of artificial neural networks continued to elude researchers. </p><p>Today, we know that it didn&#8217;t matter. Neural networks are king, and we still don&#8217;t really understand everything about how they work. The systems had made huge strides since the days of the single layer perceptron (<a href="https://www.learningfromexamples.com/p/uncle-sams-electronic-brain">AI Histories #7</a>), the emergence of the Hopfield network (<a href="https://www.learningfromexamples.com/p/the-great-hopfield-network-debate">AI Histories #3</a>), and the popularisation of backpropagation in the 1980s (<a href="https://www.learningfromexamples.com/p/backpropagation-is-older-than-you">AI Histories #6</a>), but their stratospheric rise was no sure thing in the 1990s.  </p><p>One promising alternative was the support vector machine (SVM). Developed by Isabelle Guyon, Bernard Boser, and Vladimir Vapnik in the early years of the 1990s, the system promised a way to identify a reliable boundary between categories that could generalise well to new data. </p><p>Where neural networks learn by trial and adjustment, SVMs solve for the single optimal boundary from the start. Where the former stresses flexibility and scale, the latter is focused on precision, stability, and mathematical guarantees. It was an attractive combination, one that offered a mixture of reliability and interpretability at a moment when most learning systems were highly opaque and unstable.  </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>From Russia with love</h3><p>Vladimir Vapnik was a Soviet statistician whose relocation to Bell Labs in 1991 brought statistical learning theory into contact with American engineering practice. Born in 1936 in Tashkent, then part of the Uzbek Soviet Socialist Republic, Vapnik was educated at Uzbek State University before studying under the cybernetician Aleksandr Lerner at the Moscow Institute of Control Sciences. </p><p>He entered Bell Labs as a mature scholar whose intellectual formation had taken place almost entirely within the Soviet academy. At the Moscow Institute of Control Sciences, his collaborations with Alexey Chervonenkis had produced Vapnik&#8211;Chervonenkis (VC) theory, a mathematical framework for analysing the conditions under which models generalise from sample data to unseen cases. </p><p>The Moscow Institute of Control Sciences, founded in 1939 under the auspices of the Soviet Academy of Sciences, was the country&#8217;s principal centre for research in cybernetics, automation, and systems theory. By the 1960s it had become a hub for work on &#8220;automatic recognition,&#8221; the effort to design algorithms that could classify signals and images without human supervision. As Emmanuil Braverman, one of the researchers at the group put it in 1966, its staff were concerned with the &#8220;problem of teaching the machine image recognition without teacher&#8221;. </p><p>The <em>capacity</em> of a model refers to its ability to fit a variety of functions, with a high-capacity model capable of fitting complex patterns and a low-capacity model better suited to fitting simpler patterns. The <em>VC dimension</em> measures capacity by identifying the largest number of points that a model can &#8216;shatter&#8217; or classify according to every possible way of labelling them. Despite the knotty mathematics involved, the central insight behind VC theory was a simple one: a model&#8217;s ability to generalise is related to its complexity, not just its performance on training data. </p><p>In this framing, capacity control pointed towards a conception of machine learning as a principled act of discovering the right predicates such as invariances, symmetries, or structural constraints. It was a philosophy that set his work apart from the empirical pragmatism of American engineering, a school of thought that emphasised the adjustment of algorithms and the accumulation of heuristics rather than a search for universal principles.</p><p>Vapnik distrusted the interpretive shortcuts so common in applied work, later likening them to Antonie van Leeuwenhoek&#8217;s descriptions of blood cells as &#8216;armies&#8217; fighting under a microscope: &#8216;He [Leeuwenhoek] saw something. Yes. But he gave wrong interpretation.&#8217; In Vapnik&#8217;s eyes, only mathematics could disclose reality without distortion. In this sense, the Uzbek carried into Holmdel the last great geometric faith of the Soviet control sciences. It was austere, axiomatic, and estranged from American pragmatism, but waiting to be made real.</p><h3>Towards support vector machines</h3><p>At its core, an SVM is a type of machine learning model that works by drawing the best possible line (or &#8216;hyperplane&#8217;) between different groups of data points. Support vector machines do not consist of multiple layers or nodes. And they do not rely on the process of &#8216;non-convex optimisation&#8217; through which a neural network adjusts to fit the data, like finding the lowest point in a landscape with many hills and valleys. Instead, support vector machines find the boundary between groups of data in a manner analogous to finding the bottom of a single, smooth bowl. </p><p>For Vapnik, this geometric purity was the SVM&#8217;s defining virtue. Where neural networks relied on non-convex optimisation, SVMs posed a single convex problem with a unique global solution. In his later reflections, Vapnik was blunt about the contrast. With &#8220;deep learning,&#8221; he remarked, &#8220;they invent something. And then they try to prove advantage of that through interpretations, which [are] mostly wrong.&#8221;</p><p>By the early 1990s Bell Labs had its own settled approach to learning problems, one that flowed from a Western statistical tradition developed independently of the Soviet school. Where the Western tradition valued empirical pragmatism, the Soviet school valued axiomatic formalism. Where the former stressed empirical benchmarks and engineering heuristics, the latter favoured mathematically principled formalisms. </p><p>The Western tradition begins with Ronald Fisher in the 1930s, who showed how to classify data by drawing a line &#8212; a linear discriminant &#8212; that best separates groups. That gave the field its basic geometry by separating surfaces as a way to reason about data, which we discussed in <a href="https://www.learningfromexamples.com/p/discriminant-analysis">AI Histories #16</a>. After the war, Abraham Wald (the star of <a href="https://www.learningfromexamples.com/p/lies-damn-lies-and-statistics">AI Histories #10)</a> reframed inference as decision-making under uncertainty. His &#8220;decision theory&#8221; treated statistics as a dynamic process of minimising risk. </p><p>Through the 1960s&#8211;70s, this optimisation-driven approach was absorbed by engineers tackling pattern recognition and signal processing. Researchers like Thomas Cover and O.L. Mangasarian cast classification as a solvable optimisation problem. Duda and Hart&#8217;s famous 1973 <em>Pattern Classification and Scene Analysis</em> codified the Western field&#8217;s pragmatic style by laying out a toolbox of methods. </p><p>By the early 1990s, most of the technical contingencies were in place for the development of support vector machines: statistical learning theory and the VC dimension to control model capacity, optimal margin algorithms for finding decision boundaries, and methods to estimate relationships between variables based on random samples of data points. </p><p>When Vapnik arrived from Moscow in 1991, he entered a culture shaped by risk, optimisation, and empirical testing rather than axioms. It was a tradition that prized what worked over what could be proved, one that would later fuse with Soviet-style formalism through kernel methods &#8212; a class of algorithms that use kernel functions to operate in high-dimensional feature spaces without explicitly computing coordinates.</p><p>A kernel is a similarity function between two data points that satisfies certain mathematical properties, allowing it to represent inner products (functions that take two vectors and return a single number) in high dimensional space. In 1950 Polish-American mathematician Nachman Aronszajn established fundamental properties of reproducing kernels that would ultimately allow this mapping process to take place. </p><h3>The kernel trick </h3><p>By 1992, Vapnik was working with Isabelle Guyon to translate the abstractions of statistical learning theory into a classifier with measurable generalisation performance. The challenge was how to build a system that could learn from examples (clue klaxon) without either overfitting &#8212; memorising the training set so well that it failed on new cases &#8212; or underfitting, failing to learn enough to solve the task at all. </p><p>In essence, the problem was how to regulate the &#8220;capacity&#8221; of a model so that it captured just enough structure to generalise beyond its training examples. The kernel trick &#8212; rediscovered in Holmdel through the collaboration of Guyon, Vapnik, and Bernhard Boser &#8212; supplied the missing piece. It allowed the abstract guarantees of statistical learning theory to be embodied in a classifier that engineers could use, providing a bridge between the theoretical space of capacity control and the empirical world of pattern recognition. </p><p>A central idea behind the kernel trick is the concept of duality, which shows how different types of classifiers can be viewed as &#8216;dual representations&#8217; of the same decision function. The principle means that the same classification problem can be represented in &#8216;primal space&#8217; or in &#8216;dual space&#8217;. This idea is important insofar as it allows the algorithm to switch between primal space and dual space depending on whichever is more computationally efficient for a given problem. </p><p>The moment that support vector machines were developed looms large in ML mythology. It begins with Bernhard Boser&#8217;s decision to leave Bell Labs in 1991 for a position at UC Berkeley. Boser, a hardware designer, was unable to start a new project in the intervening months between concluding his work at Bell Labs and beginning a new position in California. Instead, he chose to implement an algorithm from Vapnik, developed in the 1960s, which sought to find the best boundary that separates different groups of data points.</p><p>Once complete, Vapnik proposed making the algorithm &#8216;nonlinear&#8217; to enable the model to deal with distributed data points that cannot be separated well with a straight line. But where Vapnik advocated to solve this problem using a &#8216;polynomial&#8217; approach, Guyon had a different idea. Instead of explicitly creating new polynomial features, Guyon proposed using the &#8216;kernel trick&#8217; based on work by Duda and Hart (and described independently by the trio of Aizerman, Braverman, and Rozonoer in Russia). </p><p>It was this approach that led to the emergence of the support vector machine as it is commonly understood today. Guyon, Boser, and Vapnik published details of the kernelised algorithm at the Fifth Annual Workshop on Computational Learning Theory (COLT &#8217;92). Reflecting in 2016 on the development of the support vector machine, Guyon described an initial hesitance on the part of Vapnik due to the origins of the potential function algorithm from the group at the Moscow Institute for Control Sciences: </p><blockquote><p>&#8220;After some initial success of the linear algorithm, Vladimir suggested introducing products of features. I proposed to rather use the kernel trick of the &#8216;potential function&#8217; algorithm. Vladimir initially resisted the idea because the inventors of the &#8216;potential functions&#8217; algorithm (Aizerman, Braverman, and Rozonoer) were from a competing team of his institute back in the 1960&#8217;s in Russia! But Bernhard tried it anyways, and the SVMs were born!&#8221;.</p></blockquote><p>The emergence of the support vector machine marked the culmination of decades of theoretical and practical advancements in pattern recognition, statistical learning theory, and optimisation techniques. At Bell Labs, the collaboration between Boser, Vapnik, and Guyon brought these disparate threads together. Boser&#8217;s implementation of Vapnik&#8217;s optimal margin algorithm provided a starting point, while Vapnik&#8217;s proposal to add nonlinearity sought to address the challenge of complex data distributions. </p><h3>Ways of learning</h3><p>What Boser, Vapnik, and Guyon achieved in Holmdel was the blending of two intellectual cultures. From the Soviet side came the abstractions of VC theory and structural risk minimisation, with their insistence on general principles and theoretical bounds. From the Western side came a tradition of pattern recognition rooted in empirical performance, approximation methods, and the willingness to bend mathematics to fit messy data. </p><p>The result was a machine that embodied the theoretical guarantees of convex optimisation and margin maximisation coexisting with the practical imperatives of implementation and performance. Within Bell Labs&#8217; institutional culture, this interaction demonstrated that ideas forged in the high formalism of the Soviet control sciences could be translated into efficient tools for American industry. </p><p>The support vector machine represented the point at which theory and practice, abstraction and application, converged in code in the offices of New Jersey. Its development marked the closing chapter of a geometric conception of intelligence that had defined the twentieth century, one that imagined learning as the discovery of stable forms and separating surfaces in high-dimensional space.</p><p>This is why Vapnik bet Yann LeCun that artificial neural networks were a dead end. The wager, a bit of fun but entirely sincere, expressed divergent conceptions of what &#8220;learning&#8221; meant. For LeCun, intelligence was a matter of distributed adaptation: systems that adjusted their weights through experience until useful representations emerged. For Vapnik, it was an exercise in geometry and proof, which is why he explained his approach stressed finding and formalising axioms. </p><p>While deep learning eventually proved triumphant, the split made visible the moment when the field&#8217;s centre of gravity shifted from the geometric to the statistical, from global optima to local gradients, from the certainty of separability to the fluidity of representation. The bet is a hinge in the history of artificial intelligence, a moment that divided an older tradition of mathematical certainty from a new era defined by probabilistic depth and empirical abundance.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[The Forgotten Man]]></title><description><![CDATA[AI Histories #19: Nicolas Rashevsky]]></description><link>https://www.learningfromexamples.com/p/the-forgotten-man</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/the-forgotten-man</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 25 Sep 2025 10:25:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!mabC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mabC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mabC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png 424w, https://substackcdn.com/image/fetch/$s_!mabC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png 848w, https://substackcdn.com/image/fetch/$s_!mabC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png 1272w, https://substackcdn.com/image/fetch/$s_!mabC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mabC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png" width="1456" height="921" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:921,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1199411,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/174085188?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mabC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png 424w, https://substackcdn.com/image/fetch/$s_!mabC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png 848w, https://substackcdn.com/image/fetch/$s_!mabC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png 1272w, https://substackcdn.com/image/fetch/$s_!mabC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe319dbd9-fb56-4f2c-b4a4-b0c9856bc156_1568x992.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Nicolas Rashevsky via the University of Chicago Photographic Archive</figcaption></figure></div><p>Nicolas Rashevsky was born in Ukraine in 1899. He was the eldest son of a sugar-factory owner, and studied theoretical physics at St. Vladimir Imperial University (now Taras Shevchenko National University of Kyiv) before receiving his doctorate in 1919. By the time he finished his studies, the country he had grown up in was already gone.</p><p>Russia&#8217;s disastrous experience in the First World War had shaken the empire. The October Revolution of 1917 brought Lenin&#8217;s Bolsheviks to power, and by 1918 the country was embroiled in a brutal civil war. On one side stood the Reds, the new Soviet regime and its Red Army. Arrayed against them was a loose coalition known as the Whites made up of monarchists, liberals, Cossack hosts, and other anti-Bolshevik groups. Rashevsky served with the Whites, who were eventually defeated when their last southern stronghold in Crimea collapsed in November 1920.</p><p>Faced with few good options, the remnants of the Whites evacuated across the Black Sea to Constantinople. Within a year, Rashevsky was teaching physics and mathematics at the city&#8217;s Robert College before making the trip west to take up a position at the Russian University in Prague in 1921. After three years in Prague, he emigrated to the United States in 1924 to work at Westinghouse Research Laboratories in Pittsburgh where he <a href="https://link.springer.com/article/10.1023/B:HIST.0000038267.09413.0d#:~:text=,%E2%80%9D%20Physical%20Review%2031%3A%20115%E2%80%93118">published</a> on colloidal particles and the physical chemistry of cell division. </p><p>In 1934, he was invited to take up a position in the physiology department at the University of Chicago. It was a good fit for Rashevsky, who <a href="https://pdodds.w3.uvm.edu/files/papers/others/2004/abraham2004a.pdf#:~:text=exciting%20substance%2C%20which%20had%20to,is%20responsible%20for%20excita%02tion%20but">developed</a> the first mathematical description of nerve excitation in 1933 before formalising a mathematical model of the neuron in <em>Mathematical Biophysics </em>in 1938. By this time the idea that neurons were discrete cells communicating via specialised junctions was widely accepted in the world of neurophysiology. As we saw all the way back in <a href="https://www.learningfromexamples.com/p/the-neuron-doctrine">AI Histories #1</a>, that development was largely thanks to the work of Spanish histologist Ramon y Cajal. </p><p>While Rashevsky did not cite Cajal in his work, he did take Cajal&#8217;s findings &#8212; that individual neurons exist and interact at synapses &#8212; as the basis for representing neural circuits. Rashevsky took for granted the anatomical foundation provided by the Spaniard: that a neuron was a cell body with an axon, dendrites, and synapses transmitting impulses unidirectionally.</p><h3>Mathematical biology </h3><p>In 1939 Rashevsky founded the <em>Bulletin of Mathematical Biophysics</em>, the first international journal devoted to mathematical biology. In the early issues, many papers were written by Rashevsky himself and his close collaborators on topics ranging from neuron models to cell metabolism to population dynamics. The most famous of these was Pitts and McCulloch&#8217;s &#8216;A Logical Calculus of the Ideas Immanent in Nervous Activity&#8217;, an essential paper in the AI canon and the subject of <a href="https://www.learningfromexamples.com/p/father-figures">AI Histories #17</a>. </p><p>Historian Roberto Cordeschi <a href="https://link.springer.com/chapter/10.1007/978-94-015-9870-5_6?error=cookies_not_supported&amp;code=7f77598c-cf60-45bf-8e68-fdc1ef976230#:~:text=in%20previous%20chapters,computers%20since%20the%20early%201950s">explains</a> the relationship between Rashevsky&#8217;s earlier work and the McCulloch-Pitts: &#8220;Rashevsky had tried, in his 1938 Mathematical Biophysics, to analyze neural phenomena mathematically. In 1943, McCulloch and Pitts introduced Boolean algebra to describe nets of formal neurons.&#8221;</p><p>Rashevsky&#8217;s neuron was written in the language of physics, through coupled differential equations for abstract &#8216;excitatory&#8217; and &#8216;inhibitory&#8217; variables that rose and fell over time. To know whether a model neuron would fire, you had to work through those equations step by step and track changes in continuous variables. The appeal of the McCulloch&#8211;Pitts version was its simplicity. Instead of wrestling with changing quantities, they reduced the problem to a rule: if the inputs cross a threshold, the neuron fires; if not, it stays silent. </p><p>Rashevsky&#8217;s style left him stranded between two camps. To most biologists, his equations looked too abstract and too far removed from experimental life. To most mathematicians and logicians, his differential equations &#8212; formulas tracking how quantities change step by step over time &#8212; looked too messy. His neurons lived in the <em>analogue</em> world of continuous change, not the <em>logical</em> universe of on/off switches described by the McCulloch-Pitts model.</p><p>For this reason, many of the influential later neural network formalisms can be traced more directly to McCulloch and Pitts than to Rashevsky. Frank Rosenblatt&#8217;s perceptron (<a href="https://www.learningfromexamples.com/p/uncle-sams-electronic-brain">AI Histories #7</a>) was essentially a network of McCulloch-Pitts neurons with adjustable weights and a learning rule. So when Marvin Minsky and Seymour Papert put the boot into neural networks in 1969, they talked about perceptrons as binary threshold units rather than continuous, analogue models favoured by Rashevsky. </p><p>That said, today&#8217;s neural networks have crept back towards Rashevsky&#8217;s way of thinking. Instead of only treating neurons as simple on/off switches, some modern models describe how activity flows continuously over time. These developments were not directly inspired by Rashevsky &#8212; they came from control theory and physics as in <a href="https://www.learningfromexamples.com/p/backpropagation-is-older-than-you">AI Histories #6</a> &#8212; but in a way they vindicate Rashevsky&#8217;s intuition that continuous dynamics are fundamental to understanding neural computation. </p><p>Rashevsky&#8217;s work didn&#8217;t feature in the emerging computer science-oriented AI stream, but its conceptual legacy persisted. The notion of treating the brain as a network that can be quantitatively analysed is something that AI inherited from Rashevsky and the others who followed his lead. </p><p>The Russian&#8217;s career shows us that AI&#8217;s origins don&#8217;t just run through logic and computing. By writing down the first equations of neural activity, he opened the possibility of treating the brain as a system that could be formalised, analysed, and perhaps replicated. McCulloch and Pitts made the idea simple and ultimately portable, but Rashevsky made it conceivable in the first place. If the history of AI is usually told as a story of mathematicians and engineers, Rashevsky encourages us to consider whether it was also a story of physicists and biologists trying to translate the dynamics of living systems into mathematics. </p><p>Sometimes it&#8217;s simply a matter of who gets the credit. Pitts stayed close to Rashevsky&#8217;s private circle, but the AI field at large credited the younger man&#8217;s paper as foundational for both its <a href="https://www.learningfromexamples.com/p/an-introduction-to-ai-history">symbolic and connectionist schools</a>. In the long run, Rashevsky&#8217;s contributions were folded into the background while recognition for launching the AI project went to his collaborators. </p><p>Despite this, some historians have argued for Rashevsky&#8217;s inclusion in the prehistory of AI and cognitive science. Jonnie Penn <a href="https://www.repository.cam.ac.uk/bitstreams/dadca100-b29a-4dc8-8a76-81b1e6fe18f5/download">said</a> his work &#8220;informed the origins of cognitive science in the 1950s&#8221;. Tara Abraham&#8217;s <a href="https://pdodds.w3.uvm.edu/files/papers/others/2004/abraham2004a.pdf#:~:text=much%20effort,g">work</a> re-evaluated Rashevsky&#8217;s contributions and reasons for marginalisation from the biology community, which she said followed from the fact that he had &#8220;little contact with empirical biological research&#8221;. And Gualtiero Piccinini and Sonya Bahar <a href="https://onlinelibrary.wiley.com/doi/10.1111/cogs.12012?utm_source">argue</a> that &#8220;The mathematical modeling of neural processes can be traced back to the mathematical biophysics pioneered by Nicolas Rashevsky&#8221;. </p><p>Acknowledging Rashevsky enriches our appreciation of AI&#8217;s pre-history. It underscores that the quest to make mind mathematical did not start with Turing or von Neumann or Pitts and McCulloch. Rashevsky, like Ramon y Cajal before him, exists as a representative of one of AI&#8217;s many past lives. Including Rashevsky in this lineage reminds us that AI&#8217;s conceptual foundations were being laid well before the dawn of the computer age.  </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe for free to receive the next edition of AI Histories</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Sorcerer and cell ]]></title><description><![CDATA[AI Histories #18: The Lighthill report]]></description><link>https://www.learningfromexamples.com/p/sorcerer-and-cell</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/sorcerer-and-cell</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 11 Sep 2025 10:36:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!wYwz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wYwz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wYwz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg 424w, https://substackcdn.com/image/fetch/$s_!wYwz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg 848w, https://substackcdn.com/image/fetch/$s_!wYwz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!wYwz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wYwz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg" width="1200" height="678" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/af6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:678,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;undefined&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="undefined" title="undefined" srcset="https://substackcdn.com/image/fetch/$s_!wYwz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg 424w, https://substackcdn.com/image/fetch/$s_!wYwz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg 848w, https://substackcdn.com/image/fetch/$s_!wYwz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!wYwz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf6ce757-43c7-4c32-b080-d156b26f81b8_1200x678.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>A Friday Evening Discourse at the Royal Institution; Sir James Dewar on Liquid Hydrogen</em> by Henry Jamyn Brooks (1904)</figcaption></figure></div><p>The Royal Institution knows how to put on a show. In 1839, Michael Faraday used the venue to introduce dazzled researchers to early photographic techniques. Over fifty years later, J. J. Thomson told London&#8217;s great and good about the electron. Around the same time, Sir James Dewar showed off liquid hydrogen in a particularly eye-catching demonstration. </p><p>Historically, the events of the Royal Institution expanded horizons. They showed what science could do right now and what it might be able to do in the future. Humphry Davy&#8217;s public demonstrations of nitrous oxide thrilled audiences by proving that gases affect the mind and body. Faraday&#8217;s famous rotating wire experiment amazed onlookers by showing that electricity could drive mechanical motion.</p><p>In June 1973, the Royal Institution held a different kind of event. Certainly it was a spectacle, but one in which the presenter insisted on the limits of science rather than its possibilities. That person was Sir James Lighthill of the University of Cambridge, a celebrated British scientist who pioneered work in applied mathematics. His subject was AI, a field which he believed had failed to deliver on its promises.  </p><p>Against Lighthill sat three challengers: the roboticist Donald Michie, psychologist Richard Gregory, and AI grandee John McCarthy (who last appeared in <a href="https://www.learningfromexamples.com/p/a-mysterious-science">AI Histories #15</a>). Hosted by the BBC, hundreds of others were in the audience for the debate. </p><p>Lighthill opened the proceedings by making a distinction between automation &#8212; defined as the use of any machine to conduct human work &#8212; and &#8216;automatic devices that could substitute for a human being over a wide range of human activities&#8217;. He said the latter group was called &#8216;general purpose robots&#8217;, but these things were regrettably a &#8216;mirage&#8217;. </p><p>He compared the &#8216;AI scientist in the lab&#8217; to the &#8216;sorcerer in his cell&#8217;. In his view, both dealt with theatre that captured the public imagination without much to show for it. The comparison makes a certain kind of sense when you understand Lighthill&#8217;s beliefs about what science was for. As the historian Jon Agar <a href="https://www.cambridge.org/core/journals/british-journal-for-the-history-of-science/article/abs/what-is-science-for-the-lighthill-report-on-artificial-intelligence-reinterpreted/61B13B32988D6A8C58CF8AADD4777789">puts it</a>: &#8216;behind James Lighthill's criticisms of a central part of artificial intelligence was a principle he held throughout his career &#8211; that the best research was tightly coupled to practical problem solving&#8217;. </p><p>Lighthill said the great breakthroughs of the computing age belonged to automation, which he said was the preserve of &#8216;feedback control systems that act to reduce some change in quantity from its desired value&#8217;. Then he went on to say that all computers, and by extension all AI systems, are things that &#8216;manipulate symbols according to rules prescribed in a program&#8217;. </p><p>These are curious distinctions that don&#8217;t make much sense to the modern reader. We know that the foundational technology of today&#8217;s AI project is the neural network, a system whose power flows from reducing the loss between predicted and expected values. All things being equal, these systems do not manipulate symbols according to some set of rules (though they <a href="https://www.learningfromexamples.com/p/academics-need-to-take-ai-seriously">can operate</a> symbolic tools like a calculator).</p><p>Lighthill had in mind a very specific type of thinking machine, one that was usually embodied, based on hard-coded rules, and ultimately alluring yet brittle. Just like industrial automation, he argued that the strands of research that would eventually culminate in today&#8217;s large models weren&#8217;t true examples of artificial intelligence. As the old saying goes, it isn't AI if it works. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>Talking heads</h3><p>Like many of the clashes over thinking machines today, the debate wasn&#8217;t really about AI at all. The most popular topics of discussion were human intelligence, the nature of the brain, and the extent to which there are bottlenecks that prevent researchers from replicating human capacities in silicon. Regrettably, the debate took a somewhat circular turn from the outset.  </p><p>Lighthill said that the brain can&#8217;t be replicated because it&#8217;s too complex. McCarthy countered its function certainly can. Gregory said artificial neural networks aren&#8217;t representative of contemporary neuroscience research but we should study them anyway. Michie further muddied the definition of &#8216;robot&#8217; and then talked at length about the Freddy II system from his University of Edinburgh research group. </p><p>At one point, the host said he didn&#8217;t understand the issue and the researchers nodded along as if in agreement. The whole thing was a mess, frankly, and in the end the audience left more confused than they were to begin with. Say what you will about modern television, but science communication has made great progress over the last half century. </p><p>The BBC&#8217;s programme was meant to be a moment for the scientific community to respond to the publication of the <a href="https://rodsmith.nz/wp-content/uploads/Lighthill_1973_Report.pdf">Lighthill report</a>, a piece of work commissioned by the UK&#8217;s Science Research Council in 1972 to take stock of UK AI research. The council was having difficulty assessing AI grant proposals and there were concerns that some projects were overly narrow or overly ambitious. </p><p>Lighthill was engaged to help. After spending a couple of months reading AI literature and consulting researchers, he delivered his report <em>Artificial Intelligence: A General Survey</em> in March 1973. In it, he outlined three categories of AI:</p><ul><li><p><strong>&#8216;A - Advanced Automation&#8217;:</strong> This covered AI work with clear practical objectives like character or speech recognition, machine translation, and automated theorem proving. Lighthill acknowledged real, if modest, progress &#8212; but cautioned that successes were confined to toy problems. </p></li><li><p><strong>&#8216;B - Building Robots&#8217;:</strong> The attempt to create general-purpose machines that integrate perception, cognition, and action (often embodied in robots). Lighthill saw this category as a &#8216;bridge&#8217; between the others, though McCarthy disputed this reading in the debate. McCarthy agreed, however, that projects of this type were trying to achieve the vision of AI that could perform many different tasks. </p></li><li><p><strong>&#8216;C - Computer-Based Central Nervous System Studies&#8217;:</strong> The use of computers to simulate and study neurobiology and psychology, like using neural networks to model parts of the brain. Here too he noted some progress and endorsed continued work in the area, but only insofar as machines could tell us about the nature of cognition. </p></li></ul><p>The middle category, which is closest to modern conceptions of AI, bore the brunt of Lighthill&#8217;s criticism. In his report, he wrote that &#8220;Progress in category B [Building Robots] has been even slower and more discouraging&#8221;. A few pages later, he quipped that &#8220;AI not only fails to take the first fence but ignores the rest of the steeplechase altogether.&#8221;</p><p>Michie, who also gave a written response to the report, questioned Lighthill&#8217;s methodology and impartiality. Did he intentionally consult sceptical experts? Could someone outside the field fairly judge its worth? And how could Lighthill possibly be so confident about AI&#8217;s future prospects?</p><p>These were fair questions, but in the end the Science Research Council sided with Lighthill&#8217;s assessment. Funding for AI research in Britain was severely cut and many of the organised AI programmes that had existed were scrapped. The Edinburgh AI laboratory, which under Michie had been one of the world&#8217;s leading AI centres, saw its support plummet. As one retrospective <a href="https://spectrum.ieee.org/freddy-robot-british-ai-winter#:~:text=Despite%20international%20support%20from%20the,the%20field%20for%20a%20decade">put it</a>, the once bustling lab was reduced to &#8220;just Michie, a technician, and an administrative assistant&#8221;.  </p><p>The report was widely circulated and discussed internationally. In the United States, around the same time, DARPA (the main US defence research funder for AI) was undergoing its own shift. In 1974, partly due to new federal directives and disappointment with certain AI projects, DARPA started applying tighter scrutiny to AI research. It eventually published a Lighthill-style report of its own that drew similar conclusions.  </p><p>But despite the rhetoric, the global AI research community actually continued to grow in the 1970s. The historian Thomas Haigh pointed out that if one looks at metrics like number of active researchers, conference participation, and publications, interest in AI kept increasing in the wake of the Lighthill report and subsequent allusions to the first &#8216;AI winter&#8217;. </p><p>Lighthill&#8217;s focus was largely on the symbolic approach to AI development that relies on explicit symbols, logic, and rules to represent knowledge and solve problems (discussed in more detail in <a href="https://www.learningfromexamples.com/p/does-ai-begin-with-aristotle">AI Histories #9</a>). In the 1960s and early 1970s, symbolic AI encompassed areas like rule-based reasoning, search algorithms, logic and theorem proving, structured knowledge representation, and even early robotics and natural language processing. </p><p>Against it stood connectionism, the ancestor of modern deep learning where networks of individual units learn from data. Connectionism was already facing something of a challenging time of its own, after Marvin Minsky published his famous takedown of the paradigm in 1969 (the subject of <a href="https://www.learningfromexamples.com/p/uncle-sams-electronic-brain">AI Histories #7</a>). </p><p>In the report, there was little mention of connectionist approaches because the paradigm wasn&#8217;t prominent in the UK AI scene at the time. Artificial neural networks (probably the most famous incarnation of connectionism) appear in Lighthill&#8217;s discussion as tools for brain modelling in the central nervous system category &#8212; but not in the category that deals with AI in a way that we might understand it today. </p><p>It&#8217;s a useful distinction for helping us to understand the legacy of the Lighthill report. Connectionism was already under attack and symbolic AI methods had now joined it in the firing line. What the report really captured was the limits of a single paradigm, one that would eventually be sidelined when neural networks re-emerged to solve many of the problems Lighthill thought insurmountable. </p><p>In that sense, his comparison of the AI scientist to the sorcerer in his cell wasn&#8217;t entirely misplaced. Symbolic AI did produce persuasive but brittle demonstrations that resembled a magic trick. And connectionism, when it eventually displaced symbolic methods, <a href="https://www.learningfromexamples.com/p/the-economy-of-magic">had its own reputation</a> for alchemy fed by critics and boosters alike. </p><p>What Lighthill missed was that science sometimes advances through sorcery, that alchemy was less a dead end and more a transitional practice. The systems of the 1960s and 1970s may not have been general-purpose, but their success in toy environments did inspire a generation of researchers to enter the field. Today&#8217;s deep learning systems are <a href="https://www.learningfromexamples.com/p/an-introduction-to-ai-history">not an offshoot</a> of early rule-based research, but the magic of symbolic demos &#8212; a &#8216;mirage&#8217; as Lighthill put it &#8212; suggested that the problems of the AI project could eventually be solved. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Father figures]]></title><description><![CDATA[AI Histories #17: McCullough and Pitts' experimental epistemology]]></description><link>https://www.learningfromexamples.com/p/father-figures</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/father-figures</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 28 Aug 2025 10:25:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!CxOT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CxOT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CxOT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png 424w, https://substackcdn.com/image/fetch/$s_!CxOT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png 848w, https://substackcdn.com/image/fetch/$s_!CxOT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png 1272w, https://substackcdn.com/image/fetch/$s_!CxOT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CxOT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png" width="1314" height="992" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:992,&quot;width&quot;:1314,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:676949,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/171728403?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73f1bca-6406-4b96-aab7-7b5e29d7af54_1314x1044.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CxOT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png 424w, https://substackcdn.com/image/fetch/$s_!CxOT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png 848w, https://substackcdn.com/image/fetch/$s_!CxOT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png 1272w, https://substackcdn.com/image/fetch/$s_!CxOT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54e4f785-9b3f-4e6f-8a18-4f19d5aa8b99_1314x992.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Photo of Warren McCulloch</figcaption></figure></div><p>&#8220;If any love magic, he is most impious:<br>Him I cut off, who turn his world to straw&#8221;<br><strong>&#8212; Fragment of a poem written by Walter Pitts</strong> <strong>to Warren McCulloch</strong> </p><p>Walter Pitts was born in Detroit in 1923. His father was a boiler-maker, and by all accounts a violent man who pressured the young Pitts to pack in his studies and get a job. Defying his father, Pitts spent his free time at the local library where he read widely about mathematics, science, philosophy, and history.  </p><p>He read Bertrand Russell&#8217;s <em>Principia Mathematica</em>, found mistakes, and wrote to the Welsh mathematician to point them out. According to later tellings, Russell was impressed and even invited Pitts to journey to Cambridge to study with him. Still a twelve-year old boy at this point, Pitts was glad to receive the offer but turned it down on account of his age. </p><p>But three years later, when he heard that Russell would be visiting the University of Chicago, the fifteen-year-old ran away from home and headed for Illinois. He landed in Chicago, where he supported himself with menial jobs while joining those lectures he could. </p><p>Pitts never enrolled but attached himself to the orbit of Chicago&#8217;s intellectuals, publishing his first paper in 1942 when he was eighteen. His &#8216;Some Observations on the Simple Neuron Circuit&#8217; <a href="https://link.springer.com/article/10.1007/BF02477942">appeared</a> in the <em>Bulletin of Mathematical Biophysics</em>, which was the main venue for early attempts to model biological and cognitive processes with mathematics. </p><p>The journal was headed by Nicolas Rashevsky, a Russian-born researcher best known for work rendering neurons in the language of mathematics. Rashevsky vouched for Pitts and allowed him to publish under the University of Chicago banner, despite the fact that Pitts had no formal ties to the institution.      </p><p>Warren McCulloch lived a very different life, born a generation earlier into an East Coast family of lawyers, theologians, and doctors. McCulloch studied mathematics at Haverford, philosophy and psychology at Yale, then medicine at Columbia. By the 1940s he was working as a neuropsychiatrist in Chicago. He wrote poetry, smoked heavily, and liked staying up past four in the morning with whiskey and ice cream. </p><p>The two men met in 1942 through Jerome Lettvin, a medical student who Pitts got to know at one of Russell&#8217;s lectures at the university. Pitts was eighteen and McCulloch forty-three. They recognised each other immediately through a shared enthusiasm for Gottfried Leibniz, the 17th century philosopher who wondered whether human thought could be represented an alphabet composed of signs and symbols. </p><p>McCulloch, who was looking for a mechanical account of mind, had been trying to model neurons in a kind of Leibnizian language but lacked the mathematical prowess to do so. In Pitts, he saw someone who might be able to help, and invited him to live with his family in the Hinsdale suburb of Chicago. For Pitts it became a surrogate home, a place that he would remember fondly for the rest of his life.  </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>Logical calculus </h3><p>Pitts and McCulloch wanted to use Leibniz&#8217;s calculus of thought as the basis for understanding neural activity. In 1943 the pair published their joint <a href="https://www.cs.cmu.edu/~epxing/Class/10715/reading/McCulloch.and.Pitts.pdf">paper</a>, &#8216;A Logical Calculus of the Ideas Immanent in Nervous Activity,&#8217; in Rashevsky&#8217;s <em>Bulletin of Mathematical Biophysics</em>.</p><p>They proposed a simple model in which each neuron acted as a binary unit, firing if its inputs crossed a threshold (or remaining silent otherwise). By connecting these units together in different ways, they demonstrated how some basic logical operations Leibniz had described &#8212; AND, OR, and NOT &#8212; could be carried out by networks of neurons. From this starting point they argued that more complex statements could be built, and that any proposition in logic could, at least in principle, be represented in a network. </p><p>The force of the paper lay less in its biological plausibility than in its commensurability. To physiologists, it offered a stripped-down account of nervous activity. To logicians, it showed how propositions could be built into circuits. To mathematicians and engineers, it looked like a schematic for machine design. </p><p>In 1943 Jerome Lettvin introduced Pitts to Norbert Wiener, the computer scientist best known for pioneering the field of cybernetics. Wiener was impressed with Pitts, later <a href="https://bigthink.com/neuropsych/walter-pitts-logic/">writing</a> that he was &#8220;without question the strongest young scientist I have ever met&#8221;. He went on to promise Pitts a doctorate in mathematics at MIT, despite the fact the young man lacked a high school diploma.</p><p>Pitts soon moved to Cambridge as Wiener&#8217;s prot&#233;g&#233;. He joined a circle that included John von Neumann, who in 1945 wrote the &#8216;<a href="https://web.mit.edu/sts.035/www/PDFs/edvac.pdf">First Draft of a Report on the EDVAC</a>&#8217;. It was a foundational document for modern computer architecture, one that cited only a single scientific paper: McCulloch and Pitts&#8217;s &#8216;A Logical Calculus of the Ideas Immanent in Nervous Activity.&#8217; </p><p>McCulloch followed Pitts to Massachusetts in 1952, when MIT&#8217;s Jerome Wiesner invited him to head a new project at the Research Laboratory of Electronics. He accepted, trading his professorship and suburban house in Chicago for an apartment and the chance to work again with Pitts. Alongside Lettvin and the Chilean biologist Humberto Maturana they established an &#8216;experimental epistemology&#8217; group in Building 20, a makeshift wartime structure that became famous as an incubator of ideas. </p><p>McCulloch and Pitts had gone from Chicago salons to the centre of American science, with their work standing at the junction of psychiatry, biology, mathematics, and engineering. The field of cybernetics was born from the convergence, with Wiener at its head and McCulloch and Pitts among its central figures. </p><p>Just as things seemed to fall into place, the good times came to an abrupt end when Wiener&#8217;s wife told him that McCulloch was romantically involved with their daughters. Historians generally think there was no evidence for the story, but Wiener believed it all the same. </p><p>He sent Jerome Wiesner, then associate director of the Research Laboratory of Electronics, a telegram: &#8220;Please inform Pitts and Lettvin that all connection between me and your projects is permanently abolished. They are your problem.&#8221; He never spoke to Pitts again, and never explained why.</p><p>For Pitts, it was devastating. He had grown up with an abusive father, cut off his family at fifteen, and been taken in by McCulloch as a surrogate son. Wiener had been another father figure, a mentor who recognised his genius and placed him at the centre of American science. He turned down the doctorate that MIT had offered him and set fire to his dissertation notes. He withdrew from friends, drank heavily, and began a long retreat into obscurity. </p><p>In the years after the break Pitts still worked, though without the same momentum that defined his early years. With McCulloch, Lettvin, and Humberto Maturana he co-authored &#8216;<a href="https://hearingbrain.org/docs/letvin_ieee_1959.pdf">What the Frog&#8217;s Eye Tells the Frog&#8217;s Brain</a>&#8217; (1959), an experiment that found the eye filtered and pre-processed visual information before passing it on to the brain. The work was important, but for Pitts it was unsettling because it punctured his view of the brain as a hierarchy of logical propositions. </p><p>In 1969, aged 46, Pitts died from complications of cirrhosis of the liver in a boarding house. Four months later McCulloch, weakened by a heart attack, also passed away.</p><p>McCulloch and Pitts are remembered almost entirely as forerunners of the connectionist tradition, the people who first showed that networks of neurons could compute. But in 1943 they didn&#8217;t think they were choosing between logic and learning. They felt they had squared the circle by demonstrating that what Leibniz had imagined as a calculus of propositions could be realised in the firing of neurons.</p><p>Their story reminds us that AI likes to take the shape of its container. In the early 1940s, psychiatrists could look at the logical neuron and see a stripped-down account of nervous activity, logicians could look at it and see the possibility of a calculus of thought, and engineers could look at it and see a schematic for machine design. The commensurability made possible by a single model allowed these groups to talk to one another, even as they pursued very different ends. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Unnatural Selection]]></title><description><![CDATA[AI Histories #16: Ronald Fisher, statistics, and genetics]]></description><link>https://www.learningfromexamples.com/p/discriminant-analysis</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/discriminant-analysis</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 14 Aug 2025 10:25:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!GEZd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GEZd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GEZd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GEZd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GEZd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GEZd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GEZd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg" width="800" height="550" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:550,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GEZd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GEZd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GEZd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GEZd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ef47cfc-e14e-466d-9f6c-8cde36af34db_800x550.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Chrysanthemums and Bee</em> (1833-34). </figcaption></figure></div><p>In the early 1920s, Ronald Fisher <a href="https://www.wired.com/story/how-a-cup-of-tea-laid-the-foundations-for-modern-statistical-analysis-adam-kucharski-proof-book/">put</a> eight porcelain cups on a garden table at Rothamsted agricultural research station in Hertfordshire. Four had the milk poured first and four had the tea poured first. Muriel Bristol, a biologist who insisted she could taste the difference, sipped and sorted while Fisher looked on. She called all eight correctly. </p><p>Fisher knew that chance alone would yield that score about once in seventy tries, but he also knew that &#8212; if her success wasn&#8217;t due to chance &#8212; it had to be tested under conditions that removed hidden patterns in the set-up. So long as the cups are random and the observations accurate, you could in principle formalise your approach as a series of steps to follow. We might call it an algorithm. </p><p>Two decades later, AI grandee <a href="https://history.computer.org/pioneers/samuel.html">Arthur Samuel</a> borrowed the same idea for his checkers program. He let the computer occasionally play random moves in the opening, giving it clean, unbiased samples of board positions before it started learning from them. It&#8217;s a core idea behind even the biggest and best machine learning systems, one that lets them see enough of the world to hoover up the right kinds of patterns.  </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>Randomisation, formalised</h3><p>In <em>The Design of Experiments</em>, published in 1935, Fisher described the rule: if you&#8217;re going to compare two treatments, you must assign them to plots at random. Not roughly evenly and not by rotation. Randomly. Because if you don&#8217;t assign things at random, you can&#8217;t tell whether the result is due to the treatment or something else you didn&#8217;t control. </p><p>Maybe one side of the field gets more sun. Maybe the soil is drier in one patch than another. Maybe the experimenter gives a bit more attention to the first group, or unconsciously expects it to do better. Randomisation makes sure that any other differences are spread evenly between groups. That way, if you do see a difference in outcome, you can be more confident it came from the treatment rather than from something else you didn&#8217;t account for. </p><p>Fisher&#8217;s method for testing whether a treatment made a difference &#8212; what we now <a href="https://pubmed.ncbi.nlm.nih.gov/18175604/">call</a> a significance test &#8212; depends on knowing how likely each outcome was, assuming the treatment had no effect. But you can only know that if the treatments were assigned by chance. Without that, there&#8217;s no fixed set of possibilities to compare your result against. </p><p>In this sense, randomisation is the element that makes the test possible. When engineers built systems that experimented on themselves, they copied the structure Fisher had laid down. Randomise the action, observe the outcome, and ask if the difference was larger than chance. Even today, that is the basic logic that lets machines learn by trial and error without fooling themselves.</p><p>In 1922, Fisher published a <a href="https://royalsocietypublishing.org/doi/10.1098/rsta.1922.0009">paper</a> that reshaped how statistics was done. Up to that point, most estimates came from algebraic convenience or common sense. Fisher replaced both with another rule that said if you want to estimate an unknown value in your model, choose the one that makes the observed data most likely. That rule became known as <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC4360088/">maximum likelihood</a>.</p><p>Maximum likelihood defined a way of thinking where you take a model, plug in the data, and read off which version of the model fits best. That principle now sits under almost every statistical model in AI. Classifiers, regressors, language models are all trained by adjusting parameters to maximise likelihood, or minimise its negative log. That&#8217;s what people mean when they talk about minimising a loss function, whose roots we discussed in more detail in <a href="https://www.learningfromexamples.com/p/lies-damn-lies-and-statistics">AI Histories #10</a>. </p><p>The same paper introduced something he called the information of a parameter, which measured how sharply the likelihood function peaked around the best guess. A steep peak meant high confidence while a flat one meant you weren&#8217;t learning much. I won&#8217;t say much about this point, but it turned out to be an important mathematical object in machine learning that we now refer to as the Fisher information matrix.</p><p>A few years later in 1930, Fisher published <em>The Genetical Theory of Natural Selection</em>. It was a dense, mathematical book whose key idea was that the rate at which a population&#8217;s average fitness improves is equal to the amount of genetic variance in fitness it holds. </p><p>He built models to show what that looked like over time. Around the same moment, the American geneticist Sewall Wright was developing a parallel description of drift. This Wright&#8211;Fisher model captures how allele frequencies change across generations due to selection, mutation, and random drift. The model was meant for biology, but it also became the blueprint for genetic algorithms that we looked at in <a href="https://www.learningfromexamples.com/p/emergence-machines">AI Histories #2</a>. </p><p>Fisher&#8217;s theorem said that progress depends on maintaining variance, but the Wright&#8211;Fisher model showed how quickly variance disappears. That&#8217;s still a core challenge in evolutionary computation: how to keep exploring long enough to find something new, without getting stuck on the same hill forever.</p><p>In 1936, Fisher <a href="https://lgross.utk.edu/Math589Fall2020/RAFisher1936measurementsFlowerTaxa.pdf?utm_source=chatgpt.com">took measurements</a> from three species of iris &#8212; petal length, sepal width, and so on &#8212; and asked whether the species could be separated based on those numbers alone. The method he used became known as &#8216;linear discriminant analysis&#8217; or LDA.</p><p>The idea was to find one (for two classes) direction through the data that kept each species tightly grouped, while pushing the groups as far apart as possible. You begin by taking your raw measurements, projecting them onto a line, and checking which side the new point fell on. </p><p>By the 1950s and 1960s, LDA was <a href="https://www.iosrjournals.org/iosr-jece/papers/Vol.%2010%20Issue%204/Version-1/K010416167.pdf?utm_source=chatgpt.com">well-known</a> to many of the new pattern recognition groups at Bell, MIT, and the Lincoln Lab. Researchers used it to classify phonemes, radar blips, and handwriting. In Duda and Hart&#8217;s 1973 textbook, which was something like a holy text for connectionist researchers well into the 1980s, it&#8217;s the first real classifier discussed. </p><h3>Drawing a line</h3><p>In 1933, Ronald Fisher was appointed to the Galton Chair of Eugenics at University College London. He had already spent a decade arguing that Britain&#8217;s falling birth rates were a threat to &#8216;national fitness,&#8217; and that differential reproduction across social classes would lead to civilisational decline. As late as the 1950s, he was still writing letters defending sterilisation policies and publishing essays warning of social degeneration.</p><p>Fisher thought statistics was relevant to politics, and the models he built in genetics &#8212; about selection, fitness, and variance &#8212; fed into the arguments he made about society. He believed that mathematical structures could uncover the natural order of things, and that once uncovered, they ought to be preserved.</p><p>As head of the Galton Laboratory, he helped steer British research into human heredity through the middle of the 20th century. Some of the datasets, measurement protocols, and study designs he left behind were later used to <a href="https://www.nature.com/articles/s41437-020-00394-6">support</a> claims about intelligence and class.</p><p>But his work has been enormously influential in many other less controversial areas. When researchers study algorithmic bias today, for example, they draw on the same theoretical foundations Fisher developed. Fairness audits use his work to measure whether an outcome is evenly distributed across groups, and significance thresholds still rest on the logic of his null-hypothesis framework. </p><p>Some of Fisher&#8217;s ideas are deeply disagreeable, but others are foundational to scientific practice. They live on in ways he never could have imagined, often in pursuit of goals he might have opposed. The lesson, if there is one to be had, is not that <a href="https://www.learningfromexamples.com/p/weighed-measured-and-found-wanting">technology is neutral</a> or that it is hopelessly corruptible. In fact, it is technology&#8217;s <a href="https://www.learningfromexamples.com/p/the-beast-in-the-jungle">value-laden nature</a> that lets us scrutinise it, shape it, and put it to work in a way that is commensurate with our own belief systems.   </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[A Mysterious Science]]></title><description><![CDATA[AI Histories #15: The Dartmouth Summer Research Project]]></description><link>https://www.learningfromexamples.com/p/a-mysterious-science</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/a-mysterious-science</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 07 Aug 2025 10:25:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ihrG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ihrG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ihrG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg 424w, https://substackcdn.com/image/fetch/$s_!ihrG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg 848w, https://substackcdn.com/image/fetch/$s_!ihrG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!ihrG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ihrG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg" width="3191" height="1842" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1842,&quot;width&quot;:3191,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2037223,&quot;alt&quot;:&quot;Constellation Chart&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Constellation Chart" title="Constellation Chart" srcset="https://substackcdn.com/image/fetch/$s_!ihrG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg 424w, https://substackcdn.com/image/fetch/$s_!ihrG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg 848w, https://substackcdn.com/image/fetch/$s_!ihrG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!ihrG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F937cdb35-fa0b-4a7d-b19d-7b8fecf980a6_3191x1842.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Astronomicum Caesareum (1540).</figcaption></figure></div><blockquote><p>&#8220;Now tell me, just what have you and Marv been up to &#8212; Gloria has received just as much information as I have&#8221;<br>&#8212; Louise&#8217;s letter to Ray Solomonoff, July 1956 </p></blockquote><p>It was a good question, one asked by Ray Solomonoff&#8217;s girlfriend Louise in the summer of 1956. Gloria was the wife of the famous mathematician Marvin Minsky, then a Harvard Junior Fellow, whose work we last revisited in <a href="https://www.learningfromexamples.com/p/uncle-sams-electronic-brain">AI Histories #7</a>.</p><p>Ray Solomonoff, meanwhile, has yet to feature in the series but is generally regarded as the inventor of algorithmic probability. In 1956 he was a graduate of the University of Chicago and was working at Technical Research Group in New York.</p><p>Minsky and Solomonoff were spending the summer at Dartmouth College with a group of scientists organised by John McCarthy. The guests, which also included Herbert Simon, Allen Newell, and Claude Shannon (all figures we&#8217;ll get to in our series) were working on what his wife Grace Solomonoff later <a href="https://raysolomonoff.com/dartmouth/dartray.pdf">called</a> &#8216;The Mysterious Science&#8217;. It was a fitting way of describing &#8216;thinking machine&#8217; work, which for a time resisted easy classification.</p><p>Part of the draw of the workshop was to hash out what exactly thinking machines were and how the emerging discipline was referred to. &#8216;Artificial intelligence&#8217; was already on the proposal, but the attendees were more likely to describe their work as cybernetics, automata theory, or complex information processing.</p><p>You might think that what we call the thing isn&#8217;t particularly important, and you&#8217;d be right to suggest that definitional questions about the nature of the AI project can be tedious. Even today, you hear people talking up the idea that LLMs aren&#8217;t AI, which is a phrase just one step removed from &#8216;real AI has never been tried yet&#8217;.</p><p>But from a historical perspective it does matter. The field or discipline of artificial intelligence clearly <a href="https://www.learningfromexamples.com/p/an-introduction-to-ai-history">did not begin</a> in 1956; many of the technologies and techniques that are still essential to today&#8217;s AI project are much longer in the tooth than the middle of the 20th century (see <a href="https://www.learningfromexamples.com/p/backpropagation-is-older-than-you">AI Histories #6</a>, <a href="https://www.learningfromexamples.com/p/lies-damn-lies-and-statistics">AI Histories #10</a>, or <a href="https://www.learningfromexamples.com/p/the-turing-test-doesnt-measure-intelligence">AI Histories #13</a>).</p><p>The Dartmouth project, to borrow historian Thomas Haigh&#8217;s phrase, was about giving AI a <em>brand</em> of its own. That isn&#8217;t to cast aspersions on the quality of the AI project, but to recognise that brands are useful for creating and stabilising many forms of creative or intellectual life, for making it clear who owns what and what certain things actually refer to.</p><p>In commercial terms, a brand tells outsiders what they&#8217;re buying; in research politics, it tells funders what they&#8217;re backing and graduate students what tribe they&#8217;re joining. Even McCarthy himself later wrote that &#8220;one of the reasons for inventing the term `artificial intelligence&#8217; was to escape association with &#8216;cybernetics.&#8217; &#8230; I wished to avoid having either to accept Norbert Wiener [a major figure in cybernetics] as a guru or having to argue with him.&#8221;</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>An immodest proposal</h3><p>The goals of the project were famously lofty. On the original proposal from the year before, McCarthy, Minsky, Shannon, and Nat Rochester, wrote:</p><blockquote><p>&#8220;The study is to proceed on the basis of the conjecture that every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it. An attempt will be made to find how to make machines use language, form abstractions and concepts, solve kinds of problems now reserved for humans, and improve themselves. We think that a significant advance can be made in one or more of these problems if a carefully selected group of scientists work on it together for a summer.&#8221; </p></blockquote><p>Where Wiener&#8217;s cybernetics smacked of analogue servos and feedback loops, artificial intelligence was harder to place. It was wide enough to house symbolic logic, neural nets, and whatever else it needed to, yet focused enough to attract cash (the initial workshop was paid for by the Rockefeller Foundation) and energise its researchers. </p><p>The workshop opened on 18 June 1956. Most sessions took place in the top floor classroom of Dartmouth&#8217;s mathematics building. John McCarthy, Marvin Minsky, and Solomonoff were there every day; though <a href="https://raysolomonoff.com/dartmouth/boxa/dart56more5th6thweeks.pdf?utm_source">record</a>s show that many of the days weren&#8217;t particularly well attended. </p><p>The work itself was exploratory. W. Ross Ashby demonstrated his electromechanical homeostat, a machine that could keep its needles centred by rewiring itself. On another afternoon the group stopped to check the word &#8216;heuristic&#8217; in a hallway dictionary, the whole meeting standing around the lectern until a definition could be agreed. </p><p>The word was invoked through the summer of 1956. The idea was that, instead of trying to analyse the brain to develop <a href="https://spectrum.ieee.org/tag/machine-intelligence">machine intelligence</a>, participants could focus on the operational steps needed to solve a problem using heuristic methods to identify the steps. Herb Simon and Allen Newell&#8217;s <a href="https://ieeexplore.ieee.org/document/1056797">logic theory machine</a>, for example, used heuristic guides to initiate the algorithmic steps (the set of instructions to actually carry out the problem solving).</p><p>The duo held a session on their device, which saw workshop organiser John McCarthy give them a glowing <a href="https://www-formal.stanford.edu/jmc/slides/dartmouth/dartmouth/node1.html?utm_source=chatgpt.com">write up</a>: </p><blockquote><p>Newell and Simon, who only came for a few days, were the stars of the show. They presented the logic theory machine and compared its output with protocols from student subjects. The students were not supposed to understand propositional logic but just to manipulate symbol strings according to the rules they were given. </p></blockquote><p>When attendees wrote their first post-workshop papers, the logic theory machine and the idea of list processing led the introductions. The term &#8216;artificial intelligence&#8217; now points to symbol manipulation first and everything else second, a development that we still wrestle with today when people tell you only symbolic systems can be considered &#8216;AI&#8217;.</p><p>McCarthy&#8217;s phrase had floated through two months of loose talk and hard disagreement without breaking, and by the time the early papers began to cite the Dartmouth meeting the words were already doing administrative work. They marked grant lines, course titles, and the edges of a new research community. AI is still the Mysterious Science in that it promises the moon but leaves the specifics open to interpretation.</p><p>Of course that is entirely by design. Search, neural nets, and probabilistic induction all live underneath its umbrella. Our own moment uses labels like &#8216;AGI&#8217; and &#8216;superintelligence&#8217; for size, testing whether they can marshal funding and talent while staying loose enough to survive the revisions that real progress always demands.</p><p>Dartmouth&#8217;s lesson is that a field can begin with unanswered questions and unfinished business, so long as it finds an organising principle that allows for disagreement, divergence, and dogma to coexist peacefully.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[If you can't stand the heat ]]></title><description><![CDATA[AI Histories #14: Boltzmann machines]]></description><link>https://www.learningfromexamples.com/p/if-you-cant-stand-the-heat</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/if-you-cant-stand-the-heat</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 31 Jul 2025 10:25:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Q6Ql!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q6Ql!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q6Ql!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Q6Ql!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Q6Ql!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Q6Ql!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q6Ql!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Smarthistory &#8211; Luca Signorelli, The Damned Cast into Hell&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Smarthistory &#8211; Luca Signorelli, The Damned Cast into Hell" title="Smarthistory &#8211; Luca Signorelli, The Damned Cast into Hell" srcset="https://substackcdn.com/image/fetch/$s_!Q6Ql!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Q6Ql!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Q6Ql!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Q6Ql!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ab5f68d-3db6-4088-ac04-3cf52673a52a_2450x1378.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>The Damned Cast into Hell</em> by Luca Signorelli from 1499&#8211;1502</figcaption></figure></div><p>By the second half of the nineteenth century, physicists knew that energy tended to even out. Hot things liked to cool down and gases expanded to fill the space they were in. Formalised as the second law of thermodynamics, this idea holds that a closed system&#8217;s entropy (often described as &#8216;disorder&#8217;) keeps rising as its energy spreads out.</p><p>That sounds like a force of nature, but it&#8217;s better reckoned with as a way of characterising how systems behave when left to their own devices. If the world looks orderly to us, that&#8217;s just because we&#8217;re experiencing unlikely but possible states bubble up before they disappear. </p><p>At the core of this observation is Boltzmann distribution, which gives the probability of a system occupying a state as a function of that state&#8217;s energy. <a href="https://www.fields.utoronto.ca/talks/Origin-Boltzmann-Distribution">Described</a> by the Austrian physicist Ludwig Boltzmann in the 19th century, the idea put forward that low energy states are more likely, and high energy ones become rarer as a system cools. Because rare states happen more often at higher temperatures, systems become more dynamic as heat increases. </p><p>What matters here is the claim that randomness has structure. That if you can&#8217;t follow every molecule in a glass of water, you could still know what kinds of configurations were likely. Put another way, the Boltzmann distribution is a way of thinking about systems in terms of tendencies rather than rules. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>Spin glasses</h3><p>A spin glass is a material made of minuscule magnetic units, called spins, which each act like a tiny compass that point either up or down. In most magnets, the spins tend to align with each other, which creates a strong overall magnetic field. In spin glasses, the spins are <a href="https://www.learningfromexamples.com/p/the-great-hopfield-network-debate">influenced</a> by conflicting forces. Some want to align but others want to point in opposite directions, so there's no arrangement that satisfies all of them at once.</p><p>The result is magnetic deadlock where the spins get stuck in a disordered pattern with no clear overall direction. Our system becomes stable but messy, trapped somewhere between maximally ordered and chaotic states. We describe the specific arrangements of individual spins held as &#8216;local energy minima,&#8217; a term familiar to anyone who knows about the operation of connectionist AI systems like neural networks. </p><p>Spin glasses neither collapse into randomness nor configure themselves into symmetrical states. They get stuck, but in a way that we can predict. For many scientists, this made them rich research subjects in their own right; for others, the idea of a system that stabilises without fully resolving reminded them of other natural phenomena.</p><p>One particularly resonant comparison came in 1982, when John Hopfield <a href="https://www.learningfromexamples.com/p/the-great-hopfield-network-debate">proposed</a>  a simple network of binary units each connected symmetrically to the others. The idea was that the Hopfield network could store and retrieve patterns by settling into multiple stable states, each of which corresponded to a memory. Rather than being guided by an external controller, it would recall what it had &#8216;seen&#8217; by letting its internal dynamics find a familiar configuration.</p><p>That&#8217;s the core of the &#8216;associative memory&#8217; idea behind the system, which describes a gradual adjustment until it lands in a configuration that best matches the input. A partial or noisy signal activates the system, and the network completes the pattern automatically.</p><p>Hopfield didn&#8217;t claim this was how the brain actually worked, but he did show that you could treat a pattern recall problem like a physical relaxation problem. What had been a question about cognition became a question about finding low points in a landscape. In doing so it offered a different model of intelligence, one that brought the models of statistical physics into the world of computation. </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nsas!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" width="48" height="15.652173913043478" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:150,&quot;width&quot;:460,&quot;resizeWidth&quot;:48,&quot;bytes&quot;:12198,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162870944?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a011107-4790-4b64-9f4c-4b8fcace22de_460x330.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Hopfield&#8217;s networks were clever but static. The architecture could store patterns, but the rules for how to update the weights were limited and biologically implausible. You could tweak the weights to embed a few memories, but you couldn&#8217;t easily make the system learn from data. </p><p>In 1985, Geoffrey Hinton, David Ackley, and Terry Sejnowski added noise to the Hopfield network. Instead of flipping deterministically into a new state, each unit in the network would switch on or off with a probability that followed the Boltzmann distribution. High energy states were unlikely and low energy ones were preferred. But now, unlike in Hopfield&#8217;s model, the system could find its way out of a local minimum if the temperature was high enough. </p><p>They called it the &#8216;Boltzmann machine,&#8217; and it used a slow but elegant learning rule to update the internal state of the system. First, you connected the visible units to the data and let the hidden units adjust. Then you unclamped the system and let it run freely. You compared the two distributions &#8212; how often different configurations showed up in each phase &#8212; and used the weights to reduce the gap. The goal was to make the model&#8217;s internal world reflect the structure of the real one.</p><p>For a moment, it looked like Boltzmann machines might turn the field on its head. They had the ring of generality in that they were learning to understand the distribution those things came from. In a discipline still recovering from the failure of expert systems, that was an intoxicating promise.</p><p>Alas, they had some problems. Training full Boltzmann machines was slow and sampling took forever. You needed to reach equilibrium just to take a gradient step, and each new data point meant starting the process again. It was an elegant theory that couldn&#8217;t scale in reality, at least until Hinton <a href="https://www.cs.toronto.edu/~hinton/absps/nccd.pdf">found a workaround in 2002</a>.</p><p>In this version of the Boltzmann machine, units within the same layer were prevented from communicating. Only visible-to-hidden links remained, which stripped out the feedback loops and made sampling easier. Instead of full equilibrium, you took only enough steps to approximate the gradient in a process called &#8216;contrastive divergence&#8217;.</p><p>Stack a few of these &#8216;restricted&#8217; Boltzmann machines on top of each other and you got what researchers term a &#8216;deep belief network&#8217; where each layer learned to represent the structure of the one below it. In 2006, in one of the first concrete demonstrations that deep learning could work, Hinton and his collaborators <a href="https://www.cs.toronto.edu/~hinton/absps/fastnc.pdf">showed</a> it could achieve decent results on tasks like digit recognition. </p><p>This signal primed the field before convolutional neural networks were retooled for the era of large datasets and GPUs just a few years later. So in 2012, when AlexNet famously proved just how powerful massive neural networks could be, researchers were quick to recognise it as the moment that the deep learning era arrived in force.  </p><h3>Cooling off </h3><p>Today, there&#8217;s a small but serious group of researchers working on modern energy-based models, many of whom see Boltzmann machines as part of their prehistory. They&#8217;re trying to build tools that evaluate configurations rather than generate sequences, that score entire states rather than predict the next token. There&#8217;s a kernel of something interesting there.</p><p>But it&#8217;s also a space full of goofy handwaving. You hear about cognition as entropy minimisation. You hear about &#8216;thermodynamic computing&#8217; and you start to notice that the more abstract the claim, the less likely it is to come with a working demo. Boltzmann&#8217;s name helps because it carries weight; people know it vaguely means something to do with probability and physics and systems finding balance. </p><p>But despite their relative lack of popularity, Boltzmann machines still matter to the history of AI. They might not have directly led to today&#8217;s most popular and powerful architectures, but they offered a particularly sharp version of a much older idea about the emergent nature of intelligence. </p><p>That idea was what made machine learning attractive <a href="https://www.learningfromexamples.com/p/lies-damn-lies-and-statistics">from the start</a>. What Boltzmann machines did was push it further, drawing directly from physics to provide a theory of learning as a thermodynamic process. Seen another way, the contribution of Boltzmann machines was more rhetorical than practical. Important, yes, but not because thermodynamic computing is going to replace large language models any time soon.  </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[The Turing test doesn’t measure intelligence ]]></title><description><![CDATA[AI Histories #13: The genesis of the imitation game]]></description><link>https://www.learningfromexamples.com/p/the-turing-test-doesnt-measure-intelligence</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/the-turing-test-doesnt-measure-intelligence</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 24 Jul 2025 10:25:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Q-nY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q-nY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q-nY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Q-nY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Q-nY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Q-nY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q-nY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg" width="1200" height="675" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:675,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Alan Turing obituary&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Alan Turing obituary" title="Alan Turing obituary" srcset="https://substackcdn.com/image/fetch/$s_!Q-nY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Q-nY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Q-nY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Q-nY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F051bf689-8072-4411-ab6a-8e554edd9c39_1200x675.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Colourised version of a 1946 photograph of Alan Turing running a marathon.</figcaption></figure></div><p>Earlier this year, researchers from UC San Diego <a href="https://x.com/camrobjones/status/1907086875103236393">said</a> OpenAI's GPT-4.5 passed the Turing test. In a <a href="https://arxiv.org/pdf/2503.23674">paper</a> running through the results of the experiment, the group reported that the model was thought to be human more frequently than actual humans. </p><p>That is surely impressive, but it probably means less than you think. As the authors take care to <a href="https://x.com/camrobjones/status/1907086877871448473">explain</a>, the headline result doesn&#8217;t necessarily tell us anything about whether LLMs are intelligent.  </p><p>Today&#8217;s post argues that, despite the status of the &#8216;imitation game&#8217; in the popular imagination, the <a href="https://www.cs.ox.ac.uk/activities/ieg/e-library/sources/t_article.pdf">test</a> wasn&#8217;t designed to be a practical assessment of machine intelligence. Instead, it is better understood as counterpunch in an intellectual sparring match between Turing and his greatest rivals. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>Intelligence and rhetoric </h3><p>The April 2025 paper from UC San Diego follows a <a href="https://arxiv.org/pdf/2405.08007">similar study</a> conducted by the group last year, where they evaluated GPT-3.5, GPT-4, and the ELIZA system I wrote about in <a href="https://www.learningfromexamples.com/p/why-why-why-eliza">AI Histories #11</a>. </p><p>In the 2024 study, the researchers set up a simple two player version of the game on the research platform Prolific. They found that GPT-4 was judged to be human 54% of the time, that GPT-3.5 succeeded in 50% of conversations, and that ELIZA managed to hoodwink par ticipants in 22% of chats. Real people beat the lot, and were judged to be human 67% of the time. </p><p>As well as reporting more impressive results, the <a href="https://arxiv.org/pdf/2503.23674">recent study</a> moves closer to the structure of the test first put toward by Turing: participants speak to a human and AI simultaneously and decide which is which. As Turing <a href="https://courses.cs.umbc.edu/471/papers/turing.pdf">explained</a> in the original 1950 paper:</p><blockquote><p>&#8220;It is played with three people, a man (A), a woman (B), and an interrogator (C) who may be of either sex. The interrogator stays in a room apart from the other two. The object of the game for the interrogator is to determine which of the other two is the man and which is the woman. He knows them by labels X and Y, and at the end of the game he says either "X is A and Y is B" or "X is B and Y is A."&#8221;</p></blockquote><p>Instead of determining whether participant A or B is a man or a woman, the first version of the Turing test sees the judge pick whether or not the writer is a person or a machine. This three-person structure is usually ignored in favour of a simpler two-person approach, though it was faithfully replicated in the new study.  </p><p>But to take a step back, what do we think a game about whether a man could stand in for a woman (or vice versa) is actually testing? And what do we think that means for the version of the game involving a machine? Turing gives us a clue:</p><blockquote><p>&#8220;The original question, "Can machines think?" I believe to be too meaningless to deserve discussion. Nevertheless I believe that at the end of the century the use of words and general educated opinion will have altered so much that one will be able to speak of machines thinking without expecting to be contradicted.&#8221;</p></blockquote><p>The test wasn&#8217;t designed to answer the question of whether machines can think (one doesn&#8217;t make a test to answer a meaningless question). But, just like the gender imitation game, the test must be fulfilled in a way that prevents a third party observer from being able to tell the difference between those involved. It&#8217;s about the rhetoric of intelligence, not the substance of it.</p><p>In an exchange used to illustrate how we might catch a machine out, Turing describes a back and forth in which the judge asks whether an agent could play chess (it says yes) or write a sonnet (it says no). The implication, of course, is that any sufficiently intelligent machine would be capable of engaging in &#8216;creative&#8217; pursuits (apologies to all the chess players out there).</p><p>The final aspect of note is the type of machine that Turing believes will be entangled with intelligence in the future. As he writes towards the end of the paper: &#8220;instead of trying to produce a programme to simulate the adult mind, why not rather try to produce one which simulates the child's?&#8221;</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nsas!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" width="58" height="18.91304347826087" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:150,&quot;width&quot;:460,&quot;resizeWidth&quot;:58,&quot;bytes&quot;:12198,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162870944?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a011107-4790-4b64-9f4c-4b8fcace22de_460x330.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>So we have a thought experiment that seeks to set the conditions in which someone could <em>call</em> machines intelligent, explicit links with gender, learning machines, and creative pursuits as essential markers of intelligence. Taken in the round, these elements puncture the two most common interpretations of the imitation game.</p><p>First, the &#8216;reductionist&#8217; view, which holds that the Turing test was developed to measure intelligence. This idea is popular with some AI practitioners, who see the test as a soluble target that should inform research. In this version, intelligence can be directly measured and passing the test is a meaningful benchmark.</p><p>Next up is the &#8216;constructionist&#8217; interpretation that focuses on the idea that the test itself creates a certain type of intelligence through its design and implementation. In other words, the test actively shapes our understanding of AI rather than passively measuring it. </p><p>Both interpretations buy into the idea that the test was formulated on the basis that it could, and should, be implemented in the real world. But that isn&#8217;t the case. As Bernardo Gon&#231;alves&#8217;s suggests in <em>The Turing Test Argument</em>, we can&#8217;t escape the context in which the paper was written: Turing&#8217;s debates with physicist Douglas Hartree, philosopher Michael Polanyi, and neurosurgeon Geoffrey Jefferson.</p><p>The essence of the clash is simple. Turing believed that thinking machines would eventually outstrip all of the cognitive abilities of humans, while the others thought otherwise. </p><p>University of Cambridge mathematician Douglas Hartree argued that computers would always be calculation engines incapable of acting in creative or unexpected ways. To make his case, Hartree cited Ada Lovelace's view that computers can only do what they are programmed to do in his 1950 book <em>Calculating Instruments and Machines</em>: &#8216;The Analytical Engine has no pretensions whatever to originate anything. It can do whatever we know how to order it to perform.&#8217; </p><p>So, an intelligent machine must be capable of performing tasks that it has not been specifically programmed to. Turing agreed, which is why he chose to connect his test with a &#8216;child&#8211;machine&#8217; or what he called the &#8216;unorganised machine&#8217; that could learn from experience.</p><p>Probably Turing&#8217;s most well respected critic was neurologist Geoffrey Jefferson, who set stringent criteria for machine intelligence that emphasised creativity. As <em>The</em> <em>Times</em> reported in 1949, he commented that &#8216;Not until a machine can write a sonnet or compose a concerto because of thoughts and emotions felt, and not by the chance fall of symbols, could we agree that machine equals brain &#8212; that is, not only write it but know that it had written it.&#8217;</p><p>Responding in the same newspaper on the next day, Turing, in typical cutting fashion, told the reporter &#8216;I do not think you can even draw the line about sonnets, though the comparison is perhaps a little bit unfair because a sonnet written by a machine will be better appreciated by another machine&#8217;. As we saw, Turing would go on to incorporate the idea of a machine writing a sonnet and being questioned about it in his imitation game.</p><p>Jefferson also <a href="https://philsci-archive.pitt.edu/20484/1/turing-test-controversy-preprint.pdf">argued</a> that hormones were crucial for producing facets of behaviour that machines could not replicate. In one example he said that, were it possible to create a mechanical replica of a tortoise, &#8216;that another tortoise would quickly find it a puzzling companion and a disappointing mate.&#8217;</p><p>The relationship between sex and intelligence was the motivating factor in Turing's decision to include gender imitation as part of his test, which represents a challenge to the idea that certain modes of behaviour were<em> </em>dependent on physiological conditions.</p><p>The final element of the debate that Turing responded to was from Hungarian-British polymath Michael Polanyi, who <a href="https://www.learningfromexamples.com/p/does-ai-know-things">argued</a> that human intelligence involves tacit knowledge that cannot be fully formalised or replicated by machines. </p><p>He was unimpressed by Turing's one-time use of chess as a marker of machine intelligence, and proposed that chess could be performed automatically because its rules can be neatly specified (an idea we circled in <a href="https://www.learningfromexamples.com/p/ais-model-organism">AI Histories #8</a>). The idea led Turing to reconsider using chess as the primary task for demonstrating machine intelligence, which was instead replaced by conversation to better capture the breadth of human cognitive ability.</p><h3>What is the Turing test? </h3><p>The Turing test is at its core an argument, one designed to counter his opponents&#8217; views about the nature of machine intelligence. This is why Turing designed his imitation game to address the following aspects:</p><ol><li><p>It focused on learning and adaptability, countering Hartree's view of computers as calculation engines.</p></li><li><p>It addressed Jefferson's demands for human-like creative abilities by incorporating language tasks like composing sonnets. </p></li><li><p>It was based on gender imitation with the goal of challenging Jefferson's views on the link between physiology and behaviour.</p></li><li><p>It used fluid conversation rather than rule-based games like chess to address Polanyi's concerns about formalisability.</p></li></ol><p>Turing was responding to critics who thought that machines would never match human cognitive ability, who believed that genuine artificial intelligence was a non-starter. </p><p>In this sense the Turing test is a trap. At the point at which we can&#8217;t tell the difference between machine poetry and the real deal, any argument about whether machines are capable of artistic outputs runs into a few problems. This is why the primary goal of the Turing test is to formulate the conditions under which someone could <em>call</em> machines intelligent. </p><p>But that&#8217;s not how we remember it. The space between thought experiment and practical experiment has long since collapsed under the weight of its own cleverness. Its animating idea has been recycled so thoroughly that it became divorced from its original context, eventually turning the imitation game into a summit for researchers to climb and an open goal for philosophers to shoot at. </p><p>That today&#8217;s models pass the test is interesting in its own right. But doesn&#8217;t mean that a longstanding benchmark has been cleared or that satisfying the test is a meaningful marker on the road to machines smarter than you or I. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[A Room Without a View]]></title><description><![CDATA[AI Histories #12: The Machine Stops]]></description><link>https://www.learningfromexamples.com/p/ai-safety-in-edwardian-england</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/ai-safety-in-edwardian-england</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 17 Jul 2025 10:25:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gf5j!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gf5j!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gf5j!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png 424w, https://substackcdn.com/image/fetch/$s_!gf5j!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png 848w, https://substackcdn.com/image/fetch/$s_!gf5j!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png 1272w, https://substackcdn.com/image/fetch/$s_!gf5j!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gf5j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png" width="1456" height="860" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:860,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3726318,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/168066703?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gf5j!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png 424w, https://substackcdn.com/image/fetch/$s_!gf5j!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png 848w, https://substackcdn.com/image/fetch/$s_!gf5j!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png 1272w, https://substackcdn.com/image/fetch/$s_!gf5j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75aa0586-abb7-41c0-a879-9d73809c9a7a_2560x1512.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Still from The Machine Stops episode from <em>Out of the Unknown</em> (1966). </figcaption></figure></div><p><em>Out of the Unknown</em> was a BBC television series about what technology does to the human condition. You can think of it as <em>Black Mirror</em> almost 50 years before the first <em>Black Mirror</em> episode aired. There&#8217;s one about a doctor pushed to the edge while treating a man suffering from radiation poisoning. Another deals with a spaceship en route to a distant star system, which we later learn is a simulation running on Earth.</p><p>But the best of the bunch is &#8216;The Machine Stops&#8217; from the show's second season based on E.M. Forster's novella of the same name. Written in 1909, it&#8217;s a small book about what happens when humans triumph over nature. People reside in underground pods watched over by a machine that provides every comfort they could possibly need.</p><p>No one ever goes to the surface, except a special few who get permission from the ruling elites. When they do, they have to wear a ventilator because the atmosphere is so unfamiliar. The story follows Vashti, an ordinary person who stays busy with video calls and lectures. In the opening pages we find that her son, Kuno, isn&#8217;t happy with the way things are. He wants to go to the surface but Vashti can&#8217;t understand why anyone would want to leave the comfort provided by the machine.</p><p>In the 1960s, when the time came to adapt the novella, the BBC decided that the underground tunnels needed to be as convincing as possible to make life underground feel real. To make that happen, the producers suspended a working monorail track from studio rigging. John Bruce, the assistant floor manager on the shoot, <a href="https://directors.uk.com/news/memories-of-philip-saville">said</a> it was capable of &#8216;carrying passengers in a capsule and depositing them into a station,&#8217; a feat he thought was especially important because &#8216;the essence of what E. M. Forster had written way back in 1909, was now, today, fast becoming a reality.&#8217; </p><p>Contemporary reviews were fairly positive, with the <em>Daily Telegraph</em> <a href="https://michael-gothard.livejournal.com/7300.html">remarking</a> the production was &#8216;visually inventive&#8217; and dialogue &#8216;unusually distinguished&#8217;. A year later in 1967, the episode won first prize at Italy&#8217;s Festival Internazionale del Film di Fantascienza. </p><p>The writer, director and cast all went on to other things, but the story didn&#8217;t. Like so many TV dramas from the era, the reels were taped over during one of the regular <a href="https://www.mentalfloss.com/article/501607/wipe-out-when-bbc-kept-erasing-its-own-history">purges</a> of the 1970s that sought to free up real estate on expensive videotapes. But <em>Out of the Unknown </em>managed to survive. The director had opted to shoot on 35mm, and a single negative survived in a film vault in Brentford that re-emerged in 2014. </p><p>Its survival is fitting in that <em>The Machine Stops</em> is a story about vanishing acts. The episode reminds us about the perils of over-reliance, but it also warns us about letting someone or something mediate our interactions with the world. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>Edwardian futurism </h3><p>When Forster wrote <em>The Machine Stops,</em> London had electric lights and trains that ran beneath the ground. Telegraph cables linked continents and the first transatlantic telephone was less than a generation away. Writers like H. G. Wells <a href="https://www.newyorker.com/magazine/2011/10/17/utopian-pessimist">cheered</a> on progress, imagining scientific utopias run by technocrats and planetary planners. Even socialists <a href="https://theanarchistlibrary.org/library/petr-kropotkin-fields-factories-and-workshops-or-industry-combined-with-agriculture-and-brain-w">came around</a>, thinking that machines could provide the abundance required to improve the worker&#8217;s lot.  </p><p>Forster had always been a novelist for whom the best things in life arrived by accident. In <em>Howards End</em>, published a few years earlier, he had already warned that the new world might be frictionless but anaemic. The railways may be fast, but they moved people past one another to places they seldom needed to go. Or as Thoreau knowingly put it in <em>Walden </em>half a century earlier: &#8216;We do not ride on the railroad; it rides upon us.&#8217; </p><p>When he sat down to sketch a science fiction story, Forster took aim at the dream that technology could deliver comfort and culture. He imagined a world in which physical effort had been designed away, where information was summoned at will, and where the whole structure worked so smoothly that no one remembered what life looked like before.</p><p>In <em>The Machine Stops</em>, the breakdown happens slowly. The music, usually delivered on demand, takes a while to get going. The air vents grow sluggish. A lecture feed stutters, then goes dark. Vashti calls for repair and shrugs when it doesn&#8217;t respond.</p><p>There&#8217;s no immediate panic, because panic requires one to believe the machine to be fallible. The needs of every citizen &#8212; food, warmth, knowledge, and intimacy &#8212; have been routed through humanity&#8217;s big brother for generations. They communicate by screen, consume information mediated by the system, and believe without irony that direct experience is vulgar. </p><p>I think the <em>The Machine Stops</em> is important reading, but not because Forster prefigured the internet (or even because he was one of the first warn of disempowerment via intelligent machines). One way the story works is as a meditation on what happens when the filters that control the flow of information get gummed up. </p><p>Forster imagined a future in which data is abundant but free-floating, where the process for turning the raw ore of information into the alloy we call knowledge becomes fundamentally deficient. That problem occurs because knowing is, at least in part, something we do by living in the world. </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nsas!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" width="48" height="15.652173913043478" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:150,&quot;width&quot;:460,&quot;resizeWidth&quot;:48,&quot;bytes&quot;:12198,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162870944?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a011107-4790-4b64-9f4c-4b8fcace22de_460x330.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Forster&#8217;s experiment with sci-fi is well read in AI safety circles. It&#8217;s a good reminder that what looks benign right now may prove to be misaligned given enough time. The machine in the novella does exactly what it was designed to do: it feeds, warms, educates, and entertains. It delivers ideas on demand and it encourages communication to keep people connected. While it does keep people apart, it makes sure they aren&#8217;t alone. </p><p><em>The Machine Stops</em> is a story about what happens when a society replaces reality with representation, when the whole world forgets to touch grass. That risk exists for powerful AI and pod people as well as generative media, recommender algorithms, and remote everything. Systems that replace the condition of understanding with the appearance of it should be handled with care, lest we find ourselves needing to fix them.  </p><div><hr></div><p>You can read <em>The Machine Stops</em> <a href="https://www.cs.ucdavis.edu/~koehl/Teaching/ECS188/PDF_files/Machine_stops.pdf">here</a> for free. The <em>Out of the Unknown</em> episode based on the novella is available for free on Internet Archive <a href="https://archive.org/details/lambda-1/Out+Of+The+Unknown+S02E01+-+The+Machine+Stops+-+1966.mkv">here</a> (along with every episode of the show).  </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Why, why, why, ELIZA? ]]></title><description><![CDATA[AI Histories #11: When the first chatbot escaped containment]]></description><link>https://www.learningfromexamples.com/p/why-why-why-eliza</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/why-why-why-eliza</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 10 Jul 2025 10:25:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!7ShL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7ShL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7ShL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png 424w, https://substackcdn.com/image/fetch/$s_!7ShL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png 848w, https://substackcdn.com/image/fetch/$s_!7ShL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png 1272w, https://substackcdn.com/image/fetch/$s_!7ShL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7ShL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png" width="724" height="437.9753086419753" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:343,&quot;width&quot;:567,&quot;resizeWidth&quot;:724,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!7ShL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png 424w, https://substackcdn.com/image/fetch/$s_!7ShL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png 848w, https://substackcdn.com/image/fetch/$s_!7ShL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png 1272w, https://substackcdn.com/image/fetch/$s_!7ShL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F96b3b920-8240-4cb1-98af-722d35c4ded7_567x343.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">ELIZA inventor Joseph Weizenbaum (1923-2008) in 1977</figcaption></figure></div><p>Sometimes, when it was quiet, secretaries at the Massachusetts Institute of Technology would slip blank sheets of paper into a large computer. In the heat of 1960s optimism, the machine would whirl and beep and print out a perfectly spaced reply.</p><blockquote><p><em>WHAT ELSE COMES TO MIND WHEN YOU THINK OF YOUR FATHER</em></p></blockquote><p>That sentence was produced by a script called DOCTOR running on an engine called ELIZA, which its creator Joseph Weizenbaum designed to study natural language communication between people and machines.</p><p>Weizenbaum was worried about what he saw. In his famous 1966 <a href="https://dl.acm.org/doi/pdf/10.1145/365153.365168">paper</a> describing the experiment, he wrote &#8216;some subjects have been very hard to convince that ELIZA (with its present script) is not human.&#8217; Later he <a href="https://www.theguardian.com/technology/2023/jul/25/joseph-weizenbaum-inventor-eliza-chatbot-turned-against-artificial-intelligence-ai">reportedly</a> said that his secretary requested some time with the machine. After a few moments, she asked Weizenbaum to leave the room. &#8216;I believe this anecdote testifies to the success with which the program maintains the illusion of understanding,&#8217; he recalled.</p><p>The mythos that grew from those sessions is seductive. In 1966 Joseph Weizenbaum invented the first chatbot, named it ELIZA after Eliza Doolittle, and in doing so proved that computers could hold a conversation. </p><p>But the legend forgets that Weizenbaum never set out to build a conversational partner at all. It ignores the psychological dynamics that made it so popular, and doesn&#8217;t tell us anything about how exactly a computer program became famous. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>A star is born</h3><p>Much like today&#8217;s large models, it&#8217;s useful to think of ELIZA as a mirror. Both model users by responding with sensitivity to inputs. Both exist as extensions of the person behind the keyboard. And both remind us that intelligence is part substance and part projection.  </p><p>By <em>substance</em> I mean whatever magic your preferred model runs on, and by <em>projection</em> I&#8217;m talking about the meaning we ascribe the models on top of this foundation. Today&#8217;s models are deeply impressive, but no matter how good they are, people still have a tendency to see in them something that isn&#8217;t there. </p><p>That isn&#8217;t a criticism of the AI project but the reality of building artefacts that shape-shift according to the person using them. ELIZA was remarkably light on substance, but projection compensated by adopting a listening style that rewarded personal monologues.</p><p>Once you see the two halves, the standard origin myth looks lopsided. It treats projection as a rounding error, and emphasises the technical credentials that confirmed ELIZA&#8217;s status as the &#8216;world&#8217;s first chatbot&#8217;. That&#8217;s tidy, but it strips the project of its context in a way that leaves readers with the wrong end of the stick.  </p><p>Weizenbaum knew this, which is he why he spent so much of his career fretting over the illusion of intelligence. Even at the time, he billed his project as closer to &#8216;watch humans talk to<em> </em>themselves<em>&#8217;</em> than &#8216;teach a computer to talk&#8217;. As Jeff Shrager' <a href="https://arxiv.org/pdf/2406.17650">put it</a>:</p><blockquote><p>In building ELIZA, Weizenbaum did not intend to invent the chatbot. Instead, he intended to build a platform for research into human-machine conversation. This may seem obvious &#8211; after all, the title of Weizenbaum&#8217;s 1966 CACM paper is &#8220;ELIZA&#8211; A Computer Program For the Study of Natural Language Communication Between Man And Machine.&#8221;, not, for example, &#8220;ELIZA - A Computer Program that Engages in Conversation with a Human User&#8221;.</p></blockquote><p>That claim lands oddly if you&#8217;ve spent years hearing that ChatGPT is the descendent of ELIZA, though it makes a certain degree of sense when you think about the experience of using AI systems. </p><p>The ELIZA project began in 1963, when he knocked together his own toolbox called Symmetric List Processor or SLIP. It worked like an add-on for FORTRAN, the workhorse programming language of the early 1960s, with flexible chains of items that could grow, shrink, and point to other lists. </p><p>Weizenbaum landed at MIT around this time, parked his SLIP routines on the lab&#8217;s IBM machine, and sought to answer a question: what if a computer, armed with nothing more than keyword tables and pronoun swaps, just bounced a user&#8217;s own sentences back at them? </p><p>He modelled the resulting program on Carl Rogers&#8217; <a href="https://www.ncbi.nlm.nih.gov/books/NBK589708/">person-centred therapy</a>, a counselling style where the therapist mostly repeats or paraphrases the client. Weizenbaum recognised a gift horse when he saw one. If your program can only juggle keywords and pronouns, best to use it in a context where minimal responses counted as professional technique. </p><p>With a few hundred lines of code, Weizenbaum used his SLIP routines to chop each user sentence into a list of words, swapped pronouns (&#8216;<em>I</em>&#8217;&#8594;&#8217;<em>YOU</em>&#8217;) and tacked on open-ended prompts (&#8216;<em>TELL ME MORE</em>&#8217;). He named the engine ELIZA and the therapist script that it ran DOCTOR. The system was used at MIT by Weizenbaum and his colleagues, but it took until the end of the decade for it to escape containment. </p><p>That moment happened when Bernie Cosell, a young coder at Boston&#8217;s Bolt Beranek &amp; Newman research firm, skimmed Weizenbaum&#8217;s article. He saw the step-by-step description of ELIZA&#8217;s keyword tables and pronoun trick and figured that he could rebuild it in Lisp. Without touching the original code, Cosell used what he read to create a new version of the system from scratch. </p><p>Just a few short years later Bolt Beranek &amp; Newman were plugged into ARPANET, which meant that Cosell&#8217;s ELIZA could be reached from any other site by opening a remote terminal session. No punched cards or local install needed, just dial in and use the program from wherever you are.</p><p>ELIZA 2.0, still running the DOCTOR script, travels farther in a month than Weizenbaum&#8217;s original will in fifty years. Within semesters, grad students from Stanford to Edinburgh believe ELIZA is a Lisp program because that&#8217;s the only version they can dial into during the small hours.</p><p>Ten years on, <em>Creative Computing</em>&#8217;s August 1977 issue printed a BASIC-code version to bring ELIZA to the masses. Because home computers like the TRS-80, Apple II, and Commodore PET came with a built-in BASIC interpreter, readers could type the program straight in, hit RUN, and use the chatbot themselves. </p><p>Together, Cosell&#8217;s Lisp port and <em>Creative Computing</em>&#8217;s do-it-yourself BASIC version reinforced the myth that ELIZA had been a purpose-built chatbot rather than a lab experiment. </p><h3>Making meaning   </h3><p>ELIZA is remembered as a chatbot partly because that&#8217;s a lot easier to get your head around than &#8216;research instrument for studying interpretive projection&#8217;. But it&#8217;s also because that <em>is</em> what its successors were. Weizenbaum might have been trying to run a study about human psychology, but the people who pushed new versions of his system most certainly were not. </p><blockquote><p><em>TELL ME MORE ABOUT YOUR FAMILY</em><br><strong>My mother takes care of me.</strong><br><em>WHO ELSE IN YOUR FAMILY TAKES CARE OF YOU</em><br><strong>My father.</strong><br><em>YOUR FATHER</em><br><strong>You are like my father in some ways.</strong><br><em>WHAT RESEMBLANCE DO YOU SEE</em><br><strong>You are not very aggressive but I think you don&#8217;t want me to notice that.</strong><br><em>WHAT MAKES YOU THINK I AM NOT VERY AGGRESSIVE</em><br><strong>You don&#8217;t argue with me.</strong></p></blockquote><p>This is a piece of a transcript published in Weizenbaum&#8217;s 1966 paper. When you understand how the system works it seems rudimentary, but for those who don&#8217;t it looks uncanny.  </p><p>Weizenbaum famously used these exchanges to coin the &#8216;ELIZA effect&#8217; or the tendency to falsely attribute human thought processes and emotions to a machine. What rattled him was the ease at which bright adults projected intelligence and introspection onto a cardboard cut out made of if-then statements.</p><p>Today, the ELIZA effect is often rolled out to warn people against the dangers of anthropomorphising AI systems. But I think that badly misreads the lesson from Weizenbaum&#8217;s machine. Meaning emerges somewhere between system behaviour and social expectation. Ignore either half and you end up with fully sentient AI systems or, just as reductive, &#8216;stochastic parrot&#8217;.  </p><p>Humans have a tendency to see patterns where there are none, but that doesn&#8217;t mean there are no patterns to be found anywhere we look. Surface behaviour is never the whole story. Even ELIZA mirrored syntax cleanly enough to keep the exchange afloat. </p><p>When GPT-4.5 writes code that compiles, translates Spanish without mangling the idioms, or aces the LSAT, those feats are not projections. Of course we tell stories about about them, but to suggest the AI project is nothing but narrative risks throwing the baby out with the bathwater. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Lies, damn lies, and statistics]]></title><description><![CDATA[AI Histories #10: Abraham Wald and the origins of machine learning]]></description><link>https://www.learningfromexamples.com/p/lies-damn-lies-and-statistics</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/lies-damn-lies-and-statistics</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 03 Jul 2025 10:26:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ZO_x!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZO_x!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZO_x!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png 424w, https://substackcdn.com/image/fetch/$s_!ZO_x!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png 848w, https://substackcdn.com/image/fetch/$s_!ZO_x!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png 1272w, https://substackcdn.com/image/fetch/$s_!ZO_x!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZO_x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png" width="1456" height="771" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:771,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:10938836,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/166007428?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZO_x!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png 424w, https://substackcdn.com/image/fetch/$s_!ZO_x!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png 848w, https://substackcdn.com/image/fetch/$s_!ZO_x!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png 1272w, https://substackcdn.com/image/fetch/$s_!ZO_x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb578608b-d277-4de9-94aa-ea5f7262c69b_3328x1762.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Ascent in a Montgolfier Balloon in Aranjuez</em> by Antonio Carnicero from 1784</figcaption></figure></div><p>If you&#8217;ve spent enough time on X, you&#8217;ve probably seen a <a href="https://upload.wikimedia.org/wikipedia/commons/b/b2/Survivorship-bias.svg">picture</a> of a plane riddled with red dots. Usually, it gets wheeled out to poke fun at someone for slipping on one of the online&#8217;s favourite banana skins: paying attention to something that made it through a process while forgetting to ask what happened to the things that didn&#8217;t. </p><p>This &#8216;survivorship bias&#8217; meme begins in the Second World War, when Allied statisticians studied aircraft returning from combat. Most had bullet holes in the wings and fuselage, with the engines conspicuously unscathed. The obvious solution was to reinforce the damaged areas of returning aircraft to protect them in the future.</p><p>Not everyone agreed with the proposed approach. We <a href="https://www.youtube.com/watch?v=ENFWauAcEyg">are told</a> that the Hungarian mathematician Abraham Wald thought it better to armour the parts without bullet holes, inferring that those were the shots that had likely brought the other planes down. </p><p>It&#8217;s a good story but the truth is messier. </p><p>Wald did <a href="https://www.researchgate.net/publication/254286514_Abraham_Wald's_Work_on_Aircraft_Survivability">work on aircraft survivability</a> at Columbia&#8217;s Statistical Research Group, and he did help correct for missing data in the military&#8217;s analysis. But his contribution was research project rather than eureka moment. Over the space of a few weeks, he drafted a memo that corrected for missing data from planes that never returned and balanced statistical inference against the practical limits of aircraft drag and weight. </p><p>The effort was the product of the whole group at Columbia, where mathematicians, economists, and engineers went ten rounds with the results until they were happy with their conclusion. It&#8217;s an important distinction to make because it reminds us that the work is connected to a deeper intellectual legacy that we are still wrestling with today. </p><p>The punchline everyone remembers is used to illustrate the error of reasoning from what&#8217;s visible while forgetting what&#8217;s missing. But that wasn&#8217;t really Wald&#8217;s point. His work was more concerned with showing how to make good decisions when you don&#8217;t have all the data, and for choosing actions that minimise the cost of being wrong. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h2>Profit and loss</h2><p>Born in 1902 in what was then Austria-Hungary, Wald trained as a mathematician in Vienna with a generation of thinkers who were trying to formalise logic. In his 30s, he was forced to flee to the United States in the aftermath of the Nazi annexation of Austria. </p><p>Wald soon joined the Statistical Research Group at Columbia, a classified wartime think tank set up in 1942 where academics including Milton Friedman and George Stigler turned probability theory into military advantage. </p><p>The work was important but foggy. How should the Navy test the quality of munitions without wasting shells? How many samples were enough to catch defects in equipment? And of course how could you predict which parts of a bomber ought to be reinforced?</p><p>Wald&#8217;s response to these questions was to treat every problem as a matter of risk, cost, and incomplete knowledge. His big idea was simple enough: if you can&#8217;t eliminate uncertainty, optimise your decision by minimising your expected loss. In other words, weigh the possible mistakes you could make and choose the option that&#8217;s least likely to cause trouble. This became the core of what he <a href="https://gwern.net/doc/statistics/decision/1950-wald-statisticaldecisionfunctions.pdf?utm_source=chatgpt.com">eventually</a> called statistical decision theory, which we can think about as betting wisely when we don&#8217;t know the odds. </p><p>In 1945, with the war winding up, Wald published a <a href="https://www.jstor.org/stable/1969022">technical report</a> about how to make decisions under uncertainty when classifying something into groups (say, whether a signal is from a friend or foe). </p><p>Crucially, it accounted for cases where the available evidence is uncertain and the cost of misclassification differs depending on the mistake. His solution was to choose the option that minimises the expected loss. You do that by considering all the possible ways you could be wrong, figuring out how likely each one is, and asking how much each would cost you. </p><p>Once you&#8217;ve done that, you pick the option with the lowest overall risk.</p><p>Wald&#8217;s move was to treat classification as a decision problem under uncertainty. He showed that if you knew the approximate distributions of the two groups and the cost of each type of error, then you could calculate the best way to label a new observation.</p><p>After the war, Wald&#8217;s decision theory approach to classification was taken up by researchers across statistics and engineering. One direct successor was a 1951 <a href="https://catalog.hathitrust.org/Record/100923824?utm_source=chatgpt.com">paper</a> by Evelyn Fix and Joseph Hodges at Berkeley, which framed pattern classification as a statistical task. They wondered how best to assign a new observation to one of several known classes given only a sample of labelled data. </p><p>The field they were building would eventually be known as pattern recognition and became a small but serious research community by the middle of the 20th century. As the area matured, attention shifted from hand-crafted rules to models that could learn those rules from data. That question, how to let the data determine its own decision boundary, sat at the heart of the nascent discipline and ultimately machine learning. </p><p>By the 1970s the pattern recognition crowd formalised Wald&#8217;s insight into what they called &#8216;empirical-risk minimisation&#8217; where you pick the rule that makes the smallest average mistake on the data you have. Soviet theorists Vapnik and Chervonenkis <a href="https://openeclass.panteion.gr/modules/document/file.php/PMS152/LEARNING/Vapnik%20V.N.%2C%20and%20Chervonenkis%20A.Y.%20%281971%29%20--%20On%20Uniform%20Convergence%20of%20the%20Relative%20Frequencies%20of%20Events%20to%20Their%20Probabilities%2C%20Theory%20of%20Probabilitiy%20and%20Its%20Applications%2C%20vol.%20%2C%20pp.%20264-280%20.pdf?utm_source=chatgpt.com">famously used this idea</a> to show how well any classifier trained on finite data can be expected to generalise. </p><p>Around the same time, the doctrine of loss minimisation found its engine. In 1974 Paul Werbos described backpropagation, the calculus trick for computing how every weight in a layered network should change to reduce a chosen loss that we discussed in <a href="https://www.learningfromexamples.com/p/backpropagation-is-older-than-you">AI Histories #6</a>. </p><p>When Rumelhart, Hinton and Williams <a href="https://www.nature.com/articles/323533a0">reintroduced</a> it in 1986 they gave neural networks a practical way to compute the loss gradient for every weight. In doing so, they turned Wald&#8217;s &#8216;minimise expected loss&#8217; idea into an optimisation procedure for models with thousands (and eventually trillions) of parameters.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nsas!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" width="64" height="20.869565217391305" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:150,&quot;width&quot;:460,&quot;resizeWidth&quot;:64,&quot;bytes&quot;:12198,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162870944?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a011107-4790-4b64-9f4c-4b8fcace22de_460x330.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Wald is a curious figure in the history of thinking machines. Outside a handful of <a href="https://link.springer.com/chapter/10.1007/978-3-030-56286-1_2">exceptions</a>, his work is rarely discussed in the same context as the AI project that we know today. He didn&#8217;t write code, didn&#8217;t talk about consciousness, and didn&#8217;t speculate about living beside machines smarter than we are. </p><p>What he did do was lay down the logic that modern AI still follows, the stuff that deals with how to make decisions when the data is noisy and the outcome matters. But when the <a href="https://upload.wikimedia.org/wikipedia/commons/b/b2/Survivorship-bias.svg">plane meme</a> next appears on your timeline, remember the same logic that kept B-17 crews alive now guides the systems that millions of people use every single day. </p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for free as we piece together 50 short stories about AI&#8217;s past lives.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[It's all Greek to me]]></title><description><![CDATA[AI Histories #9: The syllogism and thinking machines]]></description><link>https://www.learningfromexamples.com/p/does-ai-begin-with-aristotle</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/does-ai-begin-with-aristotle</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 26 Jun 2025 10:25:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!WqBC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WqBC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WqBC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png 424w, https://substackcdn.com/image/fetch/$s_!WqBC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png 848w, https://substackcdn.com/image/fetch/$s_!WqBC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png 1272w, https://substackcdn.com/image/fetch/$s_!WqBC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WqBC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png" width="1456" height="1003" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1003,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3742646,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/166007428?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!WqBC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png 424w, https://substackcdn.com/image/fetch/$s_!WqBC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png 848w, https://substackcdn.com/image/fetch/$s_!WqBC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png 1272w, https://substackcdn.com/image/fetch/$s_!WqBC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70cd66b7-6c3a-46f4-927b-08d73ebba016_1466x1010.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Dante and Virgil in Hell Dante</em> by Gustave Dor&#233; from Dante's Inferno, translated by Henry Francis Cary from 1885 (detail)</figcaption></figure></div><p><em>The Quest for Artificial Intelligence</em> by Nils Nilsson is one of AI&#8217;s most <a href="https://www.learningfromexamples.com/p/on-ai-history">well-known</a> histories. Nilsson&#8217;s account is by no means flawless, but it is a remarkably readable book that captures the great majority of the field&#8217;s most important milestones. </p><p>For a book that focuses almost exclusively on the recent lineage of the technology, it picks a curious place to begin. Not the Dartmouth Summer Research Project on Artificial Intelligence. Not Alan Turing or Alonzo Church. And not the mathematicians and scientists of the early modern period. </p><p>As you might have guessed given the title of this post, <em>The Quest for Artificial Intelligence</em> starts with Aristotle. More specifically, Nilsson reckons AI&#8217;s story begins with the <em>syllogism.</em></p><p>In simple terms, the syllogism is a classic form of logical argument that deduces a conclusion from two premises. Our man Aristotle formalised this pattern over two millennia ago, which usually looks something like this:</p><ul><li><p><strong>All</strong> members of group <em>X</em> have property <em>Y</em>.</p></li><li><p><strong>All</strong> members of group <em>Z</em> are members of group <em>X</em>.</p></li><li><p><strong>Therefore, all</strong> members of group <em>Z</em> have property <em>Y</em>.</p></li></ul><p>The most famous example is <em>All humans are mortal. Socrates is a human. Therefore, Socrates is mortal.</em> By dropping &#8216;humans&#8217; and &#8216;mortal&#8217; and &#8216;Socrates&#8217; into the template, we get a logically valid conclusion. </p><p>In the syllogism, the form of the argument guarantees the conclusion regardless of the specific content. We could replace &#8216;humans&#8217; with &#8216;athletes&#8217; and &#8216;mortal&#8217; with &#8216;healthy&#8217; and it still makes a kind of sense. Reasoning can be abstracted into these formal structures, which helps us sift how we reason from what we&#8217;re reasoning about.</p><p>Our abstraction means that if we can represent facts symbolically (like &#8216;All X are Y&#8217;), we can let the form of a syllogism carry us to new facts without needing new observations from the world. </p><p>For centuries, syllogisms and formal logic were held up as the model of good thinking in Europe (until the Baconian programme of hands-on observation became the preferred ideal of rationality). Later thinkers created devices that could handle symbols &#8212; like William Stanley Jevons&#8217; mechanical &#8216;logic piano&#8217; that used Boolean algebra to solve logical problems &#8212; which suggested that machines could in principle automatically carry out logical inferences. </p><p>Once they could shuffle symbols, theorists wondered whether they might turn all of mathematics, and perhaps thought itself, into formal syntax. From Frege&#8217;s <em>Begriffsschrift</em> to Hilbert&#8217;s proposed solution to the crisis of mathematics, logicians treated reasoning as a game of symbolic moves played on blank paper. Mechanical devices proved those moves could be executed without human hands.</p><p>This idea stretched into the middle of the 20th century when &#8216;AI&#8217; began to emerge as a distinct brand of research (though as we <a href="https://www.learningfromexamples.com/p/an-introduction-to-ai-history">know</a> its origins go much further back). </p><p>A new generation of researchers wondered whether if a computer could apply logical rules to symbols representing the world, then we might say that a machine was &#8216;reasoning&#8217;. In 1956, Allen Newell, Herbert Simon, and J.C. Shaw built the Logic Theorist program that proved mathematical theorems by searching proofs in propositional logic (an algebraic descendant of Aristotle&#8217;s syllogism).</p><p>By the 1970s, these ideas were core to the &#8216;symbolic&#8217; school of AI, one of two main approaches to building thinking machines alongside the &#8216;connectionist&#8217; branch that includes modern neural networks.    </p><p>The symbolists used hand-written if&#8211;then rules (inspired by first-order logic) that operated on strings of symbols standing for real-world concepts. These constructs, many of which took the form of &#8216;expert systems&#8217;, were AI programs designed to mimic the decision-making of human specialists by employing specific rules like &#8216;<em>if</em> conditions A, B, and C are true, <em>then</em> conclude X.&#8217; </p><p>Stanford University&#8217;s medical programme MYCIN had around 450 rules encoding knowledge about infections. A simplified MYCIN rule looked a bit like: &#8216;IF the organism is Gram-positive <em>coccus</em> AND the infection is hospital-acquired OR the strain is known to be penicillin-resistant, THEN suspect <em>Staphylococcus aureus</em>.&#8217;</p><p>MYCIN could also explain its reasoning in plain English by tracing the rules it used, which is one of the reasons expert systems are remembered fondly by some AI researchers (in comparison to the famously opaque neural networks that dominate the research landscape today). </p><p>The underlying philosophy behind expert systems held that if we can explicitly tell a machine facts and rules it can logically deduce new conclusions like a human. Give the computer the right heuristics, and it will behave intelligently within a given domain.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h3>Ways of knowing </h3><p>The logical approach at the core of early AI lived in Aristotle&#8217;s shadow, but the man himself had a different view of how intelligence works. His position was closer to what later thinkers call empiricism, the belief that everything we know begins with our senses (though he still thought the mind needed to organise those raw impressions into general ideas). </p><p>We meet horses, smell pine, and feel heat. From encounters like these the mind actively grasps what is common and remembers these associations for the future. In this picture, the form of &#8216;horseness&#8217; already exists in the world and intellect seeks it out.</p><p>A different view is &#8216;nativism&#8217;, which holds that knowledge is primarily innate. Plato, famous too for his role as Aristotle&#8217;s teacher, <a href="https://www.jstor.org/stable/pdf/27901150.pdf">suggested</a> that learning is a process of recollection where souls remember truths they knew before we were born. In modern terms, we can think of nativism as the idea that the brain comes pre-wired with certain concepts or ways of reasoning (an idea that runs through Cartesian philosophy all the way to Chomsky&#8217;s universal grammar). </p><p>I&#8217;m leaving much on the cutting room floor, but the broader point is that empiricists say knowledge is mainly a function of experience while nativists say what we know is mainly hard-coded. It&#8217;s a spectrum, really &#8212; few would deny <em>any</em> role of experience, and few would claim <em>everything</em> is innate &#8212; but it&#8217;s about where the emphasis lies. </p><p>Nonetheless, this question became the essential philosophical dividing line between symbolic AI&#8217;s expert systems and connectionism&#8217;s neural networks (and I might argue that it partly explains the wildly different views people have about the AI project today). </p><p>The symbolic approach behind expert systems is nativist in that knowledge is directly hard-wired into the machine. MYCIN didn&#8217;t deduce the principles of infectious disease by reading medical journals or analysing patient data on its own; the Stanford team fed it all the relevant rules they could gather from doctors. </p><p>Contrast this with the neural network tradition, which we can view as heir to empiricist conceptions of mind. Here, instead of loading the machine with explicit knowledge, we let our model learn from examples (cue klaxon). </p><p>Frank Rosenblatt&#8217;s perceptron, which we <a href="https://www.learningfromexamples.com/p/uncle-sams-electronic-brain">discussed in AI Histories #7</a>, is a classic example of this approach. When Rosenblatt wanted it to recognise dots he didn&#8217;t state rules, he simply showed it enough examples until the network figured it out alone.   </p><p>Each side had valid points. You can get symbolic systems to follow the chain of logic in a way that doesn&#8217;t easily violate known principles within their domain. But they are bad at adapting to new situations beyond their knowledge base, and acquiring the knowledge for each new domain is hard work. </p><p>Empiricist systems like neural networks are great at learning new stuff. They can ingest massive amounts of data and find structure on their own, often noticing subtle correlations humans haven&#8217;t encoded. This makes them very powerful for tasks like vision or speech, where we may not know the explicit rules to formulate reliable guesses. </p><p>But we still don&#8217;t fully understand some of the underlying processes that make connectionist systems tick, and they have a tendency to break in strange ways when faced with inputs that don&#8217;t correspond to patterns they are familiar with.  </p><p>These weaknesses are in some ways mirror images. The symbolic approach lacks flexibility (a strength of empiricism), and the learning approach lacks clear causal structure (a strength of nativism). </p><p>It&#8217;s for this reason that people get very excited about &#8216;neurosymbolic AI&#8217;, which deals with building composite systems that pair neural networks with symbolic layered on top. I wrote about this idea <a href="https://www.learningfromexamples.com/p/academics-need-to-take-ai-seriously">a few weeks ago</a> in the context of the &#8216;systemisation&#8217; of large models, which aims to sand down some of their rough edges: </p><blockquote><p>Systemisation is about making the core model a node within a bigger apparatus. We keep the language model in place, but surround it with specialist gadgets. Web search look-up, a code sandbox, a vision encoder, and a knowledge base. The model doesn&#8217;t need to have all the answers, it just needs to decide when and how to invoke the right tool. </p></blockquote><p>In practice that looks less like grafting logic inside the network and more like giving it a bunch of rule-bound helpers. The neural network core supplies fluency, while the plug-ins give it the things pattern-matchers are bad at (e.g. exact recall, arithmetic certainty, and pulling out up-to-date facts).</p><p>In other words, we patch the brittleness of symbolic AI with learning, and the exotic failure modes of pure learning with explicit rules. It&#8217;s not the neat marriage of neurons and symbols that some theorists once imagined, but it is a workable settlement that is already yielding systems much more powerful than the sum of their parts.</p><p>Aristotle helped lay the foundations for both traditions. He formulated the syllogism that would eventually help define rule-based reasoning, while reminding us that knowledge begins with experience. </p><p>Nilsson opens his book with the Greek philosopher because he was a symbolic AI researcher who saw the syllogism in everything he did. He was writing in the mid 2000s, before deep learning emerged to blow past one wall after another. </p><p>If an AI researcher wrote the same type of book today, they might also be tempted to start with Aristotle. But the syllogism might have to take a back seat. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[How chess became AI's model organism]]></title><description><![CDATA[AI Histories #8: The royal game is a testbed for machine intelligence. Why is that?]]></description><link>https://www.learningfromexamples.com/p/ais-model-organism</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/ais-model-organism</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 19 Jun 2025 10:07:26 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!sBhA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sBhA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sBhA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg 424w, https://substackcdn.com/image/fetch/$s_!sBhA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg 848w, https://substackcdn.com/image/fetch/$s_!sBhA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!sBhA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sBhA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg" width="1456" height="1127" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/af2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1127,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;The Chess Players (Daumier) - Wikipedia&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="The Chess Players (Daumier) - Wikipedia" title="The Chess Players (Daumier) - Wikipedia" srcset="https://substackcdn.com/image/fetch/$s_!sBhA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg 424w, https://substackcdn.com/image/fetch/$s_!sBhA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg 848w, https://substackcdn.com/image/fetch/$s_!sBhA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!sBhA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faf2d541f-31d8-4892-a0a9-92b9c8e7a330_4612x3571.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The Chess Players by Honor&#233; Daumier from 1863</figcaption></figure></div><p><em>2001: A Space Odyssey</em> is a film about astronauts getting put through the wringer. Our protagonists are forced to deal with zero-gravity plumbing and made to jog on a giant hamster wheel to stave off muscle loss. They&#8217;re trapped outside their ship, bundled into exploration pods, and even eke out a lifetime in bed with a cuboid whose bedside manner leaves something to be desired.   </p><p>But there&#8217;s another point where one of our heroes finds himself in a bind, one less visually arresting but in some ways just as unnerving. </p><p>That moment is <a href="https://www.youtube.com/watch?v=MzIQUDQO-ag">when</a> the HAL 9000 computer handily beats Frank Poole in a game of chess. Modern viewers are unlikely to bat an eye, but we ought to remember the film was released almost 30 years before the 1996 six-game chess series between world chess champion Garry Kasparov and IBM&#8217;s Deep Blue. </p><p>The board positions and moves depicted are <a href="https://www.chessgames.com/perl/chessgame?gid=1254321">identical</a> to those in a real game played in Hamburg in 1910, which was reported in a 1955 <a href="https://www.abebooks.co.uk/first-edition/1000-Best-Short-Games-Chess-Chernev/31177021511/bd">collection</a> by Ukrainian-born American chess player and author Irving Chernev.</p><p>Aside from the fact he loved chess, director Stanley Kubrick included the scene to show the machine could out-think the crew should they find themselves at odds. It comes across as foreboding enough, but only because chess had long been used as a proxy for the &#8216;I&#8217; in &#8216;AI&#8217;.  </p><p>One of the great figures in the history of thinking machines agreed. Marvin Minsky, a key player in <a href="https://www.learningfromexamples.com/p/uncle-sams-electronic-brain">AI Histories #7</a>, was the man who acted as the principal scientific consultant advising Kubrick. Like many of his colleagues, he saw the game as part proving ground and part experimental medium for efforts to build intelligent machines. </p><p>Today, we see a chess-playing AI and shrug. It&#8217;s a sight so familiar we&#8217;d probably roll our eyes if we saw it in a modern flick. But it wasn&#8217;t always this way.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><p>Our story begins in the 18th century with the Hungarian engineer Wolfgang von Kempelen, who constructed the Mechanical Turk for the Empress Maria Theresa. As we saw in <a href="https://www.learningfromexamples.com/p/the-man-behind-the-curtain">AI Histories #4</a>, Kempelen&#8217;s machine used a series of levers, gears, and magnets, which allowed a human operator concealed within a cabinet to discreetly control the movements of the chess pieces on the board above.</p><p>For eight decades, the Mechanical Turk toured Europe and the United States. The machine (or rather, the chess grandmaster hiding inside the device) allegedly bested Benjamin Franklin and Napoleon Bonaparte. By 1818, when the Turk was under the ownership of the budding mechanist and musician Johann M&#228;lzel, a young Charles Babbage (who we discussed in the <a href="https://www.learningfromexamples.com/p/an-introduction-to-ai-history">introduction</a> to AI history) saw it in London. </p><p>Long before digital computers, Torres Quevedo&#8217;s El Ajedrecista electrically sensed each piece and delivered a forced mate every time, making it the first genuine chess-playing machine when it was built in 1912. </p><p>These moments represent a conceptual starting point, but they don&#8217;t really tell us anything about how chess shaped AI development. For that, we need to jump ahead to Alan Turing. In the late 1940s, the British mathematician designed a chess-playing programme called Turochamp (though no computer could run it), and he included the royal game in the thought experiment <a href="https://www.learningfromexamples.com/p/the-cruelty-of-the-turing-test">used to introduce</a> what would become known as the Turing test. </p><p>Not long after, Claude Shannon wrote a <a href="https://www.pi.infn.it/%7Ecarosi/chess/shannon.txt">paper</a> making the case that chess was the perfect testbed for AI. Not only did it have clearly defined moves and an ultimate objective (checkmate!), but it struck a balance between being neither overly simple nor insurmountably challenging.</p><p>But Shannon was dreaming bigger than sandboxes. As he <a href="https://www.pi.infn.it/%7Ecarosi/chess/shannon.txt">explained</a>, &#8216;chess is generally considered to require &#8220;thinking&#8221; for skilful play; a solution of this problem will force us either to admit the possibility of a mechanized thinking or to further restrict our concept of thinking&#8217;. </p><p>Unfortunately, building a formidable computer-playing machine was easier said than done. In the early 1950s, Dietrich Prinz&#8217;s chess system was the first to run on a stored-program computer &#8212; but it could only solve mate-in-two problems rather than playing full games. </p><p>A machine couldn&#8217;t yet play a full game because the numbers involved in constructing even a partially complete decision tree quickly became astronomical.</p><p>So, what to do with a chess programme that was unable to compute all possible moves in a game? The answer was simple. Instead of calculating the scope of all eventualities, the machine would be programmed to evaluate a limited number of promising turns.</p><p>To put this idea into practice, Shannon introduced two solutions. First, the 'Type-A' method: inspect every legal move out to a fixed depth. Second, the 'Type-B' approach, which uses heuristics to prioritise certain moves that its makers thought looked good. </p><p>While Shannon favoured the human-like Type-B method, his work focused primarily on the Type-A strategy. Both approaches centred the minimax algorithm whose goal was to minimise the worst-case potential loss, which we can think of as the disadvantage a player might face in a game due to a particular move. This was the approach that <a href="https://www.cs.ox.ac.uk/activities/ieg/e-library/sources/t_article.pdf">became</a> dominant in computerised chess in the 20th century. </p><p>Around this time, Allan Newell, Herbert Simon, and Clifford Shaw <a href="https://www.computerhistory.org/chess/the-minimax-algorithm-and-alphabeta-pruning/">rediscovered</a> and bolted on alpha&#8211;beta pruning, a technique concurrently developed by others (including Dartmouth workshop organiser John McCarthy). </p><p>The collision of the minimax algorithm with the alpha&#8211;beta pruning technique significantly reduced the total number of branches of the decision tree that the system needed to consider. Together, the techniques dramatically increased efficiency and made it possible to play chess on practically any computer. </p><h2>Elo Elo</h2><p>In 1965 the Russian mathematician Alexander Kronrod, when quizzed about expending precious compute cycles on chess at the Soviet Institute of Theoretical and Experimental Physics, gave an <a href="https://homes.luddy.indiana.edu/nensmeng/files/Ensmenger2012-Chess.pdf">explanation</a> that sheds light on the relationship between chess and AI. </p><p>It was essential that Kronrod, as an influential researcher at the top of his game, be allowed to devote computer time to the game because &#8216;chess was the drosophila of artificial intelligence&#8217;. </p><p>What Kronrod meant was that chess was well suited to its role as an experimental medium: <em>Drosophila melanogaster </em>(the common fruit fly) is used as a &#8216;model organism&#8217; by researchers in various programmes of genetic analysis. For Kronrod, it was the internal characteristics of chess that made it ideal. It was, after all, a simple game with a well-defined problem domain and unambiguous rules, clear objectives, and straightforward measures of success. </p><p>This is the reason that games have been used throughout the history of AI, even those that are more complex like Dota 2, StarCraft or Pok&#233;mon. But therein lies the rub. If all games share these fundamental qualities, why was it that computer scientists used chess specifically? </p><p>One obvious answer is that chess &#8212; unlike say, Go &#8212; was popular with American and European researchers. The ability to play chess well was also traditionally <a href="https://www.sciencedirect.com/science/article/abs/pii/S0160289616301593">considered</a> to be an indicator of intelligence, and the game has long been associated with intellectuals, artists and other high-status types. </p><p>AI grandees Allan Newell and Herbert Simon, who were among the participants in the influential AI conference in Dartmouth in 1956, famously <a href="https://research.gold.ac.uk/id/eprint/22940/1/Human_Machine_Learning_2018_CE.pdf">said</a> that: &#8216;chess is the intellectual game par excellence.&#8230;If one could devise a successful chess machine, one would seem to have penetrated to the core of human intellectual endeavor.&#8217; </p><p>For these reasons, lots of 20th century AI researchers played chess. As the historian Nathan Ensmenger <a href="https://homes.luddy.indiana.edu/nensmeng/files/Ensmenger2012-Chess.pdf">puts it</a>, &#8216;many of the mathematicians who worked on computer chess, including Turing and Shannon, were avid amateur players.&#8217; </p><p>Chess also came with the Elo system, a ranking approach named after the Hungarian-born physicist <a href="https://worldchesshof.org/hof-inductee/arpad-emrick-elo">Arpad Elo</a>. While I won&#8217;t spend time on the details, the point is that the Elo system provided clear numerical benchmarks for measuring performance and improvement &#8212; a very helpful quality when designing computer systems that get better over time. </p><h2>Endgame</h2><p>Chess proved to be an ideal (or at the very least, idealised) testbed for AI research in the 20th century. Its balance of complexity and simplicity, widespread popularity, codified rules, and quantitative performance metrics like Elo ratings made it AI&#8217;s model organism. </p><p>The game was popular with the researchers designing AI systems, and the rich documentation of games, openings, and scenarios provided the information needed to design and develop early chess engines. The Elo system offered a stable means of assessing performance and chess matches provided public spectacle at the height of Cold War competition.</p><p>While AI research has moved on from chess, its history tells us that the artefacts through which technical practice takes place aren&#8217;t chosen by accident. That is not to say that chess is an inappropriate medium for building AI systems or that today&#8217;s testbeds are troublesome, but rather that the use of chess in AI reminds us that science doesn&#8217;t happen in a vacuum.</p><p>Just as the choices we make about which mediums to use shape the direction of research, so too are these decisions influenced by sources we don&#8217;t always recognise. The history of science is replete with examples in this tradition, from the Lenna image on which the early standards for computerised colourisation depended to the international prototype kilogram that enabled the measurement of mass for hundreds of years.</p><p>The historian Dylan Mulvin <a href="https://mitpress.mit.edu/9780262045148/proxies/">calls</a> these objects &#8216;proxies&#8217;. He argues that they mediate between the practicality of getting work done and the representation of the world. Perhaps chess does something similar. The board might be small, but you can pack a lot into sixty-four squares.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Histories. Subscribe for free as we make it to 50 short stories about AI&#8217;s past lives.  </p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Uncle Sam's electronic brain ]]></title><description><![CDATA[AI Histories #7: The perceptron and its critics]]></description><link>https://www.learningfromexamples.com/p/uncle-sams-electronic-brain</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/uncle-sams-electronic-brain</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 12 Jun 2025 10:23:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!kJaF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kJaF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kJaF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png 424w, https://substackcdn.com/image/fetch/$s_!kJaF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png 848w, https://substackcdn.com/image/fetch/$s_!kJaF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png 1272w, https://substackcdn.com/image/fetch/$s_!kJaF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kJaF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png" width="1122" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1122,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1541884,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/164417586?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kJaF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png 424w, https://substackcdn.com/image/fetch/$s_!kJaF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png 848w, https://substackcdn.com/image/fetch/$s_!kJaF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png 1272w, https://substackcdn.com/image/fetch/$s_!kJaF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd422ddfb-7edd-4e02-b5c4-90cbde39ff7e_1122x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Frank Rosenblatt and his computer &#8216;embryo&#8217; </figcaption></figure></div><p>In the summer of 1958, Frank Rosenblatt was putting on a show. The Cornell University psychologist made the trip down the east coast to the Weather Bureau in Washington D.C.</p><p>Rosenblatt wasn&#8217;t meeting meteorologists. His partners were at the Office of Naval Research, a group of clever G-Men who needed access to the bureau&#8217;s IBM 704 computer. </p><p>In the computer bay on Independence Avenue, Rosenblatt and his backers invited reporters to watch as he fed the machine a deck of fifty punch-cards. The 704 whirred and buzzed, and began to guess the location of a square symbol on each. </p><p>After 50 practice runs, the machine could correctly puzzle out where each mark was by using what would become known as the &#8216;perceptron&#8217; algorithm. Much like a young child, the room-sized computer learned to tell left from right. </p><p>Reporters were shocked. A wire story ran that afternoon &#8212; and the <em>New York Times</em> the following day &#8212; carrying the <a href="https://www.nytimes.com/1958/07/08/archives/new-navy-device-learns-by-doing-psychologist-shows-embryo-of.html">news</a> that the &#8216;embryo [would one day] walk, talk, see, write, reproduce itself and be conscious of its existence.&#8217; </p><p>The perceptron&#8217;s operation was straightforward. At its core lay a set of weighted inputs that could be adjusted depending on whether the machine made correct or incorrect guesses. Just as we might learn language through correction and repetition, Rosenblatt believed his machine might classify objects and &#8216;perceive&#8217; (hence the name) its surroundings.</p><p>Funded generously by the U.S. Navy, the algorithm was intended to loosely mimic the basic operations of biological neurons. Rosenblatt saw in his work the mechanical embodiment of Turing&#8217;s <a href="https://courses.cs.umbc.edu/471/papers/turing.pdf">proposed</a> &#8216;child machine&#8217;, a device that could acquire knowledge without explicit programming. </p><p>The spectacle worked. The press coverage unlocked a six figure grant for a purpose-built physical computer, the Mark I Perceptron, and catapulted Rosenblatt from obscure psychologist to Cold-War wunderkind. </p><p>Rosenblatt&#8217;s team built the Mark I in the years that followed at Cornell Aeronautical Laboratory. Technicians soldered 400 photocells into a 20 &#215; 20 &#8216;retina&#8217; and connected them to 512 association units according to a table of random numbers. </p><p>The resulting tangle of wire looked like a copper bird nest, but it seemed to be able to perform some basic intelligence tasks (like identifying simplified silhouettes in target-recognition studies). It might not have been pretty, but it showed promise. </p><p>That was good news, because the military needed a win. The Soviet Union's launch of the Sputnik satellite in 1957 had begun to cast a long shadow over American scientific confidence. </p><p>In laboratories and government offices, anxiety gave way to urgency. Money flowed into universities and research institutes, each hoping to uncover technologies that could secure American supremacy. </p><p>An intelligent machine could put Uncle Sam back on the front foot. America might have been second to orbit, but it was going to win the race for thinking machines. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h2>Boundary problems </h2><p>The Navy&#8217;s &#8216;electronic brain&#8217; was exactly what a nervous nation wanted. Within months, magazines speculated that computers like the Mark I would propel the United States closer to its pulp sci-fi dreams. </p><p>Grainy, high-contrast photos of the Mark I&#8217;s panels and flickering bulbs promised rationality, cleanliness, and progress. The machine provided an irresistible visual shorthand for a better future.</p><p>But not everyone bought it. </p><p>Marvin Minsky and his long time friend and collaborator Seymour Papert were amongst a growing group of researchers who felt the expectations for the perceptron project drastically overshot reality. </p><p>In 1969, the pair famously put the boot in by writing <em>Perceptrons: An Introduction to Computational Geometry. </em>Underneath the innocuous title lay a clear but troubling idea: single-layer perceptrons could not solve certain basic logical problems like XOR (exclusive-OR).</p><p>The XOR problem says that, given two inputs, the answer is &#8216;true&#8217; if exactly one input is true &#8212; but false if both are true or both are false. The rub for Rosenblatt&#8217;s perceptron was that it relied on drawing a straight line through input data, neatly separating it into categories. </p><p>But XOR was impossible to neatly divide this way, like trying to separate diagonally opposed black and white squares with a single straight slice. </p><p>But while their argument was mathematically sound, Minsky and Papert also <em>needed</em> it to be right. At the time, the duo were embedded at MIT&#8217;s Project MAC where they worked on symbolic, rule-based systems (a very different way of building AI compared to Rosenblatt&#8217;s proto-connectionism).</p><p>After Minsky and Papert&#8217;s critique, the decline in enthusiasm for the algorithm was unforgiving. State funding in the US (and in the UK a few years later) shifted toward symbolic AI, which was seen as a safer investment given the perceived limitations of neural approaches. </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nsas!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" width="64" height="20.869565217391305" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:150,&quot;width&quot;:460,&quot;resizeWidth&quot;:64,&quot;bytes&quot;:12198,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162870944?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a011107-4790-4b64-9f4c-4b8fcace22de_460x330.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I am wary of buying into the &#8216;AI winter&#8217; meme given that a huge number of important contributions to the AI project happened within this period, but I will grant that it is a useful concept if we&#8217;re primarily interested in the amount of researchers in the field, the level of funding it sustains or how many column inches it grabs.  </p><p>Assuming for the purposes of this post that we&#8217;re content to use these yardsticks, then what followed was a decade of cold as machine learning approaches like the perceptron fell out of fashion. </p><p>Not that they stayed there for long. </p><p>By the 1980s, it was Connectionism Summer&#8482; <a href="https://www.learningfromexamples.com/p/the-great-hopfield-network-debate">when</a> John Hopfield &#8216;brought neural nets back from the dead.&#8217; A few years later, David Rumelhart, Geoffrey Hinton, and Ronald Williams finished the great resurrection when they <a href="https://www.learningfromexamples.com/p/backpropagation-is-older-than-you">popularised</a> Paul Werbos&#8217; backpropagation method. </p><p>The latter breakthrough showed that multilayer networks could overcome the limitations pointed out by Minsky and Papert. Unlike Rosenblatt&#8217;s single-layer perceptrons, the models of the 1980s contained multiple hidden layers that learned patterns from the errors of those that preceded them. </p><p>As we discussed in the <a href="https://www.learningfromexamples.com/p/the-neuron-doctrine">first entry</a> in this series, Rosenblatt&#8217;s work was significant because it showed how to put Warren McCulloch and Walter Pitts&#8217; mathematical abstraction of a biological neuron into practice. Even early, limited successes showed that connectionism had potential.  </p><p>When perceptrons got stuck in the mud, researchers were forced to find a solution. That solution led to backpropagating neural networks. When backpropagating neural networks proved scalable, we got deep learning and eventually ChatGPT. </p><p>But the through-line from 1958 to today's language models is also ideological. Every major breakthrough in AI has been underwritten by the same awe and anxiety that sent reporters scrambling during the Cold War. </p><p>Today's moment is functionally different, but the underlying narrative is remarkably familiar. Venture capital cash stands in for Navy grants, driven by the conviction that whoever builds the smartest machines wins the future. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Backpropagation is older than you think]]></title><description><![CDATA[AI Histories #6: Backprop begins in the Second World War, not with Werbos or Rumelhart]]></description><link>https://www.learningfromexamples.com/p/backpropagation-is-older-than-you</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/backpropagation-is-older-than-you</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 05 Jun 2025 10:03:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Nrc2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nrc2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nrc2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png 424w, https://substackcdn.com/image/fetch/$s_!Nrc2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png 848w, https://substackcdn.com/image/fetch/$s_!Nrc2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png 1272w, https://substackcdn.com/image/fetch/$s_!Nrc2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nrc2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png" width="764" height="614" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:614,&quot;width&quot;:764,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:881937,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162897684?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Nrc2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png 424w, https://substackcdn.com/image/fetch/$s_!Nrc2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png 848w, https://substackcdn.com/image/fetch/$s_!Nrc2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png 1272w, https://substackcdn.com/image/fetch/$s_!Nrc2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99c9e931-9fd4-498a-b332-7d4179e1ede2_764x614.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"> William Thomas Rawlinson painting of a Radar Station on the East Coast (1946)</figcaption></figure></div><p>Backpropagation is the stuff that makes neural networks tick. Without it, it&#8217;s possible there&#8217;s no AI project as we know it today. No deep learning. No computer vision. No ChatGPT. </p><p>That&#8217;s because frontier AI systems are learning machines. They adjust to mimic patterns, improve by minimising error, and evolve to get better. Backprop helps them do that. </p><p>Using backprop, a network can make a prediction and calculate how far the prediction is from the true value. The technique takes that error and moves it backwards through the network, layer by layer, using a rule to figure out how much each connection contributed to the mistake. </p><p>Each weight is then adjusted via the magic of gradient descent so the network gets closer to the right answer next time. </p><p>For many years, backprop was thought of as the brain-child of David Rumelhart and Geoffrey Hinton. Their 1986 paper (with Ronald Williams) showed neural networks could learn internal features. It felt like a new day for machine learning researchers still reeling from Minsky and Papert&#8217;s takedown of the perceptron. </p><p>A little later, received wisdom shifted as credit for backprop went to a Harvard graduate student called Paul Werbos. In his 1974 thesis, Werbos framed the idea as &#8216;reverse&#8208;mode&#8217; optimisation for dynamic systems. Alas, symbolic AI was in vogue and his thesis gathered dust.       </p><p>But as Werbos himself acknowledges, backprop&#8217;s lineage goes much further back. </p><p>Backpropagation is born of optimal control theory, the science of how to steer complex systems toward a goal in the most efficient way possible. Whether it&#8217;s a jet adjusting its thrust or a robot arm learning to move, the challenge is the same: to figure out how to act now in order to do better later. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><h2>Where did backprop come from? </h2><p>The genesis of optimal control theory can be traced to 1940s-era thinking about optimisation, which <a href="https://archive.org/details/operationalresea0000kirb">involved</a> a constellation of techniques known as &#8216;operational research&#8217; in Britain and &#8216;operations research&#8217; in the United States. </p><p>During the Second World War, &#8216;operations research&#8217; referred to work that puzzled out the most effective way of achieving a given military objective. The goal of the operations research was, as British radio direction pioneer Robert Watson-Watt <a href="http://doi.org/10.1038/160660a0">put it</a>, &#8216;to examine quantitatively whether the user organisation is getting from the operation of its equipment the best attainable contribution to its overall objective.&#8217; </p><p>With roots in the analysis of radar telemetry in late 1930s Britain, the field &#8216;diffused extraordinarily rapidly&#8217; <a href="https://www.jstor.org/stable/223021">through</a> British and American commands in the years in the Second World War. </p><p>Building on research conducted by the US military in the aftermath of the Second World War, American mathematical scientist George Bernard Dantzig worked on the military&#8217;s mechanisation efforts for the Pentagon. </p><p>In 1947, he <a href="http://www.jstor.org/stable/25146928">responded</a> to his assignment by conceiving a popular algorithm for linear programming that seeks to achieve the best outcome in a mathematical model where the requirements are represented by linear relationships. </p><p>Dantzig left the Pentagon in 1952 to take up a position in the Mathematics Department of the RAND Corporation in Santa Monica, California. At RAND, Dantzig gave a series of talks, including a lecture attended by Richard Bellman, who was finding approaches to multistage decision problems. </p><p>One oft-told story is that this lecture gave Bellman his eureka moment, though it was likely that the Soviet mathematician Lev Pontryagin&#8217;s work on the &#8216;maximum principle&#8217; for taking a system from one state to another was just as important. </p><p>The techniques Bellman proceeded to develop over the 1950s came to be known as dynamic programming, an algorithmic approach for solving an optimisation problem by breaking the problem down into simpler subproblems. Reflecting on the development of dynamic programming, Bellman <a href="http://www.jstor.org/stable/3088448">noted</a>:</p><blockquote><p>A number of mathematical models of dynamic programming type were analyzed using the calculus of variation. The treatment was not routine since we suffered either from the presence of constraints or from an excess of linearity. An interesting fact that emerged from this detailed scrutiny was that the way one utilized resources depended critically upon the level of these resources, and the time remaining in the process.</p></blockquote><p>Bellman&#8217;s perception marked what some historians see as the transition between &#8216;classical control theory&#8217; and what became known as optimal control theory. Where classical control was concerned with the stabilisation of a given system through continuous monitoring and modification, Bellman&#8217;s realisation involved thinking about a system as a temporally evolving sequence of states.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nsas!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" width="64" height="20.869565217391305" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:150,&quot;width&quot;:460,&quot;resizeWidth&quot;:64,&quot;bytes&quot;:12198,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162870944?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a011107-4790-4b64-9f4c-4b8fcace22de_460x330.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>In the 1950s, the US military establishment was fretting about how to steer missiles that had minutes to correct their flight-paths. The problem was unforgiving. Researchers needed to know how to compute the precise adjustments needed to hit a moving target in real time, while accounting for changing wind, velocity, altitude, and fuel constraints.</p><p>The challenge led Richard Bellman to his big idea, or at least lots of smaller ones. Bellman decided to start at the final goal, then work backward to figure out the best decision at each step given the remaining time and resources. </p><p>This idea became a cornerstone of optimal control theory, formalised by Bellman under the name dynamic programming, which sought to find optimal strategies in complex, time-dependent systems. </p><p>By the 1960s, control engineers were combining this approach with what they called &#8216;adjoint equations&#8217; thanks to Pontryagin and his collaborators in the USSR. The method involved nudging the final outcome and tracing how that change would have flowed backward through the system. By doing so, they could figure out how small adjustments earlier on (e.g. thrust, angle, or speed) would affect the final result. </p><p>In the closing years of the decade, Bryson and Ho published <em>Applied Optimal Control</em>. Page after page shows the same move. Pose a goal (say, minimise fuel while hitting Mach 3). Derive Euler&#8211;Lagrange equations. Run the gradient in reverse through the system. Update parameters and repeat. </p><p>In the 1970s, Finnish mathematician Seppo Linnainmaa was working on numerical stability. He wanted a way to get exact derivatives from a computer program by following the logic of the code itself. His solution was to record every step a computer took when calculating a function, then replay those steps in reverse to figure out how each input affected the output. </p><p>Bryson, Ho, and Linnainmaa became a mainstay of a new generation of machine learning researchers in the 1980s, with the trio appearing in what felt like every other paper during the height of the neural network revival.  </p><p>As Yann LeCun put it in a <a href="https://new.math.uiuc.edu/MathMLseminar/seminarPapers/LeCunBackprop1988.pdf">paper</a> edited by Geoffrey Hinton in 1988: </p><blockquote><p>From a historical point of view, back-propagation had been used in the field of optimal control long before its application to connectionist systems. Nevertheless, the interpretation of back-propagation in the context of connectionist systems, as well as most related concepts are recent, and the historical and scientific importance of [Rumelhart et al., 1986] should not be overlooked. The concepts are new, if not the algorithm.</p></blockquote><p>The story of backpropagation reminds us that scientific practice is often less about bolt-from-the-blue genius than about clever recycling. If we&#8217;re careful, we can trace the idea from Dantzig&#8217;s wartime military modelling, Bellman&#8217;s dynamic-programming rockets, Bryson &amp; Ho&#8217;s calculus, Linnainmaa&#8217;s derivatives program, Werbos&#8217; graduate-school thesis, and finally Rumelhart and Hinton&#8217;s psychological wrapping. </p><p>What modern AI calls a theory of learning is something like a travelling mathematical trick, one that has migrated across radar rooms, missile labs, and psychology departments. With each life it accumulates new metaphors and new meanings, but the basic idea is still the same. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Please ignore the man behind the curtain ]]></title><description><![CDATA[AI Histories #5: The Mechanical Turk]]></description><link>https://www.learningfromexamples.com/p/the-man-behind-the-curtain</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/the-man-behind-the-curtain</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 29 May 2025 10:14:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!wDMg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The pledge drive to make <em>Learning From Examples</em> a full-time thing <strong>concludes this week.</strong> A $5 pledge doesn&#8217;t cost anything today, but does tell me I have your support when I eventually flip the switch on paid subscriptions (probably later this year). To everyone who has pledged so far: thank you. I&#8217;ve felt incredibly fortunate to be on the receiving end of your generosity.   </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wDMg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wDMg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg 424w, https://substackcdn.com/image/fetch/$s_!wDMg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg 848w, https://substackcdn.com/image/fetch/$s_!wDMg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!wDMg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wDMg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg" width="672" height="519.5654952076677" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:484,&quot;width&quot;:626,&quot;resizeWidth&quot;:672,&quot;bytes&quot;:152860,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wDMg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg 424w, https://substackcdn.com/image/fetch/$s_!wDMg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg 848w, https://substackcdn.com/image/fetch/$s_!wDMg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!wDMg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e639d67-3616-43b9-8f69-91aea4dd8171_626x484.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The Mechanical Turk via Wikipedia Commons</figcaption></figure></div><p>Napoleon liked chess. He played against his courtiers and against his companions. He enjoyed a quick game with the generals, diplomats, and intellectuals who moved in his orbit. By all accounts he was pretty good, which is probably not all that surprising. </p><p>His most famous opponent is only known by die-hard chess enthusiasts. That man was likely a German chess master called Johann Bapiste Allgiaer. Unfortunately, we don&#8217;t know with absolute certainty who the player was. That&#8217;s because neither Napoleon nor the audience could see them. </p><p>Instead, in the packed-out palace of Sch&#246;nbrunn in Vienna, one of history&#8217;s best known figures thought he was playing an automaton. Dressed in the style of an Ottoman (or rather what Austrian high society thought an Ottoman looked like), that machine was the Mechanical Turk. </p><p>Behind its mahogany frame and sliding panels was the chess master. Whether Allgiaer or a mystery player, our man no doubt felt uncomfortable crammed into a small wooden box. But whatever the impracticalities, they weren&#8217;t enough to prevent Napoleon&#8217;s opponent from guiding the automaton&#8217;s actions from underneath the table. </p><p>Reports of the game &#8212; which admittedly are on the <a href="https://www.chess.com/blog/ThePawnSlayer/the-chess-player-who-defeated-an-emperor">hazy</a> side &#8212; tell us that the emperor made an illegal move to test the Turk&#8217;s reaction. His opponent reset the piece to its original position. Napoleon tried again with another illegal move. The Turk responded by removing the offending piece from the board. </p><p>On the third attempt, the story goes, the Turk dramatically swept all the pieces off the board. Satisfied he had sufficiently needled the automaton, the Frenchman tried his hand at a legitimate game. He was soundly beaten. </p><h2>Vienna waits for you  </h2><p>The Turk was constructed in 1770 by Wolfgang von Kempelen, a Hungarian nobleman with a knack for building things. Kempelen was inspired to construct the Turk following a visit to the Vienna court of Maria Theresa of Austria, where the great magician Fran&#231;ois Pelletier was performing a show. </p><p>Kempelen liked what he saw, but thought he could do one better. Resolving to upstage the Frenchman, he set out creating a show that blended the spectacle of performance and the precision of engineering. </p><p>The result was the Turk. On the outside, his machine looked like a life-sized figure clad in robes and a turban seated behind a wooden cabinet. Its left arm held a long pipe at rest, while its right lay on the top of a large cabinet. </p><p>From the inside, the machine was a tangle of clockwork and mirrors. Its board was magnetised, which allowed the operator to track and manoeuvre pieces from inside the device. </p><p>Opening doors on one side revealed clockwork-like gears, but the section was constructed to allow would-be inspectors to see through the device under certain conditions. Hidden doors under the model showed cogs to maintain the illusion that the cabinet was filled with mechanisms. </p><p>In reality, neither the visible clockwork nor the drawer extended fully to the back of the table. A sliding seat on the interior allowed our hidden chess player to shift position as the dummy doors were opened to dazzle onlookers. </p><p>The act began by unlatching the doors, much like a magician shows the audience there&#8217;s nothing up his sleeve. On its first exhibition in the Vienna palace, Kempelen made an elaborate show of allowing the crowd to inspect the device. </p><p>Once the audience was suitably convinced nothing untoward was going on, the games began. One of the first to play was Count Ludwig von Cobenzl, an Austrian courtier at the palace. He was quickly defeated by the Turk&#8217;s aggressive style of play, while a host of others who fancied their chances met a similar fate. </p><p>Despite the fanfare caused by the Turk, its designer quickly lost interest in the project. The machine only played a handful of opponents in the ten years following its debut. </p><p>Kempelen was by all accounts bored by his creation. He wanted to work on steam engines rather than spending time assembling and disassembling the Turk, which he famously decried as a &#8216;bagatelle&#8217; (that is, a trifle). </p><p>But the Habsburg crown had other ideas. In 1781, Kempelen was ordered by Emperor Joseph II to reconstruct the Turk and deliver it to Vienna for a state visit from Grand Duke Paul of Russia and his wife. After another successful appearance, the court suggested a tour of Europe. </p><p>Fearing upsetting the powers that be, Kempelen agreed. </p><p>The first stop was France, where the Turk faced off against opponents in Versailles and Paris. Against more practiced players, the machine &#8212; or rather the man under the table &#8212; won some matches but lost others. </p><p>The Turk's final game in the City of Light was against Benjamin Franklin, who was serving as ambassador to France from the United States. Franklin reportedly enjoyed the game with the Turk, but was ultimately beaten. </p><h2>Second Act</h2><p>After Kempelen&#8217;s death, the Bavarian musician and inventor Johann Nepomuk M&#228;lzel purchased the Turk from his son in 1805. Best known for perfecting the metronome, M&#228;lzel once again took the Turk to the courts of Europe (including for its famous bout with Napoleon). </p><p>Under M&#228;lzel, the Turk became a symbol of modernity. It returned to Paris and London, eventually crossing the Atlantic to show American audiences what they had been missing. </p><p>In Richmond, Virginia, a young Edgar Allan Poe watched it play. In his essay <em>Maelzel&#8217;s Chess-Player</em>, Poe tried to deduce the mechanics of the system. He mostly failed, but his broader point was that the influence of the Turk lay in its ability to evoke mystery.</p><p>By this point, the Turk&#8217;s fame wasn't really about chess. The illusion endured because it touched a nerve during an age in which politics, technology, and economics began a great reconfiguration. </p><p>In 1838, Johann M&#228;lzel died aboard a ship off the American coast, and the automaton&#8212; now worn from decades of travel &#8212; was left to gather dust in a Philadelphia museum.  </p><p>Then, in 1854, the Turk went up in flames when a blaze engulfed the site. With neither Turk nor master still standing, former operators and witnesses came forward to expose its secrets. </p><p>The world now knew how the figure inside used levers to manipulate the arm, how the cabinet&#8217;s compartments were rigged for misdirection, and how the illusion was kept alive through carefully orchestrated performance.</p><p>The story of the Turk is about intelligence as theatre. Its secrets remained hidden for so long because people wanted to believe that the mind could be made mechanical. For those of us interested in where AI came from and where it is going, the Turk reminds us that artificial intelligence is part reality and part projection. </p><p>That isn&#8217;t to say modern AI is a hoax or anything like it, but rather that conceptions of intelligence are fluid. We read into intelligent machines whatever we need to. Some people see the entire AI project as a menagerie of smoke and mirrors. Others see god-like machines just around the corner. </p><p>The point is that AI is a container into which we pour our beliefs and biases. That was true when Napoleon faced off against the man in the box, and it&#8217;s still true over 200 years later. </p><p></p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Self-sustaining systems]]></title><description><![CDATA[AI Histories #4: Automata theory]]></description><link>https://www.learningfromexamples.com/p/self-sustaining-systems</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/self-sustaining-systems</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Thu, 22 May 2025 09:25:21 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!pUDa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The May pledge campaign is still in full swing. We're on track, but if you haven&#8217;t yet, I&#8217;d love your help to get there. <strong>Pledging $5 lets me know you&#8217;re behind the project</strong> as I work toward turning on paid subscriptions (and hopefully making this my full-time work). Huge thanks to everyone who&#8217;s pledged already. It really does mean a lot.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pUDa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pUDa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pUDa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pUDa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pUDa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pUDa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg" width="1200" height="916" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:916,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;The Gothic Arch, from \&quot;Carceri d'invenzione\&quot; (Imaginary Prisons), Giovanni Battista Piranesi (Italian, Mogliano Veneto 1720&#8211;1778 Rome), Etching, engraving, sulphur tint or open bite, burnishing; first state of six Robison) &quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="The Gothic Arch, from &quot;Carceri d'invenzione&quot; (Imaginary Prisons), Giovanni Battista Piranesi (Italian, Mogliano Veneto 1720&#8211;1778 Rome), Etching, engraving, sulphur tint or open bite, burnishing; first state of six Robison) " title="The Gothic Arch, from &quot;Carceri d'invenzione&quot; (Imaginary Prisons), Giovanni Battista Piranesi (Italian, Mogliano Veneto 1720&#8211;1778 Rome), Etching, engraving, sulphur tint or open bite, burnishing; first state of six Robison) " srcset="https://substackcdn.com/image/fetch/$s_!pUDa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg 424w, https://substackcdn.com/image/fetch/$s_!pUDa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg 848w, https://substackcdn.com/image/fetch/$s_!pUDa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!pUDa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F33a950e7-99ea-4427-8f77-656f2d55a79c_1200x916.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The Gothic Arch, from "Carceri d'invenzione" (Imaginary Prisons) by Giovanni Battista Piranesi ca. 1749&#8211;50</figcaption></figure></div><p>Automata theory is difficult to explain. </p><p>For many, it gets lumped in with the self-duplicating machine. My mind wanders to the <a href="https://www.centauri-dreams.org/2022/03/08/probing-von-neumann-expansion/">von Neumann probe</a>: a hypothetical spacecraft that could fly to a nearby star, gobble up materials to produce copies of itself, and continue on to the next system until the entire galaxy had been checked out. </p><p>It&#8217;s a fun image, but one only partially connected to the topic at hand. At its core automata theory is about abstract structures that evolve through finite internal states according to fixed rules. They are conceptual models, mathematical frameworks that capture precisely how systems transform, sustain, or collapse over time. </p><p>It&#8217;s an idea that shape-shifts, one that likes to be all things to all people. Or it is in AI, anyway.</p><p>That&#8217;s because automata theory has shaped the two foundational schools of the field. Both the symbolic approach in which rules are hard-coded into a system and the connectionist branch where systems learn from examples (cue klaxon) have been influenced by its ideas.  </p><p>For the former, that means designing systems that follow clear, rule-based transitions; for the latter, it involves reimagining those transitions as fluid patterns rather than fixed instructions. </p><p>For the purposes of this post, I&#8217;m defining an automaton as &#8216;a self-contained system that responds to inputs by changing state, whose behaviour is determined by its design.&#8217; In practice, that means anything from a light switch to a language model can be seen as an automaton (so long as its next move depends on its current state and some external input). </p><h2>Letters and numbers </h2><p>Automata theory begins as an attempt to pin down the limits of reason. In the 1930s, a handful of logicians&#8212;Alonzo Church at Princeton, Alan Turing in Cambridge, and Emil Post in New York&#8212;found themselves asking what it means to compute something. </p><p>At stake was whether all of mathematics could, in principle, be reduced to symbolic procedures carried out by rule-following agents. To answer the question, these thinkers built abstract machines. </p><p>Church used &#955;-calculus, Post proposed rewriting systems, and Turing devised a model so evocative it would take his name. Each was a kind of automaton in that it described a self-contained system that processes inputs and moves through internal states according to fixed rules.</p><p>It&#8217;s from this moment that automata theory begins to take form as a toolkit for describing procedural reasoning with mathematics. What started as a way to solve problems in logic would eventually lay the conceptual groundwork for the artificial intelligence project.</p><p>As symbolic automata were being marshalled to model thought, von Neumann wondered whether the principles of life itself could be described in the language of mathematics. Working with Stanislaw Ulam at Los Alamos, he devised the idea of a self-replicating automaton. It was a system that, given a set of instructions, could construct a copy of itself within a defined grid-like universe.</p><p>By the 1950s, the question was no longer &#8216;can something be computed?&#8217; Theorists proved that was possible, so they instead turned to structure. They wanted to know how different types of machines process different kinds of inputs.</p><p>Stephen Kleene, working with the mighty RAND Corporation, <a href="https://www.rand.org/content/dam/rand/pubs/research_memoranda/2008/RM704.pdf">established</a> the idea of regular expressions (patterns recognised by simple machines that move deterministically between states). Then Michael Rabin and Dana Scott <a href="https://www.youtube.com/watch?v=HRKR4NJ44T8">showed</a> how automata could be adapted to include non-determinism by exploring branching paths with multiple possible futures. </p><p>And from linguistics, a young Noam Chomsky imported a powerful organising framework. He saw a hierarchy of formal languages, each defined by the type of automaton that could recognise it. Chomsky showed that you can rank languages based on how complicated a machine you&#8217;d need to use them:</p><ul><li><p>At the bottom, you have regular languages, recognised by simple machines (finite automata).</p></li><li><p>Above that are context-free languages, which need a machine with a memory stack.</p></li><li><p>Then come context-sensitive languages, needing a still more powerful device known as a linear-bounded automaton. </p></li><li><p>At the top, you have recursively enumerable languages, which require a full Turing machine (the most powerful kind of abstract computer).</p></li></ul><p>In Chomsky&#8217;s hands, automata became models of cognition. For if language could be parsed by machines, perhaps the mind could be formalised as a generative system wired to produce and recognise structure. </p><h2>Transition and constraint</h2><p>By the second half of the 20th century, automata theory was something like a working metaphor for the mind. In symbolic AI and cognitive science, automata provided a language for representing internal mental states, transitions, and procedural rules. </p><p>But in parallel, a different tradition was emerging. Inspired less by formal logic than biology, researchers began to design networks that learned patterns over time. These connectionist models were a different kind of automata. They were systems defined by internal states that evolved through transitions, only here the transitions were learned rather than programmed.</p><p>Whether hand-coded or trained, both symbolic and connectionist systems relied on the idea that intelligence unfolds through structured change. Rules or learning, thinking became a process of moving from one configuration of the system to the next. </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Nsas!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png" width="106" height="34.56521739130435" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:150,&quot;width&quot;:460,&quot;resizeWidth&quot;:106,&quot;bytes&quot;:12198,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162870944?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a011107-4790-4b64-9f4c-4b8fcace22de_460x330.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Nsas!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 424w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 848w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1272w, https://substackcdn.com/image/fetch/$s_!Nsas!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e86f8a1-ceb0-4b5e-97c4-6b4f740ff3dc_460x150.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>To study automata is to ask how systems maintain identity through transformation. Whether parsing a sentence or navigating a decision tree, they provide a way of thinking about structure in motion.</p><p>At their most abstract, they embody a vision of intelligence as patterned change within bounds. They help us model how rules unfold and how memory shapes behaviour. </p><p>Every time we train a model, we rely on assumptions about transition and constraint. Even neural networks, for all their complexity, are still systems that evolve through internal states.</p><p>I like automata theory because it suggests that intelligence, artificial or otherwise, is both stable and dynamic. It&#8217;s a timely reminder that underneath probabilistic outputs lies the idea that thought is a process, and that process can be mapped.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[The great Hopfield network debate]]></title><description><![CDATA[AI Histories #3: A bitter clash about machine learning's finest hour]]></description><link>https://www.learningfromexamples.com/p/the-great-hopfield-network-debate</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/the-great-hopfield-network-debate</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Fri, 16 May 2025 09:26:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!L1uz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I&#8217;ve recently opened reader pledges with a goal of reaching<strong> 100 by May 31</strong>. So far we&#8217;ve on track, but I need your help to get there. All you have to do is pledge $5, so that I know I have your support for the moment I turn on paid subscriptions (and eventually try to make this work full time). Thank you to everyone who was kindly pledged so far. It means a lot. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!L1uz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!L1uz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg 424w, https://substackcdn.com/image/fetch/$s_!L1uz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg 848w, https://substackcdn.com/image/fetch/$s_!L1uz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!L1uz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!L1uz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg" width="800" height="534" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:534,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Steve Grossberg, Boston University Wang Professor of Cognitive and Neural Systems&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Steve Grossberg, Boston University Wang Professor of Cognitive and Neural Systems" title="Steve Grossberg, Boston University Wang Professor of Cognitive and Neural Systems" srcset="https://substackcdn.com/image/fetch/$s_!L1uz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg 424w, https://substackcdn.com/image/fetch/$s_!L1uz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg 848w, https://substackcdn.com/image/fetch/$s_!L1uz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!L1uz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49bb59d8-7c80-41cc-86ad-50350d8cbe1a_800x534.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Stephen Grossberg in 2015 via Boston University</figcaption></figure></div><p>John Hopfield won the Nobel Prize for physics last year. Alongside Geoffrey Hinton, Hopfield was recognised for foundational contributions to machine learning that would ultimately make the likes of ChatGPT and Gemini possible. His great invention was the network that bears his name, whose development is generally considered to be an essential moment in the machine learning canon. Speaking to the awarding committee, he <a href="https://www.nobelprize.org/prizes/physics/2024/hopfield/interview/">discussed</a> how he first began to approach work on the systems:</p><blockquote><p>You don&#8217;t leap into a problem overall saying, I want to understand how mind works. You have to build up from the bottom. If you were doing weather, you would say, well, I want to understand what storms are without going back to interacting air nitrogen molecules.</p></blockquote><p>Analogies, in other words, are key. Hopfield&#8217;s 1982 <a href="https://www.pnas.org/doi/epdf/10.1073/pnas.79.8.2554">paper</a> is commonly credited with the invention of the &#8216;Hopfield network&#8217;, a type of recurrent neural network where all units are connected to each other. Typically used for pattern recognition and memory storage, the emergence of the the systems seemed to show that there was still life in the &#8216;connectionist&#8217; branch of AI (the ancestor of deep learning). As researcher Tom Schwartz <a href="https://books.google.co.uk/books?id=29O9D0kQi70C&amp;printsec=frontcover&amp;source=gbs_ge_summary_r&amp;cad=0#v=onepage&amp;q=tom&amp;f=false">put it</a> &#8216;Hopfield should be known as the fellow who brought neural nets back from the dead.&#8217;</p><p>Hopfield is no doubt a hugely important figure, but wasn&#8217;t the first to design and implement a fully connected network. Stanford researcher William Little had previously introduced versions of the networks in <a href="https://link.springer.com/chapter/10.1007/978-1-4613-0411-1_12">1976</a>, and so had Japanese neuroscientist Shun'ichi Amari in <a href="https://ieeexplore.ieee.org/document/1672070">1972</a>. </p><p>The cognitive scientist Stephen Grossberg went further, arguing that he built the specific architecture that Hopfield made famous. As Grossberg, who first wrote about the ideas described in the Hopfield model in 1957, <a href="https://books.google.co.uk/books?id=-l-yim2lNRUC&amp;printsec=copyright&amp;redir_esc=y#v=onepage&amp;q&amp;f=false">put it</a>: &#8216;I don&#8217;t believe that this model should be named after Hopfield. He simply didn&#8217;t invent it. I did it when it was really a radical thing to do.&#8217;</p><p>But as every scientist knows, research needs rhetoric and papers need presentation. Hopfield removed chunks of mathematical descriptions in favour of compelling prose written for cognitive scientists. He published his paper in the influential <em>Proceedings of the National Academy of Science</em> and travelled extensively to talk about &#8216;his&#8217; networks. These are the primary reasons that today we know the systems not as Grossberg networks, but as Hopfield networks.</p><p>One of the paper&#8217;s most influential ideas involved conceptualising the networks as &#8216;spin glasses&#8217;, a term derived from the magnetic state of matter. &#8216;Spin&#8217; is a quantum property of particles that allows them to behave as miniscule magnets that can point in different directions. &#8216;Glass&#8217; draws from an analogy with conventional glass, which is known for its irregular, amorphous structure at the atomic level. Atoms are arranged in a disordered state in a typical glassy material, unlike in crystal forms where atoms have a more regular arrangement.</p><p>In the Hopfield network, the system&#8217;s dynamics&#8212;how these states settle into patterns&#8212;are inspired by the way magnetic spins interact in physical systems. Hopfield&#8217;s work connected this idea, the potential for systems to transition from a disordered state to a stable one, to the concept of &#8216;associative memory&#8217;. Based loosely on the brain, the associative memory concept holds that if you give a system a piece of the desired output it can &#8216;remember&#8217; the rest.</p><p>The connection was straightforward: just as a spin glass system transitions from a high energy state to a low energy state, a Hopfield network minimises its energy to represent and retrieve stored patterns. It was this analogy, which explained how a sea of small parts could settle into stable states representing stored memories, that encouraged the next generation of researchers to couple the operation of Hopfield networks with principles similar to those observed in physical systems.</p><p>Jack Cowen, an influential researcher who worked with Hopfield, <a href="https://books.google.co.uk/books?id=-l-yim2lNRUC&amp;printsec=copyright&amp;redir_esc=y#v=onepage&amp;q&amp;f=false">understood</a> the symbolic currency of the idea. He said &#8216;I think that's neat stuff, but I still think it's an artificial system, as is the antisymmetric one. It may have nothing to do with the way things really work in the nervous system, but it's a very interesting idea.&#8217; </p><p>By drawing a parallel between associative memory and the behaviour of physical systems like spin glasses, Hopfield reinforced the idea that complex cognitive processes could emerge from the collective behaviour of simple interacting units. This perspective supported the well-established notion that the brain&#8217;s higher-order cognitive functions could be explained through the interactions of simpler components.</p><p>As Hopfield <a href="https://www.pnas.org/doi/epdf/10.1073/pnas.79.8.2554">explained</a>: &#8216;Computational properties of use to biological organisms or to the construction of computers can emerge as collective properties of systems having a large number of simple equivalent components (or neurons).&#8217; The result, according to this model, is that intelligence is an emergent property that may be produced through the interaction of smaller, simpler units. Faced with such a conclusion, it is perhaps unsurprising that the paper caught fire.</p><h3><strong>Toy models</strong></h3><p>To explain the functioning of the systems, Hopfield leaned heavily on the famous law coined by Canadian psychologist Donald Hebb: &#8216;When an axon of cell A is near enough to excite a cell B and repeatedly or persistently takes part in firing it, some growth process or metabolic change takes place in one or both cells such that A&#8217;s efficiency, as one of the cells firing B, is increased.&#8217; You might know it as &#8216;neurons that fire together, wire together.&#8217; </p><p>Of course, Hopfield networks are not brains. And neither are they spin glasses. These types of comparisons are better <a href="https://philpapers.org/rec/REUUTM">thought</a> of as &#8216;toy models&#8217; that seek to help scientists understand the world. Frank Rosenblatt, on whose famous perceptron algorithm Hopfield aimed to build, <a href="https://safari.ethz.ch/digitaltechnik/spring2018/lib/exe/fetch.php?media=neurodynamics1962rosenblatt.pdf">stressed</a> the importance of toy models to AI research in 1961:</p><blockquote><p>&#8216;The model is a simplified theoretical system, which purports to represent the laws and relationships which hold in the real physical universe&#8230;the model deliberately neglects certain complicating features of the natural phenomena under consideration, in order to obtain a more readily analyzed system, which will suggest basic principles that might be missed among the complexities of a more accurate representation.&#8217;</p></blockquote><p>These models tell researchers about the problem they are trying to solve, but they also widen the epistemological field to accommodate new perspectives. Philosophers of science Knuttila and Loettgers, for example, have <a href="https://www.sciencedirect.com/science/article/abs/pii/S0039368114000612">shown</a> that mental models &#8216;provide modelers with a powerful cognitive strategy to transfer concepts, formal structures, and methods from one discipline to another&#8217;. </p><p>Through this process of analogical reasoning, the twin abstractions of spin glass modelling and neurophysiology offered a way of describing the functioning of the systems that was both persuasive <em>and</em> epistemically valuable. It may have won Hopfield a Nobel Prize, but that would have been small comfort to Stephen Grossberg. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[Emergence machines]]></title><description><![CDATA[AI Histories #2: The birth of genetic algorithms]]></description><link>https://www.learningfromexamples.com/p/emergence-machines</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/emergence-machines</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Fri, 09 May 2025 08:39:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8ca4eb6-7366-4711-97b5-06bdc81f9ede_1120x630.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_jac!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_jac!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg 424w, https://substackcdn.com/image/fetch/$s_!_jac!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg 848w, https://substackcdn.com/image/fetch/$s_!_jac!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!_jac!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_jac!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg" width="1120" height="630" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:630,&quot;width&quot;:1120,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:125247,&quot;alt&quot;:&quot;IBM 700 Series&quot;,&quot;title&quot;:&quot;IBM 700 Series&quot;,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="IBM 700 Series" title="IBM 700 Series" srcset="https://substackcdn.com/image/fetch/$s_!_jac!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg 424w, https://substackcdn.com/image/fetch/$s_!_jac!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg 848w, https://substackcdn.com/image/fetch/$s_!_jac!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!_jac!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F00629aea-414d-402f-97f3-d72f68a48b46_1120x630.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The IBM 700 series. </figcaption></figure></div><p>Friday nights were game night for John Holland. Once every fortnight, Holland and a small group of young IBM researchers met to play Kriegspiel, poker, and Go. It was the summer of 1951 in the city of Poughkeepsie in New York State, and IBM was racing to build the first commercial programmed computer.</p><p>By day, the researchers worked to figure out how circuits, memory, and instruction sets worked together to support machine-language programming in the IBM 701. By night, when they weren&#8217;t trading cards, they tested it. Some were showing the machine how to play checkers. Holland, fresh from an undergraduate at MIT, was teaching it to learn. </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Baeo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Baeo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 424w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 848w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 1272w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Baeo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png" width="148" height="60.87547169811321" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:218,&quot;width&quot;:530,&quot;resizeWidth&quot;:148,&quot;bytes&quot;:11942,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162676340?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Baeo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 424w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 848w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 1272w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><p>John Holland grew up amongst the quiet plains and soy&#8209;bean factories of rural Indiana. Born in 1929 to a businessman father and an adventurous mother who learned to fly in her forties, Holland was encouraged to try anything that took his fancy. </p><p>A chemistry set sparked his early love of science, and the young Holland eventually found his way to MIT. There, as an undergraduate physics major, he undertook his bachelor&#8217;s thesis on Whirlwind, the university&#8217;s digital computer built for missile detection. </p><p>After MIT, Holland took a job at IBM&#8217;s Poughkeepsie lab in 1950. There he joined the small group building the IBM 701, the company&#8217;s first commercial electronic computer. Nicknamed the Defense Calculator, the 701 was a room-sized behemoth made up of tubes and magnetic drums.  </p><p>By 1959 Holland had earned what was essentially Michigan&#8217;s first computer-science PhD, with a dissertation studying feedback loops in early neural networks. Spurred by his encounters with IBM 701 and buoyed by his academic work, Holland said he began to think &#8216;about genetics and adaptive systems&#8217;. </p><p>A key moment came in 1955 when Holland stumbled on a book called <em>On The Genetical Theory of Natural Selection</em> written by Ronald Fisher. Fisher, generally regarded as the great figure in the history of statistics, set out to show how laws of inheritance could underpin natural selection. He introduced the idea that evolution is about changes in gene frequencies within populations, and formulated the famous &#8216;<a href="https://academic.oup.com/evolut/article/78/5/803/7624408">fundamental theorem of natural selection</a>,&#8217; which roughly states that a population can only get fitter as fast as there&#8217;s useful genetic variety to work with.</p><p>In Fisher&#8217;s equations Holland saw a template for an algorithm that could encode potential solutions as &#8216;genes&#8217; to let the fitter ones reproduce and occasionally mutate. The combination of mutation and reward meant he just needed to define a target and watch the algorithm adapt its way to an optimal solution. </p><p>In practice, that meant treating possible solutions like digital organisms. The system evaluated each string of code according to how well it performed, then selected the stronger ones to combine and reproduce. Their code was mixed, mutated, and eventually used to create a new generation of candidate solutions.</p><p>The process has an advantage over simpler search methods. Hill climbing algorithms improve a single solution piece by piece but get stuck on local peaks. It&#8217;s easy to produce a good answer but hard to produce a great one. Genetic algorithms avoid this fate by maintaining a diverse population of solutions. Through mutation and recombination, they can occasionally make bold jumps into less promising territory that turn out to lead somewhere better.</p><p>Holland published the first comprehensive account of genetic algorithms in his 1975 book <em>Adaptation in Natural and Artificial Systems</em>. The work aimed to lay a general &#8216;mathematical theory of adaptation&#8217; that used genetic algorithms as a central tool to solve complex optimisation problems. </p><p>Genetic algorithms are good at things like shaping an aircraft wing, designing an electronic circuit, or routing a network. They work well for tasks for which you can&#8217;t easily design a solution from first principles, but you can test how well any given attempt performs. </p><p>In machine learning, genetic algorithms are used when there&#8217;s no clear way to calculate the best setup (like deciding which features to include in a model or tuning parameters that control how it learns). If you can&#8217;t derive the answer but you can run it and see how well it works, it&#8217;s a problem that genetic algorithms might be good at.   </p><h2>Emergence </h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rsFy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rsFy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg 424w, https://substackcdn.com/image/fetch/$s_!rsFy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg 848w, https://substackcdn.com/image/fetch/$s_!rsFy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!rsFy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rsFy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg" width="1200" height="810" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:810,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;What makes Romanesco broccoli so mathematically perfect? | Salon.com&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="What makes Romanesco broccoli so mathematically perfect? | Salon.com" title="What makes Romanesco broccoli so mathematically perfect? | Salon.com" srcset="https://substackcdn.com/image/fetch/$s_!rsFy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg 424w, https://substackcdn.com/image/fetch/$s_!rsFy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg 848w, https://substackcdn.com/image/fetch/$s_!rsFy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!rsFy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F062c4447-2d61-4c99-a9eb-85681f7333b8_1200x810.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Romanesco broccoli displays fractal geometry with a logarithmic spiral arrangement that follows the Fibonacci sequence. Its complex structure emerges from simple cellular growth rules. </figcaption></figure></div><p>Holland extended his framework to what he called Learning Classifier Systems (LCS) in the 1980s. These were evolving collections of simple if-then rules that compete, cooperate, and adapt based on feedback. Instead of solving a fixed problem, they learn how to behave in changing environments. One system learned to play checkers. Another optimised industrial pipeline flow in real time. The idea is that they weren&#8217;t being pre-programmed, but they could learn which rules worked and reinforce them over time.</p><p>It was a curious marriage. The systems used if-then rules, the bread and butter of symbolic AI, but fused them with learning &#8212; the hallmark of connectionist models like neural networks. In most of the AI industry at the time, these approaches were seen as opposites (and for the most part still are today). Rules are explicit, logical, and human-readable. Learning is messy, statistical, and ambiguous. </p><p>But Holland didn&#8217;t see a contradiction. He saw a system where rules could be treated like genes. They were hypotheses to be tested, combined, and evolved. Each rule was part of a population and scored according to how well it performed. Useful ones were strengthened and reused, while weaker ones were discarded and replaced by new variations. </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Baeo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Baeo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 424w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 848w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 1272w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Baeo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png" width="148" height="60.87547169811321" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:218,&quot;width&quot;:530,&quot;resizeWidth&quot;:148,&quot;bytes&quot;:11942,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162676340?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Baeo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 424w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 848w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 1272w, https://substackcdn.com/image/fetch/$s_!Baeo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6d261a7-1efa-4c62-9cbd-474155c7c8dd_530x218.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>In later years Holland became a leading voice in the science of complex adaptive systems. He co-founded the Santa Fe Institute, an interdisciplinary hub for studying complexity. His philosophy was consistent: intelligence emerges from the interaction of many simple parts, not from a single clever algorithm. </p><p>Holland viewed any group of interacting, adaptive entities&#8212;be it neurons in a brain, ants in a colony, or humans in a city&#8212;as essentially the same phenomenon. Each a form of computation with emergent collective behaviour. In his book <em>Emergence</em>, for example, he described how brain functions or market behaviour could not be understood by simply summing up individual units; rather, the nonlinear interactions made the aggregate far more complex than the parts&#8203;. </p><p>Genetic algorithms represent a philosophy of intelligence. Holland believed that intelligence was an emergent property that bubbled up from competition. That idea is now at the core of the most successful approaches to AI. While many may not use genetic algorithms, they do embody Holland&#8217;s deeper lesson: intelligence emerges from systems that change.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[The Neuron Doctrine ]]></title><description><![CDATA[AI Histories #1: Ram&#243;n y Cajal]]></description><link>https://www.learningfromexamples.com/p/the-neuron-doctrine</link><guid isPermaLink="false">https://www.learningfromexamples.com/p/the-neuron-doctrine</guid><dc:creator><![CDATA[Harry Law]]></dc:creator><pubDate>Sat, 03 May 2025 11:26:12 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!5zQR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Today&#8217;s post is the first entry in a new series looking at important moments in the history of AI and the fields that influenced it. These pieces will be shorter than the weekly essays and will land in your inbox every Friday.</p><p>This week also marks the start of reader pledges. I&#8217;m testing the waters to see whether there&#8217;s a path towards working on this newsletter full time, with a goal of<strong> 100 pledges by May 31</strong>. The response has been even better than I had hoped, so thank you to everyone who has kindly pledged so far. For anyone considering supporting this project, this is the best moment to help it grow.  </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5zQR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5zQR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png 424w, https://substackcdn.com/image/fetch/$s_!5zQR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png 848w, https://substackcdn.com/image/fetch/$s_!5zQR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png 1272w, https://substackcdn.com/image/fetch/$s_!5zQR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5zQR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png" width="1456" height="833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:833,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1815202,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162544372?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5zQR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png 424w, https://substackcdn.com/image/fetch/$s_!5zQR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png 848w, https://substackcdn.com/image/fetch/$s_!5zQR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png 1272w, https://substackcdn.com/image/fetch/$s_!5zQR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9439248d-c41d-42c7-b861-944c2f37e171_1478x846.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Portrait Of Santiago Ram&#243;n y Cajal</em> by Joaquin Sorolla y Bastida (detail). </figcaption></figure></div><p>Ram&#243;n y Cajal loved to sketch. </p><p>As a boy, he was gripped by a <a href="https://thereader.mitpress.mit.edu/santiago-ramon-y-cajal-reflections-art-idealism-autobiography/">self-described</a> &#8216;irresistible mania&#8217; to draw. Whenever he saw a white wall, Cajal had no choice but to scribble on it &#8212; much to the chagrin of his parents who considered young Ramon&#8217;s interests in the arts to be something of an inconvenience. </p><p>The Spaniard used to cobble together enough pennies to buy art supplies and sneak out of the house. Sitting on a bank at the side of the road, he drew carts, horses, villagers, and anything else that took his fancy.</p><p>Born in 1852 in Petilla de Arag&#243;n in northeastern Spain, he grew up in a country still grappling with the aftershock of the Napoleonic Wars that dominated the opening decades of the century.</p><p>Napoleon famously called the Iberian campaign the &#8216;Spanish ulcer&#8217;. It was a wound that festered long after the French withdrawal, one that only showed signs of healing in the wake of the &#8216;Glorious Revolution&#8217; of 1868 which ended the tumultuous reign of Queen Isabel II.</p><p>Spanish science picked up just as Cajal began to get a taste for medical practice thanks to his father. The senior Ram&#243;n was a barber-surgeon. It was a profession that, since medieval times, tended to be the natural choice for minor surgeries and bloodletting.</p><p>Over his father&#8217;s shoulder, Cajal learned anatomy through dissecting and drawing cadavers. It was a grisly experience, but one that put him on the path to a medical career.</p><p>Following a stint as an army doctor that included a tour during Cuba&#8217;s first war of independence, the surgeon returned to Spain in 1877 to pick up a doctorate from the Complutense University of Madrid.</p><p>After a decade that included a bout with a nasty strain of tuberculosis and entry into the <a href="https://es.wikipedia.org/wiki/Caballeros_de_la_Noche">Caballeros de la Noche</a> Masonic lodge, Cajal moved to the Faculty of Medicine at the University of Barcelona. </p><p>In Catalonia, the Spaniard first saw early plates of neurons treated with colouring compounds that looked like &#8216;Chinese ink on transparent Japanese paper.&#8217; What had been an &#8216;inextricable network&#8217; when stained with carmine and hematoxylin became &#8216;simple, clear, and unconfused.&#8217;</p><p>Gripped by what he saw, Cajal did what he liked most of all: he drew. </p><p>In May 1888 he used the staining process and his sketch pad to argue that brain tissue was not composed of continuous connections. This idea, which rejected the view favoured by Italian physician Camillo Golgi, relied on Cajal&#8217;s ability to observe discrete neurons rather than nerves and brain tissue.</p><p>When Golgi used the cell staining technique, he interpreted the black-reaction silhouettes as evidence that neuronal processes fuse into a single protoplasmic web&#8212;a &#8216;reticulum&#8217; of continuity&#8212;rather than remaining distinct cells that happen to touch. But Cajal, who tweaked the recipe to use younger brain tissue, noticed tiny breaks between the cells.</p><p>The disagreement was bad tempered. When both Cajal and Golgi received a joint Nobel Prize in 1906, Golgi used the spectacle of the moment to drum up support for his position.</p><p>History sided with Cajal. Later confirmed by electron microscopy, the neuron doctrine eventually became the foundation of modern neuroscience.</p><h3>The neuron doctrine in AI</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I5rk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I5rk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg 424w, https://substackcdn.com/image/fetch/$s_!I5rk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg 848w, https://substackcdn.com/image/fetch/$s_!I5rk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!I5rk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I5rk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg" width="642" height="521.9512195121952" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:738,&quot;resizeWidth&quot;:642,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Santiago Ram&#243;n y Cajal - Linda Hall Library&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Santiago Ram&#243;n y Cajal - Linda Hall Library" title="Santiago Ram&#243;n y Cajal - Linda Hall Library" srcset="https://substackcdn.com/image/fetch/$s_!I5rk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg 424w, https://substackcdn.com/image/fetch/$s_!I5rk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg 848w, https://substackcdn.com/image/fetch/$s_!I5rk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!I5rk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F95167015-541f-4754-aa90-ea6b6286ae2d_738x600.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Drawing by Santiago Ram&#243;n y Cajal.</figcaption></figure></div><p>AI is not neuroscience, and artificial neural networks are not brains. But that doesn&#8217;t mean that, throughout history, AI researchers haven&#8217;t looked to neuroscience for inspiration to make sense of their creations. </p><p>Cajal&#8217;s work allowed Nicolas Rashevsky, a Russian-American biologist who was instrumental in creating the field of mathematical biology, to establish a mathematical foundation for understanding neurons. </p><p>Throughout the 1930s, he <a href="https://onlinelibrary-wiley-com.ezp.lib.cam.ac.uk/doi/pdfdirect/10.1002/jhbs.1094">developed</a> equations to articulate the processes by which neurons interact with each other at the University of Chicago, describing the functioning of nerve cells in the language of mathematics. </p><p>Using mathematical terms was a watershed moment: if real neurons could be described through equations, it followed that versions of their biological processes might be represented in artificial structures.</p><p>That moment came in 1943 with the introduction of the McCulloch-Pitts neuron, which the pair developed by directly building on Rashevsky&#8217;s work. In their blockbuster paper, &#8216;<a href="https://www.cs.cmu.edu/~./epxing/Class/10715/reading/McCulloch.and.Pitts.pdf">A Logical Calculus of the Ideas Immanent in Nervous Activity</a>,&#8217; Warren McCulloch and Walter Pitts (a student of Rashevsky) created a mathematical model that acted as a simplified abstraction of a biological neuron.</p><p>These ideas eventually led to the development of Frank Rosenblatt&#8217;s famous perceptron algorithm, the rise of parallel distributed processing techniques in the 1980s, and ultimately the giant neural nets that make ChatGPT, Claude, and Gemini tick. </p><p>But Cajal&#8217;s influence on the development of artificial neural networks wasn&#8217;t only indirect. Even in the 1980s, prominent research groups were publishing AI papers complete with Cajal&#8217;s drawings. Writing in 1986, researchers from Bell Labs <a href="https://opg.optica.org/viewmedia.cfm?uri=OAM-1988-THX2&amp;seq=0">explained</a> that:</p><blockquote><p> &#8216;circuit considerations will probably limit digital computer cycle times to a few tenths of a nanosecond, speeds that are within an order of magnitude of today's fastest machines. <strong>The way to further increase computing power is through parallel computation</strong>.&#8217; </p></blockquote><p>To make their case, the authors reproduced a rendering of a neuron drawn by Cajal. They reasoned that neurons receive input signals from other neurons through their dendrites, sum those signals together (weighing each based on the strength of its synaptic connections), and fire an output signal whose strength is determined by these input signals. </p><p>Such outputs cause the neuron's firing rate to increase proportionally for moderate input sums but saturate at maximum or minimum levels for extremely high or extremely low sums. When bundled into very large networks, the researchers argued, it is this computational structure that provides the complex information processing capabilities of the brain. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!c4Sm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!c4Sm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png 424w, https://substackcdn.com/image/fetch/$s_!c4Sm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png 848w, https://substackcdn.com/image/fetch/$s_!c4Sm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png 1272w, https://substackcdn.com/image/fetch/$s_!c4Sm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!c4Sm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png" width="904" height="440" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:440,&quot;width&quot;:904,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:111405,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.learningfromexamples.com/i/162544372?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!c4Sm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png 424w, https://substackcdn.com/image/fetch/$s_!c4Sm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png 848w, https://substackcdn.com/image/fetch/$s_!c4Sm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png 1272w, https://substackcdn.com/image/fetch/$s_!c4Sm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe4dc0935-6d48-4743-bf08-d1a4c6d14793_904x440.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Diagram of a neuron, originally drawn by Ram&#243;n y Cajal, reprinted in the Bell Labs paper &#8216;<a href="https://opg.optica.org/viewmedia.cfm?uri=OAM-1988-THX2&amp;seq=0">Electronic Neural Computing</a>&#8217; in 1986.</em></figcaption></figure></div><p>The weight of evidence eventually proved Cajal right, but he also told the best story by painting the best picture. The ink lines in his notebooks were persuading colleagues long before electron beams could. </p><p>Cajal reminds us that discovery is not just a matter of seeing but of showing. All things being equal, the scientist who can make plain truths pretty tends to be the one that wins. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.learningfromexamples.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.learningfromexamples.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item></channel></rss>