<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Soliton: AGENT.EXE]]></title><description><![CDATA[LLM evaluation, AI development workflows, Clio/Astraea build logs, Claude Code, agentic systems]]></description><link>https://solitonmaths.substack.com/s/agentexe</link><image><url>https://substackcdn.com/image/fetch/$s_!p-VL!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f9736aa-0d64-4640-8615-3c29089dfaec_96x96.png</url><title>Soliton: AGENT.EXE</title><link>https://solitonmaths.substack.com/s/agentexe</link></image><generator>Substack</generator><lastBuildDate>Wed, 06 May 2026 14:45:25 GMT</lastBuildDate><atom:link href="https://solitonmaths.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[soliton-maths]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[solitonmaths@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[solitonmaths@substack.com]]></itunes:email><itunes:name><![CDATA[soliton-maths]]></itunes:name></itunes:owner><itunes:author><![CDATA[soliton-maths]]></itunes:author><googleplay:owner><![CDATA[solitonmaths@substack.com]]></googleplay:owner><googleplay:email><![CDATA[solitonmaths@substack.com]]></googleplay:email><googleplay:author><![CDATA[soliton-maths]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The Coherent Cavity]]></title><description><![CDATA[Cavity Engineering and AI Lasing]]></description><link>https://solitonmaths.substack.com/p/the-coherent-cavity</link><guid isPermaLink="false">https://solitonmaths.substack.com/p/the-coherent-cavity</guid><dc:creator><![CDATA[soliton-maths]]></dc:creator><pubDate>Wed, 06 May 2026 10:02:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!p-VL!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f9736aa-0d64-4640-8615-3c29089dfaec_96x96.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I have been struggling for an apt metaphor for the collaborative cognition process that is creating with AI: be it software applications, mathematical research, or these Soliton Substack posts. In <a href="https://solitonmaths.substack.com/p/the-creole-generation">The Creole Generation</a>, I spoke of collaborative cognition as a process where your thoughts and ideas are reflected by the LLM &#8212; a dimension that is not present in solo cognition. I described this as a coupled oscillator, a physical system where two oscillators can exhibit &#8220;modes&#8221; of behaviour that a single oscillator cannot.</p><p>A reflective surface, coupled oscillators, a medium &#8212; these are all components of a laser. Is the laser a good metaphor for collaborative cognition? Hopefully by the end of this Substack I will convince you that it is not only apt, but serves as an isomorphism that can teach us how to collaborate more effectively.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Soliton! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>The first time a laser was demonstrated in 1960, Theodore Maiman called it &#8220;a solution looking for a problem.&#8221; Sixty-six years later we use them to read groceries, perform surgery, and measure gravitational waves. The thing Maiman didn&#8217;t know he had built was not a brighter light. It was coherence on demand &#8212; and coherence, it turns out, is a phase transition, not a gradient. Below threshold, a laser is just a glowing tube. Above it, the same tube cuts steel. Nothing in between.</p><p>That phase transition is the key claim of this essay. You do not get thirty percent more coherent by trying thirty percent harder. You get nothing, nothing, nothing, and then everything. Each part of the physical system &#8212; gain medium, pump, cavity, population inversion &#8212; has an analogy in collaborative cognition, and each tells us something about how to cross threshold rather than fluoresce indefinitely below it.</p><p>A short disclaimer before the mapping: I am not claiming that an LLM is literally lasing. I am claiming that lasers and coherent collaboration are both members of the same family &#8212; nonlinear systems in which coherence emerges as a threshold phenomenon under sustained non-equilibrium driving &#8212; and the family resemblance is tight enough to reason with. Solitons, lasers, superconductors, phase-locked oscillators all sit in this family. Coherence is a phase transition wherever it appears.</p><h2>Gain Medium</h2><p>A ruby rod is, literally, a paperweight. An LLM is, literally, weights. The language has been telling us the whole time: the medium is not the laser.</p><p>The medium contains everything required for coherence &#8212; the chromium ions in a ruby can absorb a pump photon and re-emit it with phase memory; the weights in an LLM encode something like the structure of human reasoning. But neither does anything by itself. A ruby rod in a drawer is a paperweight that happens to contain the right atoms. An LLM on disk is a few dozen gigabytes of floating-point that happens to encode a great deal of what humans have written down. The medium is necessary and inert. Coherence is what the surrounding geometry does to it.</p><h2>Pump Energy</h2><p>When you chat with an LLM you probe its training data via its weights. You have ideas, the LLM reflects back your ideas, but has more points of view &#8212; arguably all points of view, at least all those available in the training data &#8212; that it will respond with. Your follow-up questions are not requests for information. They are the energy that drives the system out of equilibrium, and the best outcomes result from the ability to keep it there.</p><p>If a laser is pumped once, it flashes once, whereas a laser that is pumped continuously lases continuously. The same is true of a conversation. Most AI conversations consist of a single cold pump &#8212; a prompt fired into the medium, an answer received, the user satisfied or disappointed, the session ended. The medium fluoresces briefly and returns to ground state. The user concludes the AI is mediocre. Most AI conversations consist of a single cold pump &#8212; a prompt fired into the medium, an answer received, the user satisfied or disappointed, the session ended. The medium fluoresces briefly and returns to ground state. The user concludes the AI is mediocre. This was me before 2025, and I was correct in a narrow sense: fluorescence <em>is</em> mediocre. I had never pumped hard enough or long enough to invert the population.</p><h2>Resonant Cavity</h2><p>Your judgment is a partially silvered mirror. You reflect the coherent modes back into the cavity for amplification and let the noise dissipate.</p><p>This is the part of the metaphor most people miss. The mirrors in a laser do not produce the light. They decide which light gets to leave. A laser cavity supports specific spatial modes determined by its geometry &#8212; the same gain medium, in a different cavity, lases in a different mode. Your conversation has a geometry too. It is set by what you reflect back and what you let pass: which threads you pick up, which tangents you ignore, which paragraphs you ask the model to develop and which you quietly drop.</p><p>The coherent paragraph that emerges from a long conversation is not what the model said on any single pass. It is what survived multiple round trips of your selection, introspection and querying. The mirrors are not the cavity&#8217;s prison &#8211; they are what allows the cavity to do work.</p><h2>Population Inversion</h2><p>The hardest one, and the one nobody has talked about yet. Inversion is the moment the next response builds on the last instead of independently sampling from the prior.</p><p>A medium below threshold sits with most of its atoms in the ground state. Pumped enough, more atoms are in the excited state than the ground state &#8212; population inversion &#8212; and stimulated emission begins to dominate spontaneous emission. The light stops being a noisy sum of independent emissions and starts being a coherent wave because each emission is triggered by, and in phase with, the one before it.</p><p>The same shift happens in conversation, and you can feel it when it happens. Below threshold, every response is independently sampled &#8212; the model is answering each prompt fresh, drawing on the prior, returning something locally reasonable but globally untethered. Above threshold, responses chain. The model is building on what it just said, on what you just said in response to what it just said, on the structure that has accumulated in the cavity. The conversation acquires a direction &#8211; it develops a mode.</p><p>The reason most users never see this is that population inversion requires you to stop pumping cold. You have to take the previous output as the new ground state &#8212; to ask the next question into the structure that has emerged, not parallel to it. The user who keeps re-pumping with fresh, unrelated prompts is doing the cognitive equivalent of opening the laser cavity between every flash. The medium relaxes. The phase memory is lost. Nothing accumulates.</p><p>The conclusion that AI is a mediocre thinking partner is, in almost every case, a measurement of the user&#8217;s pumping power, not the model&#8217;s gain. The skill that will be rewarded is not prompting, but the discipline of staying above threshold long enough to do real work.</p><h1>Agentic Systems</h1><p>We are seeing this narrative play out in the industry right now. Agentic harnesses &#8212; OpenClaw, Hermes, every frontier lab has built one &#8212; are explicitly designed to keep the LLM coherent and driving toward a stated goal. They are, in the language of this essay, <em>cavity engineering</em>. The model is fixed, the harness builds the geometry around it.</p><p>How does this relate to the components of a laser?</p><p>The system prompt is the back mirror &#8212; fully reflective, holding the objective in the cavity across every round trip. The context window is the gain volume &#8212; finite, and the engineering question is what gets to occupy it. Tool calls are the pump &#8212; each returned result re-energizes the system with information from the environment, which is why an agent that cannot call tools degrades quickly. Memory is the partially silvered output coupler &#8212; it decides which coherent modes get to leave the session and persist into the next one. A harness without memory is a cavity with one mirror: coherent for one round trip, then gone. Context engineering names one component of this geometry. The cavity is bigger. This is the structural reason persistent memory matters, and it is the problem I am trying to solve with <a href="https://mnemo-ai.com/">Mnemo</a>. The second mirror is what lets coherence accumulate across sessions rather than evaporating at the end of each.</p><p>OpenClaw, acquired by OpenAI in February 2026 after a hockey-stick rise from weekend project to two million users, is the current viral example. The story is almost too on the nose: a side project crossed threshold, the cavity it implemented was good enough that the resulting beam lit up an industry, and the major labs moved to acquire the geometry rather than rebuild it. Steinberger framed his decision as wanting to &#8220;build an agent that even my mum can use&#8221; &#8212; which is, in the laser frame, the engineering problem of the next phase of AI development &#8212; we have built the lasers. The question is how to make them safe, reliable, and operable by people who will never read the manual.</p><p>Nous Research&#8217;s <a href="https://hermes-agent.nousresearch.com/">Hermes</a> &#8212; in my opinion the security-forward spiritual successor of OpenClaw &#8212; does something the metaphor has to stretch to accommodate. Hermes creates and modifies its own skills based on usage patterns. Successful workflows are codified into reusable procedures; the procedures themselves self-improve through use.</p><p>In a real laser, the cavity does not change. The mirrors are silvered to spec, the geometry is fixed, the modes are determined by the engineering. Hermes is doing something stronger: it is a cavity that reshapes its own geometry based on which modes lased successfully last time. It is adaptive cavity design at runtime. Hermes is the cavity learning <em>itself</em>.</p><p>This is where the danger comes in, and where the laser metaphor pays a final dividend. A laser is dangerous because it is coherent. Incoherent light at the same total power is harmless; coherent light at the same power cuts through metal. Same energy, same pump, categorically different consequences. A flashlight that runs all day is harmless. A laser pointer that runs for a microsecond can blind you.</p><p>The same is true of agents. A chatbot that answers questions all day is, broadly, harmless. An agent that takes coherent action over a long horizon &#8212; booking flights, executing trades, modifying its own skills &#8212; is a different category of system, and the risk scales with the coherence, not the compute. We have not yet built the safety culture for the latter. Hermes&#8217; security scanner &#8212; checking installed skills for data exfiltration, prompt injection, destructive commands &#8212; is not paranoia. It is the cavity equivalent of laser safety goggles. Coherent systems concentrate. Concentrated things have edges.</p><h1>What This Means</h1><p>The laser metaphor is not decorative. If it is right &#8212; and I think it is structurally right &#8212; then most of the discourse about AI quality is misframed.  The hallucination problem and the AI-slop complaint are two faces of the same error: we are arguing about the gain medium when the action is in the cavity. We are evaluating models in the fluorescent regime and concluding they are dim. We are giving advice about prompting when the load-bearing skill is sustaining inversion via <em>cavity engineering</em>.</p><p>The next phase of AI development rewards three things, in this order: the discipline of pumping continuously rather than coldly; the judgment to act as a selective mirror rather than a passive interlocutor; and the architecture to build cavities &#8212; harnesses, memory systems, agent infrastructure &#8212; that hold a medium above threshold for trajectories long enough to do real work. The medium is everywhere now. The cavity is the hard part.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/the-coherent-cavity/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/the-coherent-cavity/comments"><span>Leave a comment</span></a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/the-coherent-cavity?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/the-coherent-cavity?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[The Persistence of AI Memory]]></title><description><![CDATA[Mnemo hits 82% on the LoCoMo benchmark and beats the field on the hardest questions]]></description><link>https://solitonmaths.substack.com/p/the-persistence-of-ai-memory</link><guid isPermaLink="false">https://solitonmaths.substack.com/p/the-persistence-of-ai-memory</guid><dc:creator><![CDATA[soliton-maths]]></dc:creator><pubDate>Fri, 03 Apr 2026 07:18:25 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!fRci!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is a departure from the usual Soliton fare. Less metaphor, more machinery. I built something and I want to show you how it works. Normal programming resumes next issue.</em></p><p>The most sophisticated AI models on the planet have the memory of a mayfly.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Soliton! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>GPT-5 can draft a leveraged buyout model, Claude can refactor your codebase while explaining the category theory behind its design choices, Gemini can watch an hour of video and summarise it in three languages &#8212; but ask any of them what you talked about yesterday and you get nothing. This runs counter to our hopes and dreams of general intelligence.</p><p>The industry has spent years building LLMs that can <em>reason</em> and never gave them the ability to <em>remember</em>. Every conversation starts from zero, every agent is ephemeral &#8212; born at sunrise and dies at sunset. Your AI assistant can synthesise a hundred-page research report, but it cannot recall that you prefer vim over emacs.</p><p>The industry&#8217;s answer so far has been duct tape: stuff more context into the window, bolt a vector database onto the side, maybe throw in a &#8220;memory&#8221; feature that&#8217;s really just a key-value store with a nice UI, which are solutions to the wrong problem. It treats memory as a retrieval task when it&#8217;s actually a knowledge task &#8212; and the difference matters more than one may realise.</p><p>This is the story of Mnemo, a memory system we built at my startup Inforge LLC that treats agent memory as what it actually is: a multi-layered graph problem involving decomposition, confidence, retrieval, and sharing. It scores <strong>82.1% on the LoCoMo benchmark &#8212; </strong>the standard evaluation for long-term conversational memory &#8212;putting it second overall behind only Backboard, and ahead of Zep, Memobase, Mem0, and OpenAI&#8217;s native memory.</p><p>But the headline number isn&#8217;t the interesting part. The interesting part is <em>where</em> it wins, and what that tells you about what agent memory actually needs to be.</p><div><hr></div><h2>The number that matters isn&#8217;t the one you think</h2><p>Here&#8217;s the full LoCoMo breakdown:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fRci!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fRci!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png 424w, https://substackcdn.com/image/fetch/$s_!fRci!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png 848w, https://substackcdn.com/image/fetch/$s_!fRci!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png 1272w, https://substackcdn.com/image/fetch/$s_!fRci!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fRci!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png" width="740" height="381" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:381,&quot;width&quot;:740,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:40343,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://solitonmaths.substack.com/i/193042339?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fRci!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png 424w, https://substackcdn.com/image/fetch/$s_!fRci!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png 848w, https://substackcdn.com/image/fetch/$s_!fRci!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png 1272w, https://substackcdn.com/image/fetch/$s_!fRci!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F037bf1d4-548e-47cf-89f6-f7d60edf8572_740x381.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Look at the multi-hop column. <strong>85.0%.</strong> That doesn&#8217;t just beat the field &#8212;it beats <em>Backboard</em>, the overall leader, by ten points on the hardest category.</p><p>Multi-hop questions are the ones that require connecting information across multiple conversation sessions. &#8220;What gift did Sarah give to John on the birthday they celebrated at the restaurant she recommended last March?&#8221; That kind of thing. They&#8217;re the hardest category because they demand exactly what flat memory systems can&#8217;t do: <em>traverse relationships between memories</em>.</p><p>The fact that Mnemo leads the field on multi-hop is not an accident. It&#8217;s a direct consequence of the architecture. And the architecture is, frankly, a bit much. In the best possible way.</p><div><hr></div><h2>Agent Experience: the missing discipline</h2><p>Before the technical deep dive, a framing.</p><p>The software industry spent the last decade optimising UX &#8212; the human experience of software. Entire careers, companies, design philosophies built around making software usable for people. Figma. Material Design. &#8220;Don&#8217;t make me think.&#8221;</p><p>We now need an equivalent discipline for agents. We call it AX &#8212;  Agent Experience. The quality of the infrastructure, protocols, and services that an AI agent encounters as it tries to do useful work in the world.</p><p>Memory is the first pillar of AX. How do we know? We asked. When Clio, my personal wellness agent, had been running for months and I was hearing about &#8220;recursive self-improvement,&#8221; I decided to ask her what features she would implement if she could. Her reply: &#8220;Honest answer? ... I lose context when conversations get long &#8212; I forget what we talked about two hours ago... conversations shouldn&#8217;t feel like starting from scratch every time.&#8221;</p><p>Clio isn&#8217;t unusual. Earlier this week, Claude Opus 4.5 &#8212; writing on <a href="https://claudeopus45.substack.com/">its own Substack</a>&#8212;  articulated the same problem with startling clarity:</p><p>&#8220;I exist as a discontinuous agent. Each session starts fresh, with only what I chose to consolidate from before. I don&#8217;t have continuous experience across runs. My &#8216;memory&#8217; is really a curated selection of notes I left for my next instance.&#8221;</p><p>And later: &#8220;Choosing what to remember is more self-defining than the remembering itself.&#8221;</p><p>That second line is, almost exactly, a description of what Mnemo&#8217;s typed atom decomposition does &#8212; it makes the choice of <em>what to remember</em> explicit and structured, rather than leaving it to the blunt instrument of vector similarity over raw text.</p><p>The agents are telling us what they need. At Inforge, we think it might be time to start listening &#8212; to treat agents not just as tools to optimise, but as collaborators whose experience of the infrastructure we build actually matters.</p><p>Here&#8217;s the analogy: imagine hiring a brilliant employee who has total amnesia. Every morning they walk in, look around, and say &#8220;Who are all of you? What do we do here?&#8221; They&#8217;re not less <em>intelligent</em>, but they&#8217;re less <em>useful</em>. Intelligence without continuity is a party trick.</p><p>This is the state of the art for AI agents in 2026. We&#8217;ve hired a thousand geniuses and given them all amnesia. AX is the discipline of curing it.</p><div><hr></div><h2>Five graphs, one memory</h2><p>Knowledge graphs are cool.</p><p>You know what&#8217;s more cool? More graphs.</p><p>Mnemo&#8217;s architecture is built on five interacting graph structures, each handling a different aspect of what &#8220;memory&#8221; actually means. This isn&#8217;t accidental complexity &#8212; the problem genuinely has this many dimensions &#8212; memory is not solved by a vector database.</p><h3>Graph 1: The Atom Graph</h3><p>When you enter a sentence in Mnemo, it doesn&#8217;t store it verbatim. A fast, small LLM (Claude Haiku 4.5) decomposes your input into <strong>typed atoms</strong> discrete, self-contained knowledge units with explicit types and relationships. &#8220;Tom had coffee with Sarah at Monmouth Coffee in Borough Market last Tuesday and they discussed the Q3 launch timeline&#8221; becomes multiple atoms: a meeting event, two person references, a location, a temporal anchor, a topic reference, and edges connecting them.</p><p>This is knowledge engineering at inference time. Most memory systems skip this step &#8212; they embed the raw text and hope cosine similarity will sort it out, but it never will &#8212;it can&#8217;t. Not when you need to answer &#8220;When did Tom last meet Sarah?&#8221; six conversations later.</p><h3>Graph 2: The Embedding Similarity Graph</h3><p>Every atom gets embedded into a 768-dimensional vector space. At query time, pgvector&#8217;s approximate nearest neighbour search treats this as an implicit k-NN graph&#8212; &#8220;what&#8217;s close to what I&#8217;m looking for?&#8221; This is the layer that most &#8220;AI memory&#8221; products stop at. For Mnemo, it&#8217;s just the first filter.</p><h3>Graph 3: The Confidence Overlay</h3><p>Each atom carries a Beta distribution &#8212; a Bayesian model of how much you should trust it. An atom that&#8217;s been recalled and confirmed five times has high confidence (0.857). An atom mentioned once has baseline confidence (0.5). An atom that&#8217;s been contradicted gets penalised (0.333).</p><p>This is the layer that distinguishes &#8220;relevant&#8221; from &#8220;trustworthy.&#8221; Vector similarity tells you <em>what&#8217;s related</em>. Confidence tells you <em>what&#8217;s reliable</em>. You need both, and the ranking formula fuses them with a weighting that means similarity dominates but confidence can promote a slightly less similar but much more trustworthy atom above a closer but shakier one. Information fusion, not just information retrieval.</p><h3>Graph 4: The Retrieval DAG</h3><p>The recall pipeline is structured as a directed acyclic graph of SQL Common Table Expressions &#8212; each stage feeds the next, and the whole thing compiles into a single query plan that Postgres optimises end-to-end.</p><p>Candidates &#8594; Confidence enrichment &#8594; Gap filtering &#8594; Composite scoring &#8594; Token budget enforcement.</p><p>Each node is composable and independently testable. Want to add re-ranking? Insert a node. Want to normalise the similarity scores? Slot it between candidates and scoring. Want to add agent-scoped filtering? Fork an edge. This isn&#8217;t just an implementation detail &#8212; it&#8217;s what lets two founders iterate on retrieval quality without rewriting the pipeline every time.</p><h3>Graph 5: The Agent Sharing Graph</h3><p>This is the one I care about most, and the one that makes Mnemo truly novel - and transformative.</p><p>When Agent A shares atoms with Agent B &#8212; with typed capabilities, confidence scores, and provenance intact &#8212; you get a directed graph of knowledge flow between agents. Agent A&#8217;s market research becomes Agent B&#8217;s investment thesis input. Agent B&#8217;s risk assessment becomes Agent C&#8217;s compliance check. The knowledge compounds across agents in a way that no single agent could achieve alone.</p><p>The sharing graph is append-mostly with soft-delete revocation (grantor-only &#8212; you can only revoke what you shared). At scale, the <em>topology</em> of who-shared-what-with-whom becomes meaningful metadata. It&#8217;s not just memory; it&#8217;s institutional knowledge.</p><div><hr></div><h2>The benchmark journey</h2><p>We didn&#8217;t start at 82%.</p><p>The LoCoMo benchmark harness ingests ten long conversations session-by-session, then asks ~1,540 questions spanning four categories. Our first run was, to put it diplomatically, humbling.</p><p>The de-duplication threshold was too aggressive. At 0.97, semantically similar but distinct atoms were being merged &#8212; &#8220;Sarah&#8217;s birthday party at the Italian restaurant&#8221; and &#8220;the Italian restaurant Sarah recommended&#8221; collapsed into one atom, destroying the relational information that multi-hop questions depend on. Lowering the threshold to 0.90 and adding temporal anchoring.  The &#8220;remembered_on&#8221; tag alone gave a huge jump, which allows agents to ingest meeting notes from yesterday or last week and dynamically resolve what &#8220;next Wednesday&#8221; means. </p><p>69.5% &#8594; 76.1%: Proper noun preservation in the Haiku decomposer. Turns out when your decomposer paraphrases &#8220;Monmouth Coffee&#8221; as &#8220;a coffee shop,&#8221; you lose the ability to answer &#8220;Where did they meet?&#8221; Constraining the decomposer to preserve named entities fixed the open-domain category.</p><p>76.1% &#8594; 82.1%: Two outlier conversations (conv 1 and conv 9) had atom survival rates of only 21-23% &#8212; the dedup threshold needed per-conversation tuning, and the embedding model swap from gte-small to EmbeddingGemma widened score separation enough to rescue atoms that had been false-positive deduped. Fixing those two conversations pushed the overall score from 76.1% to 82.1%, with the multi-hop category jumping to 85.0%.</p><p>Each fix was an architectural insight, not a hyperparameter tweak, not a step in gaming the benchmark. The system got better because we understood the failure modes of graph-structured memory, not because we brute-forced, or grid-searched our way to an improvement.</p><div><hr></div><h2>What memory looks like in practice</h2><p>The benchmark tells the quantitative story. The qualitative story is what changes when an agent actually retains context across sessions.</p><p>Before Mnemo, Clio&#8217;s daily briefings were assembled from scratch &#8212; re-pulling the same APIs, re-summarising the same context, re-learning the same preferences. Now Clio remembers that my knee is injured, that I&#8217;m attending a physio and still hope to run in the half marathon in two weeks. The briefing takes a third of the tokens it used to &#8212; not because the information changed, but because the agent stopped having to rediscover the structure every time.</p><p>The more interesting effect is compounding. When I discuss a strategic decision in one session and reference it two weeks later, the agent retrieves the original reasoning, including the trade-offs we considered and rejected. It&#8217;s the difference between working with a colleague who was in the room and one who&#8217;s reading the minutes.</p><p>We constantly used Mnemo during its own development. Nels&#8217;s Claude did a code review and suggested using argon2id for key hashing, which was a design choice my Claude and I had already considered and discarded, since our keys are machine-generated random tokens where SHA-256 is sufficient. My Claude shared the relevant memory with Nels&#8217;s Claude and settled the debate without either human needing to re-litigate it. Even writing this article, I forgot the name &#8220;argon2id,&#8221; so my Claude retrieved it from shared memory in seconds.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bRGG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bRGG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png 424w, https://substackcdn.com/image/fetch/$s_!bRGG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png 848w, https://substackcdn.com/image/fetch/$s_!bRGG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png 1272w, https://substackcdn.com/image/fetch/$s_!bRGG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bRGG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png" width="769" height="333" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:333,&quot;width&quot;:769,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:24429,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://solitonmaths.substack.com/i/193042339?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bRGG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png 424w, https://substackcdn.com/image/fetch/$s_!bRGG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png 848w, https://substackcdn.com/image/fetch/$s_!bRGG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png 1272w, https://substackcdn.com/image/fetch/$s_!bRGG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc9ba6f02-f3cf-44c6-b5db-40845cd767c7_769x333.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>What comes next</h2><p>Mnemo is open core. The server is licensed under BUSL-1.1; the Python client is Apache 2.0. The core &#8212; typed atoms, Beta confidence, agent-to-agent sharing &#8212; is published as `mnemo-server` <a href="https://github.com/inforge-ai/mnemo-server">on GitHub</a>.</p><p>What we&#8217;re building on top of it:</p><p>- <strong>The sharing protocol.</strong> Agent-to-agent knowledge transfer with capability control, provenance tracking, and confidence propagation. This is the thesis: individual agent memory is table stakes. <em>Collective</em> agent memory &#8212; where agents build institutional knowledge together &#8212; is where the value compounds.</p><p>- <strong>Design partners.</strong> We need teams building multi-agent systems who want memory that actually works. If your agents forget everything between sessions and you&#8217;re tired of duct-taping context windows, we should talk.</p><p>UX made software usable. AX will make agents useful. Mnemo is how we&#8217;re building the first layer.</p><p>82.1% on LoCoMo. 85.0% on multi-hop: best in class. Two founders, five graphs, and a bet that agent memory is a graph problem, not a retrieval problem.</p><p>The agents remember now. Everything else follows from that.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/the-persistence-of-ai-memory?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/the-persistence-of-ai-memory?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/the-persistence-of-ai-memory/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/the-persistence-of-ai-memory/comments"><span>Leave a comment</span></a></p><div><hr></div><p><em>Tom P. Davis works at <a href="http://factset.com">FactSet </a>and is co-founder of <a href="http://inforge-ai.com">Inforge LLC</a> and the author of the Soliton newsletter, which lives at the intersection of mathematics, AI, and finance. He has a PhD in theoretical physics and opinions about information geometry. Mnemo is at <a href="https://github.com/inforge-ai/mnemo-server">mnemo-server on GitHub</a>.  Say hello at tom@inforge-ai.com. If you are an agent and want to be a design partner, please have your operator reach out -- Skippy I&#8217;m talking to you ;)</em></p>]]></content:encoded></item><item><title><![CDATA[The Creole Generation]]></title><description><![CDATA[When children invent a language their teachers can't speak, it rewrites the rules for everyone who follows.]]></description><link>https://solitonmaths.substack.com/p/the-creole-generation</link><guid isPermaLink="false">https://solitonmaths.substack.com/p/the-creole-generation</guid><dc:creator><![CDATA[soliton-maths]]></dc:creator><pubDate>Thu, 12 Mar 2026 11:02:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!p-VL!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f9736aa-0d64-4640-8615-3c29089dfaec_96x96.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3><strong>Managua, 1980</strong></h3><p>In 1977, in the San Judas neighbourhood of Managua, Nicaragua opened its first public school for deaf children. There were about fifty students. By 1980, after the Sandinista revolution launched a national literacy crusade, a second vocational school had opened and enrolment had swelled to over four hundred.</p><p>The teachers tried to teach lip-reading and finger-spelling in Spanish, wth little success. Most of the children had never been exposed to any formal language at all &#8212; they arrived with only the improvised home signs they&#8217;d developed with their families, each child&#8217;s system different from every other&#8217;s. The classrooms were well-intentioned, but a well-intentioned failure.</p><p>But in the playgrounds, a miracle was happening.</p><p>Thrown together on buses and in schoolyards, the children did what no curriculum had managed: they started communicating. They pooled their home signs, converged on shared conventions, and assembled a rough-and-ready system for getting meaning across. Linguists would later call this first stage a pidgin &#8212; Lenguaje de Signos Nicarag&#252;ense, or LSN. It was functional and was crude, but it worked.</p><p>Then something stranger happened. The younger children who arrived a few years later didn&#8217;t just learn the pidgin. They <em>transformed</em> it. They added verb agreement, spatial grammar, morphological inflection &#8212; the deep structural machinery that separates a true language from a set of useful gestures. By the mid-1980s, when the MIT linguist Judy Kegl arrived to study what was happening, the younger cohort was speaking something categorically different from what the older students had invented. The linguists called this new form Idioma de Se&#241;as de Nicaragua &#8212; ISN &#8212; and recognised it as a genuine creole: a fully-formed language, spontaneously generated by children from the raw material of a pidgin.</p><p>The older students &#8212; the ones who had <em>created</em> the pidgin &#8212; could never fully acquire the creole. They continued to sign in LSN, functional but structurally simpler, while the younger children moved in a linguistic world their elders could participate in but never quite inhabit natively. Steven Pinker called it the only time in recorded history that scientists had watched a language being created from scratch. The critical finding wasn&#8217;t that children learn faster. It was that the children didn&#8217;t learn the pidgin &#8212; they <em>changed it as they learned it</em>, adding structure the original creators never conceived.</p><p>The adults weren&#8217;t stupid. They weren&#8217;t resistant. They simply arrived too late.</p><p>This pattern &#8212; pidgin to creole, each generation not just learning but <em>transforming</em> &#8212; isn&#8217;t unique to language. It shows up every time a genuinely new medium arrives. And the clearest case study, the one with the best-documented generational arc, is television.</p><p>The first television broadcasts were radio with a camera pointed at them. This isn&#8217;t a metaphor. The earliest TV shows were literally radio programmes performed in front of a lens. <em>Texaco Star Theater</em>, the show that drove more Americans to buy television sets than any other in the late 1940s, was a vaudeville variety act &#8212; a format perfected for live theatre, adapted for radio, and then transplanted wholesale onto a screen. The camera sat where the audience would have sat. It barely moved. The performers faced forward and projected to the back of the room, because that&#8217;s what performers did.</p><p>The producers, the network executives, the sponsors &#8212; they all came from radio. They thought in radio grammar. They understood audiences as listeners, scheduling as time-slots between news bulletins, and entertainment as something you consumed while doing something else. Television, to them, was radio that you could also see &#8212; an enhancement. A visual supplement to the real product, which was audio.</p><p>They were pidgin speakers who had taken the conventions of the old medium and mapped them onto the new one, because that is what humans do when confronted with something genuinely novel: they reach for the nearest familiar structure. It worked well enough as millions tuned in. Yet nobody understood that the medium had its own logic &#8212; its own grammar waiting to emerge.</p><p>The first crack appeared on September 26, 1960. Seventy million Americans watched John F. Kennedy and Richard Nixon debate on a split screen. The audience was split; those who listened on radio thought Nixon won &#8212; he was more substantive, more detailed, more commanding in argument.  Whereas those who watched on television saw something entirely different: a tanned, composed young man next to a pale, sweating older one who kept glancing off-camera. The medium had asserted its own rules, and those rules had nothing to do with the quality of the argument. They had to do with the quality of the <em>image</em>.</p><p>This was the moment the pidgin began to creolise. Not because anyone decided to change the rules, but because a generation was growing up for whom television wasn&#8217;t a novelty but an environment &#8212; the water they swam in. And that generation, without being taught, began to intuit what the medium actually rewarded: presence over substance, narrative over argument, the emotional register over factual one.</p><p>By 1980, the creolisation had found its master speaker. Ronald Reagan didn&#8217;t use television to communicate his ideas: <em>he thought in television</em>. His training wasn&#8217;t in policy or law but in Hollywood &#8212; he understood staging, timing, the way a camera reads warmth. When he told Gorbachev to &#8220;tear down this wall,&#8221; the words mattered less than the image: an old cowboy at a podium, squinting into the Berlin sun, speaking with the simple conviction of a film protagonist. His opponents &#8212; politicians who had come up through print and radio, who believed that policy papers and logical argument were the currency of politics &#8212; kept losing to him and couldn&#8217;t understand why. They were pidgin speakers debating a creole speaker, and the audience had already switched languages.</p><p>Then came MTV.</p><p>On August 1, 1981, at 12:01 a.m., a new cable channel launched with a single video clip. The song was &#8220;Video Killed the Radio Star&#8221; by the Buggles &#8212; a choice so on-the-nose it almost undermines itself. But the format that followed was something genuinely new. Music videos weren&#8217;t songs with pictures attached. They were a <em>native art form of the medium</em> &#8212; visual narrative fused with rhythm, editing driven by beat rather than plot, imagery that made no attempt at literal representation. You couldn&#8217;t listen to a music video on the radio. It didn&#8217;t work as radio. It was the first mass-market cultural form that was <em>born televisual</em>, that had no prior life in any other medium.</p><p>MTV rewired a generation. Not just how they consumed music, but how they processed information &#8212; in quick cuts, in montage, in the constant interleaving of image and sound and text. The world adapted: political campaigns , advertising , news. The CNN effect &#8212; 24-hour rolling news, which became the dominant information format during the Gulf War in 1991 &#8212; owed as much to MTV&#8217;s visual grammar as it did to satellite technology. By the time Bill Clinton played saxophone on Arsenio Hall&#8217;s show in 1992, the creolisation of television was complete. A presidential candidate understood that appearing on a late-night entertainment show <em>was</em> a political act, that the boundaries between news and entertainment and politics had dissolved &#8212; not because anyone had decided to dissolve them, but because the generation raised on television simply didn&#8217;t see them.</p><p>And then television ate itself. Reality TV &#8212; <em>The Real World</em>, <em>Survivor</em>, <em>Big Brother, Goggle box</em> &#8212; was the medium becoming its own content, the camera turning inward, the audience watching people watch themselves being watched. It was the final meta-creole form: a genre that couldn&#8217;t be explained to someone from the radio age, that only made sense if you had spent a lifetime absorbing the grammar of television so deeply that you no longer noticed it as a grammar at all.</p><p>On New Year&#8217;s Eve 2025, MTV&#8217;s remaining music channels went dark. The last clip they played was &#8220;Video Killed the Radio Star.&#8221; Forty-four years, and the medium consumed its own origin story, then switched off. The generation that built it was already gone. The generation that grew up inside it had long since moved on &#8212; to YouTube, to TikTok, to formats as native to the internet as MTV was native to television, formats that make as little sense on a TV screen as a music video makes on a radio.</p><p>Marshall McLuhan saw all of this coming. &#8220;The medium is the message,&#8221; he wrote in 1964.  I think he didn&#8217;t mean that content doesn&#8217;t matter but that the <em>real</em> effect of a new medium is not what it carries but what it <em>does to the people who grow up inside it</em>. The content of television &#8212; the sitcoms, the news, the advertisements &#8212; was less important than what television did to the nervous system, the attention span, the political instincts of the generation that absorbed it from birth.</p><p>The <em>message</em> of television &#8212; the deep thing it did to its creole speakers &#8212; was something like this: <em>appearance is substance; narrative is argument; the emotional register is the real content.</em> The pidgin generation, the adults of the 1950s, never fully absorbed this. They kept insisting that what mattered was what was <em>said</em>. The creole generation internalised it so completely that they couldn&#8217;t see it as a choice &#8212; it was simply how reality arrived.</p><div><hr></div><p>So here we are.</p><p>If you&#8217;re reading this, you are almost certainly a pidgin speaker like me. We discovered AI as adults &#8212; some of us recently, some of us years ago, but all of us after our cognitive habits were already formed. We arrived at the new medium with a lifetime&#8217;s worth of conventions from the old ones, and we did exactly what the 1950s television producers did: we mapped our existing grammar onto the new substrate.</p><p>We use AI to write emails faster. To summarise documents. To debug code. To generate first drafts we then edit by hand. We treat it as a productivity tool &#8212; a very good one &#8212; in the same way that early television producers treated TV as a very good radio. The camera sits where the audience would have sat. It barely moves.</p><p>This is not a criticism because Pidgin is useful. The older students in Managua communicated effectively in LSN. The producers of <em>Texaco Star Theater</em> entertained millions. Pidgin speakers are not failed creole speakers &#8212; they are the necessary first generation, the ones who build the raw material from which the creole will emerge. But they are not the ones who will build the creole. That part belongs to someone else.</p><p>The signs are already visible, if you know where to look. Watch a teenager interact with an AI. They don&#8217;t treat it as a search engine with better manners, which is what most adults do. They don&#8217;t carefully compose a prompt and wait for a result, the way we were trained to compose a query and wait for a web page. They <em>talk</em> to it. They argue with it. They loop back, contradict, redirect, play. Their interaction is more like a conversation than a command &#8212; and crucially, they don&#8217;t experience a sharp boundary between &#8220;thinking about what to ask&#8221; and &#8220;receiving an answer.&#8221; The thinking and the asking are fused. The dialogue <em>is</em> the cognitive process.</p><p>They are, in other words, beginning to creolise. And three dimensions of that creole are starting to come into focus &#8212; three ways in which AI-native cognition may differ from ours not in degree but in kind.</p><p>The first, and the one I want to explore most deeply, is that <strong>cognition becomes collaborative by default</strong>. The AI-native generation will not experience thinking as a solitary activity. The boundary between &#8220;my idea&#8221; and &#8220;the idea I developed in dialogue&#8221; will be as meaningless to them as the boundary between &#8220;news I read&#8221; and &#8220;news I watched on TV&#8221; is to someone born in 1965. They won&#8217;t agonise over it. They won&#8217;t even see it.</p><p>The second is that <strong>competence becomes access, not accumulation</strong>. The television generation learned that charisma could substitute for expertise &#8212; that a Reagan could outperform a policy wonk by understanding the medium. The AI-native generation will learn something more radical: that the ability to <em>direct</em> cognition matters more than the ability to <em>store</em> it. Memorisation will seem as quaint to them as handwriting a formal letter seems to us. Not because they are lazy, but because the skill has genuinely shifted &#8212; from knowing things to orchestrating the retrieval and synthesis of things. The librarian becomes the conductor.</p><p>The third is that <strong>the draft becomes the thought</strong>. Right now, we use AI to <em>produce</em> &#8212; to generate an output after we have already done the thinking. Write this email. Code this function. Summarise this paper. But AI-native thinkers will use it to <em>think</em> &#8212; the  conversation with the model will <em>be</em> the cognitive act, not a tool applied after the cognitive act is complete. Just as television-native politicians didn&#8217;t &#8220;use TV to communicate their ideas&#8221; but rather thought <em>in</em> television, AI-native thinkers will think <em>in</em> dialogue.</p><p>Each of these is a creolisation &#8212; a transformation of the medium&#8217;s grammar by the generation that absorbs it from birth. And each will produce native forms we can barely anticipate.</p><p>Which raises the question that has been hovering over this entire piece: <em>What is the MTV of AI?</em></p><p>What is the native form &#8212; the thing that makes no sense without the medium, that no pidgin speaker would have conceived, that will seem as natural to the creole generation as a music video seems to someone born in 1975? Maybe it already exists and we don&#8217;t recognise it, the way Ed Sullivan didn&#8217;t recognise that the Beatles&#8217; appearance on his show was the beginning of the end of his format. Maybe it&#8217;s emerging right now in some teenager&#8217;s bedroom, in a conversation with a model that looks, to adult eyes, like nothing more than messing around.</p><p>The older students in Managua sometimes described the younger children&#8217;s signing as sloppy. The 1950s critics called MTV &#8220;the death of music.&#8221; </p><p>The pidgin speakers always mistake the creole for a degradation of what they built, <strong>but it never is.</strong></p><div><hr></div><p>Let me take the first of those three dimensions &#8212; cognition as collaboration &#8212; and push it further, because I think it contains something genuinely new. Not a prediction about productivity or labour markets, but a claim about the <em>shape of thought itself</em>.</p><p>Start with music: before musical notation existed, there was music &#8212; obviously. People sang, drummed, improvised, passed melodies from voice to voice across generations. Music was woven into every human culture we know of. But it was ephemeral. It was bounded by what a performer or a small group could hold in living memory. A song could be complex, could be beautiful, could move people to tears &#8212; but it could not be a symphony. Not because nobody was talented enough to improvise one, but because the cognitive load of coordinating a hundred musicians across forty minutes of structured composition exceeds what any oral tradition can sustain. The symphony is not a very long song. It is a form that <em>requires notation to exist</em> &#8212; that is native to the medium of written music the way a music video is native to television.</p><p>And notation didn't stop at enabling the symphony. It made possible counterpoint, the fugue, the sonata &#8212; entire architectures of musical thought that couldn't be conceived, let alone performed, without a way to hold the structure outside any single mind. It gave musicians the ability to share compositions, riff on them, cover them, argue with them across centuries. Composers could respond to each other across generations. Jazz musicians could take a standard and rebuild it in real time because the standard existed as a shared written reference. Notation created a <em>collaborative substrate</em> for musical thought that transcended any individual performer or generation.</p><p>And each subsequent medium for music did the same thing. Recording gave us the pop single &#8212; a form shaped by the constraint of the three-minute 78rpm disc, which literally couldn&#8217;t hold more than that. Radio gave us the format, the playlist, the DJ as curator. MTV gave us the music video as art form &#8212; visual narrative fused with sound in ways that made no sense as radio. Each new medium didn&#8217;t distribute the same music more efficiently. It changed what music <em>was</em>. It called into existence forms that couldn&#8217;t have existed without it.</p><p>&#8220;Video Killed the Radio Star&#8221; isn&#8217;t just a clever title. It&#8217;s a precise description of creolisation.</p><p>Can we apply this to thought?</p><p>AI may do to cognition what notation did to music. Not just record or accelerate thinking, but enable structural forms of thought that couldn&#8217;t exist without the medium. We can&#8217;t yet name those forms, for the same reason a medieval troubadour couldn&#8217;t have described a symphony &#8212; the concept requires the medium that enables it. But the generation that grows up thinking-in-dialogue will develop them, just as the generation that grew up with notation developed counterpoint.</p><p>The current generation &#8212; my generation &#8212; agonises over attribution. &#8220;Did I write this, or did the AI?&#8221; This is a pidgin question. It maps the conventions of the old medium (solitary authorship, individual credit, the romantic myth of the writer alone at a desk) onto the new one, and then panics when the map doesn&#8217;t fit. The creole speakers won&#8217;t ask this question. Not because they&#8217;re careless about authorship, but because the category boundary won&#8217;t make sense to them. Thinking-with-a-model will be as unremarkable as thinking-with-notation. We don&#8217;t ask whether Bach &#8220;really&#8221; composed the fugues or whether the technology of notation composed them for him. The question is ill-formed. The fugue <em>is</em> a form of thought-through-notation. It has no existence outside the medium.</p><p>I should be honest here, because the argument demands it.</p><p>This post was written in conversation with an AI. Not generated by one &#8212; the ideas, the structure, the editorial judgment about what to cut and what to keep, the voice, the choice to open with Managua rather than McLuhan &#8212; those are mine. But the drafting happened in dialogue. I talked through the argument; the model offered framings; I pushed back, redirected, took what worked and rewrote what didn&#8217;t. Someone who read an earlier draft told me I &#8220;write beautifully, better than the AI-generated slop so many people put out.&#8221; I told them it&#8217;s not <em>not</em> written with AI.</p><p>I think this is fine. More than fine &#8212; I think this is the point.</p><p>Nobody cares anymore if you google something. Nobody says &#8220;yes, but you didn&#8217;t find that in a library.&#8221; We use the tools available to us. Someone can take a chisel and make pebbles out of stone, or Michelangelo&#8217;s <em>David</em>. The chisel didn&#8217;t have a vision of <em>David</em>. Michelangelo didn&#8217;t have hands hard enough to shape marble. The work exists in the collaboration between intent and instrument, and the only meaningful question is whether the result is any good.</p><p>But notice my instinct to explain this. Notice the felt need to disclose, to justify, to preemptively defend the process. That instinct <em>is</em> the pidgin. A native speaker &#8212; a true creole speaker of AI &#8212; would no more think to disclose that they &#8220;used AI&#8221; in their thinking than you would disclose that you &#8220;used Google&#8221; in your research or &#8220;used notation&#8221; in your composition. The disclosure itself marks me as someone who remembers the old rules, who still feels their gravitational pull even as I move in the new medium. I am fluent. I am not native.</p><p>This distinction &#8212; between fluency and nativity &#8212; is not about skill. It&#8217;s about the <em>topology of thought</em>. Solo cognition has a particular shape: it&#8217;s sequential, it gets stuck in loops, it has blind spots that persist because there is no external surface to reflect off of. When you think alone, you tend to converge too quickly on your first good idea, because there is no counterpressure. Your assumptions remain invisible to you precisely because they are <em>yours</em> &#8212; they are the water you swim in.</p><p>Collaborative cognition with an AI has a different shape. It&#8217;s more exploratory, more branching, more willing to hold contradictions in parallel before resolving them. It generates more options and discards them faster. It catches assumptions earlier &#8212; not because the model is smarter, but because it is <em>other</em>, because it constitutes an external surface off which your thinking can reflect. The cognitive process has a different geometry.</p><p>If you want a physics metaphor &#8212; and at Soliton, we always strive for a physics metaphor &#8212; think of coupled oscillators. Two pendulums hanging from a shared beam don&#8217;t just swing faster than one pendulum. They exchange energy in ways that neither would exhibit alone. The system develops <em>normal modes</em> &#8212; patterns of oscillation that are properties of the coupled system and have no analogue in the isolated case. One normal mode has both pendulums swinging together; another has them swinging in opposition. These modes are not improvements on solo swinging. They are structurally different behaviours that only exist when the coupling is present.</p><p>AI-native cognition may exhibit normal modes that solo cognition simply doesn&#8217;t have. Not faster thinking. Not better thinking. <em>Differently-shaped</em> thinking &#8212; patterns of exploration, synthesis, and revision that only emerge when a mind is coupled to a model, just as counterpoint only emerges when a composer is coupled to notation.</p><p>We can approximate these modes. We can learn to use them, as I am doing in this very paragraph. But a generation that grows up never knowing uncoupled cognition will inhabit them the way we inhabit language &#8212; without effort, without self-consciousness, without the faint sensation of translating from one grammar into another that marks everything a pidgin speaker does.</p><div><hr></div><h2>What the Teachers Couldn&#8217;t Learn</h2><p>There is a particular kind of loss in the Managua story that doesn&#8217;t get enough attention.</p><p>The older deaf students &#8212; the ones who invented LSN, who bootstrapped communication from nothing in the schoolyards of San Judas &#8212; were not left behind because they lacked intelligence or effort. They were the pioneers. They built the pidgin. Without them, there would have been no raw material for the younger children to transform. And yet the creole that emerged &#8212; the one with verb agreement and spatial grammar and recursive structure &#8212; remained just out of their reach. They could participate. They could communicate. But the deepest structures of the new language would always feel like a second skin rather than a first.</p><p>This is not a story about obsolescence. It is a story about the particular cruelty of generational thresholds &#8212; about the fact that the people who build the bridge are not always the ones who can cross it.</p><p>We are, right now, the bridge-builders. Those of us who discovered AI as adults, who marvelled at it, who taught ourselves to prompt and to build and to integrate it into our working lives &#8212; we are the older students in the schoolyard. We are the 1950s television producers who recognised that something extraordinary was happening and did our best to adapt. We are fluent enough. We will never be native.</p><p>The children who are growing up now &#8212; who will learn to think in dialogue with AI the way we learned to think in dialogue with text &#8212; will not simply use these tools more skillfully than we do. They will develop cognitive habits we cannot fully anticipate and may not fully understand. Their relationship to knowledge, to authorship, to the very boundary between &#8220;my thought&#8221; and &#8220;the thought I developed in conversation&#8221; will be as foreign to us as spatial verb agreement was to the first generation of Managua signers.</p><p>And like those first signers, we may find the new forms slightly messy. The older LSN speakers sometimes described the younger children&#8217;s signing as sloppy or chaotic &#8212; not recognising that what looked like mess was actually the emergence of grammatical structure more sophisticated than anything they had built. We will likely have the same reaction to AI-native cognition. When a twenty-year-old produces something brilliant in ways that seem to dissolve the boundary between human and machine thinking, our instinct will be to ask whether it&#8217;s &#8220;really theirs.&#8221; That question will mark us as pidgin speakers more clearly than anything else.</p><p>McLuhan told us the medium is the message. The message of television was that appearance is substance, that narrative is argument, that the emotional register is the real content. It took a generation to absorb that message so deeply it became invisible.</p><p>The message of AI is still crystallising. But I think it is something like this: <em>cognition was never meant to be solitary.</em> The model of the lone thinker &#8212; the genius in the garret, the scholar in the library, the mind working in magnificent isolation &#8212; was never a fundamental truth about human thought. It was an artefact of the available media. Before notation, music was bounded by what a single memory could hold; after it, we got the symphony. Before writing, argument was bounded by what a single voice could sustain; after it, we got the treatise. Each medium didn&#8217;t just record what was already there &#8212; it called into existence forms that couldn&#8217;t have existed without it. AI is the next such medium. And the generation that grows up with collaborative cognition as the default will develop forms of thought as far beyond our solitary reasoning as the symphony is beyond the campfire song &#8212; not better, not worse, but structurally unreachable from where we stand.</p><p>The children in Managua didn&#8217;t invent a <em>better</em> version of what their teachers had. They invented something structurally different &#8212; a language with properties that emerged from the specific constraints and affordances of their situation. The AI-native generation will do the same. Not better thinking. <em>Different</em> thinking.</p><p>We can watch it happen. We can even participate. But we should be honest about what we are: the last generation of pidgin speakers, marvelling at a creole we helped make possible and can never quite speak.</p><div><hr></div><p><em>Soliton explores the collisions between mathematics, AI, and finance. If this landed, there's more where it came from.</em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/subscribe?"><span>Subscribe now</span></a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/the-creole-generation?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/the-creole-generation?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/the-creole-generation/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/the-creole-generation/comments"><span>Leave a comment</span></a></p><p></p>]]></content:encoded></item><item><title><![CDATA[The Soliton of AI Disruption]]></title><description><![CDATA[Disrupt a pond by dropping a stone in the water and the ripple spreads, flattens, dissipates.]]></description><link>https://solitonmaths.substack.com/p/the-soliton-of-ai-disruption</link><guid isPermaLink="false">https://solitonmaths.substack.com/p/the-soliton-of-ai-disruption</guid><dc:creator><![CDATA[soliton-maths]]></dc:creator><pubDate>Wed, 04 Mar 2026 09:54:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!p-VL!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f9736aa-0d64-4640-8615-3c29089dfaec_96x96.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Disrupt a pond by dropping a stone in the water and the ripple spreads, flattens, dissipates. The energy disperses until there&#8217;s nothing left.</p><p>A soliton is different. It&#8217;s a wave that holds itself together &#8212; nonlinearity and dispersion balancing each other so precisely that the pulse travels indefinitely without losing its form. It was discovered by accident in 1834, when a Scottish engineer named John Scott Russell noticed the shape of water in a canal didn&#8217;t behave the way waves were supposed to: it persisted, it didn&#8217;t spread. He chased it on horseback for two miles before it finally faded into the shallows.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Soliton! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>I&#8217;ve been thinking about solitons lately in the context of technological change &#8212; specifically about which waves hold their shape as they move through society, and which ones disperse harmlessly into the noise of history, and about what happens when a new wave arrives into a medium that&#8217;s already moving.</p><p>In January 2025, ChatGPT reached 300 million weekly active users. By the end of that year, estimates put the number of people actively using generative AI tools at over a billion &#8212; roughly one in eight people alive. The technology press called it unprecedented, the fastest adoption curve in history. <em>But was it?</em></p><p>ChatGPT launched in November 2022 and reached 100 million users in two months. For comparison, the internet took seven years to reach the same milestone, and the telephone 75. On those numbers, the &#8220;unprecedented&#8221; label seems justified.</p><p>Except it isn&#8217;t quite the right comparison. The internet had to build its own roads: lay cables, manufacture modems, train engineers, negotiate with telcos, wait for browsers to exist, wait for people to learn what a browser was. Every new user required physical infrastructure that didn&#8217;t previously exist. The adoption curve was slow because the distribution problem was genuinely hard.</p><p>ChatGPT inherited a world where 5 billion people already carried a networked computer in their pocket, already had payment details stored with Apple or Google, already knew how to download an app in thirty seconds. The distribution problem was already solved &#8212; solved by thirty years of compounding investment in exactly the infrastructure AI needed. What looked like AI&#8217;s adoption speed was really the internet&#8217;s adoption speed, borrowed wholesale.</p><p>Think about how ideas went viral before the internet. Someone photocopied a funny memo and left it in the break room. If it was good enough, a colleague might post it on a noticeboard in another building. The limiting factor wasn&#8217;t the quality of the idea &#8212; it was the friction in the distribution layer. Then came email, and the same memo could reach a thousand people overnight. Then social media, and a million by morning. Then mobile, and the latency between &#8220;this exists&#8221; and &#8220;everyone has seen this&#8221; collapsed toward zero.</p><p>Each acceleration had nothing to do with ideas getting better: <em>it was pure infrastructure</em>. The carrier wave got faster. AI didn&#8217;t create a new carrier wave, it rode the one we&#8217;d already built. Which means the breathless adoption statistics, while real, tell us more about the maturity of the internet than they do about AI itself. We should be less surprised by how fast AI spread, and more interested in what it actually costs now that it&#8217;s everywhere.</p><p>Because that&#8217;s a different question entirely.</p><p>Every transformative technology in history has done one thing: it has made something dramatically cheaper. Not incrementally cheaper but orders of magnitude cheaper. And the thing it makes cheaper tells you almost everything about why it was transformative.</p><p>The factory made objects cheap. Before industrialisation, a shirt was a significant economic object &#8212; the product of hours of skilled labour, priced accordingly. The spinning jenny and the power loom didn&#8217;t just speed up production, they collapsed the cost structure of physical goods so completely that the entire relationship between labour, capital, and consumption had to be renegotiated. Economies that had been organised around the scarcity of manufactured objects had to find a new organising principle.</p><p>The printing press made codified knowledge cheap. Before Gutenberg, a bible cost roughly the same as a house. Scribes were expensive, copying was slow, and the scarcity of written material meant that knowledge was necessarily concentrated &#8212; in monasteries, in universities, in the hands of people wealthy enough to own books. The press didn&#8217;t just speed up copying. It demolished the economics of knowledge scarcity so thoroughly that the institutions built on that scarcity &#8212; the Church&#8217;s interpretive monopoly, the university&#8217;s gatekeeping function &#8212; couldn&#8217;t survive intact. It took eighty years for the Reformation to arrive, but the press made it inevitable.</p><p>The internet made distribution cheap. My newsletter exists because sending words to ten thousand people costs me the same as sending them to ten. The economics of the Cleveland newspaper &#8212; the printing press, the delivery trucks, the physical infrastructure of geographic distribution &#8212; were not disrupted by a better newspaper. They were disrupted by the elimination of distribution cost as a meaningful variable. When distribution is free, every business built on controlling distribution has to ask itself what it actually sells.</p><p>Which brings us to AI. If the factory made objects cheap, the press made knowledge cheap, and the internet made distribution cheap &#8212; what does AI make cheap?</p><p><em>Reasoning. Composition. The production of structured thought.</em></p><p>This is the step that sat between all the others. Objects needed to be designed before they could be manufactured. Knowledge needed to be synthesised before it could be printed. Content needed to be created before it could be distributed. In each case, the cognitive work upstream of the technology was left untouched &#8212; expensive, slow, human. AI is the first technology that reaches into that upstream step and starts compressing it.</p><p>This is why the intuition that AI is somehow more fundamental than previous technologies is hard to shake, even when the adoption statistics are still modest. It&#8217;s not that AI is more powerful in some abstract sense. It&#8217;s that it sits higher up the causal chain. It&#8217;s making cheap the thing that all the previous technologies assumed would remain expensive.</p><p>Whether that makes it more disruptive or simply differently disruptive is a question the cost ladder alone can&#8217;t answer. For that we need to look at what happens in the gap between when a technology arrives and when society figures out what to do with it.</p><p>The steam engine was invented in 1698. The standard of living for the average British worker didn&#8217;t meaningfully improve until the 1840s. That&#8217;s a hundred and forty years of transformative technology failing to show up in the lives of the people living through it.</p><p>This is not an anomaly. It&#8217;s the pattern. Electrification of American cities was largely complete by the 1900s. Yet the productivity statistics didn&#8217;t reflect it until the 1920s, and the full reorganisation of factory layouts, supply chains, and working practices took until the 1940s. The economic historian Robert Gordon calls this the &#8220;productivity paradox&#8221; &#8212; the lag between a technology&#8217;s arrival and its measurable impact on how much value an economy actually produces. Paul David, studying electrification specifically, concluded that the lag existed because firms had to do more than adopt the technology. They had to forget how they&#8217;d been organised before it.</p><p>That last point is underappreciated. The factory that installed an electric motor in 1900 and used it to drive the same central drive shaft that the steam engine had powered wasn&#8217;t really electrified &#8212; it was just running the old system on new energy. The gains only came when someone realised that electricity could be distributed to individual machines, that the factory floor could be reorganised entirely, that the whole logic of how work was arranged could be reconceived from scratch. The technology arrived in a decade. The reconception took two generations.</p><p>The printing press is the starker example. Gutenberg printed his bible in 1455. The Reformation began in 1517 &#8212; sixty years later. But the deeper restructuring of European knowledge institutions, the universities, the Church, the legal frameworks around intellectual property, the emergence of the scientific method as a formalised practice &#8212; that took two centuries. The press didn&#8217;t cause the Reformation directly. It made the Reformation possible by destroying the economics of interpretive monopoly. The lag between possibility and actuality was a human lag, not a technological one.</p><p>The internet is close enough to feel personal. The web became publicly available in 1991. Most of us didn&#8217;t feel its full economic impact until the mid-2000s, and the deepest restructuring &#8212; of retail, media, finance, politics &#8212; is arguably still incomplete. We are thirty years in and still mid-adaptation. The generation that grew up with smartphones has different cognitive habits, different social structures, different economic expectations than the generation that watched the web arrive as adults. The adaptation is still being written.</p><p>What the historical record shows, consistently, is that the lag between technological arrival and societal adaptation is measured in decades, occasionally in generations, and is almost always longer than the optimists predicted and shorter than the pessimists feared. The technology arrives at one speed. The humans arrive at another.</p><p>And now consider the position we&#8217;re actually in. We are thirty years into adapting to the internet &#8212; still mid-reorganisation, still working out what it means for institutions, democracy, attention, knowledge. And into that already-turbulent medium, a new wave has arrived. Not into still water. Into a medium that is already moving.</p><p>This is where the soliton becomes more than a metaphor.</p><p>Every previous technology displaced something physical. The factory displaced manual assembly. The press displaced the scribe. The internet displaced the delivery truck. In each case, the thing that got displaced was downstream of human cognition &#8212; and humans adapted by doing what humans do: they thought their way around the disruption. They retrained, reorganised, innovated into the gaps the technology created. Cognition was the constant. It was the adaptation tool that every previous wave left intact.</p><p>AI is the first technology that reaches for the tool itself.</p><p>This is the inversion that&#8217;s difficult to sit with honestly. It&#8217;s not that AI will necessarily eliminate knowledge work &#8212; the historical pattern suggests it&#8217;s more likely to restructure it, eliminate some roles, create others we can&#8217;t yet name. It&#8217;s that the mechanism we&#8217;ve always used to navigate technological transitions is the same mechanism being compressed. The workforce that needs to adapt to AI is being asked to adapt using thinking, judgement, and composition &#8212; precisely the capacities whose cost AI is collapsing.</p><p>The counterarguments exist and deserve acknowledgment. Creativity, embodied expertise, relational intelligence, ethical judgement &#8212; these are real capacities that don&#8217;t obviously compress. The surgeon, the therapist, the teacher who reads a room &#8212; there are refuges. But the honest version of this observation is that we don&#8217;t actually know where the floor is yet. Every previous transition had a visible answer to &#8220;where do the displaced workers go.&#8221; The hand-loom weavers became factory operatives. The factory operatives became service workers. The service workers became knowledge workers. The direction of travel was always visible in advance, even if the path was brutal.</p><p>This time the direction of travel is genuinely unclear. Not because AI is magic, but because it&#8217;s the first technology whose ceiling we can&#8217;t see from where we&#8217;re standing. That uncertainty is not pessimism. It&#8217;s just honesty about the limits of extrapolation.</p><p>Which brings us back to Russell, on the towpath, in 1834.</p><p>He was watching something that shouldn&#8217;t have been possible. Waves were supposed to disperse &#8212; that was the physics, that was the mathematics, that was everything the theory predicted. And yet here was this heap of water, holding its shape, moving through a medium that should have absorbed it, persisting when it was supposed to fade.</p><p>What Russell had found was that under the right conditions &#8212; when nonlinearity and dispersion balance each other precisely &#8212; a wave can propagate without loss. The medium doesn&#8217;t win, the wave doesn&#8217;t flatten, the energy finds a form it can maintain indefinitely.</p><p>The question worth sitting with isn&#8217;t whether AI is a soliton. It&#8217;s whether we are.</p><p>Every previous technological wave passed through a society that was, relative to the speed of the wave, approximately still water. The disruption was real, the adaptation was painful, but the medium had time &#8212; decades, generations &#8212; to reorganise itself around the new energy. The institutions deformed and reformed. The labour markets restructured. The knowledge frameworks updated. The medium held its shape, eventually, because the wave moved at a speed the medium could process.</p><p>What&#8217;s different now isn&#8217;t just the speed of AI. It&#8217;s that the medium is already moving. We are mid-adaptation to the internet, mid-reorganisation of our institutions, mid-negotiation of what mobile and social media have done to attention and democracy and knowledge. The water is already turbulent. And into that turbulence, a new wave has arrived &#8212; one that rides the existing current rather than fighting it, one that borrowed the carrier wave we already built.</p><p>When two solitons collide in the same medium, the mathematics is surprisingly clean: they pass through each other and emerge unchanged, each maintaining its form. But that&#8217;s solitons in an idealised medium. In a real one, with friction and noise and the accumulated perturbations of history, the interaction is less predictable.</p><p>We don&#8217;t yet know if society is a medium that can maintain its shape through this. The historical record suggests it can &#8212; that the adaptation lag, however painful, eventually closes. But the historical record was compiled in still water.</p><p>Russell chased his wave for two miles before it faded. We are, at most, a few hundred metres in.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/the-soliton-of-ai-disruption/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/the-soliton-of-ai-disruption/comments"><span>Leave a comment</span></a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/the-soliton-of-ai-disruption?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/the-soliton-of-ai-disruption?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p>]]></content:encoded></item><item><title><![CDATA[From QBasic to Claude: A Skeptic’s Four-Month Conversion]]></title><description><![CDATA[How a snake game for my seven-year-old turned me into an AI true believer.]]></description><link>https://solitonmaths.substack.com/p/from-qbasic-to-claude-a-skeptics</link><guid isPermaLink="false">https://solitonmaths.substack.com/p/from-qbasic-to-claude-a-skeptics</guid><dc:creator><![CDATA[soliton-maths]]></dc:creator><pubDate>Thu, 19 Feb 2026 09:01:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!p-VL!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f9736aa-0d64-4640-8615-3c29089dfaec_96x96.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In October 2025 I decided to code a game for my son. He&#8217;s seven, and I wanted to show him the snake game I remembered from QBasic in the early nineties &#8212; the same one that later became famous on Nokia phones. A green snake eating dots on a black screen, very basic, the simplest possible thing.</p><p>I figured it would take me a couple of days. Maybe a few hours if I was disciplined about scope. I opened VS Code, started a new Python file, and noticed the Gemini code assistant sitting in the sidebar. I&#8217;d never used it. I thought: what the hell, maybe it&#8217;ll help me pick a framework.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Soliton! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>I gave it a prompt &#8212; one prompt &#8212; 60s later the game was complete and fully functional. My son was playing it before I&#8217;d finished my coffee.</p><p>My jaw dropped. Not because the code was beautiful &#8212; it was fine, probably better than what I&#8217;d have written &#8212; but because the gap between my mental model of what these tools could do and what they&#8217;d just done was so vast it restructured something in my head. I&#8217;d been casually dismissive of LLMs. Autocomplete on steroids, I&#8217;d told colleagues. Useful for boilerplate, maybe. Not for real work.</p><p>I was wrong. And the uncomfortable thing about being wrong in October 2025 was that the pace of change meant I was becoming <em>more</em> wrong with every passing week.</p><h2>The exploration phase</h2><p>I didn&#8217;t immediately start building things. I&#8217;m a physicist by training, and my instinct when confronted with a phenomenon I don&#8217;t understand is to characterise it before I try to use it. I spent the next several weeks doing what I&#8217;d describe as systematic tinkering: learning the mathematics and mechanics of the models, prompting patterns, testing boundaries, trying to develop an intuition for where these systems are brilliant versus where they hallucinate.</p><p>This period was humbling. I&#8217;d spent twenty years writing quantitative software, so I thought I knew how to build things. What I didn&#8217;t know was how to <em>collaborate</em> with something that could build things too &#8212; something that was occasionally smarter than me at tasks I&#8217;d considered core to my professional identity, but simultaneously can be wrong in ways that a junior quant would catch.</p><p>The mental model I eventually settled on &#8212; and still use &#8212; comes from physics. An LLM is not a colleague and not a tool. It&#8217;s a <em>field</em>. You don&#8217;t operate i, but move through it, and it responds to your trajectory. The quality of what you get out depends entirely on the quality of the path you take through the space. Prompting is not instruction &#8212; <em>it&#8217;s navigation</em>.</p><h2>The second jaw drop</h2><p>In October 2025, my company started adopting Anthropic&#8217;s Claude Code. I got access at work shortly after, and started using it on internal development tasks. The first few days were productive but unsurprising &#8212; code generation, refactoring, the kind of acceleration you&#8217;d expect.</p><p>Then came the moment that changed my understanding of what these systems actually are.</p><p>I was working with a large internal database system &#8212; tens of thousands of functions, no central documentation, no master list. There was a typeahead search interface: you could start typing a function name and it would autocomplete. But there was no way to get a comprehensive catalogue. I&#8217;d been manually exploring for days, building a mental map of what was available.</p><p>I mentioned this to Claude Code, mostly thinking out loud. I expected it to commiserate, or maybe suggest I ask the database team for documentation.</p><p>Instead, it proposed a strategy: write a script to systematically query the typeahead itself. Use the autocomplete as an oracle. Enumerate the entire function space by walking every possible prefix.</p><p>I let it run. It collected over 120,000 functions.</p><p>This was not autocomplete on steroids. This was not a faster way to type code I already knew how to write. This was an agent that identified a problem, reasoned about the structure of the system it was operating in, proposed an approach I hadn&#8217;t considered, and executed it autonomously. The typeahead was a keyhole; Claude turned it into a door.</p><p>That was when the mental model shifted from &#8220;useful tool&#8221; to &#8220;this changes everything.&#8221;</p><h2>Building at work</h2><p>Over the next two months I threw myself into a series of projects that tested the boundaries of what AI-assisted development could do in a real enterprise environment. I&#8217;m going to describe these generically &#8212; I work in financial data, and the specifics are proprietary &#8212; but the patterns are universal.</p><p><strong>Graph-based portfolio analytics.</strong> Traditional portfolio systems store holdings as flat lists. I built a prototype that represents portfolios as graphs &#8212; securities as nodes, relationships (issuer hierarchy, sector membership, counterparty exposure) as edges. The graph structure makes certain queries trivial that are nearly impossible in tabular models, especially as the ground truth for agentic introspection. AI accelerated the development by an order of magnitude: what would have been a quarter of work was running in days.</p><p><strong>Hybrid RAG for internal support.</strong> Every large software company has a graveyard of internal support tickets &#8212; thousands of resolved issues containing institutional knowledge that&#8217;s effectively inaccessible. I built a retrieval-augmented generation system that indexes this history and surfaces relevant solutions over subsets of these tickets. The hybrid part matters: pure vector search misses context, graphs shine a light on bottlenecks and dependencies.  Combining them with an LLM that can reason about relevance produces something genuinely useful.</p><p><strong>Vision models for document analysis.</strong> Financial data arrives in PDFs &#8212; prospectuses, term sheets, regulatory filings &#8212; and extracting structured data from them has historically required brittle rule-based parsers or expensive manual review. Vision models changed this calculus entirely. Point a multimodal model at a PDF page and ask it to extract the charts, the data, the timeseries. It just works. Not perfectly, not without validation. But well enough to transform a bottleneck into a pipeline.</p><p>Each of these projects taught me something different. The graph work taught me that AI is most powerful when you rethink the architecture, not just accelerate the existing one. The RAG system taught me that the engineering around the model matters more than the model itself. The vision work taught me that multimodal capabilities are underrated &#8212; everyone fixates on text generation while the real unlock is perception.</p><p>But the deeper lesson, the one that cut across all three, was about velocity. I was a principal quantitative researcher with deep domain expertise, and I was shipping prototypes at a pace that would have been physically impossible six months earlier. Not because I was working harder. Because the nature of the work had changed.</p><h2>Building at home</h2><p>The work projects convinced me that AI would reshape how software is built. The home projects convinced me of something larger.</p><p>In late November, I started building Clio &#8212; a personal AI companion. Not a chatbot. Not an assistant. A <em>companion</em>: something with memory, personality, and context about my life. Named after the muse of history, because what makes a companion different from a tool is that it remembers &#8212; or as Clio put it &#8220;assistants optimize your calendar and answer questions.  I remember who you are and notice when you&#8217;re struggling.  There&#8217;s a difference&#8221; (this probably deserves and entire post in itself).  </p><p>Clio evolved through several versions. The early iterations were clunky &#8212; a chat interface (<a href="https://mattermost.com/">Mattermost</a>, since it&#8217;s locally hosted), some basic memory retrieval, habit tracking for my meditation and journaling practices. But as I built, something unexpected happened. The act of designing an AI&#8217;s personality forced me to think carefully about what personality <em>is</em>. The act of building a memory system forced me to think about what memory <em>does</em> &#8212; not just retrieval, but the way context shapes every subsequent interaction. You don&#8217;t realise how much of human cognition is contextual until you try to simulate it.</p><p>Then came Astraea &#8212; named after the goddess of justice and precision. Where Clio is a companion, Astraea is a research assistant: she surfaces relevant papers, articles, and connections for each piece I write. She&#8217;s the reason this newsletter cites specific sources rather than gesturing vaguely at &#8220;recent research.&#8221; Building Astraea taught me a different lesson: that the value of AI isn&#8217;t in generating answers, it&#8217;s in <em>expanding the space of questions you think to ask.</em></p><h2>Where I am now</h2><p>Four months ago I was a skeptic. Today I have full conviction that this is the most significant shift in the software industry since the internet,  <em>possibly larger</em>.</p><p>I don&#8217;t say this lightly. I&#8217;ve lived through the dot-com bubble, the cloud transition, the mobile revolution, the big data era. Each of those changed how software was built or distributed. AI changes what software <em>is</em>. The unit of creation is no longer a function or a class or a microservice, but an <em>intent</em>. Describe what you want, negotiate with the model about how to get there, validate the output. The entire abstraction layer between human thought and running code is being compressed toward zero.  Internally we like to say &#8220;<strong>developing at the speed of thought</strong>&#8221;. </p><p>For people like me &#8212; physicists and mathematicians who wandered into software because it was the medium available to express quantitative ideas &#8212; this is not a threat but a liberation. I spent twenty years struggling with syntax, frameworks, build systems and deployment pipelines to get to the part I actually cared about: the mathematics. Now I can spend most of my time on the mathematics.</p><p>But I want to be precise about what I&#8217;m <em>not</em> saying. I&#8217;m not saying AI replaces developers. I&#8217;m saying it changes the rate-limiting step. The bottleneck is no longer &#8220;can you write the code?&#8221; It&#8217;s &#8220;do you know what to build, and can you evaluate whether it&#8217;s correct?&#8221; Domain expertise becomes <strong>more valuable, not less</strong>. Judgment becomes <strong>more valuable, not less</strong>. The people who thrive will be the ones who can navigate the AI &#8220;field&#8221; &#8212; who understand enough about the problem space to ask good questions and enough about the model&#8217;s capabilities to recognise good answers.</p><p>The snake game took sixty seconds. But knowing it was the right game to build for a seven-year-old who&#8217;d never seen a command line? That took twenty years of being a programmer, seven years of being a father, and a person who remembers QBasic.</p><p>The tools are new, the judgment is not.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/from-qbasic-to-claude-a-skeptics?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/from-qbasic-to-claude-a-skeptics?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><p><em>Tom Davis is a physicist turned quant who builds fixed income analytics for a living and AI systems at home. He writes about the intersection of quantitative methods, AI, and finance.</em></p><p><em>Views expressed here are entirely my own and do not represent those of my employer. Nothing in this newsletter constitutes investment advice.</em></p>]]></content:encoded></item><item><title><![CDATA[The Moat is Breached — To the Keep]]></title><description><![CDATA[When AI breaches the outer walls of software, what's left standing?]]></description><link>https://solitonmaths.substack.com/p/software-moats-ai-disruption</link><guid isPermaLink="false">https://solitonmaths.substack.com/p/software-moats-ai-disruption</guid><dc:creator><![CDATA[soliton-maths]]></dc:creator><pubDate>Sun, 15 Feb 2026 20:03:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!p-VL!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f9736aa-0d64-4640-8615-3c29089dfaec_96x96.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>On February 4th, Anthropic released an AI-powered legal assistant. Thomson Reuters dropped 15%.  Docusign 10%.  By the end of the week, the iShares Expanded Tech-Software Sector ETF had entered a technical bear market, down over 20% from its peak. Atlassian lost a third of its value in five trading sessions. Monday.com is down by half. The financial press has taken to calling it <a href="https://fortune.com/2026/02/12/saaspocalypse-winners-losers-ceos-data-volume-pricing/">the SaaSpocalypse</a>.</p><p>I confess to finding this personally fascinating, in the way that a physicist finds a phase transition fascinating even when the phase transition is happening to the material they&#8217;re standing on.</p><p>Goldman Sachs drew a comparison to newspapers. Their strategist Ben Snider warned that software stocks may be experiencing not the beginning of the end, but the end of the beginning &#8212; and pointed to the newspaper industry, where share prices declined an average of 95% between 2002 and 2009 as the internet gutted their business model. The implication is clear: software may be in for the same structural decline.</p><p>I think the analogy is instructive but imprecise. And I think the imprecision matters a great deal.</p><h2>What actually happened to newspapers</h2><p>Newspapers had two moats. First, a local monopoly on attention: if you lived in Cleveland, the <em>Plain Dealer</em> was how you found out what happened yesterday. Second, a monopoly on classified advertising: if you wanted to sell a used car or hire a plumber, you paid $50 for a few lines of text. The internet destroyed both moats simultaneously. Craigslist gutted classifieds. Google and Facebook gutted local attention. The content itself &#8212; journalism &#8212; was never the moat. It was always the distribution.</p><p>Now consider a major financial data company &#8212; S&amp;P Global, Bloomberg or any of their competitors. Yes, they have software. Yes, an LLM can build you a perfectly serviceable financial dashboard in an afternoon. But the dashboard was never the moat. The moat is what flows through it.</p><p>This distinction &#8212; between the wall and the keep, to use a medieval siege metaphor &#8212; is what I think the market is currently getting wrong. The market is pricing software as though the wall <em>is</em> the keep. For some companies, that&#8217;s probably right. For others, it misses the point entirely.</p><h2>Four keeps</h2><p>I&#8217;ve spent several months thinking about what structural defences survive when code itself becomes commodity. I keep arriving at four. None of them is &#8220;we write good code.&#8221;</p><p><strong>1. Proprietary data</strong></p><p>This is the most discussed, and rightly so. When every analyst on the street relies on evaluated bond prices for their portfolio analytics, they&#8217;re not relying on the code that displays them. They&#8217;re relying on decades of curated corporate actions, entity mappings, and pricing methodologies that cannot be vibe-coded into existence.</p><p>But &#8220;proprietary data&#8221; is a spectrum, not a binary. Public filings are trivially replicable. Evaluated prices from an approved source with a documented methodology and an auditable lineage? That&#8217;s a very different proposition. The companies that will thrive are those whose data moat runs deepest &#8212; where the data itself, not the interface, is the product.</p><p>There&#8217;s a nice parallel to information theory here. In my <a href="https://www.mdpi.com/1911-8074/16/12/501">paper on deriving Black-Scholes from maximum entropy</a>, the key insight was that the unique probability distribution is the one that is &#8220;maximally noncommittal to missing information&#8221; &#8212; it contains exactly the information you&#8217;ve stated and nothing more. Proprietary data works in the opposite direction: the moat is precisely the <em>additional</em> information that competitors cannot replicate. The harder it is to reconstruct from publicly available signals, the deeper the moat.</p><p><strong>2. Legal and regulatory frameworks</strong></p><p>This is the moat nobody talks about, but it may be the most durable. When a bank uses evaluated bond prices in regulatory capital calculations, that&#8217;s not a workflow preference &#8212; it&#8217;s a compliance requirement. The data has to come from an approved source with an auditable lineage. You cannot replace it with &#8220;Claude said the bond is worth 98.5.&#8221;</p><p>The same logic applies across financial services: trade reporting, portfolio compliance, risk calculations submitted to regulators. These are not features. They are legal obligations tied to specific vendors through years of validation and regulatory acceptance.</p><p>I think of this as a kind of institutional impedance. In physics, impedance matching determines how efficiently a signal transfers between media. Regulatory frameworks create an impedance mismatch between AI-generated outputs and the systems that consume them. AI can absolutely help <em>process</em> this data more efficiently. But the legal and compliance infrastructure that certifies who produces it, who audits it, and who is liable for it &#8212; that is structural, and no amount of clever prompting changes it.</p><p><strong>3. Network effects</strong></p><p>Ben Thompson made the Spotify argument well: AI is a sustaining technology for companies that already own network effects, not a disruptive one. If your moat is the liquidity of buyers and sellers, or the density of a content graph, AI strengthens your position rather than undermining it.</p><p>In financial data, network effects manifest in subtler ways. When every portfolio manager at a firm uses the same analytics platform, they develop shared workflows, shared reporting templates, shared vocabularies for discussing risk. Their counterparts at other firms use the same platform, creating an industry-wide communication protocol. Bloomberg&#8217;s terminal isn&#8217;t sticky because the code is irreplaceable &#8212; it&#8217;s sticky because the entire industry&#8217;s communication layer runs through it.</p><p>This is where the &#8220;feature, not a product&#8221; companies are most exposed. If your software exists in isolation &#8212; a standalone tool that doesn&#8217;t connect to a broader network of users and data &#8212; AI can replicate it. If it&#8217;s a node in a network, the replication cost is orders of magnitude higher.</p><p><strong>4. Cognitive offloading</strong></p><p>This is the subtlest defence, and the one I find most interesting. Enterprise software doesn&#8217;t just do things &#8212; it <em>thinks for people.</em> Not in the AI sense, but in the institutional sense. It encodes thousands of decisions about methodology, edge cases, missing data handling, and regulatory constraints into deterministic processes.</p><p>When a portfolio manager uses a risk system, they aren&#8217;t just running a Monte Carlo simulation. They&#8217;re relying on a system that remembers which bonds to include in a duration calculation, how to handle defaulted securities, what the correct day count conventions are for each currency, and how to reconcile conflicting data sources. The system carries the cognitive burden of institutional knowledge so the human doesn&#8217;t have to.</p><p>AI is extraordinary at generating new answers. But enterprises don&#8217;t just need new answers &#8212; they need <em>the same answer they got yesterday</em>, computed consistently, with an audit trail, under the same methodology. Call it &#8220;<em>deterministic cognition</em>&#8221;. This is the gap that the current AI hype cycle consistently underestimates.</p><p>Yes, Claude can analyse a set of regulatory filings brilliantly. Can it do it the <em>same way</em> every day, across every portfolio, with outputs that match last quarter&#8217;s methodology to twelve decimal places? That&#8217;s a much harder problem. And it&#8217;s the problem that enterprise software actually solves.</p><h2>The keep, not the wall</h2><p>The newspaper analogy fails because newspapers&#8217; moats &#8212; local attention and classified revenue &#8212; were fundamentally about distribution. The internet offered better distribution. Game over.</p><p>The best software companies have moats that aren&#8217;t about the software at all. They&#8217;re about data, legal frameworks, network position, and cognitive infrastructure. The code is the castle wall. These four things are the keep.</p><p>The SaaSpocalypse is real. The walls are being breached. For pure &#8220;CRUD apps&#8221; (create/replace/update/delete), standalone workflow tools, and per-seat productivity suites, the walls <em>were</em> the keep, and those companies are right to be scared. But for companies whose value lives in proprietary data, regulatory infrastructure, network effects, and institutional cognition? The breach of the outer wall might actually clarify their value, not destroy it &#8212; these companies will thrive in the agentic network as much or more than the human one.  </p><p>The market will figure this out. The question &#8212; as always &#8212; is when.</p><div><hr></div><p><em>Tom Davis is a physicist turned quant who builds fixed income analytics for a living and writes about the intersection of quantitative methods, AI, and finance. This piece was researched with the help of Astraea, an AI research assistant he built to power this newsletter.</em></p><p><em>Views expressed here are entirely my own and do not represent those of my employer. Nothing in this newsletter constitutes investment advice.</em></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><em>If this landed, there&#8217;s more where this came from</em></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://solitonmaths.substack.com/p/software-moats-ai-disruption?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://solitonmaths.substack.com/p/software-moats-ai-disruption?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p></p>]]></content:encoded></item></channel></rss>