<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[lance.tech]]></title><description><![CDATA[CEO @ OpenPay; Venture Partner @ Bessemer Venture Partners]]></description><link>https://www.lance.tech</link><image><url>https://substackcdn.com/image/fetch/$s_!QDmE!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2be09467-a5e0-47a2-af6a-0fd13e372215_1026x1026.png</url><title>lance.tech</title><link>https://www.lance.tech</link></image><generator>Substack</generator><lastBuildDate>Wed, 06 May 2026 11:13:11 GMT</lastBuildDate><atom:link href="https://www.lance.tech/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Lance Co Ting Keh]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[lancecotingkeh@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[lancecotingkeh@substack.com]]></itunes:email><itunes:name><![CDATA[Lance Co Ting Keh]]></itunes:name></itunes:owner><itunes:author><![CDATA[Lance Co Ting Keh]]></itunes:author><googleplay:owner><![CDATA[lancecotingkeh@substack.com]]></googleplay:owner><googleplay:email><![CDATA[lancecotingkeh@substack.com]]></googleplay:email><googleplay:author><![CDATA[Lance Co Ting Keh]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[AI Infra Roadmap: Five frontiers for 2026]]></title><description><![CDATA[The first generation of AI infrastructure companies unlocked the &#8220;brains&#8221; for intelligence. The next generation will unleash these engines of intelligence into the real-world.]]></description><link>https://www.lance.tech/p/ai-infra-roadmap-five-frontiers-for</link><guid isPermaLink="false">https://www.lance.tech/p/ai-infra-roadmap-five-frontiers-for</guid><dc:creator><![CDATA[Lance Co Ting Keh]]></dc:creator><pubDate>Mon, 30 Mar 2026 19:56:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!4ZtI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Originally published in Bessemer&#8217;s <a href="https://www.bvp.com/atlas/ai-infrastructure-roadmap-five-frontiers-for-2026">Atlas</a>; co-authored by Janelle Teng Wade, Talia Goldberg, David Cowan, Grace Ma, Bhavik Nagda, Brandon Nydick, Bar Weiner</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4ZtI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4ZtI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png 424w, https://substackcdn.com/image/fetch/$s_!4ZtI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png 848w, https://substackcdn.com/image/fetch/$s_!4ZtI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png 1272w, https://substackcdn.com/image/fetch/$s_!4ZtI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4ZtI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png" width="886" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e774e3c9-7989-4436-88ec-8c84cf520640_886x816.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:886,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:865789,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.lance.tech/i/192651853?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4ZtI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png 424w, https://substackcdn.com/image/fetch/$s_!4ZtI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png 848w, https://substackcdn.com/image/fetch/$s_!4ZtI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png 1272w, https://substackcdn.com/image/fetch/$s_!4ZtI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe774e3c9-7989-4436-88ec-8c84cf520640_886x816.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The first generation of AI was built for a world where the model was the product, and progress meant bigger weights, more data, and stellar benchmarks. AI infrastructure mirrored this reality, fueling the rise of giants in foundation models, compute capacity, training techniques, and data ops. This was the focus of our <a href="https://www.bvp.com/atlas/roadmap-ai-infrastructure">2024 AI Infrastructure Roadmap</a>, which drove our investments in companies such as <a href="https://anthropic.com/">Anthropic</a>, <a href="https://fal.ai/">Fal AI</a>, <a href="https://supermaven.com/">Supermaven</a> (acquired by <a href="https://cursor.com/">Cursor</a>), and <a href="https://vapi.ai/">VAPI</a> as the AI infrastructure revolution unfolded.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.lance.tech/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading lance.tech! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>But the landscape has changed. Big labs are moving beyond chasing benchmark gains to designing AI that interfaces with the real world, and enterprises are graduating from POCs to production. The infrastructure that got us here, which was optimized for scale and efficiency, won&#8217;t get us to the next phase. What&#8217;s needed now is infrastructure for grounding AI in operational contexts, real-world experience, and continuous learning.</p><p>The stage is being set for a new wave of AI infrastructure tools to enable AI to operate in the real world. We&#8217;ve identified five frontiers that will define this next wave, each addressing a structural limitation that needs to be solved beyond model scaling.</p><h2><strong>Five cutting-edge frontiers for next-gen AI infrastructure</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tFXX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tFXX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg 424w, https://substackcdn.com/image/fetch/$s_!tFXX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg 848w, https://substackcdn.com/image/fetch/$s_!tFXX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!tFXX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tFXX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!tFXX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg 424w, https://substackcdn.com/image/fetch/$s_!tFXX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg 848w, https://substackcdn.com/image/fetch/$s_!tFXX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!tFXX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F35d56441-8a48-43f8-a24c-1b7b9f1f87f7_1600x900.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3><strong>1. &#8220;Harness&#8221; infrastructure</strong></h3><p>As AI deployments shift from single models to compound systems, infrastructure designed to &#8220;harness&#8221; models becomes more important than ever.</p><p>Take memory and context management. Most enterprise AI systems suffer from organizational amnesia. While basic Retrieval-Augmented Generation (RAG) solved the connection problem between models and data sources, compound AI systems now require more sophisticated memory infrastructure. Enterprises hold vast amounts of historical data and organizational knowledge, from proprietary documents to CRM records, that AI systems must access to avoid hallucinations and stay grounded in company-specific reality.</p><p>Reliable AI deployment depends not just on raw model horsepower, but on orchestrating components like knowledge retrieval, cross-session context management, and planning. As models become commoditized, differentiation shifts to the memory and context layer. What developers once built from scratch &#8212; custom vector databases and retrieval systems &#8212; is now emerging as its own infrastructure category. Startups and <a href="https://research.google/blog/turboquant-redefining-ai-efficiency-with-extreme-compression/">Big Tech alike</a> now offer plug-and-play semantic layers that maintain conversation context, user preferences, and long-term memory across sessions.</p><p>Novel evaluation and observability present another critical infrastructure challenge &#8212; one that didn&#8217;t exist in prior software development paradigms. Consider teams shipping conversational AI agents to production. Traditional monitoring tracks completion rates, latency, error codes, and thumbs up/down feedback. But conversational AI fails differently. When a chatbot gives a confident wrong answer, gradually drifts from the user&#8217;s actual question, or misunderstands the request while producing something plausible, users often don&#8217;t react. No complaint, no thumbs down, no error signal. The conversation looks fine in dashboards, and AI just quietly failed.</p><p>An estimate <strong><a href="https://arxiv.org/abs/2603.15423">78% of AI failures are invisible</a></strong> &#8212; AI gets something wrong, but no one catches it. Not the user, not traditional monitoring, not even a sentiment analysis. These failures cluster into recurring patterns:</p><ul><li><p><strong>The confidence trap</strong> &#8212; AI is confidently wrong, and the user accepts it</p></li><li><p><strong>The drift</strong> &#8212; AI gradually answers a different question than what was asked</p></li><li><p><strong>The silent mismatch</strong> &#8212; AI misunderstands but produces something plausible enough that the user doesn&#8217;t push back</p></li></ul><p>These patterns persist across 93% of cases even with more powerful models, because they stem from interaction dynamics &#8212; how models present outputs and how users communicate intent &#8212; not capability gaps.</p><p>New infrastructure is emerging to address this. Platforms like <a href="http://bigspin.ai/">Bigspin.ai</a> provide not just pre-deployment testing but real-time production monitoring of model outputs against golden datasets and user feedback. We&#8217;re also moving beyond traditional analytics toward semantic metrics, with new platforms such as <a href="https://www.braintrust.dev/">Braintrust</a> and <a href="https://judgmentlabs.ai/">Judgment Labs</a>, as well as techniques such as LLM-as-a-judge, that are emerging for high-quality evals and metrics definition.</p><p><strong>These examples illustrate evolving needs for AI harness infrastructure. For more on environments, runtime, orchestration, protocols, and frameworks, see our <a href="https://www.bvp.com/atlas/roadmap-developer-tooling-for-software-3-0">Software 3.0 roadmap</a>.</strong></p><h3><strong>2. Continual learning systems</strong></h3><p>Today&#8217;s AI models face a fundamental constraint: frozen weights prevent true learning after deployment. While context management strategies like compaction are powerful, and we see many big labs use them for long-running agents, in-context learning enables only surface-level adaptation through rote memorization, not the acquisition of new skills. It also becomes prohibitively expensive as contexts grow, since the KV cache scales linearly with added context. From both technical and economic perspectives, it&#8217;s infeasible to build AI systems that remember everything and continuously improve over years of use.</p><p>This is where continual learning offers a solution. It enables AI to accumulate knowledge and skills across tasks over time, maintaining earlier capabilities while acquiring new ones. Unlike traditional models trained once and deployed statically, continual learning systems evolve in production &#8212; getting smarter with each interaction while avoiding catastrophic forgetting. Researchers and practitioners are pursuing this through innovations at both pre-training and post-training stages.</p><p>Architectural approaches fundamentally rethink how models learn:</p><ul><li><p><a href="https://learning-machine.ai/">Learning Machine</a> is building models that continuously learn during inference, as humans do. Through a new architecture and training paradigm, models will master the meta skill of &#8220;how to learn&#8221;, enabling adaptation to individual users and enterprises post-deployment</p></li><li><p>Core Automation is fundamentally rethinking transformer architecture to build systems where memory emerges naturally from novel attention mechanisms</p></li><li><p><a href="https://arxiv.org/pdf/2512.23675">Stanford and Nvidia&#8217;s TTT-E2E</a> uses a sliding-window Transformer that continues learning at test time through next-token prediction on its context &#8211; compressing that context into its weights. During training, the model learns how to better update its own weights at inference, making the approach end-to-end</p></li></ul><p>Near-term, production-ready solutions are also emerging:</p><ul><li><p>Engram&#8217;s <a href="https://scalingintelligence.stanford.edu/pubs/cartridges/">&#8220;cartridges&#8221; methodology</a> stores long contexts in small KV caches trained offline once, then reused across different user requests during inference</p></li><li><p>Sublinear Systems and foundation model labs are racing to address context limitations through novel techniques</p></li></ul><p><strong>The spectrum of approaches we&#8217;ve seen for continual learning ranges from high-risk architectural moonshots that could redefine the field entirely to production-ready techniques that incrementally improve existing transformers. We&#8217;re eager to meet founders across this spectrum.</strong></p><p>Production deployment of continual learning requires new governance primitives that don&#8217;t yet exist in standard ML workflows. Rollback mechanisms enable reversion to stable checkpoints when updates introduce regressions, requiring full lineage tracking of weights, data, and hyperparameters. Isolation techniques allow safe experimentation without affecting core capabilities. Creating benchmarks, beyond needle-in-the-haystack tests, to gauge the performance of continual learning systems versus in-context learning will also be critical.</p><h3><strong>3. Reinforcement learning platforms</strong></h3><p>With data quality fundamentally determining AI capabilities, the old machine learning axiom of &#8220;garbage in, garbage out&#8221; has never been more relevant. Data platforms such as <a href="https://www.mercor.com/">Mercor</a>, <a href="https://www.turing.com/">Turing</a>, and <a href="https://www.micro1.ai/">micro1</a> have been instrumental in the AI revolution&#8217;s first wave by mobilizing human expertise to create high-quality datasets. But we believe that as AI systems evolve from pattern recognition to autonomous decision-making, a critical limitation has emerged: human-generated labeled data is no longer enough to enable production-grade AI. It cannot teach AI systems how to navigate complex, multi-step tasks with delayed consequences and compounding decisions.</p><p>This is where reinforcement learning (RL) becomes essential, as AI must learn through interaction rather than static datasets to ground the AI in &#8220;experience.&#8221; Leveraging an RL stack is now a cornerstone of AI infra tooling to teach agents complex behaviors without the cost and risk of real-world trial and error. Platforms in this emerging stack include:</p><p>Environment building and experience curation</p><p><a href="https://www.bespokelabs.ai/">Bespoke Labs</a>, <a href="https://deeptune.com/">Deeptune</a>, <a href="http://fleet.so/">Fleet</a>, <a href="https://www.habitat.inc/">Habitat</a>, <a href="https://matrices.ai/">Matrices</a>, <a href="https://www.mechanize.work/">Mechanize,</a> <a href="https://openreward.ai/">OpenReward</a>, <a href="https://www.phinity.ai/">Phinity</a>, <a href="https://www.preferencemodel.com/">Preference Model</a>, <a href="https://proximal.ai/">Proximal,</a> <a href="https://www.sepalai.com/">SepalAI</a>, <a href="https://steadyworks.ai/">Steadyworks, </a><a href="https://veris.ai/">Veris</a>, <a href="https://vmax.ai/">VMax</a></p><p>RL-as-a-service</p><p><a href="https://appliedcompute.com/">Applied Compute</a>, <a href="https://cgft.io/">cgft</a>, <a href="https://www.withmetis.ai/">Metis</a>, <a href="https://osmosis.ai/">osmosis, </a><a href="https://trajectory.ai/">Trajectory</a></p><p>Platform infrastructure</p><p><a href="https://www.agilerl.com/">AgileRL</a>, <a href="https://www.hud.ai/">Hud</a>, <a href="https://www.isidor.ai/">Isidor</a>, <a href="https://openpipe.ai/blog/announcing-6-7m-seed-raise">OpenPipe</a>, <a href="https://www.primeintellect.ai/">Prime Intellect</a>, <a href="https://thinkingmachines.ai/tinker/">Tinker</a></p><h3><strong>4. Inference inflection point</strong></h3><p>Model deployment and inference optimization emerged as a critical infrastructure layer in our 2024 roadmap, when vendors like <a href="https://fal.ai/">Fal</a>, <a href="https://www.together.ai/">Together</a>, <a href="https://www.baseten.co/">Baseten</a>, and <a href="https://fireworks.ai/">Fireworks</a> pioneered efficient serving solutions. At that time, capital-intensive model training consumed the majority of compute resources across the AI stack. Today, we&#8217;re witnessing a fundamental shift in the compute center of gravity. As AI agents and applications transition from prototype to production at scale, inference workloads now rival, and in many cases exceed, training in both compute demand and economic importance. As NVIDIA&#8217;s <a href="https://www.youtube.com/watch?v=jw_o0xr8MWU">Jensen Huang stated in his GTC 2026 keynote</a>, &#8220;Finally, AI is able to do productive work, and therefore the inflection point of inference has arrived.&#8221;</p><p><strong>This inflection point reflects a maturing market where the cost and performance of running AI systems continuously matter just as much as the initial investment in building them.</strong></p><p>A new generation of infrastructure startups is addressing this production imperative through specialized optimization across the inference stack. Companies like <a href="https://www.tensormesh.ai/">TensorMesh</a> are leveraging <a href="https://lmcache.ai/">LMCache</a> to eliminate redundant re-computation, <a href="https://www.radixark.ai/">RadixArk</a> is advancing SGLang-based routing and scheduling for multi-turn conversations, and <a href="https://inferact.ai/">Inferact</a> is pushing vLLM performance boundaries for high-throughput serving. <a href="https://gimletlabs.ai/">Gimlet Labs</a> and even hyperscalers like <a href="https://developer.nvidia.com/blog/inside-nvidia-groq-3-lpx-the-low-latency-inference-accelerator-for-the-nvidia-vera-rubin-platform/">NVIDIA</a> are working on heterogeneous inference innovations purpose-built for complex agentic systems. These innovations translate cutting-edge systems research into measurable production gains: faster response times and lower costs.</p><p>We&#8217;re also seeing innovations in inference for novel deployments, with edge and on-device as one prime example. As AI proliferates all sectors of the economy, from robotics to consumer, AI deployments need to meet users where they are, which isn&#8217;t always cloud-based. We&#8217;re seeing companies such as <a href="https://www.webai.com/">WebAI</a>, <a href="https://femto.ai/">FemtoAI</a>, <a href="https://www.polargrid.ai/">PolarGrid</a>, <a href="https://aizip.ai/">Aizip Mirai</a>, and <a href="https://openinfer.io/">OpenInfer</a> build at the very &#8220;edge&#8221; of what&#8217;s possible for on-device AI deployments in consumer devices. On-device innovations from model vendors such as <a href="https://www.perceptron.inc/">Perceptron</a> are also important for physical AI, and we expect more in the space as we outlined in <a href="https://www.bvp.com/atlas/intelligent-robotics-the-new-era-of-physical-ai">our thinking on intelligent robotics</a>.</p><p><a href="https://www.bvp.com/atlas/defense-tech-roadmap-five-frontiers-for-2026">Edge AI is also critical for industries such as defense</a>, where comms are jammed or denied; companies such as <a href="https://www.turbineone.com/">TurbineOne</a>, <a href="https://www.bvp.com/news/dominion-dynamics-forging-the-future-of-interoperable-attritable-systems-for-arctic-and-allied-defense">Dominion Dynamics</a>, <a href="https://picogrid.com/">Picogrid</a>, and <a href="https://breakerindustries.com/">Breaker</a> are leading the charge on providing the infrastructure tooling for warfighters to harness the power of AI even in the most austere environments.</p><h3><strong>5. World models</strong></h3><p><a href="https://www.bvp.com/atlas/roadmap-ai-infrastructure#1-Innovations-in-scaling-novel-model-architectures-and-specialized-purpose-foundation-models">The model layer is one of the most dynamic and hotly contested layers within the AI infrastructure stack</a>. While LLMs have taken over language intelligence, a new class of models &#8212; world models &#8212; has emerged to deliver intelligence for the physical world.</p><p>As AI moves from our screens to our physical realities, new challenges arise: how does an AI &#8220;brain&#8221; develop intuition for physics and the world if it has no &#8220;body&#8221;? World models offer a solution. At the core, these are AI systems trained on real-world data &#8212; video, sensors, GPS, and more &#8212; that learn to predict how the world evolves given a current situation and action. Rather than describing reality, they simulate it.</p><p>Out of this newer research, three broad architectural paradigms have emerged. In practice, companies are also beginning to explore hybrids that combine elements of each:</p><ul><li><p><strong>Video-based world models</strong> from companies such as <a href="https://reka.ai/">Reka</a> and <a href="https://decart.ai/">Decart</a> frame the problem as one of video generation, predicting future frames directly in pixel space. Because they generate outputs step-by-step, they can operate in real time and respond dynamically to new inputs, making them well-suited for interactive environments. Though they still struggle with maintaining physical consistency over longer horizons, they produce visually compelling outputs</p></li><li><p><strong>Explicit 3D representation models</strong> from companies such as <a href="https://www.worldlabs.ai/">World Labs</a> take a different path, constructing persistent 3D scene representations that deliver strong spatial coherence at a lower inference cost. For now, these environments are pre-generated and static, but World Labs has signaled that real-time interactivity is on its roadmap</p></li><li><p><strong>Latent predictive models</strong>, based on Joint Embedding Predictive Architectures (JEPA) pioneered by <a href="https://amilabs.xyz/">AMI Labs</a>, avoid pixel generation altogether by forecasting future states in a compressed latent space. This approach is significantly more compute-efficient and sidesteps many visual failure modes, but comes with reduced interpretability. While each paradigm has seen meaningful progress, important gaps remain &#8212; how these are resolved will shape the path to the broader commercialization of world models</p></li></ul><p>This commercial opportunity for world models is expansive. We recently shared our view of <a href="https://www.bvp.com/atlas/can-world-models-unlock-general-purpose-robotics">world models in robotics</a>, as this sector has been among the most visible early applications. By generating unlimited synthetic training environments, world models solve the data scarcity problem that has bottlenecked physical AI for decades. Autonomous driving is proving this as Waymo and Wayve use world models to simulate rare edge cases that no real-world test program could economically replicate. The same core capability unlocks even more, such as high-stakes simulation in defense, healthcare, industrial operations, and enterprise planning.</p><p>World models are not a vertical-specific kind of tool &#8212; they&#8217;re a new substrate for machine intelligence, analogous to what LLMs did for text-based reasoning. The industries that build on top of them early will have a significant head start on deploying agents that work in the real world. We&#8217;re excited about companies building the architectures and simulators that make world models possible across industries.</p><h2><strong>Building infrastructure for AI to experience and enter the real world</strong></h2><p>While the first generation of AI infrastructure companies built the engines of intelligence &#8212; the models, compute clusters, and training pipelines that proved AI&#8217;s capability &#8212; the next generation must build the nervous system and harnesses that allow AI to sense, remember, adapt, and operate continuously in the real world. These frontiers represent more than incremental improvements to existing infrastructure. The companies building in these spaces aren&#8217;t just optimizing latency or reducing costs; they&#8217;re solving the fundamental challenges that separate impressive demos from reliable systems that create enduring value.</p><p>We believe 2026 will be the year when AI infrastructure&#8217;s center of gravity definitively shifts, reimagining what AI-native operations look like for this year and beyond. <strong>We&#8217;re particularly excited to work with founders who are pursuing these endeavors. To get in touch with us, please contact aiinfra@bvp.com.</strong></p><p><strong><sup>Disclaimer</sup></strong><sup>: The information presented here is for general informational and educational purposes only and does not constitute investment advice, a recommendation, or an offer or solicitation to buy or sell any securities or investment products. The information presented is also not intended as advertising material under the Investment Advisers Act. Certain companies discussed may be current or former portfolio companies. BVP may still have a financial interest in these companies. Any discussion of specific companies, securities, or investment strategies should not be considered a recommendation to take any particular action. Past performance is not indicative of future results. All investments involve risk, including possible loss of principal. Market conditions and investment returns can fluctuate significantly. Please visit https://www.bvp.com/legal for more information.</sup></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.lance.tech/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading lance.tech! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[ChipAgents: Agentic AI for Chip Design]]></title><description><![CDATA[Bessemer Venture Partners leads ChipAgents&#8217; $21M Series A to transform chip design with agentic AI.]]></description><link>https://www.lance.tech/p/chipagents-agentic-ai-for-chip-design</link><guid isPermaLink="false">https://www.lance.tech/p/chipagents-agentic-ai-for-chip-design</guid><dc:creator><![CDATA[Lance Co Ting Keh]]></dc:creator><pubDate>Fri, 20 Mar 2026 16:27:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!26RW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!26RW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!26RW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png 424w, https://substackcdn.com/image/fetch/$s_!26RW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png 848w, https://substackcdn.com/image/fetch/$s_!26RW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png 1272w, https://substackcdn.com/image/fetch/$s_!26RW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!26RW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png" width="794" height="620" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:620,&quot;width&quot;:794,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:519623,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.lance.tech/i/191598731?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!26RW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png 424w, https://substackcdn.com/image/fetch/$s_!26RW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png 848w, https://substackcdn.com/image/fetch/$s_!26RW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png 1272w, https://substackcdn.com/image/fetch/$s_!26RW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2d926df-189a-4c3f-8443-9c04bf4a6989_794x620.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p><em>Originally published in Bessemer&#8217;s <a href="https://www.bvp.com/news/noda-ai-building-the-future-of-operational-collaborative-autonomy-at-the-frontlines">Atlas</a>; co-authored by Lance Co Ting Keh, Jason Scheller, David Cowan </em></p><p>For years, chip design and verification have remained among the most complex and resource-intensive challenges in engineering. Hardware teams spend months navigating fragmented electronic design automation (EDA) toolchains, long simulation runtimes, and manual verification workflows that stretch development cycles and inflate costs. Designing at the Register Transfer Level (RTL), where logic is described in languages like Verilog or VHDL, demands both software-like precision and deep hardware intuition. Each change must be validated across thousands of simulation cases, with engineers writing extensive test benches and debugging waveform outputs line by line. Design and verification can consume anywhere from 60-80% of total chip development time, yet these painstaking efforts are indispensable, forming the foundation for everything from AI accelerators and data-center processors to automotive controllers.</p><p><a href="https://chipagents.ai/">ChipAgents</a> marks a breakthrough in this long-standing bottleneck. The company&#8217;s agentic AI platform reimagines how chips are designed, debugged, and verified, enabling faster design cycles, automated verification, and seamless collaboration between human engineers and AI. Built to feel native within the environments and tools engineers already rely on, ChipAgents fits seamlessly into existing design flows while introducing new levels of intelligence and automation. By transforming how teams generate RTL code, debug complex systems, and validate designs, ChipAgents is unlocking a new era of speed, quality, and creativity in hardware engineering. Early customers report that tasks which once required weeks of manual effort can now be completed in days without compromising correctness or performance.</p><p>This is why Bessemer&#8217;s Lance Co Ting Keh, Jason Scheller, and David Cowan are excited to lead ChipAgents&#8217; $21 million Series A. With oversight from industry veterans like Wally Rhines, Ra&#250;l Camposano, and Jack Harding, the company&#8217;s vision goes beyond innovation. To learn more, we sat down with founder and CEO William Wang for his thoughts about how ChipAgents is defining the future of EDA.</p><h2><strong>Q&amp;A with the founder of ChipAgents</strong></h2><p><strong>Tell us about yourself.</strong></p><p>Before starting ChipAgents.ai, I spent most of my career in academia and currently serve as the Duncan and Suzanne Mellichamp Endowed Chair in Artificial Intelligence and Design at the University of California, Santa Barbara. Over the past nine years at UCSB, I&#8217;ve built the Natural Language Processing research group into a leading center for AI innovation, collaborating with nearly all major technology companies to advance fundamental algorithms in natural language processing, artificial intelligence, and machine learning.</p><p><strong>What inspired you to start ChipAgents?</strong></p><p>My journey toward founding ChipAgents started long before the deep learning revolution. Back in 2011, during my PhD at Carnegie Mellon University, we were experimenting with recurrent neural network language models, well before the deep learning era and even before ImageNet transformed the field.</p><p>My doctoral research at CMU focused on theorem proving and formal methods. We developed an approximate personalized PageRank algorithm that made inference for theorem proving <em>locally groundable</em>, meaning that inference time became independent of the size of the underlying database. That early work on scalable, interpretable reasoning planted the seeds for what would eventually become ChipAgents&#8217; mission in verification.</p><p>At UCSB, my group continued pushing the boundaries of reasoning and learning. We created DeepPat, the industry&#8217;s first deep reinforcement learning framework for reasoning models in 2017, which demonstrated how neural networks can learn to reason through complex knowledge graphs. By 2024, it became clear that these technologies had matured to a point where they could fundamentally transform how we design and optimize chips. That belief, combining decades of AI research with the opportunity to reinvent EDA, was what inspired me to start ChipAgents.</p><p><strong>What exactly does ChipAgents do?</strong></p><p>At ChipAgents, we&#8217;re building intelligent AI agents that transform how chips are designed and verified. Our mission is to bring the power of agentic AI into the core of the EDA process, making chip design faster, more reliable, and far more scalable. Traditionally, verification, ensuring that a chip design behaves exactly as intended &#8212; has been one of the most complex, time-consuming, and expensive stages of semiconductor development. Modern chips can contain billions of transistors, and verifying every interaction has required massive engineering effort. At ChipAgents, we&#8217;re using AI reasoning agents that can analyze, understand, and even prove design properties automatically, dramatically reducing verification bottlenecks.</p><p>Our technology allows these agents to read design specifications, reason about logical correctness, and generate verification artifacts or proofs, using the same class of language and reasoning models that have revolutionized natural language understanding. This creates a new level of automation, one that goes beyond pattern matching or simulation, toward true semantic understanding of hardware behavior.</p><p>ChipAgents builds on more than a decade of AI research. From my early work at Carnegie Mellon on theorem proving and locally groundable inference, to our development of deep reinforcement learning reasoning models, we&#8217;ve been exploring how AI can perform structured, explainable reasoning. Now, that foundation enables us to apply these methods to one of the most critical challenges in modern computing: making chip verification intelligent, adaptive, and orders of magnitude more efficient.</p><p><strong>Tell us about the momentum you&#8217;re seeing at ChipAgents.</strong></p><p>We&#8217;re seeing incredible momentum &#8212; sales are up more than 50x year-over-year in ARR, and usage has grown over 60x. It&#8217;s a super exciting time to witness how AI is transforming chip design and verification at a fundamental level. We&#8217;re still at the early stages of this revolution, but the acceleration we&#8217;re seeing from customers and partners makes it clear that intelligent agents are redefining what&#8217;s possible in the semiconductor industry.</p><p><strong>How might chip companies redefine their competitive advantage when the barrier isn&#8217;t hardware capability but AI workflow mastery?</strong></p><p>I believe the entire workflow of chip design and verification will change in the era of agentic AI. In the past, engineers spent enormous time implementing specifications, writing test plans, and generating test stimuli by hand. But as AI becomes a core part of the workflow, competitive advantage will shift from manual implementation to how effectively teams can orchestrate AI agents &#8212; how they write better prompts, configure intelligent workflows, and verify AI-generated collateral with confidence. In other words, success won&#8217;t just depend on who has the most powerful chips, but on who can best collaborate with AI to accelerate innovation, ensure correctness, and continuously optimize the design cycle.</p><p><strong>What is your vision for the future of ChipAgents?</strong></p><p>Our vision is to go beyond point tools in EDA and create a truly end-to-end agentic AI design and verification system, from RTL to GDS. We want to build an integrated workflow where AI agents can understand design intent, reason across abstraction levels, and continuously improve through feedback.</p><p>Ultimately, we aim to close the loop between pre-silicon and post-silicon, using real-world performance and validation data to inform the next generation of design. This creates a virtuous cycle of agentic AI, where the system not only automates design and verification but also learns and optimizes itself over time. That&#8217;s how we see the future of chip design: intelligent, adaptive, and self-improving.</p><p><strong>What is a common misconception about the future of AI? Or something that people are not thinking about, but should be?</strong></p><p>One of the biggest misconceptions about AI is that it&#8217;s here to replace engineers. In reality, AI will transform the nature of engineering work, not eliminate it. There&#8217;s often a psychological barrier when people imagine AI &#8220;taking over&#8221; their roles, but at ChipAgents, our mission is the opposite: to enable engineers to 10x (even 100x productivity in the future), not replace them. We&#8217;re giving engineers the tools to shift from implementation to innovation, to spend less time on tedious, repetitive tasks and more time designing new architectures, exploring new ideas, and creating better chips. The future of AI in chip design isn&#8217;t about automation for its own sake; it&#8217;s about augmenting human creativity and enabling engineers to operate at an entirely new level of abstraction and productivity.</p>]]></content:encoded></item><item><title><![CDATA[DeepSeek R1 ]]></title><description><![CDATA[and Some Implications for Startups]]></description><link>https://www.lance.tech/p/deepseek-r1</link><guid isPermaLink="false">https://www.lance.tech/p/deepseek-r1</guid><dc:creator><![CDATA[Lance Co Ting Keh]]></dc:creator><pubDate>Sun, 26 Jan 2025 06:26:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!dcL9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4005061-2e01-4e60-a9a6-6544722c9a23_1260x660.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>How We&#8217;ve Been Training Foundation Models</h2><p>If you wanted to train a state-of-the-art foundation model over the past few years, the playbook was pretty well established. Start with pre-training: grab a massive, largely unlabeled corpus of internet-scale text, throw it into a gigantic transformer&#8212;and let it soak up the statistical patterns of language. This stage is computationally expensive but crucial&#8212;it&#8217;s what gives models like GPT-4, Llama, and Claude their broad generalization abilities.</p><p>Once you&#8217;ve got a decently capable language model, the next step is Supervised Fine-Tuning (SFT). Here, you refine the model on high-quality labeled datasets for specific tasks. The idea dates back to GPT-1 (2018), which introduced fine-tuning as a critical step for transfer learning in large-scale transformers. Other key milestones include DecaNLP (2018) and T5 (2019), which further popularized the notion of multi-task learning leading to fine-tuning. ULMFiT (2018) was another early example demonstrating how pre-trained language models could be adapted efficiently to downstream tasks.</p><p>Finally, there&#8217;s RLHF&#8212;Reinforcement Learning from Human Feedback. This phase is what makes modern chatbots feel human-like. RLHF starts with human annotators ranking multiple model outputs, and then the model is trained via reinforcement learning (typically PPO) to generate responses that align better with human preferences.</p><p>This stack&#8212;pre-training &#8594; SFT &#8594; RLHF&#8212;has been the  standard. DeepSeek is now challenging that paradigm.</p><p></p><h2>The DeepSeek Breakthrough: Reinforcement Learning First</h2><p>DeepSeek-R1 flips the script. Instead of fine-tuning on a curated dataset before reinforcement learning, DeepSeek-R1-Zero skips supervised fine-tuning entirely. The model is trained via pure RL from the get-go, incentivizing reasoning capabilities directly. This is a big deal.</p><h3>What&#8217;s Special About R1-Zero?</h3><p>Without any SFT, DeepSeek-R1-Zero still manages to exhibit strong reasoning capabilities, particularly in Chain-of-Thought (CoT) style problem-solving. The model achieves impressive benchmarks purely through RL optimization, with no initial human-labeled data.</p><p>A few notable highlights:</p><ul><li><p><strong>CoT emerges naturally</strong>&#8212;not because it was explicitly trained on examples, but because RL incentivized longer reasoning chains.</p></li><li><p><strong>Reflection as an emergent behavior</strong>&#8212;the model learns to step back, reassess its approach, and self-correct, all through RL.</p></li><li><p><strong>GRPO (Group Relative Policy Optimization)</strong>&#8212;an RL technique that avoids the traditional critic model, instead using a ranking-based approach to improve efficiency.</p></li><li><p><strong>Majority voting and test-time compute</strong>&#8212;DeepSeek-R1-Zero leverages <strong>two key test-time compute strategies</strong> to enhance accuracy and reasoning depth:</p></li></ul><ol><li><p><strong>Extended Generation Length</strong>&#8212;The model autonomously <strong>allocates more compute per response</strong>, generating longer reasoning chains (hundreds to thousands of tokens) as it refines its thought process. This emergent behavior leads to <strong>higher accuracy on complex reasoning tasks</strong>.</p></li><li><p><strong>Majority Voting</strong>&#8212;For benchmarks like AIME 2024, <strong>16 candidate responses were generated per question</strong>, and accuracy was computed based on majority voting. </p></li></ol><p>Benchmark results for R1-Zero show remarkable performance gains across multiple reasoning tasks. On AIME 2024, it achieved a pass@1 score of 71.0%, and when using majority voting, the accuracy jumped to 86.7%, rivaling OpenAI&#8217;s o1-0912. The model also demonstrated strong performance on math (MATH-500 at 95.9%) and coding (Codeforces percentile of 60.0%). These results underscore how powerful reinforcement learning alone can be in developing advanced reasoning capabilities without the need for initial supervised fine-tuning.</p><h3>DeepSeek-R1: Adding a Cold Start</h3><p>DeepSeek-R1 improves upon R1-Zero by incorporating a small amount of high-quality &#8220;cold start&#8221; data before RL. This helps stabilize training and improve readability. The pipeline looks like this:</p><ol><li><p><strong>Cold Start Data</strong>&#8212;800K carefully curated high-quality samples.</p><ul><li><p>600K are generated via rejection sampling from an RL-trained checkpoint.</p></li><li><p>200K cover task-specific data (e.g., factual QA, writing, self-cognition, and translation).</p></li><li><p>The rejection sampling process refines outputs to ensure higher reasoning coherence.</p></li></ul></li><li><p><strong>Two RL stages</strong>&#8212;</p><ul><li><p><strong>Stage 1: Reasoning Optimization</strong>&#8212;The first RL stage focuses purely on enhancing reasoning capabilities, driving emergent CoT-style responses and improving problem-solving efficiency.</p></li><li><p><strong>Stage 2: Alignment with Human Preferences</strong>&#8212;Once reasoning optimization stabilizes, a second RL stage fine-tunes outputs for readability and human alignment, reducing incoherence while preserving strong reasoning skills.</p></li></ul></li><li><p><strong>SFT after RL</strong>&#8212;Unlike traditional approaches, DeepSeek uses SFT to refine the RL-learned behaviors rather than kickstarting them.</p></li></ol><h3>Distillation: Making It Even More Efficient</h3><p>One of the most exciting results from DeepSeek-R1 is that <strong>distillation alone, without reinforcement learning, still produces highly capable models</strong>. By using the same 800K high-quality samples curated from RL-trained outputs, DeepSeek was able to train distilled versions of both Llama and Qwen models that demonstrated remarkable reasoning performance.</p><ul><li><p>Llama and Qwen models distilled&#8212;Distillation was successfully applied to both Llama and Qwen, producing models that outperformed their non-distilled counterparts in reasoning tasks. The distilled Qwen-32B and Llama-70B models demonstrated superior pass@1 scores in math, coding, and general reasoning benchmarks.</p></li><li><p><strong>Distillation alone yields strong Chain-of-Thought (CoT) reasoning</strong>&#8212;The distilled models retained CoT capabilities without needing explicit reinforcement learning, highlighting how easily these reasoning skills transfer.</p></li><li><p><strong>Distillation vs RL on Qwen</strong>&#8212;DeepSeek compared training RL directly on Qwen vs distilling R1&#8217;s output into Qwen. The latter consistently outperformed RL-only models, suggesting that knowledge transfer from a stronger model is more effective than RL on smaller models alone.</p></li></ul><h3>Prompting Simplicity</h3><p>A major advantage across all DeepSeek models is <strong>prompting simplicity</strong>. Unlike traditional models that require extensive prompt engineering, DeepSeek models leverage a straightforward "Think-Answer" format, making them easier to integrate into applications with minimal adaptation.</p><div id="datawrapper-iframe" class="datawrapper-wrap outer" data-attrs="{&quot;url&quot;:&quot;https://datawrapper.dwcdn.net/ethVO/1/&quot;,&quot;thumbnail_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b4005061-2e01-4e60-a9a6-6544722c9a23_1260x660.png&quot;,&quot;thumbnail_url_full&quot;:&quot;&quot;,&quot;height&quot;:337,&quot;title&quot;:&quot;Benchmark Results&quot;,&quot;description&quot;:&quot;&quot;}" data-component-name="DatawrapperToDOM"><iframe id="iframe-datawrapper" class="datawrapper-iframe" src="https://datawrapper.dwcdn.net/ethVO/1/" width="730" height="337" frameborder="0" scrolling="no"></iframe><script type="text/javascript">!function(){"use strict";window.addEventListener("message",(function(e){if(void 0!==e.data["datawrapper-height"]){var t=document.querySelectorAll("iframe");for(var a in e.data["datawrapper-height"])for(var r=0;r<t.length;r++){if(t[r].contentWindow===e.source)t[r].style.height=e.data["datawrapper-height"][a]+"px"}}}))}();</script></div><p></p><h2>What This Might Mean for Startups</h2><ul><li><p><strong>Simple prompt</strong>&#8212;DeepSeek&#8217;s "Think-Answer" format significantly reduces the need for complex prompt engineering. Efforts in tuning prompts could instead be pushed up the stack into distilling reasoning models.</p></li><li><p><strong>Ease of training CoT</strong> - With RL-first approaches yielding strong reasoning capabilities, smaller teams may be able to bypass expensive supervised fine-tuning, allowing them to deploy domain-specific reasoning models quickly. The bar for training high-quality reasoning models has dropped significantly.</p></li><li><p><strong>Distillation on a small dataset</strong>&#8212;The ability to fine-tune reasoning models effectively with a relatively small dataset (800K samples) demonstrates that high performance can be achieved without massive data collection. Spending resources building a high-quality data moat seems correct.</p></li><li><p><strong>Test-time compute optimization</strong>&#8212; The trend continues. Product teams should delineate which AI tasks require high-quality, infrequent responses (where test-time compute strategies like extended generation and majority voting can be leveraged) versus high-throughput, latency-sensitive tasks better served by smaller, fine-tuned models.</p></li></ul><p>The research community is moving quickly to validate and expand on these findings, with teams like <a href="https://github.com/huggingface/open-r1">Hugging Face already diving in</a>. Expect rapid iteration, fresh insights, and potential breakthroughs as more teams test and refine DeepSeek&#8217;s approach. Hats off to the DeepSeek team for this work!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.lance.tech/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading lance.tech! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Thoughts on OpenAI's o3 ]]></title><description><![CDATA[and where I'd apply expensive but effective CoT models today]]></description><link>https://www.lance.tech/p/thoughts-on-openais-o3</link><guid isPermaLink="false">https://www.lance.tech/p/thoughts-on-openais-o3</guid><dc:creator><![CDATA[Lance Co Ting Keh]]></dc:creator><pubDate>Sun, 29 Dec 2024 00:11:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bwlh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello World! I&#8217;ve long told myself that I should start writing more publicly and what better time to do so than to talk briefly about the <a href="https://www.youtube.com/watch?v=SKBG1sqdyIU">latest OpenAI model, o3.</a></p><p></p><h2><strong>The ARC-AGI benchmark</strong></h2><p>Francois Collet, the creator of Keras, argued in 2019 that task-specific benchmarks are doomed to overfit in some shape or form. Overfitting can arise <em>explicitly</em> from sufficient data coverage in a domain or <em>implicitly</em> from biases in the training or evaluation sets, feature over-reliance, or other algorithmic artifacts. To combat this, he released a benchmark called the Abstract Reasoning Corpus (ARC) in his paper <a href="https://arxiv.org/abs/1911.01547">On the Measure of Intelligence</a>. This benchmark is designed to evaluate the general reasoning abilities of AI systems beyond narrow tasks. It consists of abstract pattern-recognition problems that require understanding, generalization, and creativity to solve.</p><p>The <a href="https://github.com/fchollet/ARC-AGI">benchmark</a> emphasizes reasoning over statistical pattern-matching, measuring how well AI systems can approach tasks they haven't been explicitly trained on.</p><p>The following excerpt and figure describes the dataset well: &#8220;ARC-AGI consists of unique training and evaluation tasks. Each task contains input-output examples. The puzzle-like inputs and outputs present a grid where each square can be one of ten colors. A grid can be any height or width between 1x1 and 30x30&#8221;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bwlh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bwlh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png 424w, https://substackcdn.com/image/fetch/$s_!bwlh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png 848w, https://substackcdn.com/image/fetch/$s_!bwlh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png 1272w, https://substackcdn.com/image/fetch/$s_!bwlh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bwlh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png" width="1456" height="733" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:733,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bwlh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png 424w, https://substackcdn.com/image/fetch/$s_!bwlh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png 848w, https://substackcdn.com/image/fetch/$s_!bwlh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png 1272w, https://substackcdn.com/image/fetch/$s_!bwlh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63dd27d1-8bf0-46c3-8cde-514709a49822_1510x760.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Source: <a href="https://arcprize.org/arc">https://arcprize.org/arc</a></em></figcaption></figure></div><p>While I don&#8217;t consider ARC-AGI to be the ultimate benchmark for AGI, I have long believed that it remains an exceptionally robust measure of a model&#8217;s general reasoning versatility that clearly gets around memorization and demonstrates the ability to solve novel, unstructured problems.</p><p>The community has experimented with a number of approaches to attacking the ARC-AGI benchmark, including both deep learning and &#8220;classical&#8221; statistical learning approaches. Not surprisingly most of the recent ones were LLM based. The ARC team published a <a href="https://arxiv.org/abs/2412.04604">technical paper</a> talking about some of the best 2024 approaches. We&#8217;ll cover some of these approaches in another post, but the best results before o3 hovered around ~55% (on the private held-out set), achieved in the work titled <a href="https://arxiv.org/abs/2411.02272)">Combining Induction and Transduction for Abstract Reasoning</a> by Li, et. al. Human performance is reported to be around 85%. </p><p></p><h2><strong>o3 results</strong></h2><p>o3 scored an amazing 75.7% on ARC-AGI&#8217;s semi-private eval set in the low compute mode and 87.5% in the high compute mode. There are a few disclaimers here. o3 was reportedly fine-tuned on 300 of 400 ARC training tasks and the runs were on the semi-private eval set (presumably because of the inability to run o3 in complete isolation [no internet]). Nevertheless, it still outperforms in other well known benchmarks:</p><ul><li><p><a href="https://www.swebench.com/">SWE bench</a> - simulates real-world software engineering issues by providing a codebase and an issue description. o3 scored 71.7%, up from <a href="https://aide.dev/blog/sota-bitter-lesson">62.2% by the Aide team</a>, and 49% by Claude 3.5 (sonnet).</p></li><li><p><a href="https://codeforces.com/blog/entry/22260?mobile=false">Codeforces</a> - competitive programming problems (similar to those in the Informatics Olympiad [IOI]) . o3 scores an Elo of 2727, up from 1891 rating for o1.</p></li><li><p>AIME 2024 - math problems - o3 made just one error (whopping my personal best of three errors in high school!).</p></li><li><p><a href="https://arxiv.org/abs/2311.12022">GPQA</a> - multiple choice in bio, physics and chemistry. o3 scores 87.7% up from 78% from o1 and Claude.</p></li><li><p><a href="https://epoch.ai/frontiermath">EpochAI&#8217;s FrontierMath</a> - hard math problems. o3 scored 25.2% up from 2%</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1RIp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1RIp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png 424w, https://substackcdn.com/image/fetch/$s_!1RIp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png 848w, https://substackcdn.com/image/fetch/$s_!1RIp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png 1272w, https://substackcdn.com/image/fetch/$s_!1RIp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1RIp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png" width="1456" height="763" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:763,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1RIp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png 424w, https://substackcdn.com/image/fetch/$s_!1RIp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png 848w, https://substackcdn.com/image/fetch/$s_!1RIp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png 1272w, https://substackcdn.com/image/fetch/$s_!1RIp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1a1d26d-5a28-41c3-b008-3096229b80b7_1600x838.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Source: <a href="https://www.youtube.com/live/SKBG1sqdyIU?si=Pqx8uANCZHldABM9">OpenAI&#8217;s announcement video</a></em></figcaption></figure></div><h6></h6><p>These are very impressive results across the board and we have a lot to be excited about. Given what we&#8217;ve seen with o1 and now o3, it seems that Chain of Thought (CoT) models drive the best performance for complex multi-step tasks given an unlimited inference cost and latency budget. Even Anthropic&#8217;s Claude, which to our knowledge was not trained directly to decompose problems, still encourages the use of <a href="https://www.anthropic.com/news/prompt-improver">CoT via prompting</a>.</p><p></p><h2><strong>CoT models for startups today</strong></h2><p>At OpenPay and in my role as a Venture Partner for Bessemer Venture Partners, I think a lot about how the latest models and frameworks affect startups. The very nature of CoT involves taking a complex query and breaking it down into step-by-step reasoning. The obvious trade-off here is the move towards inference-time-compute (as opposed to training) to achieve performance. Even in the &#8220;low compute&#8221; setting, running inference on o3 allegedly costs a whopping ~$20 per query. Models will only get cheaper and faster in time, but until then what might this mean for some AI use cases today:</p><ul><li><p>Agentic frameworks and architectures - these high cost CoT models are clearly not usable for most sub-task/tool calling components of agentic architectures today, especially because for most subtasks smaller models will do just fine. However, high value tasks that are less frequently called or called at the top of the stack such as planning or meta prompting are prime candidates for these high performance, high cost models.</p></li><li><p>Code Generation and Dev tooling - the cost of running o3 is prohibitive for quick and iterative conversation right now, but there are many use cases where escalating to what should be closer to an L7 engineer clears the threshold of value. I think there can be good product experiences designed around this, especially human-in-the-loop experiences that know how to clarify and re-plan.</p></li><li><p>High Value Vertical SaaS - We all know models will keep getting better at generalizing and reasoning over time, unlocking more and more use cases. But high value use cases that are relatively asynchronous, tolerant to error and where elegant product experiences can be built could be early beneficiaries. Fields like legal, research and consulting all clear the bar.</p></li><li><p>Data labeling and Post training - more fields can benefit from training vertical CoT models, which in turn will require human annotated &#8220;chains-of-thought&#8221; labels (step-by-step reasoning labels, intermediate annotations, etc) for post-training. Companies can probably be built around the next generation of labeling tools to properly support these complex labeling tasks. This might be extended further to provide the end-to-end post-training loop.</p></li><li><p>Synthetic data - a good friend and mentor Alex Kvamme thinks there is a lot to be built here, and I agree. Synthetic data both in terms of training CoT models and distilling CoT models into simpler task specific models.</p></li></ul><p>What a time to be alive! 2025 will be a big year for AI and the builders around it. Thanks for reading.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.lance.tech/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading lance.tech! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>