<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Entropy]]></title><description><![CDATA[Entropy]]></description><link>https://www.moreentropy.com</link><image><url>https://substackcdn.com/image/fetch/$s_!qjq3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ac09fe3-3e30-4a16-ab41-86c636ad8216_1024x1024.png</url><title>Entropy</title><link>https://www.moreentropy.com</link></image><generator>Substack</generator><lastBuildDate>Thu, 16 Apr 2026 20:18:07 GMT</lastBuildDate><atom:link href="https://www.moreentropy.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Fraser]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[moreentropy@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[moreentropy@substack.com]]></itunes:email><itunes:name><![CDATA[Fraser]]></itunes:name></itunes:owner><itunes:author><![CDATA[Fraser]]></itunes:author><googleplay:owner><![CDATA[moreentropy@substack.com]]></googleplay:owner><googleplay:email><![CDATA[moreentropy@substack.com]]></googleplay:email><googleplay:author><![CDATA[Fraser]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Generalist - Scaling AI in Robotics ]]></title><description><![CDATA[Making general-purpose robots a reality.]]></description><link>https://www.moreentropy.com/p/generalist-scaling-ai-in-robotics</link><guid isPermaLink="false">https://www.moreentropy.com/p/generalist-scaling-ai-in-robotics</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Wed, 05 Nov 2025 19:46:16 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0f465777-17ed-489d-9c4e-0b99dae045ad_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>We have seen what happens when the scaling hypothesis is applied to language. In 2019, AI could complete a sentence. By 2020 it was finishing a paragraph. Today, it&#8217;s working in complex codebases for hours and <a href="https://scottaaronson.blog/?p=9183">helping prove theorems</a> in quantum complexity theory.</p><p>We are seeing what happens when the scaling hypothesis is applied to biology. First, AI could generate functional proteins that were of little use. Now they&#8217;re <a href="https://www.nature.com/articles/s41586-025-09298-z">generating functional CRISPR systems</a> completely from scratch. Soon, AI will make biology programmable.</p><p>The same trajectory is unfolding in robotics. It continues to amaze me that the transformer, when paired with enough data, works across domains. As Ilya said, &#8220;These models, they just want to learn&#8221;.</p><p>In late 2022 and early 2023 there were a series of influential robotics papers published by Google. <a href="https://research.google/blog/palm-e-an-embodied-multimodal-language-model/">PaLM-E</a> was impossible to ignore. The paper showed that with the right data a generalist model beats a specialist model. It suggested that the transformer would take us to a generalized robotics foundation model.</p><p><a href="https://www.linkedin.com/in/peteflorence">Pete Florence</a> and <a href="https://www.linkedin.com/in/andyzenglinked">Andy Zeng</a> were co-authors of PaLM-E and together with <a href="https://www.linkedin.com/in/andrew-barry-aa88219">Andy Barry</a>, Pete&#8217;s PhD classmate at MIT, are the co-founders of <a href="https://generalistai.com/">Generalist</a>, a startup advancing the frontier of AI in robotics.</p><p>Yesterday, <a href="https://generalistai.com/blog/nov-04-2025-GEN-0">Generalist shared GEN-0</a>. It is the highest-dexterity model in the world, the beginning of an era where capabilities predictably scale with data, and the first step toward realizing their mission of making general purpose-robots a reality. Along with the model they are sharing the first scaling laws in robotics, a new model architecture, and news of over 270,00 hours of real-world manipulation data from 1,000s of homes, warehouses, and workplaces.</p><p>Pete, Andy, and Andy have built a team that feels equivalent in talent and raw ambition to early OpenAI. And their progress reflects this. At OpenAI I consistently had my hair blown back by the engineering challenges that were solved in service of unblocking scale. Generalist has been no different. The stories are already legendary, including building custom hardware and infrastructure &#8211; such as laying new dedicated internet lines &#8211; to handle the scale of the data volume they&#8217;re collecting for their models.</p><p>On a personal note, it&#8217;s a joy to be reunited with my friend Evan Morikawa, part of the founding team at Generalist and my partner in building the product development org at OpenAI.</p><p>I am beyond thrilled to be involved with Andy, Andy, Pete, and the entire team at Generalist. As part of Spark&#8217;s investment I have joined the board. Generalist is well on their way to enabling a future where we all benefit from robots in our homes and work.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.moreentropy.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Miracle Is Behind Us: Thinking about Eras of AI to Understand What's Coming]]></title><description><![CDATA[People still don&#8217;t appreciate the miracle that is next token prediction and the autoregressive transformer.]]></description><link>https://www.moreentropy.com/p/the-miracle-is-behind-us-thinking</link><guid isPermaLink="false">https://www.moreentropy.com/p/the-miracle-is-behind-us-thinking</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Tue, 01 Apr 2025 16:17:15 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/16e5b9f4-7929-4cc0-b287-47a6b2b57fb6_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>People still don&#8217;t appreciate the miracle that is next token prediction and the autoregressive transformer.</p><p>We are starting to grasp the impact of &#8220;generative ai&#8221; when applied to language and code. In the next couple of years, language models will do hours of independent work on our behalf and build complex software applications from scratch. Immense value is being created as the AI&#8217;s capabilities improve dramatically, year after year.</p><p>Everything we&#8217;re seeing with language is playing out in other fields and the future is both obvious and profound. Within a decade, AI will design molecules that cure many diseases and robots will drive the marginal cost of labor for many tasks to zero.&nbsp;</p><p>Hyperbole to some, obvious to others. The most amazing thing is that the breakthrough to enable this has already happened &#8211; next token prediction and the autoregressive transformer. The miracle is behind us. The future is foreseeable. We just need the data.<br></p><p><em><strong>&#8220;These models, they just want to learn&#8221; - Ilya Sutskever</strong></em></p><p>We&#8217;ve seen that if you take a sequence of text &#8211; &#8220;The cow jumped over the&#8221; &#8211; that an AI model can learn what comes next. Do this with a massive amount of data and compute and the capabilities of the AI become superhuman.&nbsp;</p><p>Similarly with code. Take a sequence &#8211; &#8220;for i in range(5):&#8221; &#8211; and an AI model can learn what comes next. The AI&#8217;s capabilities improve as the amount of training data and compute increase.</p><p>It intuitively makes sense. To the AI, &#8220;the cow jumped&#8221; is no different from &#8220;for i in range(5)&#8221;, in the sense that both are sequences that have meaning and the model learns the underlying meaning and rules of the &#8220;language&#8221; as the amount of training data increases.</p><p>Both biology and robotics can be expressed as simple sequences &#8211; `M-D-K-K-Y-S` and `"joint_positions": [0.2, 1.1, -0.5, 0.8, 1.2, -0.3]`, respectively. The next couple of years are predictable: the models, they just want to learn. And with enough data, we will see revolutions in biology and robotics. In fact, we already are. We just don&#8217;t grasp the impact yet.<br></p><p><em><strong>Eras: Understanding the trajectory of AI</strong></em></p><p>It&#8217;s helpful to think about <em>eras</em> of AI, and the transition from toy-like models to AI that is useful and, eventually, AI that is valuable.</p><p>In 2018-2020, early language models like GPT-2 could complete a sentence of text, but it was often nonsensical. The models were toy-like. They were not <em>useful</em> and certainty weren&#8217;t <em>valuable</em>. But to those who understood what had been done, the toy models pointed to an exciting future.</p><p>As the amount of training data and compute increased, and the models entered the second era, early versions of GPT-3 became <em>useful</em> in constrained use-cases, such as writing short ad copy. It took entering the third era, with GPT-3.5 and GPT-4, before models started to become <em>valuable</em>. And the commercial value has continued to increase with model advances like Sonnet 3.7 and o3.</p><p>The models are now completing tasks that take hours of time for highly trained knowledge workers to do. As new models arrive, and we enter new eras, the commercial value delivered by these models grows significantly. At launch, ChatGPT was priced at $20 / mo. Today, access to the best models costs $200 / mo. OpenAI is discussing prices of $20,000 / mo for future models.&nbsp;</p><p>Each new era has expanded the capabilities of what the models can do, expanded the use-cases where the models are useful, and expanded how valuable the models are.</p><p>In 2019, looking at the capabilities of GPT-2, seeing it fail to coherently finish a single sentence, it was hard for many to see what was coming. But the models, they just want to learn.&nbsp;</p><p>Code has followed a similar trajectory. From generating short code snippets, rife with errors, to developing entire applications from scratch in a few short years. Had you looked at the early coding models it would have been hard to foresee the explosive growth in revenue that we&#8217;re seeing from products like Lovable or Replit. But the pattern with code has been identical to language: from toy to <em>useful</em> to <em>valuable</em>, era by era as the models get larger.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TItu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TItu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png 424w, https://substackcdn.com/image/fetch/$s_!TItu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png 848w, https://substackcdn.com/image/fetch/$s_!TItu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png 1272w, https://substackcdn.com/image/fetch/$s_!TItu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TItu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png" width="1456" height="673" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:673,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TItu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png 424w, https://substackcdn.com/image/fetch/$s_!TItu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png 848w, https://substackcdn.com/image/fetch/$s_!TItu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png 1272w, https://substackcdn.com/image/fetch/$s_!TItu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F49753cde-0f1b-4b0f-9691-63647978abec_1600x740.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We&#8217;re two eras into the same trajectory in biology. Early transformer-based models, trained on a small amount of data, showed that a model could generate sequences for simple proteins that were sometimes functionally correct. Neither useful nor valuable and easily dismissible. Toy-like and no different from early GPT-2 demos. Scale up the model, enter the second era, and the model can <a href="https://www.nytimes.com/2024/04/22/technology/generative-ai-gene-editing-crispr.html">generate a protein that&#8217;s functionally identical to a molecule worth $60b</a>. Certainly usable, perhaps not yet valuable. But given it&#8217;s the same underlying mechanic, next token prediction and an autoregressive transformer, what do you think is going to happen as the model scales and enters the 3rd and 4th eras? What about the 7th and 8th era? We are on a path where models are learning the underlying language of nature. Biology is becoming programmable. Era after era, the models are becoming more useful and more valuable.</p><p>When you read Google&#8217;s papers from 2023, <a href="https://research.google/blog/palm-e-an-embodied-multimodal-language-model/">PALM-E</a> and <a href="https://robotics-transformer2.github.io/assets/rt2.pdf">RT-2</a>, where they show that an autoregressive transformer, trained on next token prediction, can control a robot, it may be easy to miss what&#8217;s coming. But when more recent work, from <a href="https://www.physicalintelligence.company/blog/pi0">Physical Intelligence</a> and <a href="https://storage.googleapis.com/deepmind-media/gemini-robotics/gemini_robotics_report.pdf">DeepMind</a>, show that generalization and capability improve as data scales, it&#8217;s hard to ignore that we&#8217;re walking down the same path. Yes, it&#8217;s a toy today. Yes, it&#8217;s not useful today. Yes, it&#8217;s far from valuable today. But we&#8217;re in the first era with robotics and it&#8217;s working the same way it worked for language and code. The trajectory will be no different.</p><p>The miracle is behind us. It&#8217;s all going to work. We just need the data.</p><p></p>]]></content:encoded></item><item><title><![CDATA[Elicit - Automating Research]]></title><description><![CDATA[Automating the ability to understand what is known is just the first step towards Elicit&#8217;s mission to radically increase good reasoning in the world.]]></description><link>https://www.moreentropy.com/p/elicit-automating-research</link><guid isPermaLink="false">https://www.moreentropy.com/p/elicit-automating-research</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Wed, 26 Feb 2025 14:40:15 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/bd50ca1c-5b55-4eab-99e5-93244894ddf7_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Einstein said the important thing is to never stop questioning, a truth that fuels scientific discovery, the heartbeat of progress. The potential for AI to positively shape the future is perhaps greatest in accelerating research itself &#8211; enhancing our ability to ask better questions, find relevant information faster, and draw connections that might otherwise remain hidden.</p><p><a href="https://elicit.com">Elicit</a> is building an AI Research Assistant that is used by over 400,000 researchers each month to accelerate their research and automate time-consuming tasks like summarizing papers, extracting data, and synthesizing findings.&nbsp;</p><p>Today, Elicit is introducing new features that use AI to automatically generate <a href="https://blog.elicit.com/systematic-review/">Systematic Reviews</a> and <a href="https://www.google.com/url?q=http%3A%2F%2Fblog.elicit.com%2Felicit-reports%2F&amp;sa=D&amp;ust=1740984900000000&amp;usg=AOvVaw2R6SPlqHLLCwKpanot6qpL">Research Reports</a>, delivering in minutes what has traditionally taken months of manual work. For a given research question, Elicit searches for potentially relevant papers from a corpus of 125 million, writes its own screening criteria and filters the studies based on the inclusion criteria, extracts key metrics and data from the filtered set of papers, and then synthesizes the findings into a report. All automatically via AI.&nbsp;</p><p>Automating the ability to understand what is known is just the first step towards Elicit&#8217;s mission to radically increase good reasoning in the world. The plan is to help automate discovery and the scientific method while generalizing scientific rigor to many more domains.</p><p>The future envisioned by Elicit's co-founders, <a href="https://x.com/stuhlmueller/">Andreas Stuhlm&#252;ller</a> and <a href="https://x.com/jungofthewon">Jungwon Byun</a>, sees Elicit being used by medical researchers to quickly synthesize findings across thousands of studies to identify promising treatments. By policy makers who rapidly assess the evidence base for interventions. By scientists who can navigate vast literature landscapes to discover overlooked connections and accelerate breakthroughs. By removing the friction from the research process, and guiding humans to better decisions, Elicit transforms how knowledge accumulates and flows, turning what was once a painstaking process of discovery into something that unfolds in real time, with positive implications for human progress.</p><p>To bring this vision to reality, Andreas and Jungwon have built a talented team across engineering, machine learning, and product design. We were impressed by the quality of everything Elicit does - from their product to their job descriptions, the team ships quality, quickly.&nbsp;</p><p><a href="https://www.sparkcapital.com">Spark</a> is proud to support Andreas, Jungwon, and the team at Elicit as they work to automate research and build the infrastructure for good reasoning at scale. Together with <a href="https://www.footwork.vc">Footwork</a> we are co-leading a $22M investment in Elicit joined by <a href="https://www.fiftyyears.com">Fifty Years</a>, <a href="https://www.basisset.com">Basis Set</a>, and <a href="https://www.mythos.vc">Mythos</a>.&nbsp;</p><p>With this funding Elicit will expand beyond academic research, bringing systematic, evidence-based AI decision-making to organizations across healthcare, policy-making, scientific research, and many other domains. In a world increasingly shaped by artificial intelligence, Elicit&#8217;s approach &#8211; using AI to help radically improve reasoning &#8211; represents the kind of thoughtful innovation we need. We&#8217;re excited to join Elicit on this journey.</p><p></p>]]></content:encoded></item><item><title><![CDATA[Hallway Chat]]></title><description><![CDATA[A decade ago or so Nabeel and Bijan started recording informal conversations about startups, technology, and products.]]></description><link>https://www.moreentropy.com/p/hallway-chat</link><guid isPermaLink="false">https://www.moreentropy.com/p/hallway-chat</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Fri, 31 May 2024 16:07:18 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d1d4e4b6-60b7-4989-92de-f862f93ae43a_800x800.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A decade ago or so <a href="https://x.com/nabeel">Nabeel</a> and <a href="https://x.com/bijan">Bijan</a> started recording informal conversations about startups, technology, and products. It was authentic and unpolished, their passion and personalities a joy to listen to.</p><p>A few months ago Nabeel and I <a href="https://hallwaychat.co">revised Hallway Chat</a>. Our goal is simple - to share some of our discussions as we try to make sense of this moment and explore products and technologies. </p><p>Here are some of my favorite episodes:</p><p><a href="https://hallwaychat.co/episodes/launch-stories-of-chatgpt">Launch stories of ChatGPT</a> where I shared some stories on the one year anniversary.</p><p>Exploring how <a href="https://hallwaychat.co/episodes/voice-note-apps-the-three-layers-of-ai-coding-a-wishlist-of-products">AI will impact software development</a>.</p><p>Thinking through AI models and when there may be <a href="https://hallwaychat.co/episodes/ai-models-weighing-commodity-from-state-of-the-art">differentiation vs commodification</a>.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.moreentropy.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Profluent - Scaling AI in Biology]]></title><description><![CDATA[Scale the transformer in language, learn the blueprint of language. Scale the transformer in biology, learn the blueprint of nature.]]></description><link>https://www.moreentropy.com/p/profluent-scaling-ai-in-biology</link><guid isPermaLink="false">https://www.moreentropy.com/p/profluent-scaling-ai-in-biology</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Thu, 16 May 2024 15:12:22 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b713c6aa-616a-4ee6-be69-2f5d057c6ccc_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In 2019 I sent a rather cold email to Ilya Sutskever. In hindsight it&#8217;s rather humorous &#8212; I thought he&#8217;d be interested in reading Rich Sutton&#8217;s <em><a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">The Bitter Lesson</a> </em>along with my thoughts on how dramatically I believed the world was going to change as people scaled AI. He responded warmly asking to meet at Stable Cafe.</p><p>This was part of a two week stretch that is among the most electric of my career. I would read research papers late into the evening only stopping to eagerly tell my wife about the elegance of masked language modeling and what it pointed to. I tinkered with fine-tuning BERT, joyously taunting the best ML Engineer I knew that he was going to be unemployed soon (he&#8217;s now at OpenAI).</p><p>We have seen what happens when the scaling hypothesis is applied to language. </p><p>Last summer I had the joy of experiencing this feeling again when I met Ali Madani and learned what&#8217;s happening as the scaling hypothesis is pursued in biology. </p><p>In January 2023 Ali and his team <a href="https://www.nature.com/articles/s41587-022-01618-2">published research</a> detailing for the first time how large AI models trained on biological data can generate complete proteins that don&#8217;t exist in nature yet function as well as proteins that evolved over millions of years. This model was equivalent in size to GPT-2, a model that was released in 2019. </p><p>Large AI models in biology exhibit similar traits to what has been observed in language models &#8212; the base model can be fine-tuned to improve performance in specific domains and new capabilities emerge with scale.</p><p>Last month the <a href="https://www.nytimes.com/2024/04/22/technology/generative-ai-gene-editing-crispr.html">New York Times covered a recent release from Ali&#8217;s company, Profluent</a>. Profluent has trained an LLM on a massive amount of biological data and have an AI model that can design gene editors that do not exist in nature. Profluent has now shown that a gene editor completely designed from scratch by AI can successfully edit human DNA. And they are open sourcing an AI-generated gene editor for free use. This is truly remarkable for many reasons, but particularly for the future it points to as the scaling hypothesis plays out in biology.</p><p>I am so delighted to be involved with Ali and Profluent as they advance the frontier of AI research in biology and work to make biology programmable. You can <a href="https://medium.com/spark-capital/welcome-profluent-773117378ffa">read more about our investment on Spark&#8217;s site</a>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.moreentropy.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The case for open AI]]></title><description><![CDATA[The memo that leaked from Google (which is real) makes the case for an open future for AI.]]></description><link>https://www.moreentropy.com/p/the-case-for-open-ai</link><guid isPermaLink="false">https://www.moreentropy.com/p/the-case-for-open-ai</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Fri, 05 May 2023 17:24:43 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/174aa3f4-c35b-4f6c-bbe3-d71a80576d0c_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The memo that <a href="https://www.semianalysis.com/p/google-we-have-no-moat-and-neither">leaked from Google</a> (which is real) makes the case for an open future for AI. I have increasing conviction that aspects of this future will unfold but I disagree with the conclusion. I believe we are heading to a world where there&#8217;s an oligopoly for the largest, most capable models accompanied by a thriving open ecosystem of smaller models.</p><p>Let&#8217;s get this out of the way upfront: I&#8217;m an OpenAI shareholder; I&#8217;m now actively investing into the future I foresee as a VC at Spark. I&#8217;m biased, have conflict etc etc.&nbsp;</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.moreentropy.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Join 1,750 others who subscribe to Entropy</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3><strong>AGI or Bust</strong></h3><p>I don&#8217;t see how there&#8217;s anything but a small number of groups providing the largest, most capable models. At this point it seems clear that it will be OpenAI and Anthropic and then tbd on whether anyone else will join them.&nbsp;</p><p>The window to recreate the dynamics that have led to these groups emerging is closing. These are technical and research problems on the scale of history&#8217;s largest projects, requiring the capital requirements of such projects, while being executed at the velocity of a ruthless startup.&nbsp;</p><p>Even in terms of &#8220;recreating the dynamics&#8221; it&#8217;s worth remembering that OpenAI and Anthropic are essentially the same, evolving via a <a href="https://openai.com/blog/organizational-update">mitosis-like event</a> (what a scene this moment will make when the book is eventually written). Which is to say, despite an oligopoly forming, there remains zero evidence that <em>anyone</em> is able to recreate the required dynamics. Further, the talent required to train models of this nature is becoming scarcer, as the flow of talent accelerates toward a small number of orgs.</p><p>On some cadence - tick, tick, tick - OpenAI and other AGI or Bust companies will release the world&#8217;s largest models, capabilities will step forward, and the world will gasp as AI accomplishes previously impossible feats.&nbsp;</p><p>There&#8217;s no chance open source will move in lock-step, let alone surpass these private labs when it comes to training the largest models. But open source will still play a critical role in the future of AI.</p><h3>Open AI</h3><p>Today, the state of open source AI is rather dismal, in the sense that while performance on benchmarks against comparable models is good, performance in production settings is not. This will all get fixed, and relatively quickly. Here&#8217;s why I&#8217;m optimistic.</p><p><strong>Open source efforts are just getting started and accelerating quickly.</strong></p><p>Llama was released just two months ago. Pipelines, tools, data sets, etc are all evolving quickly. Performance looks good on academic benchmarks because these models originated from research-motivated researchers and evals for production aren&#8217;t really a thing (yet). But product-motivated researchers and developers will continue to iteratively move the ecosystem toward solutions that work well in production.</p><p><strong>The world is the R&amp;D lab for open source.</strong></p><p>Talent may be flowing toward a few private labs, but they will only ever have a sliver of the total talent available across the world. The permissionless innovation that open source enables means that professors at Berkeley, alongside pseudo-anonymous students in Bangladesh, are all advancing the state of the ecosystem. There will be an explosion of experiments, creating toolsets and datasets, but also integrations with the weights themselves, that tailor open models for very specific use-cases.&nbsp;</p><p><strong>Many product use-cases will have a ceiling in terms of the AI capabilities they require.</strong></p><p>As GPT-5, -6, -7, &#8230; arrive, new use-cases will be unlocked by the advancing capabilities. But with every generational step forward, some use-cases will no longer be able to absorb the increase in capability and other measures will dictate which solution is optimal.&nbsp;</p><p>We&#8217;re already seeing this. While GPT-4 performs <a href="https://www.reddit.com/r/MBA/comments/11rg9ra/gpt4_passes_basically_every_exam_and_doesnt_just/">breathtakingly well on standard exams</a>, it wouldn&#8217;t make sense to use this model in production at scale for a simple classification task. Replit is showing us that the same may already be true for more complex tasks such as code completion. Sometimes it will make sense to pay up (in every sense) to use the most general, most capable model. Many times it won&#8217;t, as your needs will be best met by a narrow model that&#8217;s tailored to your specific use-case.</p><p><strong>AI-enabled products will provide the optimal user experience by going &#8216;full stack&#8217;.</strong></p><p>History has shown us that delivering the best user experience matters. And the best user experience is the result of every product detail being considered and constructed with intention. In AI-enabled products, this means that control, and the ability to make opinionated decisions across all aspects of the model, will be important for delivering the optimal product experience. As the open ecosystem evolves, it will increasingly be easy for product teams to go &#8220;full stack&#8221; with their AI model to deliver the best experience.</p><p>For example, the nature of the training data - such as form, style, length, etc - impacts what the AI generates and what the AI generates is increasingly a large part of the user experience. Product teams will start to care deeply about the training data because it will have a material impact on the user experience.</p><h3><strong>The likely future</strong></h3><p>We&#8217;re heading toward a world where a small number of players have a defensible oligopoly on the most capable, most general models. Product teams will use these models when required. Adjacent to this will be a vibrant, open ecosystem of smaller models, tailored and customizable to specific product needs.&nbsp;</p><p>There&#8217;s a likely future where the aggregate volume from open models is orders of magnitude larger than the private ones. What&#8217;s unclear is how much of the value will accrue to the open ecosystem.</p><p>_</p><h4>Misc</h4><ol><li><p>I <a href="https://medium.com/spark-capital/welcoming-fraser-to-spark-b8c12d1b6785">joined Spark Capital</a> and am leading investments in seed and series a startups. I love entrepreneurship and technology. I love that a small team can ship and change the world. As a former founder, I&#8217;m excited to support founders building remarkable products.</p></li><li><p>I&#8217;m helping <a href="https://twitter.com/yoheinakajima">Yohei</a> with BabyAGI, an <a href="https://spark-capital.notion.site/BabyAGI-4d5ad0edffc24caba6396d1774f79ca3">open source effort to build AI agents.</a> I believe in a future where AI assistants autonomously support us across a broad and diverse set of tasks and I want to help support the open ecosystem working to realize this future.</p></li><li><p>We&#8217;re hiring a <a href="https://www.sparkcapital.com/ai-hacker-in-residence">hacker in residence at Spark</a>. Come build with me.</p></li></ol>]]></content:encoded></item><item><title><![CDATA[Novelty vs utility, science vs belief, and where we are with AI-first products]]></title><description><![CDATA[Entropy #3]]></description><link>https://www.moreentropy.com/p/novelty-vs-utility-science-vs-belief</link><guid isPermaLink="false">https://www.moreentropy.com/p/novelty-vs-utility-science-vs-belief</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Wed, 01 Feb 2023 19:02:15 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/68a62868-61db-44e1-9993-7125ceffc99b_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://proofinprogress.com/posts/2023-02-01/the-ai-crowd-is-mad.html">The AI Crowd is Mad</a> by Tim Daubensch&#252;tz is worth a read. I agree that the LLM discussion could benefit from more nuance. I agree that costs need to come down (but of course they will). I find the contemplated future for AI odd. And, in terms of startups and products, I want to use the post to frame where I believe we are with modern AI and what we should expect.</p><h4>Novelty vs Utility.</h4><p>Tim argues that we&#8217;re collectively experiencing survivorship bias: seeing the algorithmically promoted highlights of ChatGPT on social media that are &#8220;truly stunning&#8221; is not giving us a reasonable sense of performance. </p><p>I agree. Further, much of what we&#8217;re seeing is novelty rather than utility. But even if the highlight reel is predominantly novelty, it is indeed stunning:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://twitter.com/tqbf/status/1598513757805858820" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6AKu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png 424w, https://substackcdn.com/image/fetch/$s_!6AKu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png 848w, https://substackcdn.com/image/fetch/$s_!6AKu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png 1272w, https://substackcdn.com/image/fetch/$s_!6AKu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6AKu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png" width="1180" height="1254" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1254,&quot;width&quot;:1180,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1197717,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://twitter.com/tqbf/status/1598513757805858820&quot;,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6AKu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png 424w, https://substackcdn.com/image/fetch/$s_!6AKu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png 848w, https://substackcdn.com/image/fetch/$s_!6AKu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png 1272w, https://substackcdn.com/image/fetch/$s_!6AKu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5188dca7-aec1-49c0-8449-ea82aeddc8ea_1180x1254.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>However, scroll through a twitter search for what ChatGPT is doing with code, or appreciate what <a href="https://twitter.com/intercom/status/1620425558546063360">Intercom did with ChatGPT in less than 8 weeks</a>, and it&#8217;s clear that there&#8217;s meaningful value adjacent to the fun.</p><p>The degree to which the highlight reel is stunning and the mix of novelty:utility matters, but what matters more is the rate of change for both and how much further the technology can improve.</p><h4>Belief vs Science.</h4><p>In November 2019 OpenAI released the full version of GPT-2. I remember playing with <a href="https://transformer.huggingface.co/doc/gpt2-large">Write With Transformer</a>, an accessible UI for interacting with GPT-2, and using Colab to fine-tune the model to <a href="https://avc.com/2019/12/aivc/">create toy-like experiences</a>. It was pure novelty and not very compelling. But I was stunned. The model was completing sentences that - sort of? - made sense. It didn&#8217;t take an overly active imagination to see what could come next.</p><p>Tim&#8217;s <a href="https://proofinprogress.com/posts/2023-02-01/the-ai-crowd-is-mad.html">post</a> highlights this tweet from Sam, taking issue with Sam&#8217;s use of &#8216;trust&#8217;, and stating &#8220;it is clear to me that these opinions are not anchored in reproducible observations&#8221;.</p><div class="twitter-embed" data-attrs="{&quot;url&quot;:&quot;https://twitter.com/sama/status/1599111626191294464?ref_src=twsrc%5Etfw%7Ctwcamp%5Etweetembed%7Ctwterm%5E1599111626191294464%7Ctwgr%5E7202de96603710defd1f0c04c591ee50602f21d1%7Ctwcon%5Es1_&amp;ref_url=https%3A%2F%2Fproofinprogress.com%2Fposts%2F2023-02-01%2Fthe-ai-crowd-is-mad.html&quot;,&quot;full_text&quot;:&quot;interesting to me how many of the ChatGPT takes are either \&quot;this is AGI\&quot; (obviously not close, lol) or \&quot;this approach can't really go that much further\&quot;.\n\ntrust the exponential. flat looking backwards, vertical looking forwards.&quot;,&quot;username&quot;:&quot;sama&quot;,&quot;name&quot;:&quot;Sam Altman&quot;,&quot;profile_image_url&quot;:&quot;&quot;,&quot;date&quot;:&quot;Sat Dec 03 18:41:36 +0000 2022&quot;,&quot;photos&quot;:[],&quot;quoted_tweet&quot;:{},&quot;reply_count&quot;:0,&quot;retweet_count&quot;:796,&quot;like_count&quot;:9022,&quot;impression_count&quot;:0,&quot;expanded_url&quot;:{},&quot;video_url&quot;:null,&quot;belowTheFold&quot;:true}" data-component-name="Twitter2ToDOM"></div><p>One may look at the progress over the past 3 years, from GPT-2 to ChatGPT, and decide that they need more evidence before they&#8217;ll extrapolate. Fortunately, that evidence exists in the form of scaling laws (initially published by <a href="https://arxiv.org/abs/2001.08361v1">OpenAI in 2020</a>, revised by <a href="https://www.lesswrong.com/posts/6Fpvch8RR29qLEWNH/chinchilla-s-wild-implications">DeepMind in early 2022</a>). Much like other technology trends, things are going to get better in forecastable ways for a while.</p><h4>Where We Are and What&#8217;s to Come</h4><p>Tim references Carlota Perez&#8217;s framework for technological revolutions and financial capital, which is useful as a way to think about our place in time and what&#8217;s to come.</p><p>I believe we&#8217;re in the midst of the irruption phase and that the frenzy is still to come. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!otHa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!otHa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png 424w, https://substackcdn.com/image/fetch/$s_!otHa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png 848w, https://substackcdn.com/image/fetch/$s_!otHa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png 1272w, https://substackcdn.com/image/fetch/$s_!otHa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!otHa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png" width="1456" height="912" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:912,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2184958,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!otHa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png 424w, https://substackcdn.com/image/fetch/$s_!otHa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png 848w, https://substackcdn.com/image/fetch/$s_!otHa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png 1272w, https://substackcdn.com/image/fetch/$s_!otHa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1620fe24-8c14-49cf-812b-37c26ec85b38_1922x1204.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Apparently half of the current <a href="https://twitter.com/gokulr/status/1620577528867074048">YC class is building with ChatGPT</a>, which is exactly what we should expect at this point (as Replit&#8217;s <a href="https://twitter.com/amasad/status/1620656939263995912?s=20&amp;t=FMFxYm6C67vyhshfIdE0YA">Amjad rightly argues</a>). We&#8217;re still figuring out what&#8217;s possible with this technology and developing the tools and infrastructure that make it easier to build with it. All while the underlying technology is rapidly advancing. Which is to say, it&#8217;s still early.</p><p>AI-enabled products will dramatically change the world, but not as soon as some think. While most applications built in the short-term will be inane, some will become transformative companies. AI is already good enough to deliver meaningful utility with certain product experiences. And while the degree of novelty remains high today&#8230; trust the exponential.</p><div><hr></div><p><a href="https://stevebick.bandcamp.com/track/cimylm">Soundtrack while writing</a>. </p><p>Follow me on Twitter: @<a href="https://twitter.com/fraser">fraser</a></p><p>Join 1,400 other readers of Entropy and follow along as we try to make sense of modern AI and AI-enabled products:</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.moreentropy.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.moreentropy.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Startups and the technique behind ChatGPT]]></title><description><![CDATA[Entropy #2]]></description><link>https://www.moreentropy.com/p/startups-and-the-technique-behind</link><guid isPermaLink="false">https://www.moreentropy.com/p/startups-and-the-technique-behind</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Thu, 26 Jan 2023 19:05:21 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/68926a4b-4054-4e35-a718-cafa9a3f2a5a_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hello to all 1,151 of you and welcome to the 166 who signed up in the past two days! If you haven&#8217;t yet, subscribe to follow my efforts to tell the story of modern AI:</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.moreentropy.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.moreentropy.com/subscribe?"><span>Subscribe now</span></a></p><p>&#127925;<a href="https://soundcloud.com/bigbootiemix/bb22">Soundtrack while writing</a></p><div><hr></div><p>Tomorrow is the anniversary of OpenAI&#8217;s &#8220;<em><a href="https://openai.com/blog/instruction-following/">Aligning Language Models to Follow Instructions</a></em>&#8221;, a seminal <a href="https://arxiv.org/abs/2203.02155">paper that I co-authored</a> (just kidding &#8211; I had no reason to be listed as an author but <a href="https://twitter.com/ryan_t_lowe">Ryan</a> is as kind as he is intelligent).</p><p>The technique pioneered by Ryan, <a href="https://twitter.com/longouyang">Long</a>, and team was used to <a href="https://openai.com/blog/chatgpt/">create ChatGPT</a> and understanding what was done is useful in appreciating the opportunity for startups building differentiated product experiences with AI.</p><p><strong>What was done</strong></p><p>First, they basically did GPT Cosplay. Humans pretended to be the AI and wrote &#8220;good&#8221; replies to different ways people may want to interact with the model. For example, a human would be shown a news article and asked to summarize it. That person would read the article and write a summary. All of these examples were then used to fine-tune a model.</p><p>They then had people look at multiple AI-generated replies for a given query and rate them from best to worst. <em>&#8220;Here&#8217;s one way AI could respond, and another, and another, &#8230; which do you like best?&#8221; </em>These ratings were used to train a second model, which predicts how &#8220;good&#8221; the AI-generated content is.</p><p>One model generates content and, almost like a coach, the other modeI evaluates how well it performed (<em>&#8220;you did pretty good on this one!&#8221;)</em>. This feedback is then used to improve the model that generates content.</p><p><strong>Results</strong></p><p>The proof is in the pudding with ChatGPT. But the results from the original paper are profound: a model trained this way outperforms a model that&#8217;s 100X larger.&nbsp;</p><p><strong>Why this matters</strong></p><p>The amount of data used to achieve the results in the paper was relatively small. They had people write ~10,000 &#8220;good&#8221; responses and make ~30,000 ratings. And since the data was spread across a range of use-cases &#8211; from copywriting to Q&amp;A, summarization to classification and others &#8211; there was an even smaller amount of data for any given use-case. This technique is obtainable for startups.</p><p>A product is the result of many opinionated decisions that need to be considered together to create the optimal user experience. This is one area where the difference between research and product is felt. What does a &#8220;good&#8221; response mean, absent from the product experience it&#8217;s enabling? Certainly a good summary for an app targeted at medical professionals would differ from what&#8217;s best for an app targeted at readers of the news. And the best summary for either would account for the product&#8217;s UI. Startups that care deeply about the user experience will care deeply about the data used to fine-tune their model.</p><p>Scale is <a href="https://scale.com/blog/chatgpt-reinforcement-learning">already offering</a> the infrastructure and human contractors to do this. Inevitably there will be robust open source tools for these pipelines. A small amount of data, distributed across a range of use-cases, led to a 100X improvement. As startups embrace this technique, in a focused and opinionated way, we&#8217;re going to see transformative products emerge.</p><div><hr></div><p>Follow me on Twitter: <a href="https://twitter.com/fraser">@fraser</a></p><p>If you&#8217;re building a startup with AI, email me!</p>]]></content:encoded></item><item><title><![CDATA[Testing by Vibes & Control Over the UX]]></title><description><![CDATA[Entropy #1]]></description><link>https://www.moreentropy.com/p/testing-by-vibes-and-control-over</link><guid isPermaLink="false">https://www.moreentropy.com/p/testing-by-vibes-and-control-over</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Tue, 24 Jan 2023 19:40:28 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5ac09fe3-3e30-4a16-ab41-86c636ad8216_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.moreentropy.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.moreentropy.com/subscribe?"><span>Subscribe now</span></a></p><p>Hello to the 985 of you who have signed up! As mentioned, Entropy is an experiment and I&#8217;ll be iterating based on what I like and what I hear from you all. Please do share feedback.</p><p>&#127925;<a href="https://soundcloud.com/altego_music/unholy-sam-smith-ft-kim-petras-altego-remix">Soundtrack while writing</a></p><div><hr></div><h2><strong>Testing by Vibes</strong></h2><p>Even more than AI safety, the state of benchmarking for AI-enabled products feels akin to the practice of bloodletting: something we&#8217;ll look back on and wonder what we were collectively thinking.</p><p>Here&#8217;s a passage <a href="https://youtu.be/pFsj9V6-1NE?t=1782">from a discussion</a> last week between W&amp;B&#8217;s CEO <a href="https://twitter.com/l2k">Lukas Biewald</a> and Replit&#8217;s CEO <a href="https://twitter.com/amasad">Amjad Masad</a>, moderated by Replit&#8217;s <a href="https://twitter.com/ornelladotcom">Ornella Altunyan</a>. It&#8217;s lengthy, but it&#8217;s worth the read as it touches on another important topic &#8211; the role of open source in AI-enabled products.</p><blockquote><p><strong>Ornella</strong>: Something that is so interesting about this boom that we&#8217;re having right now is that open source is keeping pace with the large research universities and Google&#8217;s of the world. Why do you think that is?</p><p><strong>Lukas</strong>: ML, unlike other software development, has this incredibly strong academic lineage where open source is the default. So much of the ML is coming directly from research and everyone in research is opening up their stuff&#8230; This is a fantastic thing. It makes everything available to everyone in the world.</p><p><strong>Amjad</strong>: I agree with that. This generation of models lends itself really nicely to open source. They're essentially stateless programs, so you give them input and you get an output. They're very portable, they can run on any machine. Especially diffusion models are very small, so that really lends itself nicely to open source. I think it's been hard to replicate the really large models. Some of them are doing well on benchmarks but then when you try to use them they're actually not very good.&nbsp;</p><p><strong>Lukas</strong>: Which ones are you talking about?</p><p><strong>Amjad</strong>: All of them! GLM, Bloom. All of them. I respect everything they&#8217;ve done but they&#8217;re really not that great. Maybe they&#8217;re still iterating on them. Maybe here the data advantage starts to show at very large parameters. That being said, I&#8217;m pretty sure they&#8217;ll catch up pretty soon.&nbsp;</p><p><strong>Lukas</strong>: I also think there's a lot to make a model between one that runs well on a benchmark versus one that does something actually useful for a real human. That's not a trivial undertaking.</p><p>&#8230;</p><p><strong>Ornell</strong>: Speaking of open source, I was wondering if Amjad could walk us through the approach to leveraging open source to build tools like Ghostwriter.</p><p><strong>Amjad</strong>: When we started working on it, it wasn't really clear that there was going to be this big open source movement that would help startups like ours build Copilot-like experiences. One of the main reasons we wanted to self-host it, and not just rely on an API, is because we really want to control the latency and really want to control the quality, want to be able to improve it over time, want to be able to collect data and feedback, and iterate on it.&nbsp;</p><p>We felt like for some use cases, very large models or chat-type use cases, it's fine to use an API. But for something this close to the user &#8230; it was very important for us to own the experience end-to-end.&nbsp;</p><p>At first we literally tried every model out there. And we found Codegen from Salesforce. When you&#8217;re evaluating some technologies they&#8217;re never perfect for your use-case but you see a glimpse, &#8220;ok I can see this working&#8221;. Then it was like two weeks of manic work trying to make it faster. &#8230; and it started feeling like something we could use. Eventually we did some fine-tuning and figured out how to improve the model on the fly based on user feedback. Now we&#8217;re training a model from scratch. But that&#8217;s after we shipped. We shipped a product, people love it, we&#8217;re collecting a lot more data, we know what&#8217;s missing, we know what features we want, and we have our first model trained from scratch, and we&#8217;ll have something to announce in a month or two.</p><p><em>[discussion on the sweet spot model size for Replit&#8217;s product experience, balancing capabilities and latency]</em></p><p><strong>Amjad</strong>: The tricky thing is again what we talked about with benchmarking is that you could do well on benchmarks and then go try in the real world and it's actually made the model performance worse. We had one case, for example, where it was doing better on python after some fine-tuning but then it forgot how to write jsx. We call it testing by vibes.</p><p>Lukas: [laughing rather hard] Oh no! That doesn&#8217;t sound like a best practice, Amjad.</p><p>Amjad: We kind of check the vibes of the model and if it passes the vibe check it goes into an a/b test to check the acceptance rate. If it inches upward, we&#8217;re doing something well. If it&#8217;s neural, maybe we didn&#8217;t do something useful. Would love to get more objective about it but we haven&#8217;t found a way, other than building up more and more benchmarks over time.</p></blockquote><p>From a benchmarking perspective, there are a number of interesting things here. It&#8217;s rather wild to consider that state of the art is &#8220;testing by vibes&#8221;, yet this is true. You sort of poke around to get a ~feeling~ and once you&#8217;ve spelunked enough you yolo into production.</p><p>Another observation is that academic benchmarks don&#8217;t adequately measure performance for a given product experience. And even within a single product experience, performance of a new model can vary. With one fine-tuned model, Replit saw performance improve in one area, python, but diminish in another, jsx.&nbsp;</p><p>Finally, an important aspect highlighted by Amjad is the idea that capability is only one measure of performance that matters when building a product. Latency and cost matter too and the optimal capability/latency tradeoff is specific to the product experience being delivered.</p><p>All of this is to say that evaluating a model&#8217;s performance today is a nuanced, labour-intensive, error-prone process. The incentive to solve this problem is high because evaluating performance is critical to controlling the product experience.</p><p></p><h2><strong>Control Over the Product Experience</strong></h2><p><em>&#8220;It was very important for us to own the experience end-to-end.&#8221;</em> - Amjad</p><p>To craft the user experience exactly as they wanted and to iterate and improve the ux over time, Replit self-hosted an open source model (amusingly, an open source model from Salesforce). They shipped this product, gathered feedback, insights, and, importantly, data. And now they&#8217;ve trained their first model from scratch. While the degree of complexity has undoubtedly increased, the degrees of freedom that they have in controlling the product experience has too. To paraphrase Alan Kay, perhaps people who are really serious about product should make their own models. The open source world is increasingly enabling this for those who want to take on the cost and complexity.</p><div><hr></div><h3><strong>Building in the open</strong></h3><p>I was happily surprised by the response to my <a href="https://twitter.com/Fraser/status/1617552406178299904?s=20&amp;t=-4WP-FN5jHYdF5MTG2nEBQ">Tweets yesterday</a>. As of now, there have been 347,000 views, 985 subscribers on Substack, and 192 people who have expressed interest in <a href="https://www.moreentropy.com/about">investing with me</a> in founders who are building with AI. My plan is to tell the story of modern AI and those who are building ambitious products with it. By doing so, I hope to better understand where the world is going and to help create a better future. Thanks for your support!</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.moreentropy.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Entropy! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Coming soon]]></title><description><![CDATA[This is Entropy.]]></description><link>https://www.moreentropy.com/p/coming-soon</link><guid isPermaLink="false">https://www.moreentropy.com/p/coming-soon</guid><dc:creator><![CDATA[Fraser]]></dc:creator><pubDate>Mon, 23 Jan 2023 13:36:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!qjq3!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ac09fe3-3e30-4a16-ab41-86c636ad8216_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is Entropy.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.moreentropy.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.moreentropy.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item></channel></rss>