<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Richards Tu's Blog]]></title><description><![CDATA[some personal thoughts :)]]></description><link>https://blog.richardstu.com</link><image><url>https://substackcdn.com/image/fetch/$s_!XZx4!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F127b7f9a-94f6-4e5a-ba83-158a035eef31_896x896.png</url><title>Richards Tu&apos;s Blog</title><link>https://blog.richardstu.com</link></image><generator>Substack</generator><lastBuildDate>Wed, 22 Apr 2026 08:36:57 GMT</lastBuildDate><atom:link href="https://blog.richardstu.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Richards Tu]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[richardstu19999@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[richardstu19999@substack.com]]></itunes:email><itunes:name><![CDATA[Richards Tu]]></itunes:name></itunes:owner><itunes:author><![CDATA[Richards Tu]]></itunes:author><googleplay:owner><![CDATA[richardstu19999@substack.com]]></googleplay:owner><googleplay:email><![CDATA[richardstu19999@substack.com]]></googleplay:email><googleplay:author><![CDATA[Richards Tu]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[2026 and beyond]]></title><description><![CDATA[I did a podcast episode in mid-January talking about what I think is coming for agents and AI in 2026 and beyond.]]></description><link>https://blog.richardstu.com/p/2026-and-beyond</link><guid isPermaLink="false">https://blog.richardstu.com/p/2026-and-beyond</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Sun, 01 Feb 2026 16:02:23 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6c544501-1214-4f22-9aaa-e7426a687c59_5056x3392.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I did a podcast episode in mid-January talking about what I think is coming for agents and AI in 2026 and beyond. This blog is basically the written version of that, covering what I discussed there, plus a few things I didn&#8217;t get to mention. So if you&#8217;ve already listened, consider this as some warp up and extended cut; if you haven&#8217;t, please go and check that out~</p><p>This is my attempt at opening up the next few years - what I expect to see, mainly around agents. I&#8217;ve done some reasoning on what will keep happening this year and beyond, and I also took a sneak peek into the farther future. I know I&#8217;ve done this kind of thing several times before, but I think it&#8217;s still worth redoing since AI is evolving so fast. Things look pretty different from even a year ago.</p><h2><strong>2026</strong></h2><p>For 2026, I think one of the main topics will still be agents, but they will be more personalized and truly useful agents.</p><h3><strong>Proactive agent</strong></h3><p>The whole journey has evolved like this: first we had chatbots, only used for conversations. Then people wanted them to access external data, so we got basic tool use - searching the web, finding live data. As models became more powerful, they could handle more tools and become more robust. Now, we gave them reasoning capabilities, and they became agents.</p><p>Expectations are growing. People want agents to do more personalized things, to be more useful, to know more about them. But the limitation is clear: current agents only do tasks explicitly requested by humans. They lack self-initiating ability. We say agents will help us take actions and save time, but the reality is that current agents, whether text-based, GUI-based, or combined, are slow. After you send a request like &#8220;shop some groceries for me&#8221;, you need to wait a while. That&#8217;s counter-intuitive to how we want them to save us from tedious daily work.</p><p>To improve this, agents should really be able to do tasks on its own; in another words, they have to understand how we use them, and then use that understanding to do tasks in the background without waiting for humans to initiate and intervene. They need to prepare what we want in advance. This makes me think of them as a more advanced form of autocomplete. The only difference is the scale of the task. Ordinary autocomplete, like Cursor&#8217;s Tab mode, handles lines of code across files, while task-based agents, like Manus, handle entire tasks.</p><p>To get there, they need to learn your usage patterns. For example: the agent knows you always ask it to summarize your email on Monday morning, so it automatically does that in the future; or it knows when you&#8217;re running low on groceries and handles that for you.</p><p>But we also need to make sure they&#8217;re not annoying, which means timing also matters - it shouldn&#8217;t be too intrusive or too hidden. Otherwise it&#8217;s useless. This means the UI and UX need to change. We can&#8217;t just have input-box-only interfaces anymore. Gmail&#8217;s AI Inbox from last week is a good example: it didn&#8217;t fundamentally change how you interact with Gmail, but it added AI features that actually boost productivity. AI-powered tools don&#8217;t necessarily need an obvious input box; they should be bound to the task context itself.</p><p>If this develops well, it will significantly boost people&#8217;s productivity with agents, and people will start believing in them more.</p><h3><strong>Memory</strong></h3><p>The second key piece is memory. People are giving models higher expectations, and models need to know users better to feel genuinely useful. This ties back directly to what I mentioned about proactive agents.</p><p>Currently there are a few general solutions for memory. Looking at it from a product perspective, there are basically three types:</p><ol><li><p>Model uses a tool to store things into a memory space (ChatGPT, Gemini, Claude, Kimi, Qwen, etc.)</p></li><li><p>Model uses a conversation search tool to find specific topics from past chats (Claude, ChatGPT)</p></li><li><p>System summarizes user interactions daily, then extracts new information into detailed summarized memory (Claude)</p></li></ol><p>These are pretty good, and I&#8217;ve seen promising performance from products like Claude with their memory system. But for broader general agents, it can, and should, be better. Memory isn&#8217;t just limited to basic information about us, it&#8217;s connected to our general preferences across life: shopping style, coding style, travel style, and a lot more. These affect how well an agent can complete tasks within your expectations. But it&#8217;s annoying to repeatedly mention or re-state preferences. So how products &#8220;form&#8221; these memories needs innovation too.</p><p>One approach I&#8217;ve been thinking about: hand off your apps and websites to an agent to explore first. An agent will always be better at learning your preferences than asking you to describe them, and you definitely don&#8217;t want to repeat yourself over and over; so you give the agent login access, it browses and checks though your previous orders, learns and summarizes your preferences, i.e. what do you usually order for groceries? which airline do you always prefer? The agent then summarizes these into specialized documentation. Each time it goes to that specific app or website, the relevant instructions load automatically, ensuring the model already knows what it needs to know. This doesn&#8217;t require any special model capability - it just needs the product or environment (&#8221;agent harness&#8221;) to be optimized to push what the model knows further.</p><p>The same approach would apply to many other use cases. And unlike recent skills or similar features, this doesn&#8217;t require user or model to pay extra attention. There won&#8217;t be cases where the model ignores specific preferences, because they&#8217;re loaded by default.</p><p>The above are all product-based, but we could also think from a more foundational side. Sometimes models don&#8217;t realize the importance of using user knowledge, so they just skip it. (I should note: my ideas here may be wrong since there&#8217;s no clear experimentation showing these work yet.)</p><ol><li><p>We could rely on SAEs (sparse autoencoders) from mechanistic interpretability. Anthropic has used these in some of their research. Generally, SAEs can find activated feature points inside a model when it&#8217;s generating specific tokens. If we could use this technique to detect a model&#8217;s tendency to seek external knowledge, including user memories, then when that tendency is high, we could auto-inject relevant knowledge after that token. The model receives it and generates more useful responses.</p></li><li><p>We could use fewer, more specialized experts in MoE models. For example, a model with only three or four experts, each for a specific action: one for thinking/reasoning, another for tool use, and a final one for responding. Maybe one more for orchestrating which expert to use at each step.</p></li></ol><p>There could be more innovation in memory on the model side.</p><p>Either way, we&#8217;re going to see a lot of surprises around proactive agents and memory. The key question now is how models can really boost productivity - because I think that&#8217;s where they can bring the most economic impact before they can truly impact society at large, they should first have huge, noticeable impact on individuals.</p><h2><strong>Trends</strong></h2><p>Also, I think there would be some continuing trends that will keep happening or starting to shift in next year or two.</p><h3><strong>Model as a product</strong></h3><p>The first continuing trend is <strong>model as a product</strong>. This has been a long-running pattern; and I think it has two slightly different side:</p><ol><li><p>Model having unique abilities that can directly become a new product or feature (like GPT-Image, Nano Banana, Sora 2, Genie 3, etc.)</p></li><li><p>Model having is quite strong that people can build geneneral products around it with some engineering work (like early Manus on Claude-3.7 Sonnet, Claude Code, etc.)</p></li></ol><p>Among these, I think the new <a href="https://deepmind.google/models/genie/">Genie 3</a>, Google&#8217;s latest world model, we got in public last week has huge potential. You can create the world, control how you &#8220;walk around&#8221; inside it - the whole thing is customizable. It&#8217;s going to be much more fun than video models like Sora. And since it can generate interactive worlds, it has potential to become one of the first reliable generative games. I didn&#8217;t play many games before, but if we get solid products based on robust world models, I might start - creating my own experiences sounds really fun lol. Some examples I saw on X:</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;143deca2-2399-4456-8378-7eb5f80df400&quot;,&quot;duration&quot;:null}"></div><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;0ea5eafd-86de-4703-bb51-8179ceece890&quot;,&quot;duration&quot;:null}"></div><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;63017a72-4a9f-44f2-bb14-1c3c8ee0cac5&quot;,&quot;duration&quot;:null}"></div><h3><strong>Agent capacity</strong></h3><p>The second trend is <strong>agent capacity</strong>. Models will become more robust, that&#8217;s the clear trajectory. They&#8217;ll handle more long-tail tasks, able to do even complex tasks. They would help humans more in accelerating not just SWE work but also AI research itself; maybe even automate some of it. We&#8217;ve already seen huge potential here, both in scientific research and other areas. And we have benchmarks tracking this, like <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">METR Time Horizo</a>, <a href="https://andonlabs.com/evals/vending-bench-2">VendingBench</a>, and many more.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WGxe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WGxe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png 424w, https://substackcdn.com/image/fetch/$s_!WGxe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png 848w, https://substackcdn.com/image/fetch/$s_!WGxe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png 1272w, https://substackcdn.com/image/fetch/$s_!WGxe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WGxe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png" width="1456" height="767" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:767,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:237845,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.richardstu.com/i/186475470?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WGxe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png 424w, https://substackcdn.com/image/fetch/$s_!WGxe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png 848w, https://substackcdn.com/image/fetch/$s_!WGxe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png 1272w, https://substackcdn.com/image/fetch/$s_!WGxe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F177efa96-4474-42ba-85b6-b47d1018f6b5_2730x1438.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YZN5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YZN5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png 424w, https://substackcdn.com/image/fetch/$s_!YZN5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png 848w, https://substackcdn.com/image/fetch/$s_!YZN5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png 1272w, https://substackcdn.com/image/fetch/$s_!YZN5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YZN5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png" width="1456" height="1031" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/be9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1031,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:409122,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.richardstu.com/i/186475470?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YZN5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png 424w, https://substackcdn.com/image/fetch/$s_!YZN5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png 848w, https://substackcdn.com/image/fetch/$s_!YZN5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png 1272w, https://substackcdn.com/image/fetch/$s_!YZN5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe9b7863-abc2-4433-8b26-0e2455462f5d_3328x2356.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The curve is going up and will continue pretty steadily.</p><h3><strong>Model alignment</strong></h3><p>The third, and one of the most important, is <strong>model alignment</strong>. As models become more capable and people put them into more production environments, the consequences of bad intentions become catastrophic. If a model is able to help scientist in building nuclear fusion reactor, then it can help bad people build nuclear weapons; if a model is able to help companies develop madicine, then it can create bio-weapons as well; knowledge are connected anyway. I had wrote about my thoughts on this before, and there&#8217;re a lot of researches on this, but one approach I find promising is the new <a href="https://www.anthropic.com/constitution">Claude Constitution</a>. <a href="https://model-spec.openai.com/2025-12-18.html">OpenAI&#8217;s Model Spec</a> is similar, but more rule-based: what you should do, what you shouldn&#8217;t. The Constitution is more about teaching the model how to be good and do good things - less like rules, more like parents teaching a child (I remember Dario described it as a letter &#8220;from a deceased parent sealed until adulthood&#8221;). I think this is a promising direction, and I expect more companies to explore this kind of approach.</p><h3><strong>Human-AI interaction</strong></h3><p>The last thing is how <strong>human-AI interaction</strong> will change. Right now, we interact with AI through apps, APIs, websites - all limited to mobile and desktop. I think a really good new portal is AI glasses, because they can see what you see, hear what you hear. And they can have their own ecosystem position - they don&#8217;t need to replace phones or anything else. They can add something new: a different way to interact and co-live with AI; unlike Humane AI Pin or Rabbit r1, which tried to replace the phone and couldn&#8217;t.</p><p>Since AI glasses can sense almost everything we can, they&#8217;d be a great add-on for the proactive agents I mentioned. They could recommend things or help complete tasks based on your real-world environment. Better memory systems become important here too.</p><p>We&#8217;re already seeing some products. For example, <a href="https://pickle.com/1">Pickle 1</a> looks kinda promising - I&#8217;ve already pre-ordered, waiting to see how it goes. And it seems Google is working on something too, mentioned by Demis at Davos 2026. But these are still early.</p><p>We can focus on the previous parts for now; the glasses stuff is more a matter of hardware, software, and ecosystem catching up.</p><h2><strong>Future</strong></h2><p>I&#8217;ve written about the future many times before, but AI moves so fast that things look quite different from even a year ago. So I think it&#8217;s still worth sharing what I expect for the farther future - I have some new thoughts after seeing recent posts, interviews, and doing my own thinking.</p><p>Before I go further, I should mention Dario&#8217;s new essay, <em><a href="https://www.darioamodei.com/essay/the-adolescence-of-technology">The Adolescence of Technology</a></em>. It&#8217;s a serious piece that maps out the risks we&#8217;re facing and how we might address them. I have a lot of respect for how he approaches these problems - careful, concrete, not as a doomer. If you haven&#8217;t read it, I&#8217;d recommend it. What I&#8217;m writing here is more of a personal take, from someone who will live through this transition; and again, it&#8217;s all my personal thoughts, so can be incorrect.</p><h3><strong>What I want to see</strong></h3><p>The world on the other side: one where survival anxiety isn&#8217;t the default mode of human life. Where scientific progress in medicine, climate, and longevity happens way faster than before. Where people can pursue what actually matters to them, not just what pays bills.</p><p>Dario calls this <a href="https://www.darioamodei.com/essay/machines-of-loving-grace">&#8220;Machines of Loving Grace&#8221;</a>. I think he&#8217;s right about what&#8217;s possible. The real question is whether we can get through the middle part without everything falling apart.</p><p>I&#8217;ve imagined this good future a lot. Robots handling physical labor. Abundance making material scarcity less relevant. People free from the constant pressure of &#8220;making a living&#8221; and able to actually live. It sounds utopian, but I don&#8217;t think it&#8217;s impossible - just hard to get to, and need a lot of efforts.</p><h3><strong>Some hard questions</strong></h3><p><strong>If AI creates more value than you, what&#8217;s your purpose?</strong></p><p>This will be the lived experience of a lot of people soon. Companies will do the math: AI is faster, cheaper, better. The rational move is to let people go. If that happens at scale, the whole &#8220;AI benefits humanity&#8221; thing falls apart. You can&#8217;t really benefit from something that made you economically irrelevant and gave you nothing back.</p><p>I think to prevent this, companies and society need some kind of common understanding: even if AI creates more value, we should still preserve humans in the foreseeable future. After a company takes what they need for operations, they should return that value to the workers who got replaced, which should be more like a social contract. The value came from somewhere.</p><p>This is really hard to execute. No enforcement mechanism, no clear policy, and competitive pressure pushes against it. But that&#8217;s why the journey is difficult. The tech is arriving faster than our social systems can adapt; that&#8217;s why I said, it should be us to adapt the development of theses advanced systems. We&#8217;ve almost never seen these together, and our existing frameworks aren&#8217;t built for it.</p><p><strong>Meaning without work</strong></p><p>Even if we solve the material side - even if displaced workers get income - there&#8217;s still the meaning problem. People don&#8217;t just want stuff. They want to matter, to be needed. Work used to provide that, even when the work itself was boring.</p><p>I&#8217;ve thought about this a lot, and there are a lot of discussions out there. In a world where AI handles most cognitive tasks, we&#8217;ll need new structures for purpose. Creative work, community, exploration, caregiving - things that matter to us even if they don&#8217;t maximize GDP. But this won&#8217;t happen automatically, we have to build it intentionally.</p><p>Maybe this sounds abstract, but it&#8217;s actually pretty concrete. What would you do if you didn&#8217;t have to work? Not vacation-mode &#8220;what would you do&#8221;, but actually, long-term, what would give your life structure and meaning? For me, I think it&#8217;s exploring unknowns, experiencing different places, maybe creating things. But a lot of people haven&#8217;t had the chance to even think about that question, because survival comes first.</p><p>The transition will force us to answer it, and I think the answer will be different for everyone, which is kinda the point. Freedom to figure out what matters to you, rather than having it dictated by economic necessity.</p><p><strong>The transition itself</strong></p><p>I think it&#8217;s quite obvious that this transition won&#8217;t be peaceful. I&#8217;ve said before that millions will lose jobs, and society might break down in parts. That&#8217;s what history told us, the industrial revolution caused massive suffering before things got better. This could be similar, but faster and broader.</p><p>The question is whether we can make the transition as humane as possible. Not &#8220;acceptable sacrifice for progress&#8221; - that framing has been used to justify a lot of harm historically. More like: we acknowledge it will be hard, and we try to take care of each other through it.</p><h3><strong>Why I&#8217;m still optimistic</strong></h3><p>I know the risks are huge. I&#8217;ve read a lot of doomer takes, and I get where they&#8217;re coming from. Powerful AI in the wrong hands, misaligned objectives, societal collapse, and a lot more; these are indeed real.</p><p>But there are a lot of researchers working on alignment and interpretability. Some companies (like Anthropic and DeepMind) actually taking safety and relevant problems seriously. The new Claude Constitution trying to teach models to be good, not just follow rules. People having these conversations instead of ignoring them. That matters.</p><p>I&#8217;ve thought about how to hold both things - the hopeful vision and knowing that getting there will be rough. Honestly, it comes down to something simple: I believe our world can be much better, and I want to see that happen. Maybe help build it. That&#8217;s the faith I keep.</p><p>The years ahead will be hard, maybe it will take us decades. But I keep coming back to: so what? why be afraid?</p><div><hr></div><p>There&#8217;s still more to write, but I think this is enough for now, I&#8217;ll save the rest for a future post.</p><p>Anyway, I wish the world getting better and better in 2026 and beyond.</p>]]></content:encoded></item><item><title><![CDATA[Think beyond current reasoning models]]></title><description><![CDATA[Published on September 3, 2025]]></description><link>https://blog.richardstu.com/p/think-beyond-current-reasoning-models</link><guid isPermaLink="false">https://blog.richardstu.com/p/think-beyond-current-reasoning-models</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Wed, 03 Sep 2025 23:05:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d10b458f-054e-4d4d-8436-b5508f94b54a_1344x896.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Updated on November 20, 2025</strong>: The latest <a href="https://deepmind.google/models/gemini-image/pro/">Nano Banana Pro</a> from Google DeepMind (which based on their latest Gemini-3 Pro model) actually has <a href="https://x.com/m__dehghani/status/1991529191981084716">the ability to think in image</a>, and this indeed brings huge leap forward regarding to image quality and other aspects. Although I have a little doubt that this is actually a dual model underhood, like Gemini-3 Pro prompting an image generation model during its reasoning process, but anyway it is still quite impressive. I&#8217;m even more excited to see how this could evolve further to other modalities in the future.</p><div><hr></div><p>For some time, I have been thinking about how we could push the frontier of current reasoning models forward, not just their performance, but also other &#8220;use-cases&#8221; or &#8220;features&#8221;. So I came up with these two questions, which haven&#8217;t been mentioned by others (maybe?):</p><ol><li><p>How could the current reasoning model paradigm affect or enhance multimodal models?</p></li><li><p>How could so-called &#8220;hybrid reasoning models&#8221; (possibly) work?</p></li></ol><p>And around these two topics, I have some personal thoughts. I&#8217;m not aiming to be definitively correct, but just rather to share my own opinions.</p><div><hr></div><h2>1. How could the current reasoning model paradigm affect or enhance multimodal models?</h2><p>People are shocked by how Gemini-2.5 Flash Image (aka Nano-Banana) performs: image generation, editing, and so much more. A lot of models today could do more than just output text, they could also create audio and image. So I&#8217;m thinking of whether we could integrate those multimodal models with the reasoning models. Instead of only thinking in text, could they also think in image and audio? If we could let a reasoning model to use multimodal content not just in its final response but also during its internal thinking, the results could be much better as well. This may sound confusing, so here&#8217;re a few examples:</p><ul><li><p><strong>audio modality</strong>: A user is preparing a speech and wants the model to generate an audio sample. gpt&#8209;realtime or gemini&#8209;2.5 may help, but the quality cannot be ensured. If the model can reason about what emotions to convey at specific points, produce draft audio to simulate, and iterate, the result quality may improve. For example:</p></li></ul><pre><code><code>user:
&#9;I&#8217;ll have a speech about the future of our planet earth next week at the UN. Could you please give me a sample for it?

model (CoT):
The user is asking: &#8220;...&#8221; So they want a sample speech. 
&#9;I could give a text script, but it will not be that helpful. I guess give them an example audio would be better. It may also require some emotions. I should first get up structures and contents first.
&#9;...
&#9;Alright, so I am done with initial composing, let me explore how emotions should be, and I will do some drafts, then it will good to go..
&#9;Well, I should have hope, urgency, inspiration, responsibility, solidarity, empathy, ...
&#9;OK, so for hope, I think I should use upraising tone, just like this:
&#9;[a piece of voice with upraising tone]
&#9;Cool, sounds good, let&#8217;s continue:
&#9;...
&#9;Alright, I think I got all emotions done. Let me make a first draft:
&#9;[full composed speech with correct emotions]
&#9;...

model (response):
&#9;Here&#8217;s the speech I prepared for you:
&#9;[full final speech audio]
</code></code></pre><ul><li><p><strong>image modality</strong>: Although the latest Nano&#8209;Banana can create or edit amazing images, it still struggles in complex scenarios. This makes sense because the model outputs the image in just a single shot. If it could take some time to think and draft first, like humans do, the results might be better. For example:</p></li></ul><pre><code><code>user:
&#9;Please create a cinematic image of a interstellar war. it needs to be epic.

model (CoT):
&#9;The user is asking me to generate a image of ...
&#9;So let me first imagine the scene: a massive fleet of sleek starships engaged in battle over a luminous nebula, laser beams and plasma torpedoes streaking across space, explosions creating bright fireballs, a giant battlecruiser in the foreground with detailed hull, background with distant stars and a planet partially visible ...
&#9;Cool, let me create a sketch first:
&#9;[a sketch of the scene with only few lines]
&#9;...
&#9;I see the planet in the image doesn&#8217;t seem real, let me change the position:
&#9;[a final-prepared version of the image with all elements and colors]
&#9;...
&#9;Great, it seems perfect, let me give this to the user.
&#9;
model (response):
&#9;Here&#8217;s the image you want:
&#9;[the final version of the image]
</code></code></pre><ul><li><p><strong>or even video modality</strong> <em>(I know there&#8217;s no language model that could output both text and video, but maybe it is a good idea)</em></p><p>Today&#8217;s video generation models, like Veo-3, are great as well; however, they all have some huge weaknesses: short generation, not able to comply with complex story, etc. So giving the model the ability to think in audio, image and more, could enhance it greatly. For example:</p></li></ul><pre><code><code>user:
&#9;please give me a short movie about space exploration.
&#9;
model (CoT):
&#9;The user asks for a short movie ...
&#9;I plan to write the script first, then draw some certain frames, create some suitable BGMs. This seems good. Let me proceed with the some ideas.
&#9;...
&#9;Phew. I think this is amazing, user is gonna love it. So next, frames.
&#9;For the opening, I think this would be good:
&#9;[a sketch frame for the first part of the story]
&#9;...
&#9;That&#8217;s great. I&#8217;m all done. Now, music.
&#9;[some background audios for specific frames]
&#9;...
&#9;Sounds good, then let me go for the final video part.
&#9;...
&#9;
model (response):
&#9;Here&#8217;s the movie you want:
&#9;[final movie]
</code></code></pre><h2>2. How could so-called &#8220;hybrid reasoning models&#8221; (possibly) work?</h2><p>So the definition for hybrid reasoning models is that the models which can either respond directly or think deeply before respond (or even able to decide when to think more on its own). And currently, there&#8217;re only a few models have such ability: Claude-3.7, Claude-4, DeepSeek-v3.1, Qwen-3, and some more (GPT-5 doesn&#8217;t count for now, because it has a router).</p><p>Claude is close-sourced, so we don&#8217;t know how their thinking toggle actually works (maybe similar to other models). For DeepSeek-v3.1 and Qwen-3&#8217;s non-thinking mode, they are just prefilled with a blank thinking block (like &lt;think&gt; &lt;/think&gt;). This is a good and quick way to let the model skip thinking and respond directly, but&#8230;. well, seems like the result may not be that satisfying for Qwen team, and they separated thinking and non-thinking mode into two models soon (<a href="https://x.com/Alibaba_Qwen/status/1947344511988076547">here</a>).</p><p>But, what if, we let the model always think, but for different reason? Here&#8217;s what I mean:</p><p>In basic sense, we could train the model to know how to react with different setting (thinking mode on, off, or auto). In inference time, we will let the model know (like via system prompt) what user pick. And in different mode, the model will always take a look at the current setting, and decide what it needs to do on its own. This means, there will be fewer manual intervention, the model just knows what to do.</p><p>Example behavior:</p><ul><li><p>Thinking mode on:</p></li></ul><pre><code><code>system: ... &lt;thinking_mode&gt;on&lt;/thinking_mode&gt;

user:
&#9;...
&#9;
model (CoT):
&#9;Let me see. I saw the thinking mode has been set with on, which means I should take more time to continue thinking afterwards.
&#9;The user asks ...

model (response):
&#9;...
</code></code></pre><ul><li><p>Thinking mode off:</p></li></ul><pre><code><code>system: ... &lt;thinking_mode&gt;off&lt;/thinking_mode&gt;

user:
&#9;...
&#9;
model (CoT):
&#9;Hmmm... I see the thinking mode is off. This means I should start responding directly. Yes, no more thinking. Start to respond now.
&#9;
model (response):
&#9;...
</code></code></pre><ul><li><p>Thinking mode with auto:</p></li></ul><pre><code><code>system: ... &lt;thinking_mode&gt;auto&lt;/thinking_mode&gt;

user:
&#9;What is 1+1?
&#9;
model (CoT):
&#9;I see an auto thinking mode is been put. So basically I just need to decide how much I need to think. Hmmm... Let me see.
&#9;The user asks: &#8220;What is 1+1?&#8221; This is trivial: 2. Any other things? No. Just a number is fine. Respond now.

model (response):
&#9;2.</code></code></pre><ul><li><p>or</p></li></ul><pre><code><code>system: ... &lt;thinking_mode&gt;auto&lt;/thinking_mode&gt;

user:
&#9;...
&#9;
model (CoT):
&#9;... Oh my god, this is hard. According to my setting, I guess I should take more time to think further for the message.
&#9;The user asks ...

model (response):
&#9;...
</code></code></pre><p>And we could also use RL to let the model learn such behaviors:</p><ol><li><p>For the general &#8220;react-to-setting&#8221; behavior, we could set a verifier that look for whether model really correctly react to model&#8217;s setting, because the model may usually use similar phrasing or meaning when reacting to the setting.</p></li><li><p>For whether model correctly react to the setting (like stop thinking when off, continue when on), we could reward/penalize the model with its afterward behavior, like if model still insists in thinking forward when the setting is set to off, penalty would be casted.</p></li><li><p>For the auto thinking, I think we could use a prelabelled dataset with &#8220;easy/hard&#8221; labels, then reward on the output that correctly deal with the question under auto setting.</p></li></ol><p>This may not really work, but may be a path that worth exploration. Why? Take a look at OpenAI&#8217;s o-series models and GPT-5-thinking model. Their reasoning effort are all been controlled by internal parameter called &#8220;juice&#8221; (GPT-5 even has a param called &#8220;oververbosity&#8221; which is for final response verbosity control). Also, Anthropic uses something like &lt;max_thinking_length&gt; to tell the model how long it should think in all. So via curated data and RL training, the model may also gain ability in thinking more adaptively and efficiently.</p>]]></content:encoded></item><item><title><![CDATA[Looking Ahead to 2025]]></title><description><![CDATA[Published on February 6, 2025]]></description><link>https://blog.richardstu.com/p/looking-ahead-to-2025</link><guid isPermaLink="false">https://blog.richardstu.com/p/looking-ahead-to-2025</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Thu, 06 Feb 2025 17:22:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5b1c5081-ec93-4a48-ada4-5fde4a8f1a21_1344x896.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Although it&#8217;s February now, I still think it would be quite nice to wrap up 2024 together with the beginning of 2025 and look forward into the new year.</p><p>The past thirteen months have been great, with a lot of good things happening. We&#8217;ve made huge progress across models&#8217; multimodal, reasoning, and agentic abilities, which are all important components on my own imaginary roadmap to capable AI systems that would have a huge impact on our species (or what people call &#8220;AGI&#8221;).</p><div><hr></div><h2>On multimodal model</h2><p>This is an interesting topic. The reason for its importance is that I strongly believe letting models &#8220;feel&#8221; the world in many different ways is an important key to helping them better understand physics, the world, and the whole universe &#8212; text does not include everything in &#8220;language&#8221;; &#8220;language&#8221; is diverse, it&#8217;s much richer than text.</p><p>Currently, I think the existing tokenizer is what limits the model from moving forward. In fact, there are many simple tasks that we, as humans, would definitely not get wrong; however, even the strongest LLM currently available (i.e., o1-pro) would easily get stuck. For example:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-WoE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-WoE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png 424w, https://substackcdn.com/image/fetch/$s_!-WoE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png 848w, https://substackcdn.com/image/fetch/$s_!-WoE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png 1272w, https://substackcdn.com/image/fetch/$s_!-WoE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-WoE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png" width="1456" height="1001" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1001,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:236541,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://richardstu19999.substack.com/i/175494597?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-WoE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png 424w, https://substackcdn.com/image/fetch/$s_!-WoE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png 848w, https://substackcdn.com/image/fetch/$s_!-WoE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png 1272w, https://substackcdn.com/image/fetch/$s_!-WoE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F285f60be-6650-4abb-8667-a54e4749ae94_1554x1068.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So I think we have to come up with a real multimodal model that could get rid of those limitations of the current visual encoders/tokenizers and really understand the image. This is a SUPER basic ability that models need to have.</p><p>Beyond multimodal input, we have multimodal output. This showed up in GPT-4V back in May and the new Gemini 2. The reason I think it is also pretty cool is that it&#8217;s better than letting a model write prompts and use DALL&#183;E or Midjourney to create images, because there are a lot of limitations with traditional text-to-image models. They sometimes get stuck with complex things, and they don&#8217;t understand what they are drawing. However, the models with true multimodal output can know what they need to generate, and humans can let them iterate on those results. What&#8217;s more, we could do more fun things with such ability, like:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xxy6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xxy6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png 424w, https://substackcdn.com/image/fetch/$s_!xxy6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png 848w, https://substackcdn.com/image/fetch/$s_!xxy6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png 1272w, https://substackcdn.com/image/fetch/$s_!xxy6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xxy6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png" width="1148" height="1434" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1434,&quot;width&quot;:1148,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:436543,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://richardstu19999.substack.com/i/175494597?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xxy6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png 424w, https://substackcdn.com/image/fetch/$s_!xxy6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png 848w, https://substackcdn.com/image/fetch/$s_!xxy6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png 1272w, https://substackcdn.com/image/fetch/$s_!xxy6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff4889e87-cb6a-453b-a6e5-6cb0440781ad_1148x1434.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Pretty cool, right? And since you can actually let the model generate or edit a picture for you, everyone can do PS work - they don&#8217;t need to actually have such expertise, which could be super convenient.</p><h2>On reasoning models</h2><p>This is the hottest topic from the last few months, and I&#8217;ve already written about it in last August. Up till now, we have several reasoning/thinking models in hand (o-series model, R1, Gemini Thinking models and a lot of others from research).</p><p>Thanks to RL, progress is really fast; for example, from o1 to o3 with roughly about 3 months, the model is able to solve a bunch of AGI-ARC tasks, and we could expect more crazy things in the coming months.</p><p>I think the idea of giving the model more time to respond is great. However, sometimes the model will have an overthinking problem, which is sometimes time- and compute-consuming. For example, when you ask R1 &#8220;1+1&#8221;, it will think for seconds (~100 tokens):</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hP7f!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hP7f!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png 424w, https://substackcdn.com/image/fetch/$s_!hP7f!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png 848w, https://substackcdn.com/image/fetch/$s_!hP7f!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png 1272w, https://substackcdn.com/image/fetch/$s_!hP7f!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hP7f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:576153,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://richardstu19999.substack.com/i/175494597?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hP7f!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png 424w, https://substackcdn.com/image/fetch/$s_!hP7f!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png 848w, https://substackcdn.com/image/fetch/$s_!hP7f!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png 1272w, https://substackcdn.com/image/fetch/$s_!hP7f!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb9aef4b-0f0c-4672-927b-01f941659949_1716x1332.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>So that&#8217;s why I would say the model being able to control when they need to think is also another important ability, which may be the next focus for researchers. But before that, we need to make general reasoning ability (beyond math and coding) better. BTW, in the blog I wrote last year, I mentioned system 1 and system 2 thinking patterns of humans. I still think that would be useful, though directly applying that into the current models would not be feasible; we could still borrow some ideas.</p><p>Besides, another interesting thing mentioned in R1&#8217;s paper is that the model sometimes uses mixed language in its thinking process. I think as we continually scale RL and test-time compute, we would even see models generate nonsense or scrambled text, while the result will not be affected at all. I think that would be a time when we say, &#8220;Okay, RL just works.&#8221; (But this would be a total disaster for Anthropic and some AI doomers lol :P)</p><h2>On agents</h2><p>Besides reasoning, I think this is another term that everyone addicts to use. I still remember that last year almost all products said they had some &#8220;AI agents&#8221; stuff (and I blacklist every product that says so).</p><p>There are only a few real agents in my mind, like Project Astra from DeepMind, Operator and Deep Research from OpenAI. These tools are really the AI systems that could take reasonable actions for you.</p><p>Here, the definition I gave is that only when you have a good reasoner or your model can reason well, then you can call the tool or the system built upon your model an agent. I think that should be something that we should expect, instead of those weird and fancy tools that you click, then summarize some emails or things like that.</p><p>Although the products that claim they have agents in the past year are not acceptable to me, their idea is somehow kind of good&#8212;what they need is a better base model, i.e., o3-mini: fast, cheap, and capable.</p><p>Another core feature that would really push these agents forward should be in-thinking tool use. When o1 with tool use was released, I was concerned about whether o1&#8217;s tool use was like thinking, calling a tool, then responding directly; but now with o3, my concerns have vanished. o3&#8217;s tool use flow is like thinking, tool-using, rethinking (maybe with another few turns), then responding. In fact, I saw great benefits from this pattern in o3-mini with web browsing and Deep Research powered by fine-tuned o3. And I expect to see more agents from OpenAI and from other research labs.</p><p>Phew~ that&#8217;s all I wanna say. January is just a starting point, and we&#8217;re gonna have a wild ride in upcoming months! Just buckle up for it.</p>]]></content:encoded></item><item><title><![CDATA[Envisioning Our Future with AI]]></title><description><![CDATA[Published on October 12, 2024]]></description><link>https://blog.richardstu.com/p/envisioning-our-future-with-ai</link><guid isPermaLink="false">https://blog.richardstu.com/p/envisioning-our-future-with-ai</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Sat, 12 Oct 2024 10:33:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/06470d6d-9283-40d9-b2dd-b34a310ac14f_1344x896.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<blockquote><p>This blog is inspired by Dario&#8217;s Reflecting on Dario&#8217;s &#8220;Machines of Loving Grace&#8221;</p></blockquote><div><hr></div><p><em><strong>Disclaimer:</strong> I&#8217;m not a professional in any of the domains mentioned below. My conclusions are drawn from my observations and my understanding of AI, so errors are inevitable.</em></p><p>I think the most inspiring term in the whole blog would be <strong>&#8220;country of geniuses in a datacenter&#8221;</strong>. I can&#8217;t think of any term better than this for describing the powerful AI.</p><p>So within my limited knowledge capacity, I would envision how our future with AI would be in two aspect:</p><ol><li><p><strong>Life with AI:</strong> Our life with AI, I think, would definitely be more convenient. Let&#8217;s first talk about just the models. Starting from this, we can see that a lot of things can be accompanied and assisted by AI. Because I think as AI becomes more capable, we should increasingly consider it as a companion, not just a tool. It&#8217;s like an assistant that helps us every day and knows what we know, sees what we see, hears what we hear, and helps us plan what we plan. Our way of interaction with machine is changing.</p><p>Another aspect is robotics, which would be another part making our lives more convenient. Just look at what Elon showed us at Tesla&#8217;s &#8220;We, Robot&#8221; event. I think it was a fantastic event, and many people posted videos about Tesla&#8217;s Optimus robot from the event. And I think it&#8217;s so powerful. It can answer questions, dance, and even give you drinks. I think it&#8217;s incredibly convenient, and as its price decreases, it will become generally available in everyone&#8217;s home, just like we have televisions. And our family robots could help us do many things, i.e. houseworks, which can help us save time for doing something more meaningful.</p></li><li><p><strong>Science with AI:</strong> Let&#8217;s move on to discuss how AI will change the current paradigm of scientific research. Looking back over the last century, we&#8217;ve seen numerous breakthroughs in fields such as biology, chemistry, physics, and a lot more. Notably, if you&#8217;ve been following the 2024 Nobel Prizes, you&#8217;ll have noticed that both the Physics and Chemistry awards were given for AI-related achievements. However, I believe this is just the beginning.</p><p>This trend indicate increasing number of people recognizing how AI will impact scientific research, and wider acceptance of applying AI into science. Inevitably, I think AI will have its own profound impact on science. As <a href="https://blog.richardstu.com/my-few-thoughts-on-ai-ethics">I&#8217;ve written before</a>, science with AI will be much more advanced than it was previously, and the innovation and research will continually speed up. A lot of new progresses will be made. The future of scientific research with AI integration will be fascinating, and I&#8217;m eagerly looking forward to it.</p><p>So in this way, I very much agree with what Dario said that we&#8217;re going to shrink the research progress of the next 50 to 100 years into 5 or 10 years. I think this assumption is pretty accurate; however, since I&#8217;m just a normal person without any insights about the progress inside those companies, I cannot comment much on this. But anyway, I&#8217;m still quite optimistic about it.</p></li></ol><p>I believe we will have a future where AI is seamlessly integrated into our daily lives and all other aspects of society. In this future, many global problems present today, like diseases and global warming, will possibly be addressed. The boundary between artificial and human intelligence will gradually blur, as there would be no discernible difference. Eventually, this will foster a symbiotic relationship that drives unprecedented advancements in science, technology, and social progress. As we navigate this new era together, our harmonious coexistence with AI will guide us along the path...</p><p><strong>Good time is coming, just STAY ALIVE.</strong></p>]]></content:encoded></item><item><title><![CDATA[My Few Thoughts on OpenAI's o1 family models]]></title><description><![CDATA[Published on September 15, 2024]]></description><link>https://blog.richardstu.com/p/my-few-thoughts-on-openais-o1-family</link><guid isPermaLink="false">https://blog.richardstu.com/p/my-few-thoughts-on-openais-o1-family</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Sun, 15 Sep 2024 06:30:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1f60335f-dfcc-4f2f-a285-f5729dcd5afb_1344x896.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Thinking Models Are Good Models</strong></p><p>This time, OpenAI&#8217;s <a href="https://openai.com/index/learning-to-reason-with-llms/">latest o1 family models</a> (o1, o1-preview, and o1-mini) are indeed very powerful, with remarkably impressive performance. I think the most noteworthy points are: 1. They possess extremely strong logical reasoning abilities; 2. The models come with built-in CoT (Chain of Thoughts), requiring minimal user prompting.</p><p>I consider these two points very important because in the past, when faced with complex mathematical problems, these language models were often just &#8220;guessing answers&#8221; rather than truly reasoning step by step. But this time it&#8217;s different. OpenAI has specifically integrated RL and CoT based on GPT-4o, as well as added special &#8220;Reasoning tokens&#8221;, making the model truly &#8220;think&#8221;.</p><p>For example, when I asked o1-preview and o1-mini to calculate 279563 multiplied by 356104, they were both able to &#8220;think&#8221; first, self-reflect and correct during the thinking process, and then give the answer. Upon verification, both were correct. In the past, this task would have yielded completely incorrect results from LLMs - they would either guess or make fatal logical errors between steps. The same improvement was evident when I gave them the most challenging math problems in 2024 Gaokao in China (check result <a href="https://photos.app.goo.gl/JE9UakvNAtdd5wsJ7">here</a>). Additionally, I tested them with several questions from the national competitions this year, and the results were also very good. So we can see that this iteration of models performs very strongly on reasoning-related tasks. Logically rigorous and self-consistent reasoning is a necessary condition for us to reach the next level of AI, namely Agent, after all, the latter needs to be able to act on behalf of humans. We can&#8217;t allow them to make any mistakes, otherwise, it could lead to catastrophic consequences. (Read more about Agent and Model Autonomy <a href="https://blog.richardstu.com/my-few-thoughts-on-agents">here</a>)</p><p>Moreover, we can notice the prompt &#8220;thought for x seconds&#8221;. For instance, o1-preview&#8217;s &#8220;thinking&#8221; time is relatively long, while mini&#8217;s &#8220;thinking&#8221; time is shorter (because the latter has been specifically fine-tuned on competitive math problems). I think the potential this brings is limitless. Now it&#8217;s &#8220;thinking&#8221; for a few seconds or minutes, in the future it will be &#8220;thinking&#8221; for months, making more complex reasoning and analysis, and obtaining more accurate and logical results.</p><p>These two phenomena remind me of System 1 and 2 that I mentioned in <a href="https://blog.richardstu.com/does-llm-really-have-reasoning-ability-repost-from-my-x">this article</a> I wrote about possible future improvements in model reasoning abilities. These two concepts were originally used to describe human brain thinking, but now I see that o1 also has this characteristic. By definition, System 1 is responsible for intuitive fast thinking, like 1+1=2; System 2 is responsible for complex thinking that requires reasoning, such as complex mathematical problems, etc. More precise thinking can yield better and deeper results. This reminds me of <a href="https://blog.richardstu.com/my-few-thoughts-on-ai-ethics">another article I wrote</a>, where I mentioned that if future models could reach the level of Nobel Prize winners, we could have hundreds of such AI copies form a research group and give them months to &#8220;think&#8221; and conduct research. Now it seems that models are already very strong in logical reasoning, biology, and other fields, so I believe the probability of this phenomenon occurring is very high. I&#8217;m looking forward to seeing AI assist humans in developing important drugs, discovering new materials, and even proving mathematical theorems in the future.</p><p>The future of humanity looks bright at the moment.</p>]]></content:encoded></item><item><title><![CDATA[My Few Thoughts on LLM's Reasoning Ability]]></title><description><![CDATA[Published on August 19, 2024]]></description><link>https://blog.richardstu.com/p/my-few-thoughts-on-llms-reasoning</link><guid isPermaLink="false">https://blog.richardstu.com/p/my-few-thoughts-on-llms-reasoning</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Mon, 19 Aug 2024 08:00:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4010e510-14eb-40a8-826e-769a13ba0265_1344x896.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>People are having debate on this topic on X these days. Some ppl say that LLM can definitely reason because it can help us do math and code on some extent; but some other guys argue that LLM can&#8217;t reason and they are not designed for it, what they do is just recite in training data.</p><p>Frankly speaking, I don&#8217;t think that the LLMs can&#8217;t reasoning; in fact, I think there can be three possible ways to help them get this ability:</p><ol><li><p><strong>Scaling is all we need:</strong> We could literally do nothing to the current LLM, what we need to do is just to continue scaling (compute, data, and model size), and just let the model learn and understand the underlying logic pattern and syntax inside the training data, as the model becoming more and more complex during scaling; then we just wait for the &#8220;miracles&#8221; to happen, but it won&#8217;t happened till the end of the possibly exponential scaling curve;</p></li><li><p><strong>Human&#8217;s thinking pattern is all we need:</strong> We could try to apply the System 1 and 2 thinking to the current LLM: System 1 thinks fast and intuitively, best for quick decision, similar to the current LLMs; while System 2 thinks slowly and deliberately, which can be the perfect system for LLMs to solve all those kind of complex and reasoning-required tasks.</p></li><li><p><strong>Tree search is all we need:</strong> We could implement the tree search into the current LLM, and we have seen deepseek-prover-v1.5 succeeded with MCTS and AlphaProof-2 succeeded with hybrid approach with tree search usage; both of these achieve great results, which means it is effective for model&#8217;s complex problem solving ability. (since I&#8217;m not quite familiar with this, I really recommend <a href="https://www.notion.so/AI-Search-The-Bitter-er-Lesson-44c11acd27294f4495c3de778cd09c8d?pvs=21">&#8220;The Bitter-er Lesson&#8221;</a> by <a href="https://x.com/aidan_mclau">Aidan Mclau</a>, it can be really helpful to get the importance and potential of tree search!)</p></li></ol>]]></content:encoded></item><item><title><![CDATA[My Few Thoughts on Agents and Model's Autonomous Behavior]]></title><description><![CDATA[Published on August 12, 2024]]></description><link>https://blog.richardstu.com/p/my-few-thoughts-on-agents-and-models</link><guid isPermaLink="false">https://blog.richardstu.com/p/my-few-thoughts-on-agents-and-models</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Mon, 12 Aug 2024 06:00:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/b149fc1a-2f56-467b-9a80-fcbc77b826ac_1344x896.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is definitely one of the hottest topic these days lol. And I personally think that except the agent itself, its autonomous behavior is also really intriguing, since they&#8217;re both related to how capable the base model is and also how dangerous the model can be.</p><h2>For agent stuff:</h2><p>We have seen a lot of products that claim to be agentic in recent months, but I think while some of them are really cool, most of them are just hyping around the market without any real value that can bring to consumers. To introduce briefly, an agent is an LLM-based system that can be used to act on human behalf and interact with the real physical world. For most of the time, the system need to take long series of actions to complete a complex task, for example, planning a trip, finding the best school for a kid, or even building a house, so that the error rate of the base LLM needs to be extremely low, since a single error can lead to potential disaster. However, the current models are still far from that, we can see several products that seem pretty powerful yet often fail in the middle of the process, and we still need some scaling to make the model be more capable in long sequences of actions and reasoning before we can consider it as an agent. When it comes to actually building these agents, there are a bunch of technical hurdles we need to overcome. Like, how do we make sure the agent can keep track of what it&#8217;s doing over a long period of time? The real world is messy and uncertain - how do we teach an AI to deal with that? And then there&#8217;s the whole can of worms that is getting these systems to play nice with all the different APIs and external systems out there. It&#8217;s not just about making the model bigger - we need to tackle these practical issues too. Another thing that&#8217;s been on my mind is the ethical side of all this. If we&#8217;ve got AI agents running around doing stuff for us, who&#8217;s responsible when things go wrong? And how do we make sure we can actually understand why the AI is making the decisions it&#8217;s making? We can&#8217;t just have a black box making important choices. Plus, we&#8217;ve seen how AI can pick up and amplify human biases - that could cause some real problems if we&#8217;re not careful. I think without solving or understanding these hurdles, we can&#8217;t build a good and reliable agent that can be droped in production.</p><h2>For model autonomous behavior:</h2><p>Although this topic is not that practical and touchable to most of us, I still think it&#8217;s kinda more interesting to talk about. First of all, I should point out that the autonomous behavior from an LLM or agent is dangerous. Why? Because it basically means that the model is doing things that are out of our expectations. It can hide valneralble mistakes in the process, and it can also huge potential harm to the real world. For example, let&#8217;s say we have another global outage in the future, similar to the recent CrowdStrike incident, but on a much larger scale. Then, we let a powerful agent find the bug and fix it. We give it 3 hours to do it. After it&#8217;s done, it just says &#8220;Okay, I&#8217;m done, everything is fixed!&#8221;, but in fact we don&#8217;t know what exactly it is doing; the agent may write another script that may cause another outage, all of which we don&#8217;t know, and these types of events are the dangers. This whole autonomous behavior thing gets even trickier when you start thinking about how we might control it. We need to come up with some serious safety measures and control mechanisms. Maybe we need some kind of AI oversight system, or hard limits on what actions an AI can take without human approval. But then you run into the problem of potentially limiting the AI&#8217;s effectiveness. It&#8217;s a real balancing act. And let&#8217;s not forget about the regulatory side of things. Right now, the laws around AI are pretty fuzzy, but you can bet that&#8217;s going to change as these systems get more advanced and more widely used. We might end up with some kind of AI licensing system, or mandatory safety tests. It&#8217;s going to be interesting to see how that works. So it&#8217;s clear that to build a helpful and effective agent, we not only need to make the model more powerful by scaling it up, but also need to asure the model isn&#8217;t gonna do bad or unexpected things, which enforce us to know the deep mechnism of the models and how they work, and which is the interpretability research doing by several labs. In this way, I think we&#8217;re going to see a lot more focus on human-AI collaboration rather than fully autonomous agents or something, at least in the short term. It&#8217;s a way to leverage the strengths of AI while keeping a human in the loop for safety and decision making process. But long-term? Nobody knows. The potential impact on society is huge, and it could go in a lot of different directions depending on how we handle the development and deployment of these technologies. And by the way, I recommend the readers to check a org called METR (which is the lab that is really good at model threat eval), they release an eval for model autonomous behavior: here.</p><p>Frankly speaking, while the idea of having AI agents to handle complex tasks for us seems really cool, we&#8217;ve got a long way to go before we can truly rely on them. It gonna take a lot of careful thought and development to get it right.</p>]]></content:encoded></item><item><title><![CDATA[My Few Thoughts on Compute Scaling]]></title><description><![CDATA[Published on August 8, 2024]]></description><link>https://blog.richardstu.com/p/my-few-thoughts-on-compute-scaling</link><guid isPermaLink="false">https://blog.richardstu.com/p/my-few-thoughts-on-compute-scaling</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Thu, 08 Aug 2024 12:00:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4eb2896a-3200-4e78-b836-eda603775eeb_1344x896.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>To scale, or not to scale?</em> It&#8217;s a really interesting topic. Scaling Laws is a famous law in AI and ML, and the people&#8217;s opinion&#8217;s on it is also diverse. So I wanna share some thoughts on compute scaling, which is a part of the Scaling Laws.</p><p>I believe there&#8217;s still significant scaling potential for LLM training in terms of computational power. This requires simultaneous efforts in both funding and resources, and the current trend looks promising. While we should be cautious about computational scaling turning into an arms race between companies, I think competition is exactly what we need right now. The key is to focus on training efficiency and avoid the paperclip effect. Otherwise, even if we pour all of humanity&#8217;s resources into it, we won&#8217;t see significant results. This could end up having a catastrophic impact on the global ecosystem, completely contradicting our vision of building an AGI system that benefits all of humanity. To be honest, even with high training efficiency, model training is incredibly expensive. A couple of months ago, Microsoft and OpenAI announced plans to invest $100 billion in building a massive computational center. In an interview last month, Anthropic CEO Dario mentioned that their current investments are sufficient for training the next generation of models, but he&#8217;s unsure about next year. He predicts that model training next year could potentially cost tens to hundreds of billions of dollars.</p><p>Some might argue that instead of focusing so much on scaling up computational power to improve model capabilities, we should research more efficient model architectures. However, I think you need to ensure short-term research results with enormous potential and feasibility. Otherwise, for the big players, you&#8217;re essentially gambling - and the reality is they can&#8217;t afford to gamble. Once you fall behind, it&#8217;s very difficult to catch up. Now it&#8217;s just a matter of seeing who will have the last laugh in this long run.</p>]]></content:encoded></item><item><title><![CDATA[My Few Thoughts on AI Ethics]]></title><description><![CDATA[Published on July 10, 2024]]></description><link>https://blog.richardstu.com/p/my-few-thoughts-on-ai-ethics</link><guid isPermaLink="false">https://blog.richardstu.com/p/my-few-thoughts-on-ai-ethics</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Sun, 07 Jul 2024 12:00:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/5b750b11-68f1-493a-afae-cde07692686d_1344x896.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It&#8217;s so clear that we are on the fast lane of AI development; and as I previously wrote, AI safety is becoming a huge concern for a lot of people. But I think we should also focus more on AI Ethics. Obviously, we will face or are facing the situation that the pace of development of those frontier AI system greatly surpasses the one of adaptation of human society. However, the problem is: Should human society adapt to the development of AI, or should the development of AI adapt to the human society?</p><p><strong>I think it should be us to adapt the development of the AI system.</strong></p><p>Currently, one of the greatest concern is about unemployment. My opinion is that general deployment of advanced models would make a lot of people lose their jobs, for example, those who do repetitive text-based work (transcriptionist, proofreader, etc.). This would cause great negative impact to our society. But we can try to slow down the speed of unemployment rate by bring more job opportunities. I suppose in the near future, we would need more people who are adept in guiding the AI systems to do something that they&#8217;re not so good at, or trying to bring them into some specific domains (Chemistry, Biology, etc.). So it comes to my next point, I believe we would get something really great when combining the most frontier models with different domains. And I would say <strong>AI + Science &#8805; Science</strong>. I believe that we might have models that are as knowledgeable and creative as top scientists by 2025-26, including Nobel Prize winners or heads of research labs; for example, if we had a million copies of such AI models, they can collaborate with each others, and just work like a research group. This unlike human scientists who definitely would get tired, AI systems would not, you can just turn the server on and let them do research in the background. The research group they form would have greater efficiency comparing to human one, while also free human scientists from heavy (unnecessary maybe?) work. And more importantly, may accelerate the rate of scientific discoveries significantly. We can actually use AI to help us solve some of the most challenging or unsolved problems in the world, such as climate change, cancer, etc. I think it would not only be a great way to let human society adapt AI development, but also largely accelerate the development of human society. Except of this, I think we should also consider the existential crisis that AI would possibly bring. As those AI models become more advanced in areas like creativity, reasoning, and emotional intelligence, some of us may question what makes us truly unique or special as a species. Besides, AI raises profound questions about the nature of consciousness, intelligence, and what it means to be a sentient being. What&#8217;s more, AI&#8217;s potential to solve many of our problems might lead some to question the meaning and purpose of our existence if our traditional roles are diminished.</p><p>I think these are complex and nuanced questions without simple answers; and only time would give us the answer.</p>]]></content:encoded></item><item><title><![CDATA[My Few Thoughts on AI Security and AGI]]></title><description><![CDATA[Published on June 6, 2024]]></description><link>https://blog.richardstu.com/p/my-few-thoughts-on-ai-security-and</link><guid isPermaLink="false">https://blog.richardstu.com/p/my-few-thoughts-on-ai-security-and</guid><dc:creator><![CDATA[Richards Tu]]></dc:creator><pubDate>Sat, 06 Jul 2024 14:27:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/359756a6-3206-4e63-bf73-219fcb2e7dab_1344x896.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h2>On AI Security</h2><p>I think the security is very important to the future AI development. I have argued with another person for quite a long time. He thinks that the current AI systems are not capable enough to have threat to us, while I think we should take precautions, we should get full prepared for any kind of thing that could happen in the future. I mean, I didn&#8217;t believe the &#8220;catastrophic consequences&#8221; shown in the Terminator in the past, but for today, I would fear bad things that would come along due to the rapid development. I really believe that stuff like alignment, security post-training is the lifeguard of humanity. They can set a guardrail for the model capacity, for the &#8220;monster&#8221; inside; like locking the model in a cage and we, human, can study it from outside w/out letting it harm us. But this does not mean that I&#8217;m denying or rejecting the development of AI systems, I think the current ones are great; gpt-4o, claude-3-opus, and upcoming llama-3-400b are all quite awesome. I just want them to be more secure, while being capable, which means finding a perfect balance between these two. And OpenAI just announced that new Security Committee will be formed, I hope it works ;) <em>By the way, I really love Anthropic; I love their models, I love their research, especially the recent one about model interpretability, I think the ability to &#8220;enable&#8221; the features in the model really pave a new way for model capacity discovery. I was surprised when I first read the blog.</em></p><h2>On AGI</h2><p>It&#8217;s a pretty abstract concept. I mean, how &#8220;general&#8221; is &#8220;general&#8221;? Even ASL classification by Anthropic is kinda fuzz. I think, to achieve &#8220;AGI&#8221; does not necessarily mean that we need to have an AI system that exceeds human in all domains. We just need one that exceeds average human ability in most domains. But before that, I think we should make sure the AI system can <em>really</em> understand:</p><ul><li><p>Riddles</p></li><li><p>Jokes</p></li><li><p>Memes</p></li><li><p>Idioms</p></li><li><p>Stuff related to special cultural context</p></li></ul><p>Although these seem not so important, I think they can reflect the model&#8217;s basic and crucial language ability. They are the system that predicts the next words, so unless they really understand the pattern under words, they won&#8217;t be able to nail the points I mentioned. And if the model has strong capacity in next-word-prediction or really understands the underlying mystery between the words, it would possibly be really capable overall. (I learnt it from a podcast from Ilya hahaha) But then again, what would possibly happen if an AI system really exceeds humans in not just average but all domains? Will we be replaced? Only time will tell. The only thing we can do is to get fully prepared, and make sure the future frontier will reach a balance.</p><p>I hope AGI can really benefits humanity in the future.</p>]]></content:encoded></item></channel></rss>