<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Joe Dong | Physical AI Deep Dives]]></title><description><![CDATA[Deep dives on robotics and physical AI papers, system design, and the commercial implications that matter.]]></description><link>https://newsletter.joedong.ai</link><image><url>https://substackcdn.com/image/fetch/$s_!wK-Z!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F87948048-3349-497d-9bef-88fb4ad4bc2f_1024x1024.png</url><title>Joe Dong | Physical AI Deep Dives</title><link>https://newsletter.joedong.ai</link></image><generator>Substack</generator><lastBuildDate>Tue, 12 May 2026 15:36:44 GMT</lastBuildDate><atom:link href="https://newsletter.joedong.ai/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Joe]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[joedong42@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[joedong42@substack.com]]></itunes:email><itunes:name><![CDATA[Joe Dong]]></itunes:name></itunes:owner><itunes:author><![CDATA[Joe Dong]]></itunes:author><googleplay:owner><![CDATA[joedong42@substack.com]]></googleplay:owner><googleplay:email><![CDATA[joedong42@substack.com]]></googleplay:email><googleplay:author><![CDATA[Joe Dong]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[System Thinking and Model Thinking in Robot Learning]]></title><description><![CDATA[Both schools are betting on the same variable -- DATA.]]></description><link>https://newsletter.joedong.ai/p/system-thinking-and-model-thinking</link><guid isPermaLink="false">https://newsletter.joedong.ai/p/system-thinking-and-model-thinking</guid><dc:creator><![CDATA[Joe Dong]]></dc:creator><pubDate>Mon, 11 May 2026 14:02:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!q2RN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Two things landed on my feed in the same week &#8212; Danfei Xu&#8217;s interview with whynotTV[1] and the latest Genesis release[2] &#8212; and the word that kept showing up in both was the same. Not a new architecture. Not a better loss. It&#8217;s the <strong>System.</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-Sne!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-Sne!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png 424w, https://substackcdn.com/image/fetch/$s_!-Sne!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png 848w, https://substackcdn.com/image/fetch/$s_!-Sne!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png 1272w, https://substackcdn.com/image/fetch/$s_!-Sne!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-Sne!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png" width="1008" height="481" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:481,&quot;width&quot;:1008,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:115064,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/197169351?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-Sne!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png 424w, https://substackcdn.com/image/fetch/$s_!-Sne!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png 848w, https://substackcdn.com/image/fetch/$s_!-Sne!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png 1272w, https://substackcdn.com/image/fetch/$s_!-Sne!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7feb6394-d3da-4b96-9852-d2099df219ab_1008x481.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Two framings</h2><p><strong>Model thinking</strong> treats robot learning as a model problem. What matters is architecture, loss, tokenization, training schedule. Push the model harder; the rest of the stack is given.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.joedong.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Joe Dong | Physical AI Deep Dives! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><strong>System thinking</strong> treats robot learning as a pipeline problem. Scenes, hardware, capture rigs, training stack, evaluation, deployment &#8212; all in scope. The model is one component; push every part of the stack.</p><p>Most teams do some of both. But each team has a default &#8212; the side they invest in when they have to choose.</p><h2>The model school</h2><p>The cleanest commercial expression is the &#960;-series from Physical Intelligence[3][4][5]. Each release foregrounds a model-side mechanism &#8212; fast token serialization, knowledge insulation, conditioning on mixed-quality data &#8212; with the data engine treated as a (well-invested) input rather than the research target. Real contributions; &#960;0.7 is one of the more thoughtful pieces of model work this year.</p><p>The academic counterpart is the VLA-versus-world-model debate. World models[6] consume internet video and simulation; VLA models inherit from instruction-tuned LLMs. Both are arguments about how best to abstract robot learning into a learnable function.</p><p>In every case, the unit of innovation is the model. Data and hardware are taken as given.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4RPl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4RPl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png 424w, https://substackcdn.com/image/fetch/$s_!4RPl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png 848w, https://substackcdn.com/image/fetch/$s_!4RPl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png 1272w, https://substackcdn.com/image/fetch/$s_!4RPl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4RPl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png" width="1456" height="802" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:802,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Understanding pi0 by Physical Intelligence: A Vision-Language-Action Flow Model for General Robot Control&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Understanding pi0 by Physical Intelligence: A Vision-Language-Action Flow Model for General Robot Control" title="Understanding pi0 by Physical Intelligence: A Vision-Language-Action Flow Model for General Robot Control" srcset="https://substackcdn.com/image/fetch/$s_!4RPl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png 424w, https://substackcdn.com/image/fetch/$s_!4RPl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png 848w, https://substackcdn.com/image/fetch/$s_!4RPl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png 1272w, https://substackcdn.com/image/fetch/$s_!4RPl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732421ba-1b65-470b-8ee2-2df918ce7878_2000x1101.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>The system school</h2><p>The clearest example is the UMI line out of Stanford &#8212; UMI[7], HoMMI[8], DexUMI[9]. Each paper introduces a new capture rig as its load-bearing contribution; the policy network is whatever standard architecture fits the data. The point is not &#8220;a better model&#8221; &#8212; it is &#8220;a way to collect data that didn&#8217;t exist before.&#8221;</p><p>Same pattern in Danfei Xu&#8217;s ego-centric direction[1], the EgoScale / EgoVerse consortium work[10][11], and commercial programs like Generalist and Sunday.</p><p>These results seed lineages. UMI led to DexUMI, then HoMMI. EgoScale [10] led to EgoVerse [11]. Capture rigs and protocols outlive any single result. Each becomes someone else&#8217;s starting point.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NE-k!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NE-k!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg 424w, https://substackcdn.com/image/fetch/$s_!NE-k!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg 848w, https://substackcdn.com/image/fetch/$s_!NE-k!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!NE-k!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NE-k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg" width="1200" height="494" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:494,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Image&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Image" title="Image" srcset="https://substackcdn.com/image/fetch/$s_!NE-k!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg 424w, https://substackcdn.com/image/fetch/$s_!NE-k!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg 848w, https://substackcdn.com/image/fetch/$s_!NE-k!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!NE-k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F758e71fc-d43d-4b51-908e-6ca60182871c_1200x494.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>The common thread</h2><p>On the surface the two schools look very different. Model thinking ships architectures; system thinking ships hardware. But ask what each is actually contributing, and the answer lands in the same place.</p><p>UMI&#8217;s gripper is a way to collect manipulation data without a robot in the loop. HoMMI scales whole-body mobile capture by integrating egocentric sensing and mixed-reality interfaces. EgoScale turns egocentric video into a real training signal. Hardware is the visible artifact; data is the contribution.</p><p>Retell the model side the same way. &#960;0.7&#8217;s mixed-quality conditioning is a method for absorbing noisier human-and-robot data without performance collapse[5]. World models use non-robot video as a denser supervision substrate. VLA absorbs internet-scale image-language pretraining into a robot policy.</p><p>The variable being optimized is <strong>data</strong>. The system school competes by enlarging the supply. The model school competes by extracting more signal per unit of existing supply. Two bets on opposite sides of the same <strong>scaling law</strong>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!q2RN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!q2RN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png 424w, https://substackcdn.com/image/fetch/$s_!q2RN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png 848w, https://substackcdn.com/image/fetch/$s_!q2RN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png 1272w, https://substackcdn.com/image/fetch/$s_!q2RN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!q2RN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png" width="1008" height="504" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:504,&quot;width&quot;:1008,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:85880,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/197169351?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!q2RN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png 424w, https://substackcdn.com/image/fetch/$s_!q2RN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png 848w, https://substackcdn.com/image/fetch/$s_!q2RN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png 1272w, https://substackcdn.com/image/fetch/$s_!q2RN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F23a0c92c-83ed-4ecb-a118-91459ff47f8d_1008x504.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Why the system side has more room</h2><p><strong>Robot data is scarce</strong> by several orders of magnitude relative to internet-scale corpora[12]. When supply is that constrained, the biggest lever is enlarging it. Model-side innovations look smaller in absolute terms not because the work is worse &#8212; but because the supply they operate on hasn&#8217;t been allowed to grow yet.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hYfY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hYfY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png 424w, https://substackcdn.com/image/fetch/$s_!hYfY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png 848w, https://substackcdn.com/image/fetch/$s_!hYfY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png 1272w, https://substackcdn.com/image/fetch/$s_!hYfY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hYfY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png" width="1008" height="412" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:412,&quot;width&quot;:1008,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:115296,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/197169351?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hYfY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png 424w, https://substackcdn.com/image/fetch/$s_!hYfY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png 848w, https://substackcdn.com/image/fetch/$s_!hYfY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png 1272w, https://substackcdn.com/image/fetch/$s_!hYfY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F14baa244-6b6e-4a99-ae8c-529895178608_1008x412.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>This will not last. As capture programs, consortium datasets, and simulation pipelines mature, supply expands and the relative size of the two levers shifts. CV and NLP both went through this rhythm: architecture-driven before ImageNet and Common Crawl, extraction-driven after. The honest read is not &#8220;system has won.&#8221; It is &#8220;supply is small enough right now that the side enlarging it has the larger lever.&#8221; <strong>The regime will change.</strong></p><h2>Closing</h2><p>Both schools have already conceded, implicitly, that <strong>data is the binding constraint on robot learning</strong>. They disagree only about which side of the data equation to push on. For builders, the question is which side has more room right now. Right now, the room is on the system side. The teams that invest in both, and shift weight when the regime changes, are the ones I&#8217;d bet on.</p><p></p><h2>References</h2><p>[1] <a href="https://youtu.be/__P5yygfRRQ?si=c02lW283GCer4RM8">Xu, D. Interview on system-centric robot learning. whynotTV, 2026</a>. </p><p>[2] <a href="https://www.genesis.ai/blog/gene-26-5-advancing-robotic-manipulation-to-human-level">Genesis Embodied AI. Latest release and accompanying blog post. 2026.</a> </p><p>[3] Black, K., Brown, N., Driess, D., et al. &#960;0: A Vision-Language-Action Flow Model for General Robot Control. Physical Intelligence, 2024. https://www.physicalintelligence.company/blog/pi0</p><p>[4] Physical Intelligence. &#960;0.5: Knowledge Insulation for Robot Foundation Models. 2025.</p><p>[5] Physical Intelligence. &#960;0.7: Conditioning on Mixed-Quality Robot Data. 2026.</p><p>[6] Hafner, D., et al. DreamerV3: Mastering Diverse Domains through World Models. 2023. See also LeCun, Y., A Path Towards Autonomous Machine Intelligence. 2022.</p><p>[7] Chi, C., Xu, Z., Pan, C., et al. Universal Manipulation Interface: In-the-Wild Robot Teaching Without In-the-Wild Robots. RSS 2024. https://umi-gripper.github.io/</p><p>[8] Xu, X., Park, J., Zhang, H., et al. HoMMI: Learning Whole-Body Mobile Manipulation from Human Demonstrations. arXiv:2603.03243, 2026. </p><p>[9] Stanford REAL Lab. DexUMI: Extending UMI to Dexterous Manipulation. 2025.</p><p>[10] EgoScale: Large-Scale Egocentric Pretraining for Manipulation Policies. 2025.</p><p>[11] Punamiya, Kareer, Liu, et al. EgoVerse: An Egocentric Human Dataset for Robot Learning from Around the World. 2026. https://egoverse.ai/</p><p>[12] Open X-Embodiment Collaboration. Open X-Embodiment: Robotic Learning Datasets and RT-X Models. 2024. https://robotics-transformer-x.github.io/</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.joedong.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Joe Dong | Physical AI Deep Dives! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Robot Learning Should Be Goal-Driven and Data-First]]></title><description><![CDATA[A goal-driven view on robot learning, industrial deployment, and the data engine needed to scale dexterous robots.]]></description><link>https://newsletter.joedong.ai/p/robot-learning-should-be-goal-driven</link><guid isPermaLink="false">https://newsletter.joedong.ai/p/robot-learning-should-be-goal-driven</guid><dc:creator><![CDATA[Joe Dong]]></dc:creator><pubDate>Tue, 28 Apr 2026 14:03:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!kDtu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Generalist AI recently published a <a href="https://generalistai.com/blog/apr-07-2026-beyond-world-models">blog</a> that sparked a lot of discussion in robotics. The main takeaway is simple: robotics research should be more goal-driven than idea-driven.</p><p>Instead of starting from a method label &#8212; world model, VLA, end-to-end policy, diffusion policy &#8212; we should start from the actual objective and constraints.</p><p>What are we trying to achieve? What are the constraints today? Given those constraints, what is the most practical path forward?</p><p>This sounds more modest than chasing a single grand idea. But in robotics, I think this mindset is especially important.</p><h2>The Waymo vs. Tesla analogy</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!c2RI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!c2RI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg 424w, https://substackcdn.com/image/fetch/$s_!c2RI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg 848w, https://substackcdn.com/image/fetch/$s_!c2RI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!c2RI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!c2RI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg" width="550" height="300" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:300,&quot;width&quot;:550,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Waymo vs Tesla Austin: Robotaxi Battle Intensifies&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Waymo vs Tesla Austin: Robotaxi Battle Intensifies" title="Waymo vs Tesla Austin: Robotaxi Battle Intensifies" srcset="https://substackcdn.com/image/fetch/$s_!c2RI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg 424w, https://substackcdn.com/image/fetch/$s_!c2RI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg 848w, https://substackcdn.com/image/fetch/$s_!c2RI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!c2RI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbdc5969e-d9c0-44bc-a51c-f37247e933ce_550x300.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This reminds me of the difference between Waymo and Tesla in autonomous driving.</p><p>Tesla has always been very good at explaining its technical vision publicly: BEVFormer perception, neural planning, end-to-end learning, world models, and so on. It is a very method-forward narrative.</p><p>Waymo feels more goal-driven. The goal is to launch truly driverless service safely and reliably. Given that objective, Waymo combines whatever methods are needed: sensors, maps, SOTA foundation models, simulation, safety validation, and large-scale operations.</p><p>The approaches may look less ideologically pure, but it is practical and effective, as Waymo continuously launches truly driverless robotaxi services in more and more cities. I think robot learning needs more of this attitude. The goal is not to prove that one model class is the final answer. The goal is to make robots work under real-world constraints.</p><h2>What are the real constraints in robot learning?</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vyVN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vyVN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png 424w, https://substackcdn.com/image/fetch/$s_!vyVN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png 848w, https://substackcdn.com/image/fetch/$s_!vyVN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png 1272w, https://substackcdn.com/image/fetch/$s_!vyVN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vyVN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png" width="1456" height="818" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:818,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Re-imagining Telepresence With Humanoid Robots And VR Headsets | Hackaday&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Re-imagining Telepresence With Humanoid Robots And VR Headsets | Hackaday" title="Re-imagining Telepresence With Humanoid Robots And VR Headsets | Hackaday" srcset="https://substackcdn.com/image/fetch/$s_!vyVN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png 424w, https://substackcdn.com/image/fetch/$s_!vyVN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png 848w, https://substackcdn.com/image/fetch/$s_!vyVN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png 1272w, https://substackcdn.com/image/fetch/$s_!vyVN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F34a4a816-0f37-4f21-888f-ce47be044385_1523x856.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>For robot learning today, the biggest constraint is still data.</p><p>Compared with language models, robotics is still in a relatively low-data regime. We do not yet have anything close to internet-scale, high-quality robot interaction data.</p><p>Because of that, highly adaptive and generalized robot policies are still elusive. It is still hard to build one model that can reliably handle many tasks, many objects, and many environments.</p><p>This is why, at Chestnut Robotics, we focus first on <strong>industrial setups.</strong></p><p>Industrial environments are not easy. But they are more structured and controlled than household environments. The workspace is more predictable. The object distribution is narrower. The success criteria are clearer. Given the current state of robot learning, this makes industrial manipulation a more practical place to start.</p><p>To me, this is a goal-driven choice. We are not trying to solve the hardest version of general-purpose robotics on day one. <em>We are trying to choose a deployment setting where today&#8217;s models, hardware, and data pipelines can compound into real progress.</em></p><h2>Tesla was right about data</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WZtY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WZtY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WZtY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WZtY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WZtY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WZtY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg" width="1456" height="699" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:699,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:92367,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/195590145?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WZtY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WZtY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WZtY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WZtY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a4b5ea3-d7fd-40e2-b21f-17feca3e6079_1536x737.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>That said, Tesla also had a very important insight.</p><p>Even if one debates specific technical choices, Tesla understood early that autonomy would ultimately become a data-driven problem. Its product strategy allowed it to collect a huge amount of real-world driving data through deployed vehicles.</p><p>That data-first view is powerful. I think the same is true for robotics. The long-term winner in robot learning will not only have a better model. It will have a better data engine.</p><p>This is where robotics is harder than autonomous driving. High-quality robot data is expensive. Dexterous manipulation data is even more expensive. You need robots, operators, sensors, resets, maintenance, calibration, and a lot of trial and error. For robotic hands, the cost is even higher because contact-rich manipulation is fragile and high-dimensional.</p><p>So the key question becomes: How do we scale high-quality dexterous manipulation data?</p><h2><a href="https://chestnut.bot/">Chestnut</a>&#8217;s data strategy</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kDtu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kDtu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png 424w, https://substackcdn.com/image/fetch/$s_!kDtu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png 848w, https://substackcdn.com/image/fetch/$s_!kDtu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png 1272w, https://substackcdn.com/image/fetch/$s_!kDtu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kDtu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png" width="1456" height="727" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:727,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2207664,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/195590145?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kDtu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png 424w, https://substackcdn.com/image/fetch/$s_!kDtu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png 848w, https://substackcdn.com/image/fetch/$s_!kDtu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png 1272w, https://substackcdn.com/image/fetch/$s_!kDtu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04e9591e-32b9-4800-b5c7-413d6811cbe8_2992x1494.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>At <a href="https://chestnut.bot/">Chestnut</a>, our answer is to rethink the data collection pipeline. Instead of collecting all data directly through robot operation, we use a human-wearable exoskeleton to capture high-quality human dexterous manipulation data.</p><p>But the critical point is not just collecting human data. The critical point is conversion.</p><p>Through our co-designed robotic hand and generative inpainting algorithm, we can eventually convert human hand interaction data into high-quality robot data. This is the key merit of the system: we are not only recording humans; we are building a scalable bridge from human dexterity to robot dexterity.</p><p>This matters because human dexterous data is much easier to scale than robot dexterous data. But the biggest downside of human data is the embodiment gap between robot and human. If we can reliably convert human dexterous data into robot-quality training data, then we can change the data economics of robot learning.</p><p>That is the deeper bet. It&#8217;s a better way to create the data that makes better policies possible.</p><h2>A more modest view</h2><p>My current view is that robotics probably will not be solved by one clean architecture choice.</p><p>World models are useful. VLAs are useful. End-to-end policies are useful. Classical planning, simulation, teleoperation, and structured industrial deployment can all be useful too.</p><p>The real question is how to combine them under today&#8217;s constraints.</p><p>For us, the most important constraints are clear: robot learning needs more high-quality data, especially for dexterous manipulation, and current robot data collection is too expensive to scale.</p><p>So our focus is simple: start from structured industrial environments, build a scalable data engine, and use human dexterity as a path toward robot dexterity.</p><p>This may not sound as grand as claiming one model paradigm will solve robotics. But it feels like a practical path forward.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.joedong.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Joe Dong | Physical AI Deep Dives! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[π0.7: Prompt Engineering Comes to Robot Foundation Models]]></title><description><![CDATA[&#960;0.7&#8217;s key advance is full action context, not just a stronger model&#8212;unlocking compositional generalization.]]></description><link>https://newsletter.joedong.ai/p/07-prompt-engineering-comes-to-robot</link><guid isPermaLink="false">https://newsletter.joedong.ai/p/07-prompt-engineering-comes-to-robot</guid><dc:creator><![CDATA[Joe Dong]]></dc:creator><pubDate>Thu, 23 Apr 2026 15:00:51 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/863fe5d3-3298-4bc4-a460-49e54eaf9a8d_772x475.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!i6kY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!i6kY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png 424w, https://substackcdn.com/image/fetch/$s_!i6kY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png 848w, https://substackcdn.com/image/fetch/$s_!i6kY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png 1272w, https://substackcdn.com/image/fetch/$s_!i6kY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!i6kY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png" width="1289" height="230" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3c259baa-5389-4568-95af-26a13949f28d_1289x230.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:230,&quot;width&quot;:1289,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:43026,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/195185439?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!i6kY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png 424w, https://substackcdn.com/image/fetch/$s_!i6kY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png 848w, https://substackcdn.com/image/fetch/$s_!i6kY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png 1272w, https://substackcdn.com/image/fetch/$s_!i6kY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c259baa-5389-4568-95af-26a13949f28d_1289x230.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p><strong>&#960;0.7 expands the input to a robot policy from a single task instruction into a full context for how to act: language subtasks, subgoal images, episode metadata, and control mode. That is why it starts to show something robot learning has long been missing: compositional generalization.</strong></p><div><hr></div><h2>Overview</h2><p>Physical Intelligence recently released <strong>&#960;0.7: a Steerable Generalist Robotic Foundation Model with Emergent Capabilities</strong>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.joedong.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Joe Dong | Physical AI Deep Dives! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><a href="https://www.pi.website/blog/pi07">https://www.pi.website/blog/pi07</a></p><p><a href="https://arxiv.org/pdf/2604.15483">https://arxiv.org/pdf/2604.15483</a></p><p>My main takeaway is simple:</p><blockquote><p><strong>The key step in &#960;0.7 is not just better performance. It is that the model no longer takes only a task instruction. It takes a full prompt about how to do the task. That includes language subtasks, subgoal images, episode metadata, and control mode. Because of that, it starts to show clear signs of compositional generalization.</strong></p></blockquote><p>What matters most in this paper is not one flashy demo.</p><p>What matters is the broader recipe: <strong>absorb all usable data, then use richer prompts and context to explicitly tell the model about strategy differences, quality differences, and goal differences inside that data.</strong></p><p>That way, the model learns not only <em>what</em> to do, but also:</p><ul><li><p>how to do it</p></li><li><p>how fast to do it</p></li><li><p>how well to do it</p></li><li><p>whether mistakes happened</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!r1kd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!r1kd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png 424w, https://substackcdn.com/image/fetch/$s_!r1kd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png 848w, https://substackcdn.com/image/fetch/$s_!r1kd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png 1272w, https://substackcdn.com/image/fetch/$s_!r1kd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!r1kd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png" width="1405" height="855" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bde6a420-052f-439f-b4bd-51e162547155_1405x855.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:855,&quot;width&quot;:1405,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1097437,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/195185439?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!r1kd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png 424w, https://substackcdn.com/image/fetch/$s_!r1kd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png 848w, https://substackcdn.com/image/fetch/$s_!r1kd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png 1272w, https://substackcdn.com/image/fetch/$s_!r1kd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbde6a420-052f-439f-b4bd-51e162547155_1405x855.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>Why this matters</h2><h3>1. Robot prompt engineering becomes a core paradigm</h3><p>Many robot VLAs still use a very simple prompt. In practice, it is often just a task description.</p><p>&#960;0.7 goes much further. Its prompt can include:</p><ul><li><p>task and subtask language</p></li><li><p>subgoal images</p></li><li><p>episode metadata such as speed, quality, and mistakes</p></li><li><p>control mode such as joint control or end-effector control</p></li></ul><p>This matters because robot data is naturally messy.</p><p>Even for the same task, one operator may be fast, another careful. Some trajectories are expert demonstrations. Some are failed trajectories. Some are autonomous rollouts.</p><p><strong>Without extra context, all of these modes get mixed together. Then the model tends to learn an average behavior that looks like nobody.</strong></p><p>That is the core idea of &#960;0.7:</p><blockquote><p><strong>Use prompting to explicitly separate these modes.</strong></p></blockquote><p>My own read is that <strong>&#960;0.7 is fundamentally a diverse prompting strategy for robot models</strong>. It feels similar to prompt expansion in language models and video models.</p><p>The contribution is not just scaling parameters. It is organizing the conditioning signal in a much richer and more controllable way.</p><h3>2. Mixed-quality data is not just usable. It may be the better route.</h3><p>One strong signal from this paper is that <strong>low-quality data, failed data, autonomous rollout data, and even non-robot data do not have to be treated as noise.</strong></p><p>If the conditioning is done correctly, they can become useful training signals.</p><p>The ablations are especially important here. Without metadata, model performance can drop as the dataset gets larger but lower-quality on average. With metadata, &#960;0.7 keeps improving as more mixed-quality data is added.</p><p>That is a very important result.</p><p>It suggests that the key problem in future robot learning is not simply filtering data harder. It is:</p><blockquote><p><strong>How do we organize messy but real-world data so that the model can actually learn from it?</strong></p></blockquote><h3>3. Compositional generalization finally starts to look real</h3><p>One major weakness of current robot foundation models is that <strong>semantic generalization is often better than compositional generalization</strong>.</p><p>A model may recognize many objects and many task words. But once you change the tool, change the order, or recombine skills in a new way, performance often collapses.</p><p>&#960;0.7 is interesting because it starts to show several more convincing forms of compositional behavior:</p><ul><li><p>it can perform some new short-horizon tasks out of the box</p></li><li><p>it can handle new long-horizon tasks through step-by-step human coaching</p></li><li><p>it can distill those coaching trajectories into a higher-level language policy for autonomous execution</p></li></ul><p>That is much closer to how humans teach real skills.</p><p>We do not always collect thousands of low-level action labels first. Sometimes we explain the task step by step, then refine it into a repeatable skill.</p><div><hr></div><h2>The most impressive results</h2><h3>1. One generalist model can match specialists</h3><p>One of the strongest results in the paper is that <strong>&#960;0.7 can match, and in some cases exceed, task-specific specialist policies on several dexterous tasks</strong>.</p><p>Examples include:</p><ul><li><p>espresso making</p></li><li><p>box building</p></li><li><p>laundry folding</p></li></ul><p>This is important because the usual assumption is:</p><ul><li><p>generalist model = broader, but weaker</p></li><li><p>specialist model = stronger, but narrower</p></li></ul><p>&#960;0.7 pushes back on that tradeoff.</p><p>It suggests that with the right data recipe and the right prompting strategy, a generalist model can absorb specialist capability without giving up much performance.</p><h3>2. It is actually listening to language, not just replaying dataset bias</h3><p>The paper includes a very nice test called <strong>Reverse Bussing</strong>.</p><p>In the training data, the normal pattern is:</p><ul><li><p>trash goes into the trash</p></li><li><p>dishes go into the bussing bin</p></li></ul><p>At test time, the instruction is reversed:</p><ul><li><p>trash goes into the bussing bin</p></li><li><p>dishes go into the trash</p></li></ul><p>There is also <strong>Reverse Fridge to Microwave</strong>, which similarly goes against the dominant direction in the dataset.</p><p>&#960;0.7 performs much better than earlier models on these tasks.</p><p>That matters a lot.</p><p>Many robot models seem to follow language until language conflicts with the strongest bias in the training data. Then they fall back to the common pattern.</p><p>&#960;0.7 is still not perfect, but at least it shows that <strong>language is no longer just decoration. It is starting to genuinely steer behavior.</strong></p><h3>3. Cross-embodiment transfer starts to look real</h3><p>The paper also has a very interesting embodiment transfer result.</p><p>They collect shirt-folding and laundry-folding data on a smaller bimanual robot, then transfer the skill zero-shot to a much larger dual-UR5e platform.</p><p>What is interesting is that the new robot does <strong>not</strong> simply replay the same motion pattern.</p><p>Instead, it finds a different strategy that better fits its own geometry and dynamics. For example, the smaller robot may pick up clothes from the side, while the larger UR5e setup prefers a more top-down vertical grasp.</p><p>Even more striking, the zero-shot transferred policy gets very close to expert teleoperators on shirt folding.</p><p>That suggests the model is not just imitating source trajectories. It is learning something more like a higher-level manipulation skill.</p><div><hr></div><h2>Important details</h2><h3>1. The point is not just a bigger model</h3><p>A key detail here is that <strong>&#960;0.7 is not simply a giant video diffusion robot policy</strong>.</p><p>The main VLA is around <strong>5B</strong>:</p><ul><li><p>a Gemma3 4B VLM backbone</p></li><li><p>a MEM-style video history encoder</p></li><li><p>an <strong>860M flow-matching action expert</strong></p></li></ul><p>It also continues PI&#8217;s earlier ingredients:</p><ul><li><p>knowledge insulation</p></li><li><p>FAST tokens</p></li><li><p>flow matching objectives</p></li></ul><p>The larger branch is the one used to generate subgoal images. That branch is initialized from <strong>BAGEL 14B</strong> and is responsible for turning language subtasks into near-future multi-view goal images.</p><p>So the important point is not &#8220;just make the model bigger.&#8221;</p><p>The important point is:</p><blockquote><p><strong>Build a system with clear division of labor, then connect language, world goals, actions, and data quality through prompting.</strong></p></blockquote><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DBGg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DBGg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png 424w, https://substackcdn.com/image/fetch/$s_!DBGg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png 848w, https://substackcdn.com/image/fetch/$s_!DBGg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png 1272w, https://substackcdn.com/image/fetch/$s_!DBGg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DBGg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png" width="1383" height="669" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:669,&quot;width&quot;:1383,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:239330,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/195185439?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DBGg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png 424w, https://substackcdn.com/image/fetch/$s_!DBGg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png 848w, https://substackcdn.com/image/fetch/$s_!DBGg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png 1272w, https://substackcdn.com/image/fetch/$s_!DBGg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa77fe67b-3f9d-4277-9a9e-d899a97c2a5a_1383x669.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>2. Subgoal images are genuinely useful</h3><p>Goal conditioning is not new. But &#960;0.7 uses it in a much more systematic way.</p><p>It does not rely on a single goal image. It uses <strong>multi-view subgoals</strong>, including base views and wrist views.</p><p>That is important because the goal needs to specify both:</p><ul><li><p>what should happen to the object and the scene</p></li><li><p>what state the robot arms and grippers should reach</p></li></ul><p>The paper also makes two practical engineering choices I like:</p><ul><li><p>randomly dropping different parts of the prompt during training</p></li><li><p>mixing real future images with generated goal images during training</p></li></ul><p>Both are very pragmatic.</p><p>The first makes the model robust to different prompt subsets at test time.</p><p>The second reduces the mismatch between training and deployment, since test-time subgoals may come from the world model rather than from ground-truth future frames.</p><h3>3. The ablation story is unusually strong</h3><p>I also like the ablations in this paper.</p><p>They are not just &#8220;remove one module and report the drop.&#8221; They ask more basic questions:</p><ol><li><p><strong>Can the model really learn from mixed-quality data?</strong></p></li><li><p><strong>Does data diversity actually improve generalization?</strong></p></li></ol><p>One especially good experiment compares:</p><ul><li><p>removing the most diverse 20% of the data</p></li><li><p>removing a random 20%</p></li></ul><p>The result is very clear: removing the most diverse 20% hurts much more.</p><p>That means the important variable is not only data scale. It is <strong>task diversity</strong>.</p><div><hr></div><h2>My take</h2><h3>1. Using all data sources, especially egocentric human data, is the 2026 direction</h3><p>My view is pretty clear here.</p><p>Demonstrations, suboptimal data, autonomous rollouts, egocentric human video, and web multimodal data should all move into the same training pipeline.</p><p>The key question is no longer whether to mix them.</p><p>The key question is <strong>how to condition on them correctly</strong>.</p><h3>2. &#8220;Dexterous task&#8221; is still a bit hand-wavy</h3><p>I increasingly think dexterity should be split into at least two levels:</p><ul><li><p><strong>wrist dexterity</strong>: folding laundry, peeling vegetables, handling objects with richer contact</p></li><li><p><strong>finger dexterity</strong>: pressing small buttons, using a mouse, fine in-hand manipulation, true multi-finger regrasp</p></li></ul><p>&#960;0.7 looks very strong on the first category.</p><p>It still looks much less convincing on the second.</p><p>That boundary matters, and I think many robotics papers still blur it too much.</p><h3>3. &#8220;Embodiment&#8221; is also becoming hand-wavy</h3><p>For UMI-style or end-effector-centric approaches, embodiment often mainly means the end-effector interface.</p><p>As long as the hand pose trajectory is aligned, the arm and body can often be abstracted away and handled by IK or other lower-level controllers.</p><p>But for PI-style joint-level policies, the robot arm itself is part of the embodiment.</p><p>That is because the model directly predicts joint-level control. It must understand how this specific robot moves in order to choose a good strategy.</p><p>The laundry-folding transfer result shows this nicely. A smaller arm and a larger UR5e do not just execute the same motion with different scales. They may prefer different grasping strategies altogether.</p><p>My own view is that if UMI-like approaches are going to scale well, the robot body may need to become more human-like, at least in upper-body geometry and reachable manipulation space. That would make human hand 6D motion easier to reproduce with lower-level controllers.</p><h3>4. Goal conditioning is powerful, but I still have doubts about long-term scalability</h3><p>To be honest, I think the goal-conditioned route is very effective, but also somewhat hacky.</p><p>It makes the learning problem easier. But it also separates the system into two parts:</p><ul><li><p>one module generates the goal or subgoal</p></li><li><p>another module executes the trajectory conditioned on that goal</p></li></ul><p>This is practical, and clearly useful.</p><p>But it can also introduce suboptimality. If the generated goal is not quite right, the downstream policy may still be pulled in the wrong direction.</p><p>So my current view is:</p><p><strong>This is a strong near-term direction. I am just not yet convinced it is the final scalable form of a unified robot foundation model.</strong></p><div><hr></div><h2>My open question</h2><p>The biggest open question for me is this:</p><p><strong>How well can &#960;0.7-style generalization transfer from grippers and parallel-jaw grippers to truly dexterous hands?</strong></p><p>I am still skeptical.</p><p>The gap from grippers to dexterous hands may be much larger than the arm-morphology gap shown in this paper.</p><p>It is not just about reach, inertia, or kinematics.</p><p>The action space changes. The contact modes change. The manipulation regime changes.</p><p>If that gap turns out to be too large, then many current &#8220;cross-embodiment&#8221; conclusions may still mostly hold inside the gripper world.</p><div><hr></div><h2>Final take</h2><p>If I had to summarize &#960;0.7 in one sentence, I would say this:</p><blockquote><p><strong>It is one of the clearest demonstrations so far that the next step in robot foundation models is not just larger models, and not just cleaner data, but richer prompts, messier but more realistic data, and stronger steerability.</strong></p></blockquote><p>What excites me most is not one benchmark result.</p><p>It is that &#960;0.7 pushes robot learning away from the old interface of &#8220;collect trajectories and train a policy,&#8221; and toward a new interface of <strong>prompt, coach, steer, and compose</strong>.</p><p>That feels much closer to the path toward truly general robot models.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.joedong.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Joe Dong | Physical AI Deep Dives! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[NVIDIA EgoScale: Pretraining Dexterous Manipulation with 20,000 Hours of Egocentric Human Data]]></title><description><![CDATA[A simple but important idea: pretrain on large-scale human egocentric data first, then use a small amount of aligned human&#8211;robot data and robot demonstrations to make it work on real dexterous hands.]]></description><link>https://newsletter.joedong.ai/p/nvidia-egoscale-pretraining-dexterous</link><guid isPermaLink="false">https://newsletter.joedong.ai/p/nvidia-egoscale-pretraining-dexterous</guid><dc:creator><![CDATA[Joe Dong]]></dc:creator><pubDate>Tue, 14 Apr 2026 16:58:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8HzR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Jfh0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Jfh0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png 424w, https://substackcdn.com/image/fetch/$s_!Jfh0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png 848w, https://substackcdn.com/image/fetch/$s_!Jfh0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png 1272w, https://substackcdn.com/image/fetch/$s_!Jfh0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Jfh0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png" width="1372" height="586" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:586,&quot;width&quot;:1372,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:99202,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/194207153?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Jfh0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png 424w, https://substackcdn.com/image/fetch/$s_!Jfh0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png 848w, https://substackcdn.com/image/fetch/$s_!Jfh0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png 1272w, https://substackcdn.com/image/fetch/$s_!Jfh0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbd0b5f8-1ae2-4871-a827-311fea541932_1372x586.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Overview</h2><p>Recently NVIDIA GEAR released EgoScale. My one-line summary is:</p><blockquote><p>Use egocentric human video plus wrist and hand motion as a scalable supervision signal to pretrain a VLA. Then use a small amount of aligned human&#8211;robot data plus robot data to make it work on real dexterous manipulation.</p></blockquote><p>That is the core recipe.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.joedong.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Joe Dong | Physical AI Deep Dives! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>The most important point is that this paper does not just show transfer. It shows a scaling law. EgoScale pretrains on 20,854 hours of egocentric human video. In the scaling study, average downstream task completion rises from 0.30 at 1k hours to 0.71 at 20k hours. The human-action validation loss follows a near log-linear trend, and that offline loss tracks real robot performance closely.</p><p>So my main takeaway is simple:</p><p><strong>This paper argues that dexterous manipulation can be pretrained from large-scale human data first, then aligned and improved with a much smaller amount of robot data.</strong></p><h2>Why this matters</h2><h3>1. It systematically shows that human data can scale for dexterous manipulation</h3><p>Earlier human-to-robot work already showed that transfer is possible. EgoScale is the scale-up version. The paper explicitly argues that prior work was limited by smaller human datasets and by lower-DoF hands. EgoScale pushes pretraining to more than 20,854 hours, which is over 20&#215; larger than prior efforts, and it tests transfer on a 22-DoF dexterous hand rather than a simple gripper.</p><p>So the bigger question here is:</p><blockquote><p>Can dexterous manipulation be pretrained from large-scale human data?</p></blockquote><p>EgoScale&#8217;s answer is yes. That is why I think this paper matters. It suggests that dexterous foundation models may move from a robot-data-first route to a human-data-first route.</p><p>This is also why I see EgoScale as a natural extension of work like EgoMimic, EgoVLA, DexWild, and related human-to-robot transfer papers. The difference is not only better engineering. The difference is scale.</p><h3>2. It shows a strong link from offline metric to real robot performance</h3><p>This is the part I care about most. The paper shows that human-action validation loss is not just an offline number. It correlates strongly with downstream real-robot task completion. That makes pretraining much more predictable. If the validation loss keeps improving at scale, downstream robot performance is likely to improve too.</p><p>That is a big deal. In many robot papers, offline metrics and real-world performance are only loosely connected. Here the paper is saying something stronger: the offline pretraining metric is actually useful.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8HzR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8HzR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png 424w, https://substackcdn.com/image/fetch/$s_!8HzR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png 848w, https://substackcdn.com/image/fetch/$s_!8HzR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png 1272w, https://substackcdn.com/image/fetch/$s_!8HzR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8HzR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png" width="1456" height="807" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:807,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2234931,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/194207153?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8HzR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png 424w, https://substackcdn.com/image/fetch/$s_!8HzR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png 848w, https://substackcdn.com/image/fetch/$s_!8HzR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png 1272w, https://substackcdn.com/image/fetch/$s_!8HzR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42a4cc52-c25f-4d90-a038-bb5490803741_2082x1154.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h2>Important details</h2><h3>1. Action representation</h3><p>The paper makes a very clear action-representation choice.</p><ul><li><p><strong>For the arm / wrist:</strong> use relative wrist motion between consecutive timesteps. This avoids relying on a fixed global coordinate frame.</p></li><li><p><strong>For the fingers:</strong> retarget human hand motion into a 22-DoF dexterous hand joint space.</p></li><li><p><strong>Why this matters:</strong> the wrist-only version performs badly. The fingertip-based version is better, but still inconsistent. The retargeted joint-space hand action is the most stable choice across tasks.</p></li></ul><p>My read is simple:</p><ul><li><p><strong>Wrist motion + finger motion is becoming the main representation.</strong></p></li><li><p>The robot&#8217;s native joint format matters less.</p></li><li><p>The most transferable interface is how the wrist moves and how the fingers move.</p></li></ul><p>That is not the paper&#8217;s exact wording. But it is the clearest design message I take from it.</p><h3>2. The data pyramid</h3><p>The training recipe is the paper.</p><ul><li><p><strong>Stage I: very large human pretraining.</strong><br> The model is pretrained on 20,854 hours of egocentric human data. The main corpus spans 9,869 scenes, 6,015 tasks, and 43,237 objects. Most of it is in-the-wild first-person video. The supervision is noisy but scalable. The mixture is also complemented by 829 hours of EgoDex with more accurate wrist and hand tracking. </p></li><li><p><strong>Stage II: small aligned human&#8211;robot mid-training.</strong><br> This dataset contains 344 tabletop tasks. Each task has about 30 human trajectories and 5 robot trajectories, totaling about 50 hours of human data and 4 hours of robot data. The setup is carefully aligned: matched viewpoints, calibrated intrinsics, Vive trackers for wrist pose, and Manus gloves for in-hand motion. </p></li><li><p><strong>Stage III: task post-training.</strong><br> The policy is then fine-tuned on five dexterous tasks on the Galaxea R1 Pro. Most tasks use 100 teleoperated robot demonstrations. Shirt rolling uses 20. The five tasks are shirt rolling, card sorting, tong-based fruit transfer, bottle-cap unscrewing, and syringe liquid transfer. </p></li></ul><p>My own read of this recipe is:</p><ul><li><p><strong>Pretraining</strong> is mainly for learning broad manipulation priors: real-world dynamics, semantic grounding, robustness, and scene generalization.</p></li><li><p><strong>Mid-training</strong> is for alignment. It grounds the pretrained representation in the robot&#8217;s sensing and control space.</p></li><li><p><strong>Post-training</strong> is for task performance. It turns the prior into reliable task completion.</p></li></ul><p>So I do <strong>not</strong> read EgoScale as &#8220;robot data is no longer needed.&#8221; The paper&#8217;s recipe is more precise than that:</p><ul><li><p>Use huge, cheap human data for scale.</p></li><li><p>Use small aligned data for grounding.</p></li><li><p>Use a limited amount of robot task data for success rate.</p></li></ul><p>There is another important implication here. For Stage I, the paper suggests that <strong>scale and diversity matter more than perfect precision</strong>. The data is noisy and not sensor-aligned, but it still beats the small aligned-only baseline across most tasks. That is a very foundation-model-like result.</p><h3>3. One-shot adaptation is a strong result</h3><p>The paper also evaluates adaptation to unseen skills under very limited robot supervision.</p><ul><li><p>Start from the human-pretrained model.</p></li><li><p>Do aligned mid-training.</p></li><li><p>Post-train on a new task with only 1 robot demonstration plus 100 aligned human demonstrations.</p></li></ul><p>In that setting, the model reaches 0.88 success on fold-shirt and 0.55 success on unscrewing water bottles. Models that remove either large-scale human pretraining or aligned mid-training fail in this setting. This is a strong sign that the pretrained representation is doing real work.</p><h3>4. Wrist camera is very important</h3><p>This is another point I strongly agree with.</p><p>The robot uses three RGB cameras:</p><ul><li><p>one head-mounted egocentric camera</p></li><li><p>two wrist cameras facing the palm</p></li></ul><p>This is not elegant. But it is practical. For dexterous tasks, the wrist view is extremely useful. It sees contact details that the head camera often misses. That matters for cards, tongs, bottle caps, and syringe operations, where small contact errors cause the whole task to fail.</p><p>There is also an interesting gap here. The large-scale Stage I human pretraining is mostly ordinary egocentric video, while the robot and aligned mid-training setup use the richer head-plus-wrist camera configuration. So one obvious question is:</p><p><strong>If we also add wrist-view human data at scale, would transfer improve further?</strong></p><h2>Cross-embodiment transfer</h2><p>The paper also tests transfer to a different embodiment: the Unitree G1 with a 7-DoF tri-finger hand.</p><ul><li><p>The G1 has a shorter arm and different kinematics.</p></li><li><p>The tasks are Pen in Bin and Dish in Rack.</p></li><li><p>Human pretraining plus embodiment-specific mid-training improves performance substantially over using the G1 data alone.</p></li></ul><p>The paper&#8217;s intended message is clear: large-scale human motion learns a reusable motor prior. It is not completely tied to one dexterous hand. I think that claim is directionally right. But my own view is still narrower: this kind of pretraining should work best when the downstream robot also has reasonably dexterous hands. Human data transfers most naturally when the robot can express similar finger behaviors.</p><h2>My 2026 take</h2><p>My takeaway is straightforward.</p><ul><li><p><strong>Very low-cost human video will become a major pretraining source for embodied models.</strong></p></li><li><p><strong>Dexterous hands will become the most important end effector for transferring human manipulation data well.</strong></p></li><li><p><strong>Finger actions and wrist 6D motion will become the key representation layer.</strong></p></li><li><p><strong>The main bottleneck will move to collecting high-quality robot data cheaply for alignment and post-training.</strong></p></li></ul><p>EgoScale does not prove every one of those points by itself. But it strongly points in that direction. The paper already shows that large, noisy human data can supply strong priors, that aligned mid-training makes those priors executable, and that small amounts of robot data then go much farther than they would from scratch.</p><p>This is why I see EgoScale as a paradigm-shift paper. The pretraining data can be low precision, cheap, and massive. The later-stage data can be smaller, more expensive, and more precise. That looks much closer to the foundation-model playbook than to the old robot imitation-learning playbook.</p><h2>My questions</h2><p>I still have a few open questions.</p><ol><li><p><strong>What if large-scale pretraining also includes wrist cameras?</strong><br> The current Stage I data is mostly ordinary first-person video, while the robot setup depends heavily on wrist views. A human data collection pipeline that includes wrist-view video might reduce this perceptual gap. </p></li><li><p><strong>Can the mid-training stage be made much lighter?</strong><br> Right now it still requires matched viewpoints, Vive trackers, and Manus gloves, with humans acting in a robot-compatible workspace. That is much heavier than the Stage I recipe. It clearly works, but it still looks expensive and unnatural for large-scale collection. </p></li><li><p><strong>If high-quality robot trajectory collection becomes cheap, do we still need mid-training?</strong><br>Today, the paper shows that mid-training is very useful. It is what enables one-shot transfer in combination with human pretraining. But if robot data collection becomes dramatically cheaper and higher quality, it is worth asking whether some of the alignment role of mid-training could be replaced by more robot data directly.</p></li></ol><p>A related technical question is whether better scalable data systems could improve Stage II and Stage III at the same time. The paper already shows that alignment data and teleoperated robot data matter. So the next bottleneck may simply be building a cheaper and higher-quality way to collect them.</p><h2>Final take</h2><p>I think the main message of EgoScale is very simple.</p><p><strong>Dexterous manipulation may be moving from &#8220;robot data first&#8221; to &#8220;human data first.&#8221;</strong></p><p>The &#8220;ego&#8221; in EgoScale is not just the camera viewpoint. It is the scalable supervision source. Use egocentric human video to pretrain broad motor priors. Use a small amount of aligned data to connect that prior to the robot. Then use a limited amount of robot data to drive task success.</p><p>That is the recipe. And this paper shows that the recipe actually scales.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://newsletter.joedong.ai/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Joe Dong | Physical AI Deep Dives! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Nvidia DreamZero0: Why World Models May Become Policies]]></title><description><![CDATA[DreamZero0 is not just a stronger robotics benchmark.]]></description><link>https://newsletter.joedong.ai/p/nvidia-dreamzero0-why-world-models</link><guid isPermaLink="false">https://newsletter.joedong.ai/p/nvidia-dreamzero0-why-world-models</guid><dc:creator><![CDATA[Joe Dong]]></dc:creator><pubDate>Wed, 08 Apr 2026 23:57:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!y408!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>DreamZero0 is not just a stronger robotics benchmark. It is a serious argument for a new robotics stack.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!luSq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!luSq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png 424w, https://substackcdn.com/image/fetch/$s_!luSq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png 848w, https://substackcdn.com/image/fetch/$s_!luSq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png 1272w, https://substackcdn.com/image/fetch/$s_!luSq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!luSq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png" width="1137" height="529" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:529,&quot;width&quot;:1137,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:94164,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/193638173?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!luSq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png 424w, https://substackcdn.com/image/fetch/$s_!luSq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png 848w, https://substackcdn.com/image/fetch/$s_!luSq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png 1272w, https://substackcdn.com/image/fetch/$s_!luSq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa7234789-ca39-414c-a722-2ea57ad49c82_1137x529.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h2><strong>Key points</strong></h2><p>1. <strong>DreamZero changes the learning objective.</strong> It predicts future video and actions together. That is the core idea behind its <strong>World Action Model</strong> (WAM).</p><p>2. <strong>The bigger contribution is the data thesis.</strong> Broad, messy, non-repetitive robot data may be better for generalization than repeated demos of a narrow task set.</p><p>3. <strong>The transfer result matters.</strong> DreamZero improves from short video-only demonstrations from humans or another robot. Action labels may no longer be the only scarce resource.</p><p>4. <strong>This has direct startup implications.</strong> The next robotics moat may come less from narrow task tuning and more from world-model quality, diverse operational data, sensor standardization, and inference systems.</p><p>5. <strong>The paper is impressive, but not magic.</strong> The model is still short-horizon. It is expensive to serve. High-precision manipulation is still open.</p><h2><strong>Why this paper matters</strong></h2><p>I wanted to start this Substack with DreamZero because it makes a claim that feels bigger than a benchmark win.</p><p>Many recent robotics foundation models still follow the same recipe: images and language in, actions out. Those models can improve semantic understanding and object coverage. But they often still fail when the required motion is new. They fall back to the dominant action prior in the dataset: grasp, move, place.</p><p>DreamZero argues for a different path.</p><p>Its core claim is simple: a robot policy may generalize better if it is trained to model <strong>how the world should change</strong>, not only <strong>which action to emit next</strong>.</p><p>That shift matters because robotics is not only a semantics problem. It is also a dynamics problem. A robot must understand not only <em>what</em> an object is, but <em>how</em> the scene evolves under contact, motion, error, and recovery.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!y408!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!y408!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png 424w, https://substackcdn.com/image/fetch/$s_!y408!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png 848w, https://substackcdn.com/image/fetch/$s_!y408!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png 1272w, https://substackcdn.com/image/fetch/$s_!y408!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!y408!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png" width="1091" height="854" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:854,&quot;width&quot;:1091,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:902486,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://newsletter.joedong.ai/i/193638173?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!y408!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png 424w, https://substackcdn.com/image/fetch/$s_!y408!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png 848w, https://substackcdn.com/image/fetch/$s_!y408!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png 1272w, https://substackcdn.com/image/fetch/$s_!y408!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F974be8fb-bdbf-4a30-993c-81ceb411bae8_1091x854.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>From VLA to WAM</strong></h2><p>The paper introduces DreamZero as a <strong>World Action Model</strong> rather than a standard <strong>Vision-Language-Action</strong> model.</p><p>A standard VLA is trained in a direct way:</p><p><strong>observation + language -&gt; action</strong></p><p>DreamZero changes this to something closer to:</p><p><strong>observation + language -&gt; future world + action</strong></p><p>The model imagines the next visual states of the world while also producing the actions that would make those states happen.</p><p>This matters for one reason: <strong>future video is dense supervision</strong>.</p><p>Every future frame constrains geometry, object motion, contact, temporal consistency, and scene evolution. That is a much richer signal than action labels alone. The action head is not learning in isolation. It is learning inside a model that must maintain a plausible visual future.</p><p>This is the strongest idea in the paper. It makes the old phrase &#8220;world models can become policies&#8221; feel like a practical design choice.</p><p>Technically, the system is large. DreamZero uses a 14B robot foundation model built on a pretrained Wan video-diffusion backbone, with robot-specific state and action modules added on top. The key point is the design principle: <strong>start from a strong video model, keep its spatiotemporal prior, and align robot action with predicted world evolution</strong>.</p><h2><strong>The result that matters most: unseen motion generalization</strong></h2><p>The headline numbers are strong. But the value is in what they mean.</p><p>DreamZero performs well not only on seen tasks in new environments, but also on tasks that were absent from training. That is the hard case. Many models look competent when the object changes or the wording changes. Fewer models keep working when the motion pattern itself changes.</p><p>The paper reports that on AgiBot G1, DreamZero reaches <strong>62.2% average task progress</strong> in zero-shot evaluation on seen tasks with unseen environments and objects. A strong pretrained VLA baseline reaches <strong>27.4%</strong>. On <strong>10 tasks not present in training</strong>, DreamZero reaches <strong>39.5% average task progress</strong>, while VLA baselines drop sharply and often revert to generic pick-and-place behavior.</p><p>This is the right failure mode to focus on.</p><p>If a model hears a new instruction and responds with a familiar manipulation template, that is not robust generalization. It is action prior collapse. DreamZero appears to reduce that problem.</p><h2><strong>The deeper thesis is about data, not architecture</strong></h2><p>The most important part of the paper may be the data recipe.</p><p>DreamZero is pretrained on about <strong>500 hours of teleoperation data</strong> across <strong>22 real environments</strong>. The key point is not only the scale. It is the distribution.</p><p>The data is intentionally <strong>non-repetitive</strong>.</p><p>Instead of collecting many clean repetitions of the same task, the team collects long episodes with many coarse tasks and many transitions between them. Tasks are rotated out after enough collection, which pushes the dataset toward breadth and long-tail coverage.</p><p>This is an important claim.</p><p>A lot of robot data collection still assumes that better policies come from cleaner labels and more repeated demonstrations. DreamZero suggests that for world-model-style policies, repetition may be less valuable than diversity. The model benefits from seeing many kinds of scene changes, many kinds of interaction failures, and many kinds of partial task structure.</p><p>The paper includes a clean ablation. With the same amount of data, DreamZero trained on <strong>diverse data</strong> outperforms DreamZero trained on <strong>repetitive data</strong>, improving task progress from <strong>33% to 50%</strong> on the same evaluation.</p><p>If this result holds more broadly, it changes how we should think about data moat.</p><p>The most valuable robotics dataset may not be the cleanest benchmark dataset. It may be the operational dataset with the widest support over real work: tidying, carrying, folding, wiping, sorting, recovering, and switching tasks inside the same episode.</p><h2><strong>What this means for startups and companies</strong></h2><p>DreamZero changes the answer to a basic question:</p><p><strong>What should a robotics company scale?</strong></p><p>Under the standard VLA view, the answer is often: collect more action-labeled demos across more tasks.</p><p>Under the DreamZero view, the answer shifts. A company should try to scale:</p><ul><li><p>broad multi-view interaction data,</p></li><li><p>enough action data to ground control,</p></li><li><p>strong world-model pretraining,</p></li><li><p>and a deployment stack that can serve a large generative policy in closed loop.</p></li></ul><p>That has three clear implications.</p><p>First, <strong>data strategy changes</strong>. A company in logistics, retail, hospitality, or home robotics may gain more by collecting diverse workflow fragments than by over-optimizing one benchmark skill.</p><p>Second, <strong>video becomes more valuable</strong>. If video-only demonstrations can improve performance, then human demonstration video and cross-fleet recordings become strategic assets.</p><p>Third, <strong>systems engineering becomes part of the moat</strong>. If the policy is also a large generative video model, then inference speed, GPU scheduling, quantization, caching, and asynchronous execution all become first-class product concerns.</p><p>This is why DreamZero matters for investors as well as researchers. It points to a different stack, and therefore a different kind of company.</p><h2><strong>The transfer story matters</strong></h2><p>The cross-embodiment section is the most strategic part of the paper.</p><p>DreamZero improves unseen-task performance using short <strong>video-only demonstrations</strong> from another robot and from humans. Those demonstrations do not include action labels. They only provide visual evidence of task dynamics.</p><p>If a world model already captures much of the task structure, then video can teach the model what successful interaction should look like even when the action interface changes. In the paper, just <strong>20 minutes</strong> of video from another robot or <strong>12 minutes</strong> of human video substantially improves performance on unseen tasks.</p><p>The paper also shows few-shot adaptation to a new bimanual robot with around <strong>30 minutes of play data</strong>.</p><p>I would not overclaim here. This is not proof that transfer across radically different robots is solved. But it is strong evidence that the world prior may transfer faster than the action interface.</p><h2><strong>Two caveats that matter</strong></h2><p>The first caveat is <strong>inference cost</strong>.</p><p>DreamZero is a large video-diffusion policy. That is inherently expensive. The paper reaches about <strong>7 Hz</strong> closed-loop control through a long stack of optimizations. That is impressive. It is also a reminder that WAMs are not just a modeling choice. They are an infrastructure choice.</p><p>The second caveat is <strong>task horizon and precision</strong>.</p><p>The paper is fairly explicit that DreamZero is still closer to a short-horizon &#8220;System 1&#8221; model than a long-horizon planner. It also does not solve fine, high-precision manipulation. Tasks like tight insertion or very delicate contact remain difficult.</p><p>So I do not read this paper as &#8220;VLAs are obsolete.&#8221; I read it as something more useful: <strong>the motion-generalization frontier has moved, and direct action regression is no longer the only serious path</strong>.</p><h2><strong>Final take</strong></h2><p>My main takeaway is direct.</p><p>DreamZero matters because it proposes a better scaling story for robotics.</p><p>The paper suggests that <strong>predicting the world may be a better route to robust action than predicting actions alone</strong>. It also suggests that broad and messy real-world data may be more valuable than narrow repeated demos. And it hints that video-only transfer may become a real lever for embodiment adaptation.</p><p>If those claims continue to hold, the implications are large.</p><ul><li><p>The value of diverse precise robot data goes up.</p></li><li><p>The value of human video goes up.</p></li><li><p>The value of sensor standardization goes up.</p></li><li><p>The value of inference engineering goes up.</p></li><li><p>And the value of narrow demo farms may go down.</p></li></ul><p>That is why I think DreamZero is one of the most important robotics papers in this cycle.</p><p>Not because it solves general robotics.</p><p>But because it changes what a scalable solution might look like.</p>]]></content:encoded></item></channel></rss>