<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd"><channel><title><![CDATA[RoboPapers]]></title><description><![CDATA[Chris Paxton & Michael Cho geek out over robotic papers with paper authors. <br/><br/><a href="https://itcanthink.substack.com?utm_medium=podcast">itcanthink.substack.com</a>]]></description><link>https://itcanthink.substack.com/podcast</link><generator>Substack</generator><lastBuildDate>Wed, 08 Apr 2026 21:15:01 GMT</lastBuildDate><atom:link href="https://api.substack.com/feed/podcast/2883266.rss" rel="self" type="application/rss+xml"/><author><![CDATA[Chris Paxton and Michael Cho]]></author><copyright><![CDATA[Chris Paxton]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[itcanthink@substack.com]]></webMaster><itunes:new-feed-url>https://api.substack.com/feed/podcast/2883266.rss</itunes:new-feed-url><itunes:author>Chris Paxton and Michael Cho</itunes:author><itunes:subtitle>Robotics and AI; the future we&apos;re building and how we&apos;ll get there</itunes:subtitle><itunes:type>episodic</itunes:type><itunes:owner><itunes:name>Chris Paxton and Michael Cho</itunes:name><itunes:email>itcanthink@substack.com</itunes:email></itunes:owner><itunes:explicit>No</itunes:explicit><itunes:category text="Technology"/><itunes:category text="Science"/><itunes:image href="https://substackcdn.com/feed/podcast/2883266/f928778615871704988ed827dec56683.jpg"/><item><title><![CDATA[RoboPapers Episode 4: Vision Language Models are In-Context Value Learners ]]></title><description><![CDATA[<p><em>Note: this is an old episode of  </em><a target="_blank" href="https://x.com/robopapers"><em>RoboPapers</em></a><em>. I am uploading old episodes of our podcast here to make it easier to find them and to find our reference materials. Please follow the podcast on </em><a target="_blank" href="https://www.youtube.com/@RoboPapers"><em>Youtube</em></a><em> or on </em><a target="_blank" href="https://open.spotify.com/show/3U0Ed7poaOElItEyUPkuto"><em>Spotify</em></a><em>.</em></p><p>Value prediction is an important robotics problem, wherein we can determine how useful a state is to the successful execution of a task in the future. The key insight that Jason and his co-authors showed in this paper was simple:</p><p>* Large vision-language models already encode a ton of information about task completion from being trained on a wide range of human image and video data</p><p>* We can use these <em>on robotics tasks</em> to estimate progress, which is a useful surrogate for the “value” of any individual state.</p><p>Here’s the abstract:</p><p>Predicting temporal progress from visual trajectories is important for intelligent robots that can learn, adapt, and improve. However, learning such progress estimator, or temporal value function, across different tasks and domains requires both a large amount of diverse data and methods which can scale and generalize. To address these challenges, we present Generative Value Learning (GVL), a universal value function estimator that leverages the world knowledge embedded in vision-language models (VLMs) to predict task progress. Naively asking a VLM to predict values for a video sequence performs poorly due to the strong temporal correlation between successive frames. Instead, GVL poses value estimation as a temporal ordering problem over shuffled video frames; this seemingly more challenging task encourages VLMs to more fully exploit their underlying semantic and temporal grounding capabilities to differentiate frames based on their perceived task progress, consequently producing significantly better value predictions. Without any robot or task specific training, GVL can in-context zero-shot and few-shot predict effective values for more than 300 distinct real-world tasks across diverse robot platforms, including challenging bimanual manipulation tasks. Furthermore, we demonstrate that GVL permits flexible multi-modal in-context learning via examples from heterogeneous tasks and embodiments, such as human videos. The generality of GVL enables various downstream applications pertinent to visuomotor policy learning, including dataset filtering, success detection, and value-weighted regression -- all without any model training or finetuning.</p><p><a target="_blank" href="https://www.youtube.com/watch?v=EtwmJlDhHlY&#38;t=4s">Watch on YouTube</a></p><p><a target="_blank" href="https://open.spotify.com/episode/4SPFVlhIFdg4vsWXZFh3ke?si=nB24sGeyTJupWRvi9NrzOw">Listen on Spotify</a></p><p><a target="_blank" href="https://x.com/RoboPapers/status/1905379309939724339">Original post on Twitter/X</a></p><p><a target="_blank" href="https://generative-value-learning.github.io/">Project  Site</a></p><p></p> <br/><br/>This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit <a href="https://itcanthink.substack.com?utm_medium=podcast&#38;utm_campaign=CTA_1">itcanthink.substack.com</a>]]></description><link>https://itcanthink.substack.com/p/robopapers-episode-4-vision-language</link><guid isPermaLink="false">substack:post:168346194</guid><dc:creator><![CDATA[Chris Paxton and Michael Cho]]></dc:creator><pubDate>Tue, 15 Jul 2025 00:02:50 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/168346194/4c1605586ec155cba569b4ccfc2521b1.mp3" length="69003526" type="audio/mpeg"/><itunes:author>Chris Paxton and Michael Cho</itunes:author><itunes:explicit>No</itunes:explicit><itunes:duration>4313</itunes:duration><itunes:image href="https://substackcdn.com/feed/podcast/2883266/post/168346194/542d28d4a15e8d70176f0a4c9662e28d.jpg"/></item><item><title><![CDATA[RoboPapers Episode 3: Sim-to-Real Reinforcement Learning for Vision-Based Dexterous Manipulation on Humanoids with Toru Lin]]></title><description><![CDATA[<p>Sim-to-real training is an <a target="_blank" href="https://itcanthink.substack.com/p/bringing-robot-skills-from-simulation">important part of the future of robot skill learning.</a> But a lot of sim-to-real work focuses on <a target="_blank" href="https://ovmm.github.io/">navigation</a>, grasping, or, more recently, non-interactive robot <a target="_blank" href="https://x.com/TheHumanoidHub/status/1891881636595159223">behaviors like dancing</a>. Training <em>dexterous</em> policies for humanoids is very different, because <a target="_blank" href="https://mtmason.com/the-unstable-queen/">manipulation is a very hard problem</a> and multi-finger dexterous hands are even more difficult.</p><p>Enter this cool work from Toru Lin and colleagues. Their goal is to do long-horizon manipulation with dexterous humanoid robots:</p><p>And they discuss what is actually necessary to make this work, which involves things like real-to-sim and reward engineering as well as neural network configuration:</p><p>Abstract</p><p>Reinforcement learning has delivered promising results in achieving human- or even superhuman-level capabilities across diverse problem domains, but success in dexterous robot manipulation remains limited. This work investigates the key challenges in applying reinforcement learning to solve a collection of contact-rich manipulation tasks on a humanoid embodiment. We introduce novel techniques to overcome the identified challenges with empirical validation. Our main contributions include an automated real-to-sim tuning module that brings the simulated environment closer to the real world, a generalized reward design scheme that simplifies reward engineering for long-horizon contact-rich manipulation tasks, a divide-and-conquer distillation process that improves the sample efficiency of hard-exploration problems while maintaining sim-to-real performance, and a mixture of sparse and dense object representations to bridge the sim-to-real perception gap. We show promising results on three humanoid dexterous manipulation tasks, with ablation studies on each technique. Our work presents a successful approach to learning humanoid dexterous manipulation using sim-to-real reinforcement learning, achieving robust generalization and high performance without the need for human demonstration.</p><p><a target="_blank" href="https://toruowo.github.io/recipe/">Project page</a></p><p><a target="_blank" href="https://arxiv.org/abs/2502.20396">ArXi</a>v</p> <br/><br/>This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit <a href="https://itcanthink.substack.com?utm_medium=podcast&#38;utm_campaign=CTA_1">itcanthink.substack.com</a>]]></description><link>https://itcanthink.substack.com/p/robopapers-episode-3-sim-to-real</link><guid isPermaLink="false">substack:post:161177323</guid><dc:creator><![CDATA[Chris Paxton]]></dc:creator><pubDate>Mon, 14 Apr 2025 13:00:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/161177323/a78eb43a4db39ad1745054d3ec11f5da.mp3" length="49214840" type="audio/mpeg"/><itunes:author>Chris Paxton</itunes:author><itunes:explicit>No</itunes:explicit><itunes:duration>3076</itunes:duration><itunes:image href="https://substackcdn.com/feed/podcast/2883266/post/161177323/90d1a48b4824653d20ce762866c0b746.jpg"/></item><item><title><![CDATA[RoboPapers Episode 2: Robot Utility Models with Mahi Shafiullah]]></title><description><![CDATA[<p>One of the dreams of robotics research is being able to download and test out a model and have it *just work*. Mahi talks to us about “robot utility models,” which are essentially just this: models that you can download and test out to do useful things like opening a cabinet or picking up a can.</p><p>This work is extremely interesting and it shows up in my long post on <a target="_blank" href="https://itcanthink.substack.com/p/what-are-the-data-scaling-laws-for">scaling laws for imitation learning.</a> </p><p>Robot models, particularly those trained with large amounts of data, have recently shown a plethora of real-world manipulation and navigation capabilities. Several independent efforts have shown that given sufficient training data in an environment, robot policies can generalize to demonstrated variations in that environment. However, needing to finetune robot models to every new environment stands in stark contrast to models in language or vision that can be deployed zero-shot for open-world problems. In this work, we present Robot Utility Models (RUMs), a framework for training and deploying zero-shot robot policies that can directly generalize to new environments without any finetuning. To create RUMs efficiently, we develop new tools to quickly collect data for mobile manipulation tasks, integrate such data into a policy with multi-modal imitation learning, and deploy policies on-device on Hello Robot Stretch, a cheap commodity robot, with an external mLLM verifier for retrying. We train five such utility models for opening cabinet doors, opening drawers, picking up napkins, picking up paper bags, and reorienting fallen objects. Our system, on average, achieves 90% success rate in unseen, novel environments interacting with unseen objects. Moreover, the utility models can also succeed in different robot and camera set-ups with no further data, training, or fine-tuning. Primary among our lessons are the importance of training data over training algorithm and policy class, guidance about data scaling, necessity for diverse yet high-quality demonstrations, and a recipe for robot introspection and retrying to improve performance on individual environments. Our code, data, models, hardware designs, as well as our experiment and deployment videos are open sourced and can be found on our project website: <a target="_blank" href="https://robotutilitymodels.com">this https URL</a></p><p>You can also find <a target="_blank" href="https://robotutilitymodels.com/">the project page</a> to learn more.</p><p>If you want to learn more about scaling in robotics broadly, you can read my post on that.</p> <br/><br/>This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit <a href="https://itcanthink.substack.com?utm_medium=podcast&#38;utm_campaign=CTA_1">itcanthink.substack.com</a>]]></description><link>https://itcanthink.substack.com/p/robopapers-episode-2-robot-utility</link><guid isPermaLink="false">substack:post:160469832</guid><dc:creator><![CDATA[Chris Paxton and Michael Cho]]></dc:creator><pubDate>Thu, 03 Apr 2025 02:37:26 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/160469832/40a659dafe67d55913ac6bde49030fa7.mp3" length="64591977" type="audio/mpeg"/><itunes:author>Chris Paxton and Michael Cho</itunes:author><itunes:explicit>No</itunes:explicit><itunes:duration>4037</itunes:duration><itunes:image href="https://substackcdn.com/feed/podcast/2883266/post/160469832/880035721da36c6080d1f00d1c5137d8.jpg"/></item><item><title><![CDATA[RoboPapers Episode 1: SAM2Act with Jiafei Duan]]></title><description><![CDATA[<p>Join Chris Paxton & Michael Cho as we geek out over robotic papers with paper authors. First episode: Jiafei Duan talks about SAM2Act.</p><p>How do we build robots which can <em>remember</em> things and perform challenging, long horizon manipulation tasks with objects they haven't seen before?</p><p><a target="_blank" href="https://sam2act.github.io/">Project Website</a> <a target="_blank" href="https://arxiv.org/abs/2501.18564">ArXiV</a> <a target="_blank" href="https://www.youtube.com/watch?v=BpUThCmcklM">Youtube</a> <a target="_blank" href="https://open.spotify.com/episode/3ha2a96mOx4bvjEVFPkPxo?si=defe64c0613c4e47">Spotify</a></p><p><strong>Abstract:</strong></p><p>Robotic manipulation systems operating in diverse, dynamic environments must exhibit three critical abilities: multitask interaction, generalization to unseen scenarios, and spatial memory. While significant progress has been made in robotic manipulation, existing approaches often fall short in generalization to complex environmental variations and addressing memory-dependent tasks. To bridge this gap, we introduce <strong>SAM2Act</strong>, a multi-view robotic transformer-based policy that leverages multi-resolution upsampling with visual representations from large-scale foundation model. SAM2Act achieves a state-of-the-art average success rate of <strong>86.8% across 18 tasks</strong> in the RLBench benchmark, and demonstrates robust generalization on <em>The Colosseum</em> benchmark, with only a <strong>4.3% performance gap</strong> under diverse environmental perturbations. Building on this foundation, we propose <strong>SAM2Act+</strong>, a memory-based architecture inspired by SAM2, which incorporates a memory bank, an encoder, and an attention mechanism to enhance spatial memory. To address the need for evaluating memory-dependent tasks, we introduce <em>MemoryBench</em>, a novel benchmark designed to assess spatial memory and action recall in robotic manipulation. SAM2Act+ achieves <strong>competitive performance on </strong><strong><em>MemoryBench</em></strong>, significantly outperforming existing approaches and pushing the boundaries of memory-based robotic systems.</p> <br/><br/>This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit <a href="https://itcanthink.substack.com?utm_medium=podcast&#38;utm_campaign=CTA_1">itcanthink.substack.com</a>]]></description><link>https://itcanthink.substack.com/p/robopapers-episode-1-sam2act-with</link><guid isPermaLink="false">substack:post:159722954</guid><dc:creator><![CDATA[Chris Paxton and Michael Cho]]></dc:creator><pubDate>Tue, 25 Mar 2025 14:29:00 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/159722954/c6736cbfb61fbd7def22c5e92cca0010.mp3" length="66229123" type="audio/mpeg"/><itunes:author>Chris Paxton and Michael Cho</itunes:author><itunes:explicit>No</itunes:explicit><itunes:duration>4139</itunes:duration><itunes:image href="https://substackcdn.com/feed/podcast/2883266/post/159722954/880035721da36c6080d1f00d1c5137d8.jpg"/></item></channel></rss>