Skip to main content

Showing 1–9 of 9 results for author: Torne, M

Searching in archive cs. Search in all archives.
.
  1. arXiv:2512.16881  [pdf, ps, other

    cs.RO cs.LG

    PolaRiS: Scalable Real-to-Sim Evaluations for Generalist Robot Policies

    Authors: Arhan Jain, Mingtong Zhang, Kanav Arora, William Chen, Marcel Torne, Muhammad Zubair Irshad, Sergey Zakharov, Yue Wang, Sergey Levine, Chelsea Finn, Wei-Chiu Ma, Dhruv Shah, Abhishek Gupta, Karl Pertsch

    Abstract: A significant challenge for robot learning research is our ability to accurately measure and compare the performance of robot policies. Benchmarking in robotics is historically challenging due to the stochasticity, reproducibility, and time-consuming nature of real-world rollouts. This challenge is exacerbated for recent generalist policies, which has to be evaluated across a wide variety of scene… ▽ More

    Submitted 18 December, 2025; originally announced December 2025.

    Comments: Website: https://polaris-evals.github.io/

  2. arXiv:2511.19750  [pdf, ps, other

    cs.LG

    DISCO: A Browser-Based Privacy-Preserving Framework for Distributed Collaborative Learning

    Authors: Julien T. T. Vignoud, Valérian Rousset, Hugo El Guedj, Ignacio Aleman, Walid Bennaceur, Batuhan Faik Derinbay, Eduard Ďurech, Damien Gengler, Lucas Giordano, Felix Grimberg, Franziska Lippoldt, Christina Kopidaki, Jiafan Liu, Lauris Lopata, Nathan Maire, Paul Mansat, Martin Milenkoski, Emmanuel Omont, Güneş Özgün, Mina Petrović, Francesco Posa, Morgan Ridel, Giorgio Savini, Marcel Torne, Lucas Trognon , et al. (6 additional authors not shown)

    Abstract: Data is often impractical to share for a range of well considered reasons, such as concerns over privacy, intellectual property, and legal constraints. This not only fragments the statistical power of predictive models, but creates an accessibility bias, where accuracy becomes inequitably distributed to those who have the resources to overcome these concerns. We present DISCO: an open-source DIStr… ▽ More

    Submitted 24 November, 2025; originally announced November 2025.

  3. arXiv:2511.14759  [pdf, ps, other

    cs.LG cs.RO

    $π^{*}_{0.6}$: a VLA That Learns From Experience

    Authors: Physical Intelligence, Ali Amin, Raichelle Aniceto, Ashwin Balakrishna, Kevin Black, Ken Conley, Grace Connors, James Darpinian, Karan Dhabalia, Jared DiCarlo, Danny Driess, Michael Equi, Adnan Esmail, Yunhao Fang, Chelsea Finn, Catherine Glossop, Thomas Godden, Ivan Goryachev, Lachy Groom, Hunter Hancock, Karol Hausman, Gashon Hussein, Brian Ichter, Szymon Jakubczak, Rowan Jen , et al. (31 additional authors not shown)

    Abstract: We study how vision-language-action (VLA) models can improve through real-world deployments via reinforcement learning (RL). We present a general-purpose method, RL with Experience and Corrections via Advantage-conditioned Policies (RECAP), that provides for RL training of VLAs via advantage conditioning. Our method incorporates heterogeneous data into the self-improvement process, including demon… ▽ More

    Submitted 18 November, 2025; v1 submitted 18 November, 2025; originally announced November 2025.

  4. arXiv:2505.09561  [pdf, ps, other

    cs.RO cs.AI cs.LG

    Learning Long-Context Diffusion Policies via Past-Token Prediction

    Authors: Marcel Torne, Andy Tang, Yuejiang Liu, Chelsea Finn

    Abstract: Reasoning over long sequences of observations and actions is essential for many robotic tasks. Yet, learning effective long-context policies from demonstrations remains challenging. As context length increases, training becomes increasingly expensive due to rising memory demands, and policy performance often degrades as a result of spurious correlations. Recent methods typically sidestep these iss… ▽ More

    Submitted 19 May, 2025; v1 submitted 14 May, 2025; originally announced May 2025.

    Comments: Videos are available at https://long-context-dp.github.io

  5. arXiv:2412.01770  [pdf, ps, other

    cs.RO cs.AI cs.LG

    Robot Learning with Super-Linear Scaling

    Authors: Marcel Torne, Arhan Jain, Jiayi Yuan, Vidaaranya Macha, Lars Ankile, Anthony Simeonov, Pulkit Agrawal, Abhishek Gupta

    Abstract: Scaling robot learning requires data collection pipelines that scale favorably with human effort. In this work, we propose Crowdsourcing and Amortizing Human Effort for Real-to-Sim-to-Real(CASHER), a pipeline for scaling up data collection and learning in simulation where the performance scales superlinearly with human effort. The key idea is to crowdsource digital twins of real-world scenes using… ▽ More

    Submitted 11 October, 2025; v1 submitted 2 December, 2024; originally announced December 2024.

  6. arXiv:2407.16677  [pdf, other

    cs.RO cs.LG

    From Imitation to Refinement -- Residual RL for Precise Assembly

    Authors: Lars Ankile, Anthony Simeonov, Idan Shenfeld, Marcel Torne, Pulkit Agrawal

    Abstract: Recent advances in Behavior Cloning (BC) have made it easy to teach robots new tasks. However, we find that the ease of teaching comes at the cost of unreliable performance that saturates with increasing data for tasks requiring precision. The performance saturation can be attributed to two critical factors: (a) distribution shift resulting from the use of offline data and (b) the lack of closed-l… ▽ More

    Submitted 12 December, 2024; v1 submitted 23 July, 2024; originally announced July 2024.

    Comments: Project website: https://residual-assembly.github.io

  7. arXiv:2403.03949  [pdf, other

    cs.RO cs.AI cs.LG

    Reconciling Reality through Simulation: A Real-to-Sim-to-Real Approach for Robust Manipulation

    Authors: Marcel Torne, Anthony Simeonov, Zechu Li, April Chan, Tao Chen, Abhishek Gupta, Pulkit Agrawal

    Abstract: Imitation learning methods need significant human supervision to learn policies robust to changes in object poses, physical disturbances, and visual distractors. Reinforcement learning, on the other hand, can explore the environment autonomously to learn robust behaviors but may require impractical amounts of unsafe real-world data collection. To learn performant, robust policies without the burde… ▽ More

    Submitted 23 November, 2024; v1 submitted 6 March, 2024; originally announced March 2024.

    Comments: Project page: https://real-to-sim-to-real.github.io/RialTo/

  8. arXiv:2310.20608  [pdf, other

    cs.LG cs.AI cs.RO

    Autonomous Robotic Reinforcement Learning with Asynchronous Human Feedback

    Authors: Max Balsells, Marcel Torne, Zihan Wang, Samedh Desai, Pulkit Agrawal, Abhishek Gupta

    Abstract: Ideally, we would place a robot in a real-world environment and leave it there improving on its own by gathering more experience autonomously. However, algorithms for autonomous robotic learning have been challenging to realize in the real world. While this has often been attributed to the challenge of sample complexity, even sample-efficient techniques are hampered by two major challenges - the d… ▽ More

    Submitted 31 October, 2023; originally announced October 2023.

    Comments: Project website https://guided-exploration-autonomous-rl.github.io/GEAR/

  9. arXiv:2307.11049  [pdf, other

    cs.LG cs.AI cs.RO

    Breadcrumbs to the Goal: Goal-Conditioned Exploration from Human-in-the-Loop Feedback

    Authors: Marcel Torne, Max Balsells, Zihan Wang, Samedh Desai, Tao Chen, Pulkit Agrawal, Abhishek Gupta

    Abstract: Exploration and reward specification are fundamental and intertwined challenges for reinforcement learning. Solving sequential decision-making tasks requiring expansive exploration requires either careful design of reward functions or the use of novelty-seeking exploration bonuses. Human supervisors can provide effective guidance in the loop to direct the exploration process, but prior methods to… ▽ More

    Submitted 20 July, 2023; originally announced July 2023.