Skip to main content

Showing 1–8 of 8 results for author: Glossop, C

Searching in archive cs. Search in all archives.
.
  1. arXiv:2511.14759  [pdf, ps, other

    cs.LG cs.RO

    $π^{*}_{0.6}$: a VLA That Learns From Experience

    Authors: Physical Intelligence, Ali Amin, Raichelle Aniceto, Ashwin Balakrishna, Kevin Black, Ken Conley, Grace Connors, James Darpinian, Karan Dhabalia, Jared DiCarlo, Danny Driess, Michael Equi, Adnan Esmail, Yunhao Fang, Chelsea Finn, Catherine Glossop, Thomas Godden, Ivan Goryachev, Lachy Groom, Hunter Hancock, Karol Hausman, Gashon Hussein, Brian Ichter, Szymon Jakubczak, Rowan Jen , et al. (31 additional authors not shown)

    Abstract: We study how vision-language-action (VLA) models can improve through real-world deployments via reinforcement learning (RL). We present a general-purpose method, RL with Experience and Corrections via Advantage-conditioned Policies (RECAP), that provides for RL training of VLAs via advantage conditioning. Our method incorporates heterogeneous data into the self-improvement process, including demon… ▽ More

    Submitted 18 November, 2025; v1 submitted 18 November, 2025; originally announced November 2025.

  2. arXiv:2509.19480  [pdf, ps, other

    cs.RO cs.LG

    OmniVLA: An Omni-Modal Vision-Language-Action Model for Robot Navigation

    Authors: Noriaki Hirose, Catherine Glossop, Dhruv Shah, Sergey Levine

    Abstract: Humans can flexibly interpret and compose different goal specifications, such as language instructions, spatial coordinates, or visual references, when navigating to a destination. In contrast, most existing robotic navigation policies are trained on a single modality, limiting their adaptability to real-world scenarios where different forms of goal specification are natural and complementary. In… ▽ More

    Submitted 23 September, 2025; originally announced September 2025.

    Comments: 9 pages, 7 figures, 6 tables

  3. arXiv:2508.13446  [pdf, ps, other

    cs.RO

    CAST: Counterfactual Labels Improve Instruction Following in Vision-Language-Action Models

    Authors: Catherine Glossop, William Chen, Arjun Bhorkar, Dhruv Shah, Sergey Levine

    Abstract: Generalist robots should be able to understand and follow user instructions, but current vision-language-action (VLA) models struggle with following fine-grained commands despite providing a powerful architecture for mapping open-vocabulary natural language instructions to robot actions. One cause for this is a lack of semantic diversity and language grounding in existing robot datasets and, speci… ▽ More

    Submitted 18 August, 2025; originally announced August 2025.

  4. arXiv:2505.05592  [pdf, ps, other

    cs.RO cs.CV cs.LG eess.SY

    Learning to Drive Anywhere with Model-Based Reannotation

    Authors: Noriaki Hirose, Lydia Ignatova, Kyle Stachowicz, Catherine Glossop, Sergey Levine, Dhruv Shah

    Abstract: Developing broadly generalizable visual navigation policies for robots is a significant challenge, primarily constrained by the availability of large-scale, diverse training data. While curated datasets collected by researchers offer high quality, their limited size restricts policy generalization. To overcome this, we explore leveraging abundant, passively collected data sources, including large… ▽ More

    Submitted 21 November, 2025; v1 submitted 8 May, 2025; originally announced May 2025.

    Comments: 9 pages, 8 figures, 6 tables

    Journal ref: IEEE Robotics and Automation Letters 2025

  5. arXiv:2410.03603  [pdf, other

    cs.RO

    LeLaN: Learning A Language-Conditioned Navigation Policy from In-the-Wild Videos

    Authors: Noriaki Hirose, Catherine Glossop, Ajay Sridhar, Dhruv Shah, Oier Mees, Sergey Levine

    Abstract: The world is filled with a wide variety of objects. For robots to be useful, they need the ability to find arbitrary objects described by people. In this paper, we present LeLaN(Learning Language-conditioned Navigation policy), a novel approach that consumes unlabeled, action-free egocentric data to learn scalable, language-conditioned object navigation. Our framework, LeLaN leverages the semantic… ▽ More

    Submitted 4 October, 2024; originally announced October 2024.

    Comments: 23 pages, 9 figures, 5 tables, Conference on Robot Learning 2024

  6. arXiv:2402.19432  [pdf, other

    cs.RO

    Pushing the Limits of Cross-Embodiment Learning for Manipulation and Navigation

    Authors: Jonathan Yang, Catherine Glossop, Arjun Bhorkar, Dhruv Shah, Quan Vuong, Chelsea Finn, Dorsa Sadigh, Sergey Levine

    Abstract: Recent years in robotics and imitation learning have shown remarkable progress in training large-scale foundation models by leveraging data across a multitude of embodiments. The success of such policies might lead us to wonder: just how diverse can the robots in the training set be while still facilitating positive transfer? In this work, we study this question in the context of heterogeneous emb… ▽ More

    Submitted 29 February, 2024; originally announced February 2024.

    Comments: 16 pages, 9 figures

    MSC Class: 68T40 ACM Class: I.2.9

  7. arXiv:2310.07896  [pdf, other

    cs.RO cs.CV cs.LG

    NoMaD: Goal Masked Diffusion Policies for Navigation and Exploration

    Authors: Ajay Sridhar, Dhruv Shah, Catherine Glossop, Sergey Levine

    Abstract: Robotic learning for navigation in unfamiliar environments needs to provide policies for both task-oriented navigation (i.e., reaching a goal that the robot has located), and task-agnostic exploration (i.e., searching for a goal in a novel setting). Typically, these roles are handled by separate models, for example by using subgoal proposals, planning, or separate navigation strategies. In this pa… ▽ More

    Submitted 11 October, 2023; originally announced October 2023.

    Comments: Project page https://general-navigation-models.github.io/nomad/

  8. arXiv:2210.15199  [pdf, other

    cs.RO

    Characterising the Robustness of Reinforcement Learning for Continuous Control using Disturbance Injection

    Authors: Catherine R. Glossop, Jacopo Panerati, Amrit Krishnan, Zhaocong Yuan, Angela P. Schoellig

    Abstract: In this study, we leverage the deliberate and systematic fault-injection capabilities of an open-source benchmark suite to perform a series of experiments on state-of-the-art deep and robust reinforcement learning algorithms. We aim to benchmark robustness in the context of continuous action spaces -- crucial for deployment in robot control. We find that robustness is more prominent for action dis… ▽ More

    Submitted 27 October, 2022; originally announced October 2022.

    Comments: 18 pages, 15 figures