Important
This repository has moved from bdaiinstitute/lottery_tickets to rai-opensource/lottery_tickets.
| Original Policy (i.e: Initial noise sampled from gaussian) | 🎫 Golden Ticket (i.e: Optimized fixed initial noise) | |
|---|---|---|
|
|
|
|
|
|
|
|
|
| (left) Baseline policy (Gaussian sampling) vs. (right) 🎫 golden-ticket policy using a fixed initial noise. (top) is frankasim, (middle) is 🤗 SmolVLA + LIBERO, (bottom) is DPPO + robomimic. Each row uses a different golden ticket that was optimized for that model. | ||
This is a repository for testing the lottery ticket hypothesis for robot control: The performance of a pretrained, frozen diffusion/flow matching policy can be improved by replacing sampling initial noise from the prior distribution (typically isotropic gaussian) with a well-chosen, constant initial noise input, which we call a golden ticket. There are three different experimental setups, where each experiment uses a unique simulation and policy class:
- franka-sim cube picking with state-based flow matching policies
- 🤗 LeRobot pretrained 🤗SmolVLA for LIBERO
- DPPO + robomimic
All three experiment setups contain their own READMEs and code for running a baseline policy, generating tickets, evaluating tickets, and links to golden tickets we have found so you can try them yourself. Each subfolder may contain other utilities, since each experiment testbed serves a different purpose:
🦾 Franka-sim involves a cube picking task with a Franka robot, where the cube randomly spawns in a ~1/2 square meter region in front of the robot. Our codebase includes an automated way to generate demonstrations, training code for behavior cloning with a flow matching policy on the collected data, and model checkpoints of policies we have already trained. We also include golden tickets for the checkpoints we provide. This is a great experimental testbed if you'd like to examine all parts of a pipeline (data collection, policy training, and inference) that result in policies with golden tickets. The small model makes it easier to do experiments with little compute. The policy and training code is all custom-written.
🤗 SmolVLA + Libero represents an experiment where a pretrained VLA checkpoint is taken (directly from LeRobot), and golden tickets are searched for over a multitude of task suites. We also include golden tickets we have found which can be evaluated. This is a good experimental testbed for examining lottery tickets with an open-source VLA, and on a multi-task setting. The policy used in our experiments comes from an off-the-shelf LIBERO checkpoint from LeRobot, so this reflects looking for lottery tickets in a model we didn't create.
✨ DPPO for robomimic includes the original DPPO robomimic checkpoints used in the DSRL project. We provide golden tickets for these policies, and code for generating new tickets and comparing against the base policy. This reflects a setting of using a model we didn't create.
We use some features of the uv package manager in our pyproject.toml.
The easiest way to get started is to install uv using these instructions.
You can then install the individual experiment setups using uv sync --extra $EXPERIMENT_NAME; see the individual READMEs for more details.
This repository is released as-is to accompany a paper submission.
If you find any bugs, corrections, or issues that should be resolved for anyone looking to reproduce the results in this repository, please file an issue and we will look at it as soon as we can.
For other improvements, including new features, we recommend creating your own fork of the repository.
@article{patil2026you,
title={You've Got a Golden Ticket: Improving Generative Robot Policies With A Single Noise Vector},
author={Patil, Omkar and Biza, Ondrej and Weng, Thomas and Schmeckpeper, Karl and Thomason, Wil and Zhang, Xiaohan and Walters, Robin and Gopalan, Nakul and Castro, Sebastian and Rosen, Eric},
journal={arXiv preprint arXiv:2603.15757},
year={2026}





