RoboSandbox¶
A sim-first sandbox for robot manipulation. Bring your own arm, objects, and tasks.
Platform support
macOS (Apple Silicon and Intel) works out of the box — no GL configuration needed.
Linux (Ubuntu 22.04/24.04) is the CI-tested platform; headless rendering needs one apt-get line (see Quickstart).
Windows is not directly supported; WSL2 running Ubuntu 22.04 works.
RoboSandbox is a small manipulation sandbox for building and testing manipulation loops without needing a giant stack around them. Load a robot, drop in a few objects, define a task, run a planner or policy, and record the result. If you want to export episodes for policy work later, that is part of the same flow.
Why This Project Exists¶
RoboSandbox is a small manipulation sandbox for learning, prototyping, and integration work.
It exists for the gap between toy demos and heavyweight robotics stacks. You can bring in a robot, define a task, run a planner or policy, record the result, and inspect the interfaces without a large simulator setup.
The project is deliberately scoped. It is meant to help you understand the workflow, try ideas quickly, and make the seams between robot, task, skills, recorder, and policy visible. It is not trying to be the final simulator you use forever.
If you outgrow RoboSandbox and move to MuJoCo, Isaac Sim, LeRobot training pipelines, or real hardware, that is success, not failure.
Who This Is For¶
RoboSandbox is a good fit if you are:
- learning how a manipulation stack fits together
- doing robotics work but want a lighter-weight way to prototype in simulation
- already comfortable with simulation and need a small, hackable integration harness
It is especially useful when you want to answer questions like:
- How do I add a new robot?
- How do I describe a task?
- What does a policy need to consume and emit?
- What gets recorded and exported?
- What breaks when I swap embodiments?
When To Move Beyond It¶
RoboSandbox is a starting point, not an end state.
You may want to move beyond it when you need:
- lower-level simulator control than the current abstractions expose
- photorealism or richer sensor simulation
- large-scale industrial workflows
- large scenes or more complex multi-robot environments
- production deployment infrastructure
The intended path is simple: start small, understand the workflow, validate the seams, then move to a heavier stack when your requirements become sharper.
user: "pick up the red cube and put it on the green cube"
│
▼
planner ─► [pick(red_cube), place_on(green_cube)]
│
▼
perception (VLM or ground truth) locates both in 3D
│
▼
motion (DLS Jacobian IK + Cartesian interpolation) executes
│
▼
recorder writes runs/<id>/video.mp4 + events.jsonl
Get started¶
Open the browser viewer — no API key, no model download.
Pick a task, type a command like pick up the red cube, click Run.
The arm plans and executes while frames stream to your browser. Hit
Record before running to save the episode to disk for training.
The built-in planner understands a small grammar — pick, place, push, pour, stack, open/close drawer, go home — which is enough to exercise the full agent loop. See the Quickstart for the install steps and the record → export → train flow.
Want richer natural language? Plug in a VLM for free-form commands and visual scene reasoning:
Both use the same agent loop — only the planner changes.
Where to go next¶
- Quickstart — install, open the viewer, record an episode. 5 minutes end-to-end.
- Start here if you want the product thesis first — Why RoboSandbox exists explains the robot loop, where modern model families fit, and what problem RoboSandbox is actually trying to solve.
- Concepts — Scenes & objects, Skills & agents, Perception & grasping, Recording & export, Real-robot bridge.
- Tutorials — Custom arm, Custom task, Custom skill, Policy replay.
- Reference — CLI, API, roadmap.
What ships in v0.1¶
- MuJoCo physics backend + built-in 6-DOF arm + bundled Franka Panda (URDF import path).
- 9 skills:
pick,place_on,push,home,pour,tap,open_drawer,close_drawer,stack. - 8 default benchmark tasks + 1 experimental, including a long-horizon
pour_can_into_bowland an articulated-drawer primitive. - 10 bundled YCB objects, drop-in via
@ycb:<id>. - Browser live viewer with record + keyboard teleop.
- LeRobot v3 parquet export + policy replay loop.
- Real-robot bridge stub — subclass, fill the hardware driver;
observation+step skills (
Home, teleop, policy rollouts) carry over unchanged. Motion-planning skills (Pick,PlaceOn,Push) still depend on MuJoCo kinematics — see the sim-to-real handoff tutorial.
See the roadmap for what is coming next.