You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Traditional robotics is built around hardware, with many interacting parts and specialized AI modules.
230
+
With machine learning taking the lead, this relationship flips around: <em> robots are components of a machine learning pipeline</em>.
231
+
</p>
232
+
233
+
<br>
234
+
<p>
235
+
Many libraries embrace this and adopt a Python- and ML-first approach, but they often lack robust robotics features and hardware support.
236
+
Robust policies require careful debugging in both simulation and hardware, which relies on classical robotics tools.
237
+
</p>
238
+
239
+
<br>
240
+
<p>
241
+
RCS bridges this gap by combining an ML-first design with the essential robotics tools.
242
+
It gives you the means to debug interfaces, validate tasks, and test directly on hardware—while remaining a lightweight pip-installable package with minimal dependencies.
We provide device APIs in C++ with automatically generated Python bindings, ensuring mirrored functionality in both languages.
261
+
A new device can be integrated into RCS in either C++ or in Python, ensuring broad hardware compatibility.
262
+
</p>
263
+
<br>
264
+
<p>
265
+
<b>Composable scenes</b>
266
+
Higher-level abstractions are built on top of our own device APIs.
267
+
They leverage Gymnasium wrappers to enable modular scene creation through composition.
268
+
</p>
269
+
<br>
270
+
<p>
271
+
<b>Layered architecture</b>
272
+
Because we build upon a minimal low-level device API, you can quickly get up and running with new hardware: implement our interface, benefit from all the wrappers and apps higher up in the stack.
<h2class="title is-3 has-text-centered">Robots Setups with Digital Twins</h2>
295
+
<h2class="title is-3 has-text-centered">Robot Setups with Digital Twins</h2>
296
+
<p>
297
+
We evaluate the usability of RCS's hardware oriented features by integrating multiple setups with different robots, grippers, cameras and touch sensors.
298
+
In total, four robots, four end-effectors, two cameras and a tactile sensor are implemented, both in simulation and on physical hardware.
299
+
</p>
300
+
<br>
262
301
263
302
<style>
264
303
/* minimal CSS */
@@ -322,14 +361,21 @@ <h2 class="title is-3 has-text-centered">Robots Setups with Digital Twins</h2>
<p>We demonstrate how RCS supports VLA research by investigating VLA generalization across multiple embodiments and assessing the benefit of simulated data for robotic foundation models.
Fig. 2: Success rate plots of different VLA comparisons.
458
-
<i>Left:</i>
459
-
The Pi Zero model fine-tuned on four datasets from different setups.
460
-
Each fine-tuning dataset contains of less then 150 episodes and each model is evaluated on 50 rollouts.
461
-
<i>Center:</i>
462
-
Different models fine-tuned on 143 episodes on our FR3 setup (real) with a down-sampled frequency of 5Hz
463
-
and evaluated on the real-world setup and the replicated simulated scene on 30 real-world and 100
464
-
simulated rollouts.
465
-
<i>Bottom:</i>
466
-
Different data mixes of synthetic and real data evaluated on the real-world setup and the simulated
467
-
scene on 30 real-world and 100 simulated rollouts.
468
-
The number denotes the amount of episodes from the respective domain used in the training mix.
506
+
Fig. 2: We fine-tune Pi Zero on four datasets from different setups.
507
+
Each dataset contains fewer than 150 episodes.
508
+
The fine-tuned models are deployed on the corresponding setups.
509
+
The robots that are more prominent in the base model's data mix achieve better results.
469
510
</div>
470
511
<imgsrc="static/images/results/sim_real_eval.svg"
471
512
alt="Success rate plot over training checkpoints." width="100%">
472
513
<divclass="content has-text-justified">
473
-
Fig. 3: Evaluation success rates measured for each checkpoint throughout the training process in the
474
-
real and replicated simulated domain. Each checkpoint is evaluated on 20 real and 100 simulated
475
-
rollouts. Left: Trained on 143 episodes on our FR3 dataset. Right: Trained on a mix of 143 episodes from
476
-
our FR3 dataset and 500 episodes from the scripted dataset of the replicated simulated domain.
514
+
Fig. 3: We investigate the impact of simulated data on VLA performance.
515
+
Our setup is replicated in simulation and used to generate 500 trajectories using a scripted policy, which is then used to complement our manually collected dataset of 143 trajectories.
516
+
The plots show the success rate of the policy, both in the simulated scene and on the hardware, as training progresses.
517
+
Success rates in simulation correlate with success rates on the physical robot—consistent with a good evaluation metric.
518
+
Adding simulated data to the training mix improves performance in both settings.
0 commit comments