You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Our pipeline utilizes advanced <strong>neural rendering with a reactive environment</strong> to generate high-fidelity multi-view observations controlled by the perturbed ego trajectory.
256
258
Furthermore, we develop a <strong>pseudo-expert trajectory generation</strong> mechanism for these newly simulated states to provide action supervision.
257
-
Upon the synthesized data, we find that a simple <strong>co-training strategy on both real-world and simulated samples</strong> can lead to significant improvements in both robustness and generalization for various planning methods on challenging real-world benchmarks, up to <strong>+6.8 EPDMS on navhard and +2.9 on navtest</strong>.
259
+
Upon the synthesized data, we find that a simple <strong>co-training strategy on both real-world and simulated samples</strong> can lead to significant improvements in both robustness and generalization for various planning methods on challenging real-world benchmarks, up to <strong>+8.6 EPDMS on navhard and +2.9 on navtest</strong>.
258
260
More importantly, such policy improvement scales smoothly by increasing simulation data only, even without extra real-world data streaming in.
259
261
We further reveal several crucial findings of such a sim-real learning system, which we term SimScale, including the design of pseudo-experts and the scaling properties for different policy architectures.
0 commit comments