Skip to yearly menu bar Skip to main content


Poster
in
Workshop: 6th Robot Learning Workshop: Pretraining, Fine-Tuning, and Generalization with Large Scale Models

World Model Based Sim2Real Transfer for Visual Navigation

Kiran Lekkala · Chen Liu · Laurent Itti

Keywords: [ World Models ] [ contrastive learning ] [ Visual Navigation ] [ Sim2Real ]


Abstract:

Sim2Real transfer has gained popularity because it helps transfer from inexpensive simulators to real world. This paper presents a novel system that fuses components in a traditional World Model into a robust system, trained entirely within a simulator, that Zero-Shot transfers to the real world. To facilitate transfer, we use an intermediary representation that are based on Bird's Eye View (BEV) images. Thus, our robot learns to navigate in a simulator by first learning to translate from complex First-Person View (FPV) based RGB images to BEV representations, then learning to navigate using those representations. Later, when tested in the real world, the robot uses the perception model that translates FPV-based RGB images to embeddings that are used by the downstream policy.The incorporation of state-checking modules using Anchor images and Mixture Density LSTM not only interpolates uncertain and missing observations but also enhances the robustness of the model when exposed to the real-world environment. We trained the model using data collected using a Differential-drive robot in the CARLA simulator. Our methodology's effectiveness is shown through the deployment of trained models onto a Real-world Differential-drive robot. Lastly we release a comprehensive codebase, dataset and models for training and deployment that are available to the public.

Chat is not available.