ICCV'23 Workshop on Robot Learning and SLAM

Oct 02 - Oct 06, 2023, Paris

This workshop aims to present the latest results on the theory and practice of both traditional and modern techniques for robot learning, robot perception, and SLAM. It will be consist of a seminar and a SLAM challenge. A series of contributed and invited talks by academic leaders and renowned researchers will discuss ground-breaking perception and mapping methods for long-term autonomy based on current cutting-edge traditional solutions and modern learning methods. The workshop will also discuss the current challenges and future research directions and will include posters and spotlight talks to facilitate interaction between the speakers and the audience. We plan to have a hybrid format with in-person speakers/attendees and a live broadcast to convey the message to a broader audience.


  Join our Mailing List     Subscribe to the AirLab YouTube Channel     Subscribe to the SAIR Lab YouTube Channel     Join our Discord


Seminar

Session 1 (8:30-10:00 AM)

Presenter Session Title Time YouTube Link
Welcome message by organizers & overview of workshop

8:30 - 8:40 AM

Letizia Gionfrida

Postdoctoral Research Fellow, Biorobotics Lab

Harvard University

TDB

8:40 - 9:00 AM

John Leonard

Professor of Mechanical and Ocean Engineering

Massachusetts Institute of Technology

TBD

9:00 - 9:20 AM

Luca Carlone

Associate Professor in the Department of Aeronautics and Astronautics

Massachusetts Institute of Technology

From SLAM to Spatial Perception: Hierarchical Models, Certification, and Self-Supervision

9:20 - 9:40 AM

Sebastian Scherer

Associate Research Professor, Robotics Institute

Carnegie Mellon University

TBD

9:40 - 10:00 AM

Coffee Break and Posters (10:00 - 10:25 AM)

Session 2 (10:25 AM - 12:00 PM)

Presenter Session Title Time YouTube Link
Spotlight Talk 1

10:25 - 10:40 AM

Tim Barfoot

Professor

University of Toronto

TBD

10:40 - 11:00 AM

Chen Wang

Assistant Professor in Department of Computer Science and Engineering

University of Buffalo

TBD

11:20 - 11:40 AM

Andrew Davison

Professor of Robot Vision

Imperial College London

TBD

11:20 - 11:40 AM

Daniel Cremers

Professor of Informatics and Mathematics

Technical University of Munich

TBD

11:40 - 11:59 AM

Lunck Break and Posters (12:00 - 13:00 PM)

Session 3 (1:00 - 3:20 PM)

Presenter Session Title Time YouTube Link
SubT-MRS SLAM Challenge Summary

1:00 - 1:20 PM

TartanAir Challenge Summary

1:20 - 1:40 PM

Marc Pollefeys

Full Professor

ETH Zurich

TBD

1:40 - 2:00 PM

Spotlight Talk 2

2:00 - 2:20 PM

Shubham Tulsiani

Assistant Professor, Robotics Institute

Carnegie Mellon University

TBD

2:20 - 2:40 PM

Ayoung Kim

Associate Professor

Seoul National University

TBD

2:40 - 2:55 PM

Michael Kaess

Associate Professor in Robotics Institute

Carnegie Mellon University

TBD

2:55 - 3:10 PM

Concluding Remarks

3:10 - 3:20 PM

Challenge

Pushing SLAM Towards All-weather Environments

Robust odometry system is an indispensable need of autonomous robots operating navigation, exploration, and locomotion in unknown environments. In recent years, various robots are being deployed in increasingly complex environments for a broad spectrum of applications such as off-road driving, search-and-rescue in extreme environments, and robotic rovers on planetary missions. Despite the progress made, most of state estimation algorithms are still vulnerable in long-term operation and still struggle in these scenarios. A key necessity in progressing SLAM for long-term autonomy is the availability of high-quality datasets including various challenging scenerios.

To push the limits of robust SLAM and robust perception, we will organize a SLAM challenge and evaluate the performance from virtual to real world robotics. For virtual environments,

Multi Degradation: The dataset contains a broad set of perceptually degraded environments such as darkness, airbone obscurats conditions such as fog, dust, smoke and lack of prominent perceptual features in self-similar areas (Figure2).

  • Multi Robots: The dataset is collected by various heterogeneous robots including aerial, wheeled and legged robots over multiple seasons. Most importantly, our dataset also provide the extrinsic and communication signal between robots which allows the map could be merged in the single world frame. These features are very important for resarcher to study multi agent SLAM.
  • Multi Spectral: The dataset also contains different modalities not only visual, LiDAR, and inertial sensors but also the thermal data which is beyond the human eye.
  • Multi Motion: The existing popular datasets such as KITTI and Cityscapes only covers very limited motion patterns, which are mostly moving straight forward plus small left or right turns. This regular motion is too simple to sufficiently test a visual SLAM algorithm. Our dataset covers much more diverse motion combinations in 3D space, which is significantly more difficult than existing datasets.
  • Multi Dynamic: Our dataset contains dynamic objects including human, vehicles, dust and snow.
  • Friendly to learning methods: Our dataset not only provide the benchmark for traditional method but also will provide the benchmark for learning based methods.

The dataset can be used for a number of visual tasks, including optical flow, visual odometry, lidar odometry, thermal odometry and multi agent odometry. Preliminary experiments show that methods performing well on established benchmarks such as KITTI does not achieve satisfactory results on SubT-MRS dataset. In this competition, we will focus on robustness and efficiency. We will provide an evaluation metric (same with the KITTI dataset) and evaluation website for comparison and submission.

The workshop has an associated new benchmark dataset (Subt-MRS datasets and TartanAir V2 datasets) that we will publish three months before the workshop.

01/June - 01/July Release Datasets Stage:

  • Release the SubT-MRS Datasets and TartanAir V2 Datasets

01/July - 30/Sep Competition Start

  • Organize Competition and Submission Results

01/Oct - 02/Oct Competition End

  • 9:00 - 9:30 AM: Subt-MRS SLAM Challenge Summary
  • 9:30 - 10:00 AM: TartanAir V2 SLAM Challenge Summary

Organizers & Committee


</td> </table> ## Workshop Registration

Chen Wang

Assistant Professor, Spatial AI & Robotics Lab

University at Buffalo

Shibo Zhao

PhD Candidate

Carnegie Mellon University

Wenshan Wang

Project Scientist, Robotics Institute

Carnegie Mellon University

Yuheng Qiu

PhD Candidate

Carnegie Mellon University

Letizia Gionfrida

Postdoctoral Research Fellow, Biorobotics Lab

Harvard University

Sebastian Scherer

Research Associate Professor, Robotics Institute

Carnegie Mellon University

Ayoung Kim

Robust Perception and Mobile Robotics Lab

Seoul National University

Yi Du

Ph.D. student, Spatial AI & Robotics Lab

University at Buffalo

Shaoshu Su

Ph.D. student, Spatial AI & Robotics Lab

University at Buffalo

Shuang Ma

Senior researcher, Autonomous Systems Research Team

Microsoft

Rushan Jiang

Undergradate Student

Carnegie Mellon University

Haoxiang Sun

Undergradate Student

Carnegie Mellon University

Yao He

Research Associate

Carnegie Mellon University

Tian Hao Wu

Research Associate

Carnegie Mellon University

Tingting Da

Undergraduate Student

Carnegie Mellon University

Damanpreet Singh

Graduate Student

Carnegie Mellon University

</tr>

YuanJun Gao

Research Associate

Carnegie Mellon University

Jay Karhade

MSR Student

Carnegie Mellon University