This workshop aims to present the latest results on the theory and practice of both traditional and modern techniques for robot learning, robot perception, and SLAM. It will be consist of a seminar and a SLAM challenge. A series of contributed and invited talks by academic leaders and renowned researchers will discuss ground-breaking perception and mapping methods for long-term autonomy based on current cutting-edge traditional solutions and modern learning methods. The workshop will also discuss the current challenges and future research directions and will include posters and spotlight talks to facilitate interaction between the speakers and the audience. We plan to have a hybrid format with in-person speakers/attendees and a live broadcast to convey the message to a broader audience.
Seminar
Session 1 (8:30-10:00 AM)
Presenter | Session Title | Time | YouTube Link | |
---|---|---|---|---|
Welcome message by organizers & overview of workshop |
8:30 - 8:40 AM |
|
||
![]() |
Postdoctoral Research Fellow, Biorobotics Lab Harvard University |
TDB |
8:40 - 9:00 AM |
|
![]() |
Professor of Mechanical and Ocean Engineering Massachusetts Institute of Technology |
TBD |
9:00 - 9:20 AM |
|
![]() |
Associate Professor in the Department of Aeronautics and Astronautics Massachusetts Institute of Technology |
From SLAM to Spatial Perception: Hierarchical Models, Certification, and Self-Supervision |
9:20 - 9:40 AM |
|
![]() |
Associate Research Professor, Robotics Institute Carnegie Mellon University |
TBD |
9:40 - 10:00 AM |
|
Coffee Break and Posters (10:00 - 10:25 AM)
Session 2 (10:25 AM - 12:00 PM)
Presenter | Session Title | Time | YouTube Link | |
---|---|---|---|---|
Spotlight Talk 1 |
10:25 - 10:40 AM |
|
||
![]() |
Professor University of Toronto |
TBD |
10:40 - 11:00 AM |
|
![]() |
Assistant Professor in Department of Computer Science and Engineering University of Buffalo |
TBD |
11:20 - 11:40 AM |
|
![]() |
Professor of Robot Vision Imperial College London |
TBD |
11:20 - 11:40 AM |
|
![]() |
Professor of Informatics and Mathematics Technical University of Munich |
TBD |
11:40 - 11:59 AM |
|
Lunck Break and Posters (12:00 - 13:00 PM)
Session 3 (1:00 - 3:20 PM)
Presenter | Session Title | Time | YouTube Link | |
---|---|---|---|---|
SubT-MRS SLAM Challenge Summary |
1:00 - 1:20 PM |
|
||
TartanAir Challenge Summary |
1:20 - 1:40 PM |
|
||
![]() |
Full Professor ETH Zurich |
TBD |
1:40 - 2:00 PM |
|
Spotlight Talk 2 |
2:00 - 2:20 PM |
|
||
![]() |
Assistant Professor, Robotics Institute Carnegie Mellon University |
TBD |
2:20 - 2:40 PM |
|
![]() |
Associate Professor Seoul National University |
TBD |
2:40 - 2:55 PM |
|
![]() |
Associate Professor in Robotics Institute Carnegie Mellon University |
TBD |
2:55 - 3:10 PM |
|
Concluding Remarks |
3:10 - 3:20 PM |
|
Challenge
Pushing SLAM Towards All-weather Environments
Robust odometry system is an indispensable need of autonomous robots operating navigation, exploration, and locomotion in unknown environments. In recent years, various robots are being deployed in increasingly complex environments for a broad spectrum of applications such as off-road driving, search-and-rescue in extreme environments, and robotic rovers on planetary missions. Despite the progress made, most of state estimation algorithms are still vulnerable in long-term operation and still struggle in these scenarios. A key necessity in progressing SLAM for long-term autonomy is the availability of high-quality datasets including various challenging scenerios.
To push the limits of robust SLAM and robust perception, we will organize a SLAM challenge and evaluate the performance from virtual to real world robotics. For virtual environments,
Multi Degradation: The dataset contains a broad set of perceptually degraded environments such as darkness, airbone obscurats conditions such as fog, dust, smoke and lack of prominent perceptual features in self-similar areas (Figure2).
- Multi Robots: The dataset is collected by various heterogeneous robots including aerial, wheeled and legged robots over multiple seasons. Most importantly, our dataset also provide the extrinsic and communication signal between robots which allows the map could be merged in the single world frame. These features are very important for resarcher to study multi agent SLAM.
- Multi Spectral: The dataset also contains different modalities not only visual, LiDAR, and inertial sensors but also the thermal data which is beyond the human eye.
- Multi Motion: The existing popular datasets such as KITTI and Cityscapes only covers very limited motion patterns, which are mostly moving straight forward plus small left or right turns. This regular motion is too simple to sufficiently test a visual SLAM algorithm. Our dataset covers much more diverse motion combinations in 3D space, which is significantly more difficult than existing datasets.
- Multi Dynamic: Our dataset contains dynamic objects including human, vehicles, dust and snow.
- Friendly to learning methods: Our dataset not only provide the benchmark for traditional method but also will provide the benchmark for learning based methods.
The dataset can be used for a number of visual tasks, including optical flow, visual odometry, lidar odometry, thermal odometry and multi agent odometry. Preliminary experiments show that methods performing well on established benchmarks such as KITTI does not achieve satisfactory results on SubT-MRS dataset. In this competition, we will focus on robustness and efficiency. We will provide an evaluation metric (same with the KITTI dataset) and evaluation website for comparison and submission.
The workshop has an associated new benchmark dataset (Subt-MRS datasets and TartanAir V2 datasets) that we will publish three months before the workshop.
01/June - 01/July Release Datasets Stage:
- Release the SubT-MRS Datasets and TartanAir V2 Datasets
01/July - 30/Sep Competition Start
- Organize Competition and Submission Results
01/Oct - 02/Oct Competition End
- 9:00 - 9:30 AM: Subt-MRS SLAM Challenge Summary
- 9:30 - 10:00 AM: TartanAir V2 SLAM Challenge Summary
Organizers & Committee
![]() Assistant Professor, Spatial AI & Robotics Lab University at Buffalo |
![]() PhD Candidate Carnegie Mellon University |
![]() Project Scientist, Robotics Institute Carnegie Mellon University |
![]() PhD Candidate Carnegie Mellon University |
![]() Postdoctoral Research Fellow, Biorobotics Lab Harvard University |
![]() Research Associate Professor, Robotics Institute Carnegie Mellon University |
![]() Robust Perception and Mobile Robotics Lab Seoul National University |
![]() Ph.D. student, Spatial AI & Robotics Lab University at Buffalo |
![]() Ph.D. student, Spatial AI & Robotics Lab University at Buffalo |
![]() Senior researcher, Autonomous Systems Research Team Microsoft |
![]() Undergradate Student Carnegie Mellon University |
![]() Undergradate Student Carnegie Mellon University |
![]() Research Associate Carnegie Mellon University |
![]() Research Associate Carnegie Mellon University |
![]() Undergraduate Student Carnegie Mellon University |
![]() Graduate Student Carnegie Mellon University </tr> |
![]() Research Associate Carnegie Mellon University |
![]() MSR Student Carnegie Mellon University |