Unifying Visual SLAM: From Fragmented Datasets to Scalable, Real-World Solutions

Workshop at Robotics: Science and Systems 2025, Los Angeles, California, USA

Saturday, June 21


News

  • Mar 24, 2025 - We are excited to host the first workshop on Unifying Visual SLAM at RSS 2025!

Overview

Is Visual SLAM Tackling the Right Challenges? While fields like image segmentation and reinforcement learning have thrived by leveraging vast amounts of unstructured data, visual SLAM remains constrained by a handful of curated benchmarks. This limitation hinders the development of scalable, robust systems capable of operating in a wide range of complex, real-world environments.

Fragmentation across datasets, pipelines, and evaluation metrics remains a core obstacle. Each dataset follows its own structure, making reproducibility and benchmarking difficult. Researchers frequently spend time adapting to dataset-specific requirements instead of driving fundamental advancements. Additionally, compiling baseline systems is a challenge due to the lack of standardized pipelines, undocumented dataset-specific issues, and inconsistent failure cases. This fragmentation significantly slows down progress in the field.

To move forward, SLAM needs unified dataset formats, common pipelines for baselines, and evaluation metrics that go beyond traditional ATE. Standardization will enable scalable benchmarking and the development of more generalizable SLAM systems that function reliably in real-world scenarios.

Workshop Goal: This workshop aims to bring researchers together to discuss best practices, establish dataset and pipeline standards, and streamline SLAM development. We have invited experts who have created benchmarks that foster SLAM challenges, unify SLAM practices, and develop tools that support not only SLAM but broader robotic applications.

One key output of the workshop will be a curated “Unifying Visual SLAM” list of development tools, datasets, pipelines, and benchmarks—compiled by organizers, speakers, and attendees—to serve as a future reference for the research community. By reducing implementation overhead, improving reproducibility, increasing the amount and diversity of benchmarks and fostering collaboration, this workshop seeks to advance SLAM’s ability to process large-scale data and build more scalable, real-world solutions for robotics and computer vision.


Invited Speakers

Hermann Blum

Hermann Blum

University of Bonn & Lamarr Institute

Angela Dai

Angela Dai

Technical University of Munich

Tobias Fischer

Tobias Fischer

Queensland University of Technology

Wenshan Wang

Wenshan Wang

Carnegie Mellon University

Luigi Freda

Luigi Freda

Dexory

Shehryar Khattak

Shehryar Khattak

NASA Jet Propulsion Lab


Call for Papers

We invite short papers of novel or recently published research relevant to the topics of the workshop.

  • Short papers are 2+n pages (2 pages content + references)
  • Submissions must follow the IEEE Conference double column format
  • All accepted papers will be presented as posters at the workshop and published on the workshop website.
  • Please indicate whether your paper falls into the ‘novel’ or ‘previously published’ category. Novel research papers are encouraged and can expect more substantial review feedback on their work. This is provided as a service to authors of novel papers and does not diminish the chance of acceptance.
  • All accepted submissions will be considered for the best presentation award, where 3 finalists will be selected for 5-minute plenary presentations. While all submissions are eligible, novelty will be considered in finalist selection.
  • Submissions are single blind and will be reviewed by members of the (extended) workshop committee.
  • Submissions can optionally be accompanied by a video.
  • All accepted submissions will be considered for three awards: Best Paper, Best Poster, and Best Open-Source Contribution, each receiving a 150 USD gift card.

Invited topics

We invite submissions including, but not limited to:

  • SLAM Systems and Applications: Novel approaches in SLAM, Visual Odometry, 3D Reconstruction, and Visual Place Recognition (VPR), with applications in robotics, autonomous vehicles, AR/VR, and more.
  • Tool Integration and Standardization: Open-source tools and libraries for SLAM, improving baseline evaluations, and streamlining benchmarking processes.
  • Benchmarking and Reproducibility: Standardization of datasets and evaluation methodologies, addressing challenges such as low-light conditions, harsh weather, and dynamic environments.
  • New Evaluation Metrics and Benchmarking Tools: Methods and platforms to assess localization and mapping performance more effectively.
  • Future Directions and Industrial Applications: The role of AI and machine learning in SLAM, and real-world applications in industry, commercial robotics, and emerging technologies.

Submissions Portal

A submission portal will open via CMT on April 7.

Submissions Timeline

April 7 Call for submissions
May 5 Submissions due
June 2 Notification of acceptance
June 21 Workshop at RSS!

Schedule

Time Planned Event Comments
08:00 Opening Remarks Organizing Committee
08:05 PySLAM and SlamPlay Luigi Freda
08:30 ROS2WASM: Bringing the Robot Operating System to the Web Tobias Fischer
09:00 Isaac ROS Visual SLAM Tomasz Bednarz
09:30 Tartanair and Subt-mrs datasets to push the limits of visual SLAM Wenshan Wang
10:00 Poster Session/Coffee Break
10:30 Scannet++: A high-fidelity dataset of 3d indoor scenes Angela Dai
11:00 Simplifying visual SLAM for large-scale and multi-device solutions: Do we really need maps? Hermann Blum
11:30 Present and future of SLAM in extreme environments Shehryar Khattak
12:00 Unifying Visual SLAM: From Fragmented Datasets to Scalable, Real-World Solutions

Organizers

Alejandro Fontan

Alejandro Fontan

Queensland University of Technology

Lukas Schmid

Lukas Schmid

Massachusetts Institute of Technology

Olga Vysotska

Olga Vysotska

ETH Zurich

Mubariz Zaffar

Mubariz Zaffar

TU Delft

Linfei Pan

Linfei Pan

ETH Zurich

Gokul B. Nair

Gokul B. Nair

Queensland University of Technology

Javier Civera

Javier Civera

Universidad de Zaragoza

Michael Milford

Michael Milford

Queensland University of Technology


Unifying Visual SLAM

Resource Description Comments
PySLAM A python implementation of a Visual SLAM pipeline that supports monocular, stereo and RGBD cameras. -
slamplay A collection of powerful tools to start playing and experimenting with SLAM in C++. -