Unifying Visual SLAM: From Fragmented Datasets to Scalable, Real-World Solutions

Workshop at Robotics: Science and Systems 2025, Los Angeles, California, USA

Saturday, June 21


News

  • April 7, 2025 - Submissions Portal is now open! Submissions are due by May 5. Please submit your paper via CMT.
  • Mar 24, 2025 - We are excited to host the first workshop on Unifying Visual SLAM at RSS 2025!

Overview

Is Visual SLAM Tackling the Right Challenges? While fields like image segmentation and reinforcement learning have thrived by leveraging vast amounts of unstructured data, visual SLAM remains constrained by a handful of curated benchmarks. This limitation hinders the development of scalable, robust systems capable of operating in a wide range of complex, real-world environments.

Fragmentation across datasets, pipelines, and evaluation metrics remains a core obstacle. Each dataset follows its own structure, making reproducibility and benchmarking difficult. Researchers frequently spend time adapting to dataset-specific requirements instead of driving fundamental advancements. Additionally, compiling baseline systems is a challenge due to the lack of standardized pipelines, undocumented dataset-specific issues, and inconsistent failure cases. This fragmentation significantly slows down progress in the field.

To move forward, SLAM needs unified dataset formats, common pipelines for baselines, and evaluation metrics that go beyond traditional ATE. Standardization will enable scalable benchmarking and the development of more generalizable SLAM systems that function reliably in real-world scenarios.

Workshop Goal: This workshop aims to bring researchers together to discuss best practices, establish dataset and pipeline standards, and streamline SLAM development. We have invited experts who have created benchmarks that foster SLAM challenges, unify SLAM practices, and develop tools that support not only SLAM but broader robotic applications.

One key output of the workshop will be a curated “Unifying Visual SLAM” list of development tools, datasets, pipelines, and benchmarks—compiled by organizers, speakers, and attendees—to serve as a future reference for the research community. By reducing implementation overhead, improving reproducibility, increasing the amount and diversity of benchmarks and fostering collaboration, this workshop seeks to advance SLAM’s ability to process large-scale data and build more scalable, real-world solutions for robotics and computer vision.


Invited Speakers

Hermann Blum

Hermann Blum

University of Bonn & Lamarr Institute

Angela Dai

Angela Dai

Technical University of Munich

Tobias Fischer

Tobias Fischer

Queensland University of Technology

Wenshan Wang

Wenshan Wang

Carnegie Mellon University

Luigi Freda

Luigi Freda

Dexory

Shehryar Khattak

Shehryar Khattak

NASA Jet Propulsion Lab


Call for Papers

We invite short papers of novel or recently published research relevant to the topics of the workshop.

  • Short papers are 2+n pages (2 pages content + references)
  • Submissions must follow the IEEE Conference double column format
  • All accepted papers will be presented as posters at the workshop and published on the workshop website.
  • Please indicate whether your paper falls into the ‘novel’ or ‘previously published’ category. Novel research papers are encouraged and can expect more substantial review feedback on their work. This is provided as a service to authors of novel papers and does not diminish the chance of acceptance.
  • All accepted submissions will be considered for the best presentation award, where 3 finalists will be selected for 5-minute plenary presentations. While all submissions are eligible, novelty will be considered in finalist selection.
  • Submissions are single blind and will be reviewed by members of the (extended) workshop committee.
  • Submissions can optionally be accompanied by a video.
  • All accepted submissions will be considered for three awards: Best Paper, Best Poster, and Best Open-Source Contribution, each receiving a 150 USD gift card.

Invited topics

We invite submissions including, but not limited to:

  • SLAM Systems and Applications: Novel approaches in SLAM, Visual Odometry, 3D Reconstruction, and Visual Place Recognition (VPR), with applications in robotics, autonomous vehicles, AR/VR, and more.
  • Tool Integration and Standardization: Open-source tools and libraries for SLAM, improving baseline evaluations, and streamlining benchmarking processes.
  • Benchmarking and Reproducibility: Standardization of datasets and evaluation methodologies, addressing challenges such as low-light conditions, harsh weather, and dynamic environments.
  • New Evaluation Metrics and Benchmarking Tools: Methods and platforms to assess localization and mapping performance more effectively.
  • Future Directions and Industrial Applications: The role of AI and machine learning in SLAM, and real-world applications in industry, commercial robotics, and emerging technologies.

Submissions Portal

Please submit your paper via CMT.

Submissions Timeline

April 7 Call for submissions
May 5 Submissions due
June 2 Notification of acceptance
June 21 Workshop at RSS!

Schedule

Time Planned Event Comments
08:00 Opening Remarks Organizing Committee
08:05 pySLAM and slamplay: Modular, Extensible SLAM Tools for Rapid Prototyping and Integration Luigi Freda
08:30 ROS2WASM: Bringing the Robot Operating System to the Web Tobias Fischer
09:00 Isaac ROS Visual SLAM Tomasz Bednarz
09:30 Tartanair and Subt-mrs: Pushing the Limits of Visual SLAM in Challenging Environments Wenshan Wang
10:00 Poster Session/Coffee Break
10:30 Scannet++: A high-fidelity dataset of 3d indoor scenes Angela Dai
11:00 Large-scale and multi device visual SLAM: CroCoDL dataset and map-free solutions Hermann Blum
11:30 Present and future of SLAM in extreme environments Shehryar Khattak
12:00 Unifying Visual SLAM: From Fragmented Datasets to Scalable, Real-World Solutions

Organizers

Alejandro Fontan

Alejandro Fontan

Queensland University of Technology

Lukas Schmid

Lukas Schmid

Massachusetts Institute of Technology

Olga Vysotska

Olga Vysotska

ETH Zurich

Mubariz Zaffar

Mubariz Zaffar

TU Delft

Linfei Pan

Linfei Pan

ETH Zurich

Gokul B. Nair

Gokul B. Nair

Queensland University of Technology

Javier Civera

Javier Civera

Universidad de Zaragoza

Michael Milford

Michael Milford

Queensland University of Technology


Unifying Visual SLAM

State-of-the-art Visual SLAM
MASt3R-SLAMMASt3R-SLAM: Real-Time Dense SLAM with 3D Reconstruction Priors
MonoGSGaussian Splatting SLAM
DPVODeep Patch Visual Odometry/SLAM
DROID-SLAMDROID-SLAM: Deep Visual SLAM for Monocular, Stereo, and RGB-D Cameras
ORB-SLAM2ORB-SLAM2: an Open-Source SLAM System for Monocular, Stereo and RGB-D Cameras
DSODSO: Direct Sparse Odometry
Courses / Educational Materials / Tutorials
SLAM HandbookSLAM Handbook
SLAM-CourseSLAM-Course (2013/14; Cyrill Stachniss)
SLAM - 5 Minutes with CyrillSLAM - 5 Minutes with Cyrill
Software Frameworks/Libraries
VSLAM-LABVSLAM-LAB: A Comprehensive Framework for Visual SLAM Baselines and Datasets
VPR-methodsVPR-methods-evaluation
PySLAMA python implementation of a Visual SLAM pipeline that supports monocular, stereo and RGBD cameras.
slamplayA collection of powerful tools to start playing and experimenting with SLAM in C++.
Datasets
CroCoDLAugmented reality visual localization benchmark with data from legged robots, and evaluating human-robot, cross-device mapping and localization.
Foundational Papers
PTAMParallel Tracking and Mapping for Small AR Workspace
MonoSLAMMonoSLAM: Real-Time Single Camera SLAM
Talks
TalkDaniel Cremers | Deep and Direct Visual SLAM | Tartan SLAM Series
TalkFrom SLAM to Spatial AI - Andrew Davison Robotics Today
TalkGraph-based representations for Spatial-AI | Andrew Davison | Tartan SLAM Series
Others
Awesome SLAM DatasetsCollection of SLAM-related datasets
Structure-from-Motion
GLOMAPGLOMAP: Global Structure-from-Motion Revisited
COLMAPCOLMAP: General-purpose Structure-from-Motion (SfM) and Multi-View Stereo (MVS) pipeline with a graphical and command-line interface
Other Tools
PixiPixi: A package management tool for developers
Spectacular AISpectacular AI SDK: Platform-independent software solution for real-time 3D mapping and 6-DoF pose tracking