Is Visual SLAM Tackling the Right Challenges? While fields like image segmentation and reinforcement learning have thrived by leveraging vast amounts of unstructured data, visual SLAM remains constrained by a handful of curated benchmarks. This limitation hinders the development of scalable, robust systems capable of operating in a wide range of complex, real-world environments.
Fragmentation across datasets, pipelines, and evaluation metrics remains a core obstacle. Each dataset follows its own structure, making reproducibility and benchmarking difficult. Researchers frequently spend time adapting to dataset-specific requirements instead of driving fundamental advancements. Additionally, compiling baseline systems is a challenge due to the lack of standardized pipelines, undocumented dataset-specific issues, and inconsistent failure cases. This fragmentation significantly slows down progress in the field.
To move forward, SLAM needs unified dataset formats, common pipelines for baselines, and evaluation metrics that go beyond traditional ATE. Standardization will enable scalable benchmarking and the development of more generalizable SLAM systems that function reliably in real-world scenarios.
Workshop Goal: This workshop aims to bring researchers together to discuss best practices, establish dataset and pipeline standards, and streamline SLAM development. We have invited experts who have created benchmarks that foster SLAM challenges, unify SLAM practices, and develop tools that support not only SLAM but broader robotic applications.
One key output of the workshop will be a curated “Unifying Visual SLAM” list of development tools, datasets, pipelines, and benchmarks—compiled by organizers, speakers, and attendees—to serve as a future reference for the research community. By reducing implementation overhead, improving reproducibility, increasing the amount and diversity of benchmarks and fostering collaboration, this workshop seeks to advance SLAM’s ability to process large-scale data and build more scalable, real-world solutions for robotics and computer vision.
We invite short papers of novel or recently published research relevant to the topics of the workshop.
We invite submissions including, but not limited to:
Please submit your paper via CMT.
April 7 | Call for submissions |
May 5 | Submissions due |
June 2 | Notification of acceptance |
June 21 | Workshop at RSS! |
Time | Planned Event | Comments |
---|---|---|
08:00 | Opening Remarks | Organizing Committee |
08:05 | pySLAM and slamplay: Modular, Extensible SLAM Tools for Rapid Prototyping and Integration | Luigi Freda |
08:30 | ROS2WASM: Bringing the Robot Operating System to the Web | Tobias Fischer |
09:00 | Isaac ROS Visual SLAM | Tomasz Bednarz |
09:30 | Tartanair and Subt-mrs: Pushing the Limits of Visual SLAM in Challenging Environments | Wenshan Wang |
10:00 | Poster Session/Coffee Break | |
10:30 | Scannet++: A high-fidelity dataset of 3d indoor scenes | Angela Dai |
11:00 | Large-scale and multi device visual SLAM: CroCoDL dataset and map-free solutions | Hermann Blum |
11:30 | Present and future of SLAM in extreme environments | Shehryar Khattak |
12:00 | Unifying Visual SLAM: From Fragmented Datasets to Scalable, Real-World Solutions |
State-of-the-art Visual SLAM | |
---|---|
MASt3R-SLAM | MASt3R-SLAM: Real-Time Dense SLAM with 3D Reconstruction Priors |
MonoGS | Gaussian Splatting SLAM |
DPVO | Deep Patch Visual Odometry/SLAM |
DROID-SLAM | DROID-SLAM: Deep Visual SLAM for Monocular, Stereo, and RGB-D Cameras |
ORB-SLAM2 | ORB-SLAM2: an Open-Source SLAM System for Monocular, Stereo and RGB-D Cameras |
DSO | DSO: Direct Sparse Odometry |
Courses / Educational Materials / Tutorials | |
---|---|
SLAM Handbook | SLAM Handbook |
SLAM-Course | SLAM-Course (2013/14; Cyrill Stachniss) |
SLAM - 5 Minutes with Cyrill | SLAM - 5 Minutes with Cyrill |
Software Frameworks/Libraries | |
---|---|
VSLAM-LAB | VSLAM-LAB: A Comprehensive Framework for Visual SLAM Baselines and Datasets |
VPR-methods | VPR-methods-evaluation |
PySLAM | A python implementation of a Visual SLAM pipeline that supports monocular, stereo and RGBD cameras. |
slamplay | A collection of powerful tools to start playing and experimenting with SLAM in C++. |
Datasets | |
---|---|
CroCoDL | Augmented reality visual localization benchmark with data from legged robots, and evaluating human-robot, cross-device mapping and localization. |
Foundational Papers | |
---|---|
PTAM | Parallel Tracking and Mapping for Small AR Workspace |
MonoSLAM | MonoSLAM: Real-Time Single Camera SLAM |
Talks | |
---|---|
Talk | Daniel Cremers | Deep and Direct Visual SLAM | Tartan SLAM Series |
Talk | From SLAM to Spatial AI - Andrew Davison Robotics Today |
Talk | Graph-based representations for Spatial-AI | Andrew Davison | Tartan SLAM Series |
Others | |
---|---|
Awesome SLAM Datasets | Collection of SLAM-related datasets |
Structure-from-Motion | |
---|---|
GLOMAP | GLOMAP: Global Structure-from-Motion Revisited |
COLMAP | COLMAP: General-purpose Structure-from-Motion (SfM) and Multi-View Stereo (MVS) pipeline with a graphical and command-line interface |
Other Tools | |
---|---|
Pixi | Pixi: A package management tool for developers |
Spectacular AI | Spectacular AI SDK: Platform-independent software solution for real-time 3D mapping and 6-DoF pose tracking |