About the Workshop

As computer vision systems increasingly transition into real-world applications, reliable and scalable localization across heterogeneous devices becomes critical. The CroCoDL workshop brings together researchers from computer vision, robotics, and augmented reality to address the unique challenges of cross-device, multi-agent localization in complex, real-world environments. With a focus on 3D vision, visual localization, egocentric and embodied AI, and AR/VR/MR, this workshop aims to foster dialogue around bridging the gap between academic benchmarks and real-world deployment. This inaugural edition features leading experts from academia and industry and introduces CroCoDL, a new large-scale benchmark dataset capturing synchronized sensor data from smartphones, mixed-reality headsets, and legged robots across diverse environments. Through invited talks, a paper track, and an open competition, the workshop will highlight recent advances and open challenges in localization under domain shifts, sensor diversity, and dynamic scene conditions. By uniting communities working on structure-from-motion, neural rendering, and embodied AI, CroCoDL offers a platform to drive innovation toward robust, scalable localization systems capable of operating across devices, agents, and perspectives.

Invited Speakers

Ayoung Kim
Prof. Ayoung Kim

Professor at Seoul National University

Ayoung Kim leads the Robust Perception for Mobile Robotics Lab at Seoul National University. Her research aims to develop robust and reliable perception systems that enhance mobile robot navigation in complex and dynamic environments. She has made significant contributions to the field, particularly in visual localization, sensor fusion, and learning-based perception techniques.

Torsten Sattler
Dr. Torsten Sattler

Senior Researcher at Czech Technical University in Prague

Torsten Sattler is a senior researcher at the Czech Technical University in Prague (CTU), where he leads the Spatial Intelligence group. He works towards making 3D computer vision algorithms such as 3D reconstruction and visual localization more robust and reliable through machine learning models trained on scene understanding and 3D computer vision tasks.

Gabriela Csurka
Dr. Gabriela Csurka

Principal Research Scientist at Naver Labs Europe

Gabriela Csurka is a principal research scientist at Naver Labs Europe. Her work bridges fundamental research and real-world applications in robotics and augmented reality. Her contributions to the field span various topics, including feature learning, visual place recognition, and cross-domain adaptation.

unknown

Stay tuned ...

Call for Papers

We invite 8-page full papers for inclusion in the proceedings, as well as 4-page extended abstracts. Extended abstracts may present either new or previously published work; however, they will not be included in the official proceedings.

Please note that 4-page extended abstracts generally do not conflict with the dual submission policies of other conferences. In contrast, 8-page full papers, if accepted, will appear in the proceedings and are therefore subject to the dual submission policy. This means they must not be under review or accepted at another conference at the same time.

All submissions must be anonymous and comply with the official ICCV 2025 guidelines.

Topics of Interest

  • 3D Reconstruction
  • Visual localization & structure-from-motion
  • Image retrieval
  • Implicit Scene Representations
  • Egocentric & embodied AI
  • Domain shift & sensor diversity
  • Real-world deployment at scale

Submission Timeline for 8-page full papers:

  • Submission Portal: OpenReview
  • Paper Submission Opens: 16th of May, 2025
  • Paper Submission Deadline: 6th of June, 2025 16th of June, 2025
  • Notification to Authors: 25th of June, 2025 7th of July, 2025
  • Camera-ready Submission: 25th of July, 2025 15th of August, 2025

Submission Timeline for 4-page abstracts:

  • Submission Portal: OpenReview
  • Abstract Submission Opens: 16th of May, 2025
  • Abstract Submission Deadline: 26th of September, 2025
  • Notification to Authors: 10th of October, 2025
  • Camera-ready Submission: 17th of October, 2025

Challenge

unknown

The workshop challenge is centered around a newly accepted dataset at CVPR 2025 – CroCoDL: Cross-device Collaborative Dataset for Localization (pre-rebuttal version available at link). To advance research in visual co-localization, we introduce CroCoDL, a significantly larger andmore diverse dataset and benchmark, as shown in Figure 1. CroCoDL is the first dataset to incorporate sensor recordings from both robots and mixed-reality headsets and covering a wider range of real-world environments than any existing cross-device visual localization dataset. It includes synchronized sensor streams from three primary devices: hand-held smartphones, head-mounted HoloLens 2, and the legged robot Spot.

unknown

For this challenge, we have selected three large-scale locations—Hydrology, Succulent, and Design Museum—where we will release mapping and query splits. These splits will be used to evaluate visual localization performance in a crossdevice setup, meaning that the map is generated using data from one device, while the goal is to localize images taken by a different device within this map. The primary evaluation metric for submissions will be single-image localization recall at 50 cm and 10 degrees of pose error.

The competition will be split into two tracks:

  • T1. Traditional structure-from-motion (SfM) methods. This track includes approaches that may incorporate learned components such as feature extractors, matchers (including dense matchers), and keypoint refinement techniques. However, these learned components should not be trained or fine-tuned specifically for the challenge scenes.

  • T2. End-to-end learned methods. This track includes methods such as coordinate scene regression, pose regression, and some feed-forward SfM that need to be trained or fine-tuned for each scene.

  • To be eligible for a prize, participants must provide code that can reproduce their results and commit to releasing it as open-source. This code will be used to validate their submissions.

    Timeline for the Challenge:

    • Stay tuned...

    Organizers


    Dr. Zuria Bauer

    Postdoc
    ETH Zurich


    JProf. Hermann Blum

    Junior Professor
    University Bonn


    Dr. Mihai Dusmanu

    Senior Scientist
    Microsoft


    Linfei Pan

    PhD student
    ETH Zurich
     


    Dr. Qunjie Zhou

    Research Scientist
    Nvidia
     


    Prof. Marc Pollefeys

    Professor, ETH Zurich
    Director
    Spatial AI Lab, Microsoft

    Partners