| | --- |
| | license: other |
| | license_name: license |
| | license_link: LICENSE |
| | --- |
| | <div align="center"> |
| | <h1>WildCross: A Cross-Modal Large Scale Benchmark for Place Recognition and Metric Depth Estimation in Natural Environments</h1> |
| |
|
| | [**Joshua Knights**](https://scholar.google.com/citations?user=RxbGr2EAAAAJ&hl=en)<sup>1,2</sup> 路 **Joseph Reid**<sup>1</sup> 路 [**Mark Cox**](https://scholar.google.com/citations?user=Bk3UD4EAAAAJ&hl=en)<sup>1</sup> |
| | <br> |
| | [**Kaushik Roy**](https://bit0123.github.io/)<sup>1</sup> 路 [**David Hall**](https://scholar.google.com/citations?user=dosODoQAAAAJ&hl=en)<sup>1</sup> 路 [**Peyman Moghadam**](https://scholar.google.com.au/citations?user=QAVcuWUAAAAJ&hl=en)<sup>1,2</sup> |
| |
|
| | <sup>1</sup>DATA61, CSIRO   <sup>2</sup>Queensland University of Technology |
| | <br> |
| |
|
| | <a href=""><img src='https://img.shields.io/badge/arXiv-WildCross-red' alt='Paper PDF'></a> |
| | <a href=''><img src='https://img.shields.io/badge/Project_Page-WildCross-green' alt='Project Page'></a> |
| | <a href='https://doi.org/10.25919/5fmy-yg37'><img src='https://img.shields.io/badge/Dataset_Download-WildCross-blue'></a> |
| | </div> |
| |
|
| | This repository contains the pre-trained checkpoints for a variety of tasks on the WildCross benchmark |
| |
|
| |  |
| |
|
| | If you find this repository useful or use the WildCross dataset in your work, please cite us using the following: |
| | ``` |
| | @inproceedings{wildcross2026, |
| | title={{WildCross: A Cross-Modal Large Scale Benchmark for Place Recognition and Metric Depth Estimation in Natural Environments}}, |
| | author={Joshua Knights, Joseph Reid, Kaushik Roy, David Hall, Mark Cox, Peyman Moghadam}, |
| | booktitle={Proceedings-IEEE International Conference on Robotics and Automation}, |
| | pages={}, |
| | year={2026} |
| | } |
| | ``` |
| |
|
| | ## Download Instructions |
| | Our dataset can be downloaded through the [**CSIRO Data Access Portal**](https://doi.org/10.25919/5fmy-yg37). Detailed instructions for downloading the dataset can be found in the README file provided on the data access portal page. |
| |
|
| |
|
| | ## Training and Benchmarking |
| | Here we provide pre-trained checkpoints for a variety of tasks on WildCross. |
| |
|
| | **Visual Place Recognition** |
| | ### Checkpoints |
| | | Model | Checkpoint Folder| |
| | |------------|------------| |
| | | NetVlad | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/VPR/NetVLAD) | |
| | | MixVPR | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/VPR/MixVPR) | |
| | | SALAD | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/VPR/SALAD) | |
| | | BoQ | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/VPR/BoQ) | |
| |
|
| | **Cross Modal Place Recognition** |
| | ### Checkpoints |
| | | Model | Checkpoint Folder| |
| | |------------|------------| |
| | | Lip-Loc (ResNet50) | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/crossmodal/resnet50) | |
| | | Lip-Loc (Dino-v2) | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/crossmodal/dinov2) | |
| | | Lip-Loc (Dino-v3) | [Link](https://huggingface.co/CSIRORobotics/WildCross/tree/main/crossmodal/dinov3) | |
| |
|
| | **Metric Depth Estimation** |
| | ### Checkpoints |
| | | Model | Checkpoint Folder| |
| | |------------|------------| |
| | | DepthAnythingV2-vits | [Link](https://huggingface.co/CSIRORobotics/WildCross/resolve/main/DepthAnythingV2/finetuned/vits.pth) | |
| | | DepthAnythingV2-vitb | [Link](https://huggingface.co/CSIRORobotics/WildCross/resolve/main/DepthAnythingV2/finetuned/vitb.pth) | |
| | | DepthAnythingV2-vitl | [Link](https://huggingface.co/CSIRORobotics/WildCross/resolve/main/DepthAnythingV2/finetuned/vitl.pth) | |
| |
|
| | For instructions on how to use these checkpoints for training or evaluation, further instructions can be found on the [WildCross GitHub repository](). |