Skip to content

openxrlab/xrslam

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

47 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation



DocumentationactionsLICENSE

Introduction

OpenXRLab Visual-inertial SLAM Toolbox and Benchmark. It is a part of the OpenXRLab project.

XRSLAM-PC.mp4
XRSLAM-Demo.mp4

Major Features

  • Robust and Lightweight optimization-based Visual Inertial Odometry
  • Both desktop and mobile platforms are supported
  • An interactive and real-time AR application on iPhone
  • Visual localization module for running XRARDemo on prebuilt scenarios
  • Compared to other state-of-the-art systems, XRSLAM achieves competitive accuracy

Installation

We provide detailed installation tutorial for XRSLAM, users can install from scratch or use provided Dockerfile.

Getting Started

Please refer to quick start for the basic usage of XRSLAM.

License

The license of our codebase is Apache-2.0. Note that this license only applies to code in our library, the dependencies of which are separate and individually licensed. We would like to pay tribute to open-source implementations to which we rely on. Please be aware that using the content of dependencies may affect the license of our codebase. Some supported methods may carry additional licenses.

FAQ

Please refer to FAQ for frequently asked questions.

Citation

If you use this toolbox or benchmark in your research, please cite this project.

@misc{xrslam,
    title={OpenXRLab Visual-inertial SLAM Toolbox and Benchmark},
    author={XRSLAM Contributors},
    howpublished = {\url{https://github.com/openxrlab/xrslam}},
    year={2022}
}

If you use the Robust Visual-Inertial Odometry in your research, please cite:

@article{li2023rd,
  title={RD-VIO: Robust Visual-Inertial Odometry for Mobile Augmented Reality in Dynamic Environments},
  author={Li, Jinyu and Pan, Xiaokun and Huang, Gan and Zhang, Ziyang and Wang, Nan and Bao, Hujun and Zhang, Guofeng},
  journal={arXiv preprint arXiv:2310.15072},
  year={2023}
}

Contributing

We appreciate all contributions to improve XRSLAM. Please refer to CONTRIBUTING.md for the contributing guideline.

Acknowledgement

XRSLAM is an open source project that is contributed by researchers and engineers from both the academia and the industry. We appreciate all the contributors who implement their methods or add new features, as well as users who give valuable feedbacks. We wish that the toolbox and benchmark could serve the growing research community by providing a flexible toolkit to reimplement existing methods and develop their own new models.

Projects in OpenXRLab

  • XRPrimer: OpenXRLab foundational library for XR-related algorithms.
  • XRSLAM: OpenXRLab Visual-inertial SLAM Toolbox and Benchmark.
  • XRSfM: OpenXRLab Structure-from-Motion Toolbox and Benchmark.
  • XRLocalization: OpenXRLab Visual Localization Toolbox and Server.
  • XRMoCap: OpenXRLab Multi-view Motion Capture Toolbox and Benchmark.
  • XRMoGen: OpenXRLab Human Motion Generation Toolbox and Benchmark.
  • XRNeRF: OpenXRLab Neural Radiance Field (NeRF) Toolbox and Benchmark.