- Notifications
You must be signed in to change notification settings - Fork65
[CVPR2024] SchurVINS: Schur Complement-Based Lightweight Visual Inertial Navigation System
License
bytedance/SchurVINS
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
We propose a filter-based VINS framework named SchurVINS, which guarantees both high accuracy by building a complete residual model and low computational complexity with Schur complement. Technically, we first formulate the full residual model where Gradient, Hessian and observation covariance are explicitly modeled. Then Schur complement is employed to decompose the full model into ego-motion residual model and landmark residual model. Finally, Extended Kalman Filter (EKF) update is implemented in these two models with high efficiency. Experiments on EuRoC and TUM-VI datasets show that SchurVINS notably outperforms state-of-the-art (SOTA) methods in both accuracy and computational complexity. The main contributions include:
- An equivalent residual model is proposed to deal with hyper high-dimension observations, which consists of gradient, Hessian and the corresponding observation covariance. This method is of great generality in EKF systems.
- A lightweight EKF-based landmark solver is proposed toestimate position of landmarks with high efficiency.
- A novel EKF-based VINS framework is developed toachieve ego-motion and landmark estimation simultaneously with high accuracy and efficiency.
The code is licensed under GPLv3.
The SchurVINS is developed on SVO2.0(rpg_svo_pro_open), and thus its license is retained at the beginning of the related files.
Related Publication:
Yunfei Fan, Tianyu Zhao, Guidong Wang. SchurVINS: Schur Complement-Based Lightweight Visual Inertial Navigation System. (Accepted by CVPR 2024).PDF.
We have tested the codebase inUbuntu 18.04.
The following dependencies are needed:
# For Ubuntu 18.04 + Melodicsudo apt-get install python-catkin-tools python-vcstoolsudo apt-get install libglew-dev libopencv-dev libyaml-cpp-dev
sudo apt-get install libblas-dev liblapack-dev libsuitesparse-dev
Clone and build the repository:
cd ~/catkin_ws/srcgit clone https://github.com/bytedance/SchurVINS.gitsource ~/catkin_ws/devel/setup.bashmkdir -p ~/catkin_ws/src/SchurVINS/resultsmkdir -p ~/catkin_ws/src/SchurVINS/logsvcs-import < ./SchurVINS/dependencies.yamltouch minkindr/minkindr_python/CATKIN_IGNOREcatkin build
We provide examples to run SchurVINS withEuRoC dataset.
Open three terminals, launch the schur_vins , rviz and play the bag file respectively. Take MH_01 for example
roslaunch svo_ros euroc_vio_stereo.launchrviz -d svo_ros/rviz_config.rvizrosbag play YOUR_PATH_TO_DATASET/MH_01_easy.bag
There might be minor differences between the released version and the results in the paper. Please note that multi-thread performance has some randomness due to CPU utilization.On EuRoC datasets, please note to skip the first few seconds of datasets with the following configurations for reproducibility.
MH1 | MH2 | MH3 | MH4 | MH5 | V11 | V12 | V13 | V21 | V22 |
30s | 30s | 10s | 13s | 15s | 0s | 0s | 2s | 0s | 0s |
If you discover a potential security issue in this project, or think you mayhave discovered a security issue, we ask that you notify Bytedance Security via oursecurity center orvulnerability reporting email.
This work incorporates the well-known SVO2.0 open-source code. We extend our gratitude to the authors of the software.
If you found this code/work to be useful in your own research, please considering citing the following information. Additionally, please considering citing SVO2.0(rpg_svo_pro_open) since SchurVINS is developed on it.
@inproceedings{fan2024schurvins,title={SchurVINS: Schur Complement-Based Lightweight Visual Inertial Navigation System},author={Fan, Yunfei and Zhao, Tianyu and Wang, Guidong},booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},pages={17964--17973},year={2024}}
Our team is hiring FTEs with background in Deep Learning, SLAM, and 3D Vision. We are based in Beijing and Shanghai. If you are interested, please send your resume to frank.01[AT]bytedance[DOT]com.