UL-VIO:
Ultra-lightweight Visual-Inertial Odometry
with Noise Robust Test-time Adaptation

1Columbia University, 2Seoul National University
European Conference on Computer Vision (ECCV) 2024

Abstract

Data-driven visual-inertial odometry (VIO) has received highlights for its performance since VIOs are a crucial compartment in autonomous robots. However, their deployment on resource-constrained devices is non-trivial since large network parameters should be accommodated in the device memory. Furthermore, these networks may risk failure post-deployment due to environmental distribution shifts at test time. In light of this, we propose UL-VIO -- an ultra-lightweight (<1M) VIO network capable of test-time adaptation (TTA) based on visual-inertial consistency. Specifically, we perform model compression to the network while preserving the low-level encoder part, including all BatchNorm parameters for resource-efficient test-time adaptation. It achieves 36X smaller network size than state-of-the-art with a minute increase in error -- 1% on the KITTI dataset. For test-time adaptation, we propose to use the inertia-referred network outputs as pseudo labels and update the BatchNorm parameter for lightweight yet effective adaptation. To the best of our knowledge, this is the first work to perform noise-robust TTA on VIO. Experimental results on the KITTI, EuRoC, and Marulan datasets demonstrate the effectiveness of our resource-efficient adaptation method under diverse TTA scenarios with dynamic domain shifts.

Motivation


Motivation

Deploying VIO networks on mobile autonomous platforms poses a significant challenge due to the limited memory and computing capacity of such devices. More importantly, accessing off-chip DRAM memory requires two to three orders of magnitude more power compared to on-chip memory access, thereby imposing a significant limitation on the size of the networks that can be deployed on these platforms.

Yet another concern for mobile VIO platforms is that they may suffer from post-deployment performance degradation when encountering out-of-distribution (OoD) data at test time. For example, a network trained on clean camera image sequences might be prone to failure when the image distribution shifts due to environmental conditions, e.g., shadow, snow, and rain.




Model Compression


overall pipeline

We target sub-million parameter count for the model to be accommodated in the on-chip memory of a mobile platform. Commercial mobile processors like Apple A16 and Qualcomm Snapdragon only possess a few MB of on-chip memory. We reduce the size of the visual encoder while maintaining the BatchNorm (BN) parameter size for test-time adaptation since tuning BN is a preferred method for adaptation.

We summarize our approach and its effects in the following:

  • Add an AveragePool after the last convolutional layer in $E_\text{visual}$. This gives us $117 \times$ reduction in $E_\text{visual}$.
  • Reduce the channel size in $E_\text{inertial}$ since the parameter number is quadratically proportional to it, attaining $8 \times$ compression in $E_\text{inertial}$.
  • Replace the LSTM with fully connected layers for the $D_\text{fused}$, resulting in $161 \times$ downsizing in $D_\text{fused}$.



  • Model Compression - Results


    relationship map generation

    Our compressed result gives $36.45 \times$ lower model size than that of the target state-of-the-art baseline, NASVIO, while having a minute increase in relative translation/rotation errors $\{ t_{rel}, r_{rel} \} = \{ 1.11\%, 1.05^{\circ} \}$ against the art. For similarly-sized NASVIO maintaining the architecture, we achieve translation/rotation error reduction of $\{ 3.80\%, 1.94^{\circ} \}$.




    Trajectory Results


    relationship map generation

    Our network performs comparably to others on Seq. $07$ and outperforms others on Seq. $10$.




    Multi-modal Consistency-based TTA


    relationship map generation
    • Visual encoder deduces visual features from pairwise images.
    • Inertial encoder does so from the inertial input.
    • Decoder predicts pose transformation from fused features.



    • Dictionary-based Visual Encoder Adaptation


      relationship map generation

      Only the weights of the visual encoder are modified during adaptation while the weights of other modules are fixed. As shown here, domain distinctive features from the early layers of the visual encoder are utilized for domain shift detection. The visual encoder hosts an auxiliary dictionary to store and update learnable BN parameters corresponding to different noise types.




      Motivation for Multi-modal Consistency


      relationship map generation

      Although the inertial-inferred pose estimates exhibit sub-par performance compared to that of vision, it is unaffected by the weather conditions. When we simulate adversarial weather conditions on KITTI-C, we observe that the fused-feature-based poses become much more erroneous than the inertial-referred poses. This demonstrates a strong correlation ($r = 0.86$) between the inertial-inferred output and the ground truth.




      Comparison Against Fine-tuned Baselines


      relationship map generation

      We demonstrate the effectiveness of our TTA method by comparing it with networks fine-tuned with adversarial noises. Except for one case, e.g., multiplicative noise, our TTA method has the best or second-best accuracy. This case assumes stationary domain shift. Here, we fine-tuned the baseline model, trained initially on the noise-free source domain, by introducing the corresponding visual corruption.




      Continual TTA


      relationship map generation

      We report VIO results for dynamically corrupted vision inputs on KITTI Seq. $07$ with and without TTA. The sequence starts with clean images until $t_0 = 22$s. After $t_0$, the system instead receives blurred images, which continues until $t_1 = 88$s. Then, the distribution shift is removed, and the image input returns to the uncorrupted source domain. Such a domain shift results in a pose-wise $t_{rmse}$ increase from $0.022$ m to $0.133$ m. TTA reduces the error by $29.7 \%$ to $0.093$ m.




      Continual TTA w/ Dynamic Noise Shifts


      relationship map generation

      We perform vision corruptions to KITTI and EuRoC datasets with methods from ImageNet-C. With continual TTA on KITTI, our UL-VIO achieves $18\%$ reduction in pose-wise $t_{rmse}$ on average. The domain-discriminative TTA governs $K$ sets of lightweight BN parameters adequately switched based on domain matching with high $ddf$ acc. of $99.6\%$.




      BibTeX

      @inproceedings{park2024ulvio,
            title={UL-VIO: Ultra-lightweight Visual-Inertial Odometry with Noise Robust Test-time Adaptation},
            author={Park, Jinho and Chun, Se Young and Seok, Mingoo},
            booktitle={European Conference on Computer Vision},
            year={2024},
            organization={Springer}
            }