Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Hierarchical place recognition with omnidirectional images and curriculum learning-based loss functions

Created by
  • Haebom

Author

Marcos Alfaro, Juan Jos e Cabrera, Mar ia Flores, Oscar Reinoso, Luis Pay a

Outline

This paper addresses visual localization (VPR), essential for the safe navigation of mobile robots. It utilizes a deep learning model fine-tuned with a triplet loss function that integrates panoramic images and a curriculum learning strategy. By progressively presenting more challenging examples during training, the model learns more discriminative and robust feature representations, overcoming the limitations of conventional contrastive loss functions. After training, VPR is performed in two stages: coarse (room search) and fine (localization). The proposed method is evaluated in various indoor and outdoor environments, and tested against common challenges encountered in real-world operating conditions, such as severe lighting changes, dynamic visual effects such as noise and occlusion, and limited training data.

Takeaways, Limitations

Takeaways:
The curriculum-based triplet loss function consistently outperforms the standard contrastive loss function.
It has proven to be a viable solution for real-world robotics applications by achieving high recognition accuracy even under challenging conditions such as extreme lighting changes, dynamic visual effects, and limited training data.
The code used in the experiment is made public, increasing reproducibility.
Limitations:
The specific Limitations is not specified in the paper.
👍