Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

PCoreSet: Effective Active Learning through Knowledge Distillation from Vision-Language Models

Created by
  • Haebom

Author

Seongjae Kang, Dong Bok Lee, Hyungjoon Jang, Dongseop Kim, Sung Ju Hwang

Outline

This paper proposes the ActiveKD framework, which integrates knowledge distillation (KD) into active learning (AL). It aims to minimize annotation costs by leveraging the zero-shot and few-shot capabilities of large-scale vision-language models (VLMs). The key is to leverage the VLM's structural prediction bias to capture generalizable output patterns beneficial to student learning. To achieve this, we propose a sample selection strategy called Probabilistic CoreSet (PCoreSet), which maximizes coverage in the probability space rather than the feature space. PCoreSet probabilistically selects diverse unlabeled samples, facilitating efficient transfer of teacher knowledge under a limited annotation budget.

Takeaways, Limitations

Takeaways:
A new attempt at combining active learning and knowledge distillation.
A novel learning method is presented that exploits the structural prediction bias of large-scale VLMs.
Proposing PCoreSet, a sample selection strategy in probability space.
Demonstrated improved performance over existing methods on various datasets.
We verified PCoreSet's excellent performance in a variety of student and teacher network configurations.
Ensuring reproducibility through code disclosure.
Limitations:
Performance may be low in early active learning rounds.
Massive VLM dependency.
Potentially limited to specific datasets and models.
👍