Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Evaluating Identity Leakage in Speaker De-Identification Systems

Created by
  • Haebom

Author

Seungmin Seo, Oleg Aulov, Afzal Godil, Kevin Mangold

Outline

This paper presents a new benchmark to quantitatively evaluate the residual identity leakage problem of speaker anonymization techniques that conceal speakers' identities while maintaining their identifiable information. We measure residual identity leakage using three error rates: the Equal Error Rate, the Cumulative Match Characteristic hit rate, and embedding space similarity using Canonical Correlation Analysis and Procrustes Analysis. Our evaluation results show that all state-of-the-art speaker anonymization systems leak identity information, with even the best performing system only slightly better than random guessing, and the worst performing system achieving a 45% accuracy rate among the top 50 speakers based on CMC. This highlights the persistent privacy risks of current speaker anonymization techniques.

Takeaways, Limitations

Takeaways:
Quantitatively revealing the residual identity leakage problem of existing state-of-the-art speaker anonymization systems.
Reaffirming the privacy risks of speaker anonymization technology and suggesting future research directions.
The proposed benchmark provides a useful tool for evaluating the performance of speaker anonymization systems.
Limitations:
Further validation is needed to determine whether the proposed benchmark is applicable to all types of speaker anonymization systems.
Further research is needed on the generalizability of the dataset used for evaluation.
Lack of performance evaluation in real-world environments.
👍