[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

DynamicID: Zero-Shot Multi-ID Image Personalization with Flexible Facial Editability

Created by
  • Haebom

Author

Xirui Hu, Jiahao Wang, Hao Chen, Weizhan Zhang, Benqi Wang, Yikun Li, Haishun Nan

Outline

This paper focuses on the generation of personalized human images depicting specific identities from reference images. While existing methods have achieved high-fidelity identity preservation, they are limited to single-ID scenarios and lack face editing capabilities. In this paper, we present DynamicID, a tuning-free framework that supports single-ID and multi-ID personalization generation with high fidelity and flexible face editing capabilities. Key innovations include Semantic-Activated Attention (SAA), which minimizes the interference of the base model when injecting ID features and achieves multi-ID personalization without multiple ID samples during training; Identity-Motion Reconfigurator (IMR), which effectively separates and reconfigures facial motion and ID features to support flexible face editing; a task-separated training paradigm that reduces data dependency; and VariFace-10k dataset, where 10,000 unique individuals are represented by 35 different face images each. Experimental results show that DynamicID outperforms state-of-the-art methods in terms of identity fidelity, face editing capabilities, and multi-ID personalization capabilities.

Takeaways, Limitations

Takeaways:
We present a novel framework (DynamicID) that provides high-quality single- and multi-ID personalized image generation without tuning.
Achieving high face editability and identity preservation with Semantic-Activated Attention (SAA) and Identity-Motion Reconfigurator (IMR).
Reducing data dependency and improving performance by leveraging the task-separated training paradigm and the VariFace-10k dataset.
Effectively solves the problems of single ID and low face editability of existing methods, which are Limitations.
Limitations:
Further review is needed regarding the size and diversity of the VariFace-10k dataset.
There may be bias toward certain races or genders.
Lack of discussion of ethical considerations and potential exploits in real-world applications.
👍