Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

SSGaussian: Semantic-Aware and Structure-Preserving 3D Style Transfer

Created by
  • Haebom

Author

Jimin Xu, Bosheng Qin, Tao Jin, Zhou Zhao, Zhenhui Ye, Jun Yu, Fei Wu

Outline

This paper proposes a novel 3D style transfer pipeline that leverages knowledge from pre-trained 2D diffusion models to address the challenges of existing 3D style transfer methods, which struggle to effectively extract and transfer high-dimensional style semantics and suffer from structural ambiguity in the resulting style application, making object identification difficult. This pipeline consists of two steps: generating stylized renderings of dominant viewpoints and then transferring them to 3D representations. Specifically, cross-view style alignment enables feature interactions across multiple dominant viewpoints, and instance-level style transfer effectively transfers consistency between stylized dominant viewpoints to the 3D representation, resulting in structurally and visually consistent stylization results. Experimental results on various scenes demonstrate that the proposed method outperforms existing state-of-the-art methods.

Takeaways, Limitations

Takeaways:
We present a novel pipeline that effectively solves the 3D style transfer problem by leveraging a pre-trained 2D diffusion model.
Cross-view style alignment and instance-level style transfer techniques simultaneously improve style fidelity and instance-level consistency.
It showed superior performance than existing methods in various scenes (front, 360-degree environment, etc.).
Limitations:
Because of the high dependence on the 2D diffusion model, the quality of the 3D style transfer results may be affected by the performance of the 2D diffusion model.
Computational cost may be high (although not explicitly stated, this is expected given the complexity of the 2D diffusion model and 3D representation processing Limitations)
Further research is needed to determine the generalization performance of the proposed method. (Although not explicitly stated, it is anticipated that further research will be necessary, as superior performance on a specific dataset does not guarantee superior performance on all datasets.)
👍