[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

ACME: Adaptive Customization of Large Models via Distributed Systems

Created by
  • Haebom

Author

Ziming Dai, Chao Qiu, Fei Gao, Yunfeng Zhao, Xiaofei Wang

Outline

In this paper, we propose ACME, an adaptive customization method for large-scale models based on transformers that utilizes distributed systems to solve the data privacy and response delay issues that occur when deploying large-scale language models in cloud environments. ACME performs progressively fine-grained collaborative model customization using a bidirectional single-loop distributed system to avoid the inefficiency of centralized methods. To improve the suitability for user heterogeneity, we identify the Pareto Front under model size constraints to customize the backbone generation, and then use personalized architecture aggregation based on data distribution to improve header generation and model to accommodate data heterogeneity. Evaluation results on various datasets show that ACME achieves a cost-effective model under model size constraints, reduces data transmission by 6% compared to centralized systems, improves the average accuracy by 10% compared to the baseline model, and increases the performance indicator by about 30%.

Takeaways, Limitations

Takeaways:
Presenting an efficient method for customizing large-scale language models using distributed systems
Adaptive model optimization considering user and data heterogeneity
Reduced data transfer and improved accuracy compared to centralized methods
Create cost-effective models
Limitations:
ACME's performance gains may be limited to specific datasets.
Additional verification of scalability and stability in real cloud environments is needed.
Generalizability studies for various types of transformer models are needed.
👍