Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

MoSE: Skill-by-Skill Mixture-of-Experts Learning for Embodied Autonomous Machines

Created by
  • Haebom

Author

Lu Xu, Jiaqian Yu, Xiongfeng Peng, Yiwei Chen, Weiming Li, Jaewook Yoo, Sunghyun Chunag, Dongwook Lee, Daehyun Ji, Chao Zhang

Outline

To meet the growing demand for efficient and intelligent embedded AI systems, this paper proposes MoSE, a novel Mixed Expert (MoE) method. To address the challenges of existing MoE models, which require massive training data and complex optimization processes, MoSE mimics human learning and inference processes, performing skill-by-skill, step-by-step learning. It facilitates skill-by-skill learning by defining and annotating specific skills, allowing experts to identify the competencies required for various scenarios and inference tasks. It builds a hierarchical skill dataset and pretrains routers to encourage step-by-step inference, integrating auxiliary tasks such as perception-prediction-planning for autonomous driving (AD) and high- and low-level planning for robots into a single pass without additional computational overhead. It effectively scales diverse expertise with fewer than 3 billion sparsely activated parameters, outperforming existing models on both AD corner-case inference and robot inference tasks with fewer parameters (less than 40%).

Takeaways, Limitations

Takeaways:
We present a novel MoE method (MoSE) that significantly improves the inference and learning efficiency of embedded AI systems.
Efficient technology-specific learning and acquisition of diverse expertise through technology-centric routing mechanisms.
Step-by-step inference driven by hierarchical technical datasets and pre-trained routers.
Integrate various auxiliary operations without additional computational cost.
Achieve superior performance with fewer parameters than existing models.
Limitations:
Further verification of the generalization performance of the proposed MoSE model is needed.
Consider the effort and cost of annotating and defining specific technologies.
Further research is needed on the applicability and scalability to various embedded AI systems.
Robustness and safety evaluation in real environments is required.
👍