Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

LaMoGen: Laban Movement-Guided Diffusion for Text-to-Motion Generation

Created by
  • Haebom

Author

Heechang Kim, Gwanghyun Kim, Se Young Chun

Diverse Human Motion Generation with Laban-Inspired Control

Outline

This paper addresses multi-sensory human motion generation, a critical challenge in diverse fields such as computer vision, human-computer interaction, and animation. While using diffusion models for text-motion synthesis has successfully generated high-quality motion, detailed expressive motion control remains a challenging task. This is due to the lack of motion style diversity in datasets and the difficulty of expressing quantitative characteristics in natural language. This study aims to generate interpretable and expressive human motion by integrating methods for quantifying Laban effort and shape components into a text-based motion generation model. The proposed zero-shot, inference-time optimization method guides the motion generation model to acquire the desired Laban effort and shape components without additional motion data by updating the text embeddings of a pretrained diffusion model during the sampling phase.

Takeaways, Limitations

Leveraging the quantification of Laban Effort and Shape components to generate interpretable and expressive human motion.
Control desired motion properties without additional data through zero-shot, inference-time optimization methods.
Successfully manipulate motion properties according to the target Laban tag to maintain a variety of expressiveness and motion identity.
For specific references to Limitations in this paper, please refer to the original text. (Not included in the abstract.)
👍