超要約: ロボの顔で歌うSingingBot!感情表現豊かに人間と仲良くなれるよ💖
✨ ギャル的キラキラポイント✨ ● ロボが歌で感情表現!まるで人間みたい🥰 ● アバター使って表情豊かに!口パクもバッチリ👌 ● EDRで感情表現を数値化!スゴくない?✨
詳細解説いくよ~!
背景 ロボットの顔って、あんまり感情表現できなかったじゃん? でも、歌って感情を伝えるのに最強の手段じゃん? だから、歌で感情表現できるロボを作ろう!って研究が始まったんだって🎵
続きは「らくらく論文」アプリで
Equipping robotic faces with singing capabilities is crucial for empathetic Human-Robot Interaction. However, existing robotic face driving research primarily focuses on conversations or mimicking static expressions, struggling to meet the high demands for continuous emotional expression and coherence in singing. To address this, we propose a novel avatar-driven framework for appealing robotic singing. We first leverage portrait video generation models embedded with extensive human priors to synthesize vivid singing avatars, providing reliable expression and emotion guidance. Subsequently, these facial features are transferred to the robot via semantic-oriented mapping functions that span a wide expression space. Furthermore, to quantitatively evaluate the emotional richness of robotic singing, we propose the Emotion Dynamic Range metric to measure the emotional breadth within the Valence-Arousal space, revealing that a broad emotional spectrum is crucial for appealing performances. Comprehensive experiments prove that our method achieves rich emotional expressions while maintaining lip-audio synchronization, significantly outperforming existing approaches.