タイトル & 超要約 音楽の気持ちをAIが理解!大規模データセットと最新技術で音楽感情認識を爆上げ🚀
ギャル的キラキラポイント✨ ● 2496曲の音楽データで感情を分析!まるでギャルの恋愛相談みたい?😎 ● 「DAMER」ってフレームワークが神!感情のブレを抑えるとか天才じゃん?💡 ● 音楽ストリーミングとかメンタルヘルスケアで大活躍の予感!エモい未来✨
詳細解説
リアルでの使いみちアイデア💡
続きは「らくらく論文」アプリで
Music Emotion Recogniser (MER) research faces challenges due to limited high-quality annotated datasets and difficulties in addressing cross-track feature drift. This work presents two primary contributions to address these issues. Memo2496, a large-scale dataset, offers 2496 instrumental music tracks with continuous valence arousal labels, annotated by 30 certified music specialists. Annotation quality is ensured through calibration with extreme emotion exemplars and a consistency threshold of 0.25, measured by Euclidean distance in the valence arousal space. Furthermore, the Dual-view Adaptive Music Emotion Recogniser (DAMER) is introduced. DAMER integrates three synergistic modules: Dual Stream Attention Fusion (DSAF) facilitates token-level bidirectional interaction between Mel spectrograms and cochleagrams via cross attention mechanisms; Progressive Confidence Labelling (PCL) generates reliable pseudo labels employing curriculum-based temperature scheduling and consistency quantification using Jensen Shannon divergence; and Style Anchored Memory Learning (SAML) maintains a contrastive memory queue to mitigate cross-track feature drift. Extensive experiments on the Memo2496, 1000songs, and PMEmo datasets demonstrate DAMER's state-of-the-art performance, improving arousal dimension accuracy by 3.43%, 2.25%, and 0.17%, respectively. Ablation studies and visualisation analyses validate each module's contribution. Both the dataset and source code are publicly available.