iconLogo
Published:2025/12/17 14:28:23

動画生成AI、進化が止まらない!✨

超要約: 動画生成AI、もっと可愛く!ノイズ減&変なとこに反応しないようにしたよ💖

ギャル的キラキラポイント✨

● 動画の評価、簡単にした!「合格/不合格」でOK🙆‍♀️ ● 変なとこばっか評価しちゃうAIを、賢くしたよ!😎 ● 動画生成AI、もっともっとキレイになるってこと!😍

詳細解説

背景: 最近の動画生成AIはすごいけど、まだ課題がいっぱい!人間の好みを反映させる「リワードモデル (RM)」っていうのが重要なんだけど、アノテーション(評価)のノイズとか、変なとこにAIが引っかかる「リワードハッキング」とか、問題があったの😥

方法: SoliRewardっていうフレームワーク(枠組み)を使ったよ!単一項目バイナリ(二択)で評価して、変なとこに反応しないように工夫したみたい。HPQAアーキテクチャっていう、賢いRMも作ったみたい💖

続きは「らくらく論文」アプリで

SoliReward: Mitigating Susceptibility to Reward Hacking and Annotation Noise in Video Generation Reward Models

Jiesong Lian / Ruizhe Zhong / Zixiang Zhou / Xiaoyue Mi / Yixue Hao / Yuan Zhou / Qinglin Lu / Long Hu / Junchi Yan

Post-training alignment of video generation models with human preferences is a critical goal. Developing effective Reward Models (RMs) for this process faces significant methodological hurdles. Current data collection paradigms, reliant on in-prompt pairwise annotations, suffer from labeling noise. Concurrently, the architectural design of VLM-based RMs, particularly their output mechanisms, remains underexplored. Furthermore, RM is susceptible to reward hacking in post-training. To mitigate these limitations, we propose SoliReward, a systematic framework for video RM training. Our framework first sources high-quality, cost-efficient data via single-item binary annotations, then constructs preference pairs using a cross-prompt pairing strategy. Architecturally, we employ a Hierarchical Progressive Query Attention mechanism to enhance feature aggregation. Finally, we introduce a modified BT loss that explicitly accommodates win-tie scenarios. This approach regularizes the RM's score distribution for positive samples, providing more nuanced preference signals to alleviate over-focus on a small number of top-scoring samples. Our approach is validated on benchmarks evaluating physical plausibility, subject deformity, and semantic alignment, demonstrating improvements in direct RM evaluation metrics and in the efficacy of post-training on video generation models. Code and benchmark will be publicly available.

cs / cs.LG / cs.CV