超絶賢いレコメンドシステムで、ITビジネスをブチ上げ!
タイトル & 超要約 CEMG!協調とマルチモーダル(色んな情報)を合体して、レコメンド爆速化!
ギャル的キラキラポイント✨ ● 他の人の"いいね!"を参考に、アイテム(商品とか)のオススメ度を計算するんだって! ● 画像とか文章とか、色んな情報を合体させて、アイテムの魅力を最大限に伝えるよ! ● LLM(大規模言語モデル)って、最強のAIを使って、パーソナライズ(自分に合った)レコメンドを実現💖
詳細解説
続きは「らくらく論文」アプリで
Generative recommendation models often struggle with two key challenges: (1) the superficial integration of collaborative signals, and (2) the decoupled fusion of multimodal features. These limitations hinder the creation of a truly holistic item representation. To overcome this, we propose CEMG, a novel Collaborative-Enhaned Multimodal Generative Recommendation framework. Our approach features a Multimodal Fusion Layer that dynamically integrates visual and textual features under the guidance of collaborative signals. Subsequently, a Unified Modality Tokenization stage employs a Residual Quantization VAE (RQ-VAE) to convert this fused representation into discrete semantic codes. Finally, in the End-to-End Generative Recommendation stage, a large language model is fine-tuned to autoregressively generate these item codes. Extensive experiments demonstrate that CEMG significantly outperforms state-of-the-art baselines.