iconLogo
Published:2026/1/7 3:09:54

GIFTでLLM爆上げ!✨ 重要度認識ファインチューニングって何?

超要約:拡散モデルのLLMを、トークンの重要度に応じて調整して、もっと賢くする研究だよ!💖

● 生成の精度がアップ!文章がめっちゃ自然になる🎵 ● 学習が効率的になって、コスパも最強🌟 ● いろんなITサービスで、もっとすごい事ができるようになるかも⁉

詳細解説いくよ~!

背景 最近話題のLLM(大規模言語モデル)って、拡散モデルを使って文章とか画像を作ってるんだよね!✨ でも、従来の学習方法だと、文章を作る時にどの単語が大事か分からなくて、ちょっと残念な結果になることもあったみたい😭

続きは「らくらく論文」アプリで

GIFT: Guided Importance-Aware Fine-Tuning for Diffusion Language Models

Guowei Xu / Wenxin Xu / Jiawang Zhao / Kaisheng Ma

Diffusion models have recently shown strong potential in language modeling, offering faster generation compared to traditional autoregressive approaches. However, applying supervised fine-tuning (SFT) to diffusion models remains challenging, as they lack precise probability estimates at each denoising step. While the diffusion mechanism enables the model to reason over entire sequences, it also makes the generation process less predictable and often inconsistent. This highlights the importance of controlling key tokens that guide the direction of generation. To address this issue, we propose GIFT, an importance-aware finetuning method for diffusion language models, where tokens are assigned different importance weights based on their entropy. Derived from diffusion theory, GIFT delivers substantial gains: across diverse settings including different mainstream training datasets ranging from 1k to 10k in size, utilizing LoRA or full parameter fine-tuning, and training on base or instruct models, GIFT consistently achieves superior overall performance compared to standard SFT on four widely used reasoning benchmarks (Sudoku, Countdown, GSM8K, and MATH-500).

cs / cs.CL