iconLogo
Published:2025/12/3 22:06:08

離散データ生成モデル、ついに非漸近的(ひぜんきんきんてき)な収束保証!🎉

超要約:離散データ(画像とかテキストとか)を生成するAIモデルが、ちゃんと動くって保証されたってコト!✨

🌟 ギャル的キラキラポイント✨ ● 離散データ(とぎれとぎれのデータ)向けのAIモデルが、もっとちゃんと動くように💖 ● 計算量(AIの頭脳の負担)を減らして、スピーディーになったよ😎 ● 色んな種類の離散データに対応できるようになったから、マジ卍!😍

詳細解説 ● 背景 最近のAIブームで、画像とか文章とか、色んなもんをAIが作れるようになってきたじゃん? でも、そういうデータ(離散データ)を生成するAIは、まだ不安定だったり、計算量が多かったりしたんだよね🥺

● 方法 「マスク付き拡散モデル」と「ランダムウォーク拡散モデル」っていう2つのモデルを使って、ちゃんと収束する(良い結果が出る)ことを理論的に証明したんだって! 計算量も減らしたらしい!😳

続きは「らくらく論文」アプリで

Non-Asymptotic Convergence of Discrete Diffusion Models: Masked and Random Walk dynamics

Giovanni Conforti / Alain Durmus / Le-Tuyet-Nhi Pham / Gael Raoul

Diffusion models for continuous state spaces based on Gaussian noising processes are now relatively well understood, as many works have focused on their theoretical analysis. In contrast, results for diffusion models on discrete state spaces remain limited and pose significant challenges, particularly due to their combinatorial structure and their more recent introduction in generative modelling. In this work, we establish new and sharp convergence guarantees for three popular discrete diffusion models (DDMs). Two of these models are designed for finite state spaces and are based respectively on the random walk and the masking process. The third DDM we consider is defined on the countably infinite space $\mathbb{N}^d$ and uses a drifted random walk as its forward process. For each of these models, the backward process can be characterized by a discrete score function that can, in principle, be estimated. However, even with perfect access to these scores, simulating the exact backward process is infeasible, and one must rely on approximations. In this work, we study Euler-type approximations and establish convergence bounds in both Kullback-Leibler divergence and total variation distance for the resulting models, under minimal assumptions on the data distribution. In particular, we show that the computational complexity of each method scales linearly in the dimension, up to logarithmic factors. Furthermore, to the best of our knowledge, this study provides the first non-asymptotic convergence guarantees for these noising processes that do not rely on boundedness assumptions on the estimated score.

cs / cs.LG / stat.ML