タイトル & 超要約
RevDEQ(リバーシブルディーイキュー)でAIが超進化✨ 計算コスト削減&高性能化でIT業界を盛り上げる研究だよ!
ギャル的キラキラポイント✨
● 勾配計算が正確に!学習が安定して、モデルがもっと賢くなるんだって💖 ● 正則化(モデルを安定させる方法)が不要に!パラメーターが減って、メモリも節約できちゃう✌️ ● 言語処理とか画像認識とか、色んなAIタスクで既存モデルより優秀!マジ卍じゃん?
続きは「らくらく論文」アプリで
Deep Equilibrium Models (DEQs) are an interesting class of implicit model where the model output is implicitly defined as the fixed point of a learned function. These models have been shown to outperform explicit (fixed-depth) models in large-scale tasks by trading many deep layers for a single layer that is iterated many times. However, gradient calculation through DEQs is approximate. This often leads to unstable training dynamics and requires regularisation or many function evaluations to fix. Here, we introduce Reversible Deep Equilibrium Models (RevDEQs) that allow for exact gradient calculation, no regularisation and far fewer function evaluations than DEQs. We show that RevDEQs significantly improve performance on language modelling and image classification tasks against comparable implicit and explicit models.