ギャル的キラキラポイント✨ ● データが多少荒れてても大丈夫!どんな状況にも強いAIを作る方法だよ♡ ● Adamっていう優秀な最適化手法をベースにしてるから、使いやすさもバッチリ👌 ● 医療とか金融とか、色んな分野で役立つこと間違いなし!未来が楽しみじゃん?
詳細解説
リアルでの使いみちアイデア💡
もっと深掘りしたい子へ🔍 キーワード
続きは「らくらく論文」アプリで
While traditional Deep Learning (DL) optimization methods treat all training samples equally, Distributionally Robust Optimization (DRO) adaptively assigns importance weights to different samples. However, a significant gap exists between DRO and current DL practices. Modern DL optimizers require adaptivity and the ability to handle stochastic gradients, as these methods demonstrate superior performance. Additionally, for practical applications, a method should allow weight assignment not only to individual samples, but also to groups of objects (for example, all samples of the same class). This paper aims to bridge this gap by introducing ALSO $\unicode{x2013}$ Adaptive Loss Scaling Optimizer $\unicode{x2013}$ an adaptive algorithm for a modified DRO objective that can handle weight assignment to sample groups. We prove the convergence of our proposed algorithm for non-convex objectives, which is the typical case for DL models. Empirical evaluation across diverse Deep Learning tasks, from Tabular DL to Split Learning tasks, demonstrates that ALSO outperforms both traditional optimizers and existing DRO methods.