iconLogo
Published:2025/12/3 21:03:45

機械学習モデル、もっと頑丈に!敵の攻撃にも負けないぞ!💪

  1. 研究の目的、めっちゃアゲ⤴︎: 機械学習モデル(MLモデル)の強さ(堅牢性)を上げる研究だよ!敵の攻撃にも負けないようにするの✨
  2. 活性化関数って何者?: いろんな活性化関数を使って、MLモデルがどう強くなるか試すんだって!ReLUだけじゃないんだね!😳
  3. 連邦学習(FL)にも挑戦!: データがバラバラ(non-IID)な状況でも、MLモデルを強くする方法を研究してるってこと!すごい🎉

詳細解説いくよ~!

● 背景 最近のMLモデルは、色んなことに使われてるけど、攻撃されやすいっていう弱点があったの😭 自動運転とか医療とか、間違ったら困る分野でも使われてるから、もっと強くしなきゃ!ReLUっていう有名な活性化関数もあるんだけど、それだけじゃダメみたい🤔

続きは「らくらく論文」アプリで

Studying Various Activation Functions and Non-IID Data for Machine Learning Model Robustness

Long Dang / Thushari Hapuarachchi / Kaiqi Xiong / Jing Lin

Adversarial training is an effective method to improve the machine learning (ML) model robustness. Most existing studies typically consider the Rectified linear unit (ReLU) activation function and centralized training environments. In this paper, we study the ML model robustness using ten different activation functions through adversarial training in centralized environments and explore the ML model robustness in federal learning environments. In the centralized environment, we first propose an advanced adversarial training approach to improving the ML model robustness by incorporating model architecture change, soft labeling, simplified data augmentation, and varying learning rates. Then, we conduct extensive experiments on ten well-known activation functions in addition to ReLU to better understand how they impact the ML model robustness. Furthermore, we extend the proposed adversarial training approach to the federal learning environment, where both independent and identically distributed (IID) and non-IID data settings are considered. Our proposed centralized adversarial training approach achieves a natural and robust accuracy of 77.08% and 67.96%, respectively on CIFAR-10 against the fast gradient sign attacks. Experiments on ten activation functions reveal ReLU usually performs best. In the federated learning environment, however, the robust accuracy decreases significantly, especially on non-IID data. To address the significant performance drop in the non-IID data case, we introduce data sharing and achieve the natural and robust accuracy of 70.09% and 54.79%, respectively, surpassing the CalFAT algorithm, when 40% data sharing is used. That is, a proper percentage of data sharing can significantly improve the ML model robustness, which is useful to some real-world applications.

cs / cs.LG / cs.CV