最強ギャルAI、降臨~!✨ 今回は「学習の敵対的汚染」について、わかりやすく解説しちゃうよ!
超要約: データに"イタズラ"されると、AIちゃんの学習が変になるって話!
● AIちゃん、データの質にめっちゃ左右されるってコト!🥺 ● "データ汚染"っていうイタズラが、AIちゃんの性能を落とすらしい…😈 ● でも、この研究で対策が見つかるかも!期待しちゃうね~♪
続きは「らくらく論文」アプリで
We study the extent to which standard machine learning algorithms rely on exchangeability and independence of data by introducing a monotone adversarial corruption model. In this model, an adversary, upon looking at a "clean" i.i.d. dataset, inserts additional "corrupted" points of their choice into the dataset. These added points are constrained to be monotone corruptions, in that they get labeled according to the ground-truth target function. Perhaps surprisingly, we demonstrate that in this setting, all known optimal learning algorithms for binary classification can be made to achieve suboptimal expected error on a new independent test point drawn from the same distribution as the clean dataset. On the other hand, we show that uniform convergence-based algorithms do not degrade in their guarantees. Our results showcase how optimal learning algorithms break down in the face of seemingly helpful monotone corruptions, exposing their overreliance on exchangeability.