タイトル & 超要約 NERモデル拡張の秘密㊙️ 解明! 内部の動きを分析して、AIをもっと賢くする研究だよ!
ギャル的キラキラポイント✨ ● AIちゃんの頭の中、のぞき見👀!内部で何が起きてるか分析するんだって! ● カタストロフィック・フォギング(既存知識の忘却)を防ぐ方法が見つかるかも💖 ● 個人情報保護にも役立つ!AIが賢くなると、もっと安全になるってことね😉
詳細解説
リアルでの使いみちアイデア💡
続きは「らくらく論文」アプリで
Extending Named Entity Recognition (NER) models to new PII entities in noisy spoken-language data is a common need. We find that jointly fine-tuning a BERT model on standard semantic entities (PER, LOC, ORG) and new pattern-based PII (EMAIL, PHONE) results in minimal degradation for original classes. We investigate this "peaceful coexistence," hypothesizing that the model uses independent semantic vs. morphological feature mechanisms. Using an incremental learning setup as a diagnostic tool, we measure semantic drift and find two key insights. First, the LOC (location) entity is uniquely vulnerable due to a representation overlap with new PII, as it shares pattern-like features (e.g., postal codes). Second, we identify a "reverse O-tag representation drift." The model, initially trained to map PII patterns to 'O', blocks new learning. This is resolved only by unfreezing the 'O' tag's classifier, allowing the background class to adapt and "release" these patterns. This work provides a mechanistic diagnosis of NER model adaptation, highlighting feature independence, representation overlap, and 'O' tag plasticity. Work done based on data gathered by https://www.papernest.com