iconLogo
Published:2026/1/4 15:13:15

AI見破る!文章の未来を照らす!Info-Maskで信頼性爆上げ作戦🚀

超要約: AIが作った文章と人間が書いた文章が混ざったやつ(混合型改ざんテキスト)を、Info-Maskって技術でめっちゃ精度良く見破れるようにしたよ!

✨ ギャル的キラキラポイント ✨

● AIが作った文章を、まるでメイクみたいに「ここはAI!」って見抜けるようになるんだって!👀✨ ● 敵対的攻撃(AIを騙す攻撃)にも強いから、安心してAI技術を使えるようになるよ💪💖 ● 「人間が解釈できる」ってのがアツい!なんでAIって判断したか、理由も教えてくれるの💕

詳細解説いくよ~!

続きは「らくらく論文」アプリで

DAMASHA: Detecting AI in Mixed Adversarial Texts via Segmentation with Human-interpretable Attribution

L. D. M. S. Sai Teja / N. Siva Gopala Krishna / Ufaq Khan / Muhammad Haris Khan / Atul Mishra

In the age of advanced large language models (LLMs), the boundaries between human and AI-generated text are becoming increasingly blurred. We address the challenge of segmenting mixed-authorship text, that is identifying transition points in text where authorship shifts from human to AI or vice-versa, a problem with critical implications for authenticity, trust, and human oversight. We introduce a novel framework, called Info-Mask for mixed authorship detection that integrates stylometric cues, perplexity-driven signals, and structured boundary modeling to accurately segment collaborative human-AI content. To evaluate the robustness of our system against adversarial perturbations, we construct and release an adversarial benchmark dataset Mixed-text Adversarial setting for Segmentation (MAS), designed to probe the limits of existing detectors. Beyond segmentation accuracy, we introduce Human-Interpretable Attribution (HIA overlays that highlight how stylometric features inform boundary predictions, and we conduct a small-scale human study assessing their usefulness. Across multiple architectures, Info-Mask significantly improves span-level robustness under adversarial conditions, establishing new baselines while revealing remaining challenges. Our findings highlight both the promise and limitations of adversarially robust, interpretable mixed-authorship detection, with implications for trust and oversight in human-AI co-authorship.

cs / cs.CL