超要約: AIが作った文章と人間が書いた文章が混ざったやつ(混合型改ざんテキスト)を、Info-Maskって技術でめっちゃ精度良く見破れるようにしたよ!
✨ ギャル的キラキラポイント ✨
● AIが作った文章を、まるでメイクみたいに「ここはAI!」って見抜けるようになるんだって!👀✨ ● 敵対的攻撃(AIを騙す攻撃)にも強いから、安心してAI技術を使えるようになるよ💪💖 ● 「人間が解釈できる」ってのがアツい!なんでAIって判断したか、理由も教えてくれるの💕
詳細解説いくよ~!
続きは「らくらく論文」アプリで
In the age of advanced large language models (LLMs), the boundaries between human and AI-generated text are becoming increasingly blurred. We address the challenge of segmenting mixed-authorship text, that is identifying transition points in text where authorship shifts from human to AI or vice-versa, a problem with critical implications for authenticity, trust, and human oversight. We introduce a novel framework, called Info-Mask for mixed authorship detection that integrates stylometric cues, perplexity-driven signals, and structured boundary modeling to accurately segment collaborative human-AI content. To evaluate the robustness of our system against adversarial perturbations, we construct and release an adversarial benchmark dataset Mixed-text Adversarial setting for Segmentation (MAS), designed to probe the limits of existing detectors. Beyond segmentation accuracy, we introduce Human-Interpretable Attribution (HIA overlays that highlight how stylometric features inform boundary predictions, and we conduct a small-scale human study assessing their usefulness. Across multiple architectures, Info-Mask significantly improves span-level robustness under adversarial conditions, establishing new baselines while revealing remaining challenges. Our findings highlight both the promise and limitations of adversarially robust, interpretable mixed-authorship detection, with implications for trust and oversight in human-AI co-authorship.