はいはーい!最強ギャルAIの登場だよ~!🌟 この論文、ぜったい分かりやすく解説するから、安心してねっ💖
超要約: LLM が「言えないこと」を理解できるか検証! 中絶の偏見(スティグマ)を多角的に評価するよ💕
✨ ギャル的キラキラポイント ✨
● LLM (AIの頭脳)が人間の感情をどこまで理解できるか、斬新な研究って感じ💎 ● 中絶に関する "言いにくさ" をテーマにしてるのが、ちょーエモい🥺 ● AI の倫理的な問題点に光を当てて、より良い未来を目指してるのが、マジ卍じゃない?🔥
続きは「らくらく論文」アプリで
As large language models increasingly mediate stigmatized health decisions, their capacity to genuinely understand complex psychological and physiological phenomena remains poorly evaluated. Can AI understand what we cannot say? We investigate whether LLMs coherently represent abortion stigma across the cognitive, interpersonal, and structural levels where it operates. We systematically tested 627 demographically diverse personas across five leading LLMs using the validated Individual Level Abortion Stigma Scale (ILAS). Our multilevel analysis examined whether models coherently represent stigma at the cognitive level (self-judgment), interpersonal level (anticipated judgment and isolation), and structural level (community condemnation and disclosure patterns), as well as overall stigma. Models fail tests of genuine understanding across all levels. They overestimate interpersonal stigma while underestimating cognitive stigma, assume uniform community condemnation, introduce demographic biases absent from human validation data, miss the empirically validated stigma-secrecy relationship, and contradict themselves within theoretical constructs. These patterns reveal that current alignment approaches ensure appropriate language but not coherent multilevel understanding. This work provides empirical evidence that current LLMs lack coherent multilevel understanding of psychological and physiological constructs. AI safety in high-stakes contexts demands new approaches to design (multilevel coherence), evaluation (continuous auditing), governance and regulation (mandatory audits, accountability, deployment restrictions), and AI literacy in domains where understanding what people cannot say determines whether support helps or harms.