iconLogo
Published:2026/1/7 2:58:19

LRMの「無理!」を解決!アブステイン(回答拒否)能力爆上げ作戦💖

超要約: LRMが「わかんない!」って言えるようにする研究だよ✨

✨ ギャル的キラキラポイント ✨

● LRM(大規模言語モデル)の弱点克服!賢く「I don't know」って言えるようにするんだって👏 ● 内部の動きをチェックして、変な答えを出しそうになったらストップ!賢い~💖 ● AIチャットとかが、もっと安心して使えるようになるかもね😉

詳細解説いくよ~!

続きは「らくらく論文」アプリで

Answering the Unanswerable Is to Err Knowingly: Analyzing and Mitigating Abstention Failures in Large Reasoning Models

Yi Liu / Xiangyu Liu / Zequn Sun / Wei Hu

Large reasoning models (LRMs) have shown remarkable progress on complex reasoning tasks. However, some questions posed to LRMs are inherently unanswerable, such as math problems lacking sufficient conditions. We find that LRMs continually fail to provide appropriate abstentions when confronted with these unanswerable questions. In this paper, we systematically analyze, investigate, and resolve this issue for trustworthy AI. We first conduct a detailed analysis of the distinct response behaviors of LRMs when facing unanswerable questions. Then, we show that LRMs possess sufficient cognitive capabilities to recognize the flaws in these questions. However, they fail to exhibit appropriate abstention behavior, revealing a misalignment between their internal cognition and external response. Finally, to resolve this issue, we propose a lightweight, two-stage method that combines cognitive monitoring with inference-time intervention. Experimental results demonstrate that our method significantly improves the abstention rate while maintaining the overall reasoning performance.

cs / cs.AI / cs.CL