ギャル的キラキラポイント✨ ● AIが賢く(かしこく)なるほど、考えすぎちゃう問題(オーバーシンキング)を解決するんだって!✨ ● 思考(しこう)するかしないかをAIが自分で判断(はんだん)!まるで、私達みたいに賢い~!💖 ● 報酬(ほうしゅう)ハッキングっていう、ズルを防いで、AIの能力(のうりょく)を最大限(さいだいげん)に引き出すんだって!😉
詳細解説
リアルでの使いみちアイデア💡
もっと深掘りしたい子へ🔍 キーワード
続きは「らくらく論文」アプリで
Large reasoning models (LRMs) have attracted much attention due to their exceptional performance. However, their performance mainly stems from thinking, a long Chain of Thought (CoT), which significantly increase computational overhead. To address this overthinking problem, existing work focuses on using reinforcement learning (RL) to train hybrid reasoning models that automatically decide whether to engage in thinking or not based on the complexity of the query. Unfortunately, using RL will suffer the the reward hacking problem, e.g., the model engages in thinking but is judged as not doing so, resulting in incorrect rewards. To mitigate this problem, existing works either employ supervised fine-tuning (SFT), which incurs high computational costs, or enforce uniform token limits on non-thinking responses, which yields limited mitigation of the problem. In this paper, we propose Thinking-Based Non-Thinking (TNT). It does not employ SFT, and sets different maximum token usage for responses not using thinking across various queries by leveraging information from the solution component of the responses using thinking. Experiments on five mathematical benchmarks demonstrate that TNT reduces token usage by around 50% compared to DeepSeek-R1-Distill-Qwen-1.5B/7B and DeepScaleR-1.5B, while significantly improving accuracy. In fact, TNT achieves the optimal trade-off between accuracy and efficiency among all tested methods. Additionally, the probability of reward hacking problem in TNT's responses, which are classified as not using thinking, remains below 10% across all tested datasets.