ギャルのみんな~!最強のLLM(大規模言語モデル)も、たま~にウソついちゃうこと、あるよね?😱 でも心配しないで!FaithLensを使えば、LLMがなんでウソついたのか、理由までバッチリわかるんだから✨
✨ ギャル的キラキラポイント ✨
● ウソ発見器!🤖 LLMの「幻覚」(ウソのことね)を、見つけ出すスゴ技! ● なんでウソついたか説明!🤔 理由がわかれば、対策もできるじゃん? ● 低コスト&高性能!💰 お財布にも優しいのに、GPT-4並みの実力だよ!
詳細解説いくよ~!
続きは「らくらく論文」アプリで
Recognizing whether outputs from large language models (LLMs) contain faithfulness hallucination is crucial for real-world applications, e.g., retrieval-augmented generation and summarization. In this paper, we introduce FaithLens, a cost-efficient and effective faithfulness hallucination detection model that can jointly provide binary predictions and corresponding explanations to improve trustworthiness. To achieve this, we first synthesize training data with explanations via advanced LLMs and apply a well-defined data filtering strategy to ensure label correctness, explanation quality, and data diversity. Subsequently, we fine-tune the model on these well-curated training data as a cold start and further optimize it with rule-based reinforcement learning, using rewards for both prediction correctness and explanation quality. Results on 12 diverse tasks show that the 8B-parameter FaithLens outperforms advanced models such as GPT-4.1 and o3. Also, FaithLens can produce high-quality explanations, delivering a distinctive balance of trustworthiness, efficiency, and effectiveness.