超要約: RLエージェントの行動を、LLMでわかりやすく説明するスゴ技!
✨ ギャル的キラキラポイント ✨
● RLエージェントの謎行動を、LLM(すごい言葉のプロ)で解き明かす!🤖➡️🗣️ ● 難しかったRLの世界が、SySLLMのおかげで超見やすくってウケる!📖✨ ● AIの説明がわかりやすくなると、色んなビジネスがもっと楽しくなる予感!💖
詳細解説いくよ~!
続きは「らくらく論文」アプリで
Explaining reinforcement learning agents is challenging because policies emerge from complex reward structures and neural representations that are difficult for humans to interpret. Existing approaches often rely on curated demonstrations that expose local behaviors but provide limited insight into an agent's global strategy, leaving users to infer intent from raw observations. We propose SySLLM (Synthesized Summary using Large Language Models), a framework that reframes policy interpretation as a language-generation problem. Instead of visual demonstrations, SySLLM converts spatiotemporal trajectories into structured text and prompts an LLM to generate coherent summaries describing the agent's goals, exploration style, and decision patterns. SySLLM scales to long-horizon, semantically rich environments without task-specific fine-tuning, leveraging LLM world knowledge and compositional reasoning to capture latent behavioral structure across policies. Expert evaluations show strong alignment with human analyses, and a large-scale user study found that 75.5% of participants preferred SySLLM summaries over state-of-the-art demonstration-based explanations. Together, these results position abstractive textual summarization as a paradigm for interpreting complex RL behavior.