超要約: 質問をSQLに変換する技術を、データ不足でも精度爆上げするCA(Companion Agents)が提案されたよ!🚀
✨ ギャル的キラキラポイント ✨
● データ不足でもOK! データベースの知識を補完(ほかん)して、質問への答えを見つけ出すのがスゴすぎ💖 ● CAちゃんたちは3人組! データベースの構造(こうぞう)とかを分析して、質問に合った答えを生成するよ🎶 ● IT企業が抱えるデータ分析のハードルを下げて、誰でもデータ活用できるようになる未来が来るかも😍
詳細解説
続きは「らくらく論文」アプリで
Large-scale Text-to-SQL benchmarks such as BIRD typically assume complete and accurate database annotations as well as readily available external knowledge, which fails to reflect common industrial settings where annotations are missing, incomplete, or erroneous. This mismatch substantially limits the real-world applicability of state-of-the-art (SOTA) Text-to-SQL systems. To bridge this gap, we explore a database-centric approach that leverages intrinsic, fine-grained information residing in relational databases to construct missing evidence and improve Text-to-SQL accuracy under annotation-scarce conditions. Our key hypothesis is that when a query requires multi-step reasoning over extensive table information, existing methods often struggle to reliably identify and utilize the truly relevant knowledge. We therefore propose to "cache" query-relevant knowledge on the database side in advance, so that it can be selectively activated at inference time. Based on this idea, we introduce Companion Agents (CA), a new Text-to-SQL paradigm that incorporates a group of agents accompanying database schemas to proactively mine and consolidate hidden inter-table relations, value-domain distributions, statistical regularities, and latent semantic cues before query generation. Experiments on BIRD under the fully missing evidence setting show that CA recovers +4.49 / +4.37 / +14.13 execution accuracy points on RSL-SQL / CHESS / DAIL-SQL, respectively, with larger gains on the Challenging subset +9.65 / +7.58 / +16.71. These improvements stem from CA's automatic database-side mining and evidence construction, suggesting a practical path toward industrial-grade Text-to-SQL deployment without reliance on human-curated evidence.