iconLogo
Published:2025/12/16 10:51:18

ゼロショットFMのコスパ調査!🚀

超要約: ゼロショット型AIの性能と、使うお金(ハードウェアコスト)の関係を調べた研究だよ!💰✨

🌟 ギャル的キラキラポイント✨ ● ゼロショットって、学習なしで使えるAIのこと!✨ ● パフォーマンスだけじゃなく、お金も大事って話💰 ● いろんなモデルを比較してて、参考になる~!🤔

詳細解説いくね~!💖

背景 最近、色んな情報から学習しなくても使える「ゼロショット型Foundation Model(FM)」がすごい勢いで進化してるんだけど…🤔✨ 性能は良いけど、使うのにどれくらいお金かかるのか、ちゃんと調べた研究って少なかったの!

続きは「らくらく論文」アプリで

Light-Weight Benchmarks Reveal the Hidden Hardware Cost of Zero-Shot Tabular Foundation Models

Ishaan Gangwani / Aayam Bansal

Zero-shot foundation models (FMs) promise training-free prediction on tabular data, yet their hardware footprint remains poorly characterized. We present a fully reproducible benchmark that reports test accuracy together with wall-clock latency, peak CPU RAM, and peak GPU VRAM on four public datasets: Adult-Income, Higgs-100k, Wine-Quality, and California-Housing. Two open FMs (TabPFN-1.0 and TabICL-base) are compared against tuned XGBoost, LightGBM, and Random Forest baselines on a single NVIDIA T4 GPU. The tree ensembles equal or surpass FM accuracy on three datasets while completing full-test batches in <= 0.40 s and <= 150 MB RAM, using zero VRAM. TabICL achieves a 0.8 percentage-point gain on Higgs but requires roughly 40,000 times more latency (960 s) and 9 GB VRAM. TabPFN matches tree-model accuracy on Wine and Housing but peaks at 4 GB VRAM and cannot process the full 100k-row Higgs table. These results quantify the substantial hardware-versus-accuracy trade-offs in current tabular FMs and provide an open baseline for future efficiency-oriented research.

cs / cs.LG / cs.AI