🌟 ギャル的キラキラポイント✨ ● LLMの「使えるけど高い!」を解決する、画期的なフレームワークが登場💖 ● モデルの各部分の「重要度」に着目して、無駄をなくす作戦なの💎 ● 性能を落とさずに、計算コストを大幅削減できる可能性大って、すごくない?😍
詳細解説いくよ~!
背景 LLMって、すごーく賢いけど、使うのにお金かかるじゃん?💰 賢さ(精度)とコスト(計算量)はトレードオフの関係で、両立させるのが難しかったの!🤔
続きは「らくらく論文」アプリで
Constructing a Pareto set is pivotal for navigating the capability--efficiency trade-offs in Large Language Models (LLMs). However, existing merging techniques remain inadequate for this task. Coarse-grained, model-level methods yield only a sparse set of suboptimal solutions, while fine-grained, layer-wise approaches suffer from the curse of dimensionality, rendering the search space computationally intractable. To resolve this dichotomy, we propose Structural Importance Prior Bayesian Model Merging (SIP-BMM), a framework that automatically constructs the LLM Pareto set. SIP-BMM renders high-dimensional layer-wise search tractable by introducing an importance-aware Sparse Axis-Aligned Subspace Bayesian Optimization (SAASBO) strategy. By leveraging a structural importance prior derived from task-vector differences, our method guides SAASBO to automatically identify critical layers, thereby dramatically reducing the effective dimensionality without sacrificing the granularity of full-model control. The entire process is automated within an evolutionary loop driven by the Log-Noisy Expected Hypervolume Improvement ($q$NEHVI) acquisition function. Experiments demonstrate that SIP-BMM discovers a stronger and denser Pareto front than competitive baselines, enabling agile model selection tailored to diverse operational constraints. Code is available at: https://github.com/MiLab-HITSZ/2026-SIPBMM.