iconLogo
Published:2026/1/2 15:59:42

タイトル & 超要約:線形ソルバーをRLで爆速化!計算コスト削減!

ギャル的キラキラポイント✨ ● 精度と速さを両立する、まさに神テク✨ ● RL(強化学習)で自動調整!天才かよ😳 ● クラウド、AI、色んな分野で大活躍の予感!

詳細解説 ● 背景 線形ソルバー(方程式を解く計算ソフト)の性能UPを目指す研究だよ!精度を上げると計算時間もコストも上がる… って悩みを、強化学習(RL)で解決しちゃうんだって!🎉

● 方法 計算ステップごとに最適な精度をRLで選ぶよ!まるでゲームみたい🎮✨ 線形システムの特性(状況)を考慮して、最適な精度設定を学習!まるでAIのパーフェクトガイド📚

● 結果 計算コストを抑えつつ、高精度をキープ!クラウドとかAIとか、色んな分野で計算爆速になるってこと🚀🌟

続きは「らくらく論文」アプリで

Precision Autotuning for Linear Solvers via Contextual Bandit-Based RL

Erin Carson / Xinye Chen

We propose a reinforcement learning (RL) framework for adaptive precision tuning of linear solvers, and can be extended to general algorithms. The framework is formulated as a contextual bandit problem and solved using incremental action-value estimation with a discretized state space to select optimal precision configurations for computational steps, balancing precision and computational efficiency. To verify its effectiveness, we apply the framework to iterative refinement for solving linear systems $Ax = b$. In this application, our approach dynamically chooses precisions based on calculated features from the system. In detail, a Q-table maps discretized features (e.g., approximate condition number and matrix norm)to actions (chosen precision configurations for specific steps), optimized via an epsilon-greedy strategy to maximize a multi-objective reward balancing accuracy and computational cost. Empirical results demonstrate effective precision selection, reducing computational cost while maintaining accuracy comparable to double-precision baselines. The framework generalizes to diverse out-of-sample data and offers insight into utilizing RL precision selection for other numerical algorithms, advancing mixed-precision numerical methods in scientific computing. To the best of our knowledge, this is the first work on precision autotuning with RL and verified on unseen datasets.

cs / cs.LG / cs.NA / math.NA