iconLogo
Published:2026/1/7 6:09:09

トランスフォーマー、物理に挑む!計算爆速で未来をハッピーに💅💕

超要約: Transformer(変換器)の計算を早くする研究だよ!物理シミュレーションがもっと身近になるかも✨

ギャル的キラキラポイント✨

● Transolver(トランスソルバー)の弱点、見つけちゃった☆ スライス注意(Slice Attention)が邪魔してたみたい! ● 線形注意(Linear Attention)ってやつを使ったら、計算がめっちゃ早くなったよ!🤩 ● 色んな分野で使える!計算時間が短縮されて、製品開発とか色々捗るってことね♪

詳細解説

背景 Transformerって、AI界隈(かいわい)ですごく人気者じゃん?🌟 特に、難しい計算とかを速く解ける「Neural Operator(ニューラルオペレータ)」ってのが注目されてるの。Transolverもその仲間入り!物理の問題を解くのに使えるんだけど、ちょっと計算が遅いのが悩みだったみたい😢

方法 Transolverが使ってる「物理的注意機構(Physics-Attention)」ってのが、どうも計算を遅くしてる原因っぽいってことが判明!👀 特に「スライス注意」ってのが悪さしてるみたい。そこで、もっとシンプルな「線形注意」を使った新しい方法「LinearNO」を開発したんだって!

続きは「らくらく論文」アプリで

Transolver is a Linear Transformer: Revisiting Physics-Attention through the Lens of Linear Attention

Wenjie Hu / Sidun Liu / Peng Qiao / Zhenglun Sun / Yong Dou

Recent advances in Transformer-based Neural Operators have enabled significant progress in data-driven solvers for Partial Differential Equations (PDEs). Most current research has focused on reducing the quadratic complexity of attention to address the resulting low training and inference efficiency. Among these works, Transolver stands out as a representative method that introduces Physics-Attention to reduce computational costs. Physics-Attention projects grid points into slices for slice attention, then maps them back through deslicing. However, we observe that Physics-Attention can be reformulated as a special case of linear attention, and that the slice attention may even hurt the model performance. Based on these observations, we argue that its effectiveness primarily arises from the slice and deslice operations rather than interactions between slices. Building on this insight, we propose a two-step transformation to redesign Physics-Attention into a canonical linear attention, which we call Linear Attention Neural Operator (LinearNO). Our method achieves state-of-the-art performance on six standard PDE benchmarks, while reducing the number of parameters by an average of 40.0% and computational cost by 36.2%. Additionally, it delivers superior performance on two challenging, industrial-level datasets: AirfRANS and Shape-Net Car.

cs / cs.LG / cs.AI