超要約:量子技術でTransformer(変換器)を超強化!IT界に革命起こすよ☆
✨ ギャル的キラキラポイント ✨ ● 量子技術(すっごい計算ができるやつ)でTransformerを進化させた!最先端って感じ💖 ● トレーニングが安定して、表現力も爆上がり!賢くて使いやすいって最強じゃん? ● IT業界の画像認識とか、色んな分野で大活躍の予感!未来が楽しみだね🤩
詳細解説いくよ~!
背景 Transformerって、AI界のスーパースター✨ 自然言語処理とか画像認識で大活躍してるんだよね! でも、計算が不安定になったり、表現力に限界があったり… もっともっとスゴくなれるポテンシャルを秘めてるんだけど、課題もあったの。
続きは「らくらく論文」アプリで
At the core of the Transformer, the softmax normalizes the attention matrix to be right stochastic. Previous research has shown that this often de-stabilizes training and that enforcing the attention matrix to be doubly stochastic (through Sinkhorn's algorithm) consistently improves performance across different tasks, domains and Transformer flavors. However, Sinkhorn's algorithm is iterative, approximative, non-parametric and thus inflexible w.r.t. the obtained doubly stochastic matrix (DSM). Recently, it has been proven that DSMs can be obtained with a parametric quantum circuit, yielding a novel quantum inductive bias for DSMs with no known classical analogue. Motivated by this, we demonstrate the feasibility of a hybrid classical-quantum doubly stochastic Transformer (QDSFormer) that replaces the softmax in the self-attention layer with a variational quantum circuit. We study the expressive power of the circuit and find that it yields more diverse DSMs that better preserve information than classical operators. Across multiple small-scale object recognition tasks, we find that our QDSFormer consistently surpasses both a standard ViT and other doubly stochastic Transformers. Beyond the Sinkformer, this comparison includes a novel quantum-inspired doubly stochastic Transformer (based on QR decomposition) that can be of independent interest. Our QDSFormer also shows improved training stability and lower performance variation suggesting that it may mitigate the notoriously unstable training of ViTs on small-scale data.