超要約: 量子ニューラルネットをSWAPテスト回路で強化!表現力UPで未来がアツい🔥
🌟 ギャル的キラキラポイント ● 量子コンピュータでAIが進化! ● SWAPテスト回路ってなんかカワイイ💖 ● IT業界に革命が起きる予感…!
背景 量子コンピュータで動くAI、QNN(量子ニューラルネットワーク)ってあるじゃん?😎 でも、従来のQNNは表現力(性能みたいなもの)がイマイチだったの。古典的なAI技術を活かしきれてなかったからなんだよね💦
方法 SWAPテスト回路っていう、2つの量子状態(情報みたいなもの)の似てる度合いを測る回路に着目👀 これを応用して、表現力を上げる新しいQNNのアーキテクチャを考えたんだって! 古典的なAIの技術も取り入れられるようにしたんだってさ!
続きは「らくらく論文」アプリで
Quantum neural networks (QNNs) based on parametrized quantum circuits are promising candidates for machine learning applications, yet many architectures lack clear connections to classical models, potentially limiting their ability to leverage established classical neural network techniques. We examine QNNs built from SWAP test circuits and discuss their equivalence to classical two-layer feedforward networks with quadratic activations under amplitude encoding. Evaluation on real-world and synthetic datasets shows that while this architecture learns many practical binary classification tasks, it has fundamental expressivity limitations: polynomial activation functions do not satisfy the universal approximation theorem, and we show analytically that the architecture cannot learn the parity check function beyond two dimensions, regardless of network size. To address this, we introduce generalized SWAP test circuits with multiple Fredkin gates sharing an ancilla, implementing product layers with polynomial activations of arbitrary even degree. This modification enables successful learning of parity check functions in arbitrary dimensions as well as binary n-spiral tasks, and we provide numerical evidence that the expressivity enhancement extends to alternative encoding schemes such as angle (Z) and ZZ feature maps. We validate the practical feasibility of our proposed architecture by implementing a classically pretrained instance on the IBM Torino quantum processor, achieving 84% classification accuracy on the three-dimensional parity check despite hardware noise. Our work establishes a framework for analyzing and enhancing QNN expressivity through correspondence with classical architectures, and demonstrates that SWAP test-based QNNs possess broad representational capacity relevant to both classical and potentially quantum learning tasks.