论文标题

朝着物理上可实现的量子神经网络

Toward Physically Realizable Quantum Neural Networks

论文作者

Heidari, Mohsen, Grama, Ananth, Szpankowski, Wojciech

论文摘要

最近对量子神经网络(QNN)以及它们在不同领域的应用引起了重大兴趣。 QNN的当前解决方案对其可伸缩性提出了重大挑战,以确保满足量子力学的假设,并且网络在物理上可以实现。 QNNS的指数状态空间对培训程序的可扩展性构成了挑战。无克隆的原理禁止制作多个训练样本的副本,并且测量值假设导致了非确定性损失函数。因此,尚不清楚依赖于每个样本的几个副本进行训练QNN的重复测量的现有方法的物理可靠性和效率尚不清楚。本文提出了一个新的QNN模型,依赖于量子感知器(QPS)传递功能的带限制的傅立叶扩展来设计可扩展的训练程序。通过随机量子随机梯度下降技术增强了该训练程序,从而消除了对样品复制的需求。我们表明,即使由于量子测量引起的非确定性,这种训练程序在预期中也会收敛到真正的最小值。我们的解决方案具有许多重要的好处:(i)使用具有浓缩傅立叶功率谱的QPS,我们表明可以使QNN的训练程序可扩展; (ii)它消除了重新采样的需求,从而与无禁止的规则保持一致; (iii)增强了整体培训过程的数据效率,因为每个数据样本每年都要处理一次。我们为我们的模型和方法的可伸缩性,准确性和数据效率提供了详细的理论基础。我们还通过一系列数值实验来验证方法的实用性。

There has been significant recent interest in quantum neural networks (QNNs), along with their applications in diverse domains. Current solutions for QNNs pose significant challenges concerning their scalability, ensuring that the postulates of quantum mechanics are satisfied and that the networks are physically realizable. The exponential state space of QNNs poses challenges for the scalability of training procedures. The no-cloning principle prohibits making multiple copies of training samples, and the measurement postulates lead to non-deterministic loss functions. Consequently, the physical realizability and efficiency of existing approaches that rely on repeated measurement of several copies of each sample for training QNNs are unclear. This paper presents a new model for QNNs that relies on band-limited Fourier expansions of transfer functions of quantum perceptrons (QPs) to design scalable training procedures. This training procedure is augmented with a randomized quantum stochastic gradient descent technique that eliminates the need for sample replication. We show that this training procedure converges to the true minima in expectation, even in the presence of non-determinism due to quantum measurement. Our solution has a number of important benefits: (i) using QPs with concentrated Fourier power spectrum, we show that the training procedure for QNNs can be made scalable; (ii) it eliminates the need for resampling, thus staying consistent with the no-cloning rule; and (iii) enhanced data efficiency for the overall training process since each data sample is processed once per epoch. We present a detailed theoretical foundation for our models and methods' scalability, accuracy, and data efficiency. We also validate the utility of our approach through a series of numerical experiments.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源