Kernel methods in Quantum Machine Learning (QML) have recently gained significant attention as a potential candidate for achieving a quantum advantage in data analysis. Among other attractive properties, when training a kernel-based model one is guaranteed to find the optimal model's parameters due to the convexity of the training landscape. However, this is based on the assumption that the quantum kernel can be efficiently obtained from a quantum hardware. In this work we study the trainability of quantum kernels from the perspective of the resources needed to accurately estimate kernel values. We show that, under certain conditions, values of quantum kernels over different input data can be exponentially concentrated (in the number of qubits) towards some fixed value, leading to an exponential scaling of the number of measurements required for successful training. We identify four sources that can lead to concentration including: the expressibility of data embedding, global measurements, entanglement and noise. For each source, an associated concentration bound of quantum kernels is analytically derived. Lastly, we show that when dealing with classical data, training a parametrized data embedding with a kernel alignment method is also susceptible to exponential concentration. Our results are verified through numerical simulations for several QML tasks. Altogether, we provide guidelines indicating that certain features should be avoided to ensure the efficient evaluation and the trainability of quantum kernel methods.
翻译:Qantum 机器学习(QML) 的量子内核方法最近作为在数据分析中实现量子优势的潜在候选方而受到极大关注。 除其他有吸引力的特性外,当培训以内核为基础的模型,由于培训环境的微妙性,保证能够找到最佳模型参数时,这是基于量子内核能够有效地从量子硬件中获取的假设。 在这项工作中,我们从准确估计内核值所需的资源的角度研究量子内核的可训练性。 我们表明,在某些条件下,不同投入数据的量子内核值可以指数性地集中到某些固定值(qubits数量),导致培训成功培训所需的测量数量指数的指数化。 我们确定可以导致集中的四个来源包括:数据嵌入、全球测量、纠结和噪音的清晰度。 对于每一种来源,都从分析得出量子内核的关联集中。 最后,我们表明,在处理古典数据时,对准性内核数据内核的内核值值值值值值值值值值值值值值值值值值值的数值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值的值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值值