Papers
Topics
Authors
Recent
Search
2000 character limit reached

QAHAN: A Quantum Annealing Hard Attention Network

Published 30 Dec 2024 in quant-ph | (2412.20930v1)

Abstract: Hard Attention Mechanisms (HAMs) effectively filter essential information discretely and significantly boost the performance of machine learning models on large datasets. Nevertheless, they confront the challenge of non-differentiability, which raises the risk of convergence to a local optimum. Quantum Annealing (QA) is expected to solve the above dilemma. We propose a Quantum Annealing Hard Attention Mechanism (QAHAM) for faster convergence to the global optimum without the need to compute gradients by exploiting the quantum tunneling effect. Based on the above theory, we construct a Quantum Annealing Hard Attention Network (QAHAN) on D-Wave and Pytorch platforms for MNIST and CIFAR-10 multi-classification. Experimental results indicate that the QAHAN converges faster, exhibits smoother accuracy and loss curves, and demonstrates superior noise robustness compared to two traditional HAMs. Predictably, our scheme accelerates the convergence between the fields of quantum algorithms and machine learning, while advancing the field of quantum machine vision.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (1)

Collections

Sign up for free to add this paper to one or more collections.