Back to News
quantum-computing

Reservoir computing on an analog Rydberg-atom quantum computer

Amazon Braket
Loading...
16 min read
0 likes
⚡ Quantum Brief
QuEra Computing and Amazon Braket demonstrated quantum reservoir computing (QRC) on Rydberg-atom processors, achieving competitive performance in image classification and time series prediction tasks. The QRC approach uses fixed quantum dynamics for feature mapping, requiring only linear readout training—reducing computational costs compared to classical methods. Tests on MNIST and tomato disease datasets showed near-parity with neural networks. For pharmaceutical research, QRC outperformed classical models on small datasets, maintaining accuracy with as few as 100 training samples. UMAP analysis revealed clearer molecular activity clustering in quantum embeddings. Experiments highlighted noise resilience, with position encoding most sensitive to atom fluctuations. Global detuning performed worse due to thermalization limits, while local encoding offered better control. Researchers can replicate results via Amazon Braket tutorials, with QuEra’s work suggesting near-term quantum advantage for niche ML applications where data is scarce or interpretability is critical.
Reservoir computing on an analog Rydberg-atom quantum computer

Summarize this article with:

This post shows how quantum reservoir computing (QRC) can tackle machine learning challenges using Rydberg-atom quantum computers. Readers will learn how QRC works, see its performance on image classification and time series prediction tasks, and understand when it outperforms classical methods—particularly for small datasets in pharmaceutical research. After decades of progress, machine learning (ML) has become a foundation of modern technology, powering applications in areas such as computer vision, financial market prediction, and natural language processing. Despite these advances, ML still struggles with problems of growing scale and complexity. To overcome these limitations, there has been growing interest in exploiting the principles of quantum mechanics to design quantum machine learning (QML) algorithms to tackle ML problems such as image classification or those involving quantum data [1]. Among the many QML approaches proposed, quantum reservoir computing (QRC) has emerged as a promising approach for implementation on near-term quantum hardware [2-6].In this post, we dive deep into recent research done by QuEra Computing and collaborators on QRC [7,8]. Ref. [7] presents a quantum reservoir learning algorithm using Rydberg-atom quantum computers for machine learning tasks. The authors implemented the algorithm experimentally for various tasks such as image classification and time series prediction. The algorithm was subsequently applied numerically in Ref. [8] to gauge the performance of QRC for molecular property learning. The authors observed robust QRC performance on small datasets relevant for pharmaceutical research. Interested readers can reproduce the results on Amazon Braket by following the authors’ tutorials [9, 10]. Reservoir computing is a machine learning paradigm, where the relation between input signals and outputs is modeled by the time dynamics of a non-linear system with fixed parameters called a reservoir. The reservoir maps the input to a high-dimensional space, known as the configuration space of the reservoir, then a readout layer is trained to characterize the time-dependent states of the reservoir and map it to the desired output. Compared to other machine learning methods, one benefit of reservoir computing is that only the readout layer requires training while the reservoir parameters remain fixed, resulting in low training cost. Additionally, reservoir computing is amenable to analog hardware implementations, either classical or quantum, enabling access to the inherent computational capabilities of diverse physical systems [11-13].As an example of classical reservoir computing (CRC), we consider classifying the Modified National Institute of Standards and Technology (MNIST) images of handwritten digits using a chain of Nq classical spins. Each spin is described by a time-dependent vector Si(t) = [Sxi(t), Syi(t), Szi(t)] where 1 ≤ i ≤ Nq labels the spins. To encode an image, we convert it into a Nq-dimensional feature vector x (0 ≤ xi ≤ 1) and set the site-dependent longitudinal magnetic field as Δi = Δmaxxi. With that, we have set up a reservoir whose dynamics are determined by the features of the image. We then sample observables from the evolution of the reservoir, such as the Z-component of all spins Szi(t) and the correlation functions Szi(t)Szj>i(t), which yield a data-embedding vector. Finally, we build such data-embedding vectors for all images and train a linear readout (e.g., linear regression) to perform classification.From the preceding example, we see that a reservoir can be viewed as a programmable analog computer that is used as a subroutine to perform a given machine learning task. The classical spin chain is not the only choice for the reservoir, and a natural alternative is a quantum spin system, which can be viewed as an analog quantum computer. In Ref. [7], the authors considered an analog quantum computer based on Rydberg atoms and implemented a QRC algorithm utilizing the quantum dynamics of the system.The QRC algorithm workflow is summarized in Figure 1. Like its classical counterpart described earlier, it consists of three steps. The algorithm starts by converting the input data, whether an image for classification or a time series for prediction, into a feature vector readily encoded to a two-dimensional Rydberg system. Each Rydberg atom is a two-level system with tunable position ri and is subject to a local detuning Δi(t), which is analogous to the site-dependent longitudinal magnetic field used in the CRC example, and a global detuning Δ(t), analogous to a uniform longitudinal magnetic field. After the feature vector is encoded into these parameters, in the second step, the system then evolves over varying time periods and can be probed by measuring a set of local Pauli-Z observables including and i(t)>. These observables form the data-embedding vectors in a high dimensional space where linear separability becomes prominent. Hence, one can use the data-embedding vectors to train a simple classification model, such as a support vector machine (SVM), in the last step of the workflow.Figure 1 – Overview of the QRC algorithm with Rydberg atoms.Although QRC shares the same workflow as CRC, using quantum systems as reservoirs enables access to a state space beyond product states and enables long-range quantum correlations that are unavailable classically. As a result, QRC can take advantage of the superposition and entanglement unique to quantum systems to tackle challenging machine learning tasks with data of increasing complexity and intricate patterns.In Ref. [7], the authors applied the QRC algorithm to the MNIST classification task by encoding images into either the positions or the local detunings of atoms in a chain. Here we focus on the former method. For two atoms with positions ri and rj, the interaction strength between them is given bywhere is the Rydberg interaction constant. The atoms in the chain are initialized to be uniformly separated by a distance d, and we use position encoding to modulate the atom positions such that the interaction between neighboring atoms readsHere is a task-dependent encoding scale. Once the atom arrangements are determined, the Rydberg system then evolves (for five successive timesteps), after which we measure the time-dependent observables. This process is repeated for all images in the training set, and an SVM is trained with all the measured observables.The test accuracy of the QRC algorithm using position encoding is shown in Figure 2, where the authors compare the performance of QRC with other machine learning methods for the 3/8-MNIST binary classification task. The simulations and experiments are carried out using a chain with nine atoms. The exact simulation of QRC shows similar accuracy in predicting the downscaled images in the test set compared to CRC simulation and a four-layer feedforward neural network (NN). The authors also compare the test accuracy from a linear SVM, which is obtained by removing the reservoir from the workflow (see Figure 1) and training the SVM directly with the feature vectors. The relatively poor performance of the linear SVM illustrates the critical role of introducing nonlinearity via the reservoir, either classical or quantum, in the algorithms.Figure 2 – Comparison of the performance of QRC with other machine learning methods for 3/8-MNIST binary classification task.The authors also consider a different classification task: classifying tomato diseases based on leaf images. Instead of using a one-dimensional chain of atoms, position encoding is applied to a two-dimensional atom array, as shown in Figure 3(a). In this setup, each of the up to Nq = 108 atoms represents a pixel in the downscaled image. As noted, the performance of the QRC algorithm improves when more atoms are used in the encoding. This is indeed the case as shown in Figure 3(b). Although the four-layer NN (with ~20,000 hidden parameters) shows better overall performance than QRC for most of the range of Nq considered, QRC shows better scaling with respect to the number of atoms used in the encoding. Upon increasing the number of shots per datapoint from 100 to 400, the QRC test accuracy improves significantly across all system sizes and reaches similar performance as the four-layer NN for Nq ≥ 90. Note that there are known classical methods outperform the linear SVM or four-layer NN used in the benchmark. Therefore, the proof-of-concept results obtained here do not indicate that QRC performs better than NN or other state-of-the-art classical machine learning methods.Figure 3 – QRC for classifying tomato diseases based on leaf images. (a) Illustration of 2D position encoding. (b) Comparison with other machine learning methods. Here Nq corresponds to the number of pixels in the downscaled image.Because the computational power of reservoir computing comes from the time dynamics of physical systems, it is natural to apply the framework to problems such as time series prediction. In such applications, reservoir computing, which can be viewed as a special case of recurrent neural networks, uses the time dynamics of one physical system to simulate that of another physical system [11-13]. Guided by such intuition, the authors apply the QRC algorithm to a time series forecasting task in Ref. [7], where they study all the encoding methods listed in Figure 1. Here, we describe in more detail the encoding method with global detuning.The time series forecasting task considers the light intensity of a laser in the chaotic regime, as shown in Figure 4(a). Time points between 1000 and 1400 are used as a training set, and subsequent time points are used as a test set. The authors preprocess the training data by moving a sliding window of length Ns across it. In each window the feature vector is defined as x=[lk, lk+1, …, lk+Ns-1], where lj is the intensity value at the j-th time point. This is the so-called one-step prediction task. They then use the feature vector to define the global detuning at different time points as Δ(ti)= Δmaxxi, as illustrated in Figure 4(a), which controls the dynamics of the Rydberg atoms. The same process is repeated for subsequent sliding windows to obtain a data-embedding vector for training the final readout layer. With global detuning encoding, the authors use a chain with Nq atoms that are equally separated from each other. It is important to note that the encoding capability of global detuning encoding is independent of the number of atoms (Nq), which is in sharp contrast to the other two encoding methods. The impact of such a difference to their prediction accuracies is shown in Figure 4(b).The results of the time series forecasting task are shown in Figure 4(b), which compares the three different QRC encodings, for both simulation and experiments. The metric normalized mean-square error (NMSE), quantifies the discrepancy between the predicted and true time series in the test set, and lower values of NMSE mean better prediction accuracies. We can see that encoding with global detuning (labeled as “pulse”) performs worse than the other two encoding methods, position and local, in both simulation and experiments. This is not surprising because the position and local encoding methods provide higher degrees of control for the Rydberg system which results in a reservoir with a more complex configuration space and higher expressibility. The encoding capacity of global detuning is limited by thermalization, a constraint not present in the other encoding methods. Interested readers can find more on the differences in these encoding methods in Refs. [3] and [7].Figure 4 – Performance of QRC for time series forecasting. (a) Illustration of feature vector construction with Ns=9. (b) Comparison of three encoding methods for both simulation and experiment (lower is better).In Ref. [7], the authors also discuss the comparative quantum kernel advantage of QRC for a certain synthetic dataset. They consider QRC as a kernel and apply the quantum kernel method to the 3/8-MNIST binary classification task, followed by retraining and testing the QRC and CRC with the obtained synthetic dataset. Despite the synthetic nature of the resulting dataset, the experiment shows the existence of a dataset for which QRC exhibits better test accuracy compared to CRC.In Ref. [7], the authors also discuss extensively the effects of experimental noise on the QRC algorithm. For example, the difference between the local pulse and position encodings shown in Figure 4(b) can be partially attributed to the shot-to-shot atom position fluctuations. Although these fluctuations are present for all experiments, the position encoding is more susceptible to such noise. Further, quantum dynamics tends to thermalize in the long-time limit, for both the global and local pulse encodings, and this could cause a significant fraction of the feature vector to be encoded in a lossy manner. Ref. [7] provides extensive simulations showing parameter regimes where QRC performs robustly despite noise.In Ref. [8], the authors explore a separate application of QRC in pharmaceutical research: predicting molecular properties from a set of chemical descriptors. They numerically simulate QRC and compare its performance to classical machine learning methods using the Merck Molecular Activity Challenge datasets (MMACD). Their findings suggest that QRC performs better than classical methods when the training dataset is limited, which could be valuable for pharmaceutical applications. Using uniform manifold approximation and projection (UMAP) analysis, the QRC-embedded data exhibits clearer clustering patterns related to molecular activity, demonstrating that quantum reservoir embeddings appear more interpretable in lower dimensions. The research indicates that QRC-enhanced models may offer advantages in biomedical data science, particularly for cases that have small training sets and require interpretable, robust models.The comparison of the performance of QRC and various classical methods are shown in Figure 5. In Figure 5(a), the predictive abilities of the methods are quantified with mean squared error (MSE, smaller is better). The MSE of classical baselines rises steeply as the number of training records decreases, while the QRC embeddings—particularly with two-body observables (labeled as “QRC Embedding” for clarity)—retain comparatively low MSE even at 100 samples. This indicates that QRC performance deteriorates more slowly under data scarcity. Figure 5(b) compares the UMAP 2D projections for the classical reservoir embedding and QRC embedding (with two-body observables). Molecular activity labels for the former method are intermixed with weak cluster separation, making activity patterns hard to distinguish. By contrast, the same labels for QRC form clearer, more compact clusters, indicating that QRC embeddings yield more interpretable, low-dimensional representations that better reflect molecular activity.Figure 5 – Comparison of performances of QRC and classical methods for molecular property prediction. (a) Comparison for the performance on a MMACD dataset. (b) Comparison UMAP 2D projections.This research demonstrates that quantum reservoir computing on Rydberg-atom systems can match or exceed classical methods for specific ML tasks, particularly when training data is limited. Readers working with small datasets in pharmaceutical research or need interpretable models may find in QRC advantages worth exploring.

At Amazon Braket, we aim to enable researchers to more easily test their algorithms and make the most out of various quantum devices. To get started with Amazon Braket, refer to our example notebooks. Researchers from accredited institutions can apply for credits by submitting a research proposal here.[1] D. García, et al. (2022).

Systematic Literature Review: Quantum Machine Learning and its applications.

Computer Science Review 51, 100619 (2024).[2] K. Fujii, et al. (2016). Harnessing disordered ensemble quantum dynamics for machine learning.

Physical Review Applied 8, 024030 (2017).[3] R. Bravo, et al. (2022).

Quantum Reservoir Computing Using Arrays of Rydberg Atoms. PRX Quantum 3, 030325.[4] J. Dudas, et al. (2022). Quantum reservoir neural network implementation on coherently coupled quantum oscillators. npj Quantum Information 9, 1, 64 (2023).[5] T. Yasuda, et al. (2023). Quantum reservoir computing with repeated measurements on superconducting devices. arXiv: 2310.06706.[6] P. Mujal, et al. (2021). Opportunities in quantum reservoir computing and extreme learningmachines. Adv. Quantum Technol. 2100027 (2021).[7] M. Kornjača, et al. (2024). Large-scale quantum reservoir learning with an analog quantum computer. arXiv: 2407.02553.[8] D. Beaulieu, et al. (2024).

Robust Quantum Reservoir Computing for Molecular Property Prediction. Journal of Chemical Information and Modeling 65, 16, 8475 (2025).[9] https://github.com/QuEraComputing/QRC-tutorials[10] https://github.com/QuEraComputing/robust-qrc-molecular-prediction[11] H. Jaeger (2010). The “echo state” approach to analysing and training recurrent neural networks. GMD Report 148, German National Research Center for Information Technology, 2001.[12] W. Maass (2011).

Liquid State Machines: Motivation, Theory, and Applications. Computability in Context. February 2011, 275-296.[13] G. Tanaka, et al. (2019). Recent advances in physical reservoir computing: A review. Neural Networks, 115, 100 (2019).Mao Lin is a Scientist at Amazon Braket. His past research focused on theoretical condensed matter physics, particularly topological phases of matter. He studied at University of Illinois where he received his PhD in Physics, and at National University of Singapore where he received his BS in Physics.Daniel Beaulieu is a data scientist who has worked with multiple public sector clients to provide data science solutions delivering actionable insights. He is focused on exploring quantum AI and machine learning technical use cases and building out proof-of-concepts for creating AI cloud solutions which interface with quantum computers. Dan is currently pursuing a Ph.D. in quantum computing.Milan Kornjaca is a Senior Quantum Scientist in QuEra's Algorithms and Applications team. He did his PhD in condensed matter physics at Iowa State University, initially joining QuEra as an intern focusing on simulating quantum phases of matter on QuEra hardware and subsequently as a full-time scientist with a focus on quantum simulation and quantum machine learning. Milan is passionate about harnessing classically intractable quantum dynamics for real-world problem solving.Peter Komar is an Applied Scientist in the Amazon Braket team where he works on integration of new quantum devices to the Braket service and researches their applications. He holds a PhD in physics from Harvard University where he conducted theoretical research in the field of quantum metrology, searching for practical ways to achieve ultimate measurement accuracy, set by the Heisenberg limit. Peter is passionate about all means of extracting maximal information from limited data.Dr. Shengtao Wang is currently the Head of Quantum Algorithms & Applications at QuEra Computing Inc., which is a leader in commercializing quantum computers using neutral atoms. Dr. Wang is an expert in the development of near-term quantum algorithms and applications, in the areas of quantum optimization, quantum simulation, and quantum machine learning. Prior to his position at QuEra, Dr. Wang was a postdoctoral scholar at the department of Physics at Harvard University, during which he developed near-term quantum optimization algorithms implementable on today’s neutral-atom quantum computers. He received his PhD in Physics at the University of Michigan – Ann Abor.Yaroslav is an Applied Scientist at Amazon Braket. Yaroslav completed his PhD degree in Theoretical Condensed Matter Physics from University of New South Wales (Australia), where he focused on topics encompassing high-temperature superconductivity, quantum phase transitions, and topological phenomena in exotic quantum matter. After completing his PhD, Yaroslav worked as a visiting scientist at Los Alamos National Lab and as a postdoc at Joint Quantum Institute advising tech startups.

At Amazon Braket, Yaroslav’s primary interests span across quantum compilation, benchmarking, applications of machine learning methods to quantum physics / computing, and GenAI. In his free time Yaroslav enjoys hiking, playing chess, reading about AI, and practicing martial arts.

Read Original

Tags

neutral-atom
drug-discovery
quantum-machine-learning
quantum-computing
quantum-hardware

Source Information