**New approach has implications for the future of quantum information science.**

In a recent issue of Physical Review A, Argonne researchers reported a new method for alleviating the effects of “noise” in quantum information systems, a challenge scientists around the globe are working to meet in the race toward a new era of quantum technologies. The new method has implications for the future of quantum information science, including quantum computing and quantum sensing.

Many current quantum information applications, such as carrying out an algorithm on a quantum computer, suffer from “decoherence” — a loss of information due to “noise,” which is inherent to quantum hardware. Matthew Otten, a Maria Goeppert Mayer Fellow at Argonne, and Stephen Gray, group leader of Theory and Modeling at the Center for Nanoscale Materials, a U.S. Department of Energy Office of Science User Facility, have developed a new technique that recovers this lost information by repeating the quantum process or experiment many times, with slightly different noise characteristics, and then analyzing the results.

After gathering results by running the process many times in sequence or parallel, the researchers construct a hypersurface where one axis represents the result of a measurement and the other two (or more) axes represent different noise parameters. This hypersurface yields an estimate of the noise-free observable and gives information about the effect of each noise rate.

“It’s like taking a series of flawed photographs,” said Otten. “Each photo has a flaw, but in a different place in the picture. When we compile all the clear pieces from the flawed photos together, we get one clear picture.”

Applying this technique effectively reduces quantum noise without the need for additional quantum hardware.

“This is a versatile technique that can be done with separate quantum systems undergoing the same process at the same time,” said Otten.

“One could create several small quantum devices and run them in parallel,” said Gray. “Using our method, one would combine the results on the hypersurface and generate approximate noise-free observables. The results would help extend the usefulness of the quantum devices before decoherence sets in.”

“We successfully performed a simple demonstration of our method on the Rigetti 8Q-Agave quantum computer,” said Otten. “This class of methods will likely see much use in near-term quantum devices.”

The researchers’ work described above appears in Physical Review A and is entitled “Recovering noise-free quantum observables.”

Otten and Gray have also developed a similar and somewhat less computationally complex process to achieve noise-reduction results based on correcting one qubit at a time to approximate the result for all qubits being simultaneously corrected. A qubit, or quantum bit, is the equivalent in quantum computing to the binary digit or bit used in classical computing.

“In this approach, we assume that the noise can be reduced on each qubit individually, which, while experimentally challenging, leads to a much simpler data processing problem and results in an estimate of the noise-free result,” noted Otten.

This second method was recently published in Nature Partner Journals Quantum Information: “Accounting for errors in quantum algorithms via individual error reduction.”

This research was performed at the Center for Nanoscale Materials, a U.S. Department of Energy User Facility at Argonne, and was supported by the U.S. Department of Energy, Office of Science. Bebop, a high-performance computing cluster operated by the Laboratory Computing Resource Center at Argonne, was used to perform simulations that helped hone the new method and demonstrate it in situations that are not currently available with quantum hardware.

Source: ANL