Debunking algorithmic qubits

March 1, 2024
Executive Summary: Quantinuum’s H-Series computers have the highest performance in the industry, verified by multiple widely adopted benchmarks including quantum volume  We demonstrate that an alternative benchmark called algorithmic qubits is deeply flawed, hiding computer performance behind a plurality voting trick and gate compilations that are not widely useful.

Recently a new benchmark called algorithmic qubits (AQ) has started to be confused with quantum volume measurements. Quantum volume (QV) was specifically designed to be hard to “game,” however the algorithmic qubits test turns out to be very susceptible to tricks that can make a quantum computer look much better than it actually is. While it is not clear what can be done to fix the algorithmic qubits test, it is already clear that it is much easier to pass than QV and is a poor substitute for measuring performance. It is also important to note that algorithmic qubits are not the same as logical qubits, which are necessary for full fault-tolerant quantum computing.

Fig. 1: Simulations of the algorithmic qubits (AQ) test with only two-qubit gate errors for two hypothetical machines.  The machines are identical except one has much higher two qubit gate fidelity. The test was run with three different options: (Base) Running the exact circuits as specified by the algorithmic qubits Github repository, (Gate compilation) Running circuits with custom Pytket compiler passes to reduce two-qubit gate counts, and (Gate compilation + plurality voting) Running the compiled circuits and also applying plurality voting error mitigation with voting over 25 random variants each with 100 shots. Note that the quantum volume (QV) of the machines most closely tracks to the “base” case without compilation and plurality voting, but even that base case of AQ can overestimate the QV of the machine.  

To make this point clear, we simulated what algorithmic qubits data would look like for two machines, one clearly much higher performing than the other. We applied two tricks that are typically used when sharing algorithmic qubits results: gate compilation and error mitigation with plurality voting. From the data above, you can see how these tricks are misleading without further information. For example, if you compare data from the higher fidelity machine without any compilation or plurality voting (bottom left) to data from the inferior machine with both tricks (top right) you may incorrectly believe the inferior machine is performing better. Unfortunately, this inaccurate and misleading comparison has been made in the past.  It is important to note that algorithmic qubits uses a subset of algorithms from a QED-C paper that introduced a suite of application oriented tests and created a repository to test available quantum computers.  Importantly, that work explicitly forbids the compilation and error mitigation techniques that are causing the issue here.

As a demonstration of the perils of AQ as a benchmark, we look at data obtained on both Quantinuum’s H2-1 system as well as publicly available data from IonQ’s Forte system.

Fig. 2: Algorithmic qubit data with gate compilation but without plurality voting error mitigation.  Data from smaller qubit and gate counts was omitted from the Quantinuum data as those points do not tend to influence the AQ score.  H2-1 has a measured quantum volume of 216.  Based on this publicly available data from Forte, combined with the AQ simulation data above, we estimate the Forte quantum volume is around 25, although spread in qubit fidelities and details of circuit compilation could skew this estimate.

We reproduce data without any error mitigation from IonQ’s publicly released data in association with a preprint posted to the arXiv, and compare it to data taken on our H2-1 device. Without error mitigation, IonQ Forte achieves an AQ score of 9, whereas Quantinuum H2-1 achieves AQ of 26. Here you can clearly see improved circuit fidelities on the H2-1 device, as one would expect from the higher reported 2Q gate fidelities (average 99.816(5)% for Quantinuum’s H2-1 vs 99.35% for IonQ’s Forte). However, after you apply error mitigation, in this case plurality voting, to both sets of data the picture changes substantially, hiding each underlying computer’s true capabilities.

Fig. 3: Algorithmic qubit data with gate compilation and plurality voting error mitigation. For the H2-1 data plurality voting is done over 25 variants each with 20 shots for every test and qubit number. For Forte it is not clear to us exactly what plurality voting strategy was employed.

Here the H2-1 algorithmic performance still exceeds Forte (from the publicly released data), but the perceived gap has been reduced by error mitigation.  

“Error mitigation, including plurality voting, may be a useful tool for some near-term quantum computing but it doesn’t work for every problem and it’s unlikely to be scalable to larger systems. In order to achieve the lofty goals of quantum computing we’ll need serious device performance upgrades. If we allow error mitigation in benchmarking it will conflate the error mitigation with the underlying device performance. This will make it hard for users to appreciate actual device improvements that translate to all applications and larger problems,” explained Dr. Charlie Baldwin, a leader in Quantinuum’s benchmarking efforts.

There are other issues with the algorithmic qubits test. The circuits used in the test can be reduced to very easy-to-run circuits with basic quantum circuit compilation that are freely available in packages like pytket. For example, the largest phase estimation and amplitude estimation tests required to pass AQ=32 are specified with 992 and 868 entangling gates respectively but applying pytket optimization reduces the circuits to 141 and 72 entangling gates. This is only possible due to choices in constructing the benchmarks and will not be universally available when using the algorithms in applications. Since AQ reports the precompiled gate counts this also may lead users to expect a machine to be able to run many more entangling gates than what is actually possible on the benchmarked hardware.

What makes a good quantum benchmark? Quantum benchmarking is extremely useful in charting the hardware progress and providing roadmaps for future development. However, quantum benchmarking is an evolving field that is still an open area of research. At Quantinuum we believe in testing the limits of our machine with a variety of different benchmarks to learn as much as possible about the errors present in our system and how they affect different circuits. We are open to working with the larger community on refining benchmarks and creating new ones as the field evolves.

To learn more about the Algorithmic Qubits benchmark and the issues with it, please watch this video where Dr. Charlie Baldwin walks us through the details, starting at 32:40.

About Quantinuum

Quantinuum, the world’s largest integrated quantum company, pioneers powerful quantum computers and advanced software solutions. Quantinuum’s technology drives breakthroughs in materials discovery, cybersecurity, and next-gen quantum AI. With over 500 employees, including 370+ scientists and engineers, Quantinuum leads the quantum computing revolution across continents. 

Blog
November 13, 2025
From Memory to Logic

By Dr. Noah Berthusen

The earliest works on quantum error correction showed that by combining many noisy physical qubits into a complex entangled state called a "logical qubit," this state could survive for arbitrarily long times. QEC researchers devote much effort to hunt for codes that function well as "quantum memories," as they are called. Many promising code families have been found, but this is only half of the story.

Being able to keep a qubit around for a long time is one thing, but to realize the theoretical advantages of quantum computing we need to run quantum circuits. And to make sure noise doesn't ruin our computation, these circuits need to be run on the logical qubits of our code. This is often much more challenging than performing gates on the physical qubits of our device, as these "logical gates" often require many physical operations in their implementation. What's more, it often is not immediately obvious which logical gates a code has, and so converting a physical circuit into a logical circuit can be rather difficult.

Some codes, like the famous surface code, are good quantum memories and also have easy logical gates. The drawback is that the ratio of physical qubits to logical qubits (the "encoding rate") is low, and so many physical qubits are required to implement large logical algorithms. High-rate codes that are good quantum memories have also been found, but computing on them is much more difficult. The holy grail of QEC, so to speak, would be a high-rate code that is a good quantum memory and also has easy logical gates. Here, we make progress on that front by developing a new code with those properties.

Building on prior error correcting codes

A recent work from Quantinuum QEC researchers introduced genon codes. The underlying construction method for these codes, called the "symplectic double cover," also provided a way to obtain logical gates that are well suited for Quantinuum's QCCD architecture. Namely, these "SWAP-transversal" gates are performed by applying single qubit operations and relabeling the physical qubits of the device. Thanks to the all-to-all connectivity facilitated through qubit movement on the QCCD architecture, this relabeling can be done in software essentially for free. Combined with extremely high fidelity (~1.2 x10-5) single-qubit operations, the resulting logical gates are similarly high fidelity.

Given the promise of these codes, we take them a step further in our new paper. We combine the symplectic double codes with the [[4,2,2]] Iceberg code using a procedure called "code concatenation". A concatenated code is a bit like nesting dolls, with an outer code containing codes within it---with these too potentially containing codes. More technically, in a concatenated code the logical qubits of one code act as the physical qubits of another code.

The new codes, which we call "concatenated symplectic double codes", were designed in such a way that they have many of these easily-implementable SWAP-transversal gates. Central to its construction, we show how the concatenation method allows us to "upgrade" logical gates in terms of their ease of implementation; this procedure may provide insights for constructing other codes with convenient logical gates. Notably, the SWAP-transversal gate set on this code is so powerful that only two additional operations (logical T and S) are necessary for universal computation. Furthermore, these codes have many logical qubits, and we also present numerical evidence to suggest that they are good quantum memories.

Concatenated symplectic double codes have one of the easiest logical computation schemes, and we didn’t have to sacrifice rate to achieve it. Looking forward in our roadmap, we are targeting hundreds of logical qubits at ~ 1x 10-8 logical error rate by 2029. These codes put us in a prime position to leverage the best characteristics of our hardware and create a device that can achieve real commercial advantage.

technical
All
Blog
November 12, 2025
Quantinuum at SC25: Advancing the Integration of Quantum and High-Performance Computing

Every year, the International Conference for High Performance Computing, Networking, Storage, and Analysis (SC) brings together the global supercomputing community to explore the technologies driving the future of computing.

Join Quantinuum at this year’s conference, taking place November 16th – 21st in St. Louis, Missouri, where we will showcase how our quantum hardware, software, and partnerships are helping define the next era of high-performance and quantum computing.

Visit Quantinuum in the Expo Hall

The Quantinuum team will be on-site at booth #4432 to showcase how we’re building the bridge between HPC and quantum.

  • Live demo unit of our quantum hardware
  • Our new Helios replica, providing an up-close look at the design behind our next-generation system
  • The Helios chip, highlighting the innovation driving the world’s most advanced trapped-ion quantum computers

From Monday through Wednesday, our quantum computing experts will host daily tutorials at our booth on Helios, our next-generation hardware platform, Nexus, our all-in-one quantum computing platform, and Hybrid Workflows, featuring the integration of NVIDIA CUDA-Q with Quantinuum Systems.

Register for a tutorial

Speaking Sessions at SC25

Join our team as they share insights on the opportunities and challenges of quantum integration within the HPC ecosystem:

Panel Session: The Quantum Era of HPC: Roadmaps, Challenges and Opportunities in Navigating the Integration Frontier
November 19th | 10:30 – 12:00pm CST

During this panel session, Kentaro Yamamoto from Quantinuum, will join experts from Lawrence Berkeley National Laboratory, IBM, QuEra, RIKEN, and Pawsey Supercomputing Research Centre to explore how quantum and classical systems are being brought together to accelerate scientific discovery and industrial innovation.

BoF Session: Bridging the Gap: Making Quantum-Classical Hybridization Work in HPC
November 19th | 5:15 – 6:45pm CST

Quantum-classical hybrid computing is moving from theory to reality, yet no clear roadmap exists for how best to integrate quantum processing units (QPUs) into established HPC environments. In this Birds of a Feather discussion, co-led by Quantinuum’s Grahame Vittorini and representatives from BCS, DOE, EPCC, Inria, ORNL NVIDIA, and RIKEN we hope to bring together a global community of HPC practitioners, system architects, quantum computing specialists and workflow researchers, including participants in the Workflow Community Initiative, to assess the state of hybrid integration and identify practical steps toward scalable, impactful deployment.

events
All
Blog
November 5, 2025
Helios Delivers Quantum Advantage with Real-World Impact

Quantinuum’s real world experiment, on the world’s most powerful quantum computer, is the largest of its kind— so large that no amount of classical computing could match it

Figure 1. Real image (not an artist’s depiction) of 98 single atoms (atomic ions) used for computation inside Quantinuum’s Helios quantum computer. The atomic ions are cooled to a fraction of a degree above absolute zero, so that their quantum state can be carefully controlled and manipulated to perform calculations that are very difficult, if not impossible, for classical supercomputers. 

In 1911, a student working under famed physicist Heike Kamerlingh Onnes made a discovery that would rewire our understanding of electricity. The student was studying the electrical resistance of wires, a seemingly simple question that held secrets destined to surprise the world. 

Kamerlingh Onnes had recently succeeded in liquefying helium, a feat so impressive it earned him the Nobel Prize in Physics two years later. With this breakthrough, scientists could now immerse other materials in a cold bath of liquid Helium, cooling things to unprecedented temperatures and observing their behavior.

Many theories existed about what would happen to a wire at such low temperatures. Lord Kelvin predicted that electrons would freeze in place, making the resistance infinite and stopping the conduction of electricity. Others expected resistance to decrease linearly with temperature—a hypothesis that led to thermometer designs still in use today.

When the student cooled a mercury wire to 3.6 degrees above absolute zero, he found something remarkable: the electrical resistivity suddenly vanished.

Onnes quickly devised an ingenious experiment: as a diligent researcher, he knew that he needed to validate these surprising findings. He took a closed loop of wire, set a current running through it, and watched as it flowed endlessly without fading—a type of perpetual motion that seemed to defy everything we know about physics. And so, superconductivity was born. 

More than a century later, all known superconductors still require extreme conditions like brutal cold or high pressure. If we could instead design a material that superconducts at room temperature, and under normal conditions, our world would be profoundly reshaped.  “Room temperature superconductivity”, as it is generally called, would enable a raft of technological breakthroughs from affordable MRI machines to nearly lossless power grids.

Designing such a material means answering many open questions, and scientists are pursuing diverse strategies to find answers. One promising approach is light-induced superconductivity. In one astonishing study, researchers at the Max Planck Institute in Hamburg used light to entice a material that normally superconducts at roughly -180 °C to superconduct at room temperature - but only for a few picoseconds. This effect raised new questions: how does light achieve something that scientists have been grappling with for decades? What is the microscopic mechanism behind this phenomenon? Could understanding it unlock practical room-temperature superconductors?

Nature’s language is mathematics and mathematics is the language of the world’s most powerful quantum computer, Helios

Physics is a surprisingly profound field when you stop to think about it. At its core lies the idea that nature speaks the language of mathematics—and that by discovering the right equations, we can reveal her secrets. As bold as that sounds, history has proven it true time and again. Whenever we peek behind the veil; mathematics is there.

To understand a phenomena like superconductivity, physicists first need a mathematical model, or a set of equations that describe how it works. With the right model, they can predict and even design new superconductors that operate under more practical conditions. This is a key frontier in the search for room temperature superconductors, one of science’s holy grails.

Since the discovery of superconductivity, a lot of work has gone into finding this right model – one that can act as a sort of ‘Rosetta stone’ for harnessing this phenomenon. One of the best bets for describing high temperature superconductors like the one in the Hamburg study is called the “non-equilibrium Fermi-Hubbard” model, which describes how electrons interact and move in a crystal. 

A surprising element of models that describe superconductivity is the prediction that electrons ‘pair up’ when the material becomes superconducting, dancing around in a waltz, two at a time. These pairs are referred to as “cooper pairs” after the famous physicist Leon Cooper. Now, scientists studying superconductors look for “pairing correlations”, a key signature of superconductivity.

Even armed with the Fermi-Hubbard model, light-induced superconductivity has been very difficult to study. The world’s most powerful supercomputers can only handle very small versions, limiting their utility. Even quantum platforms, like analog simulators, limit researchers to observing ‘average’ quantities and obscuring the microscopic details that are crucial for unravelling this mystery.

Light-induced superconductivity has proved challenging to study with quantum computers as well, as doing so requires low error rates, many qubits, and extreme flexibility to measure the fickle symptoms of superconductivity.

That was, until now: Quantinuum’s Helios is one of the first machines in the world able to handle the complexity of the non-equilibrium Fermi-Hibbard model at scales previously out of reach. 

Hopping across the lattice and connecting the dots

Before Helios, we were limited to small explorations of this model, stalling research on this critical frontier. Now, with Helios, we have a quantum computer uniquely suited for this problem. With a novel fermionic encoding and using up to 90 qubits (72 system qubits plus 18 ancilla), Helios can simulate the dynamics of a 6×6 lattice — a system so large that its full quantum state spans over 2^72 dimensions.

Figure 2. The Helios chip, which generates tiny electromagnetic fields to trap single atomic ions hovering above the chip to be used for computation.

Using Helios to study a system like this offers researchers a sort of “qubit-based laboratory.” Capable of handling complex quantum mechanical effects better than classical computers, Helios allows researchers to thoroughly explore phenomena like this without wasting expensive laboratory time and materials, or spending lots of money and energy running it on a supercomputer. 

Our qubit-based laboratory is a dream come true for several reasons. First, it allows arbitrary state preparation – preparing states far from equilibrium, a challenging task for classical computers. Second, it allows for meaningfully long ‘dynamical simulation’ – seeing how the state evolves in time as entanglement spreads and complexity increases. This is notoriously difficult for classical computers, in part due to their difficulty with handling distinctly quantum phenomena like entanglement. Finally, it allows for flexible measurements and experimental parameters – you can measure any observable, including critical “off-diagonal” observables that carry the signature of superconductivity, and simulate any system, such as those with laser pulses or electric fields. 

This last point is the most significant. While analog quantum simulators, like cold atom systems, can take snapshots of atom positions or measure densities, they struggle with off-diagonal observables—the very ones that signal the formation of Cooper pairs in superconductors.

Breaking new ground: a light-induced pairing

In our work, we've simulated three different regimes of the Fermi-Hubbard model and successfully measured non-zero superconducting pairing correlations — a first for any quantum computing platform.

We began by preparing a low-energy state of the model at half-filling — a standard benchmark for testing quantum simulations. Then, using simulated laser pulses or electric fields, we perturbed the system and observed how it responded.

After these perturbations, we measured a notable increase in the so-called “eta” pairing correlations, a mathematical signature of superconducting behavior. These results prove that our computers can help us understand light-induced superconductivity, such as the results from the Max Planck researchers. However, unlike those physical experiments, Helios offers a new level of control and insight. By tuning every aspect of the simulation — from pulse shape, to field strength, to lattice geometry — researchers can explore scenarios that are completely inaccessible to real materials or analog simulators.

Looking to a future where superconductors permeate our lives

Why does any of this matter? If we could predict which materials will become superconducting — and at what temperature, field, or current — it would transform how we search for new superconductors. Instead of trial-and-error in the lab, scientists could design and test new materials digitally first, saving huge amounts of time and money.

In the long run, Helios and its successors will become essential tools for materials science — not just confirming theories but generating new ones. And perhaps, one day, they’ll help us crack the code behind room-temperature superconductors.

Until then, the quantum revolution continues, one entangled pair at a time.

technical
All
corporate
All