Quantinuum President and COO Tony Uttley announced three major accomplishments during his keynote address at the IEEE Quantum Week event in Colorado last week.
The three milestones, representing actionable acceleration for the quantum computing eco-system, are: (i) new arbitrary angle gate capabilities on the H-series hardware, (ii) another QV record for the System Model H1 hardware, and (iii) over 500,000 downloads of Quantinuum’s open-sourced TKET, a world-leading quantum software development kit (SDK).
The announcements were made during Uttley’s keynote address titled, “A Measured Approach to Quantum Computing.”
These advancements are the latest examples of the company’s continued demonstration of its leadership in the quantum computing community.
“Quantinuum is accelerating quantum computing’s impact to the world,” Uttley said. “We are making significant progress with both our hardware and software, in addition to building a community of developers who are using our TKET SDK.”
This latest quantum volume measurement of 8192 is particularly noteworthy and is the second time this year Quantinuum has published a new QV record on their trapped-ion quantum computing platform, the System Model H1, Powered by Honeywell.
A key to achieving this latest record is the new capability of directly implementing arbitrary angle two-qubit gates. For many quantum circuits, this new way of doing a two-qubit gate allows for more efficient circuit construction and leads to higher fidelity results.
Dr. Brian Neyenhuis, Director of Commercial Operations at Quantinuum, said, “This new capability allows for several user advantages. In many cases, this includes shorter interactions with the qubits, which lowers the error rate. This allows our customers to run long computations with less noise.”
These arbitrary angle gates build on the overall design strength of the trapped-ion architecture of the H1, Neyenhuis said.
“With the quantum-charged coupled device (QCCD) architecture, interactions between qubits are very simple and can be limited to a small number of qubits which means we can precisely control the interaction and don’t have to worry about additional crosstalk,” he said.
This new gate design represents a third method for Quantinuum to improve the efficiency of the H1 generation, said Dr. Jenni Strabley, Senior Director of Offering Management at Quantinuum.
“Quantinuum’s goal is to accelerate quantum computing. We know we have to make the hardware better and we have to make the algorithms smarter, and we’re doing that,” she said. “Now we can also implement the algorithms more efficiently on our H1 with this new gate design.”
Currently, researchers can do single qubit gates – rotations on a single qubit – or a fully entangling two-qubit gate. It’s possible to build any quantum operation out of just those building blocks.
With arbitrary angle gates, instead of just having a two-qubit gate that's fully entangling, scientists can use a two-qubit gate that is partially entangling.
“There are many algorithms where you want to evolve the quantum state of the system one tiny step at a time. Previously, if you wanted a tiny bit of entanglement for some small time step, you had to entangle it all the way, rotate it a little bit, and then unentangle it almost all the way back,” Neyenhuis said. “Now we can just add this tiny little bit of entanglement natively and then go to the next step of the algorithm.”
There are other algorithms where this arbitrary angle two-qubit gate is the natural building block, according to Neyenhuis. One example is the quantum Fourier transform. Using arbitrary angle two-qubit gates cuts the number of two-qubit gates (and the overall error) in half, drastically improving the fidelity of the circuit. Researchers can use this new gate design to run harder problems that resulted in catastrophic errors in previous experiments.
“By going to an arbitrary angle gate, in addition to cutting the number of two-qubit gates in half, the error we get per gate is lower because it scales with the amplitude of that gate,” Neyenhuis said.
This is a powerful new capability, particularly for noisy intermediate-scale quantum algorithms. Another demonstration from the Quantinuum team was to use arbitrary angle two-qubit gates to study non-equilibrium phase transitions, the technical details of which are available on the arXiv here.
“For the algorithms that we are going to want to run in this NISQ regime that we're in right now, this is a more efficient way to run your algorithm,” Neyenhuis said. “There are lots of different circuits you would want to run where this arbitrary angle gate gives you a fairly significant increase in the fidelity of your overall circuit. This capability also allows for a speed up in the circuit execution by removing unneeded gates, which ultimately reduces the time of executing a job on our machines.”
Researchers working with machine learning algorithms, variational algorithms, and time evolution algorithms would see the most benefit from these new gates. This advancement is particularly relevant for simulating the dynamics of other quantum systems.
“This just gave us a big win in fidelity because we can run the sort of interaction you're after natively, rather than constructing it out of some other Lego blocks,” Neyenhuis said.
Quantum volume tests require running arbitrary circuits. At each slice of the quantum volume circuit, the qubits are randomly paired up and a complex two-qubit operation is performed. This SU(4) gate can be constructed more efficiently using the arbitrary angle two-qubit gate, lowering the error at each step of the algorithm.
The H1-1’s quantum volume of 8192 is due in part to the implementation of arbitrary angle gates and the continued reduction in error rates. Quantinuum’s last quantum volume increase was in April when the System Model H1-2 doubled its performance to become the first commercial quantum computer to pass Quantum Volume 4096.
This new increase is the seventh time in two years that Quantinuum’s H-Series hardware has set an industry record for measured quantum volume as it continues to achieve its goal of 10X annual improvement.
Quantum volume, a benchmark introduced by IBM in 2019, is a way to measure the performance of a quantum computer using randomized circuits, and is a frequently used metric across the industry.
Quantinuum has also achieved another milestone: over 500,000 downloads of TKET.
TKET is an advanced software development kit for writing and running programs on gate-based quantum computers. TKET enables developers to optimize their quantum algorithms, reducing the computational resources required, which is important in the NISQ era.
TKET is open source and accessible through the PyTKET Python package. The SDK also integrates with major quantum software platforms including Qiskit, Cirq and Q#. TKET has been available as an open source language for almost a year.
This universal availability and TKET’s portability across many quantum processors are critical for building a community of developers who can write quantum algorithms. The number of downloads includes many companies and academic institutions which account for multiple users.
Quantinuum CEO Ilyas Khan said, “Whilst we do not have the exact number of users of TKET, it is clear that we are growing towards a million people around the world who have taken advantage of a critical tool that integrates across multiple platforms and makes those platforms perform better. We continue to be thrilled by the way that TKET helps democratize as well as accelerate innovation in quantum computing.”
Arbitrary angle two-qubit gates and other recent Quantinuum advances are all built into TKET.
“TKET is an evolving platform and continues to take advantage of these new hardware capabilities,” said Dr. Ross Duncan, Quantinuum’s Head of Quantum Software. “We’re excited to put these new capabilities into the hands of the rapidly increasing number of TKET users around the world.”
The average single-qubit gate fidelity for this milestone was 99.9959(5)%, the average two-qubit gate fidelity was 99.71(3)% with fully connected qubits, and state preparation and measurement fidelity was 99.72(1)%. The Quantinuum team ran 220 circuits with 90 shots each, using standard QV optimization techniques to yield an average of 175.2 arbitrary angle two-qubit gates per circuit.
The System Model H1-1 successfully passed the quantum volume 8192 benchmark, outputting heavy outcomes 69.33% of the time, with a 95% confidence interval lower bound of 68.38% which is above the 2/3 threshold.
Quantinuum, the world’s largest integrated quantum company, pioneers powerful quantum computers and advanced software solutions. Quantinuum’s technology drives breakthroughs in materials discovery, cybersecurity, and next-gen quantum AI. With over 500 employees, including 370+ scientists and engineers, Quantinuum leads the quantum computing revolution across continents.
In the world of physics, ideas can lie dormant for decades before revealing their true power. What begins as a quiet paper in an academic journal can eventually reshape our understanding of the universe itself.
In 1993, nestled deep in the halls of Yale University, physicist Subir Sachdev and his graduate student Jinwu Ye stumbled upon such an idea. Their work, originally aimed at unraveling the mysteries of “spin fluids”, would go on to ignite one of the most surprising and profound connections in modern physics—a bridge between the strange behavior of quantum materials and the warped spacetime of black holes.
Two decades after the paper was published, it would be pulled into the orbit of a radically different domain: quantum gravity. Thanks to work by renowned physicist Alexei Kitaev in 2015, the model found new life as a testing ground for the mind-bending theory of holography—the idea that the universe we live in might be a projection, from a lower-dimensional reality.
Holography is an exotic approach to understanding reality where scientists use holograms to describe higher dimensional systems in one less dimension. So, if our world is 3+1 dimensional (3 spatial directions plus time), there exists a 2+1, or 3-dimensional description of it. In the words of Leonard Susskind, a pioneer in quantum holography, "the three-dimensional world of ordinary experience—the universe filled with galaxies, stars, planets, houses, boulders, and people—is a hologram, an image of reality coded on a distant two-dimensional surface."
The “SYK” model, as it is known today, is now considered a quintessential framework for studying strongly correlated quantum phenomena, which occur in everything from superconductors to strange metals—and even in black holes. In fact, The SYK model has also been used to study one of physics’ true final frontiers, quantum gravity, with the authors of the paper calling it “a paradigmatic model for quantum gravity in the lab.”
The SYK model involves Majorana fermions, a type of particle that is its own antiparticle. A key feature of the model is that these fermions are all-to-all connected, leading to strong correlations. This connectivity makes the model particularly challenging to simulate on classical computers, where such correlations are difficult to capture. Our quantum computers, however, natively support all-to-all connectivity making them a natural fit for studying the SYK model.
Now, 10 years after Kitaev’s watershed lectures, we’ve made new progress in studying the SYK model. In a new paper, we’ve completed the largest ever SYK study on a quantum computer. By exploiting our system’s native high fidelity and all-to-all connectivity, as well as our scientific team’s deep expertise across many disciplines, we were able to study the SYK model at a scale three times larger than the previous best experimental attempt.
While this work does not exceed classical techniques, it is very close to the classical state-of-the-art. The biggest ever classical study was done on 64 fermions, while our recent result, run on our smallest processor (System Model H1), included 24 fermions. Modelling 24 fermions costs us only 12 qubits (plus one ancilla) making it clear that we can quickly scale these studies: our System Model H2 supports 56 qubits (or ~100 fermions), and Helios, which is coming online this year, will have over 90 qubits (or ~180 fermions).
However, working with the SYK model takes more than just qubits. The SYK model has a complex Hamiltonian that is difficult to work with when encoded on a computer—quantum or classical. Studying the real-time dynamics of the SYK model means first representing the initial state on the qubits, then evolving it properly in time according to an intricate set of rules that determine the outcome. This means deep circuits (many circuit operations), which demand very high fidelity, or else an error will occur before the computation finishes.
Our cross-disciplinary team worked to ensure that we could pull off such a large simulation on a relatively small quantum processor, laying the groundwork for quantum advantage in this field.
First, the team adopted a randomized quantum algorithm called TETRIS to run the simulation. By using random sampling, among other methods, the TETRIS algorithm allows one to compute the time evolution of a system without the pernicious discretization errors or sizable overheads that plague other approaches. TETRIS is particularly suited to simulating the SYK model because with a high level of disorder in the material, simulating the SYK Hamiltonian means averaging over many random Hamiltonians. With TETRIS, one generates random circuits to compute evolution (even with a deterministic Hamiltonian). Therefore, when applying TETRIS on SYK, for every shot one can just generate a random instance of the Hamiltonain, and generate a random circuit on TETRIS at the same time. This simple approach enables less gate counts required per shot, meaning users can run more shots, naturally mitigating noise.
In addition, the team “sparsified” the SYK model, which means “pruning” the fermion interactions to reduce the complexity while still maintaining its crucial features. By combining sparsification and the TETRIS algorithm, the team was able to significantly reduce the circuit complexity, allowing it to be run on our machine with high fidelity.
They didn’t stop there. The team also proposed two new noise mitigation techniques, ensuring that they could run circuits deep enough without devolving entirely into noise. The two techniques both worked quite well, and the team was able to show that their algorithm, combined with the noise mitigation, performed significantly better and delivered more accurate results. The perfect agreement between the circuit results and the true theoretical results is a remarkable feat coming from a co-design effort between algorithms and hardware.
As we scale to larger systems, we come closer than ever to realizing quantum gravity in the lab, and thus, answering some of science’s biggest questions.
At Quantinuum, we pay attention to every detail. From quantum gates to teleportation, we work hard every day to ensure our quantum computers operate as effectively as possible. This means not only building the most advanced hardware and software, but that we constantly innovate new ways to make the most of our systems.
A key step in any computation is preparing the initial state of the qubits. Like lining up dominoes, you first need a special setup to get meaningful results. This process, known as state preparation or “state prep,” is an open field of research that can mean the difference between realizing the next breakthrough or falling short. Done ineffectively, state prep can carry steep computational costs, scaling exponentially with the qubit number.
Recently, our algorithm teams have been tackling this challenge from all angles. We’ve published three new papers on state prep, covering state prep for chemistry, materials, and fault tolerance.
In the first paper, our team tackled the issue of preparing states for quantum chemistry. Representing chemical systems on gate-based quantum computers is a tricky task; partly because you often want to prepare multiconfigurational states, which are very complex. Preparing states like this can cost a lot of resources, so our team worked to ensure we can do it without breaking the (quantum) bank.
To do this, our team investigated two different state prep methods. The first method uses Givens rotations, implemented to save computational costs. The second method exploits the sparsity of the molecular wavefunction to maximize efficiency.
Once the team perfected the two methods, they implemented them in InQuanto to explore the benefits across a range of applications, including calculating the ground and excited states of a strongly correlated molecule (twisted C_2 H_4). The results showed that the “sparse state preparation” scheme performed especially well, requiring fewer gates and shorter runtimes than alternative methods.
In the second paper, our team focused on state prep for materials simulation. Generally, it’s much easier for computers to simulate materials that are at zero temperature, which is, obviously, unrealistic. Much more relevant to most scientists is what happens when a material is not at zero temperature. In this case, you have two options: when the material is steadily at a given temperature, which scientists call thermal equilibrium, or when the material is going through some change, also known as out of equilibrium. Both are much harder for classical computers to work with.
In this paper, our team looked to solve an outstanding problem: there is no standard protocol for preparing thermal states. In this work, our team only targeted equilibrium states but, interestingly, they used an out of equilibrium protocol to do the work. By slowly and gently evolving from a simple state that we know how to prepare, they were able to prepare the desired thermal states in a way that was remarkably insensitive to noise.
Ultimately, this work could prove crucial for studying materials like superconductors. After all, no practical superconductor will ever be used at zero temperature. In fact, we want to use them at room temperature – and approaches like this are what will allow us to perform the necessary studies to one day get us there.
Finally, as we advance toward the fault-tolerant era, we encounter a new set of challenges: making computations fault-tolerant at every step can be an expensive venture, eating up qubits and gates. In the third paper, our team made fault-tolerant state preparation—the critical first step in any fault-tolerant algorithm—roughly twice as efficient. With our new “flag at origin” technique, gate counts are significantly reduced, bringing fault-tolerant computation closer to an everyday reality.
The method our researchers developed is highly modular: in the past, to perform optimized state prep like this, developers needed to solve one big expensive optimization problem. In this new work, we’ve figured out how to break the problem up into smaller pieces, in the sense that one now needs to solve a set of much smaller problems. This means that now, for the first time, developers can prepare fault-tolerant states for much larger error correction codes, a crucial step forward in the early-fault-tolerant era.
On top of this, our new method is highly general: it applies to almost any QEC code one can imagine. Normally, fault-tolerant state prep techniques must be anchored to a single code (or a family of codes), making it so that when you want to use a different code, you need a new state prep method. Now, thanks to our team’s work, developers have a single, general-purpose, fault-tolerant state prep method that can be widely applied and ported between different error correction codes. Like the modularity, this is a huge advance for the whole ecosystem—and is quite timely given our recent advances into true fault-tolerance.
This generality isn’t just applicable to different codes, it’s also applicable to the states that you are preparing: while other methods are optimized for preparing only the |0> state, this method is useful for a wide variety of states that are needed to set up a fault tolerant computation. This “state diversity” is especially valuable when working with the best codes – codes that give you many logical qubits per physical qubit. This new approach to fault-tolerant state prep will likely be the method used for fault-tolerant computations across the industry, and if not, it will inform new approaches moving forward.
From the initial state preparation to the final readout, we are ensuring that not only is our hardware the best, but that every single operation is as close to perfect as we can get it.
Twenty-five years ago, scientists accomplished a task likened to a biological moonshot: the sequencing of the entire human genome.
The Human Genome Project revealed a complete human blueprint comprising around 3 billion base pairs, the chemical building blocks of DNA. It led to breakthrough medical treatments, scientific discoveries, and a new understanding of the biological functions of our body.
Thanks to technological advances in the quarter-century since, what took 13 years and cost $2.7 billion then can now be done in under 12 minutes for a few hundred dollars. Improved instruments such as next-generation sequencers and a better understanding of the human genome – including the availability of a “reference genome” – have aided progress, alongside enormous advances in algorithms and computing power.
But even today, some genomic challenges remain so complex that they stretch beyond the capabilities of the most powerful classical computers operating in isolation. This has sparked a bold search for new computational paradigms, and in particular, quantum computing.
The Wellcome Leap Quantum for Bio (Q4Bio) challenge is pioneering this new frontier. The program funds research to develop quantum algorithms that can overcome current computational bottlenecks. It aims to test the classical boundaries of computational genetics in the next 3-5 years.
One consortium – led by the University of Oxford and supported by prestigious partners including the Wellcome Sanger Institute, the Universities of Cambridge, Melbourne, and Kyiv Academic University – is taking a leading role.
“The overall goal of the team’s project is to perform a range of genomic processing tasks for the most complex and variable genomes and sequences – a task that can go beyond the capabilities of current classical computers” – Wellcome Sanger Institute press release, July 2025
Earlier this year, the Sanger Institute selected Quantinuum as a technology partner in their bid to succeed in the Q4Bio challenge.
Our flagship quantum computer, System H2, has for many years led the field of commercially available systems for qubit fidelity and consistently holds the global record for Quantum Volume, currently benchmarked at 8,388,608 (223).
In this collaboration, the scientific research team can take advantage of Quantinuum’s full stack approach to technology development, including hardware, software, and deep expertise in quantum algorithm development.
“We were honored to be selected by the Sanger Institute to partner in tackling some of the most complex challenges in genomics. By bringing the world’s highest performing quantum computers to this collaboration, we will help the team push the limits of genomics research with quantum algorithms and open new possibilities for health and medical science.” – Rajeeb Hazra, President and CEO of Quantinuum
At the heart of this endeavor, the consortium has announced a bold central mission for the coming year: to encode and process an entire genome using a quantum computer. This achievement would be a potential world-first and provide evidence for quantum computing’s readiness for tackling real-world use cases.
Their chosen genome, the bacteriophage PhiX174, carries symbolic weight, as its sequencing earned Fred Sanger his second Nobel Prize for Chemistry in 1980. Successfully encoding this genome quantum mechanically would represent a significant milestone for both genomics and quantum computing.
Sooner than many expect, quantum computing may play an essential role in tackling genomic challenges at the very frontier of human health. The Sanger Institute and Quantinuum’s partnership reminds us that we may soon reach an important step forward in human health research – one that could change medicine and computational biology as dramatically as the original Human Genome Project did a quarter-century ago.
“Quantum computational biology has long inspired us at Quantinuum, as it has the potential to transform global health and empower people everywhere to lead longer, healthier, and more dignified lives.” – Ilyas Khan, Founder and Chief Product Officer of Quantinuum