InQuanto Integrates NVIDIA cuQuantum for Native GPU Support and Prepares for the Era of Quantum Supercomputing

With quantum progress accelerating, we introduce InQuanto v4.0 and explain how it supports customers and researchers exploring the application of AI, HPC and quantum computing – A.K.A. Quantum Supercomputing – to problems in chemistry and related fields

November 21, 2024

Chemistry plays a central role in the modern global economy, as it has for centuries. From Antoine Lavoisier to Alessandro Volta, Marie Curie to Venkatraman Ramakrishnan, pioneering chemists drove progress in fields such as combustion, electrochemistry, and biochemistry. They contributed to our mastery of critical 21st century materials such as biodegradable plastics, semiconductors, and life-saving pharmaceuticals. 

Advances in high-performance computing (HPC) and AI have brought fundamental and industrial science ever more within the scope of methods like data science and predictive analysis. In modern chemistry, it has become routine for research to be aided by computational models run in silico. Yet, due to their intrinsically quantum mechanical nature, “strongly correlated” chemical systems – those involving strongly interacting electrons or highly interdependent molecular behaviors – prove extremely hard to accurately simulate using classical computers alone. Quantum computers running quantum algorithms are designed to meet this need. Strongly correlated systems turn up in potential applications such as smart materials, high-temperature superconductors, next-generation electronic devices, batteries and fuel cells, revealing the economic potential of extending our understanding of these systems, and the motivation to apply quantum computing to computational chemistry. 

For senior business and research leaders driving value creation and scientific discovery, a critical question is how will the introduction of quantum computers affect the trajectory of computational approaches to fundamental and industrial science?

Introducing InQuanto v4.0

This is the exciting context for our announcement of InQuanto v4.0, the latest iteration of our computational chemistry platform for quantum computers. Developed over many years in close partnership with computational chemists and materials scientists, InQuanto has become an essential tool for teams using the most advanced methods for simulating molecular and material systems. InQuanto v4.0 is packed with powerful updates, including the capability to incorporate NVIDIA’s tensor network methods for large-scale classical simulations supported by graphical processing units (GPUs). 

When researching chemistry on quantum computers, we use classical HPC to perform tasks such as benchmarking, and for classical pre- and post-processing with computational chemistry methods such as density functional theory. This powerful hybrid quantum-classical combination with InQuanto accelerated our work with partners such as BMW Group, Airbus, and Honeywell. Global businesses and national governments alike are gearing up for the use of such hybrid “quantum supercomputers” to become standard practice. 

In a recent technical blog post, we explored the rapid development and deployment of InQuanto for research and enterprise users, offering insights for combining quantum and high-performance classical methods with only a few lines of code. Here, we provide a higher-level overview of the value InQuanto brings to fundamental and industrial research teams. 

InQuanto v4.0 – under the hood

InQuanto v4.0 is the most powerful version to date of our advanced quantum computational chemistry platform. It supports our users in applying quantum and classical computing methods to problems in chemistry and, increasingly, adjacent fields such as condensed matter physics.

Like previous versions of InQuanto, this one offers state-of-the-art algorithms, methods, and error handling techniques out of the box. Quantum error correction and detection have enabled rapid progress in quantum computing, such as groundbreaking demonstrations in partnership with Microsoft, in April and September 2024, of highly reliable “logical qubits”. Qubits are the core information-carrying components of a quantum computer and by forming them into an ensemble, they are more resistant to errors, allowing more complex problems to be tackled while producing accurate results. InQuanto continues to offer leading-edge quantum error detection protocols as standard and supports users to explore the potential of algorithms for fault-tolerant machines.

InQuanto v4.0 also marks the significant step of introducing native support for tensor networks using GPUs to accelerate simulations. In 2022, Quantinuum and NVIDIA teamed up on one of the quantum computing industry’s earliest quantum-classical collaborations. InQuanto v4.0 introduces classical tensor network methods via an interface with NVIDIA's cuQuantum SDK. Interfacing with cuQuantum enables the simulation of many quantum circuits via the use of GPUs for applications in chemistry that were previously inaccessible, particularly those with larger numbers of qubits.

“Hybrid quantum-classical supercomputing is accelerating quantum computational chemistry research. With Quantinuum’s InQuanto v4.0 platform and NVIDIA’s cuQuantum SDK, InQuanto users now have access to unique tensor-network-based methods, enabling large-scale and high-precision quantum chemistry simulations” - Tim Costa, Senior Director of HPC and Quantum Computing at NVIDIA

We are also responding to our users’ needs for more robust, enterprise-grade management of applications and data, by incorporating InQuanto into Quantinuum Nexus. This integration makes it far easier and more efficient to build hybrid workflows, decode and store data, and use powerful analytical methods to accelerate scientific and technical progress in critical fields in natural science.

Adding further capabilities, we recently announced our integration of InQuanto with Microsoft’s Azure Quantum Elements (AQE), allowing users to seamlessly combine AQE’s state-of-the-art HPC and AI methods with the enhanced quantum capabilities of InQuanto in a single workflow. The first end-to-end workflow using HPC, AI and quantum computing was demonstrated by Microsoft using AQE and Quantinuum Systems hardware, achieving chemical accuracy and demonstrating the advantage of logical qubits compared to physical qubits in modeling a catalytic reaction.

Where InQuanto takes us next

In the coming years, we expect to see scientific and economic progress using the powerful combination of quantum computing, HPC, and artificial intelligence. Each of these computing paradigms contributes to our ability to solve important problems. Together, their combined impact is far greater than the sum of their parts, and we recognize that these have the potential to drive valuable computational innovation in industrial use-cases that really matter, such as in energy generation, transmission and storage, and in chemical processes essential to agriculture, transport, and medicine.

Building on our recent hardware roadmap announcement, which supports scientific quantum advantage and a commercial tipping point in 2029, we are demonstrating the value of owning and building out the full quantum computing stack with a unified goal of accelerating quantum computing, integrating with HPC and AI resources where it shows promise, and using the power of the “quantum supercomputer” to make a positive difference in fundamental and industrial chemistry and related domains.

In close collaboration with our customers, we are driving towards systems capable of supporting quantum advantage and unlocking tangible and significant business value.

To access InQuanto today, including Quantinuum Systems and third-party hardware and emulators, visit: https://www.quantinuum.com/products-solutions/inquanto 

To get started with Quantinuum Nexus, which meets all your quantum computing needs across Quantinuum Systems and third-party backends, visit: https://www.quantinuum.com/products-solutions/nexus 

To find out more and access Quantinuum Systems, visit: https://www.quantinuum.com/products-solutions/quantinuum-systems 

About Quantinuum

Quantinuum, the world’s largest integrated quantum company, pioneers powerful quantum computers and advanced software solutions. Quantinuum’s technology drives breakthroughs in materials discovery, cybersecurity, and next-gen quantum AI. With over 500 employees, including 370+ scientists and engineers, Quantinuum leads the quantum computing revolution across continents. 

Blog
November 5, 2025
Introducing Helios: The Most Accurate Quantum Computer in the World
A large room with a large rectangular objectAI-generated content may be incorrect.
Figure 1: A rendering of the Quantinuum Helios system deployed at a customer site. 

We’re pleased to introduce Helios, a technological marvel redefining the possible. 

Building on its predecessor H2, which has already breached quantum advantage, Helios nearly doubles the qubit count and surpasses H2’s industry-leading fidelity, pushing further into the quantum advantage regime than any system before it. With unprecedented capability across its full stack, Helios is the most powerful quantum computer in the world. 

“Helios is a true marvel—a seamless fusion of hardware and software, creating a platform for discovery unlike any other.”

Dr. Rajeeb Hazra, CEO 

Helios’ groundbreaking design and advanced software stack bring quantum programming closer than ever to the ease and flexibility of classical computing—positioning Helios to accelerate commercial adoption. Even before its public debut, Helios had already demonstrated its capabilities as the world’s first enterprise-grade quantum computer. During a two-month early access program, select partners including SoftBank Corp. and JPMorgan Chase conducted commercially relevant research. We also leveraged Helios to perform large-scale simulations in high-temperature superconductivity and quantum magnetism—both with clear pathways to real-world industry applications.

Helios is now available to all customers through our cloud service and on-premise offering, including an option to integrate with NVIDIA GB200 for applications targeting specific end markets.     

A Stellar Quantum Computer 
“You would need to harvest every star in the universe to power a classical machine that could do the same calculations we did with Helios."

Dr. Anthony Ransford, Helios Lead Architect
Figure 2: Random Circuit Sampling (RCS) results on Helios. Running the same calculation classically in the same amount of time would require the power of all the stars in the visible universe.

As we detailed in a benchmarking paper, Helios sets a new standard for quantum computing performance with the highest fidelity ever released to the market. It features 98 fully connected physical qubits with single-qubit gate fidelity of 99.9975% and two-qubit gate fidelity of 99.921% across all qubit pairs—making it the most accurate commercial quantum computer in the world.  

Our fidelity shines in system-level benchmarks, such as Random Circuit Sampling (RCS), famously used by Google to demonstrate quantum supremacy when it performed an RCS task that would take a classical computer “10 septillion years” to replicate. Now, RCS serves as both a benchmark and the minimum standard for serious competitors in the market. Frequently missed in this conversation, however, is the importance of fidelity, or accuracy. That's why, when benchmarking Helios using RCS, we report the fidelity achieved by Helios on circuits of varying complexity (with complexity quantified by power requirements for classical simulation).

Our results show a classical supercomputer would require more power than the Sun—or, in fact, the combined power of all stars in the visible universe—to complete the same task in the same amount of time. In contrast, Helios achieved it using roughly the power of a single data center rack. 

Like its predecessors, H1 and H2, Helios is designed to improve fidelity and overall system performance over time while sustaining competitive leadership through the launch of its successor.

Qubits at a Crossroads
Figure 3: The Helios chip, which generates tiny electromagnetic fields to trap single atomic ions hovering above the chip, which are then used for computation. The Helios chip contains the world’s first commercial ion junction – enabling a huge jump in architectural design and opening the door to true scaling.
"When I first saw the rotatable ion storage ring with a junction and gating legs sketched on a napkin, I loved the idea for its simplicity and efficiency. Seeing it finally realized after all of the team’s hard work has been truly incredible." 

Dr. John Gaebler, Fellow and Chief Scientist, Quantinuum

The Helios ion trap uses tiny currents to generate electromagnetic fields that hold single atomic ions (qubits) hovering above the trap for computation. We introduced a first-of-its-kind “junction”, which acts like a traffic intersection for qubits, enabling efficient routing and improved reliability. This is not only the first commercial implementation of this engineering triumph but it also allows our QCCD (Quantum Charged Coupled Device) architecture to scale, with future systems featuring hundreds of junctions arranged like a city street grid.   

Illustration:The Helios QPU. Ions rotate through the ring storage to the cache and logic zones for gating. Image adapted from benchmarking paper.

Whereas predecessor systems routed qubits using “physical swaps,” requiring sequential sorting, cooling, and gating that prevented parallel operations, the Helios QPU instead resembles a classical architecture with dedicated memory, cache, and computational zones. Like a spinning hard drive, the Helios QPU rotates qubits through ring storage (memory), passes them through the junction into the cache, moves them to logic zones for gating, and moves them to the leg storage while the next batch is processed. Sorting can now be done in parallel with cooling operations, resulting in a processor that is faster and less error prone.  This parallelism will become a hallmark of Quantinuum’s future generations, enabling faster operating speeds.

Animation: This triumph of engineering demonstrates exquisite control over some of nature’s smallest particles in a way the world has never seen; one colleague likened the ions to a “little marching band.”

Quantinuum’s QCCD provides full all-to-all connectivity, giving the Helios QPU significant advantages over “fixed qubit” architectures, such as those used in superconducting systems. Its ability to physically move qubits around and entangle any qubit with any other qubit enables algorithms and error-correcting codes that are functionally impossible for fixed qubit architectures. 

A blue dot pattern on a black backgroundAI-generated content may be incorrect.
Image: Real image of 98 single Barium atoms (atomic ions) used for computation inside Quantinuum’s Helios quantum computer.

We made another “tiny” but significant change: we switched our qubits from ytterbium to barium. Whereas ytterbium largely relied on ultraviolet lasers that are expensive and hard on other components, barium can be manipulated with lasers in the visible part of the spectrum, where mature industrial technology exists, providing a more affordable, reliable and scalable commercial solution.

Barium also naturally allows the quantum computer to detect and remove a certain type of error, known as leakage, at the atomic level. By addressing this error directly, programmers can enhance the performance of their computation.

Delivered on Time – in Real Time

As announced earlier this year, Helios launched with a completely new stack equipped with a new software environment that makes quantum programming feel as intuitive as classical development. 

Our new stack also features a real-time engine that massively improves our capability. With a real-time control system, we are evolving from static, pre-planned circuits to dynamic quantum programs that respond to results on the fly. We can now, for the first time on a quantum computer, interleave GPU-accelerated classical and quantum computations in a single program. 

Our real-time engine also means we have dynamic transport – routing qubits as the moment demands reduces time to solution and diminishes the impact of memory errors.  

Programmers can now use our new quantum programming language, Guppy, to write dynamic circuits that were previously impossible. By combining Guppy with our real-time engine, developers can leverage arbitrary control flow driven by quantum measurements, as well as full classical computation—including loops, higher-order functions, early exits, and dynamic qubit allocation. Far from being mere conveniences, these capabilities are essential stepping stones toward achieving fault-tolerant quantum computing at scale—putting us decisively ahead of the competition.

Fully compatible with industry standards like QIR and tools such as NVIDIA CUDA-Q, Helios bridges classical and quantum computing more seamlessly than ever, making hybrid quantum-classical development simple, natural, and accessible, and establishing Helios as the most programmable, general-purpose quantum computer ever built.  

The Most Logical Path to Fault Tolerance

While everyone else is promising fault-tolerance, we’re delivering it. We are the only company to demonstrate a fully universal fault-tolerant gate set, we’ve demonstrated more codes than anyone else, and our logical fidelities are the best in class.

Now, with 98 physical qubits, we’ve been able to make 94 logical qubits, fully entangled in one of the largest GHZ states ever recorded. We did this with better than break-even fidelity, meaning they outperform physical qubits running the same algorithm. Built on our Iceberg code, published last year in Nature Physics, these logical qubits achieve the industry’s highest encoding efficiency, needing only two ancilla qubits per code block, or roughly a 1:1 physical-to-logical qubit ratio.

With 50 error-detected logical qubits, Helios achieved better than break-even performance, running the largest encoded simulation of quantum magnetism to date—an exceptional example of how users can leverage efficient encodings. This range and flexibility let users tailor the encoding rate to their application: fewer logical qubits deliver higher fidelity for less complex tasks, while larger sets enable more complex simulations.

Helios also produced 48 fully error-corrected logical qubits at a remarkable 2:1 encoding rate, a ratio thought impossible just a few years ago. This super high encoding rate stands in stark contrast to other notable demonstrations from industry peers. For example, the demonstration linked in the previous sentence would need a whopping 4800 qubits to make 48 logical qubits. Our 2:1 encoding rate was achieved through a clever technique called code concatenation, a breakthrough that supports single-shot error correction, transversal logic, and full parallelization—all at 99.99% state preparation and measurement fidelity. 

To extend this performance at scale, all future Quantinuum systems—starting with Helios—will integrate real-time decoding using NVIDIA Grace Hopper GPUs, treating decoding as a dynamic computational process rather than a static lookup. Errors can be corrected as computations run without slowing the logical clock rate. Combined with Guppy, NVIDIA CUDA-Q, and NVQLink, this infrastructure forms the foundation for fault-tolerant, real-time quantum computation, delivering immediate quantum advantage in the near term and a clear path to scalable error-corrected computing. 

We remain the only company to perform a fully universal fault-tolerant gate set, with more error-correcting codes and higher logical fidelities than any other company.

Helios is ready to drive practical, commercial quantum applications across industries. Its unprecedented fidelity, scalability, and programmability give users the tools to tackle problems that were previously out of reach. This is just the beginning, and we look forward to seeing what users and companies will achieve with it. 

technical
All
corporate
All
Blog
November 5, 2025
Helios Delivers Quantum Advantage with Real-World Impact

Quantinuum’s real world experiment, on the world’s most powerful quantum computer, is the largest of its kind— so large that no amount of classical computing could match it

Figure 1. Real image (not an artist’s depiction) of 98 single atoms (atomic ions) used for computation inside Quantinuum’s Helios quantum computer. The atomic ions are cooled to a fraction of a degree above absolute zero, so that their quantum state can be carefully controlled and manipulated to perform calculations that are very difficult, if not impossible, for classical supercomputers. 

In 1911, a student working under famed physicist Heike Kamerlingh Onnes made a discovery that would rewire our understanding of electricity. The student was studying the electrical resistance of wires, a seemingly simple question that held secrets destined to surprise the world. 

Kamerlingh Onnes had recently succeeded in liquefying helium, a feat so impressive it earned him the Nobel Prize in Physics two years later. With this breakthrough, scientists could now immerse other materials in a cold bath of liquid Helium, cooling things to unprecedented temperatures and observing their behavior.

Many theories existed about what would happen to a wire at such low temperatures. Lord Kelvin predicted that electrons would freeze in place, making the resistance infinite and stopping the conduction of electricity. Others expected resistance to decrease linearly with temperature—a hypothesis that led to thermometer designs still in use today.

When the student cooled a mercury wire to 3.6 degrees above absolute zero, he found something remarkable: the electrical resistivity suddenly vanished.

Onnes quickly devised an ingenious experiment: as a diligent researcher, he knew that he needed to validate these surprising findings. He took a closed loop of wire, set a current running through it, and watched as it flowed endlessly without fading—a type of perpetual motion that seemed to defy everything we know about physics. And so, superconductivity was born. 

More than a century later, all known superconductors still require extreme conditions like brutal cold or high pressure. If we could instead design a material that superconducts at room temperature, and under normal conditions, our world would be profoundly reshaped.  “Room temperature superconductivity”, as it is generally called, would enable a raft of technological breakthroughs from affordable MRI machines to nearly lossless power grids.

Designing such a material means answering many open questions, and scientists are pursuing diverse strategies to find answers. One promising approach is light-induced superconductivity. In one astonishing study, researchers at the Max Planck Institute in Hamburg used light to entice a material that normally superconducts at roughly -180 °C to superconduct at room temperature - but only for a few picoseconds. This effect raised new questions: how does light achieve something that scientists have been grappling with for decades? What is the microscopic mechanism behind this phenomenon? Could understanding it unlock practical room-temperature superconductors?

Nature’s language is mathematics and mathematics is the language of the world’s most powerful quantum computer, Helios

Physics is a surprisingly profound field when you stop to think about it. At its core lies the idea that nature speaks the language of mathematics—and that by discovering the right equations, we can reveal her secrets. As bold as that sounds, history has proven it true time and again. Whenever we peek behind the veil; mathematics is there.

To understand a phenomena like superconductivity, physicists first need a mathematical model, or a set of equations that describe how it works. With the right model, they can predict and even design new superconductors that operate under more practical conditions. This is a key frontier in the search for room temperature superconductors, one of science’s holy grails.

Since the discovery of superconductivity, a lot of work has gone into finding this right model – one that can act as a sort of ‘Rosetta stone’ for harnessing this phenomenon. One of the best bets for describing high temperature superconductors like the one in the Hamburg study is called the “non-equilibrium Fermi-Hubbard” model, which describes how electrons interact and move in a crystal. 

A surprising element of models that describe superconductivity is the prediction that electrons ‘pair up’ when the material becomes superconducting, dancing around in a waltz, two at a time. These pairs are referred to as “cooper pairs” after the famous physicist Leon Cooper. Now, scientists studying superconductors look for “pairing correlations”, a key signature of superconductivity.

Even armed with the Fermi-Hubbard model, light-induced superconductivity has been very difficult to study. The world’s most powerful supercomputers can only handle very small versions, limiting their utility. Even quantum platforms, like analog simulators, limit researchers to observing ‘average’ quantities and obscuring the microscopic details that are crucial for unravelling this mystery.

Light-induced superconductivity has proved challenging to study with quantum computers as well, as doing so requires low error rates, many qubits, and extreme flexibility to measure the fickle symptoms of superconductivity.

That was, until now: Quantinuum’s Helios is one of the first machines in the world able to handle the complexity of the non-equilibrium Fermi-Hibbard model at scales previously out of reach. 

Hopping across the lattice and connecting the dots

Before Helios, we were limited to small explorations of this model, stalling research on this critical frontier. Now, with Helios, we have a quantum computer uniquely suited for this problem. With a novel fermionic encoding and using up to 90 qubits (72 system qubits plus 18 ancilla), Helios can simulate the dynamics of a 6×6 lattice — a system so large that its full quantum state spans over 2^72 dimensions.

Figure 2. The Helios chip, which generates tiny electromagnetic fields to trap single atomic ions hovering above the chip to be used for computation.

Using Helios to study a system like this offers researchers a sort of “qubit-based laboratory.” Capable of handling complex quantum mechanical effects better than classical computers, Helios allows researchers to thoroughly explore phenomena like this without wasting expensive laboratory time and materials, or spending lots of money and energy running it on a supercomputer. 

Our qubit-based laboratory is a dream come true for several reasons. First, it allows arbitrary state preparation – preparing states far from equilibrium, a challenging task for classical computers. Second, it allows for meaningfully long ‘dynamical simulation’ – seeing how the state evolves in time as entanglement spreads and complexity increases. This is notoriously difficult for classical computers, in part due to their difficulty with handling distinctly quantum phenomena like entanglement. Finally, it allows for flexible measurements and experimental parameters – you can measure any observable, including critical “off-diagonal” observables that carry the signature of superconductivity, and simulate any system, such as those with laser pulses or electric fields. 

This last point is the most significant. While analog quantum simulators, like cold atom systems, can take snapshots of atom positions or measure densities, they struggle with off-diagonal observables—the very ones that signal the formation of Cooper pairs in superconductors.

Breaking new ground: a light-induced pairing

In our work, we've simulated three different regimes of the Fermi-Hubbard model and successfully measured non-zero superconducting pairing correlations — a first for any quantum computing platform.

We began by preparing a low-energy state of the model at half-filling — a standard benchmark for testing quantum simulations. Then, using simulated laser pulses or electric fields, we perturbed the system and observed how it responded.

After these perturbations, we measured a notable increase in the so-called “eta” pairing correlations, a mathematical signature of superconducting behavior. These results prove that our computers can help us understand light-induced superconductivity, such as the results from the Max Planck researchers. However, unlike those physical experiments, Helios offers a new level of control and insight. By tuning every aspect of the simulation — from pulse shape, to field strength, to lattice geometry — researchers can explore scenarios that are completely inaccessible to real materials or analog simulators.

Looking to a future where superconductors permeate our lives

Why does any of this matter? If we could predict which materials will become superconducting — and at what temperature, field, or current — it would transform how we search for new superconductors. Instead of trial-and-error in the lab, scientists could design and test new materials digitally first, saving huge amounts of time and money.

In the long run, Helios and its successors will become essential tools for materials science — not just confirming theories but generating new ones. And perhaps, one day, they’ll help us crack the code behind room-temperature superconductors.

Until then, the quantum revolution continues, one entangled pair at a time.

technical
All
corporate
All
Blog
October 30, 2025
Scalable Quantum Error Detection

Typically, Quantum Error Detection (QED) is viewed as a short-term solution—a non-scalable, stop-gap until full fault tolerance is achieved at scale.

That’s just changed, thanks to a serendipitous discovery made by our team. Now, QED can be used in a much wider context than previously thought. Our team made this discovery while studying the contact process, which describes things like how diseases spread or how water permeates porous materials. In particular, our team was studying the quantum contact process (QCP), a problem they had tackled before, which helps physicists understand things like phase transitions. In the process (pun intended), they came across what senior advanced physicist, Eli Chertkov, described as “a surprising result.”

While examining the problem, the team realized that they could convert detected errors due to noisy hardware into random resets, a key part of the QCP, thus avoiding the exponentially costly overhead of post-selection normally expected in QED.

To understand this better, the team developed a new protocol in which the encoded, or logical, quantum circuit adapts to the noise generated by the quantum computer. They quickly realized that this method could be used to explore other classes of random circuits similar to the ones they were already studying.

The team put it all together on System Model H2 to run a complex simulation, and were surprised to find that they were able to achieve near break-even results, where the logically encoded circuit performed as well as its physical analog, thanks to their clever application of QED.  Ultimately, this new protocol will allow QED codes to be used in a scalable way, saving considerable computational resources compared to full quantum error correction (QEC).

Researchers at the crossroads of quantum information, quantum simulation, and many-body physics will take interest in this protocol and use it as a springboard for inventing new use cases for QED.

Stay tuned for more, our team always has new tricks up their sleeves.

Learn mode about System Model H2 with this video:

technical
All