Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
Promotions
AI
Gate AI
Your all-in-one conversational AI partner
Gate AI Bot
Use Gate AI directly in your social App
GateClaw
Gate Blue Lobster, ready to go
Gate for AI Agent
Gate MCP
Gate Skills Hub
10K+ Skills
From office tasks to trading, the all-in-one skill hub makes AI even more useful.
GateRouter
Smartly choose from 30+ AI models, with 0% extra fees
Latest Breakthroughs in Quantum Computing 2024: What Actually Changed and Why It Matters
There’s a version of the quantum computing story that plays on loop every year: a breathless press release, a number so large it breaks comprehension, and then silence. 2024 felt different to researchers who actually follow this field closely. Not because of one announcement, but because of three separate breakthrough moments that happened within months of each other — each from a different company using a fundamentally different approach to the same problem. When that happens simultaneously across hardware architectures, it’s usually a sign the field is moving rather than spinning.
Here’s what actually changed in 2024, why each development matters, and what the honest caveats are.
Google Willow: The Chip That Changed the Error Correction Conversation
The biggest news of the year landed on December 9, 2024. Google’s quantum AI team unveiled Willow — a 105-qubit superconducting processor built at their dedicated fabrication facility at UC Santa Barbara — and what it demonstrated was not just a faster chip. It was proof of something the field had been trying to establish for nearly three decades.
The core achievement: as Google added more qubits to Willow, the error rate went down instead of up. That sounds simple. It isn’t. For years, the central frustration of quantum computing was that more qubits meant more noise, more instability, more errors cascading through calculations. You could build a bigger system, but it would be less reliable. Willow broke that relationship. Using its error correction architecture, the chip demonstrated what’s called “below-threshold” operation — the point at which scaling actually helps rather than hurts.
The benchmark Google ran alongside this announcement became instantly famous: Willow completed a random circuit sampling computation in under five minutes that would take today’s fastest classical supercomputer 10 septillion years — that’s 10²⁵ years, roughly a million times the current age of the universe. As Hartmut Neven, who founded Google Quantum AI in 2012, put it: “We are past the break-even point.” The full technical details were published in the peer-reviewed journal Nature, which matters: previous quantum supremacy claims have attracted legitimate criticism, and having the methodology available for scrutiny is a meaningful difference.
The official announcement and technical documentation is available directly at Google’s Quantum AI blog.
The honest caveat: Willow’s benchmark test is still narrow. Random circuit sampling proves that certain computations are classically intractable for this chip — it does not mean Willow can currently run the drug discovery or climate modeling applications that get mentioned whenever quantum computing comes up. The value of Willow is architectural: it shows that large-scale error-corrected quantum computing isn’t a theoretical ceiling anymore. It’s a demonstrated engineering path.
Microsoft and Quantinuum: The Logical Qubit Milestone
Eight months before the Willow announcement, Microsoft and Quantinuum published a result in April 2024 that got less general press but arguably more attention from researchers. They demonstrated logical qubits with error rates 800 times lower than the corresponding physical qubits they were built from — using what Microsoft called “qubit virtualization”.
The distinction between physical and logical qubits is the real dividing line in quantum computing. Physical qubits are the hardware units — they’re noisy, sensitive to temperature, vibration, electromagnetic interference, and time itself. Logical qubits are built by combining multiple physical qubits into a structure that encodes information redundantly, so errors can be detected and corrected without destroying the computation. The challenge has always been that logical qubits require so many physical qubits to build that the overhead made the whole thing impractical. An 800x error rate reduction means that logical qubits are starting to look realistic rather than theoretical.
Microsoft extended this further in November 2024. Working with Atom Computing, they successfully created and entangled 24 logical qubits using ultracold neutral ytterbium atoms — setting a new record and doing it with remarkable gate fidelities: 99.963% for single-qubit operations and 99.56% for two-qubit entangling gates. The neutral atom approach uses laser-cooled atoms held in place by optical tweezers, a completely different hardware architecture than Google’s superconducting transmons. This matters because it means multiple viable paths toward fault-tolerant quantum computing are progressing simultaneously, rather than the field betting everything on one approach.
Then in December 2024, Quantinuum went further still: entangling 50 logical qubits — another record, and a demonstration that the logical qubit era is not a future milestone but an active present.
IBM Heron R2: The Engineering Discipline Breakthrough
Google’s Willow and Microsoft’s logical qubits grabbed more headlines in 2024. IBM’s contribution was quieter but equally significant for anyone thinking about where practical quantum computing actually comes from.
In November 2024, IBM unveiled the Heron R2 processor — 156 qubits, the second generation of the Heron architecture, built with a heavy-hexagonal lattice topology. The headline qubit count matters less than what happened to performance. IBM’s 2Q gate error rates dropped to 8×10⁻⁴. The system can now execute quantum circuits with up to 5,000 two-qubit gate operations. And workloads that previously took more than 120 hours to complete on IBM’s best quantum hardware were running in approximately 2.4 hours — roughly a 50x speedup.
Earlier in 2024, IBM also completed its self-imposed “100×100 Challenge,” running a 100-qubit circuit at depth 100 on the Heron processor within hours. This is a “utility-scale” computation — one that cannot be brute-forced by classical means — and completing it represents the kind of measured, incremental proof-of-progress that IBM has built its reputation on.
The more technically significant 2024 IBM result came in a Nature paper describing a new error correction code called the “bivariate bicycle” qLDPC code. Conventional quantum error correction using surface codes requires roughly 3,000 physical qubits to encode a single reliable logical qubit. IBM’s new code achieves comparable error suppression using only 144 data qubits plus 144 ancilla qubits for error checks — a 10x reduction in overhead. That kind of efficiency gain is the kind of thing that makes fault-tolerant quantum computing look less like a distant goal and more like an engineering problem with a defined solution path.
IBM’s full hardware roadmap and current processor specifications are documented at ibm.com/quantum.
NIST and Post-Quantum Cryptography: The 2024 Breakthrough Nobody Talks About
The fourth major development of 2024 didn’t involve a quantum processor at all. In August 2024, the US National Institute of Standards and Technology (NIST) formally published the first post-quantum cryptography standards — algorithms designed to resist attacks from future quantum computers. Two of the three algorithms (ML-KEM and ML-DSA) were developed by IBM Research cryptographers in Zurich.
Why does this belong in a quantum computing breakthroughs article? Because it’s the first concrete acknowledgment by a global standards body that quantum computers capable of breaking current encryption are no longer purely theoretical. The standards exist because governments and enterprises need to start transitioning now, before cryptographically relevant quantum computers arrive. The transition timeline from standard publication to widespread deployment is typically a decade or more. NIST’s 2024 decision effectively started that clock.
For blockchain and digital asset infrastructure specifically, this is directly relevant. Current asymmetric encryption schemes protecting wallets, transactions, and smart contracts will eventually need to be replaced with quantum-resistant alternatives. BlockchainReporter’s coverage of blockchain and cryptography developments tracks this transition as it unfolds across the industry.
For a detailed breakdown of how quantum advances affect cryptocurrency security specifically, see BlockchainReporter’s analysis of quantum computing’s impact on cryptocurrency.
The Honest Assessment: What 2024 Did and Didn’t Prove
It would be easy to read the above and conclude that quantum computing has “arrived.” That framing isn’t quite right, and the researchers involved have been explicit about this.
Google’s Willow is not yet running the applications its long-term roadmap promises — drug discovery, materials science, financial optimization. It demonstrated below-threshold error correction and a benchmark result. The gap between that and a commercially useful computation is still substantial, requiring error rates significantly lower than current levels.
For context on how the crypto community is actually responding to these developments, BlockchainReporter’s coverage of expert views on quantum threats to Bitcoin provides useful perspective on the gap between theoretical risk and current reality.
Quantinuum’s 50 logical qubits can detect errors, but full error correction (detecting and fixing them without destroying the quantum state) is a harder problem that’s still being worked through. Microsoft’s Atom Computing record used neutral atoms that require extremely sophisticated laser control infrastructure that doesn’t yet exist at scale.
IBM’s Heron R2 is the most practically deployed of the 2024 systems — it’s in IBM’s quantum cloud, enterprise clients are running workloads on it, and the 100×100 benchmark demonstrates utility-scale results. But the Starling processor, IBM’s first fully error-corrected system, isn’t projected until 2029.
What 2024 did prove is more important than what it didn’t. The field stopped progressing in one direction and started progressing in all directions simultaneously — hardware, error correction, logical qubits, software efficiency, and cryptographic standards. As a research community, it started acting less like a theoretical physics discipline and more like an engineering field with milestones that can be independently checked and reproduced.
For BlockchainReporter readers tracking the convergence of quantum computing and AI that’s reshaping financial infrastructure, the latest developments in blockchain and emerging technology section covers how these shifts affect decentralized systems and digital asset security in real time.
What Comes Next: The 2025–2026 Trajectory
The 2024 breakthroughs set up a specific set of next steps that the field is now actively working through.
Google’s next milestone after Willow is achieving fault-tolerant operation — moving from below-threshold error correction to full error correction where the system can run arbitrarily long computations reliably. The Quantum Echoes algorithm, published on the Willow processor in 2025, demonstrated the first-ever verifiable quantum advantage for a real computational problem, marking a step beyond benchmark demonstrations toward application-relevant results.
Microsoft’s roadmap targets 50–100 entangled logical qubits in commercial deployments within the next few years — enough, by their own estimate, for “truly practical breakthroughs in materials science or chemistry.” Their Majorana 1 chip, introduced in 2025 and built on exotic topological qubits, represents a third architectural bet alongside superconducting and neutral atom approaches.
IBM’s Starling processor, due in 2029, aims for 100 million gates across 200 error-corrected qubits using the Gross code error correction scheme — the architecture that IBM believes will finally bridge from quantum utility to quantum advantage for commercially valuable problems.
The trajectory from 2024 points in one consistent direction: the question is no longer whether large-scale error-corrected quantum computing is possible. The 2024 breakthroughs established it’s possible across multiple hardware approaches. The question now is which approach scales fastest, and how quickly the applications that justify the investment come into focus.
This article is for informational and educational purposes only. It does not constitute financial or investment advice.