1. Introduction
In this manuscript,
Section 1 contextualizes the problem of quantum error correction in real devices [
1] and defines the logical Bell state as the case study;
Section 2 reviews prior feedback-based and topological-code approaches for Bell-state stabilization on superconducting hardware;
Section 3 details the design of surface code patches, the selection of physical qubits, and the experimental protocol implemented on the
ibm_fez,
ibm_torino, and
ibm_brisbane devices;
Section 4 presents fidelity and accuracy data with and without error correction, showing improvements of 10–20%;
Section 5 analyzes how the benefit of QEC varies with hardware generation and extracts lessons for higher-distance codes;
Section 6 summarizes the key findings and suggests directions for future work in the NISQ era [
2,
3].
Quantum error correction (QEC) [
4] is universally acknowledged as an indispensable ingredient for scalable quantum computation since it enables the preservation of delicate superposition and entanglement against decoherence and imperfect gate operations [
5]. A canonical benchmark for QEC schemes is the logical Bell state,
where
spans the logical qubit Hilbert subspace. In the simple repetition code of length
n, one encodes
thereby detecting up to
bit-flip errors at the expense of an
n-fold physical-qubit overhead [
6]. However, repetition codes cannot address simultaneous bit- and phase-flip errors.
Surface codes transcend this limitation by arranging qubits on a two-dimensional lattice and introducing two commuting stabilizer families [
7,
8]:
where
and
denote Pauli operators on physical qubit
i. The protected code subspace is the common
eigenspace of all
and
. Logical operators arise as non-trivial homological cycles [
9],
with intersection
(mod 2), ensuring
. The minimal weight of such cycles defines the code distance
d [
10].
Surface codes thus offer simultaneous protection against bit- and phase-flip errors with asymptotically favorable scaling, making them prime candidates for near-term fault-tolerant architectures.
For the design of surface-code patches, the qubit coupling maps of
ibm_fez (graph
),
ibm_torino (
) and
ibm_brisbane (
) are modeled as undirected connected graphs
where
is the number of physical qubits and
the number of couplings. These topologies remain static during our February 2025 experiments, allowing precomputation of their binary cycle-space dimension
which counts all independent simple cycles [
11]. Although every simple cycle lies in this space, only chordless 4-cycles (plaquettes) aligned with the heavy-hex layout serve as valid stabilizer patches [
12]. Geometrically, each such plaquette is a rhombus (a square rotated by
) because the hexagonal lattice does not admit axis-aligned squares.
Figure 1a highlights one rhombic plaquette on the 127-qubit
ibm_brisbane device, with its vertex set
traced in black.
Each plaquette cycle
defines a Pauli–
X stabilizer
and its dual plaquette produces a Pauli–
Z stabilizer. The full set of these 4-cycle equivalence classes generates the first homology group
which captures both the count and orientation of independent plaquette windings [
13]. In this homological framework, each class
corresponds to one logical operator: its integer winding number identifies a distinct logical-qubit degree of freedom.
Mathematically, each plaquette
P in a distance-
d surface code defines a four-qubit stabilizer operator
where
are the vertices of the 4-cycle
and
is the Pauli-
X acting on qubit
i. The dual plaquettes similarly produce
Z-stabilizers
. Collectively, these operators generate the Abelian stabilizer group [
14]
whose common
eigenspace is the logical code subspace
. A logical Bell state
satisfies
Error syndromes are obtained by projective measurement of each stabilizer. A Pauli error
E anticommutes with those
for which
yielding an outcome
and thereby flagging a defect. The vector of measurement outcomes
is decoded (e.g., via minimum-weight perfect matching) [
15] to infer a correction operator
such that
restoring the state to
without disturbing
. This procedure guarantees that all residual errors lie in the stabilizer group and thus act trivially on the encoded logical information [
16,
17].
In
Figure 1a, the single plaquette cycle
, based at
, realises the logical operators
whose common
eigenstate is the encoded qubit
. In
Figure 1b, two disjoint 4-cycles
and
(anchored at
and
, respectively) define two logical qubits with operators
and
. By preparing their joint state in the simultaneous
eigenspace of
and
, the surface code directly encodes the two-qubit Bell state
thus demonstrating entanglement at the logical level. Since
, these logical operators inherit the homotopy-class structure of the underlying graph, guaranteeing topological protection against local errors [
18,
19].
2. Related Work
Quantum Bell-state stabilization on superconducting hardware [
20] has developed along two complementary tracks: feedback-based control and topological error-correcting codes. The first clear demonstration of real-time feedback employed an ancillary transmon to perform repeated parity measurements, achieving fidelities of
over twelve stabilization cycles and proving that stabilizer protocols can be executed within qubit coherence times [
21]. Subsequent work integrated table-lookup decoders directly into cryogenic control stacks, enabling fully autonomous state locking.
A second line of research exploits the passive robustness of topological modes. Jin and Greplova [
22] showed that edge states of a superconducting resonator chain can stabilize entanglement against local parameter noise, providing a hardware-efficient route to long-lived two-qubit correlations. Most recently, Hetényi and Wootton overlaid a surface code with a Bacon–Shor code on a 133-qubit heavy-hex device [
21], producing logical Bell pairs with code distance
and a post-selected fidelity of 94% after five rounds of syndrome extraction in simulation [
19].
Previous studies either remained confined to few-qubit demonstrations or focused on a single processor architecture; reported gains are typically quoted at the logical level, obscuring the physical-qubit overhead.
A
cross-platform implementation of surface code Bell-state stabilization is reported across three IBM Quantum processors—
ibm_fez,
ibm_torino, and
ibm_brisbane—featuring 127 and 156 physical qubits. In contrast to the overlay approach of [
23], every qubit participates in a single distance-
d surface code whose syndrome extraction runs within coherence-time budgets. Averaged over five error-correction rounds, a physical-state fidelity
was achieved, and net accuracy improvements of 10% on 127-qubit lattices and 20% on 156-qubit lattices relative to uncorrected runs were measured. The protocol relies only on nearest-neighbor CZ connectivity, is fully open-sourced, and can be ported to other superconducting platforms, thereby narrowing the gap between numerical benchmarks and hardware practice.
3. Materials and Methods
Two entangled logical qubits are implemented using a distance-
topological code [
4] on three IBM quantum processors—
ibm_fez (156 qubits),
ibm_torino (133 qubits) and
ibm_brisbane (127 qubits).
This implementation requires carefully selecting appropriate physical qubits to form the surface code [
24] patches where quantum information will be encoded. Due to the quantum computer’s topology, the best-case scenario allows for the creation of a maximum of three entangled logical qubit pairs. This limitation arises from the graph’s structure: 15 physical qubits are required in a specific configuration to generate a single logical entangled resource pair (ERP) [
25]. This exemplifies one of the key disadvantages of working with logical qubits. Consequently, a 156-physical-qubit processor may effectively be reduced to just 6 logical qubits, highlighting the current limitations of the NISQ era for executing large-scale, deep quantum circuits [
26].
To create the first logical qubit, four physical qubits in a diamond arrangement are entangled. The entanglement was achieved by applying Hadamard gates (
H) followed by Controlled-NOT (
) gates between each adjacent pair of qubits in a clockwise manner, establishing cyclic entanglement [
27]. The same procedure was repeated for the second logical qubit, starting from the bottom qubit. Subsequently, one qubit from the first patch was entangled with another from the second patch to establish logical-qubit entanglement.
This procedure was executed on actual quantum hardware, requiring a transpilation process to adapt the quantum operations to each computer’s native gate set. The transpilation automatically identifies equivalent gate sequences, a process facilitated by Qiskit’s qiskit.transpiler module and its transpile() function.
The detailed methodological procedure can be seen in the following algorithm (Algorithm 1):
Algorithm 1 Quantum Circuit for Entangling Logical Qubits Process |
- 1:
Input: num_qubits, num_logical_qubits, ancillas_config - 2:
Initialize qubits in state. - 3:
Initialize quantum circuit qc with num_qubits and classical bits. - 4:
Define logical qubits: qubitL_i for to num_logical_qubits - 5:
Define ancillas for parity measurement using ancillas_config - 6:
Prepare logical qubits: - 7:
for each logical qubit set qubitL_i do - 8:
for each qubit q in qubitL_i do - 9:
Apply Hadamard gate - 10:
end for - 11:
end for - 12:
Entangle qubits within each logical group: - 13:
for each logical qubit set qubitL_i do - 14:
for to do - 15:
Apply CNOT: - 16:
end for - 17:
end for - 18:
Apply cross-connections and entanglement between logical qubits. - 19:
procedure measure_parity_Z(logical, ancillas, bit_start) - 20:
- 21:
for each (qubit, (a1, a2)) in ancillas do - 22:
Apply CNOT: and - 23:
Measure - 24:
Measure - 25:
- 26:
end for - 27:
end procedure - 28:
procedure correct_errors_X(logical, ancillas, bit_start) - 29:
- 30:
for each (qubit, (a1, a2)) in ancillas do - 31:
Apply conditional X gate if or - 32:
- 33:
end for - 34:
end procedure - 35:
Apply parity measurement and error correction for all logical qubits. - 36:
Final measurements: - 37:
for each logical qubit set qubitL_i do - 38:
for to do - 39:
Measure - 40:
end for - 41:
end for - 42:
Print the final circuit diagram.
|
As shown in
Figure 2, the pairs of entangled logical qubits can be clearly identified. While their exact positions vary slightly across different quantum processors, their relative locations remain similar due to the comparable topologies of the IBM quantum computers. Each diamond-shaped pair represents an entangled qubit pair (ERP) encoded on the surface code lattice.
This structure enables the detection and correction of bit-flip errors through the stabilizer operators
, where
represents a Pauli-
X operation on the
i-th qubit. Bit-flip errors constitute one of the predominant noise sources in current NISQ-era quantum computers [
29].
4. Testing and Results
The results are obtained by executing the quantum circuit using 15 physical qubits, with a total of 10,000 circuit repetitions (shots). Measurements are performed after applying quantum error correction to each physical qubit, and the outcomes are stored as classical bit strings.
During the post-processing stage, intermediate states that introduced measurement noise are filtered out. The cleaned measurement data was subsequently analyzed to identify instances in which the encoded logical qubit pairs corresponded to either the or states.
The analysis is expected to produce a histogram with two prominent peaks—one at and the other at . In an ideal Bell state, these outcomes each appear with 50% probability. This characteristic bimodal distribution provides a clear indicator of successful logical encoding and error-correction procedures.
Figure 3a–c present a comparative analysis of quantum state preparation accuracy with (blue) and without (red) Quantum Error Correction (QEC). The accuracy metric was computed by evaluating the deviation from the ideal 50%–50% distribution between the logical states
and
using the symmetric ratio:
where
n and
m represent the measured counts of
and
states respectively. This accuracy measure reaches unity when perfect balance (
) is achieved and decreases as the distribution becomes more skewed.
However, what is particularly interesting is identifying where the use of QEC becomes truly necessary. In the case shown in
Figure 3c, the difference is minimal compared to the other two cases, which shows that quantum error correction has a limited impact in this specific scenario where the number of qubits is lower.
While this metric effectively quantifies the proximity to the target logical EPR state, experimental implementations often encounter superposition states that may affect the results [
30]. For instance, the
Bell state represents a scenario where the logical qubits are in opposite states rather than the desired correlated configuration. This necessitates the use of quantum state fidelity, which provides a rigorous comparison between the desired and obtained density matrices.
For this case, according to [
31,
32], the orthogonality between
and
the fidelity can be written as the probability (
) of obtaining the target state
:
where
is the target density matrix
is the experimentally obtained density matrix
represents the probability of measuring the desired state
Figure 4 presents the summarized accuracy results across quantum computers. The most significant improvement was observed in
ibm_fez (156 physical qubits), which at the time of writing (April 2025) represents IBM’s largest superconducting quantum processor. This indicates that newer quantum hardware may require more sophisticated qubit stabilization techniques compared to more mature systems like
ibm_brisbane.
The results obtained in
Table 1 highlight that error correction methods remain valuable for improving computational accuracy, particularly in newer quantum processors. However, it is emphasized that these experiments only provide evidence for the preservation of a well-understood quantum state (the logical EPR state), and further research is required to generalize these findings to more complex quantum states and operations.
5. Discussion
The experimental results presented in this work reveal several important insights about quantum error correction [
30] performance across different generations of IBM quantum processors. The Bell-state fidelity measurements demonstrate that, while QEC provides measurable improvements in all tested systems, the magnitude of these improvements depends strongly on device size and maturity.
To contextualize these results,
Table 2 juxtaposes our hardware-based surface-code stabilization against the simulation study by Hetényi & Wootton. Their work achieves a post-selected logical Bell-state fidelity of
for distance-2 codes in silico with five measurement rounds but reports no pre- or post-correction accuracy baseline. In contrast, this study implements a distance-4 surface code on three IBM heavy-hex devices, where the accuracy pre/post-QEC of
on 127 qubits,
on 133 qubits, and
on 156 qubits. Bell-state fidelity after correction
in all three cases.
These improvements under NISQ-era noise confirm consistent accuracy gains, while the fidelity metric highlights the remaining challenge: a contribution factor of
still lies below the ideal threshold region (
) for scalable surface-code architectures [
14].
The most substantial QEC benefits are observed in
ibm_fez with 156 qubits, where error correction improved state-preparation accuracy by
This significant enhancement suggests that larger, newer quantum processors may initially require error-correction strategies to compensate for nascent qubit-stabilization techniques. The post-correction fidelity of 0.89 approaches the threshold commonly cited for fault-tolerant surface code implementations in superconducting hardware.
Conversely, the
ibm_brisbane system with 127 qubits showed smaller but still meaningful improvements,
reflecting a higher baseline accuracy that likely arises from accumulated engineering improvements in qubit design and control electronics.
The intermediate case of
ibm_torino with 133 qubits presents a transitional performance profile, with QEC providing
This suggests that
ibm_torino benefits from partial integration of advanced error-mitigation techniques, sitting between the earlier and later generations of devices.
Despite these advances, this contribution remains
across all three processors. For comparison, the recent on-hardware magic-state injection study reports fidelities of
and
[
29]. Simulation-based work in a heavy-hex lattice reports similar contribution factors but lacks real-device validation [
21].
A contribution factor of
still falls below the ideal threshold region (≈
–
) [
28] required for scalable surface code architectures. This gap is attributed to the following:
Residual correlated errors not fully captured by the basic surface code syndrome extraction.
Circuit depth overhead, where additional syndrome-measurement layers introduce new opportunities for decoherence and gate infidelity.
Measurement and reset errors in ancilla qubits, which degrade the effectiveness of repeated syndrome-extraction cycles.
These results demonstrate that, although on-hardware QEC and magic-state protocols achieve thresholds and fidelities beyond prior benchmarks, further improvements in device coherence, decoding algorithms, and error-suppression techniques are essential to elevate the contribution factor into the regime necessary for fully fault-tolerant quantum computation.
To further elucidate the hardware coherence limitations on
ibm_brisbane, this analysis examines qubit relaxation (
) and dephasing (
) times as reported in IBM’s backend specifications [
33]:
(Relaxation Time): The average duration a qubit remains in the excited state before decaying to the ground state , characterizing energy relaxation (units: s).
(Dephasing Time): The timescale over which a qubit maintains phase coherence between and , accounting for both relaxation and pure dephasing processes (units: s).
Measured values are significantly lower on ibm_brisbane (s, s) compared to ibm_torino (s, s) and ibm_fez (s, s). These reduced coherence times directly correlate with the smaller error-correction gains observed on ibm_brisbane, indicating that its engineering optimizations—while enhancing gate performance—also produce shorter qubit lifetimes that constrain overall QEC effectiveness.