On May 29, 2025, researchers at Japan's National Institute of Information and Communications Technology presented data at OFC 2025 that reconfigures the understanding of optical transmission constraints: 1.02 petabits per second across 1,123 miles of optical fiber. Not 1.2 petabits—the earlier figure was preliminary. The confirmed measurement stands at 1.02 Pb/s, achieving a capacity×distance product of 1.86 exabits per second per kilometer. That metric matters more than the headline speed. It quantifies sustained information density across geographic scale—the engineering challenge that determines whether breakthrough physics becomes deployable infrastructure.
19 Cores, 180 Wavelengths: The Spatial Parallelism Strategy
The system architecture rests on spatial parallelism at the physical layer. Traditional single-core fiber hits quantum noise limits: encode too much information onto one light path, and nonlinear effects degrade the signal beyond recovery.
According to NICT's OFC 2025 presentation, the team embedded 19 independent optical channels into a 0.125mm multi-core fiber (0.005 inches)—standard-cladding diameter, thinner than a human hair. One cable becomes 19 simultaneous transmission highways.
Each channel operates using 16-state quadrature amplitude modulation (16QAM)—a scheme that encodes four bits per symbol—distributed across 180 distinct wavelengths in C-band and L-band spectrum. This modulation increases information density without proportional bandwidth expansion. The dual-band amplifiers—engineered to boost signals in both frequency ranges without cross-talk—represent materials science achievement as much as telecommunications engineering.
[VISUALIZATION: Multi-core fiber cross-section showing 19-channel hexagonal packing in 0.125mm diameter, with wavelength distribution across C-band and L-band spectrum]
The 1.86 Ebit/s·km figure contextualizes raw throughput. For distributed system architects designing cloud interconnects or real-time AI training pipelines across regions, this sustained density metric determines architectural possibility. The research team demonstrated that spatial multiplexing maintains coherence at continental distances when coupled with advanced signal processing.
40% Signal Loss Reduction: MIMO Processing Across 1,123 Miles
Maintaining coherence over 1,123 miles required solving propagation challenges that typically degrade optical signals. The research team reported in technical documentation that they implemented a multiple-input multiple-output (MIMO) digital signal processor treating the multi-core fiber as a coupled system, not isolated channels.
When light leaks between cores—crosstalk—the MIMO processor doesn't filter it out. It mathematically reconstructs the original signal by analyzing interference patterns across all 19 channels simultaneously.
[DIAGRAM: Signal reconstruction flow—crosstalk analysis across 19 channels to original signal recovery, showing mathematical matrix operations and error correction]
This approach achieved 40% reduction in signal loss compared to previous records. That translates directly into fewer regeneration stations needed for long-haul deployment.
Testing used 19 recirculation loops of 53.5-mile fiber—not 21 as early reports suggested—totaling 1,123.5 miles measured distance. Each loop introduces cumulative impairments: dispersion, attenuation, nonlinear phase shifts.
Measured data from 19 recirculation loops showed the system maintained bit error rates below forward error correction threshold. Transmitted data remained recoverable without retransmission.
The MIMO processing computational requirements present a scalability frontier. MIMO computational complexity scales non-linearly with channel count—roughly O(n³) for matrix inversion operations. The current system processes 19 channels; future iterations may expand to 37, 61, or more cores in hexagonal packing arrangements. Whether this processing happens in specialized silicon, FPGAs, or evolves into software-defined optics will determine upgrade paths and cost structures for network operators.
2.5× Capacity Leap: From 402 Tb/s to 1.02 Pb/s
Previous optical transmission records reached different scales through complementary approaches. The table below contextualizes the progression:
Year | Team | Capacity | Distance | Capacity×Distance |
|---|---|---|---|---|
Early 2024 | NICT | 301 Tb/s | Standard fiber | Not disclosed |
June 2024 | NICT/Aston/Nokia | 402 Tb/s | 31 miles | 20.1 Pbit·km/s |
Nov 2024 | NICT | 430.2 Tb/s | Hybrid mode | Not disclosed |
May 2025 | NICT | 1.02 Pb/s | 1,123 miles | 1.86 Ebit·km/s |
[CHART: Optical transmission capacity records 2024–2025, showing progression from 301 Tb/s to 1.02 Pb/s with capacity×distance products plotted on logarithmic scale]
The 1.02 Pb/s achievement represents roughly 2.5× capacity increase over the 402 Tb/s record, but the significance lies in demonstrating that multi-core fiber with advanced MIMO processing remains viable at extended ranges. The leap proves multiple stacked innovations—spatial multiplexing, dual-band amplification, 16QAM modulation, MIMO processing—compound rather than interfere.
Real-Time Synchronization: What 1.86 Ebit/s·km Enables
Current fiber networks face a trade-off: push more data through by accepting higher latency, or prioritize low latency by sacrificing throughput. NICT's multi-core architecture with MIMO processing begins addressing that constraint.
Consider distributed AI model training, where gradient updates must synchronize across geographic nodes. Today's networks impose waiting periods that slow convergence. A transmission system delivering 1+ petabits per second with reduced signal loss could enable near-synchronous parameter updates across continental distances, fundamentally changing how architects design distributed learning systems.
Real-time analytics pipelines processing streaming data from distributed sensors—financial markets, climate monitoring networks, autonomous vehicle fleets—currently batch data to compensate for I/O constraints. This capacity level makes continuous, unbatched processing architecturally feasible.
For cloud service providers, this capacity supports real-time replication of exabyte-scale data lakes across geographic regions—enabling true active-active architectures without trade-offs between consistency and availability.
Financial networks could process order flow and market data without microsecond delays that currently advantage co-located infrastructure. Scientific collaborations generating massive datasets—particle physics experiments, genomic sequencing facilities, radio telescope arrays—could share raw data in near-real-time rather than shipping physical storage devices.
The capacity matters for edge computing architectures as processing distributes toward data sources.
Amplifier Energy and Scalability Constraints
The research team identifies amplifier energy optimization as priority development. Dual-band amplifiers consume more power than single-band systems. At scale—across thousands of miles of deployed fiber—that energy cost compounds.
Data centers already consume approximately 1–2% of global electricity; transmission infrastructure adds to that footprint. Throughput per watt becomes as critical as throughput per fiber when evaluating deployment viability.
[GRAPH: Throughput-per-watt comparison showing single-core vs. dual-band multi-core amplification across 621 miles, 3,107 miles, and 6,214 miles distances]
Current estimates suggest dual-band amplifier systems consume approximately 15–20% more power per amplification stage compared to single-band equivalents. For a transoceanic cable requiring amplification every 31–50 miles, this compounds across 60–120 amplifier stations.
Energy efficiency improvements could emerge from semiconductor advances in amplifier design or optimization of amplification intervals based on fiber attenuation characteristics.
MIMO processing scalability presents another frontier. Current systems require specialized ASICs or high-performance FPGAs for real-time matrix operations across 19 channels. Expanding to 37 or 61 cores would increase computational requirements by factors of 6–12×, assuming O(n³) scaling holds. Software-defined approaches offer flexibility but may introduce latency that defeats the purpose of high-throughput transmission. Future research into AI-optimized algorithms for signal processing could reduce computational overhead while maintaining signal quality.
Physical Layer Constraints and Material Science
Achieving 19 channels in 0.125mm (0.005-inch) fiber required advances in glass manufacturing precision and core placement tolerances. Each core must maintain sufficient separation to minimize crosstalk while maximizing packing density—a geometric optimization problem with material constraints.
Refractive index profiles, cladding composition, and coating materials influence propagation characteristics and mechanical durability. Deployment constraints include:
- Minimum bend radius: 30–40mm (1.2–1.6 inches) for multi-core fiber vs. 15–20mm (0.6–0.8 inches) for standard single-mode fiber
- Operating temperature range: -40°C to +70°C (-40°F to +158°F) with thermal expansion coefficients matched across core materials
- Splice loss tolerances: <0.3 dB per splice with core alignment precision within 1 micron (0.00004 inches)
- Maximum tensile load: 60–80 N during installation to prevent core deformation
Bend radius limitations become more stringent with multi-core fiber. Tighter bends induce mode coupling and macro-bending losses affecting outer cores disproportionately. Installation practices—routing through ducts, around corners, securing in splice enclosures—must account for these constraints.
Field deployment will test whether lab-demonstrated performance translates under mechanical stresses of real-world infrastructure.
Parallel innovations in fiber technology offer complementary paths. In September 2025, Microsoft-backed research published in Nature Photonics demonstrated hollow-core fiber achieving 0.091 dB/km loss at 1,550 nm with approximately 45% faster photon transit. That advancement addresses signal loss through different physics—air-core guidance versus solid-core propagation—representing an alternative approach to the efficiency challenge.
Methodology Context and Deployment Timeline
The recirculation loop approach serves dual purposes: validating signal integrity under realistic impairment conditions while enabling controlled experimentation with channel configurations and modulation schemes.
All high-capacity demonstrations operate as testbed experiments using wideband transmitters, many wavelength channels, and recirculating loops or short reels. This methodology standard across optical research allows precise measurement of cumulative degradation.
Laboratory demonstration confirms viability. Commercial deployment timeline depends on manufacturing scalability and standardization. For transoceanic cables or metro network upgrades—where physical conduit space is constrained—the cost-per-bit and space-per-bit efficiency could justify adoption within a decade. For greenfield deployments with abundant conduit, the calculus differs.
Manufacturing scalability requires automated fusion splicing equipment capable of aligning 19 cores simultaneously, quality control systems that verify refractive index uniformity across all cores during fiber draw, and spooling systems that prevent differential tension across the fiber cross-section.
Standards development through ITU-T and IEC will define interface specifications, testing procedures, and interoperability requirements before widespread adoption occurs.
The 1.02 petabits per second figure represents a proof point rather than a ceiling. Theoretical limits of what glass and light can do together remain distant. What remains is translation from laboratory demonstration to deployed infrastructure—manufacturing scale, cost optimization, standards development, and the patient work of making breakthrough engineering reliable enough for global networks.






















