Why the RAN Baseband Stack Needs Neural Inference
The O-RAN 7.2x pipeline hasn't fundamentally changed since 3G. It's time to rethink what happens inside the radio unit.
Every modern cell site runs a baseband processing pipeline that was architected decades ago. Signals come in from antennas, get digitized, and then pass through a rigid sequence of DSP blocks: channel estimation, MIMO equalization, and demapping. Each block operates independently, each requires per-scenario tuning, and all of it traditionally runs in the distributed unit (DU) — connected to the radio unit by expensive, high-bandwidth fiber.
The O-RAN Alliance's 7.2x functional split was supposed to modernize this. And it did open up the RAN to multi-vendor interoperability. But it left the fundamental processing architecture untouched. The radio unit remained a "dumb" RF frontend, and all the intelligence stayed centralized in the DU.
This creates three compounding problems that get worse as networks scale.
The Fronthaul Bottleneck
In the 7.2x split, the radio unit sends raw I/Q samples — per antenna, per subcarrier — over dedicated eCPRI fiber to the DU. For a 64-antenna massive MIMO site, that's 25–50 Gbps of fronthaul bandwidth. The cost scales linearly with antennas, making densification and massive MIMO rollouts increasingly expensive.
This isn't a theoretical concern. Operators consistently cite fronthaul cost as one of the top barriers to 5G densification. Dedicated fiber to every radio unit simply doesn't scale to the deployment densities that 5G and 6G demand.
The Power Problem
Traditional baseband DSP is computationally expensive. LMMSE channel estimation requires matrix inversions on every subframe. K-best MIMO detection involves tree searches that grow combinatorially with antenna count. These operations run iteratively and consume 500–700 mW of baseband power per radio unit — before you even count the RF chain.
Multiply that across thousands of small cells in a dense urban deployment and power consumption becomes a major operational cost. Operators are under increasing pressure — both economic and regulatory — to reduce the energy footprint of their networks.
The Rigidity Problem
The three-block DSP pipeline (channel estimation → equalization → demapping) processes signals sequentially, with each block optimized in isolation. When channel conditions change — new interference patterns, different propagation environments, varying user densities — each block needs to be reconfigured independently. There's no joint optimization across the pipeline, and adapting to new scenarios requires manual engineering effort.
This rigidity is fundamentally at odds with the vision of self-optimizing, AI-native networks that the industry is moving toward.
The Neural Receiver Approach
What if you replaced all three DSP blocks with a single neural network that performs channel estimation, equalization, and demapping jointly in one forward pass?
That's exactly what a neural receiver does. A CNN + GNN architecture with 730K parameters (quantized to INT8) takes in the resource grid and outputs soft bits (LLRs) directly — skipping the entire traditional DSP chain.
The implications are significant. Instead of sending raw I/Q samples over fronthaul, the radio unit sends only LLRs — reducing fronthaul from 25–50 Gbps down to roughly 2 Gbps. That's a 10–20× reduction, which means you can potentially use existing Ethernet infrastructure instead of dedicated fiber.
Power consumption drops to approximately 150 mW for the baseband processing — a 3–5× reduction — because a single forward pass through a small neural network is far cheaper than iterative matrix operations.
And because the network learns jointly across the entire pipeline, it can adapt to new channel conditions without manual reconfiguration. Retrain or fine-tune on new data, and the whole pipeline adapts together.
Why Now?
Three things have converged to make this practical today.
The industry has validated AI-RAN. NVIDIA invested $1 billion in Nokia specifically for AI-RAN products. SoftBank demonstrated GPU-accelerated AI-RAN with 16-layer massive MU-MIMO. The question isn't whether AI belongs in the RAN — it's where exactly it runs and in what form.
Neural network quantization has matured. INT8 inference on small models (<1M parameters) is now well-understood and deployable on edge hardware. A 730K-parameter model quantized to INT8 fits comfortably on modern FPGAs and can be compiled into fixed silicon logic.
The deployment economics demand it. The 5G infrastructure market is projected to reach $676 billion by 2034. Private 5G alone is growing at 65% CAGR. At these scales, even modest per-site savings in fronthaul and power compound into billions in reduced deployment cost.
The Gap in the Market
Most AI-RAN efforts today operate at the DU or cloud level — using GPUs to run AI workloads alongside traditional baseband processing. That's valuable, but it doesn't solve the fronthaul problem. It doesn't reduce radio unit power. It doesn't fundamentally change what happens at the antenna.
The opportunity is to build purpose-built neural inference hardware that runs inside the radio unit itself. Validate on FPGA first, then compile the forward pass into fixed inference silicon — no software overhead, sub-1W total power, deployable as an ASIC-ready IP core.
That's what we're building at neuraRAN.
We're Building This
neuraRAN is developing the first purpose-built neural receiver for the radio unit. We're pre-seed, building our FPGA prototype, and looking for investors and partners who want to shape the future of RAN.
Get in Touch →