Difference between revisions of "5G OAI Neural Receiver Testbed with USRP X410"
NeelPandeya (Talk | contribs) |
NeelPandeya (Talk | contribs) |
||
| Line 437: | Line 437: | ||
This flexibility allows the same hardware platform to serve as both gNB and UE, simplifying testbed deployment and ensuring scalability for future AI-driven enhancements. | This flexibility allows the same hardware platform to serve as both gNB and UE, simplifying testbed deployment and ensuring scalability for future AI-driven enhancements. | ||
| + | |||
| + | Note that as the sample rate increases, for example, to 122.88 Msps for a 100 MHz channel bandwidth, and also as the number number of MIMO layers increases, for example, to 2x2 and 4x4, then the computational demands increase significantly, and more and more physical cores would be needed to support this load. | ||
[[File:server_dell_5860_tower.jpg|thumb|800px|center|Dell Precision 5860 Tower Workstation]] | [[File:server_dell_5860_tower.jpg|thumb|800px|center|Dell Precision 5860 Tower Workstation]] | ||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| − | |||
| + | ==Consolidated Bill of Materials== | ||
| + | Listed below is a consolidated Bill of Materials (BoM) for all the components used in the implementation of this system. Note that the specific computers listed are not necessarily or specifically required, and that computers from other vendors may be used as long as they have a similar level of performance and capability. | ||
| + | * Core Network (CN) host system: | ||
| + | ** Dell 3680 Precision Tower, with Intel i9-10980XE CPU, with 18 physical cores, with 3.00 GHz base clock frequency. | ||
| + | ** Mellanox Nvidia ConnectX-5 MCX512A-ACAT network card, with two SFP28 ports. | ||
| + | * gNB host system: | ||
| + | ** Dell 5860 Precision Tower, with Intel Xeon W7-2495X CPU, with 24 physical cores, and with turbo frequency of 4.8 GHz. | ||
| + | ** Mellanox Nvidia ConnectX-5 MCX512A-ACAT network card, with two SFP28 ports. | ||
| + | ** Nvidia GeForce RTX 4090 GPU. | ||
| + | ** GPU power cable, 30cm, 12+4-pin male to 2x8-pin female sleeved 12v extension cord for GeForce RTX 4090. | ||
| + | ** QSFP28-to-SFP28 25 Gbps Ethernet breakout cable to connect with USRP X410. | ||
| + | *** Nvidia MCP2M00-A002E30N 100 GbE to 4x25GbE (QSFP28 to 4xSFP28) Direct Attach Copper (DAC) splitter cable ([https://network.nvidia.com/pdf/prod_cables/PB_MCP7F00-A0xxRyyz_100GbE_QSFP28_to_4x25GbE_4xSFP28_DAC_Splitter.pdf here]). | ||
| + | *** NI 100 GbE to 4x25GbE (QSFP28 to 4xSFP28) splitter cable (NI Part Number 788214-01). | ||
| + | * USRP X410 radio. | ||
| + | ** One connected to the gNB system. | ||
| + | ** One connected to the UE system. | ||
| + | * OctoClock-G CDA-2990 for synchronization between the gNB USRP X410 and the UE USRP X410. | ||
| + | * RF cables (1 meter) (Quantity 8). | ||
| + | * 20 dB RF attenuators (Quantity 2). | ||
| + | * Two-port RF splitter, 6 GHz, SMA connectors (Quantity 1). | ||
| + | * Ethernet switch, 5 ports (Quantity 1). | ||
| + | * Cat-6A RJ-45 Ethernet cables (Quantity 4). | ||
Revision as of 03:53, 31 October 2025
Contents
- 1 Application Note Number and Authors
- 2 Authors
- 3 Executive Summary
- 4 Hardware Overview
- 5 Software Overview
- 6 AI in 6G
- 7 5G To 6G Roadmap
- 8 Three Challenges of Data for AI in Wireless
- 9 Overview of Neural Receivers for 5G/6G Systems
- 10 System Validation Checklist
- 11 Wireless Scenario Configuration
- 12 Testbed Equipment List
- 13 Server Hardware Overview
- 14 Dell Precision 5860 Tower
- 15 Consolidated Bill of Materials
Application Note Number and Authors
AN-829
Authors
Bharat Agarwal and Neel Pandeya
Executive Summary
Overview
This Application Note presents a practical, system-level benchmarking platform leveraging NI USRP software-defined radios (SDRs) and the OpenAirInterface (OAI) 5G/NR stack for evaluating AI-enhanced wireless receivers in real-time. It addresses one of the key challenges in deploying AI/ML at the physical layer-ensuring reliable system performance under real-time constraints.
Motivation and Context
AI and ML techniques hold promise for improving both wireless and non-wireless KPIs across the stack, from core-level optimization (e.g., load balancing, power savings), to tightly-timed PHY/MAC innovations such as:
- ML-based digital predistortion to improve power efficiency.
- Neural receivers for channel estimation and symbol detection with improved SNR tolerance.
- Intelligent beam and positioning prediction, even under fast channel dynamics.
Consortia such as 3GPP (Release-18/19) and O-RAN are actively defining how AI/ML can be incorporated into future cellular network standards.
Neural Receiver Model
We demonstrate a real-time implementation of a neural receiver that is based on a published model architecture called DeepRX, which replaces the traditional OFDM receiver blocks (channel estimation, interpolation, equalization, detection) with a single neural network that treats the time-frequency grid data as image-like input. Model training and validation are performed using the NVIDIA Sionna link-level simulator, and training data is stored using the open SigMF format for reproducibility.
More information about the SigMF file format can be found on the project website, on the Wikipedia page, and on the GitHub page.
The original paper, "DeepRx: Fully Convolutional Deep Learning Receiver", by Mikko Honkala, Dani Korpi, Janne M.J. Huttunen, can be found here and here.
Real-Time Benchmarking Platform
To validate the performance of the neural receiver in real hardware, the prototype integrates:
- The OAI real-time 5G protocol stack (complete core, RAN, and UE) running on commodity CPUs.
- NI USRP SDR hardware as the RF front-end.
- An optional O-RAN Near-RT RIC (via FlexRIC) integration.
- Neural receiver inference performed on a GPU (e.g., Nvidia A100, RTX 4070, RTX 4090), accessed via TensorFlow RT C-API for seamless integration within OAI.
This setup enables a direct comparison between the traditional receiver baseline against the neural receiver in an end-to-end real-time system.
Benchmarking Results
Initial testing focuses on uplink performance using various MCS levels (MCS-11, MCS-15, MCS-20 are specifically highlighted in this document) and SNR ranges (5 dB to 18 dB) under a realistic fading channel profile (urban micro, 2 m/s, 45ns delay spread). Each measurement is averaged over 300 transport blocks.
Some of the key findings are listed below.
- The neural receiver shows a clear Bit Error Rate (BER) advantage at lower MCS and lower SNR.
- At higher MCS levels, the performance gap narrows (a trade-off that merits further analysis).
- A reduced uplink bandwidth was used to meet strict real-time latency requirements (500 μs slot duration with 30 KHz SCS).
- The neural receiver model complexity was reduced by 15 times (from 700K to 47K parameters) to achieve real-time GPU inference.
These results underscore the crucial balance between complexity, latency, and performance in AI-enhanced wireless physical-layer deployments.
Conclusions and Implications
The testbed demonstrates a realistic path from simulation to real-time deployment of neural receiver models. This workflow supports rapid prototyping, robust AI model validation, and exploration of architecture-performance trade-offs.
Some key takeaways are listed below.
- AI/ML models can be efficiently integrated into real-time wireless stacks using SDR hardware and GPU inference.
- Low-complexity models offer promising performance improvements while satisfying real-time constraints.
- Synchronized dataset generation and automated test workflows enable scalable ML benchmarking across scenarios.
- The framework allows researchers to investigate unexpected behaviors and robustness in AI-native wireless systems.
Ultimately, the methodology bridges AI/ML conceptual research and realistic deployment, advancing trust and utility in AI-powered future wireless systems.
Hardware Overview
The Universal Software Radio Peripheral (USRP) devices from NI (an Emerson company) are software-defined radios which are widely used for wireless research, prototyping, and education. The hardware specifications for the various USRP devices are listed elsewhere on this Knowledge Base (KB). For this Neural Receiver implementation described in this document, we use the USRP X410. The USRP X440 may also be used, with some further adjustments to the system configuration.
The resources for the USRP X410 are listed below.
The Hardware Resource page for the USRP X410 can be found here.
The product page for the USRP X410 can be found here.
The User Manual for the USRP X410 can be found here.
The resources for the USRP X440 are listed below.
The Hardware Resource page for the USRP X410 can be found here.
The product page for the USRP X410 can be found here.
The User Manual for the USRP X410 can be found here.
The USRP X410 is connected to the host computer using a single QSFP28 100 Gbps Ethernet link, or using a QSFP28-to-SFP28 breakout cable, which provides four 25 Gbps SFP28 Ethernet links. On the host computer, a 100 Gbps or 25 Gbps Ethernet network card is used to connect to the USRP.
The USRP X410 devices are synchronized with the use of a 10 MHz reference signal and a 1 PPS signal, distributed from a common source. This can be provided by the OctoClock-G (see here and here for more information).
For control and management of the USRP X410, a 1 Gbps Ethernet connection to the host computer is needed, as well as a USB serial console connection.
Further details of the hardware configuration will be discussed later in this document.
Software Overview
The software stack running on the computers used in this implementation are as listed below.
- Ubuntu 22.04.5, running on-the-metal, and not in any Virtual Machine (VM)
- UHD version 4.8.0.0
- Nvidia drivers version 535
- Nvidia CUDA version 12.2
- TensorFlow 2.14
For the OAI gNB, the OAI UE, and the FlexRIC, there will be NI-specific versions used, and these will be obtained from an NI repository on GitHub.
Note that the Data Plane Development Kit (DPDK) is not used in this implementation. However, it may may be helpful when using higher sampling rates.
Further details of the software configuration will be discussed later in this document.
AI in 6G
The figure listed below highlights the vision for sixth-generation (6G) wireless systems. Beyond incremental improvements, 6G introduces three major advances, as listed below.
- Spectrum Expansion: Extending from traditional sub-6 GHz and mmWave bands into FR3 (7 to 24 GHz) and sub-THz (up to 300 GHz), enabling ultra-wide bandwidths and unprecedented data rates.
- New Applications: Integration of non-terrestrial networks (NTN) with terrestrial infrastructure and joint communication-and-sensing (JCAS) functionalities, supporting use cases such as connected vehicles, satellite-augmented IoT, and immersive XR.
- Network Optimization: Advancements in massive MIMO, multi-user beamforming, and Open RAN disaggregation, improving spectral efficiency, flexibility, and energy sustainability.
Across these pillars, embedded and trustworthy AI is the key enabler, providing intelligence for spectrum management, adaptive receivers, and end-to-end optimization.
These trends highlight that 6G will operate in highly challenging environments with wideband sub-THz channels, dynamic non-terrestrial links, and complex multi-user MIMO topologies. Traditional linear detection techniques such as ZF or MMSE struggle to cope with hardware non-idealities, nonlinear channel distortions, and the stringent latency and reliability targets of 6G. To address these limitations, the concept of a Neural Receiver has emerged. By embedding deep learning models directly into the receiver chain, neural receivers can learn from real measured impairments, jointly optimize channel estimation and detection, and deliver significant performance gains over classical approaches. This makes neural receivers a key building block for realizing the vision of embedded, trustworthy AI in 6G physical layer design.
5G To 6G Roadmap
The figure listed below illustrates the expected timeline from ongoing 5G research through to the first 6G deployments.
- 5G (Release-16 to Release-18): 3GPP initiated 5G specification development in Release-15 and Release-16, followed by commercial deployments from 2019 onward. Work on Release-17 and Release-18 (2021 to 2024) extends 5G capabilities in areas such as URLLC, industrial IoT, and positioning.
- 5G-Advanced (Release-18 to Release-20): Industry research and specification development converge to define 5G-Advanced features. Deployments are expected around 2025 to 2027, focusing on improved energy efficiency, AI/ML-native functions, and expanded NTN integration.
- 6G (Release-20 onward): Formal 6G technology studies will begin with Release-20 in H2 2025, marked by the first official 6G workshop in March 2025. Standardization of 6G specifications is planned for Release-21 in 2027, with early 6G deployments projected for the end of the decade (around 2030).
The figure above highlights the transition from 5G deployments to the research and standardization cycles of 5G-Advanced and 6G. This staged process ensures backward compatibility, while paving the way for disruptive innovations in spectrum use, AI-native networks, and new application domains.
As shown in the figure above, the transition from 5G to 6G is not only a matter of spectrum expansion and new use cases, but also of embedding AI-native functionalities into the air interface itself. Release-20 (2025) will mark the start of 6G technology studies, providing an opportunity to evaluate disruptive physical layer techniques such as Neural Receivers. These receivers directly integrate deep learning models into the detection chain, enabling them to cope with nonlinearities, hardware impairments, and the extreme bandwidths expected in FR3 and sub-THz bands. By Release 21 (2027), as 6G specifications are defined, neural receivers and other AI-based PHY innovations will play a crucial role in realizing the vision of AI-native 6G, where intelligence is embedded from the physical layer up to the application layer.
Three Challenges of Data for AI in Wireless
This section highlights the three main challenges hindering the seamless integration of AI into wireless communication systems. The challenges are listed with increasing levels of AI readiness.
- Data Scarcity
- Meaning:
- Wireless networks often lack sufficient labeled and diverse datasets.
- Why it's a problem:
- Collecting and labeling large wireless datasets is expensive and time-consuming.
- Real-time data is sparse or kept proprietary by operators/vendors.
- Rare but critical scenarios (handover failures, deep fades, interference spikes) are underrepresented.
- Impact:
- Models trained on limited data risk poor generalization and biased decision-making.
- Meaning:
- Data Quality
- Meaning:
- Available data may not be clean, representative, or consistently labeled.
- Why it's a problem:
- Measurements are noisy due to sensors or network logging errors.
- Labeling mistakes propagate errors into AI models.
- Data is biased toward specific environments (e.g., urban, indoor) and not generalizable to others.
- Impact:
- Low-quality data reduces model reliability, leading to unstable or inaccurate predictions.
- Meaning:
- Data Relevance
- Meaning:
- Even when data exists, it may not directly match the target AI task or deployment scenario.
- Why it's a problem:
- LTE datasets may not transfer well to 5G/6G systems.
- Lab-collected data ignores mobility, blockage, or coexistence effects.
- Training distributions drift away from real-time operational data.
- Impact:
- AI performs well in simulation but degrades in live networks (gap between simulation and real-world).
- Meaning:
The takeaway is that the three challenges of Scarcity, Quality, and Relevance form the key bottleneck for wireless AI, and that addressing them requires:
- Synthetic data generation (digital twins, simulators, ray-tracing),
- Federated learning (distributed training without data centralization),
- Data curation pipelines (cleaning, validation, domain adaptation).
Overview of Neural Receivers for 5G/6G Systems
A neural receiver is a machine learning-based physical layer receiver that replaces or augments traditional signal processing blocks—such as channel estimation, equalization, and detection—with a unified, data-driven model. In contrast to conventional receivers that rely on handcrafted algorithms and strict mathematical models of the wireless channel, neural receivers learn to perform these operations jointly by training on large datasets of labeled I/Q samples or OFDM resource grids.
Comparison with Traditional Receivers
The table below explains some of the differences between components in traditional receiver architectures and components in neural receiver architectures.
| Component | Traditional Receiver | Neural Receiver |
|---|---|---|
| Channel Estimation | Least Squares (LS), MMSE estimators | Learned directly from pilot and data patterns |
| Equalization | Zero-Forcing, MMSE equalizers | Implicitly learned during training |
| Symbol Detection | QAM demodulation, hard/soft decision | Jointly learned with other tasks |
| Architecture | Modular, deterministic | End-to-end differentiable neural network |
| Input | OFDM resource grid or raw IQ samples | IQ tensors or pilot+data grid |
| Output | Estimated bits or LLRs | Bits or probabilities |
Typical Neural Receiver Architecture
A commonly used architecture like DeepRx treats the resource grid as a 2D input, similar to an image, where time and frequency correspond to axes. This allows the use of convolutional neural networks (CNNs), recurrent neural networks (RNNs), or transformer-based models.
- Input: Complex-valued OFDM resource grid, with pilot and data symbols.
- Layers: Convolutional or attention-based layers extract spatial and temporal features.
- Output: Recovered symbols or bits with associated confidence scores.
Training Process
The items below describe the training process that was used in this implementation.
The use of the SigMF data format allows for the storage of raw IQ data with comprehensive metadata, which provides context.
- Dataset: Generated from link-level simulation tools, such as Sionna, under various channel models (AWGN, LOS, NLOS, 3GPP, etc.).
- Format: Datasets are stored in the SigMF format, containing raw IQ samples with metadata.
- Loss Function: Cross-entropy or binary cross-entropy; optionally soft LLR loss for reliability-aware decoding.
- Optimizer: Adam, SGD, or custom schedulers suitable for low-SNR scenarios.
Deployment Aspects
Trained models are deployed in real-time using TensorRT run-times on edge hardware such as:
- GPUs: NVIDIA A100, RTX 4090, RTX 4070, RTX 4060.
- Integration: Plugged into OAI physical-layer receiver chains.
- Latency: Achieves under 500 μs processing delay for 30 KHz SCS, meeting real-time subframe timing requirements.
Performance Summary
The table below shows a comparison of various performance metrics between traditional receiver architectures and neural receiver architectures.
| Metric | Traditional Receiver | Neural Receiver |
|---|---|---|
| Block Error Rate (BLER) under low SNR | Higher | Lower (up to 3 dB gain) |
| Complexity | Fixed, low | Tunable, moderate |
| Latency | Very low | Real-time, under 500 μs |
| Generalization | Poor to unseen channels | Better with diverse training data |
| Interpretability | High (white box) | Lower (black box) |
Use Cases
The list below highlights some common use-cases for Neural Receivers.
- Uplink Neural PHY Receiver: Real-time decoding at gNB.
- Channel Tracking: Adaptive to fast-fading and mobility scenarios.
- Joint Equalization and Detection: Reduces end-to-end BER and BLER.
- Massive MIMO: Scalable to high-dimensional antennas with deep models.
Challenges
There are several challenges to the practical realization of Neural Receivers, as listed below.
- Requires extensive datasets for generalization.
- Less interpretable compared to traditional receiver implementations.
- Hardware deployment must meet strict real-time constraints.
- Careful calibration and interface with existing stacks (such as OAI) are needed.
Summary
The neural receiver presents a promising paradigm for future 6G systems by enabling adaptive, intelligent, and performance-enhancing physical-layer decoding using machine learning. When paired with platforms like the NI USRP radios and AI accelerators, it opens the path for real-time AI-native physical-layer design.
System Validation Checklist
To ensure a consistent and reproducible set-up for our AI-enabled wireless testbed, we employ a systematic validation checklist. The table below summarizes the key checks, commands, and expected outputs that must be verified before conducting experiments. This process guarantees that both the hardware (e.g., USRPs, GPUs) and the software (e.g., operating systems, drivers, TensorFlow, UHD) are correctly installed and aligned with the requirements.
The checklist covers three broad areas:
- Operating system and hardware readiness: Includes verification of the installed OS, BIOS version, kernel, and GPU drivers.
- USRP connectivity and configuration: Ensures that USRPs are discovered, their file systems are compatible with UHD, and network parameters (e.g., MTU size, socket memory) are tuned for high-throughput streaming.
- Software stack and runtime optimization: Covers validation of TensorFlow, NVIDIA TensorRT, and TF C-API installation, as well as disabling unnecessary system services (e.g., updates, GNOME display manager) that may negatively impact performance.
This structured approach minimizes setup errors and improves reproducibility across different machines and deployments.
| Check Item | Command | Desired Output |
|---|---|---|
| Operating System | hostnamectl |
Ubuntu 22.04.5 |
| BIOS version | sudo dmidecode -s bios-version |
Check system vendor for latest version |
| Verify GPU | lspci | grep -i nvidia |
Example: RTX 4090 may appear as 17:00.0 VGA compatible controller: NVIDIA Corporation Device 2684 (rev a1)
|
| Nvidia driver version and CUDA version | nvidia-smi |
Nvidia driver version 535.183.01 and CUDA version 12.2 |
| GPU Load | nvidia-smi |
Load in percentage |
| Kernel version | uname -r |
6.5.0-44-generic
|
| Cores operation mode | cat /sys/devices/system/cpu/cpu*/cpufreq/scaling governor |
All cores should show performance
|
| Cores clock rate | watch -n1 "grep ^[c]pu MHz" /proc/cpuinfo |
Should be larger than base clock rate and less than turbo clock rate |
| UHD version | uhd config info --print-all |
4.8.0.0
|
| IP address of USRP | uhd find devices |
List of all connected USRPs with IP addresses |
| USRP specifications | uhd usrp probe |
All USRP specifications printed without any errors |
| USRP claimed status | uhd find devices |
Claimed: false
|
| MTU of Ethernet ports | ifconfig |
9000
|
| Socket buffer sizes | cat /proc/sys/net/core/rmem max cat /proc/sys/net/core/wmem max |
62500000
|
| Disable System Update | Disable updates in OS settings | Ensures stable test environment |
| Disable GNOME Display Manager | sudo apt remove gdm gdm3 |
Prevents unnecessary excess load on system |
Wireless Scenario Configuration
The table below summarizes the wireless scenario and associated gNB and UE configuration used in the demo setup. The experiment is simplified to a single link between one base station and one user equipment, without interference from neighboring cells or other users.
The system operates in 3GPP NR band n78 with a 40 MHz bandwidth and a subcarrier spacing of 30 KHz using CP-OFDM modulation. A total of six PRBs are scheduled per UE, corresponding to an effective transmission bandwidth of approximately 2.16 MHz.
The TDD frame structure follows a 5 ms periodicity, with 7 downlink slots and 6 additional downlink symbols, complemented by 2 uplink slots and 4 uplink symbols. Uplink transmissions employ PUSCH mapping type B, spanning 13 OFDM symbols. DMRS pilot signals are configured as Type 1 with an additional position pos2, leading to DMRS placement at OFDM symbols 0, 5, and 10.
This baseline configuration provides a controlled environment for evaluating the fundamental link performance of the testbed.
| Parameter | Configuration |
|---|---|
| Wireless Scenario | Single link between one BS (gNB) and one UE No interference from neighboring cells or other UEs |
| Config File | gnb.band78.sa.fr1.106PRB.1x1.usrpx410_3300MHz.conf
|
| Operating Band | n78 |
| Waveform | CP-OFDM |
| Subcarrier Spacing (SCS) | 30 KHz |
| System Bandwidth | 40 MHz |
| Scheduled PRBs per UE | 6 |
| Scheduled Transmission Bandwidth | 2.16 MHz |
| TDD Pattern | DL/UL periodicity of 5 ms |
| Downlink Slots | 7 |
| Downlink Symbols | 6 |
| Uplink Slots | 2 |
| Uplink Symbols | 4 |
| PUSCH Mapping | Type B, duration = 13 OFDM symbols |
| PUSCH DMRS Configuration | Type 1, dmrs-AdditionalPosition = pos2 DMRS in OFDM symbols 0, 5, and 10 |
Testbed Equipment List
The table listed below details the bill of materials used in the experimental 5G testbed setup. The infrastructure is divided into separate blocks for data recording, gNB, UE, and RF hardware.
The host computers for the gNB and UE are each implemented on a Dell Precision 5860 server, which has an Intel Xeon W7-2495X CPU (24 physical cores with turbo frequency of 4.8 GHz), and 25 Gbps Ethernet network cards. The gNB system also includes an Nvidia RTX 4090 GPU to accelerate AI-based signal processing workloads. The UE system uses a much less powerful GPU, as no AI acceleration is needed.
Both gNB and UE connect to USRP X410 radios via a Mellanox Nvidia ConnectX-5 MCX512A-ACAT network card (the Intel X710-DA2 may also be used), and QSFP28-to-SFP28 breakout cables.
The OctoClock-G distributes a common 10 MHz reference signal and 1 PPS signal across all USRP devices.
This hardware configuration ensures support for wideband operation, supporting channel bandwidths from 40 MHz to 100 MHz, and multiple MIMO layers, enabling advanced AI-driven signal processing and performance evaluation in a realistic testbed.
Server Hardware Overview
The testbed is designed to operate on a range of high-performance workstation and server platforms that provide sufficient compute, I/O, and GPU acceleration for 5G/6G physical-layer and application-layer experimentation. The hardware listed below represents configurations that have been tested and validated for different use cases such as baseband processing, AI-driven link adaptation, and machine learning training. Note that other hardware configurations are possible.
Listed below are the supported, validated, and tested server platforms:
- Dell Precision 5860 Tower: This system is the current baseline workstation for gNB and UE host computer deployments. It supports PCIe Gen-4, 25 Gbps and 100 Gbps Ethernet network cards, and large DDR5 memory configurations.
- Lenovo ThinkStation P8: Verified for both inference and control-plane workloads; features comparable expansion and thermal capacity to the 5860.
- Gigabyte Server Platform: Used for high-throughput data recording and multi-GPU configurations, supporting rackmount deployment scenarios.
- Dell Precision 5820 Tower: This is a legacy system that has been succeeded by the Dell 5860. Previously used in early iterations. It is still functional, but limited to PCIe Gen-3 and DDR4 memory.
Listed below are the tested and recommended GPUs:
- Nvidia RTX 4060, 4070, 4090: Validated for real-time baseband acceleration, AI inference, and model training tasks. The RTX 4090 provides the best trade-off between power, cost, and tensor performance.
- Nvidia A100: This is a legacy GPU that was previously used for large-scale neural receiver training. It has been superseded by RTX-class GPUs for compact testbed setups.
These GPU and server combinations provide the flexibility to deploy either component (either the gNB or the UE) on the same hardware architecture, simplifying replication and scaling across the testbed. Systems with PCIe Gen-4 or higher are recommended to ensure sufficient bandwidth for 25 Gbps and 100 Gbps Ethernet cards and for real-time data streaming from USRP X410 devices.
Additional compatible systems and updated configuration guidance can be found on the USRP X410 page in the Knowledge Base (KB), which includes validated NICs, timing sources, and synchronization accessories.
Dell Precision 5860 Tower
The Dell Precision 5860 Tower workstation is used as the main server platform for the gNB and UE in the 5G/6G testbed. It provides the compute, memory, and I/O performance required for advanced AI-driven wireless communication experiments.
Some of the key features of this system are:
- High-performance workstation designed for AI-driven wireless testbeds.
- Supports multiple NICs and high-throughput data interfaces.
- Optimized for GPU acceleration (e.g., NVIDIA RTX 4090).
- Expandable PCIe slots for USRP connectivity and RF front-ends.
- Large DDR5 memory capacity for parallel workloads.
- Suitable for both gNB and UE roles in 5G/6G experiments
The Dell Precision 5860 Tower was selected as the system for the gNB because it supports: • Multiple 25 Gbps Ethernet cards for high-speed connectivity with USRP X410 radios and data recording servers. • Nvidia RTX-class GPUs to accelerate physical-layer baseband processing and AI workloads. • Large DDR5 memory (up to 1 TB) for handling real-time scheduling and parallel computation. • The Intel Xeon W7-2495X CPU, with 24 physical cores and 4.50 GHz turbo clock, to meet the demands of high sampling rates and multi-layer MIMO.
This flexibility allows the same hardware platform to serve as both gNB and UE, simplifying testbed deployment and ensuring scalability for future AI-driven enhancements.
Note that as the sample rate increases, for example, to 122.88 Msps for a 100 MHz channel bandwidth, and also as the number number of MIMO layers increases, for example, to 2x2 and 4x4, then the computational demands increase significantly, and more and more physical cores would be needed to support this load.
Consolidated Bill of Materials
Listed below is a consolidated Bill of Materials (BoM) for all the components used in the implementation of this system. Note that the specific computers listed are not necessarily or specifically required, and that computers from other vendors may be used as long as they have a similar level of performance and capability.
- Core Network (CN) host system:
- Dell 3680 Precision Tower, with Intel i9-10980XE CPU, with 18 physical cores, with 3.00 GHz base clock frequency.
- Mellanox Nvidia ConnectX-5 MCX512A-ACAT network card, with two SFP28 ports.
- gNB host system:
- Dell 5860 Precision Tower, with Intel Xeon W7-2495X CPU, with 24 physical cores, and with turbo frequency of 4.8 GHz.
- Mellanox Nvidia ConnectX-5 MCX512A-ACAT network card, with two SFP28 ports.
- Nvidia GeForce RTX 4090 GPU.
- GPU power cable, 30cm, 12+4-pin male to 2x8-pin female sleeved 12v extension cord for GeForce RTX 4090.
- QSFP28-to-SFP28 25 Gbps Ethernet breakout cable to connect with USRP X410.
- Nvidia MCP2M00-A002E30N 100 GbE to 4x25GbE (QSFP28 to 4xSFP28) Direct Attach Copper (DAC) splitter cable (here).
- NI 100 GbE to 4x25GbE (QSFP28 to 4xSFP28) splitter cable (NI Part Number 788214-01).
- USRP X410 radio.
- One connected to the gNB system.
- One connected to the UE system.
- OctoClock-G CDA-2990 for synchronization between the gNB USRP X410 and the UE USRP X410.
- RF cables (1 meter) (Quantity 8).
- 20 dB RF attenuators (Quantity 2).
- Two-port RF splitter, 6 GHz, SMA connectors (Quantity 1).
- Ethernet switch, 5 ports (Quantity 1).
- Cat-6A RJ-45 Ethernet cables (Quantity 4).
| MMM | MMM | MMM |
|---|---|---|
| mmm | mmm | mmm |
| mmm | mmm | mmm |
| mmm | mmm | mmm |
| mmm | mmm | mmm |
| mmm | mmm | mmm |
| mmm | mmm | mmm |
| mmm | mmm | mmm |
| mmm | mmm | mmm |
| mmm | mmm | mmm |
