A unified, plug-and-play toolbox for building tensor equivariant neural networks (TENN), designed to support communication system applications such as multi-user multiple-input multiple-output (MU-MIMO) precoding, user scheduling, channel estimation, detection, demodulation, and so on. More information can be found in paper "Towards Unified AI Models for MU-MIMO Communications: A Tensor Equivariance Framework", which has been accepted for publication on top journal IEEE Transactions on Wireless Communications.
- IEEE DOI:10.1109/TWC.2025.3580321 - arXiv (with detailed appendix) arXiv:2406.09022
This toolbox implements a unified framework for leveraging Tensor Equivariance (TE) into neural network design. By exploiting the inherent TE of the target mapping, it facilitates scalable and efficient learning in AI-assisted wireless communication systems. TE extends the concept of permutation equivariance to high-dimensional tensors and includes:
- Multidimensional Equivariance (MDE): Permuting each tensor dimension independently results in the same permutation at the output.
- High-Order Equivariance (HOE): The same permutation is applied across multiple dimensions simultaneously.
- Multidimensional Invariance (MDI): Output remains unchanged under permutations along specified dimensions.
- π± Scalable: Generalizes to varying input sizes without retraining.
- β‘ Efficient: Requires fewer parameters, lower computational complexity, and smaller training sets.
- π Unified: Compatible with data/model-driven approaches, as well as various learning paradigms.
- π¦ Modular: Drop-in layers for various types of equivariance.
- π‘ Application-ready: Comes with precoding and scheduling examples for MU-MIMO communications.
This toolbox includes several pluggable neural modules, each aligned with a theoretical symmetry property:
Module π§© (abbr.) | Function βοΈ | Dimensions βΎοΈ |
---|---|---|
MDE |
The equivalent linear module when any fully connected layer satisfies permutation equivariance across an arbitrary number of dimensions. |
In: Out: |
HOE |
The equivalent linear module when an arbitrary fully connected layer exhibits equivariance to identical permutations across multiple input and output dimensions. (taking 1-2-order equivariance as an example). |
In: Out: |
MDI |
A nonlinear module based on the attention mechanism that satisfies permutation invariance across an arbitrary number of dimensions. |
In: Out: |
The PrecodingTECFP
network includes:
- A multidimensional equivariant network
- A multidimensional invariant module
- A high-order equivariant module
Under the model-driven approach, the network maps the CSI to the optimal auxiliary variables in the closed-form precoding expression, solving the WMMSE precoding problem for MU-MIMO systems.
The SchedulingTEUSN
network, trained with both WMMSE and MMSE encoding methods, includes:
- A multidimensional equivariant network
- A multidimensional invariant module
Under the data-driven approach, the network maps the CSI to the optimal user scheduling indicator, solving the user scheduling optimization problem in MU-MIMO systems.
Contains core model definitions:
init_func.py
: Initialization functions, common utilities, and parameter management classesTE_models.py
: Multidimensional equivariant network and pattern-generation functionsTE_module.py
: Multidimensional equivariant and invariant modules, and high-order equivariant module
Folder for storing training and testing data.
For Example 1 (Precoding):
- Channel data files named
"data_name.mat"
with dimensions[sample_num, ue_num, rx_ant_num, tx_ant_num]
For Example 2 (Scheduling):
- Channel data files named
"data_name.mat"
with dimensions[sample_num, ue_num, rx_ant_num, tx_ant_num]
- Eta label files named
"data_name_etaMMSE"
or"data_name_etaWMMSE"
with dimensions[sample_num, snr_num, ue_num]
Contains precoding-related models and training code (Example 1):
precoding_func.py
: Training and testing functionsprecoding_models.py
: Network model definitions and related functionsprecoding_test.py
: Main testing programprecoding_train.py
: Main training program
Contains scheduling-related models and training code (Example 2):
scheduling_func.py
: Training and testing functionsscheduling_MMSE_test.py
: Testing program for MMSE-trained networkscheduling_WMMSE_test.py
: Testing program for WMMSE-trained networkscheduling_MMSE_train.py
: Training program for MMSE labelsscheduling_WMMSE_train.py
: Training program for WMMSE labelsscheduling_models.py
: Network model definitions and related functions
Stores network training results.
The MDE_Network
class implements a multidimensional equivariant neural network architecture. It is constructed by stacking MDE modules and allows for significant reduction in parameter count and computational complexity by selecting different patterns, while preserving performance as much as possible (see Section III-A of the paper for the design and selection of patterns).
from TE_models import MDE_Network
from TE_models import generate_patterns
# Define the pattern of the network
MDE_dim_list = generate_patterns(n_layer=3, n_dim=3, pattern='original')
# Initialize the network
network = MDE_Network(
d_input=4, # Input dimension
d_output=8, # Output dimension
n_layer=3, # Number of equivariant layers
d_hidden=32, # Hidden layer dimension
dim_list=MDE_dim_list # List of dimensions for equivariance
)
# Input tensor shape: [batch_size, dim_1, ..., dim_n, d_input]
# Output tensor shape: [batch_size, dim_1, ..., dim_n, d_output]
output = network(input_tensor)
The MDI_Module
class implements a multidimensional invariant neural network module.
from TE_module import MDI_Module
# Initialize the module
mdi_module = MDI_Module(
d_feature=64, # Feature dimension
num_heads=8, # Number of attention heads
dim=[1, 2] # Invariant Dimensions
)
# Input tensor shape: [batch_size, M1, M2, ..., MN, d_feature]
# Output tensor shape: [batch_size, M1, M2, ..., MK, d_feature], where dimensions in 'dim' are removed
output = mdi_module(input_tensor)
If you use this toolbox in your research, please cite our paper:
Y. Wang, H. Hou, X. Yi, W. Wang, S. Jin, βTowards unified AI models for MU-MIMO communications: A tensor equivariance framework,β IEEE Trans. Wireless Commun., Early Access, 2025.
@ARTICLE{tenntoolbox,
author = {Wang, Yafei and Hou, Hongwei and Yi, Xinping and Wang, Wenjin and Jin, Shi},
journal = {IEEE Transactions on Wireless Communications},
title = {Towards unified AI models for MU-MIMO communications: A tensor equivariance framework},
year = {2025},
volume = {},
number = {},
pages = {1-1},
doi = {10.1109/TWC.2025.3580321}
}
We are grateful to Jinshuo Zhang (Southeast University) for his dedicated efforts in preparing and refining the code for open-source release. We also sincerely thank all the reviewers and editors of this paper for their insightful comments and valuable suggestions.