RE-MIMO: Recurrent and Permutation Equivariant Neural MIMO Detection
- URL: http://arxiv.org/abs/2007.00140v2
- Date: Sat, 23 Jan 2021 10:43:53 GMT
- Title: RE-MIMO: Recurrent and Permutation Equivariant Neural MIMO Detection
- Authors: Kumar Pratik, Bhaskar D. Rao, Max Welling
- Abstract summary: We present a novel neural network for symbol detection in wireless communication systems.
It is motivated by several important considerations in wireless communication systems.
We compare its performance against existing methods and the results show the ability of our network to efficiently handle a variable number of transmitters.
- Score: 85.44877328116881
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we present a novel neural network for MIMO symbol detection.
It is motivated by several important considerations in wireless communication
systems; permutation equivariance and a variable number of users. The neural
detector learns an iterative decoding algorithm that is implemented as a stack
of iterative units. Each iterative unit is a neural computation module
comprising of 3 sub-modules: the likelihood module, the encoder module, and the
predictor module. The likelihood module injects information about the
generative (forward) process into the neural network. The encoder-predictor
modules together update the state vector and symbol estimates. The encoder
module updates the state vector and employs a transformer based attention
network to handle the interactions among the users in a permutation equivariant
manner. The predictor module refines the symbol estimates. The modular and
permutation equivariant architecture allows for dealing with a varying number
of users. The resulting neural detector architecture is unique and exhibits
several desirable properties unseen in any of the previously proposed neural
detectors. We compare its performance against existing methods and the results
show the ability of our network to efficiently handle a variable number of
transmitters with high accuracy.
Related papers
- Locality-Aware Generalizable Implicit Neural Representation [54.93702310461174]
Generalizable implicit neural representation (INR) enables a single continuous function to represent multiple data instances.
We propose a novel framework for generalizable INR that combines a transformer encoder with a locality-aware INR decoder.
Our framework significantly outperforms previous generalizable INRs and validates the usefulness of the locality-aware latents for downstream tasks.
arXiv Detail & Related papers (2023-10-09T11:26:58Z) - Permutation Equivariant Neural Functionals [92.0667671999604]
This work studies the design of neural networks that can process the weights or gradients of other neural networks.
We focus on the permutation symmetries that arise in the weights of deep feedforward networks because hidden layer neurons have no inherent order.
In our experiments, we find that permutation equivariant neural functionals are effective on a diverse set of tasks.
arXiv Detail & Related papers (2023-02-27T18:52:38Z) - Detecting Information Relays in Deep Neural Networks [0.0]
We introduce a new information-theoretic concept that proves useful in understanding and analyzing a network's functional modularity.
The relay information measures how much information groups of neurons that participate in a particular function (modules) relay from inputs to outputs.
We show that the functionality of modules correlates with the amount of relay information they carry.
arXiv Detail & Related papers (2023-01-03T01:02:51Z) - Learning to Learn with Generative Models of Neural Network Checkpoints [71.06722933442956]
We construct a dataset of neural network checkpoints and train a generative model on the parameters.
We find that our approach successfully generates parameters for a wide range of loss prompts.
We apply our method to different neural network architectures and tasks in supervised and reinforcement learning.
arXiv Detail & Related papers (2022-09-26T17:59:58Z) - Variable Bitrate Neural Fields [75.24672452527795]
We present a dictionary method for compressing feature grids, reducing their memory consumption by up to 100x.
We formulate the dictionary optimization as a vector-quantized auto-decoder problem which lets us learn end-to-end discrete neural representations in a space where no direct supervision is available.
arXiv Detail & Related papers (2022-06-15T17:58:34Z) - Modeling Structure with Undirected Neural Networks [20.506232306308977]
We propose undirected neural networks, a flexible framework for specifying computations that can be performed in any order.
We demonstrate the effectiveness of undirected neural architectures, both unstructured and structured, on a range of tasks.
arXiv Detail & Related papers (2022-02-08T10:06:51Z) - Deep Convolutional Learning-Aided Detector for Generalized Frequency
Division Multiplexing with Index Modulation [0.0]
The proposed method first pre-processes the received signal by using a zero-forcing (ZF) detector and then uses a neural network consisting of a convolutional neural network (CNN) followed by a fully-connected neural network (FCNN)
The FCNN part uses only two fully-connected layers, which can be adapted to yield a trade-off between complexity and bit error rate (BER) performance.
It has been demonstrated that the proposed deep convolutional neural network-based detection and demodulation scheme provides better BER performance compared to ZF detector with a reasonable complexity increase.
arXiv Detail & Related papers (2022-02-06T22:18:42Z) - Multitask Learning and Joint Optimization for Transformer-RNN-Transducer
Speech Recognition [13.198689566654107]
This paper explores multitask learning, joint optimization, and joint decoding methods for transformer-RNN-transducer systems.
We show that the proposed methods can reduce word error rate (WER) by 16.6 % and 13.3 % for test-clean and test-other datasets, respectively.
arXiv Detail & Related papers (2020-11-02T06:38:06Z) - Incremental Training of a Recurrent Neural Network Exploiting a
Multi-Scale Dynamic Memory [79.42778415729475]
We propose a novel incrementally trained recurrent architecture targeting explicitly multi-scale learning.
We show how to extend the architecture of a simple RNN by separating its hidden state into different modules.
We discuss a training algorithm where new modules are iteratively added to the model to learn progressively longer dependencies.
arXiv Detail & Related papers (2020-06-29T08:35:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.