The Pseudo Projection Operator: Applications of Deep Learning to
Projection Based Filtering in Non-Trivial Frequency Regimes
- URL: http://arxiv.org/abs/2111.07140v1
- Date: Sat, 13 Nov 2021 16:09:14 GMT
- Title: The Pseudo Projection Operator: Applications of Deep Learning to
Projection Based Filtering in Non-Trivial Frequency Regimes
- Authors: Matthew L. Weiss, Nathan C. Frey, Siddharth Samsi, Randy C. Paffenroth
and Vijay Gadepally
- Abstract summary: We introduce a PO-neural network hybrid model, the Pseudo Projection Operator (PPO), which leverages a neural network to perform frequency selection.
We compare the filtering capabilities of a PPO, PO, and denoising autoencoder (DAE) on the University of Rochester Multi-Modal Music Performance dataset.
In the majority of experiments, the PPO outperforms both the PO and DAE.
- Score: 5.632784019776093
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traditional frequency based projection filters, or projection operators (PO),
separate signal and noise through a series of transformations which remove
frequencies where noise is present. However, this technique relies on a priori
knowledge of what frequencies contain signal and noise and that these
frequencies do not overlap, which is difficult to achieve in practice. To
address these issues, we introduce a PO-neural network hybrid model, the Pseudo
Projection Operator (PPO), which leverages a neural network to perform
frequency selection. We compare the filtering capabilities of a PPO, PO, and
denoising autoencoder (DAE) on the University of Rochester Multi-Modal Music
Performance Dataset with a variety of added noise types. In the majority of
experiments, the PPO outperforms both the PO and DAE. Based upon these results,
we suggest future application of the PPO to filtering problems in the physical
and biological sciences.
Related papers
- FilterNet: Harnessing Frequency Filters for Time Series Forecasting [34.83702192033196]
FilterNet is built upon our proposed learnable frequency filters to extract key informative temporal patterns by selectively passing or attenuating certain components of time series signals.
equipped with the two filters, FilterNet can approximately surrogate the linear and attention mappings widely adopted in time series literature.
arXiv Detail & Related papers (2024-11-03T16:20:41Z) - On Optimal Sampling for Learning SDF Using MLPs Equipped with Positional
Encoding [79.67071790034609]
We devise a tool to determine the appropriate sampling rate for learning an accurate neural implicit field without undesirable side effects.
It is observed that a PE-equipped has an intrinsic frequency much higher than the highest frequency component in the PE layer.
We empirically show in the setting of SDF fitting, this recommended sampling rate is sufficient to secure accurate fitting results.
arXiv Detail & Related papers (2024-01-02T10:51:52Z) - Instabilities in Convnets for Raw Audio [1.5060156580765574]
We present a theory of large deviations for the energy response of FIR filterbanks with random Gaussian weights.
We find that deviations worsen for large filters and locally periodic input signals.
Numerical simulations align with our theory and suggest that the condition number of a convolutional layer follows a logarithmic scaling law.
arXiv Detail & Related papers (2023-09-11T22:34:06Z) - Spectral analysis for noise diagnostics and filter-based digital error
mitigation [0.0]
We quantify the additional, higher frequency modes in the output signal caused by device errors.
We show that filtering these noise-induced modes effectively mitigates device errors.
arXiv Detail & Related papers (2022-06-17T14:42:58Z) - SpecGrad: Diffusion Probabilistic Model based Neural Vocoder with
Adaptive Noise Spectral Shaping [51.698273019061645]
SpecGrad adapts the diffusion noise so that its time-varying spectral envelope becomes close to the conditioning log-mel spectrogram.
It is processed in the time-frequency domain to keep the computational cost almost the same as the conventional DDPM-based neural vocoders.
arXiv Detail & Related papers (2022-03-31T02:08:27Z) - Deep Frequency Filtering for Domain Generalization [55.66498461438285]
Deep Neural Networks (DNNs) have preferences for some frequency components in the learning process.
We propose Deep Frequency Filtering (DFF) for learning domain-generalizable features.
We show that applying our proposed DFF on a plain baseline outperforms the state-of-the-art methods on different domain generalization tasks.
arXiv Detail & Related papers (2022-03-23T05:19:06Z) - Filter-enhanced MLP is All You Need for Sequential Recommendation [89.0974365344997]
In online platforms, logged user behavior data is inevitable to contain noise.
We borrow the idea of filtering algorithms from signal processing that attenuates the noise in the frequency domain.
We propose textbfFMLP-Rec, an all-MLP model with learnable filters for sequential recommendation task.
arXiv Detail & Related papers (2022-02-28T05:49:35Z) - Sampling-Frequency-Independent Audio Source Separation Using Convolution
Layer Based on Impulse Invariant Method [67.24600975813419]
We propose a convolution layer capable of handling arbitrary sampling frequencies by a single deep neural network.
We show that the introduction of the proposed layer enables a conventional audio source separation model to consistently work with even unseen sampling frequencies.
arXiv Detail & Related papers (2021-05-10T02:33:42Z) - Multi-stream Convolutional Neural Network with Frequency Selection for
Robust Speaker Verification [2.3437178262034095]
We propose a novel framework of multi-stream Convolutional Neural Network (CNN) for speaker verification tasks.
The proposed framework accommodates diverse temporal embeddings generated from multiple streams to enhance the robustness of acoustic modeling.
We conduct extensive experiments on VoxCeleb dataset, and the experimental results demonstrate that multi-stream CNN significantly outperforms single-stream baseline.
arXiv Detail & Related papers (2020-12-21T07:23:40Z) - On Filter Generalization for Music Bandwidth Extension Using Deep Neural
Networks [0.40611352512781856]
We formulate the bandwidth extension problem using deep neural networks, where a band-limited signal is provided as input to the network.
Our main contribution centers on the impact of the choice of low pass filter when training and subsequently testing the network.
We propose a data augmentation strategy which utilizes multiple low pass filters during training and leads to improved generalization to unseen filtering conditions at test time.
arXiv Detail & Related papers (2020-11-14T11:41:28Z) - Noise Homogenization via Multi-Channel Wavelet Filtering for
High-Fidelity Sample Generation in GANs [47.92719758687014]
We propose a novel multi-channel wavelet-based filtering method for Generative Adversarial Networks (GANs)
When embedding a wavelet deconvolution layer in the generator, the resultant GAN, called WaveletGAN, takes advantage of the wavelet deconvolution to learn a filtering with multiple channels.
We conducted benchmark experiments on the Fashion-MNIST, KMNIST and SVHN datasets through an open GAN benchmark tool.
arXiv Detail & Related papers (2020-05-14T03:40:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.