Multi-channel U-Net for Music Source Separation
- URL: http://arxiv.org/abs/2003.10414v3
- Date: Fri, 4 Sep 2020 13:37:58 GMT
- Title: Multi-channel U-Net for Music Source Separation
- Authors: Venkatesh S. Kadandale, Juan F. Montesinos, Gloria Haro, Emilia
G\'omez
- Abstract summary: Conditioned U-Net (C-U-Net) uses a control mechanism to train a single model for multi-source separation.
We propose a multi-channel U-Net (M-U-Net) trained using a weighted multi-task loss.
- Score: 3.814858728853163
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A fairly straightforward approach for music source separation is to train
independent models, wherein each model is dedicated for estimating only a
specific source. Training a single model to estimate multiple sources generally
does not perform as well as the independent dedicated models. However,
Conditioned U-Net (C-U-Net) uses a control mechanism to train a single model
for multi-source separation and attempts to achieve a performance comparable to
that of the dedicated models. We propose a multi-channel U-Net (M-U-Net)
trained using a weighted multi-task loss as an alternative to the C-U-Net. We
investigate two weighting strategies for our multi-task loss: 1) Dynamic
Weighted Average (DWA), and 2) Energy Based Weighting (EBW). DWA determines the
weights by tracking the rate of change of loss of each task during training.
EBW aims to neutralize the effect of the training bias arising from the
difference in energy levels of each of the sources in a mixture. Our methods
provide three-fold advantages compared to C-UNet: 1) Fewer effective training
iterations per epoch, 2) Fewer trainable network parameters (no control
parameters), and 3) Faster processing at inference. Our methods achieve
performance comparable to that of C-U-Net and the dedicated U-Nets at a much
lower training cost.
Related papers
- One QuantLLM for ALL: Fine-tuning Quantized LLMs Once for Efficient Deployments [43.107261545706415]
Large Language Models (LLMs) have advanced rapidly but face significant memory demands.
Current methods typically require lengthy training to alleviate the performance degradation from quantization loss.
We make an initial attempt to extend the once-for-all framework to large language models.
arXiv Detail & Related papers (2024-05-30T16:05:15Z) - Efficient Remote Sensing with Harmonized Transfer Learning and Modality Alignment [0.0]
"Harmonized Transfer Learning and Modality alignment (HarMA)" is a method that simultaneously satisfies task constraints, modality alignment, and single-modality uniform alignment.
HarMA achieves state-of-the-art performance in two popular multimodal retrieval tasks in the field of remote sensing.
arXiv Detail & Related papers (2024-04-28T17:20:08Z) - Dynamic Pre-training: Towards Efficient and Scalable All-in-One Image Restoration [100.54419875604721]
All-in-one image restoration tackles different types of degradations with a unified model instead of having task-specific, non-generic models for each degradation.
We propose DyNet, a dynamic family of networks designed in an encoder-decoder style for all-in-one image restoration tasks.
Our DyNet can seamlessly switch between its bulkier and lightweight variants, thereby offering flexibility for efficient model deployment.
arXiv Detail & Related papers (2024-04-02T17:58:49Z) - BEND: Bagging Deep Learning Training Based on Efficient Neural Network Diffusion [56.9358325168226]
We propose a Bagging deep learning training algorithm based on Efficient Neural network Diffusion (BEND)
Our approach is simple but effective, first using multiple trained model weights and biases as inputs to train autoencoder and latent diffusion model.
Our proposed BEND algorithm can consistently outperform the mean and median accuracies of both the original trained model and the diffused model.
arXiv Detail & Related papers (2024-03-23T08:40:38Z) - Distilling from Similar Tasks for Transfer Learning on a Budget [38.998980344852846]
Transfer learning is an effective solution for training with few labels, however often at the expense of a computationally costly fine-tuning of large base models.
We propose to mitigate this unpleasant trade-off between compute and accuracy via semi-supervised cross-domain distillation.
Our methods need no access to source data, and merely need features and pseudo-labels of the source models.
arXiv Detail & Related papers (2023-04-24T17:59:01Z) - Vertical Layering of Quantized Neural Networks for Heterogeneous
Inference [57.42762335081385]
We study a new vertical-layered representation of neural network weights for encapsulating all quantized models into a single one.
We can theoretically achieve any precision network for on-demand service while only needing to train and maintain one model.
arXiv Detail & Related papers (2022-12-10T15:57:38Z) - Conv-NILM-Net, a causal and multi-appliance model for energy source
separation [1.1355370218310157]
Non-Intrusive Load Monitoring seeks to save energy by estimating individual appliance power usage from a single aggregate measurement.
Deep neural networks have become increasingly popular in attempting to solve NILM problems.
We propose Conv-NILM-net, a fully convolutional framework for end-to-end NILM.
arXiv Detail & Related papers (2022-08-03T15:59:32Z) - Controllable Dynamic Multi-Task Architectures [92.74372912009127]
We propose a controllable multi-task network that dynamically adjusts its architecture and weights to match the desired task preference as well as the resource constraints.
We propose a disentangled training of two hypernetworks, by exploiting task affinity and a novel branching regularized loss, to take input preferences and accordingly predict tree-structured models with adapted weights.
arXiv Detail & Related papers (2022-03-28T17:56:40Z) - MutualNet: Adaptive ConvNet via Mutual Learning from Different Model
Configurations [51.85020143716815]
We propose MutualNet to train a single network that can run at a diverse set of resource constraints.
Our method trains a cohort of model configurations with various network widths and input resolutions.
MutualNet is a general training methodology that can be applied to various network structures.
arXiv Detail & Related papers (2021-05-14T22:30:13Z) - Rethinking Importance Weighting for Deep Learning under Distribution
Shift [86.52964129830706]
Under distribution shift (DS) where the training data distribution differs from the test one, a powerful technique is importance weighting (IW) which handles DS in two separate steps.
In this paper, we rethink IW and theoretically show it suffers from a circular dependency.
We propose an end-to-end solution dynamic IW that iterates between WE and WC and combines them in a seamless manner.
arXiv Detail & Related papers (2020-06-08T15:03:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.