Complete and separate: Conditional separation with missing target source
attribute completion
- URL: http://arxiv.org/abs/2307.14609v1
- Date: Thu, 27 Jul 2023 03:53:53 GMT
- Title: Complete and separate: Conditional separation with missing target source
attribute completion
- Authors: Dimitrios Bralios, Efthymios Tzinis, Paris Smaragdis
- Abstract summary: We present an approach in which a model, given an input mixture and partial semantic information about a target source, is trained to extract additional semantic data.
We then leverage this pre-trained model to improve the separation performance of an uncoupled multi-conditional separation network.
- Score: 27.215800308343322
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent approaches in source separation leverage semantic information about
their input mixtures and constituent sources that when used in conditional
separation models can achieve impressive performance. Most approaches along
these lines have focused on simple descriptions, which are not always useful
for varying types of input mixtures. In this work, we present an approach in
which a model, given an input mixture and partial semantic information about a
target source, is trained to extract additional semantic data. We then leverage
this pre-trained model to improve the separation performance of an uncoupled
multi-conditional separation network. Our experiments demonstrate that the
separation performance of this multi-conditional model is significantly
improved, approaching the performance of an oracle model with complete semantic
information. Furthermore, our approach achieves performance levels that are
comparable to those of the best performing specialized single conditional
models, thus providing an easier to use alternative.
Related papers
- SynerGraph: An Integrated Graph Convolution Network for Multimodal Recommendation [1.3812010983144802]
This article presents a novel approach to multimodal recommendation systems, focusing on integrating and purifying multimodal data.
We developed a filter to remove noise from various types of data, making the recommendations more reliable.
We studied the impact of top-K sparsification on different datasets, finding optimal values that strike a balance between underfitting and overfitting concerns.
arXiv Detail & Related papers (2024-05-29T12:18:32Z) - Versatile Medical Image Segmentation Learned from Multi-Source Datasets via Model Self-Disambiguation [9.068045557591612]
We propose a cost-effective alternative that harnesses multi-source data with only partial or sparse segmentation labels for training.
We devise strategies for model self-disambiguation, prior knowledge incorporation, and imbalance mitigation to tackle challenges associated with inconsistently labeled multi-source data.
arXiv Detail & Related papers (2023-11-17T18:28:32Z) - Convolutional autoencoder-based multimodal one-class classification [80.52334952912808]
One-class classification refers to approaches of learning using data from a single class only.
We propose a deep learning one-class classification method suitable for multimodal data.
arXiv Detail & Related papers (2023-09-25T12:31:18Z) - PAMI: partition input and aggregate outputs for model interpretation [69.42924964776766]
In this study, a simple yet effective visualization framework called PAMI is proposed based on the observation that deep learning models often aggregate features from local regions for model predictions.
The basic idea is to mask majority of the input and use the corresponding model output as the relative contribution of the preserved input part to the original model prediction.
Extensive experiments on multiple tasks confirm the proposed method performs better than existing visualization approaches in more precisely finding class-specific input regions.
arXiv Detail & Related papers (2023-02-07T08:48:34Z) - Learning Data Representations with Joint Diffusion Models [20.25147743706431]
Joint machine learning models that allow synthesizing and classifying data often offer uneven performance between those tasks or are unstable to train.
We extend the vanilla diffusion model with a classifier that allows for stable joint end-to-end training with shared parameterization between those objectives.
The resulting joint diffusion model outperforms recent state-of-the-art hybrid methods in terms of both classification and generation quality on all evaluated benchmarks.
arXiv Detail & Related papers (2023-01-31T13:29:19Z) - Dataless Knowledge Fusion by Merging Weights of Language Models [51.8162883997512]
Fine-tuning pre-trained language models has become the prevalent paradigm for building downstream NLP models.
This creates a barrier to fusing knowledge across individual models to yield a better single model.
We propose a dataless knowledge fusion method that merges models in their parameter space.
arXiv Detail & Related papers (2022-12-19T20:46:43Z) - Optimal Condition Training for Target Source Separation [56.86138859538063]
We propose a new optimal condition training method for single-channel target source separation.
We show that the complementary information carried by the diverse semantic concepts significantly helps to disentangle and isolate sources of interest.
arXiv Detail & Related papers (2022-11-11T00:04:55Z) - Investigating Ensemble Methods for Model Robustness Improvement of Text
Classifiers [66.36045164286854]
We analyze a set of existing bias features and demonstrate there is no single model that works best for all the cases.
By choosing an appropriate bias model, we can obtain a better robustness result than baselines with a more sophisticated model design.
arXiv Detail & Related papers (2022-10-28T17:52:10Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.