TensorFlow Audio Models in Essentia
- URL: http://arxiv.org/abs/2003.07393v1
- Date: Mon, 16 Mar 2020 18:23:30 GMT
- Title: TensorFlow Audio Models in Essentia
- Authors: Pablo Alonso-Jim\'enez, Dmitry Bogdanov, Jordi Pons, Xavier Serra
- Abstract summary: We present a set of algorithms that employ in Essentia.
Essentia is a reference open-source C++/Python library for audio and music analysis.
- Score: 28.324123632999527
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Essentia is a reference open-source C++/Python library for audio and music
analysis. In this work, we present a set of algorithms that employ TensorFlow
in Essentia, allow predictions with pre-trained deep learning models, and are
designed to offer flexibility of use, easy extensibility, and real-time
inference. To show the potential of this new interface with TensorFlow, we
provide a number of pre-trained state-of-the-art music tagging and
classification CNN models. We run an extensive evaluation of the developed
models. In particular, we assess the generalization capabilities in a
cross-collection evaluation utilizing both external tag datasets as well as
manual annotations tailored to the taxonomies of our models.
Related papers
- Guided Flows for Generative Modeling and Decision Making [55.42634941614435]
We show that Guided Flows significantly improves the sample quality in conditional image generation and zero-shot text synthesis-to-speech.
Notably, we are first to apply flow models for plan generation in the offline reinforcement learning setting ax speedup in compared to diffusion models.
arXiv Detail & Related papers (2023-11-22T15:07:59Z) - Multi-annotator Deep Learning: A Probabilistic Framework for
Classification [2.445702550853822]
Training standard deep neural networks leads to subpar performances in multi-annotator supervised learning settings.
We address this issue by presenting a probabilistic training framework named multi-annotator deep learning (MaDL)
A modular network architecture enables us to make varying assumptions regarding annotators' performances.
Our findings show MaDL's state-of-the-art performance and robustness against many correlated, spamming annotators.
arXiv Detail & Related papers (2023-04-05T16:00:42Z) - Studying How to Efficiently and Effectively Guide Models with Explanations [52.498055901649025]
'Model guidance' is the idea of regularizing the models' explanations to ensure that they are "right for the right reasons"
We conduct an in-depth evaluation across various loss functions, attribution methods, models, and 'guidance depths' on the PASCAL VOC 2007 and MS COCO 2014 datasets.
Specifically, we guide the models via bounding box annotations, which are much cheaper to obtain than the commonly used segmentation masks.
arXiv Detail & Related papers (2023-03-21T15:34:50Z) - Trieste: Efficiently Exploring The Depths of Black-box Functions with
TensorFlow [50.691232400959656]
Trieste is an open-source Python package for Bayesian optimization and active learning.
Our library enables the plug-and-play of popular models within sequential decision-making loops.
arXiv Detail & Related papers (2023-02-16T17:21:49Z) - Assemble Foundation Models for Automatic Code Summarization [9.53949558569201]
We propose a flexible and robust approach for automatic code summarization based on neural networks.
We assemble available foundation models, such as CodeBERT and GPT-2, into a single model named AdaMo.
We introduce two adaptive schemes from the perspective of knowledge transfer, namely continuous pretraining and intermediate finetuning.
arXiv Detail & Related papers (2022-01-13T21:38:33Z) - Towards Open-World Feature Extrapolation: An Inductive Graph Learning
Approach [80.8446673089281]
We propose a new learning paradigm with graph representation and learning.
Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data.
arXiv Detail & Related papers (2021-10-09T09:02:45Z) - Learning by Distillation: A Self-Supervised Learning Framework for
Optical Flow Estimation [71.76008290101214]
DistillFlow is a knowledge distillation approach to learning optical flow.
It achieves state-of-the-art unsupervised learning performance on both KITTI and Sintel datasets.
Our models ranked 1st among all monocular methods on the KITTI 2015 benchmark, and outperform all published methods on the Sintel Final benchmark.
arXiv Detail & Related papers (2021-06-08T09:13:34Z) - TensorFlow ManOpt: a library for optimization on Riemannian manifolds [0.3655021726150367]
The adoption of neural networks and deep learning in non-Euclidean domains has been hindered until recently by the lack of scalable and efficient learning frameworks.
We attempt to bridge this gap by proposing ManOpt, a Python library for optimization on Riemannian in terms of machine learning models.
The library is designed with the aim for a seamless integration with the ecosystem, targeting not only research, but also streamlining production machine learning pipelines.
arXiv Detail & Related papers (2021-05-27T10:42:09Z) - Implementing graph neural networks with TensorFlow-Keras [1.6114012813668934]
Graph neural networks are a versatile machine learning architecture that received a lot of attention recently.
In this technical report, we present an implementation of convolution and pooling layers for Keras-Keras models.
arXiv Detail & Related papers (2021-03-07T10:46:02Z) - Captum: A unified and generic model interpretability library for PyTorch [49.72749684393332]
We introduce a novel, unified, open-source model interpretability library for PyTorch.
The library contains generic implementations of a number of gradient and perturbation-based attribution algorithms.
It can be used for both classification and non-classification models.
arXiv Detail & Related papers (2020-09-16T18:57:57Z) - Graph Neural Networks in TensorFlow and Keras with Spektral [18.493394650508044]
Spektral is an open-source Python library for building graph neural networks.
It implements a large set of methods for deep learning on graphs, including message-passing and pooling operators.
It is suitable for absolute beginners and expert deep learning practitioners alike.
arXiv Detail & Related papers (2020-06-22T10:56:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.