A novel policy for pre-trained Deep Reinforcement Learning for Speech
Emotion Recognition
- URL: http://arxiv.org/abs/2101.00738v2
- Date: Sun, 31 Jan 2021 10:06:52 GMT
- Title: A novel policy for pre-trained Deep Reinforcement Learning for Speech
Emotion Recognition
- Authors: Thejan Rajapakshe, Rajib Rana, Sara Khalifa, Bj\"orn W. Schuller,
Jiajun Liu
- Abstract summary: Reinforcement Learning (RL) is a semi-supervised learning paradigm which an agent learns by interacting with an environment.
Deep RL has gained tremendous success in gaming - such as AlphaGo, but its potential have rarely been explored for challenging tasks like Speech Emotion Recognition (SER)
In this paper, we introduce a novel policy - "Zeta policy" which is tailored for SER and apply Pre-training in deep RL to achieve faster learning rate.
- Score: 8.175197257598697
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement Learning (RL) is a semi-supervised learning paradigm which an
agent learns by interacting with an environment. Deep learning in combination
with RL provides an efficient method to learn how to interact with the
environment is called Deep Reinforcement Learning (deep RL). Deep RL has gained
tremendous success in gaming - such as AlphaGo, but its potential have rarely
being explored for challenging tasks like Speech Emotion Recognition (SER). The
deep RL being used for SER can potentially improve the performance of an
automated call centre agent by dynamically learning emotional-aware response to
customer queries. While the policy employed by the RL agent plays a major role
in action selection, there is no current RL policy tailored for SER. In
addition, extended learning period is a general challenge for deep RL which can
impact the speed of learning for SER. Therefore, in this paper, we introduce a
novel policy - "Zeta policy" which is tailored for SER and apply Pre-training
in deep RL to achieve faster learning rate. Pre-training with cross dataset was
also studied to discover the feasibility of pre-training the RL Agent with a
similar dataset in a scenario of where no real environmental data is not
available. IEMOCAP and SAVEE datasets were used for the evaluation with the
problem being to recognize four emotions happy, sad, angry and neutral in the
utterances provided. Experimental results show that the proposed "Zeta policy"
performs better than existing policies. The results also support that
pre-training can reduce the training time upon reducing the warm-up period and
is robust to cross-corpus scenario.
Related papers
- SHIRE: Enhancing Sample Efficiency using Human Intuition in REinforcement Learning [11.304750795377657]
We propose SHIRE, a framework for encoding human intuition using Probabilistic Graphical Models (PGMs)
SHIRE achieves 25-78% sample efficiency gains across the environments we evaluate at negligible overhead cost.
arXiv Detail & Related papers (2024-09-16T04:46:22Z) - Using Offline Data to Speed-up Reinforcement Learning in Procedurally
Generated Environments [11.272582555795989]
We study whether agents can leverage offline data in the form of trajectories to improve the sample-efficiency in procedurally generated environments.
We consider two settings of using IL from offline data for RL: (1) pre-training a policy before online RL training and (2) concurrently training a policy with online RL and IL from offline data.
arXiv Detail & Related papers (2023-04-18T16:23:15Z) - Flexible Attention-Based Multi-Policy Fusion for Efficient Deep
Reinforcement Learning [78.31888150539258]
Reinforcement learning (RL) agents have long sought to approach the efficiency of human learning.
Prior studies in RL have incorporated external knowledge policies to help agents improve sample efficiency.
We present Knowledge-Grounded RL (KGRL), an RL paradigm fusing multiple knowledge policies and aiming for human-like efficiency and flexibility.
arXiv Detail & Related papers (2022-10-07T17:56:57Z) - Jump-Start Reinforcement Learning [68.82380421479675]
We present a meta algorithm that can use offline data, demonstrations, or a pre-existing policy to initialize an RL policy.
In particular, we propose Jump-Start Reinforcement Learning (JSRL), an algorithm that employs two policies to solve tasks.
We show via experiments that JSRL is able to significantly outperform existing imitation and reinforcement learning algorithms.
arXiv Detail & Related papers (2022-04-05T17:25:22Z) - Accelerating Robotic Reinforcement Learning via Parameterized Action
Primitives [92.0321404272942]
Reinforcement learning can be used to build general-purpose robotic systems.
However, training RL agents to solve robotics tasks still remains challenging.
In this work, we manually specify a library of robot action primitives (RAPS), parameterized with arguments that are learned by an RL policy.
We find that our simple change to the action interface substantially improves both the learning efficiency and task performance.
arXiv Detail & Related papers (2021-10-28T17:59:30Z) - Explore and Control with Adversarial Surprise [78.41972292110967]
Reinforcement learning (RL) provides a framework for learning goal-directed policies given user-specified rewards.
We propose a new unsupervised RL technique based on an adversarial game which pits two policies against each other to compete over the amount of surprise an RL agent experiences.
We show that our method leads to the emergence of complex skills by exhibiting clear phase transitions.
arXiv Detail & Related papers (2021-07-12T17:58:40Z) - AWAC: Accelerating Online Reinforcement Learning with Offline Datasets [84.94748183816547]
We show that our method, advantage weighted actor critic (AWAC), enables rapid learning of skills with a combination of prior demonstration data and online experience.
Our results show that incorporating prior data can reduce the time required to learn a range of robotic skills to practical time-scales.
arXiv Detail & Related papers (2020-06-16T17:54:41Z) - Balancing Reinforcement Learning Training Experiences in Interactive
Information Retrieval [19.723551683930776]
Interactive Information Retrieval (IIR) and Reinforcement Learning (RL) share many commonalities, including an agent who learns while interacting.
To successfully apply RL methods to IIR, one challenge is to obtain sufficient relevance labels to train the RL agents.
Our paper addresses this issue by using domain randomization to synthesize more relevant documents for the training.
arXiv Detail & Related papers (2020-06-05T00:38:39Z) - Meta-Reinforcement Learning for Robotic Industrial Insertion Tasks [70.56451186797436]
We study how to use meta-reinforcement learning to solve the bulk of the problem in simulation.
We demonstrate our approach by training an agent to successfully perform challenging real-world insertion tasks.
arXiv Detail & Related papers (2020-04-29T18:00:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.