STREAMLINE: Streaming Active Learning for Realistic Multi-Distributional
Settings
- URL: http://arxiv.org/abs/2305.10643v1
- Date: Thu, 18 May 2023 02:01:45 GMT
- Title: STREAMLINE: Streaming Active Learning for Realistic Multi-Distributional
Settings
- Authors: Nathan Beck, Suraj Kothawade, Pradeep Shenoy, Rishabh Iyer
- Abstract summary: STREAMLINE is a novel streaming active learning framework that mitigates scenario-driven slice imbalance in working labeled data.
We evaluate STREAMLINE on real-world streaming scenarios for image classification and object detection tasks.
- Score: 2.580765958706854
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks have consistently shown great performance in several
real-world use cases like autonomous vehicles, satellite imaging, etc.,
effectively leveraging large corpora of labeled training data. However,
learning unbiased models depends on building a dataset that is representative
of a diverse range of realistic scenarios for a given task. This is challenging
in many settings where data comes from high-volume streams, with each scenario
occurring in random interleaved episodes at varying frequencies. We study
realistic streaming settings where data instances arrive in and are sampled
from an episodic multi-distributional data stream. Using submodular information
measures, we propose STREAMLINE, a novel streaming active learning framework
that mitigates scenario-driven slice imbalance in the working labeled data via
a three-step procedure of slice identification, slice-aware budgeting, and data
selection. We extensively evaluate STREAMLINE on real-world streaming scenarios
for image classification and object detection tasks. We observe that STREAMLINE
improves the performance on infrequent yet critical slices of the data over
current baselines by up to $5\%$ in terms of accuracy on our image
classification tasks and by up to $8\%$ in terms of mAP on our object detection
tasks.
Related papers
- Adapt-$\infty$: Scalable Lifelong Multimodal Instruction Tuning via Dynamic Data Selection [89.42023974249122]
Adapt-$infty$ is a new multi-way and adaptive data selection approach for Lifelong Instruction Tuning.
We construct pseudo-skill clusters by grouping gradient-based sample vectors.
We select the best-performing data selector for each skill cluster from a pool of selector experts.
arXiv Detail & Related papers (2024-10-14T15:48:09Z) - LVLane: Deep Learning for Lane Detection and Classification in
Challenging Conditions [2.5641096293146712]
We present an end-to-end lane detection and classification system based on deep learning methodologies.
In our study, we introduce a unique dataset meticulously curated to encompass scenarios that pose significant challenges for state-of-the-art (SOTA) lane localization models.
We propose a CNN-based classification branch, seamlessly integrated with the detector, facilitating the identification of distinct lane types.
arXiv Detail & Related papers (2023-07-13T16:09:53Z) - ALP: Action-Aware Embodied Learning for Perception [60.64801970249279]
We introduce Action-Aware Embodied Learning for Perception (ALP)
ALP incorporates action information into representation learning through a combination of optimizing a reinforcement learning policy and an inverse dynamics prediction objective.
We show that ALP outperforms existing baselines in several downstream perception tasks.
arXiv Detail & Related papers (2023-06-16T21:51:04Z) - Mitigating Representation Bias in Action Recognition: Algorithms and
Benchmarks [76.35271072704384]
Deep learning models perform poorly when applied to videos with rare scenes or objects.
We tackle this problem from two different angles: algorithm and dataset.
We show that the debiased representation can generalize better when transferred to other datasets and tasks.
arXiv Detail & Related papers (2022-09-20T00:30:35Z) - Basket-based Softmax [12.744577044692276]
We propose a novel mining-during-training strategy called Basket-based Softmax (BBS)
For each training sample, we simultaneously adopt similarity scores as the clue to mining negative classes from other datasets.
We demonstrate the efficiency and superiority of the BBS on the tasks of face recognition and re-identification, with both simulated and real-world datasets.
arXiv Detail & Related papers (2022-01-23T16:43:29Z) - Self-supervised Audiovisual Representation Learning for Remote Sensing Data [96.23611272637943]
We propose a self-supervised approach for pre-training deep neural networks in remote sensing.
By exploiting the correspondence between geo-tagged audio recordings and remote sensing, this is done in a completely label-free manner.
We show that our approach outperforms existing pre-training strategies for remote sensing imagery.
arXiv Detail & Related papers (2021-08-02T07:50:50Z) - Automatic Curation of Large-Scale Datasets for Audio-Visual
Representation Learning [62.47593143542552]
We describe a subset optimization approach for automatic dataset curation.
We demonstrate that our approach finds videos with high audio-visual correspondence and show that self-supervised models trained on our data, despite being automatically constructed, achieve similar downstream performances to existing video datasets with similar scales.
arXiv Detail & Related papers (2021-01-26T14:27:47Z) - Multimodal Prototypical Networks for Few-shot Learning [20.100480009813953]
Cross-modal feature generation framework is used to enrich the low populated embedding space in few-shot scenarios.
We show that in such cases nearest neighbor classification is a viable approach and outperform state-of-the-art single-modal and multimodal few-shot learning methods.
arXiv Detail & Related papers (2020-11-17T19:32:59Z) - Deep Multimodal Transfer-Learned Regression in Data-Poor Domains [0.0]
We propose a Deep Multimodal Transfer-Learned Regressor (DMTL-R) for multimodal learning of image and feature data.
Our model is capable of fine-tuning a given set of pre-trained CNN weights on a small amount of training image data.
We present results using phase-field simulation microstructure images with an accompanying set of physical features, using pre-trained weights from various well-known CNN architectures.
arXiv Detail & Related papers (2020-06-16T16:52:44Z) - On the performance of deep learning models for time series
classification in streaming [0.0]
This work is to assess the performance of different types of deep architectures for data streaming classification.
We evaluate models such as multi-layer perceptrons, recurrent, convolutional and temporal convolutional neural networks over several time-series datasets.
arXiv Detail & Related papers (2020-03-05T11:41:29Z) - A Multi-Channel Neural Graphical Event Model with Negative Evidence [76.51278722190607]
Event datasets are sequences of events of various types occurring irregularly over the time-line.
We propose a non-parametric deep neural network approach in order to estimate the underlying intensity functions.
arXiv Detail & Related papers (2020-02-21T23:10:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.