Engineering the Neural Automatic Passenger Counter
- URL: http://arxiv.org/abs/2203.01156v1
- Date: Wed, 2 Mar 2022 14:56:11 GMT
- Title: Engineering the Neural Automatic Passenger Counter
- Authors: Nico Jahn, Michael Siebert
- Abstract summary: We explore and exploit various aspects of machine learning to increase reliability, performance, and counting quality.
We show how aggregation techniques such as ensemble quantiles can reduce bias, and we give an idea of the overall spread of the results.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automatic passenger counting (APC) in public transportation has been
approached with various machine learning and artificial intelligence methods
since its introduction in the 1970s. While equivalence testing is becoming more
popular than difference detection (Student's t-test), the former is much more
difficult to pass to ensure low user risk. On the other hand, recent
developments in artificial intelligence have led to algorithms that promise
much higher counting quality (lower bias). However, gradient-based methods
(including Deep Learning) have one limitation: they typically run into local
optima. In this work, we explore and exploit various aspects of machine
learning to increase reliability, performance, and counting quality. We perform
a grid search with several fundamental parameters: the selection and size of
the training set, which is similar to cross-validation, and the initial network
weights and randomness during the training process. Using this experiment, we
show how aggregation techniques such as ensemble quantiles can reduce bias, and
we give an idea of the overall spread of the results. We utilize the test
success chance, a simulative metric based on the empirical distribution. We
also employ a post-training Monte Carlo quantization approach and introduce
cumulative summation to turn counting into a stationary method and allow
unbounded counts.
Related papers
- Adapting Vision-Language Models to Open Classes via Test-Time Prompt Tuning [50.26965628047682]
Adapting pre-trained models to open classes is a challenging problem in machine learning.
In this paper, we consider combining the advantages of both and come up with a test-time prompt tuning approach.
Our proposed method outperforms all comparison methods on average considering both base and new classes.
arXiv Detail & Related papers (2024-08-29T12:34:01Z) - Just One Byte (per gradient): A Note on Low-Bandwidth Decentralized
Language Model Finetuning Using Shared Randomness [86.61582747039053]
Language model training in distributed settings is limited by the communication cost of exchanges.
We extend recent work using shared randomness to perform distributed fine-tuning with low bandwidth.
arXiv Detail & Related papers (2023-06-16T17:59:51Z) - A Maximum Log-Likelihood Method for Imbalanced Few-Shot Learning Tasks [3.2895195535353308]
We propose a new maximum log-likelihood metric for few-shot architectures.
We demonstrate that the proposed metric achieves superior performance accuracy w.r.t. conventional similarity metrics.
We also show that our algorithm achieves state-of-the-art transductive few-shot performance when the evaluation data is imbalanced.
arXiv Detail & Related papers (2022-11-26T21:31:00Z) - Heterogeneous Ensemble Learning for Enhanced Crash Forecasts -- A
Frequentest and Machine Learning based Stacking Framework [0.803552105641624]
In this study, we apply one of the key HEM methods, Stacking, to model crash frequency on five lane undivided segments (5T) of urban and suburban arterials.
The prediction performance of Stacking is compared with parametric statistical models (Poisson and negative binomial) and three state of the art machine learning techniques (Decision tree, random forest, and gradient boosting)
arXiv Detail & Related papers (2022-07-21T19:15:53Z) - MQBench: Towards Reproducible and Deployable Model Quantization
Benchmark [53.12623958951738]
MQBench is a first attempt to evaluate, analyze, and benchmark the and deployability for model quantization algorithms.
We choose multiple platforms for real-world deployments, including CPU, GPU, ASIC, DSP, and evaluate extensive state-of-the-art quantization algorithms.
We conduct a comprehensive analysis and find considerable intuitive or counter-intuitive insights.
arXiv Detail & Related papers (2021-11-05T23:38:44Z) - Squeezing Backbone Feature Distributions to the Max for Efficient
Few-Shot Learning [3.1153758106426603]
Few-shot classification is a challenging problem due to the uncertainty caused by using few labelled samples.
We propose a novel transfer-based method which aims at processing the feature vectors so that they become closer to Gaussian-like distributions.
In the case of transductive few-shot learning where unlabelled test samples are available during training, we also introduce an optimal-transport inspired algorithm to boost even further the achieved performance.
arXiv Detail & Related papers (2021-10-18T16:29:17Z) - MURAL: Meta-Learning Uncertainty-Aware Rewards for Outcome-Driven
Reinforcement Learning [65.52675802289775]
We show that an uncertainty aware classifier can solve challenging reinforcement learning problems.
We propose a novel method for computing the normalized maximum likelihood (NML) distribution.
We show that the resulting algorithm has a number of intriguing connections to both count-based exploration methods and prior algorithms for learning reward functions.
arXiv Detail & Related papers (2021-07-15T08:19:57Z) - Text Generation with Efficient (Soft) Q-Learning [91.47743595382758]
Reinforcement learning (RL) offers a more flexible solution by allowing users to plug in arbitrary task metrics as reward.
We introduce a new RL formulation for text generation from the soft Q-learning perspective.
We apply the approach to a wide range of tasks, including learning from noisy/negative examples, adversarial attacks, and prompt generation.
arXiv Detail & Related papers (2021-06-14T18:48:40Z) - Fast Uncertainty Quantification for Deep Object Pose Estimation [91.09217713805337]
Deep learning-based object pose estimators are often unreliable and overconfident.
In this work, we propose a simple, efficient, and plug-and-play UQ method for 6-DoF object pose estimation.
arXiv Detail & Related papers (2020-11-16T06:51:55Z) - First-order Optimization for Superquantile-based Supervised Learning [0.0]
We propose a first-order optimization algorithm to minimize a superquantile-based learning objective.
The proposed algorithm is based on smoothing the superquantile function by infimal convolution.
arXiv Detail & Related papers (2020-09-30T11:43:45Z) - The Integrity of Machine Learning Algorithms against Software Defect
Prediction [0.0]
This report analyses the performance of the Online Sequential Extreme Learning Machine (OS-ELM) proposed by Liang et.al.
OS-ELM trains faster than conventional deep neural networks and it always converges to the globally optimal solution.
The analysis is carried out on 3 projects KC1, PC4 and PC3 carried out by the NASA group.
arXiv Detail & Related papers (2020-09-05T17:26:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.