Classification of integers based on residue classes via modern deep
learning algorithms
- URL: http://arxiv.org/abs/2304.01333v3
- Date: Fri, 8 Sep 2023 04:28:00 GMT
- Title: Classification of integers based on residue classes via modern deep
learning algorithms
- Authors: Da Wu, Jingye Yang, Mian Umair Ahsan, Kai Wang
- Abstract summary: We tested multiple deep learning architectures and feature engineering approaches on classifying integers based on their residues when divided by small prime numbers.
We also evaluated Automated Machine Learning platforms from Amazon, Google and Microsoft, and found that they failed on this task without appropriately engineered features.
In conclusion, feature engineering remains an important task to improve performance and increase interpretability of machine-learning models.
- Score: 3.6396223542930772
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Judging whether an integer can be divided by prime numbers such as 2 or 3 may
appear trivial to human beings, but can be less straightforward for computers.
Here, we tested multiple deep learning architectures and feature engineering
approaches on classifying integers based on their residues when divided by
small prime numbers. We found that the ability of classification critically
depends on the feature space. We also evaluated Automated Machine Learning
(AutoML) platforms from Amazon, Google and Microsoft, and found that they
failed on this task without appropriately engineered features. Furthermore, we
introduced a method that utilizes linear regression on Fourier series basis
vectors, and demonstrated its effectiveness. Finally, we evaluated Large
Language Models (LLMs) such as GPT-4, GPT-J, LLaMA and Falcon, and demonstrated
their failures. In conclusion, feature engineering remains an important task to
improve performance and increase interpretability of machine-learning models,
even in the era of AutoML and LLMs.
Related papers
- InternLM-Math: Open Math Large Language Models Toward Verifiable Reasoning [98.53491178426492]
We open-source our math reasoning LLMs InternLM-Math which is continue pre-trained from InternLM2.
We unify chain-of-thought reasoning, reward modeling, formal reasoning, data augmentation, and code interpreter in a unified seq2seq format.
Our pre-trained model achieves 30.3 on the MiniF2F test set without fine-tuning.
arXiv Detail & Related papers (2024-02-09T11:22:08Z) - In-Context Language Learning: Architectures and Algorithms [73.93205821154605]
We study ICL through the lens of a new family of model problems we term in context language learning (ICLL)
We evaluate a diverse set of neural sequence models on regular ICLL tasks.
arXiv Detail & Related papers (2024-01-23T18:59:21Z) - MISS: Multiclass Interpretable Scoring Systems [13.902264070785986]
We present a machine-learning approach for constructing Multiclass Interpretable Scoring Systems (MISS)
MISS is a fully data-driven methodology for single, sparse, and user-friendly scoring systems for multiclass classification problems.
Results indicate that our approach is competitive with other machine learning models in terms of classification performance metrics and provides well-calibrated class probabilities.
arXiv Detail & Related papers (2024-01-10T10:57:12Z) - Automated Sizing and Training of Efficient Deep Autoencoders using
Second Order Algorithms [0.46040036610482665]
We propose a multi-step training method for generalized linear classifiers.
validation error is minimized by pruning of unnecessary inputs.
desired outputs are improved via a method similar to the Ho-Kashyap rule.
arXiv Detail & Related papers (2023-08-11T16:48:31Z) - Language models are weak learners [71.33837923104808]
We show that prompt-based large language models can operate effectively as weak learners.
We incorporate these models into a boosting approach, which can leverage the knowledge within the model to outperform traditional tree-based boosting.
Results illustrate the potential for prompt-based LLMs to function not just as few-shot learners themselves, but as components of larger machine learning pipelines.
arXiv Detail & Related papers (2023-06-25T02:39:19Z) - Class-Incremental Learning: A Survey [84.30083092434938]
Class-Incremental Learning (CIL) enables the learner to incorporate the knowledge of new classes incrementally.
CIL tends to catastrophically forget the characteristics of former ones, and its performance drastically degrades.
We provide a rigorous and unified evaluation of 17 methods in benchmark image classification tasks to find out the characteristics of different algorithms.
arXiv Detail & Related papers (2023-02-07T17:59:05Z) - Winning solutions and post-challenge analyses of the ChaLearn AutoDL
challenge 2019 [112.36155380260655]
This paper reports the results and post-challenge analyses of ChaLearn's AutoDL challenge series.
Results show that DL methods dominated, though popular Neural Architecture Search (NAS) was impractical.
A high level modular organization emerged featuring a "meta-learner", "data ingestor", "model selector", "model/learner", and "evaluator"
arXiv Detail & Related papers (2022-01-11T06:21:18Z) - Fast Yet Effective Machine Unlearning [6.884272840652062]
We introduce a novel machine unlearning framework with error-maximizing noise generation and impair-repair based weight manipulation.
We show excellent unlearning while substantially retaining the overall model accuracy.
This work is an important step towards fast and easy implementation of unlearning in deep networks.
arXiv Detail & Related papers (2021-11-17T07:29:24Z) - LeanML: A Design Pattern To Slash Avoidable Wastes in Machine Learning
Projects [0.5330240017302619]
We introduce the first application of the lean methodology to machine learning projects.
We argue that lean machine learning (LeanML) can drastically slash avoidable wastes in commercial machine learning projects.
arXiv Detail & Related papers (2021-07-16T18:16:48Z) - Few-Shot Incremental Learning with Continually Evolved Classifiers [46.278573301326276]
Few-shot class-incremental learning (FSCIL) aims to design machine learning algorithms that can continually learn new concepts from a few data points.
The difficulty lies in that limited data from new classes not only lead to significant overfitting issues but also exacerbate the notorious catastrophic forgetting problems.
We propose a Continually Evolved CIF ( CEC) that employs a graph model to propagate context information between classifiers for adaptation.
arXiv Detail & Related papers (2021-04-07T10:54:51Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.