A Comparison Between Tsetlin Machines and Deep Neural Networks in the
Context of Recommendation Systems
- URL: http://arxiv.org/abs/2212.10136v1
- Date: Tue, 20 Dec 2022 10:05:36 GMT
- Title: A Comparison Between Tsetlin Machines and Deep Neural Networks in the
Context of Recommendation Systems
- Authors: Karl Audun Borgersen, Morten Goodwin, Jivitesh Sharma
- Abstract summary: Recommendation Systems (RSs) are ubiquitous in modern society and are one of the largest points of interaction between humans and AI.
Modern RSs are often implemented using deep learning models, which are infamously difficult to interpret.
The newly introduced Tsetlin Machines (TM) possess some valuable properties due to their inherent interpretability.
- Score: 4.662321040754879
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recommendation Systems (RSs) are ubiquitous in modern society and are one of
the largest points of interaction between humans and AI. Modern RSs are often
implemented using deep learning models, which are infamously difficult to
interpret. This problem is particularly exasperated in the context of
recommendation scenarios, as it erodes the user's trust in the RS. In contrast,
the newly introduced Tsetlin Machines (TM) possess some valuable properties due
to their inherent interpretability. TMs are still fairly young as a technology.
As no RS has been developed for TMs before, it has become necessary to perform
some preliminary research regarding the practicality of such a system. In this
paper, we develop the first RS based on TMs to evaluate its practicality in
this application domain. This paper compares the viability of TMs with other
machine learning models prevalent in the field of RS. We train and investigate
the performance of the TM compared with a vanilla feed-forward deep learning
model. These comparisons are based on model performance,
interpretability/explainability, and scalability. Further, we provide some
benchmark performance comparisons to similar machine learning solutions
relevant to RSs.
Related papers
- SFR-RAG: Towards Contextually Faithful LLMs [57.666165819196486]
Retrieval Augmented Generation (RAG) is a paradigm that integrates external contextual information with large language models (LLMs) to enhance factual accuracy and relevance.
We introduce SFR-RAG, a small LLM that is instruction-textual with an emphasis on context-grounded generation and hallucination.
We also present ConBench, a new evaluation framework compiling multiple popular and diverse RAG benchmarks.
arXiv Detail & Related papers (2024-09-16T01:08:18Z) - Benchmarks as Microscopes: A Call for Model Metrology [76.64402390208576]
Modern language models (LMs) pose a new challenge in capability assessment.
To be confident in our metrics, we need a new discipline of model metrology.
arXiv Detail & Related papers (2024-07-22T17:52:12Z) - Understanding the differences in Foundation Models: Attention, State Space Models, and Recurrent Neural Networks [50.29356570858905]
We introduce the Dynamical Systems Framework (DSF), which allows a principled investigation of all these architectures in a common representation.
We provide principled comparisons between softmax attention and other model classes, discussing the theoretical conditions under which softmax attention can be approximated.
This shows the DSF's potential to guide the systematic development of future more efficient and scalable foundation models.
arXiv Detail & Related papers (2024-05-24T17:19:57Z) - Investigating the Robustness of Counterfactual Learning to Rank Models: A Reproducibility Study [61.64685376882383]
Counterfactual learning to rank (CLTR) has attracted extensive attention in the IR community for its ability to leverage massive logged user interaction data to train ranking models.
This paper investigates the robustness of existing CLTR models in complex and diverse situations.
We find that the DLA models and IPS-DCM show better robustness under various simulation settings than IPS-PBM and PRS with offline propensity estimation.
arXiv Detail & Related papers (2024-04-04T10:54:38Z) - Reducing Computational Costs in Sentiment Analysis: Tensorized Recurrent
Networks vs. Recurrent Networks [0.12891210250935145]
Anticipating audience reaction towards a certain text is integral to several facets of society ranging from politics, research, and commercial industries.
Sentiment analysis (SA) is a useful natural language processing (NLP) technique that utilizes lexical/statistical and deep learning methods to determine whether different-sized texts exhibit positive, negative, or neutral emotions.
arXiv Detail & Related papers (2023-06-16T09:18:08Z) - Revisiting Class-Incremental Learning with Pre-Trained Models: Generalizability and Adaptivity are All You Need [84.3507610522086]
Class-incremental learning (CIL) aims to adapt to emerging new classes without forgetting old ones.
Recent pre-training has achieved substantial progress, making vast pre-trained models (PTMs) accessible for CIL.
We argue that the core factors in CIL are adaptivity for model updating and generalizability for knowledge transferring.
arXiv Detail & Related papers (2023-03-13T17:59:02Z) - Machine Learning Capability: A standardized metric using case difficulty
with applications to individualized deployment of supervised machine learning [2.2060666847121864]
Model evaluation is a critical component in supervised machine learning classification analyses.
Items Response Theory (IRT) and Computer Adaptive Testing (CAT) with machine learning can benchmark datasets independent of the end-classification results.
arXiv Detail & Related papers (2023-02-09T00:38:42Z) - Incremental Online Learning Algorithms Comparison for Gesture and Visual
Smart Sensors [68.8204255655161]
This paper compares four state-of-the-art algorithms in two real applications: gesture recognition based on accelerometer data and image classification.
Our results confirm these systems' reliability and the feasibility of deploying them in tiny-memory MCUs.
arXiv Detail & Related papers (2022-09-01T17:05:20Z) - Recurrent Inference Machines as inverse problem solvers for MR
relaxometry [0.0]
In this paper, we propose the use of Recurrent Inference Machines (RIMs) to perform T1 and T2 mapping.
RIMs are a neural network framework that learns an iterative inference process based on the signal model.
Inference with the RIM is 150 times faster than the Maximum Likelihood Estimator (MLE)
arXiv Detail & Related papers (2021-06-08T16:50:49Z) - Towards Robust Neural Retrieval Models with Synthetic Pre-Training [28.547347789198096]
We show that synthetic training examples generated using a sequence-to-sequence generator can be effective towards this goal.
In our experiments, pre-training with synthetic examples improves retrieval performance in both in-domain and out-of-domain evaluation on five different test sets.
arXiv Detail & Related papers (2021-04-15T22:12:01Z) - Achieving Online Regression Performance of LSTMs with Simple RNNs [0.0]
We introduce a first-order training algorithm with a linear time complexity in the number of parameters.
We show that when SRNNs are trained with our algorithm, they provide very similar regression performance with the LSTMs in two to three times shorter training time.
arXiv Detail & Related papers (2020-05-16T11:41:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.