Machine learning applications to DNA subsequence and restriction site
analysis
- URL: http://arxiv.org/abs/2011.03544v5
- Date: Fri, 11 Dec 2020 16:03:26 GMT
- Title: Machine learning applications to DNA subsequence and restriction site
analysis
- Authors: Ethan J. Moyer (1) and Anup Das (PhD) (2) ((1) School of Biomedical
Engineering, Science and Health Systems, Drexel University, Philadelphia,
Pennsylvania, USA, (2) College of Engineering, Drexel University,
Philadelphia, Pennsylvania, USA)
- Abstract summary: restriction synthesis is a novel iterative DNA synthesis method that utilizes endonucleases to synthesize a query sequence from a reference sequence.
In this work, the reference sequence is built from shorter subsequences by classifying them as applicable or inapplicable for the synthesis method using three different machine learning methods.
The sensitivity using SVMs, random forest, and CNNs are 94.9%, 92.7%, 91.4%, respectively.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Based on the BioBricks standard, restriction synthesis is a novel catabolic
iterative DNA synthesis method that utilizes endonucleases to synthesize a
query sequence from a reference sequence. In this work, the reference sequence
is built from shorter subsequences by classifying them as applicable or
inapplicable for the synthesis method using three different machine learning
methods: Support Vector Machines (SVMs), random forest, and Convolution Neural
Networks (CNNs). Before applying these methods to the data, a series of feature
selection, curation, and reduction steps are applied to create an accurate and
representative feature space. Following these preprocessing steps, three
different pipelines are proposed to classify subsequences based on their
nucleotide sequence and other relevant features corresponding to the
restriction sites of over 200 endonucleases. The sensitivity using SVMs, random
forest, and CNNs are 94.9%, 92.7%, 91.4%, respectively. Moreover, each method
scores lower in specificity with SVMs, random forest, and CNNs resulting in
77.4%, 85.7%, and 82.4%, respectively. In addition to analyzing these results,
the misclassifications in SVMs and CNNs are investigated. Across these two
models, different features with a derived nucleotide specificity visually
contribute more to classification compared to other features. This observation
is an important factor when considering new nucleotide sensitivity features for
future studies.
Related papers
- Protein sequence classification using natural language processing techniques [3.0846824529023396]
This study employs natural language processing (NLP) techniques on a dataset comprising 75 target protein classes.
We explored various machine learning and deep learning models, including K-Nearest Neighbors (KNN), Multinomial Na"ive Bayes, Logistic Regression, Multi-Layer Perceptron (MLP), Decision Tree, Random Forest, XGBoost, Voting and Stacking classifiers, Convolutional Neural Network (CNN), Long Short-Term Memory (LSTM), and transformer models.
arXiv Detail & Related papers (2024-09-06T13:16:16Z) - DNA Sequence Classification with Compressors [0.0]
Our study introduces a novel adaptation of Jiang et al.'s compressor-based, parameter-free classification method, specifically tailored for DNA sequence analysis.
Not only does this method align with the current state-of-the-art in terms of accuracy, but it also offers a more resource-efficient alternative to traditional machine learning methods.
arXiv Detail & Related papers (2024-01-25T09:17:19Z) - A novel RNA pseudouridine site prediction model using Utility Kernel and
data-driven parameters [0.7373617024876725]
Pseudouridine is the most frequent modification in RNA.
Existing models to predict the pseudouridine sites in a given RNA sequence mainly depend on user-defined features.
We propose a Support Vector Machine (SVM) Kernel based on utility theory from Economics.
arXiv Detail & Related papers (2023-11-02T08:32:10Z) - Multilayer Multiset Neuronal Networks -- MMNNs [55.2480439325792]
The present work describes multilayer multiset neuronal networks incorporating two or more layers of coincidence similarity neurons.
The work also explores the utilization of counter-prototype points, which are assigned to the image regions to be avoided.
arXiv Detail & Related papers (2023-08-28T12:55:13Z) - HyenaDNA: Long-Range Genomic Sequence Modeling at Single Nucleotide
Resolution [76.97231739317259]
We present HyenaDNA, a genomic foundation model pretrained on the human reference genome with context lengths of up to 1 million tokens at the single nucleotide-level.
On fine-tuned benchmarks from the Nucleotide Transformer, HyenaDNA reaches state-of-the-art (SotA) on 12 of 18 datasets using a model with orders of magnitude less parameters and pretraining data.
arXiv Detail & Related papers (2023-06-27T20:46:34Z) - Mutual Exclusivity Training and Primitive Augmentation to Induce
Compositionality [84.94877848357896]
Recent datasets expose the lack of the systematic generalization ability in standard sequence-to-sequence models.
We analyze this behavior of seq2seq models and identify two contributing factors: a lack of mutual exclusivity bias and the tendency to memorize whole examples.
We show substantial empirical improvements using standard sequence-to-sequence models on two widely-used compositionality datasets.
arXiv Detail & Related papers (2022-11-28T17:36:41Z) - Using Signal Processing in Tandem With Adapted Mixture Models for
Classifying Genomic Signals [16.119729980200955]
We propose a novel technique that employs signal processing in tandem with Gaussian mixture models to improve the spectral representation of a sequence.
Our method outperforms a similar state-of-the-art method on established benchmark datasets by an absolute margin of 6.06% accuracy.
arXiv Detail & Related papers (2022-11-03T06:10:55Z) - Diversifying Design of Nucleic Acid Aptamers Using Unsupervised Machine
Learning [54.247560894146105]
Inverse design of short single-stranded RNA and DNA sequences (aptamers) is the task of finding sequences that satisfy a set of desired criteria.
We propose to use an unsupervised machine learning model known as the Potts model to discover new, useful sequences with controllable sequence diversity.
arXiv Detail & Related papers (2022-08-10T13:30:58Z) - Comparing Machine Learning Algorithms with or without Feature Extraction
for DNA Classification [0.7742297876120561]
Three state-of-the-art algorithms, namely Convolutional Neural Networks, Deep Neural Networks, and N-gram Probabilistic Models, are used for the task of DNA classification.
We introduce a novel feature extraction method based on the Levenshtein distance and randomly generated DNA sub-sequences.
Four different data sets, each concerning viral diseases such as Covid-19, AIDS, Influenza, and Hepatitis C, are used for evaluating the different approaches.
arXiv Detail & Related papers (2020-11-01T12:04:54Z) - Provably Efficient Neural Estimation of Structural Equation Model: An
Adversarial Approach [144.21892195917758]
We study estimation in a class of generalized Structural equation models (SEMs)
We formulate the linear operator equation as a min-max game, where both players are parameterized by neural networks (NNs), and learn the parameters of these neural networks using a gradient descent.
For the first time we provide a tractable estimation procedure for SEMs based on NNs with provable convergence and without the need for sample splitting.
arXiv Detail & Related papers (2020-07-02T17:55:47Z) - A Systematic Approach to Featurization for Cancer Drug Sensitivity
Predictions with Deep Learning [49.86828302591469]
We train >35,000 neural network models, sweeping over common featurization techniques.
We found the RNA-seq to be highly redundant and informative even with subsets larger than 128 features.
arXiv Detail & Related papers (2020-04-30T20:42:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.