Linear chain conditional random fields, hidden Markov models, and
related classifiers
- URL: http://arxiv.org/abs/2301.01293v1
- Date: Tue, 3 Jan 2023 18:52:39 GMT
- Title: Linear chain conditional random fields, hidden Markov models, and
related classifiers
- Authors: Elie Azeraf, Emmanuel Monfrini, Wojciech Pieczynski
- Abstract summary: Conditional Random Fields (CRFs) are an alternative to Hidden Markov Models (HMMs)
We show that basic Linear-Chain CRFs (LC-CRFs) are in fact equivalent to them in the sense that for each LC-CRF there exists a HMM.
We show that it is possible to reformulate the generative Bayesian classifiers Maximum Posterior Mode (MPM) and Maximum a Posteriori (MAP) used in HMMs, as discriminative ones.
- Score: 4.984601297028258
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Practitioners use Hidden Markov Models (HMMs) in different problems for about
sixty years. Besides, Conditional Random Fields (CRFs) are an alternative to
HMMs and appear in the literature as different and somewhat concurrent models.
We propose two contributions. First, we show that basic Linear-Chain CRFs
(LC-CRFs), considered as different from the HMMs, are in fact equivalent to
them in the sense that for each LC-CRF there exists a HMM - that we specify -
whom posterior distribution is identical to the given LC-CRF. Second, we show
that it is possible to reformulate the generative Bayesian classifiers Maximum
Posterior Mode (MPM) and Maximum a Posteriori (MAP) used in HMMs, as
discriminative ones. The last point is of importance in many fields, especially
in Natural Language Processing (NLP), as it shows that in some situations
dropping HMMs in favor of CRFs was not necessary.
Related papers
- Adversarial Schrödinger Bridge Matching [66.39774923893103]
Iterative Markovian Fitting (IMF) procedure alternates between Markovian and reciprocal projections of continuous-time processes.
We propose a novel Discrete-time IMF (D-IMF) procedure in which learning of processes is replaced by learning just a few transition probabilities in discrete time.
We show that our D-IMF procedure can provide the same quality of unpaired domain translation as the IMF, using only several generation steps instead of hundreds.
arXiv Detail & Related papers (2024-05-23T11:29:33Z) - Learning Hidden Markov Models Using Conditional Samples [72.20944611510198]
This paper is concerned with the computational complexity of learning the Hidden Markov Model (HMM)
In this paper, we consider an interactive access model, in which the algorithm can query for samples from the conditional distributions of the HMMs.
Specifically, we obtain efficient algorithms for learning HMMs in settings where we have query access to the exact conditional probabilities.
arXiv Detail & Related papers (2023-02-28T16:53:41Z) - Fuzzy Cognitive Maps and Hidden Markov Models: Comparative Analysis of
Efficiency within the Confines of the Time Series Classification Task [0.0]
We explore the application of Hidden Markov Model (HMM) for time series classification.
We identify four models, HMM NN (HMM, one per series), HMM 1C (HMM, one per class), FCM NN, and FCM 1C are then studied in a series of experiments.
arXiv Detail & Related papers (2022-04-28T12:41:05Z) - Lite Unified Modeling for Discriminative Reading Comprehension [68.39862736200045]
We propose a lightweight POS-Enhanced Iterative Co-Attention Network (POI-Net) to handle diverse discriminative MRC tasks synchronously.
Our lite unified design brings model significant improvement with both encoder and decoder components.
The evaluation results on four discriminative MRC benchmarks consistently indicate the general effectiveness and applicability of our model.
arXiv Detail & Related papers (2022-03-26T15:47:19Z) - Learning Hidden Markov Models When the Locations of Missing Observations
are Unknown [54.40592050737724]
We consider the general problem of learning an HMM from data with unknown missing observation locations.
We provide reconstruction algorithms that do not require any assumptions about the structure of the underlying chain.
We show that under proper specifications one can reconstruct the process dynamics as well as if the missing observations positions were known.
arXiv Detail & Related papers (2022-03-12T22:40:43Z) - On equivalence between linear-chain conditional random fields and hidden
Markov chains [6.939768185086753]
Authors usually consider conditional random fields (CRFs) as quite different from generative models.
In some areas, like natural language processing (NLP), discriminative models have completely supplanted generative models.
We show that HMCs and linear-chain CRFs are not different but just differently parametrized models.
arXiv Detail & Related papers (2021-11-14T15:53:47Z) - Learning Circular Hidden Quantum Markov Models: A Tensor Network
Approach [34.77250498401055]
We show that c-HQMMs are equivalent to a constrained tensor network.
This equivalence enables us to provide an efficient learning model for c-HQMMs.
The proposed learning approach is evaluated on six real datasets.
arXiv Detail & Related papers (2021-10-29T23:09:31Z) - BERTifying the Hidden Markov Model for Multi-Source Weakly Supervised
Named Entity Recognition [57.2201011783393]
conditional hidden Markov model (CHMM)
CHMM predicts token-wise transition and emission probabilities from the BERT embeddings of the input tokens.
It fine-tunes a BERT-based NER model with the labels inferred by CHMM.
arXiv Detail & Related papers (2021-05-26T21:18:48Z) - Graphical Modeling for Multi-Source Domain Adaptation [56.05348879528149]
Multi-Source Domain Adaptation (MSDA) focuses on transferring the knowledge from multiple source domains to the target domain.
We propose two types of graphical models,i.e. Conditional Random Field for MSDA (CRF-MSDA) and Markov Random Field for MSDA (MRF-MSDA)
We evaluate these two models on four standard benchmark data sets of MSDA with distinct domain shift and data complexity.
arXiv Detail & Related papers (2021-04-27T09:04:22Z) - Malware Classification with GMM-HMM Models [8.02151721194722]
In this paper, we use GMM-HMMs for malware classification and we compare our results to those obtained using discrete HMMs.
For our opcode features, GMM-HMMs produce results that are comparable to those obtained using discrete HMMs.
arXiv Detail & Related papers (2021-03-03T23:23:48Z) - Robust Classification using Hidden Markov Models and Mixtures of
Normalizing Flows [25.543231171094384]
We use a generative model that combines the state transitions of a hidden Markov model (HMM) and the neural network based probability distributions for the hidden states of the HMM.
We verify the improved robustness of NMM-HMM classifiers in an application to speech recognition.
arXiv Detail & Related papers (2021-02-15T00:40:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.