On the Derivational Entropy of Left-to-Right Probabilistic Finite-State Automata and Hidden Markov Models

Probabilistic finite-state automata are a formalism that is widely used in many problems of automatic speech recognition and natural language processing. Probabilistic finite-state automata are closely related to other finite-state models as weighted finite-state automata, word lattices, and hidden...

Full description

Bibliographic Details
Main Authors: Joan Andreu Sánchez, Martha Alicia Rocha, Verónica Romero, Mauricio Villegas
Format: Article
Language:English
Published: The MIT Press 2018-03-01
Series:Computational Linguistics
Online Access:https://www.mitpressjournals.org/doi/pdf/10.1162/COLI_a_00306
id doaj-61a52c65365d49148d9c889201dbf9c5
record_format Article
spelling doaj-61a52c65365d49148d9c889201dbf9c52020-11-24T21:21:50ZengThe MIT PressComputational Linguistics1530-93122018-03-01441173710.1162/COLI_a_00306COLI_a_00306On the Derivational Entropy of Left-to-Right Probabilistic Finite-State Automata and Hidden Markov ModelsJoan Andreu Sánchez0Martha Alicia Rocha1Verónica Romero2Mauricio Villegas3Universitat Politècnica de ValènciaInstituto Tecnológico de LeónUniversitat Politècnica de ValènciaSearchInkProbabilistic finite-state automata are a formalism that is widely used in many problems of automatic speech recognition and natural language processing. Probabilistic finite-state automata are closely related to other finite-state models as weighted finite-state automata, word lattices, and hidden Markov models. Therefore, they share many similar properties and problems. Entropy measures of finite-state models have been investigated in the past in order to study the information capacity of these models. The derivational entropy quantifies the uncertainty that the model has about the probability distribution it represents. The derivational entropy in a finite-state automaton is computed from the probability that is accumulated in all of its individual state sequences. The computation of the entropy from a weighted finite-state automaton requires a normalized model. This article studies an efficient computation of the derivational entropy of left-to-right probabilistic finite-state automata, and it introduces an efficient algorithm for normalizing weighted finite-state automata. The efficient computation of the derivational entropy is also extended to continuous hidden Markov models.https://www.mitpressjournals.org/doi/pdf/10.1162/COLI_a_00306
collection DOAJ
language English
format Article
sources DOAJ
author Joan Andreu Sánchez
Martha Alicia Rocha
Verónica Romero
Mauricio Villegas
spellingShingle Joan Andreu Sánchez
Martha Alicia Rocha
Verónica Romero
Mauricio Villegas
On the Derivational Entropy of Left-to-Right Probabilistic Finite-State Automata and Hidden Markov Models
Computational Linguistics
author_facet Joan Andreu Sánchez
Martha Alicia Rocha
Verónica Romero
Mauricio Villegas
author_sort Joan Andreu Sánchez
title On the Derivational Entropy of Left-to-Right Probabilistic Finite-State Automata and Hidden Markov Models
title_short On the Derivational Entropy of Left-to-Right Probabilistic Finite-State Automata and Hidden Markov Models
title_full On the Derivational Entropy of Left-to-Right Probabilistic Finite-State Automata and Hidden Markov Models
title_fullStr On the Derivational Entropy of Left-to-Right Probabilistic Finite-State Automata and Hidden Markov Models
title_full_unstemmed On the Derivational Entropy of Left-to-Right Probabilistic Finite-State Automata and Hidden Markov Models
title_sort on the derivational entropy of left-to-right probabilistic finite-state automata and hidden markov models
publisher The MIT Press
series Computational Linguistics
issn 1530-9312
publishDate 2018-03-01
description Probabilistic finite-state automata are a formalism that is widely used in many problems of automatic speech recognition and natural language processing. Probabilistic finite-state automata are closely related to other finite-state models as weighted finite-state automata, word lattices, and hidden Markov models. Therefore, they share many similar properties and problems. Entropy measures of finite-state models have been investigated in the past in order to study the information capacity of these models. The derivational entropy quantifies the uncertainty that the model has about the probability distribution it represents. The derivational entropy in a finite-state automaton is computed from the probability that is accumulated in all of its individual state sequences. The computation of the entropy from a weighted finite-state automaton requires a normalized model. This article studies an efficient computation of the derivational entropy of left-to-right probabilistic finite-state automata, and it introduces an efficient algorithm for normalizing weighted finite-state automata. The efficient computation of the derivational entropy is also extended to continuous hidden Markov models.
url https://www.mitpressjournals.org/doi/pdf/10.1162/COLI_a_00306
work_keys_str_mv AT joanandreusanchez onthederivationalentropyoflefttorightprobabilisticfinitestateautomataandhiddenmarkovmodels
AT marthaaliciarocha onthederivationalentropyoflefttorightprobabilisticfinitestateautomataandhiddenmarkovmodels
AT veronicaromero onthederivationalentropyoflefttorightprobabilisticfinitestateautomataandhiddenmarkovmodels
AT mauriciovillegas onthederivationalentropyoflefttorightprobabilisticfinitestateautomataandhiddenmarkovmodels
_version_ 1725998033603133440