2013
DOI: 10.1111/sjos.12053
|View full text |Cite
|
Sign up to set email alerts
|

Sparse Markov Chains for Sequence Data

Abstract: Finite memory sources and variable‐length Markov chains have recently gained popularity in data compression and mining, in particular, for applications in bioinformatics and language modelling. Here, we consider denser data compression and prediction with a family of sparse Bayesian predictive models for Markov chains in finite state spaces. Our approach lumps transition probabilities into classes composed of invariant probabilities, such that the resulting models need not have a hierarchical structure as in c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
25
0
1

Year Published

2014
2014
2022
2022

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 28 publications
(26 citation statements)
references
References 32 publications
0
25
0
1
Order By: Relevance
“…In contrast to approaches that incorporate temporal layers in methods for static network descriptions, we build our approach on describing the actual dynamics. We first formulate a generative model of discrete temporal processes based on arbitraryorder Markov chains with community structure [24][25][26][27]. Since our model generates event sequences, it does not aggregate data in time windows [13][14][15]17], and, other than the Markov model assumption, needs no a priori imposed timescales.…”
Section: Introductionmentioning
confidence: 99%
“…In contrast to approaches that incorporate temporal layers in methods for static network descriptions, we build our approach on describing the actual dynamics. We first formulate a generative model of discrete temporal processes based on arbitraryorder Markov chains with community structure [24][25][26][27]. Since our model generates event sequences, it does not aggregate data in time windows [13][14][15]17], and, other than the Markov model assumption, needs no a priori imposed timescales.…”
Section: Introductionmentioning
confidence: 99%
“…Allowing the possibility of keeping only one parameter for the equal‐conditional probability classes affords both parsimony and flexibility, giving a better trade‐off of variance that arises from having too many parameters and bias from having a model that does not fit the data due to contexts that are too short. A VLMC is a special case of an SMM where in each equal‐conditional probability class γ l , the m ‐tuples have a common suffix (Jääskinen, Xiong, Koski, & Corander, ). Thus if lγj denotes the length of the longest common suffix of strings of probability equivalence class γ j , then a VLMC has lγj>0j. Figure depicts the context tree of a VLMC with Σ = {0, 1} and m = 3.…”
Section: Extension Of Amc‐based Computation To Smmsmentioning
confidence: 99%
“…SMM (Xiong, Jääskinen, & Corander, ) have also been called minimal Markov models (García & González‐López, ), sparse Markov chains (Jääskinen et al, ), and partition Markov models (Fernández, García, & González‐López, ; García & González‐López, ).…”
Section: Extension Of Amc‐based Computation To Smmsmentioning
confidence: 99%
“…In a more general context, there are also sequences of several words, which are equivalent in that sense. See, for instance, [8,9]. For this kind of data, a model should retrieve and use the redundancy to improve the quality of the estimate.…”
Section: Introductionmentioning
confidence: 99%