2003
DOI: 10.1016/s0167-2789(03)00047-2
|View full text |Cite
|
Sign up to set email alerts
|

Data compression and learning in time sequences analysis

Abstract: Motivated by the problem of the definition of a distance between two sequences of characters, we investigate the so-called learning process of typical sequential data compression schemes. We focus on the problem of how a compression algorithm optimizes its features at the interface between two different sequences A and B while zipping the sequence A + B obtained by simply appending B after A. We show the existence of a universal scaling function (the so-called learning function) which rules the way in which th… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0
1

Year Published

2004
2004
2016
2016

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 36 publications
(32 citation statements)
references
References 44 publications
0
31
0
1
Order By: Relevance
“…In Figure 3 an experiment of recognition is reported [20]. Here an unknown sequence is compared, in the sense discussed in the previous section, with a number of known strings.…”
Section: Examples and Numerical Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…In Figure 3 an experiment of recognition is reported [20]. Here an unknown sequence is compared, in the sense discussed in the previous section, with a number of known strings.…”
Section: Examples and Numerical Resultsmentioning
confidence: 99%
“…In this way the zipper "learns" the A file and, when encounters the B subsequence, tries to compress it with a coding optimized for A. If B is not too long [20,21], thus preventing LZ77 from learning it as well, the cross entropy per character can be estimated as:…”
Section: Zippersmentioning
confidence: 99%
“…This term is intuitively close to C(x + ∆y) − C(x) in Equation (9), as both aim at expressing a small fraction of y only in terms of x. Secondly, the term C(y + ∆y) − C(y)in Equation (9) is intuitively close to C(∆y) in Equation (10), where in the former a representative dictionary extracted from y is used to code the fraction ∆y, while the latter discards any limitation regarding the size of the analysed objects and considers the full string y. This solves a problem raised in [30], which investigates the optimal size for ∆y in Equation (9), which does not represent y well enough if set too small, while it uses too much information from y itself in the compression step if set too large.…”
Section: Relative Entropymentioning
confidence: 96%
“…In [11] it has been studied in detail what happens when a compression algorithm tries to optimize its features at the interface between two different sequences A and B while zipping the sequence A + B obtained by simply appending B after A. It has been shown in particular the existence of a scaling function ruling the way the compression algorithm learns a sequence B after having compressed a sequence A.…”
Section: A Remoteness Between Two Textsmentioning
confidence: 99%
“…A first field of activity [11,12] is that of segmentation problems, i.e. cases in which a unique string must be partitioned into subsequences according to some criteria to identify discontinuities in its statistical properties.…”
Section: Introductionmentioning
confidence: 99%