2005
DOI: 10.1088/0143-0807/26/5/s08
|View full text |Cite
|
Sign up to set email alerts
|

Measuring complexity with zippers

Abstract: Abstract. Physics concepts have often been borrowed and independently developed by other fields of science. In this perspective a significant example is that of entropy in Information Theory. The aim of this paper is to provide a short and pedagogical introduction to the use of data compression techniques for the estimate of entropy and other relevant quantities in Information Theory and Algorithmic Information Theory. We consider in particular the LZ77 algorithm as case study and discuss how a zipper can be u… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2005
2005
2025
2025

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 17 publications
(23 citation statements)
references
References 27 publications
0
23
0
Order By: Relevance
“…As ǫ decreases, more detailed information is described by the symbols. Although c is not a reliable estimate of h at the finer partitions, it is still an indicator of the information content of the data streams and also shows the same decrease with Re [36]. The important point is that the general behavior of c and h * is the same for partitions of all sizes.…”
Section: Turbulencementioning
confidence: 92%
See 2 more Smart Citations
“…As ǫ decreases, more detailed information is described by the symbols. Although c is not a reliable estimate of h at the finer partitions, it is still an indicator of the information content of the data streams and also shows the same decrease with Re [36]. The important point is that the general behavior of c and h * is the same for partitions of all sizes.…”
Section: Turbulencementioning
confidence: 92%
“…In order to account for the "overhead" (file headers, etc. ), a random data set is compressed and that compression ratio is used to normalize the real data [36].…”
Section: Entropy and Entropy Estimationmentioning
confidence: 99%
See 1 more Smart Citation
“…For equilibrium systems, a time series of the spin or the Edwards-Anderson autocorrelation parameter at a given site, obtained by Monte Carlo simulation, was used to locate the critical points of the 3D Edwards-Anderson spin glass [17] and the 2D and 3D Ising models [18,19], and to approximate the entropy of the 2D Ising model [19]. Data compression has also proven to be a useful tool in the definition and characterization of complexity of (mostly one-dimensional) dynamical models, such as cellular automata and dynamical systems [20][21][22][23][24][25][26], as well as for turbulence [27]. Methods based on data compression have also been used to estimate the entropy production of a nonequilibrium stationary state [28,29] and to detect the onset of chaos in biological systems [30][31][32].…”
Section: A Computable Information Densitymentioning
confidence: 99%
“…On the other hand one can construct AT by merging dictionaries coming from different original texts: merging dictionaries extracted from different texts all about the same subject or all written by the same author. In this way the AT would play the role of an archetypal text of that specific subject or that specific author [63]. The possibility to construct many different AT all representative of the same original sequence (or of a given source) allows for an alternative way to estimate the selfentropy of a source (and consequently the relative entropy between two sources as mentioned above).…”
Section: Dictionaries and Artificial Textsmentioning
confidence: 99%