2003
DOI: 10.1103/physrevlett.90.089804
|View full text |Cite
|
Sign up to set email alerts
|

Benedetto, Caglioti, and Loreto Reply:

Abstract: Benedetto, Caglioti, and Loreto Reply: In [1] Khmelev et al. claim that [2] contains many misleading statements and that Markov chain approaches represent a more attractive technique than Lempel-Ziv (LZ)-based compression schemes.The authors recall the results of the experiments reported in [3] where it is shown that Markov chain based methods outperform gzip based methods. First of all, it is important to remember that in [3] an LZ-based scheme (rarw) outperformed the Markov chain approach. The generic clai… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

3
41
0

Year Published

2004
2004
2015
2015

Publication Types

Select...
5
3

Relationship

2
6

Authors

Journals

citations
Cited by 24 publications
(44 citation statements)
references
References 4 publications
3
41
0
Order By: Relevance
“…if most of the matches occur in the English part, the expression (1) will give a measure of the relative entropy. We have checked this method on sequences for which the relative entropy is known, obtaining an excellent agreement between the theoretical value of the relative entropy and the computed value [15]. The results of our experiments on linguistic corpora turned out to be very robust with respect to large variations on the size of the file b (typically 1 − 15 Kilobytes (Kb) for a typical size of file A of the order of 32 − 64 Kb).…”
mentioning
confidence: 91%
See 1 more Smart Citation
“…if most of the matches occur in the English part, the expression (1) will give a measure of the relative entropy. We have checked this method on sequences for which the relative entropy is known, obtaining an excellent agreement between the theoretical value of the relative entropy and the computed value [15]. The results of our experiments on linguistic corpora turned out to be very robust with respect to large variations on the size of the file b (typically 1 − 15 Kilobytes (Kb) for a typical size of file A of the order of 32 − 64 Kb).…”
mentioning
confidence: 91%
“…it tends progressively to find most of the matches in the Italian part with respect to the English one, and changes its rules. Therefore if the length of the Italian file is "small enough" [15], i.e. if most of the matches occur in the English part, the expression (1) will give a measure of the relative entropy.…”
mentioning
confidence: 99%
“…It is important to remark that the relative and cross entropies are not distances (metric) in the mathematical sense, since they are not symmetric and do not satisfy in general the triangular inequality. Defining a true distance between strings is an important issue both for theoretical and practical reasons (see for some recent approaches [11,12,13] and for a short review [21]). …”
Section: Entropy and Complexitymentioning
confidence: 99%
“…Recently, a method has been proposed for the estimate of the cross entropy between two strings based on LZ77 [11]. Recalling that the cross entropy C(A|B) between two strings A and B, is given by the entropy per character of B in the optimal coding for A, the idea is that of appending the two sequences and zipping the resulting file A + B.…”
Section: Zippersmentioning
confidence: 99%
“…This statement is technically only true asymptotically but in practice exceptions grow exponentially unlikely for mixing sources. This property concerning the relative entropy was recently used to distinguish and categorize natural languages from only representative samples of their texts [15], although there the slightly different algorithm was used and adaptation to the second sequence continued during its parsing, lowering the discrimination power somewhat.…”
Section: Adaptive Dictionary-based Time-symmetry Testingmentioning
confidence: 99%