2015
DOI: 10.1109/tit.2015.2452252
|View full text |Cite
|
Sign up to set email alerts
|

Efficient and Compact Representations of Prefix Codes

Abstract: Most of the attention in statistical compression is given to the space used by the compressed sequence, a problem completely solved with optimal prefix codes. However, in many applications, the storage space used to represent the prefix code itself can be an issue. In this paper we introduce and compare several techniques to store prefix codes. Let N be the sequence length and n be the alphabet size. Then a naive storage of an optimal prefix code uses O(n log n) bits. Our first technique shows how to use O(n l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(9 citation statements)
references
References 55 publications
0
9
0
Order By: Relevance
“…Some technicalities about possible coders and decoders for P -codes can be further found in [ 18 ]. Information on effective decoding algorithms can be found in [ 19 , 20 ] and on memory-efficient representation of prefix codes can be found in [ 21 ].…”
Section: Discussionmentioning
confidence: 99%
“…Some technicalities about possible coders and decoders for P -codes can be further found in [ 18 ]. Information on effective decoding algorithms can be found in [ 19 , 20 ] and on memory-efficient representation of prefix codes can be found in [ 21 ].…”
Section: Discussionmentioning
confidence: 99%
“…We have run experiments to compare the solution of Theorem 2 (referred to as WMM in the sequel, for Wavelet Matrix Model) with the only previous encoding, that is, the one used by Claude et al [1] (denoted by TABLE). Note that our codes are not canonical, so other solutions [5] do not apply. Claude et al [1] use for encoding a single table of σL bits storing the code of each symbol, and thus they easily encode in constant time.…”
Section: Methodsmentioning
confidence: 99%
“…If the alphabet consists of σ characters and the maximum codeword length is L, then we can build an O(σ log L)-bit data structure with O(log L) query time that, given a character, returns its codeword's length and rank among codewords of that length, or vice versa. If L is at most a constant times the size of a machine word (which it is when we are considering, e.g., Huffman codes for strings in the RAM model) then in theory we can make the predecessor search and the data structure's queries constant-time, meaning we can encode and decode in constant time [5].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…The same idea is presented perhaps more lucidly in [25, Section 2.6.3], with an explicit claim that this representation requires σ lg σ + O(lg 2 n) bits 2 and achieves O(lg lg n) time per codeword. Given ℓ max is the maximum length of the codewords, an improvement, both in space and time, has been achieved by Gagie et al [13], who gave a representation of the canonical Huffman tree within…”
Section: Introductionmentioning
confidence: 99%