2011
DOI: 10.1016/j.ins.2011.07.039
|View full text |Cite
|
Sign up to set email alerts
|

Natural Language Compression on Edge-Guided text preprocessing

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2013
2013
2016
2016

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(8 citation statements)
references
References 41 publications
0
8
0
Order By: Relevance
“…It retains the original Huffman algorithm features, but process the text as a sequence of words and replaces the original words with variable-bit code-words to represent them. This word-based Huffman largely outperforms the traditional character-based one by achieving near about 25% compression ratio instead of 65%, achieved by the character-based Huffman [12].…”
Section: Natural Language Text Modeling and Compressionmentioning
confidence: 91%
See 3 more Smart Citations
“…It retains the original Huffman algorithm features, but process the text as a sequence of words and replaces the original words with variable-bit code-words to represent them. This word-based Huffman largely outperforms the traditional character-based one by achieving near about 25% compression ratio instead of 65%, achieved by the character-based Huffman [12].…”
Section: Natural Language Text Modeling and Compressionmentioning
confidence: 91%
“…LZ is a well known family of compression algorithms as these algorithms are capable of deriving more compressed text file using limited resources such as memory and also with reasonably good speed of compression and decompression. Different popular compressors are designed on the LZ platform such as gzip and p7zip [12].…”
Section: Dictionary-based Compressionmentioning
confidence: 99%
See 2 more Smart Citations
“…Several scholar works on the field of focus on the same goal as ours: Preprocess text to help it compress better. One such study applies Edge-guided text compression that is based on graphs, ordered pairs and sets [1] to transform text into a word net; the adjacencies of the word have a direct relationship with the unique graph, which is the result of the word net. Our approach has less complexity as it only involves letter repositioning, rather than complex data structures as graphs.…”
Section: Related Workmentioning
confidence: 99%