Asymmetric numeral systems: entropy coding combining speed of Huffman coding with compression rate of arithmetic coding
نویسنده
چکیده
The modern data compression is mainly based on two approaches to entropy coding: Huffman (HC) and arithmetic/range coding (AC). The former is much faster, but approximates probabilities with powers of 2, usually leading to relatively low compression rates. The latter uses nearly exact probabilities easily approaching theoretical compression rate limit (Shannon entropy), but at cost of much larger computational cost. Asymmetric numeral systems (ANS) is a new approach to accurate entropy coding, which allows to end this tradeoff between speed and rate: the recent implementation [1] provides about 50% faster decoding than HC for 256 size alphabet, with compression rate similar to provided by AC. This advantage is due to being simpler than AC: using single natural number as the state, instead of two to represent a range. Beside simplifying renormalization, it allows to put the entire behavior for given probability distribution into a relatively small table: defining entropy coding automaton. The memory cost of such table for 256 size alphabet is a few kilobytes. There is a large freedom while choosing a specific table using pseudorandom number generator initialized with cryptographic key for this purpose allows to simultaneously encrypt the data. This article also introduces and discusses many other variants of this new entropy coding approach, which can provide direct alternatives for standard AC, for large alphabet range coding, or for approximated quasi arithmetic coding.
منابع مشابه
Lightweight compression with encryption based on Asymmetric Numeral Systems
Data compression combined with effective encryption is a common requirement of data storage and transmission. Low cost of these operations is often a high priority in order to increase transmission speed and reduce power usage. This requirement is crucial for battery-powered devices with limited resources, such as autonomous remote sensors or implants. Well-known and popular encryption techniqu...
متن کاملTowards Development of Efficient Compression Techniques for Different Types of Source Data
This Chapter summarizes and concludes our research work. Our work is mainly in the area of lossless data compression using arithmetic coding. Arithmetic coding is used to compress any type of source data. It falls under entropy coding compression methods and is most widely used as a replacement of Huffman encoding entropy method due to its nearly optimal entropy. It is also used with most of th...
متن کاملIntroduction to Data Compression
3 Probability Coding 10 3.1 Prefix Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 3.1.1 Relationship to Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 3.2 Huffman Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 3.2.1 Combining Messages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 3.2.2 Minim...
متن کاملEvaluation of Huffman and Arithmetic Algorithms for Multimedia Compression Standards
Compression is a technique to reduce the quantity of data without excessively reducing the quality of the multimedia data.The transition and storing of compressed multimedia data is much faster and more efficient than original uncompressed multimedia data. There are various techniques and standards for multimedia data compression, especially for image compression such as the JPEG and JPEG2000 s...
متن کاملEntropy Coding of Quantized Spectral Components in Fdlp Audio Codec
Audio codec based on Frequency Domain Linear Prediction (FDLP) exploits auto-regressive modeling to approximate instantaneous energy in critical frequency sub-bands of relatively long input segments. Current version of the FDLP codec operating at 66 kbps has shown to provide comparable subjective listening quality results to the state-of-the-art codecs on similar bit-rates even without employin...
متن کامل