|Original author(s)||Zeev Tarantov, Steinar H. Gunderson|
|Initial release||March 18, 2011|
|Stable release||1.1.2 / February 28, 2014|
|License||Apache 2 (up to 1.0.1)/New BSD|
Snappy (previously known as Zippy) is a fast data compression and decompression library written in C++ by Google based on ideas from LZ77 and open-sourced in 2011. It does not aim for maximum compression, or compatibility with any other compression library; instead, it aims for very high speeds and reasonable compression. Compression speed is 250 MB/s and decompression speed is 500 MB/s using a single core of a Core i7 processor running in 64-bit mode. The compression ratio is 20–100% lower than gzip.
Snappy is widely used in Google projects like BigTable, MapReduce and in compression data in Google's internal RPC systems. It can be used in open-source projects like Cassandra, Hadoop, LevelDB, MongoDB, RocksDB, Lucene Decompression is tested to detect any errors in the compressed stream. Snappy does not use inline assembler and is portable.
The first bytes of the stream are the length of uncompressed data, stored as a little-endian varint, which allows for variable-length encoding. The lower seven bits of each byte are used for data and the first bit is a flag which tells if the next byte is used for the same integer.
The remaining bytes in the stream are encoded using one of four element types. The element type is encoded in the first byte (tag byte) of the element. The two lower bits of this byte is the type code:
- 00 – Literal – uncompressed data; upper 6 bits are used to store length of data; if the length of data is more 60 bytes, additional variable-length encoding is added
- 01 – Copy with length stored as 3 bits and offset stored as 11 bits; one byte after tag byte is used for part of offset;
- 10 – Copy with length stored as 6 bits of tag byte and offset stored as two-byte integer after the tag byte;
- 11 – Copy with length stored as 6 bits of tag byte and offset stored as four-byte little-endian integer after the tag byte;
The copy refers to the dictionary (just-decompressed data). The offset is the shift from the current position back to the already decompressed stream. The length is the number of bytes to copy from the dictionary. The size of the dictionary was limited by the 1.0 Snappy compressor to 32768 bytes, and updated to 65536 in version 1.1.
Example of a compressed stream
Wikipedia is a free, web-based, collaborative, multilingual encyclopedia project.
may be compressed to (shown as hex data, every compressed subpart is described):
0000000: ca02 f042 5769 6b69 7065 6469 6120 6973 ...BWikipedia is
Length of data is 0x02ca(varint) = 0x014a = 330 bytes; 0xf042 = literal of 66+1 bytes follows
0000010: 2061 2066 7265 652c 2077 6562 2d62 6173 a free, web-bas 0000020: 6564 2c20 636f 6c6c 6162 6f72 6174 6976 ed, collaborativ 0000030: 652c 206d 756c 7469 6c69 6e67 7561 6c20 e, multilingual 0000040: 656e 6379 636c 6f09 3ff0 8170 726f 6a65 encyclo.?..proje
0x09 is tag-byte of type 01 with length of 0b10= 2 +4 = 6 and offset = 0x03f = 63 or "pedia ";
0xf081 is a literal with length of 129+1 bytes
0000050: 6374 2e00 0000 0000 0000 0000 0000 0000 ct.
In this example, all common substrings with four or more characters were eliminated by the compression process. More common compressors can compress this better. Unlike compression methods such as gzip and bzip2, there is no entropy encoding used to pack alphabet into the bit stream.
Snappy distributions include C++ and C bindings. Third party-provided bindings and ports include:
- "Google Snappy–A Fast Compressing Library". InfoQ. Retrieved August 1, 2011.
- Google open sources MapReduce compression. In the name of speed // The Register, 2011-03-24
- "Snappy: A fast compressor/decompressor: Readme". Google Code. Retrieved August 1, 2011."Snappy vs lzo vs zlib".
- snappy. A fast compressor/decompressor - Project page at Google Code