Lempel-Ziv-Welch
From Wikipedia, the free encyclopedia
This article may require cleanup to meet Wikipedia's quality standards. Please improve this article if you can. (January 2008) |
Lempel-Ziv-Welch (LZW) is a universal lossless data compression algorithm created by Abraham Lempel, Jacob Ziv, and Terry Welch. It was published by Welch in 1984 as an improved implementation of the LZ78 algorithm published by Lempel and Ziv in 1978. The algorithm is designed to be fast to implement but is not usually optimal because it performs only limited analysis of the data.
Contents |
[edit] Description of the algorithm
The compressor algorithm builds a string translation table from the text being compressed. The string translation table maps fixed-length codes (usually 12-bit) to strings. The string table is initialized with all single-character strings (256 entries in the case of 8-bit characters). As the compressor character-serially examines the text, it stores every unique two-character string into the table as a code/character concatenation, with the code mapping to the corresponding first character. As each two-character string is stored, the first character is sent to the output. Whenever a previously-encountered string is read from the input, the longest such previously-encountered string is determined, and then the code for this string concatenated with the extension character (the next character in the input) is stored in the table. The code for this longest previously-encountered string is output and the extension character is used as the beginning of the next word.
The decompressor algorithm only requires the compressed text as an input, since it can build an identical string table from the compressed text as it is recreating the original text. However, an abnormal case shows up whenever the sequence character/string/character/string/character (with the same character for each character and string for each string) is encountered in the input and character/string is already stored in the string table. When the decompressor reads the code for character/string/character in the input, it cannot resolve it because it has not yet stored this code in its table. This special case can be dealt with because the decompressor knows that the extension character is the previously-encountered character.[1]
[edit] Algorithm
Compressor algorithm:
w = NIL;
while (read a char c) do
if (wc exists in dictionary) then
w = wc;
else
add wc to the dictionary;
output the code for w;
w = c;
endif
done
output the code for w;
Decompressor algorithm:
read a char k;
output k;
w = k;
while (read a char k) do
if (index k exists in dictionary) then
entry = dictionary entry for k;
else if (index k does not exist in dictionary && k == currSizeDict)
entry = w + w[0];
else
signal invalid code;
endif
output entry;
add w+entry[0] to the dictionary;
w = entry;
done
[edit] Example
[edit] Compression
The following table shows the result of executing compressor algorithm on this input:
TOBEORNOTTOBEORTOBEORNOT
Suppose that we are using 256 (8-bit) ASCII code as the default dictionary. The length of this input is 24 characters. So this input requires 24 * 8 = 192 bits to store.
c |
w |
wc |
output |
dictionary |
T |
<NIL> |
T |
||
O |
T |
TO |
T |
TO = <256> |
B |
O |
OB |
O |
OB = <257> |
E |
B |
BE |
B |
BE = <258> |
O |
E |
EO |
E |
EO = <259> |
R |
O |
OR |
O |
OR = <260> |
N |
R |
RN |
R |
RN = <261> |
O |
N |
NO |
N |
NO = <262> |
T |
O |
OT |
O |
OT = <263> |
T |
T |
TT |
T |
TT = <264> |
O |
T |
TO |
||
B |
TO |
TOB |
<256> |
TOB = <265> |
E |
B |
BE |
||
O |
BE |
BEO |
<258> |
BEO = <266> |
R |
O |
OR |
||
T |
OR |
ORT |
<260> |
ORT = <267> |
O |
T |
TO |
||
B |
TO |
TOB |
||
E |
TOB |
TOBE |
<265> |
TOBE = <268> |
O |
E |
EO |
||
R |
EO |
EOR |
<259> |
EOR = <269> |
N |
R |
RN |
||
O |
RN |
RNO |
<261> |
RNO = <270> |
T |
O |
OT |
||
OT |
<263> |
After compression we have this sequence of 9-bits codes in output:
TOBEORNOT<256><258><260><265><259><261><263>
This sequence requires 16 * 9 = 144 bits to store.
k |
w |
entry |
w+entry[0] |
output |
dictionary |
84 |
T |
T |
|||
77 |
T |
O |
TO |
O |
TO = <256> |
66 |
O |
B |
OB |
B |
OB = <257> |
69 |
B |
E |
BE |
E |
BE = <258> |
77 |
E |
O |
EO |
O |
EO = <259> |
80 |
O |
R |
OR |
R |
OR = <260> |
76 |
R |
N |
RN |
N |
RN = <261> |
77 |
N |
O |
NO |
O |
NO = <262> |
84 |
O |
T |
OT |
T |
OT = <263> |
<256> |
T |
TO |
TT |
TO |
TT = <264> |
<258> |
TO |
BE |
TOB |
BE |
TOB = <265> |
<260> |
BE |
OR |
BEO |
OR |
BEO = <266> |
<265> |
OR |
TOB |
ORT |
TOB |
ORT = <267> |
<259> |
TOB |
EO |
TOBE |
EO |
TOBE = <268> |
<261> |
EO |
RN |
EOR |
RN |
EOR = <269> |
<263> |
RN |
OT |
RNO |
OT |
RNO = <270> |
[edit] Uses
The method became widely used in the program compress, which became a more or less standard utility in Unix systems circa 1986. (It has since disappeared from many for both legal and technical reasons.) Several other popular compression utilities also used the method, or closely related ones.
It became very widely used after it became part of the GIF image format in 1987. It may also (optionally) be used in TIFF files.
LZW compression provided a better compression ratio, in most applications, than any well-known method available up to that time. It became the first widely used universal data compression method on computers. It would typically compress large English texts to about half of their original sizes.
Today, an implementation of the algorithm is contained within the popular Adobe Acrobat software program.
[edit] Example
This example shows the LZW algorithm in action, showing the status of the output and the dictionary at every stage, both in encoding and decoding the message. In order to keep things clear, let us assume that we're dealing with a simple alphabet - capital letters only, and no punctuation or spaces. This example has been constructed to give reasonable compression on a very short message; when used on real data, repetition is generally less pronounced, and so the initial parts of a message will see little compression. As the message grows, however, the compression ratio tends asymptotically to the maximum.[2] A message to be sent might then look like the following:
TOBEORNOTTOBEORTOBEORNOT#
The # is a marker used to show that the end of the message has been reached. Clearly, then, we have 27 symbols in our alphabet (the 26 capital letters A through Z, plus the # character). A computer will render these as strings of bits; 5-bit strings are needed to give sufficient combinations to encompass the entire dictionary. As the dictionary grows, the strings will need to grow in length to accommodate the additional entries. A 5-bit string gives 25 = 32 possible combinations of bits, and so when the 33rd dictionary word is created, the algorithm will have to start using 6-bit strings (for all strings, including those which were previously represented by only five bits). Note that since the all-zero string 00000 is used, and is labeled "0", the 33rd dictionary entry will be labeled 32. The initial dictionary, then, will consist of the following:
# = 00000 A = 00001 B = 00010 C = 00011 . . . Z = 11010
[edit] Encoding
If we weren't using LZW, and just sent the message as it stands (25 symbols at 5 bits each), it would require 125 bits. We will be able to compare this figure to the LZW output later. We are now in a position to apply LZW to the message.
Symbol: Bit Code: New Dictionary Entry: (= output) T 20 = 10100 O 15 = 01111 28: TO <--- Don't forget, we originally had 27 symbols, so the next one is 28th. B 2 = 00010 29: OB E 5 = 00101 30: BE O 15 = 01111 31: EO <--- start using 6-bit strings R 18 = 010010 32: OR N 14 = 001110 33: RN O 15 = 001111 34: NO T 20 = 010100 35: OT TO 28 = 011100 36: TT BE 30 = 011110 37: TOB OR 32 = 100000 38: BEO TOB 37 = 100101 39: ORT EO 31 = 011111 40: TOBE RN 33 = 100001 41: EOR OT 35 = 100011 42: RNO # 0 = 000000 43: OT#
This is somewhat clearer:
Current Next Output Value Extended Sequence Char (# of bits) Dictionary NULL T T O 20 = 5 bits 27: TO <-- This IS the 28th entry, but the initial entries are numbered 0-26 so this is #27. O B 15 = 5 bits 28: OB B E 2 = 5 bits 29: BE E O 5 = 5 bits 30: EO O R 15 = 5 bits 31: OR R N 18 = 6 bits 32: RN <-- Starting at R, 6 bits are used {floor(lg2(init_dict_size + num_chars_output)) + 1} N O 14 = 6 bits 33: NO i.e. O: floor(lg2(27 + 4)) + 1 = 5 bits -> 01111 O T 15 = 6 bits 34: OT R: floor(lg2(27 + 5)) + 1 = 6 bits -> 010010 T T 20 = 6 bits 35: TT TO B 27 = 6 bits 36: TOB BE O 29 = 6 bits 37: BEO OR T 31 = 6 bits 38: ORT TOB E 36 = 6 bits 39: TOBE EO R 30 = 6 bits 40: EOR RN O 32 = 6 bits 41: RNO OT # 34 = 6 bits 42: OT# # 0 = 6 bits Total Length = 5*5 + 12*6 = 97 bits.
In using LZW we have made a saving of 28 bits out of 125 -- we have reduced the message by almost 22%. If the message were longer, then the dictionary words would begin to represent longer and longer sections of text, allowing repeated words to be sent very compactly.
[edit] Decoding
Imagine now that we have received the message produced above, and wish to decode it. We need to know in advance the initial dictionary used, but we can reconstruct the additional entries as we go, since they are always simply concatenations of previous entries.
Bits: Output: New Entry: Full: Partial: 10100 = 20 T 28: T? 01111 = 15 O 28: TO 29: O? 00010 = 2 B 29: OB 30: B? 00101 = 5 E 30: BE 31: E? 01111 = 15 O 31: EO 32: O? <--- start using 6-bit strings 010010 = 18 R 32: OR 33: R? 001110 = 14 N 33: RN 34: N? 001111 = 15 O 34: NO 35: O? 010100 = 20 T 35: OT 36: T? 011100 = 28 TO 36: TT 37: TO? <--- for 36, only add 1st element 011110 = 30 BE 37: TOB 38: BE? of next dictionary word 100000 = 32 OR 38: BEO 39: OR? 100101 = 37 TOB 39: ORT 40: TOB? 011111 = 31 EO 40: TOBE 41: EO? 100001 = 33 RN 41: EOR 42: RN? 100011 = 35 OT 42: RNO 43: OT? 000000 = 0 #
The only slight complication comes if the newly-created dictionary word is sent immediately. In the decoding example above, when the decoder receives the first symbol, T, it knows that symbol 28 begins with a T, but what does it end with? The problem is illustrated below. We are decoding part of a message that reads ABABA:
Bits: Output: New Entry: Full: Partial: . . . 011101 = 29 AB 46: (word) 47: AB? 101111 = 47 AB? <--- what do we do here?
At first glance, this may appear to be asking the impossible of the decoder. We know ahead of time that entry 47 should be ABA, but how can the decoder work this out? The critical step is to note that 47 is built out of 29 plus whatever comes next. 47, therefore, ends with "whatever comes next". But, since it was sent immediately, it must also start with "whatever comes next", and so must end with the same symbol it starts with, namely A. This trick allows the decoder to see that 47 must be ABA.
More generally the situation occurs whenever the encoder encounters the input of the form cScSc, where c is a single character, S is a string and cS is already in the dictionary. The encoder outputs the symbol for cS putting new symbol for cSc in the dictionary. Next it sees the cSc in the input and sends the new symbol it just inserted into the dictionary. By the reasoning presented in the above example this is the only case where the newly-created symbol is sent immediately.
[edit] Python example
# Lempel-Ziv-Welch compression algorithm def compress(uncompressed): """Compress a string to a list of output symbols.""" # Build the dictionary. dict_size = 256 dictionary = {} for i in range(dict_size): dictionary[chr(i)] = chr(i) w = '' result = [] for c in uncompressed: wc = w + c if wc in dictionary: w = wc else: result.append(dictionary[w]) # Add wc to the dictionary. dictionary[wc] = dict_size dict_size += 1 w = c # Output the code for w. if w: result += [char for char in w] return result def decompress(compressed): """Decompress a list of output ks to a string.""" # Build the dictionary. dict_size = 256 dictionary = {} for i in range(dict_size): dictionary[chr(i)] = chr(i) w = result = compressed.pop(0) for k in compressed: if k in dictionary: entry = dictionary[k] elif k == len(dictionary): entry = w + w[0] else: raise ValueError, 'Bad compressed k: %s' % k result += entry # Add w+entry[0] to the dictionary. dictionary[dict_size] = w+entry[0] dict_size += 1 w = entry return result
How to use:
compressed = compress('TOBEORNOTTOBEORTOBEORNOT') print compressed decompressed = decompress(compressed) print decompressed
[edit] Patent issues
Various patents have been issued in the United States and other countries for LZW and similar algorithms. LZ78 was covered by U.S. Patent 4,464,650 by Lempel, Ziv, Cohn, and Eastman, assigned to Sperry Corporation, later Unisys Corporation, filed on August 10, 1981. Two US patents were issued for the LZW algorithm: U.S. Patent 4,814,746 by Victor S. Miller and Mark N. Wegman and assigned to IBM, originally filed on June 1, 1983, and U.S. Patent 4,558,302 by Welch, assigned to Sperry Corporation, later Unisys Corporation, filed on June 20, 1983. On June 20, 2003, this patent on the LZW algorithm expired [1].
US Patent 4,558,302 is the one that has caused the most controversy (See Graphics Interchange Format#Unisys and LZW patent enforcement).
[edit] Lempel-Ziv-Welch vs. Ziv-Lempel-Welch
Although the name of the algorithm refers to the inventors as Lempel, Ziv and Welch, some people[citation needed] claim that the intellectual property rightly goes to Ziv first, so the method should be called the Ziv-Lempel-Welch algorithm, and not the Lempel-Ziv-Welch algorithm. Others who distinguish between the algorithm and the code prefer calling the algorithm LZ and the code written by Welch as LZW.
[edit] See also
- LZWL
- LZ77 and LZ78
- Lempel-Ziv-Markov algorithm (LZMA)
- Lempel-Ziv-Storer-Szymanski (LZSS)
- DEFLATE
- Burrows-Wheeler transform
[edit] References
- ^ Welch, T. A. (June 1984). "A technique for high-performance data compression." Computer. Vol. 17, pp. 8-19.
- ^ Jacob Ziv and Abraham Lempel; Compression of Individual Sequences Via Variable-Rate Coding, IEEE Transactions on Information Theory, September 1978.
[edit] External links
- "A technique for high-performance data compression" - The original paper by Welch
- United States Patent 4,558,302 URL Retrieved on Saturday, June 03, 2006
- "LZW Data Compression", by Mark Nelson (DDJ Article with source code)
- Sad day... GIF patent dead at 20
- Bringing back LZW compression by Nathan Willis
|