forked from KolibriOS/kolibrios
e9b1c1bac6
git-svn-id: svn://kolibrios.org@6725 a494cfbc-eb01-0410-851d-a64ba20cac60
1776 lines
62 KiB
C
1776 lines
62 KiB
C
/*
|
|
Copyright (c) 1990-2008 Info-ZIP. All rights reserved.
|
|
|
|
See the accompanying file LICENSE, version 2007-Mar-04 or later
|
|
(the contents of which are also included in unzip.h) for terms of use.
|
|
If, for some reason, all these files are missing, the Info-ZIP license
|
|
also may be found at: ftp://ftp.info-zip.org/pub/infozip/license.html
|
|
*/
|
|
/* inflate.c -- by Mark Adler
|
|
version c17e, 30 Mar 2007 */
|
|
|
|
|
|
/* Copyright history:
|
|
- Starting with UnZip 5.41 of 16-April-2000, this source file
|
|
is covered by the Info-Zip LICENSE cited above.
|
|
- Prior versions of this source file, found in UnZip source packages
|
|
up to UnZip 5.40, were put in the public domain.
|
|
The original copyright note by Mark Adler was:
|
|
"You can do whatever you like with this source file,
|
|
though I would prefer that if you modify it and
|
|
redistribute it that you include comments to that effect
|
|
with your name and the date. Thank you."
|
|
|
|
History:
|
|
vers date who what
|
|
---- --------- -------------- ------------------------------------
|
|
a ~~ Feb 92 M. Adler used full (large, one-step) lookup table
|
|
b1 21 Mar 92 M. Adler first version with partial lookup tables
|
|
b2 21 Mar 92 M. Adler fixed bug in fixed-code blocks
|
|
b3 22 Mar 92 M. Adler sped up match copies, cleaned up some
|
|
b4 25 Mar 92 M. Adler added prototypes; removed window[] (now
|
|
is the responsibility of unzip.h--also
|
|
changed name to slide[]), so needs diffs
|
|
for unzip.c and unzip.h (this allows
|
|
compiling in the small model on MSDOS);
|
|
fixed cast of q in huft_build();
|
|
b5 26 Mar 92 M. Adler got rid of unintended macro recursion.
|
|
b6 27 Mar 92 M. Adler got rid of nextbyte() routine. fixed
|
|
bug in inflate_fixed().
|
|
c1 30 Mar 92 M. Adler removed lbits, dbits environment variables.
|
|
changed BMAX to 16 for explode. Removed
|
|
OUTB usage, and replaced it with flush()--
|
|
this was a 20% speed improvement! Added
|
|
an explode.c (to replace unimplod.c) that
|
|
uses the huft routines here. Removed
|
|
register union.
|
|
c2 4 Apr 92 M. Adler fixed bug for file sizes a multiple of 32k.
|
|
c3 10 Apr 92 M. Adler reduced memory of code tables made by
|
|
huft_build significantly (factor of two to
|
|
three).
|
|
c4 15 Apr 92 M. Adler added NOMEMCPY do kill use of memcpy().
|
|
worked around a Turbo C optimization bug.
|
|
c5 21 Apr 92 M. Adler added the WSIZE #define to allow reducing
|
|
the 32K window size for specialized
|
|
applications.
|
|
c6 31 May 92 M. Adler added some typecasts to eliminate warnings
|
|
c7 27 Jun 92 G. Roelofs added some more typecasts (444: MSC bug).
|
|
c8 5 Oct 92 J-l. Gailly added ifdef'd code to deal with PKZIP bug.
|
|
c9 9 Oct 92 M. Adler removed a memory error message (~line 416).
|
|
c10 17 Oct 92 G. Roelofs changed ULONG/UWORD/byte to ulg/ush/uch,
|
|
removed old inflate, renamed inflate_entry
|
|
to inflate, added Mark's fix to a comment.
|
|
c10.5 14 Dec 92 M. Adler fix up error messages for incomplete trees.
|
|
c11 2 Jan 93 M. Adler fixed bug in detection of incomplete
|
|
tables, and removed assumption that EOB is
|
|
the longest code (bad assumption).
|
|
c12 3 Jan 93 M. Adler make tables for fixed blocks only once.
|
|
c13 5 Jan 93 M. Adler allow all zero length codes (pkzip 2.04c
|
|
outputs one zero length code for an empty
|
|
distance tree).
|
|
c14 12 Mar 93 M. Adler made inflate.c standalone with the
|
|
introduction of inflate.h.
|
|
c14b 16 Jul 93 G. Roelofs added (unsigned) typecast to w at 470.
|
|
c14c 19 Jul 93 J. Bush changed v[N_MAX], l[288], ll[28x+3x] arrays
|
|
to static for Amiga.
|
|
c14d 13 Aug 93 J-l. Gailly de-complicatified Mark's c[*p++]++ thing.
|
|
c14e 8 Oct 93 G. Roelofs changed memset() to memzero().
|
|
c14f 22 Oct 93 G. Roelofs renamed quietflg to qflag; made Trace()
|
|
conditional; added inflate_free().
|
|
c14g 28 Oct 93 G. Roelofs changed l/(lx+1) macro to pointer (Cray bug)
|
|
c14h 7 Dec 93 C. Ghisler huft_build() optimizations.
|
|
c14i 9 Jan 94 A. Verheijen set fixed_t{d,l} to NULL after freeing;
|
|
G. Roelofs check NEXTBYTE macro for EOF.
|
|
c14j 23 Jan 94 G. Roelofs removed Ghisler "optimizations"; ifdef'd
|
|
EOF check.
|
|
c14k 27 Feb 94 G. Roelofs added some typecasts to avoid warnings.
|
|
c14l 9 Apr 94 G. Roelofs fixed split comments on preprocessor lines
|
|
to avoid bug in Encore compiler.
|
|
c14m 7 Jul 94 P. Kienitz modified to allow assembler version of
|
|
inflate_codes() (define ASM_INFLATECODES)
|
|
c14n 22 Jul 94 G. Roelofs changed fprintf to macro for DLL versions
|
|
c14o 23 Aug 94 C. Spieler added a newline to a debug statement;
|
|
G. Roelofs added another typecast to avoid MSC warning
|
|
c14p 4 Oct 94 G. Roelofs added (voidp *) cast to free() argument
|
|
c14q 30 Oct 94 G. Roelofs changed fprintf macro to MESSAGE()
|
|
c14r 1 Nov 94 G. Roelofs fixed possible redefinition of CHECK_EOF
|
|
c14s 7 May 95 S. Maxwell OS/2 DLL globals stuff incorporated;
|
|
P. Kienitz "fixed" ASM_INFLATECODES macro/prototype
|
|
c14t 18 Aug 95 G. Roelofs added UZinflate() to use zlib functions;
|
|
changed voidp to zvoid; moved huft_build()
|
|
and huft_free() to end of file
|
|
c14u 1 Oct 95 G. Roelofs moved G into definition of MESSAGE macro
|
|
c14v 8 Nov 95 P. Kienitz changed ASM_INFLATECODES to use a regular
|
|
call with __G__ instead of a macro
|
|
c15 3 Aug 96 M. Adler fixed bomb-bug on random input data (Adobe)
|
|
c15b 24 Aug 96 M. Adler more fixes for random input data
|
|
c15c 28 Mar 97 G. Roelofs changed USE_ZLIB fatal exit code from
|
|
PK_MEM2 to PK_MEM3
|
|
c16 20 Apr 97 J. Altman added memzero(v[]) in huft_build()
|
|
c16b 29 Mar 98 C. Spieler modified DLL code for slide redirection
|
|
c16c 04 Apr 99 C. Spieler fixed memory leaks when processing gets
|
|
stopped because of input data errors
|
|
c16d 05 Jul 99 C. Spieler take care of FLUSH() return values and
|
|
stop processing in case of errors
|
|
c17 31 Dec 00 C. Spieler added preliminary support for Deflate64
|
|
c17a 04 Feb 01 C. Spieler complete integration of Deflate64 support
|
|
c17b 16 Feb 02 C. Spieler changed type of "extra bits" arrays and
|
|
corresponding huft_build() parameter e from
|
|
ush into uch, to save space
|
|
c17c 9 Mar 02 C. Spieler fixed NEEDBITS() "read beyond EOF" problem
|
|
with CHECK_EOF enabled
|
|
c17d 23 Jul 05 C. Spieler fixed memory leaks in inflate_dynamic()
|
|
when processing invalid compressed literal/
|
|
distance table data
|
|
c17e 30 Mar 07 C. Spieler in inflate_dynamic(), initialize tl and td
|
|
to prevent freeing unallocated huft tables
|
|
when processing invalid compressed data and
|
|
hitting premature EOF, do not reuse td as
|
|
temp work ptr during tables decoding
|
|
*/
|
|
|
|
|
|
/*
|
|
Inflate deflated (PKZIP's method 8 compressed) data. The compression
|
|
method searches for as much of the current string of bytes (up to a
|
|
length of 258) in the previous 32K bytes. If it doesn't find any
|
|
matches (of at least length 3), it codes the next byte. Otherwise, it
|
|
codes the length of the matched string and its distance backwards from
|
|
the current position. There is a single Huffman code that codes both
|
|
single bytes (called "literals") and match lengths. A second Huffman
|
|
code codes the distance information, which follows a length code. Each
|
|
length or distance code actually represents a base value and a number
|
|
of "extra" (sometimes zero) bits to get to add to the base value. At
|
|
the end of each deflated block is a special end-of-block (EOB) literal/
|
|
length code. The decoding process is basically: get a literal/length
|
|
code; if EOB then done; if a literal, emit the decoded byte; if a
|
|
length then get the distance and emit the referred-to bytes from the
|
|
sliding window of previously emitted data.
|
|
|
|
There are (currently) three kinds of inflate blocks: stored, fixed, and
|
|
dynamic. The compressor outputs a chunk of data at a time and decides
|
|
which method to use on a chunk-by-chunk basis. A chunk might typically
|
|
be 32K to 64K, uncompressed. If the chunk is uncompressible, then the
|
|
"stored" method is used. In this case, the bytes are simply stored as
|
|
is, eight bits per byte, with none of the above coding. The bytes are
|
|
preceded by a count, since there is no longer an EOB code.
|
|
|
|
If the data are compressible, then either the fixed or dynamic methods
|
|
are used. In the dynamic method, the compressed data are preceded by
|
|
an encoding of the literal/length and distance Huffman codes that are
|
|
to be used to decode this block. The representation is itself Huffman
|
|
coded, and so is preceded by a description of that code. These code
|
|
descriptions take up a little space, and so for small blocks, there is
|
|
a predefined set of codes, called the fixed codes. The fixed method is
|
|
used if the block ends up smaller that way (usually for quite small
|
|
chunks); otherwise the dynamic method is used. In the latter case, the
|
|
codes are customized to the probabilities in the current block and so
|
|
can code it much better than the pre-determined fixed codes can.
|
|
|
|
The Huffman codes themselves are decoded using a multi-level table
|
|
lookup, in order to maximize the speed of decoding plus the speed of
|
|
building the decoding tables. See the comments below that precede the
|
|
lbits and dbits tuning parameters.
|
|
|
|
GRR: return values(?)
|
|
0 OK
|
|
1 incomplete table
|
|
2 bad input
|
|
3 not enough memory
|
|
the following return codes are passed through from FLUSH() errors
|
|
50 (PK_DISK) "overflow of output space"
|
|
80 (IZ_CTRLC) "canceled by user's request"
|
|
*/
|
|
|
|
|
|
/*
|
|
Notes beyond the 1.93a appnote.txt:
|
|
|
|
1. Distance pointers never point before the beginning of the output
|
|
stream.
|
|
2. Distance pointers can point back across blocks, up to 32k away.
|
|
3. There is an implied maximum of 7 bits for the bit length table and
|
|
15 bits for the actual data.
|
|
4. If only one code exists, then it is encoded using one bit. (Zero
|
|
would be more efficient, but perhaps a little confusing.) If two
|
|
codes exist, they are coded using one bit each (0 and 1).
|
|
5. There is no way of sending zero distance codes--a dummy must be
|
|
sent if there are none. (History: a pre 2.0 version of PKZIP would
|
|
store blocks with no distance codes, but this was discovered to be
|
|
too harsh a criterion.) Valid only for 1.93a. 2.04c does allow
|
|
zero distance codes, which is sent as one code of zero bits in
|
|
length.
|
|
6. There are up to 286 literal/length codes. Code 256 represents the
|
|
end-of-block. Note however that the static length tree defines
|
|
288 codes just to fill out the Huffman codes. Codes 286 and 287
|
|
cannot be used though, since there is no length base or extra bits
|
|
defined for them. Similarily, there are up to 30 distance codes.
|
|
However, static trees define 32 codes (all 5 bits) to fill out the
|
|
Huffman codes, but the last two had better not show up in the data.
|
|
7. Unzip can check dynamic Huffman blocks for complete code sets.
|
|
The exception is that a single code would not be complete (see #4).
|
|
8. The five bits following the block type is really the number of
|
|
literal codes sent minus 257.
|
|
9. Length codes 8,16,16 are interpreted as 13 length codes of 8 bits
|
|
(1+6+6). Therefore, to output three times the length, you output
|
|
three codes (1+1+1), whereas to output four times the same length,
|
|
you only need two codes (1+3). Hmm.
|
|
10. In the tree reconstruction algorithm, Code = Code + Increment
|
|
only if BitLength(i) is not zero. (Pretty obvious.)
|
|
11. Correction: 4 Bits: # of Bit Length codes - 4 (4 - 19)
|
|
12. Note: length code 284 can represent 227-258, but length code 285
|
|
really is 258. The last length deserves its own, short code
|
|
since it gets used a lot in very redundant files. The length
|
|
258 is special since 258 - 3 (the min match length) is 255.
|
|
13. The literal/length and distance code bit lengths are read as a
|
|
single stream of lengths. It is possible (and advantageous) for
|
|
a repeat code (16, 17, or 18) to go across the boundary between
|
|
the two sets of lengths.
|
|
14. The Deflate64 (PKZIP method 9) variant of the compression algorithm
|
|
differs from "classic" deflate in the following 3 aspect:
|
|
a) The size of the sliding history window is expanded to 64 kByte.
|
|
b) The previously unused distance codes #30 and #31 code distances
|
|
from 32769 to 49152 and 49153 to 65536. Both codes take 14 bits
|
|
of extra data to determine the exact position in their 16 kByte
|
|
range.
|
|
c) The last lit/length code #285 gets a different meaning. Instead
|
|
of coding a fixed maximum match length of 258, it is used as a
|
|
"generic" match length code, capable of coding any length from
|
|
3 (min match length + 0) to 65538 (min match length + 65535).
|
|
This means that the length code #285 takes 16 bits (!) of uncoded
|
|
extra data, added to a fixed min length of 3.
|
|
Changes a) and b) would have been transparent for valid deflated
|
|
data, but change c) requires to switch decoder configurations between
|
|
Deflate and Deflate64 modes.
|
|
*/
|
|
|
|
|
|
#define PKZIP_BUG_WORKAROUND /* PKZIP 1.93a problem--live with it */
|
|
|
|
/*
|
|
inflate.h must supply the uch slide[WSIZE] array, the zvoid typedef
|
|
(void if (void *) is accepted, else char) and the NEXTBYTE,
|
|
FLUSH() and memzero macros. If the window size is not 32K, it
|
|
should also define WSIZE. If INFMOD is defined, it can include
|
|
compiled functions to support the NEXTBYTE and/or FLUSH() macros.
|
|
There are defaults for NEXTBYTE and FLUSH() below for use as
|
|
examples of what those functions need to do. Normally, you would
|
|
also want FLUSH() to compute a crc on the data. inflate.h also
|
|
needs to provide these typedefs:
|
|
|
|
typedef unsigned char uch;
|
|
typedef unsigned short ush;
|
|
typedef unsigned long ulg;
|
|
|
|
This module uses the external functions malloc() and free() (and
|
|
probably memset() or bzero() in the memzero() macro). Their
|
|
prototypes are normally found in <string.h> and <stdlib.h>.
|
|
*/
|
|
|
|
#define __INFLATE_C /* identifies this source module */
|
|
|
|
/* #define DEBUG */
|
|
#define INFMOD /* tell inflate.h to include code to be compiled */
|
|
#include "inflate.h"
|
|
|
|
|
|
/* marker for "unused" huft code, and corresponding check macro */
|
|
#define INVALID_CODE 99
|
|
#define IS_INVALID_CODE(c) ((c) == INVALID_CODE)
|
|
|
|
#ifndef WSIZE /* default is 32K resp. 64K */
|
|
# ifdef USE_DEFLATE64
|
|
# define WSIZE 65536L /* window size--must be a power of two, and */
|
|
# else /* at least 64K for PKZip's deflate64 method */
|
|
# define WSIZE 0x8000 /* window size--must be a power of two, and */
|
|
# endif /* at least 32K for zip's deflate method */
|
|
#endif
|
|
|
|
/* some buffer counters must be capable of holding 64k for Deflate64 */
|
|
#if (defined(USE_DEFLATE64) && defined(INT_16BIT))
|
|
# define UINT_D64 ulg
|
|
#else
|
|
# define UINT_D64 unsigned
|
|
#endif
|
|
|
|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
|
|
# define wsize G._wsize /* wsize is a variable */
|
|
#else
|
|
# define wsize WSIZE /* wsize is a constant */
|
|
#endif
|
|
|
|
|
|
#ifndef NEXTBYTE /* default is to simply get a byte from stdin */
|
|
# define NEXTBYTE getchar()
|
|
#endif
|
|
|
|
#ifndef MESSAGE /* only used twice, for fixed strings--NOT general-purpose */
|
|
# define MESSAGE(str,len,flag) fprintf(stderr,(char *)(str))
|
|
#endif
|
|
|
|
#ifndef FLUSH /* default is to simply write the buffer to stdout */
|
|
# define FLUSH(n) \
|
|
(((extent)fwrite(redirSlide, 1, (extent)(n), stdout) == (extent)(n)) ? \
|
|
0 : PKDISK)
|
|
#endif
|
|
/* Warning: the fwrite above might not work on 16-bit compilers, since
|
|
0x8000 might be interpreted as -32,768 by the library function. When
|
|
support for Deflate64 is enabled, the window size is 64K and the
|
|
simple fwrite statement is definitely broken for 16-bit compilers. */
|
|
|
|
#ifndef Trace
|
|
# ifdef DEBUG
|
|
# define Trace(x) fprintf x
|
|
# else
|
|
# define Trace(x)
|
|
# endif
|
|
#endif
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
#ifdef USE_ZLIB
|
|
|
|
/* Beginning with zlib version 1.2.0, a new inflate callback interface is
|
|
provided that allows tighter integration of the zlib inflate service
|
|
into unzip's extraction framework.
|
|
The advantages are:
|
|
- uses the windows buffer supplied by the unzip code; this saves one
|
|
copy process between zlib's internal decompression buffer and unzip's
|
|
post-decompression output buffer and improves performance.
|
|
- does not pull in unused checksum code (adler32).
|
|
The preprocessor flag NO_ZLIBCALLBCK can be set to force usage of the
|
|
old zlib 1.1.x interface, for testing purpose.
|
|
*/
|
|
#ifdef USE_ZLIB_INFLATCB
|
|
# undef USE_ZLIB_INFLATCB
|
|
#endif
|
|
#if (defined(ZLIB_VERNUM) && ZLIB_VERNUM >= 0x1200 && !defined(NO_ZLIBCALLBCK))
|
|
# define USE_ZLIB_INFLATCB 1
|
|
#else
|
|
# define USE_ZLIB_INFLATCB 0
|
|
#endif
|
|
|
|
/* Check for incompatible combinations of zlib and Deflate64 support. */
|
|
#if defined(USE_DEFLATE64)
|
|
# if !USE_ZLIB_INFLATCB
|
|
#error Deflate64 is incompatible with traditional (pre-1.2.x) zlib interface!
|
|
# else
|
|
/* The Deflate64 callback function in the framework of zlib 1.2.x requires
|
|
the inclusion of the unsupported infback9 header file:
|
|
*/
|
|
# include "infback9.h"
|
|
# endif
|
|
#endif /* USE_DEFLATE64 */
|
|
|
|
|
|
#if USE_ZLIB_INFLATCB
|
|
|
|
static unsigned zlib_inCB OF((void FAR *pG, unsigned char FAR * FAR * pInbuf));
|
|
static int zlib_outCB OF((void FAR *pG, unsigned char FAR *outbuf,
|
|
unsigned outcnt));
|
|
|
|
static unsigned zlib_inCB(pG, pInbuf)
|
|
void FAR *pG;
|
|
unsigned char FAR * FAR * pInbuf;
|
|
{
|
|
*pInbuf = G.inbuf;
|
|
return fillinbuf(__G);
|
|
}
|
|
|
|
static int zlib_outCB(pG, outbuf, outcnt)
|
|
void FAR *pG;
|
|
unsigned char FAR *outbuf;
|
|
unsigned outcnt;
|
|
{
|
|
#ifdef FUNZIP
|
|
return flush(__G__ (ulg)(outcnt));
|
|
#else
|
|
return ((G.mem_mode) ? memflush(__G__ outbuf, (ulg)(outcnt))
|
|
: flush(__G__ outbuf, (ulg)(outcnt), 0));
|
|
#endif
|
|
}
|
|
#endif /* USE_ZLIB_INFLATCB */
|
|
|
|
|
|
/*
|
|
GRR: return values for both original inflate() and UZinflate()
|
|
0 OK
|
|
1 incomplete table(?)
|
|
2 bad input
|
|
3 not enough memory
|
|
*/
|
|
|
|
/**************************/
|
|
/* Function UZinflate() */
|
|
/**************************/
|
|
|
|
int UZinflate(__G__ is_defl64)
|
|
__GDEF
|
|
int is_defl64;
|
|
/* decompress an inflated entry using the zlib routines */
|
|
{
|
|
int retval = 0; /* return code: 0 = "no error" */
|
|
int err=Z_OK;
|
|
#if USE_ZLIB_INFLATCB
|
|
|
|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
|
|
if (G.redirect_slide)
|
|
wsize = G.redirect_size, redirSlide = G.redirect_buffer;
|
|
else
|
|
wsize = WSIZE, redirSlide = slide;
|
|
#endif
|
|
|
|
if (!G.inflInit) {
|
|
/* local buffer for efficiency */
|
|
ZCONST char *zlib_RtVersion = zlibVersion();
|
|
|
|
/* only need to test this stuff once */
|
|
if ((zlib_RtVersion[0] != ZLIB_VERSION[0]) ||
|
|
(zlib_RtVersion[2] != ZLIB_VERSION[2])) {
|
|
Info(slide, 0x21, ((char *)slide,
|
|
"error: incompatible zlib version (expected %s, found %s)\n",
|
|
ZLIB_VERSION, zlib_RtVersion));
|
|
return 3;
|
|
} else if (strcmp(zlib_RtVersion, ZLIB_VERSION) != 0)
|
|
Info(slide, 0x21, ((char *)slide,
|
|
"warning: different zlib version (expected %s, using %s)\n",
|
|
ZLIB_VERSION, zlib_RtVersion));
|
|
|
|
G.dstrm.zalloc = (alloc_func)Z_NULL;
|
|
G.dstrm.zfree = (free_func)Z_NULL;
|
|
|
|
G.inflInit = 1;
|
|
}
|
|
|
|
#ifdef USE_DEFLATE64
|
|
if (is_defl64)
|
|
{
|
|
Trace((stderr, "initializing inflate9()\n"));
|
|
err = inflateBack9Init(&G.dstrm, redirSlide);
|
|
|
|
if (err == Z_MEM_ERROR)
|
|
return 3;
|
|
else if (err != Z_OK) {
|
|
Trace((stderr, "oops! (inflateBack9Init() err = %d)\n", err));
|
|
return 2;
|
|
}
|
|
|
|
G.dstrm.next_in = G.inptr;
|
|
G.dstrm.avail_in = G.incnt;
|
|
|
|
err = inflateBack9(&G.dstrm, zlib_inCB, &G, zlib_outCB, &G);
|
|
if (err != Z_STREAM_END) {
|
|
if (err == Z_DATA_ERROR || err == Z_STREAM_ERROR) {
|
|
Trace((stderr, "oops! (inflateBack9() err = %d)\n", err));
|
|
retval = 2;
|
|
} else if (err == Z_MEM_ERROR) {
|
|
retval = 3;
|
|
} else if (err == Z_BUF_ERROR) {
|
|
Trace((stderr, "oops! (inflateBack9() err = %d)\n", err));
|
|
if (G.dstrm.next_in == Z_NULL) {
|
|
/* input failure */
|
|
Trace((stderr, " inflateBack9() input failure\n"));
|
|
retval = 2;
|
|
} else {
|
|
/* output write failure */
|
|
retval = (G.disk_full != 0 ? PK_DISK : IZ_CTRLC);
|
|
}
|
|
} else {
|
|
Trace((stderr, "oops! (inflateBack9() err = %d)\n", err));
|
|
retval = 2;
|
|
}
|
|
}
|
|
if (G.dstrm.next_in != NULL) {
|
|
G.inptr = (uch *)G.dstrm.next_in;
|
|
G.incnt = G.dstrm.avail_in;
|
|
}
|
|
|
|
err = inflateBack9End(&G.dstrm);
|
|
if (err != Z_OK) {
|
|
Trace((stderr, "oops! (inflateBack9End() err = %d)\n", err));
|
|
if (retval == 0)
|
|
retval = 2;
|
|
}
|
|
}
|
|
else
|
|
#endif /* USE_DEFLATE64 */
|
|
{
|
|
/* For the callback interface, inflate initialization has to
|
|
be called before each decompression call.
|
|
*/
|
|
{
|
|
unsigned i;
|
|
int windowBits;
|
|
/* windowBits = log2(wsize) */
|
|
for (i = (unsigned)wsize, windowBits = 0;
|
|
!(i & 1); i >>= 1, ++windowBits);
|
|
if ((unsigned)windowBits > (unsigned)15)
|
|
windowBits = 15;
|
|
else if (windowBits < 8)
|
|
windowBits = 8;
|
|
|
|
Trace((stderr, "initializing inflate()\n"));
|
|
err = inflateBackInit(&G.dstrm, windowBits, redirSlide);
|
|
|
|
if (err == Z_MEM_ERROR)
|
|
return 3;
|
|
else if (err != Z_OK) {
|
|
Trace((stderr, "oops! (inflateBackInit() err = %d)\n", err));
|
|
return 2;
|
|
}
|
|
}
|
|
|
|
G.dstrm.next_in = G.inptr;
|
|
G.dstrm.avail_in = G.incnt;
|
|
|
|
err = inflateBack(&G.dstrm, zlib_inCB, &G, zlib_outCB, &G);
|
|
if (err != Z_STREAM_END) {
|
|
if (err == Z_DATA_ERROR || err == Z_STREAM_ERROR) {
|
|
Trace((stderr, "oops! (inflateBack() err = %d)\n", err));
|
|
retval = 2;
|
|
} else if (err == Z_MEM_ERROR) {
|
|
retval = 3;
|
|
} else if (err == Z_BUF_ERROR) {
|
|
Trace((stderr, "oops! (inflateBack() err = %d)\n", err));
|
|
if (G.dstrm.next_in == Z_NULL) {
|
|
/* input failure */
|
|
Trace((stderr, " inflateBack() input failure\n"));
|
|
retval = 2;
|
|
} else {
|
|
/* output write failure */
|
|
retval = (G.disk_full != 0 ? PK_DISK : IZ_CTRLC);
|
|
}
|
|
} else {
|
|
Trace((stderr, "oops! (inflateBack() err = %d)\n", err));
|
|
retval = 2;
|
|
}
|
|
}
|
|
if (G.dstrm.next_in != NULL) {
|
|
G.inptr = (uch *)G.dstrm.next_in;
|
|
G.incnt = G.dstrm.avail_in;
|
|
}
|
|
|
|
err = inflateBackEnd(&G.dstrm);
|
|
if (err != Z_OK) {
|
|
Trace((stderr, "oops! (inflateBackEnd() err = %d)\n", err));
|
|
if (retval == 0)
|
|
retval = 2;
|
|
}
|
|
}
|
|
|
|
#else /* !USE_ZLIB_INFLATCB */
|
|
int repeated_buf_err;
|
|
|
|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
|
|
if (G.redirect_slide)
|
|
wsize = G.redirect_size, redirSlide = G.redirect_buffer;
|
|
else
|
|
wsize = WSIZE, redirSlide = slide;
|
|
#endif
|
|
|
|
G.dstrm.next_out = redirSlide;
|
|
G.dstrm.avail_out = wsize;
|
|
|
|
G.dstrm.next_in = G.inptr;
|
|
G.dstrm.avail_in = G.incnt;
|
|
|
|
if (!G.inflInit) {
|
|
unsigned i;
|
|
int windowBits;
|
|
/* local buffer for efficiency */
|
|
ZCONST char *zlib_RtVersion = zlibVersion();
|
|
|
|
/* only need to test this stuff once */
|
|
if (zlib_RtVersion[0] != ZLIB_VERSION[0]) {
|
|
Info(slide, 0x21, ((char *)slide,
|
|
"error: incompatible zlib version (expected %s, found %s)\n",
|
|
ZLIB_VERSION, zlib_RtVersion));
|
|
return 3;
|
|
} else if (strcmp(zlib_RtVersion, ZLIB_VERSION) != 0)
|
|
Info(slide, 0x21, ((char *)slide,
|
|
"warning: different zlib version (expected %s, using %s)\n",
|
|
ZLIB_VERSION, zlib_RtVersion));
|
|
|
|
/* windowBits = log2(wsize) */
|
|
for (i = (unsigned)wsize, windowBits = 0;
|
|
!(i & 1); i >>= 1, ++windowBits);
|
|
if ((unsigned)windowBits > (unsigned)15)
|
|
windowBits = 15;
|
|
else if (windowBits < 8)
|
|
windowBits = 8;
|
|
|
|
G.dstrm.zalloc = (alloc_func)Z_NULL;
|
|
G.dstrm.zfree = (free_func)Z_NULL;
|
|
|
|
Trace((stderr, "initializing inflate()\n"));
|
|
err = inflateInit2(&G.dstrm, -windowBits);
|
|
|
|
if (err == Z_MEM_ERROR)
|
|
return 3;
|
|
else if (err != Z_OK)
|
|
Trace((stderr, "oops! (inflateInit2() err = %d)\n", err));
|
|
G.inflInit = 1;
|
|
}
|
|
|
|
#ifdef FUNZIP
|
|
while (err != Z_STREAM_END) {
|
|
#else /* !FUNZIP */
|
|
while (G.csize > 0) {
|
|
Trace((stderr, "first loop: G.csize = %ld\n", G.csize));
|
|
#endif /* ?FUNZIP */
|
|
while (G.dstrm.avail_out > 0) {
|
|
err = inflate(&G.dstrm, Z_PARTIAL_FLUSH);
|
|
|
|
if (err == Z_DATA_ERROR) {
|
|
retval = 2; goto uzinflate_cleanup_exit;
|
|
} else if (err == Z_MEM_ERROR) {
|
|
retval = 3; goto uzinflate_cleanup_exit;
|
|
} else if (err != Z_OK && err != Z_STREAM_END)
|
|
Trace((stderr, "oops! (inflate(first loop) err = %d)\n", err));
|
|
|
|
#ifdef FUNZIP
|
|
if (err == Z_STREAM_END) /* "END-of-entry-condition" ? */
|
|
#else /* !FUNZIP */
|
|
if (G.csize <= 0L) /* "END-of-entry-condition" ? */
|
|
#endif /* ?FUNZIP */
|
|
break;
|
|
|
|
if (G.dstrm.avail_in == 0) {
|
|
if (fillinbuf(__G) == 0) {
|
|
/* no "END-condition" yet, but no more data */
|
|
retval = 2; goto uzinflate_cleanup_exit;
|
|
}
|
|
|
|
G.dstrm.next_in = G.inptr;
|
|
G.dstrm.avail_in = G.incnt;
|
|
}
|
|
Trace((stderr, " avail_in = %u\n", G.dstrm.avail_in));
|
|
}
|
|
/* flush slide[] */
|
|
if ((retval = FLUSH(wsize - G.dstrm.avail_out)) != 0)
|
|
goto uzinflate_cleanup_exit;
|
|
Trace((stderr, "inside loop: flushing %ld bytes (ptr diff = %ld)\n",
|
|
(long)(wsize - G.dstrm.avail_out),
|
|
(long)(G.dstrm.next_out-(Bytef *)redirSlide)));
|
|
G.dstrm.next_out = redirSlide;
|
|
G.dstrm.avail_out = wsize;
|
|
}
|
|
|
|
/* no more input, so loop until we have all output */
|
|
Trace((stderr, "beginning final loop: err = %d\n", err));
|
|
repeated_buf_err = FALSE;
|
|
while (err != Z_STREAM_END) {
|
|
err = inflate(&G.dstrm, Z_PARTIAL_FLUSH);
|
|
if (err == Z_DATA_ERROR) {
|
|
retval = 2; goto uzinflate_cleanup_exit;
|
|
} else if (err == Z_MEM_ERROR) {
|
|
retval = 3; goto uzinflate_cleanup_exit;
|
|
} else if (err == Z_BUF_ERROR) { /* DEBUG */
|
|
#ifdef FUNZIP
|
|
Trace((stderr,
|
|
"zlib inflate() did not detect stream end\n"));
|
|
#else
|
|
Trace((stderr,
|
|
"zlib inflate() did not detect stream end (%s, %s)\n",
|
|
G.zipfn, G.filename));
|
|
#endif
|
|
if ((!repeated_buf_err) && (G.dstrm.avail_in == 0)) {
|
|
/* when detecting this problem for the first time,
|
|
try to provide one fake byte beyond "EOF"... */
|
|
G.dstrm.next_in = "";
|
|
G.dstrm.avail_in = 1;
|
|
repeated_buf_err = TRUE;
|
|
} else
|
|
break;
|
|
} else if (err != Z_OK && err != Z_STREAM_END) {
|
|
Trace((stderr, "oops! (inflate(final loop) err = %d)\n", err));
|
|
DESTROYGLOBALS();
|
|
EXIT(PK_MEM3);
|
|
}
|
|
/* final flush of slide[] */
|
|
if ((retval = FLUSH(wsize - G.dstrm.avail_out)) != 0)
|
|
goto uzinflate_cleanup_exit;
|
|
Trace((stderr, "final loop: flushing %ld bytes (ptr diff = %ld)\n",
|
|
(long)(wsize - G.dstrm.avail_out),
|
|
(long)(G.dstrm.next_out-(Bytef *)redirSlide)));
|
|
G.dstrm.next_out = redirSlide;
|
|
G.dstrm.avail_out = wsize;
|
|
}
|
|
Trace((stderr, "total in = %lu, total out = %lu\n", G.dstrm.total_in,
|
|
G.dstrm.total_out));
|
|
|
|
G.inptr = (uch *)G.dstrm.next_in;
|
|
G.incnt = (G.inbuf + INBUFSIZ) - G.inptr; /* reset for other routines */
|
|
|
|
uzinflate_cleanup_exit:
|
|
err = inflateReset(&G.dstrm);
|
|
if (err != Z_OK)
|
|
Trace((stderr, "oops! (inflateReset() err = %d)\n", err));
|
|
|
|
#endif /* ?USE_ZLIB_INFLATCB */
|
|
return retval;
|
|
}
|
|
|
|
|
|
/*---------------------------------------------------------------------------*/
|
|
#else /* !USE_ZLIB */
|
|
|
|
|
|
/* Function prototypes */
|
|
#ifndef OF
|
|
# ifdef __STDC__
|
|
# define OF(a) a
|
|
# else
|
|
# define OF(a) ()
|
|
# endif
|
|
#endif /* !OF */
|
|
int inflate_codes OF((__GPRO__ struct huft *tl, struct huft *td,
|
|
unsigned bl, unsigned bd));
|
|
static int inflate_stored OF((__GPRO));
|
|
static int inflate_fixed OF((__GPRO));
|
|
static int inflate_dynamic OF((__GPRO));
|
|
static int inflate_block OF((__GPRO__ int *e));
|
|
|
|
|
|
/* The inflate algorithm uses a sliding 32K byte window on the uncompressed
|
|
stream to find repeated byte strings. This is implemented here as a
|
|
circular buffer. The index is updated simply by incrementing and then
|
|
and'ing with 0x7fff (32K-1). */
|
|
/* It is left to other modules to supply the 32K area. It is assumed
|
|
to be usable as if it were declared "uch slide[32768];" or as just
|
|
"uch *slide;" and then malloc'ed in the latter case. The definition
|
|
must be in unzip.h, included above. */
|
|
|
|
|
|
/* unsigned wp; moved to globals.h */ /* current position in slide */
|
|
|
|
/* Tables for deflate from PKZIP's appnote.txt. */
|
|
/* - Order of the bit length code lengths */
|
|
static ZCONST unsigned border[] = {
|
|
16, 17, 18, 0, 8, 7, 9, 6, 10, 5, 11, 4, 12, 3, 13, 2, 14, 1, 15};
|
|
|
|
/* - Copy lengths for literal codes 257..285 */
|
|
#ifdef USE_DEFLATE64
|
|
static ZCONST ush cplens64[] = {
|
|
3, 4, 5, 6, 7, 8, 9, 10, 11, 13, 15, 17, 19, 23, 27, 31,
|
|
35, 43, 51, 59, 67, 83, 99, 115, 131, 163, 195, 227, 3, 0, 0};
|
|
/* For Deflate64, the code 285 is defined differently. */
|
|
#else
|
|
# define cplens32 cplens
|
|
#endif
|
|
static ZCONST ush cplens32[] = {
|
|
3, 4, 5, 6, 7, 8, 9, 10, 11, 13, 15, 17, 19, 23, 27, 31,
|
|
35, 43, 51, 59, 67, 83, 99, 115, 131, 163, 195, 227, 258, 0, 0};
|
|
/* note: see note #13 above about the 258 in this list. */
|
|
/* - Extra bits for literal codes 257..285 */
|
|
#ifdef USE_DEFLATE64
|
|
static ZCONST uch cplext64[] = {
|
|
0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 2, 2, 2, 2,
|
|
3, 3, 3, 3, 4, 4, 4, 4, 5, 5, 5, 5, 16, INVALID_CODE, INVALID_CODE};
|
|
#else
|
|
# define cplext32 cplext
|
|
#endif
|
|
static ZCONST uch cplext32[] = {
|
|
0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 2, 2, 2, 2,
|
|
3, 3, 3, 3, 4, 4, 4, 4, 5, 5, 5, 5, 0, INVALID_CODE, INVALID_CODE};
|
|
|
|
/* - Copy offsets for distance codes 0..29 (0..31 for Deflate64) */
|
|
static ZCONST ush cpdist[] = {
|
|
1, 2, 3, 4, 5, 7, 9, 13, 17, 25, 33, 49, 65, 97, 129, 193,
|
|
257, 385, 513, 769, 1025, 1537, 2049, 3073, 4097, 6145,
|
|
#if (defined(USE_DEFLATE64) || defined(PKZIP_BUG_WORKAROUND))
|
|
8193, 12289, 16385, 24577, 32769, 49153};
|
|
#else
|
|
8193, 12289, 16385, 24577};
|
|
#endif
|
|
|
|
/* - Extra bits for distance codes 0..29 (0..31 for Deflate64) */
|
|
#ifdef USE_DEFLATE64
|
|
static ZCONST uch cpdext64[] = {
|
|
0, 0, 0, 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 6, 6,
|
|
7, 7, 8, 8, 9, 9, 10, 10, 11, 11,
|
|
12, 12, 13, 13, 14, 14};
|
|
#else
|
|
# define cpdext32 cpdext
|
|
#endif
|
|
static ZCONST uch cpdext32[] = {
|
|
0, 0, 0, 0, 1, 1, 2, 2, 3, 3, 4, 4, 5, 5, 6, 6,
|
|
7, 7, 8, 8, 9, 9, 10, 10, 11, 11,
|
|
#ifdef PKZIP_BUG_WORKAROUND
|
|
12, 12, 13, 13, INVALID_CODE, INVALID_CODE};
|
|
#else
|
|
12, 12, 13, 13};
|
|
#endif
|
|
|
|
#ifdef PKZIP_BUG_WORKAROUND
|
|
# define MAXLITLENS 288
|
|
#else
|
|
# define MAXLITLENS 286
|
|
#endif
|
|
#if (defined(USE_DEFLATE64) || defined(PKZIP_BUG_WORKAROUND))
|
|
# define MAXDISTS 32
|
|
#else
|
|
# define MAXDISTS 30
|
|
#endif
|
|
|
|
|
|
/* moved to consts.h (included in unzip.c), resp. funzip.c */
|
|
#if 0
|
|
/* And'ing with mask_bits[n] masks the lower n bits */
|
|
ZCONST unsigned near mask_bits[17] = {
|
|
0x0000,
|
|
0x0001, 0x0003, 0x0007, 0x000f, 0x001f, 0x003f, 0x007f, 0x00ff,
|
|
0x01ff, 0x03ff, 0x07ff, 0x0fff, 0x1fff, 0x3fff, 0x7fff, 0xffff
|
|
};
|
|
#endif /* 0 */
|
|
|
|
|
|
/* Macros for inflate() bit peeking and grabbing.
|
|
The usage is:
|
|
|
|
NEEDBITS(j)
|
|
x = b & mask_bits[j];
|
|
DUMPBITS(j)
|
|
|
|
where NEEDBITS makes sure that b has at least j bits in it, and
|
|
DUMPBITS removes the bits from b. The macros use the variable k
|
|
for the number of bits in b. Normally, b and k are register
|
|
variables for speed and are initialized at the beginning of a
|
|
routine that uses these macros from a global bit buffer and count.
|
|
|
|
In order to not ask for more bits than there are in the compressed
|
|
stream, the Huffman tables are constructed to only ask for just
|
|
enough bits to make up the end-of-block code (value 256). Then no
|
|
bytes need to be "returned" to the buffer at the end of the last
|
|
block. See the huft_build() routine.
|
|
|
|
Actually, the precautions mentioned above are not sufficient to
|
|
prevent fetches of bits beyound the end of the last block in every
|
|
case. When the last code fetched before the end-of-block code was
|
|
a very short distance code (shorter than "distance-prefetch-bits" -
|
|
"end-of-block code bits"), this last distance code fetch already
|
|
exausts the available data. To prevent failure of extraction in this
|
|
case, the "read beyond EOF" check delays the raise of the "invalid
|
|
data" error until an actual overflow of "used data" is detected.
|
|
This error condition is only fulfilled when the "number of available
|
|
bits" counter k is found to be negative in the NEEDBITS() macro.
|
|
|
|
An alternate fix for that problem adjusts the size of the distance code
|
|
base table so that it does not exceed the length of the end-of-block code
|
|
plus the minimum length of a distance code. This alternate fix can be
|
|
enabled by defining the preprocessor symbol FIX_PAST_EOB_BY_TABLEADJUST.
|
|
*/
|
|
|
|
/* These have been moved to globals.h */
|
|
#if 0
|
|
ulg bb; /* bit buffer */
|
|
unsigned bk; /* bits in bit buffer */
|
|
#endif
|
|
|
|
#ifndef CHECK_EOF
|
|
# define CHECK_EOF /* default as of 5.13/5.2 */
|
|
#endif
|
|
|
|
#ifndef CHECK_EOF
|
|
# define NEEDBITS(n) {while(k<(n)){b|=((ulg)NEXTBYTE)<<k;k+=8;}}
|
|
#else
|
|
# ifdef FIX_PAST_EOB_BY_TABLEADJUST
|
|
# define NEEDBITS(n) {while(k<(n)){int c=NEXTBYTE;\
|
|
if(c==EOF){retval=1;goto cleanup_and_exit;}\
|
|
b|=((ulg)c)<<k;k+=8;}}
|
|
# else
|
|
# define NEEDBITS(n) {while((int)k<(int)(n)){int c=NEXTBYTE;\
|
|
if(c==EOF){if((int)k>=0)break;retval=1;goto cleanup_and_exit;}\
|
|
b|=((ulg)c)<<k;k+=8;}}
|
|
# endif
|
|
#endif
|
|
|
|
#define DUMPBITS(n) {b>>=(n);k-=(n);}
|
|
|
|
|
|
/*
|
|
Huffman code decoding is performed using a multi-level table lookup.
|
|
The fastest way to decode is to simply build a lookup table whose
|
|
size is determined by the longest code. However, the time it takes
|
|
to build this table can also be a factor if the data being decoded
|
|
are not very long. The most common codes are necessarily the
|
|
shortest codes, so those codes dominate the decoding time, and hence
|
|
the speed. The idea is you can have a shorter table that decodes the
|
|
shorter, more probable codes, and then point to subsidiary tables for
|
|
the longer codes. The time it costs to decode the longer codes is
|
|
then traded against the time it takes to make longer tables.
|
|
|
|
This results of this trade are in the variables lbits and dbits
|
|
below. lbits is the number of bits the first level table for literal/
|
|
length codes can decode in one step, and dbits is the same thing for
|
|
the distance codes. Subsequent tables are also less than or equal to
|
|
those sizes. These values may be adjusted either when all of the
|
|
codes are shorter than that, in which case the longest code length in
|
|
bits is used, or when the shortest code is *longer* than the requested
|
|
table size, in which case the length of the shortest code in bits is
|
|
used.
|
|
|
|
There are two different values for the two tables, since they code a
|
|
different number of possibilities each. The literal/length table
|
|
codes 286 possible values, or in a flat code, a little over eight
|
|
bits. The distance table codes 30 possible values, or a little less
|
|
than five bits, flat. The optimum values for speed end up being
|
|
about one bit more than those, so lbits is 8+1 and dbits is 5+1.
|
|
The optimum values may differ though from machine to machine, and
|
|
possibly even between compilers. Your mileage may vary.
|
|
*/
|
|
|
|
|
|
/* bits in base literal/length lookup table */
|
|
static ZCONST unsigned lbits = 9;
|
|
/* bits in base distance lookup table */
|
|
static ZCONST unsigned dbits = 6;
|
|
|
|
|
|
#ifndef ASM_INFLATECODES
|
|
|
|
int inflate_codes(__G__ tl, td, bl, bd)
|
|
__GDEF
|
|
struct huft *tl, *td; /* literal/length and distance decoder tables */
|
|
unsigned bl, bd; /* number of bits decoded by tl[] and td[] */
|
|
/* inflate (decompress) the codes in a deflated (compressed) block.
|
|
Return an error code or zero if it all goes ok. */
|
|
{
|
|
register unsigned e; /* table entry flag/number of extra bits */
|
|
unsigned d; /* index for copy */
|
|
UINT_D64 n; /* length for copy (deflate64: might be 64k+2) */
|
|
UINT_D64 w; /* current window position (deflate64: up to 64k) */
|
|
struct huft *t; /* pointer to table entry */
|
|
unsigned ml, md; /* masks for bl and bd bits */
|
|
register ulg b; /* bit buffer */
|
|
register unsigned k; /* number of bits in bit buffer */
|
|
int retval = 0; /* error code returned: initialized to "no error" */
|
|
|
|
|
|
/* make local copies of globals */
|
|
b = G.bb; /* initialize bit buffer */
|
|
k = G.bk;
|
|
w = G.wp; /* initialize window position */
|
|
|
|
|
|
/* inflate the coded data */
|
|
ml = mask_bits[bl]; /* precompute masks for speed */
|
|
md = mask_bits[bd];
|
|
while (1) /* do until end of block */
|
|
{
|
|
NEEDBITS(bl)
|
|
t = tl + ((unsigned)b & ml);
|
|
while (1) {
|
|
DUMPBITS(t->b)
|
|
|
|
if ((e = t->e) == 32) /* then it's a literal */
|
|
{
|
|
redirSlide[w++] = (uch)t->v.n;
|
|
if (w == wsize)
|
|
{
|
|
if ((retval = FLUSH(w)) != 0) goto cleanup_and_exit;
|
|
w = 0;
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (e < 31) /* then it's a length */
|
|
{
|
|
/* get length of block to copy */
|
|
NEEDBITS(e)
|
|
n = t->v.n + ((unsigned)b & mask_bits[e]);
|
|
DUMPBITS(e)
|
|
|
|
/* decode distance of block to copy */
|
|
NEEDBITS(bd)
|
|
t = td + ((unsigned)b & md);
|
|
while (1) {
|
|
DUMPBITS(t->b)
|
|
if ((e = t->e) < 32)
|
|
break;
|
|
if (IS_INVALID_CODE(e))
|
|
return 1;
|
|
e &= 31;
|
|
NEEDBITS(e)
|
|
t = t->v.t + ((unsigned)b & mask_bits[e]);
|
|
}
|
|
NEEDBITS(e)
|
|
d = (unsigned)w - t->v.n - ((unsigned)b & mask_bits[e]);
|
|
DUMPBITS(e)
|
|
|
|
/* do the copy */
|
|
do {
|
|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
|
|
if (G.redirect_slide) {
|
|
/* &= w/ wsize unnecessary & wrong if redirect */
|
|
if ((UINT_D64)d >= wsize)
|
|
return 1; /* invalid compressed data */
|
|
e = (unsigned)(wsize - (d > (unsigned)w ? (UINT_D64)d : w));
|
|
}
|
|
else
|
|
#endif
|
|
e = (unsigned)(wsize -
|
|
((d &= (unsigned)(wsize-1)) > (unsigned)w ?
|
|
(UINT_D64)d : w));
|
|
if ((UINT_D64)e > n) e = (unsigned)n;
|
|
n -= e;
|
|
#ifndef NOMEMCPY
|
|
if ((unsigned)w - d >= e)
|
|
/* (this test assumes unsigned comparison) */
|
|
{
|
|
memcpy(redirSlide + (unsigned)w, redirSlide + d, e);
|
|
w += e;
|
|
d += e;
|
|
}
|
|
else /* do it slowly to avoid memcpy() overlap */
|
|
#endif /* !NOMEMCPY */
|
|
do {
|
|
redirSlide[w++] = redirSlide[d++];
|
|
} while (--e);
|
|
if (w == wsize)
|
|
{
|
|
if ((retval = FLUSH(w)) != 0) goto cleanup_and_exit;
|
|
w = 0;
|
|
}
|
|
} while (n);
|
|
break;
|
|
}
|
|
|
|
if (e == 31) /* it's the EOB signal */
|
|
{
|
|
/* sorry for this goto, but we have to exit two loops at once */
|
|
goto cleanup_decode;
|
|
}
|
|
|
|
if (IS_INVALID_CODE(e))
|
|
return 1;
|
|
|
|
e &= 31;
|
|
NEEDBITS(e)
|
|
t = t->v.t + ((unsigned)b & mask_bits[e]);
|
|
}
|
|
}
|
|
cleanup_decode:
|
|
|
|
/* restore the globals from the locals */
|
|
G.wp = (unsigned)w; /* restore global window pointer */
|
|
G.bb = b; /* restore global bit buffer */
|
|
G.bk = k;
|
|
|
|
|
|
cleanup_and_exit:
|
|
/* done */
|
|
return retval;
|
|
}
|
|
|
|
#endif /* ASM_INFLATECODES */
|
|
|
|
|
|
|
|
static int inflate_stored(__G)
|
|
__GDEF
|
|
/* "decompress" an inflated type 0 (stored) block. */
|
|
{
|
|
UINT_D64 w; /* current window position (deflate64: up to 64k!) */
|
|
unsigned n; /* number of bytes in block */
|
|
register ulg b; /* bit buffer */
|
|
register unsigned k; /* number of bits in bit buffer */
|
|
int retval = 0; /* error code returned: initialized to "no error" */
|
|
|
|
|
|
/* make local copies of globals */
|
|
Trace((stderr, "\nstored block"));
|
|
b = G.bb; /* initialize bit buffer */
|
|
k = G.bk;
|
|
w = G.wp; /* initialize window position */
|
|
|
|
|
|
/* go to byte boundary */
|
|
n = k & 7;
|
|
DUMPBITS(n);
|
|
|
|
|
|
/* get the length and its complement */
|
|
NEEDBITS(16)
|
|
n = ((unsigned)b & 0xffff);
|
|
DUMPBITS(16)
|
|
NEEDBITS(16)
|
|
if (n != (unsigned)((~b) & 0xffff))
|
|
return 1; /* error in compressed data */
|
|
DUMPBITS(16)
|
|
|
|
|
|
/* read and output the compressed data */
|
|
while (n--)
|
|
{
|
|
NEEDBITS(8)
|
|
redirSlide[w++] = (uch)b;
|
|
if (w == wsize)
|
|
{
|
|
if ((retval = FLUSH(w)) != 0) goto cleanup_and_exit;
|
|
w = 0;
|
|
}
|
|
DUMPBITS(8)
|
|
}
|
|
|
|
|
|
/* restore the globals from the locals */
|
|
G.wp = (unsigned)w; /* restore global window pointer */
|
|
G.bb = b; /* restore global bit buffer */
|
|
G.bk = k;
|
|
|
|
cleanup_and_exit:
|
|
return retval;
|
|
}
|
|
|
|
|
|
/* Globals for literal tables (built once) */
|
|
/* Moved to globals.h */
|
|
#if 0
|
|
struct huft *fixed_tl = (struct huft *)NULL;
|
|
struct huft *fixed_td;
|
|
int fixed_bl, fixed_bd;
|
|
#endif
|
|
|
|
static int inflate_fixed(__G)
|
|
__GDEF
|
|
/* decompress an inflated type 1 (fixed Huffman codes) block. We should
|
|
either replace this with a custom decoder, or at least precompute the
|
|
Huffman tables. */
|
|
{
|
|
/* if first time, set up tables for fixed blocks */
|
|
Trace((stderr, "\nliteral block"));
|
|
if (G.fixed_tl == (struct huft *)NULL)
|
|
{
|
|
int i; /* temporary variable */
|
|
unsigned l[288]; /* length list for huft_build */
|
|
|
|
/* literal table */
|
|
for (i = 0; i < 144; i++)
|
|
l[i] = 8;
|
|
for (; i < 256; i++)
|
|
l[i] = 9;
|
|
for (; i < 280; i++)
|
|
l[i] = 7;
|
|
for (; i < 288; i++) /* make a complete, but wrong code set */
|
|
l[i] = 8;
|
|
G.fixed_bl = 7;
|
|
#ifdef USE_DEFLATE64
|
|
if ((i = huft_build(__G__ l, 288, 257, G.cplens, G.cplext,
|
|
&G.fixed_tl, &G.fixed_bl)) != 0)
|
|
#else
|
|
if ((i = huft_build(__G__ l, 288, 257, cplens, cplext,
|
|
&G.fixed_tl, &G.fixed_bl)) != 0)
|
|
#endif
|
|
{
|
|
G.fixed_tl = (struct huft *)NULL;
|
|
return i;
|
|
}
|
|
|
|
/* distance table */
|
|
for (i = 0; i < MAXDISTS; i++) /* make an incomplete code set */
|
|
l[i] = 5;
|
|
G.fixed_bd = 5;
|
|
#ifdef USE_DEFLATE64
|
|
if ((i = huft_build(__G__ l, MAXDISTS, 0, cpdist, G.cpdext,
|
|
&G.fixed_td, &G.fixed_bd)) > 1)
|
|
#else
|
|
if ((i = huft_build(__G__ l, MAXDISTS, 0, cpdist, cpdext,
|
|
&G.fixed_td, &G.fixed_bd)) > 1)
|
|
#endif
|
|
{
|
|
huft_free(G.fixed_tl);
|
|
G.fixed_td = G.fixed_tl = (struct huft *)NULL;
|
|
return i;
|
|
}
|
|
}
|
|
|
|
/* decompress until an end-of-block code */
|
|
return inflate_codes(__G__ G.fixed_tl, G.fixed_td,
|
|
G.fixed_bl, G.fixed_bd);
|
|
}
|
|
|
|
|
|
|
|
static int inflate_dynamic(__G)
|
|
__GDEF
|
|
/* decompress an inflated type 2 (dynamic Huffman codes) block. */
|
|
{
|
|
unsigned i; /* temporary variables */
|
|
unsigned j;
|
|
unsigned l; /* last length */
|
|
unsigned m; /* mask for bit lengths table */
|
|
unsigned n; /* number of lengths to get */
|
|
struct huft *tl = (struct huft *)NULL; /* literal/length code table */
|
|
struct huft *td = (struct huft *)NULL; /* distance code table */
|
|
struct huft *th; /* temp huft table pointer used in tables decoding */
|
|
unsigned bl; /* lookup bits for tl */
|
|
unsigned bd; /* lookup bits for td */
|
|
unsigned nb; /* number of bit length codes */
|
|
unsigned nl; /* number of literal/length codes */
|
|
unsigned nd; /* number of distance codes */
|
|
unsigned ll[MAXLITLENS+MAXDISTS]; /* lit./length and distance code lengths */
|
|
register ulg b; /* bit buffer */
|
|
register unsigned k; /* number of bits in bit buffer */
|
|
int retval = 0; /* error code returned: initialized to "no error" */
|
|
|
|
|
|
/* make local bit buffer */
|
|
Trace((stderr, "\ndynamic block"));
|
|
b = G.bb;
|
|
k = G.bk;
|
|
|
|
|
|
/* read in table lengths */
|
|
NEEDBITS(5)
|
|
nl = 257 + ((unsigned)b & 0x1f); /* number of literal/length codes */
|
|
DUMPBITS(5)
|
|
NEEDBITS(5)
|
|
nd = 1 + ((unsigned)b & 0x1f); /* number of distance codes */
|
|
DUMPBITS(5)
|
|
NEEDBITS(4)
|
|
nb = 4 + ((unsigned)b & 0xf); /* number of bit length codes */
|
|
DUMPBITS(4)
|
|
if (nl > MAXLITLENS || nd > MAXDISTS)
|
|
return 1; /* bad lengths */
|
|
|
|
|
|
/* read in bit-length-code lengths */
|
|
for (j = 0; j < nb; j++)
|
|
{
|
|
NEEDBITS(3)
|
|
ll[border[j]] = (unsigned)b & 7;
|
|
DUMPBITS(3)
|
|
}
|
|
for (; j < 19; j++)
|
|
ll[border[j]] = 0;
|
|
|
|
|
|
/* build decoding table for trees--single level, 7 bit lookup */
|
|
bl = 7;
|
|
retval = huft_build(__G__ ll, 19, 19, NULL, NULL, &tl, &bl);
|
|
if (bl == 0) /* no bit lengths */
|
|
retval = 1;
|
|
if (retval)
|
|
{
|
|
if (retval == 1)
|
|
huft_free(tl);
|
|
return retval; /* incomplete code set */
|
|
}
|
|
|
|
|
|
/* read in literal and distance code lengths */
|
|
n = nl + nd;
|
|
m = mask_bits[bl];
|
|
i = l = 0;
|
|
while (i < n)
|
|
{
|
|
NEEDBITS(bl)
|
|
j = (th = tl + ((unsigned)b & m))->b;
|
|
DUMPBITS(j)
|
|
j = th->v.n;
|
|
if (j < 16) /* length of code in bits (0..15) */
|
|
ll[i++] = l = j; /* save last length in l */
|
|
else if (j == 16) /* repeat last length 3 to 6 times */
|
|
{
|
|
NEEDBITS(2)
|
|
j = 3 + ((unsigned)b & 3);
|
|
DUMPBITS(2)
|
|
if ((unsigned)i + j > n) {
|
|
huft_free(tl);
|
|
return 1;
|
|
}
|
|
while (j--)
|
|
ll[i++] = l;
|
|
}
|
|
else if (j == 17) /* 3 to 10 zero length codes */
|
|
{
|
|
NEEDBITS(3)
|
|
j = 3 + ((unsigned)b & 7);
|
|
DUMPBITS(3)
|
|
if ((unsigned)i + j > n) {
|
|
huft_free(tl);
|
|
return 1;
|
|
}
|
|
while (j--)
|
|
ll[i++] = 0;
|
|
l = 0;
|
|
}
|
|
else /* j == 18: 11 to 138 zero length codes */
|
|
{
|
|
NEEDBITS(7)
|
|
j = 11 + ((unsigned)b & 0x7f);
|
|
DUMPBITS(7)
|
|
if ((unsigned)i + j > n) {
|
|
huft_free(tl);
|
|
return 1;
|
|
}
|
|
while (j--)
|
|
ll[i++] = 0;
|
|
l = 0;
|
|
}
|
|
}
|
|
|
|
|
|
/* free decoding table for trees */
|
|
huft_free(tl);
|
|
|
|
|
|
/* restore the global bit buffer */
|
|
G.bb = b;
|
|
G.bk = k;
|
|
|
|
|
|
/* build the decoding tables for literal/length and distance codes */
|
|
bl = lbits;
|
|
#ifdef USE_DEFLATE64
|
|
retval = huft_build(__G__ ll, nl, 257, G.cplens, G.cplext, &tl, &bl);
|
|
#else
|
|
retval = huft_build(__G__ ll, nl, 257, cplens, cplext, &tl, &bl);
|
|
#endif
|
|
if (bl == 0) /* no literals or lengths */
|
|
retval = 1;
|
|
if (retval)
|
|
{
|
|
if (retval == 1) {
|
|
if (!uO.qflag)
|
|
MESSAGE((uch *)"(incomplete l-tree) ", 21L, 1);
|
|
huft_free(tl);
|
|
}
|
|
return retval; /* incomplete code set */
|
|
}
|
|
#ifdef FIX_PAST_EOB_BY_TABLEADJUST
|
|
/* Adjust the requested distance base table size so that a distance code
|
|
fetch never tries to get bits behind an immediatly following end-of-block
|
|
code. */
|
|
bd = (dbits <= bl+1 ? dbits : bl+1);
|
|
#else
|
|
bd = dbits;
|
|
#endif
|
|
#ifdef USE_DEFLATE64
|
|
retval = huft_build(__G__ ll + nl, nd, 0, cpdist, G.cpdext, &td, &bd);
|
|
#else
|
|
retval = huft_build(__G__ ll + nl, nd, 0, cpdist, cpdext, &td, &bd);
|
|
#endif
|
|
#ifdef PKZIP_BUG_WORKAROUND
|
|
if (retval == 1)
|
|
retval = 0;
|
|
#endif
|
|
if (bd == 0 && nl > 257) /* lengths but no distances */
|
|
retval = 1;
|
|
if (retval)
|
|
{
|
|
if (retval == 1) {
|
|
if (!uO.qflag)
|
|
MESSAGE((uch *)"(incomplete d-tree) ", 21L, 1);
|
|
huft_free(td);
|
|
}
|
|
huft_free(tl);
|
|
return retval;
|
|
}
|
|
|
|
/* decompress until an end-of-block code */
|
|
retval = inflate_codes(__G__ tl, td, bl, bd);
|
|
|
|
cleanup_and_exit:
|
|
/* free the decoding tables, return */
|
|
if (tl != (struct huft *)NULL)
|
|
huft_free(tl);
|
|
if (td != (struct huft *)NULL)
|
|
huft_free(td);
|
|
return retval;
|
|
}
|
|
|
|
|
|
|
|
static int inflate_block(__G__ e)
|
|
__GDEF
|
|
int *e; /* last block flag */
|
|
/* decompress an inflated block */
|
|
{
|
|
unsigned t; /* block type */
|
|
register ulg b; /* bit buffer */
|
|
register unsigned k; /* number of bits in bit buffer */
|
|
int retval = 0; /* error code returned: initialized to "no error" */
|
|
|
|
|
|
/* make local bit buffer */
|
|
b = G.bb;
|
|
k = G.bk;
|
|
|
|
|
|
/* read in last block bit */
|
|
NEEDBITS(1)
|
|
*e = (int)b & 1;
|
|
DUMPBITS(1)
|
|
|
|
|
|
/* read in block type */
|
|
NEEDBITS(2)
|
|
t = (unsigned)b & 3;
|
|
DUMPBITS(2)
|
|
|
|
|
|
/* restore the global bit buffer */
|
|
G.bb = b;
|
|
G.bk = k;
|
|
|
|
|
|
/* inflate that block type */
|
|
if (t == 2)
|
|
return inflate_dynamic(__G);
|
|
if (t == 0)
|
|
return inflate_stored(__G);
|
|
if (t == 1)
|
|
return inflate_fixed(__G);
|
|
|
|
|
|
/* bad block type */
|
|
retval = 2;
|
|
|
|
cleanup_and_exit:
|
|
return retval;
|
|
}
|
|
|
|
|
|
|
|
int inflate(__G__ is_defl64)
|
|
__GDEF
|
|
int is_defl64;
|
|
/* decompress an inflated entry */
|
|
{
|
|
int e; /* last block flag */
|
|
int r; /* result code */
|
|
#ifdef DEBUG
|
|
unsigned h = 0; /* maximum struct huft's malloc'ed */
|
|
#endif
|
|
|
|
#if (defined(DLL) && !defined(NO_SLIDE_REDIR))
|
|
if (G.redirect_slide)
|
|
wsize = G.redirect_size, redirSlide = G.redirect_buffer;
|
|
else
|
|
wsize = WSIZE, redirSlide = slide; /* how they're #defined if !DLL */
|
|
#endif
|
|
|
|
/* initialize window, bit buffer */
|
|
G.wp = 0;
|
|
G.bk = 0;
|
|
G.bb = 0;
|
|
|
|
#ifdef USE_DEFLATE64
|
|
if (is_defl64) {
|
|
G.cplens = cplens64;
|
|
G.cplext = cplext64;
|
|
G.cpdext = cpdext64;
|
|
G.fixed_tl = G.fixed_tl64;
|
|
G.fixed_bl = G.fixed_bl64;
|
|
G.fixed_td = G.fixed_td64;
|
|
G.fixed_bd = G.fixed_bd64;
|
|
} else {
|
|
G.cplens = cplens32;
|
|
G.cplext = cplext32;
|
|
G.cpdext = cpdext32;
|
|
G.fixed_tl = G.fixed_tl32;
|
|
G.fixed_bl = G.fixed_bl32;
|
|
G.fixed_td = G.fixed_td32;
|
|
G.fixed_bd = G.fixed_bd32;
|
|
}
|
|
#else /* !USE_DEFLATE64 */
|
|
if (is_defl64) {
|
|
/* This should not happen unless UnZip is built from object files
|
|
* compiled with inconsistent option setting. Handle this by
|
|
* returning with "bad input" error code.
|
|
*/
|
|
Trace((stderr, "\nThis inflate() cannot handle Deflate64!\n"));
|
|
return 2;
|
|
}
|
|
#endif /* ?USE_DEFLATE64 */
|
|
|
|
/* decompress until the last block */
|
|
do {
|
|
#ifdef DEBUG
|
|
G.hufts = 0;
|
|
#endif
|
|
if ((r = inflate_block(__G__ &e)) != 0)
|
|
return r;
|
|
#ifdef DEBUG
|
|
if (G.hufts > h)
|
|
h = G.hufts;
|
|
#endif
|
|
} while (!e);
|
|
|
|
Trace((stderr, "\n%u bytes in Huffman tables (%u/entry)\n",
|
|
h * (unsigned)sizeof(struct huft), (unsigned)sizeof(struct huft)));
|
|
|
|
#ifdef USE_DEFLATE64
|
|
if (is_defl64) {
|
|
G.fixed_tl64 = G.fixed_tl;
|
|
G.fixed_bl64 = G.fixed_bl;
|
|
G.fixed_td64 = G.fixed_td;
|
|
G.fixed_bd64 = G.fixed_bd;
|
|
} else {
|
|
G.fixed_tl32 = G.fixed_tl;
|
|
G.fixed_bl32 = G.fixed_bl;
|
|
G.fixed_td32 = G.fixed_td;
|
|
G.fixed_bd32 = G.fixed_bd;
|
|
}
|
|
#endif
|
|
|
|
/* flush out redirSlide and return (success, unless final FLUSH failed) */
|
|
return (FLUSH(G.wp));
|
|
}
|
|
|
|
|
|
|
|
int inflate_free(__G)
|
|
__GDEF
|
|
{
|
|
if (G.fixed_tl != (struct huft *)NULL)
|
|
{
|
|
huft_free(G.fixed_td);
|
|
huft_free(G.fixed_tl);
|
|
G.fixed_td = G.fixed_tl = (struct huft *)NULL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#endif /* ?USE_ZLIB */
|
|
|
|
|
|
/*
|
|
* GRR: moved huft_build() and huft_free() down here; used by explode()
|
|
* and fUnZip regardless of whether USE_ZLIB defined or not
|
|
*/
|
|
|
|
|
|
/* If BMAX needs to be larger than 16, then h and x[] should be ulg. */
|
|
#define BMAX 16 /* maximum bit length of any code (16 for explode) */
|
|
#define N_MAX 288 /* maximum number of codes in any set */
|
|
|
|
|
|
int huft_build(__G__ b, n, s, d, e, t, m)
|
|
__GDEF
|
|
ZCONST unsigned *b; /* code lengths in bits (all assumed <= BMAX) */
|
|
unsigned n; /* number of codes (assumed <= N_MAX) */
|
|
unsigned s; /* number of simple-valued codes (0..s-1) */
|
|
ZCONST ush *d; /* list of base values for non-simple codes */
|
|
ZCONST uch *e; /* list of extra bits for non-simple codes */
|
|
struct huft **t; /* result: starting table */
|
|
unsigned *m; /* maximum lookup bits, returns actual */
|
|
/* Given a list of code lengths and a maximum table size, make a set of
|
|
tables to decode that set of codes. Return zero on success, one if
|
|
the given code set is incomplete (the tables are still built in this
|
|
case), two if the input is invalid (all zero length codes or an
|
|
oversubscribed set of lengths), and three if not enough memory.
|
|
The code with value 256 is special, and the tables are constructed
|
|
so that no bits beyond that code are fetched when that code is
|
|
decoded. */
|
|
{
|
|
unsigned a; /* counter for codes of length k */
|
|
unsigned c[BMAX+1]; /* bit length count table */
|
|
unsigned el; /* length of EOB code (value 256) */
|
|
unsigned f; /* i repeats in table every f entries */
|
|
int g; /* maximum code length */
|
|
int h; /* table level */
|
|
register unsigned i; /* counter, current code */
|
|
register unsigned j; /* counter */
|
|
register int k; /* number of bits in current code */
|
|
int lx[BMAX+1]; /* memory for l[-1..BMAX-1] */
|
|
int *l = lx+1; /* stack of bits per table */
|
|
register unsigned *p; /* pointer into c[], b[], or v[] */
|
|
register struct huft *q; /* points to current table */
|
|
struct huft r; /* table entry for structure assignment */
|
|
struct huft *u[BMAX]; /* table stack */
|
|
unsigned v[N_MAX]; /* values in order of bit length */
|
|
register int w; /* bits before this table == (l * h) */
|
|
unsigned x[BMAX+1]; /* bit offsets, then code stack */
|
|
unsigned *xp; /* pointer into x */
|
|
int y; /* number of dummy codes added */
|
|
unsigned z; /* number of entries in current table */
|
|
|
|
|
|
/* Generate counts for each bit length */
|
|
el = n > 256 ? b[256] : BMAX; /* set length of EOB code, if any */
|
|
memzero((char *)c, sizeof(c));
|
|
p = (unsigned *)b; i = n;
|
|
do {
|
|
c[*p]++; p++; /* assume all entries <= BMAX */
|
|
} while (--i);
|
|
if (c[0] == n) /* null input--all zero length codes */
|
|
{
|
|
*t = (struct huft *)NULL;
|
|
*m = 0;
|
|
return 0;
|
|
}
|
|
|
|
|
|
/* Find minimum and maximum length, bound *m by those */
|
|
for (j = 1; j <= BMAX; j++)
|
|
if (c[j])
|
|
break;
|
|
k = j; /* minimum code length */
|
|
if (*m < j)
|
|
*m = j;
|
|
for (i = BMAX; i; i--)
|
|
if (c[i])
|
|
break;
|
|
g = i; /* maximum code length */
|
|
if (*m > i)
|
|
*m = i;
|
|
|
|
|
|
/* Adjust last length count to fill out codes, if needed */
|
|
for (y = 1 << j; j < i; j++, y <<= 1)
|
|
if ((y -= c[j]) < 0)
|
|
return 2; /* bad input: more codes than bits */
|
|
if ((y -= c[i]) < 0)
|
|
return 2;
|
|
c[i] += y;
|
|
|
|
|
|
/* Generate starting offsets into the value table for each length */
|
|
x[1] = j = 0;
|
|
p = c + 1; xp = x + 2;
|
|
while (--i) { /* note that i == g from above */
|
|
*xp++ = (j += *p++);
|
|
}
|
|
|
|
|
|
/* Make a table of values in order of bit lengths */
|
|
memzero((char *)v, sizeof(v));
|
|
p = (unsigned *)b; i = 0;
|
|
do {
|
|
if ((j = *p++) != 0)
|
|
v[x[j]++] = i;
|
|
} while (++i < n);
|
|
n = x[g]; /* set n to length of v */
|
|
|
|
|
|
/* Generate the Huffman codes and for each, make the table entries */
|
|
x[0] = i = 0; /* first Huffman code is zero */
|
|
p = v; /* grab values in bit order */
|
|
h = -1; /* no tables yet--level -1 */
|
|
w = l[-1] = 0; /* no bits decoded yet */
|
|
u[0] = (struct huft *)NULL; /* just to keep compilers happy */
|
|
q = (struct huft *)NULL; /* ditto */
|
|
z = 0; /* ditto */
|
|
|
|
/* go through the bit lengths (k already is bits in shortest code) */
|
|
for (; k <= g; k++)
|
|
{
|
|
a = c[k];
|
|
while (a--)
|
|
{
|
|
/* here i is the Huffman code of length k bits for value *p */
|
|
/* make tables up to required level */
|
|
while (k > w + l[h])
|
|
{
|
|
w += l[h++]; /* add bits already decoded */
|
|
|
|
/* compute minimum size table less than or equal to *m bits */
|
|
z = (z = g - w) > *m ? *m : z; /* upper limit */
|
|
if ((f = 1 << (j = k - w)) > a + 1) /* try a k-w bit table */
|
|
{ /* too few codes for k-w bit table */
|
|
f -= a + 1; /* deduct codes from patterns left */
|
|
xp = c + k;
|
|
while (++j < z) /* try smaller tables up to z bits */
|
|
{
|
|
if ((f <<= 1) <= *++xp)
|
|
break; /* enough codes to use up j bits */
|
|
f -= *xp; /* else deduct codes from patterns */
|
|
}
|
|
}
|
|
if ((unsigned)w + j > el && (unsigned)w < el)
|
|
j = el - w; /* make EOB code end at table */
|
|
z = 1 << j; /* table entries for j-bit table */
|
|
l[h] = j; /* set table size in stack */
|
|
|
|
/* allocate and link in new table */
|
|
if ((q = (struct huft *)malloc((z + 1)*sizeof(struct huft))) ==
|
|
(struct huft *)NULL)
|
|
{
|
|
if (h)
|
|
huft_free(u[0]);
|
|
return 3; /* not enough memory */
|
|
}
|
|
#ifdef DEBUG
|
|
G.hufts += z + 1; /* track memory usage */
|
|
#endif
|
|
*t = q + 1; /* link to list for huft_free() */
|
|
*(t = &(q->v.t)) = (struct huft *)NULL;
|
|
u[h] = ++q; /* table starts after link */
|
|
|
|
/* connect to last table, if there is one */
|
|
if (h)
|
|
{
|
|
x[h] = i; /* save pattern for backing up */
|
|
r.b = (uch)l[h-1]; /* bits to dump before this table */
|
|
r.e = (uch)(32 + j); /* bits in this table */
|
|
r.v.t = q; /* pointer to this table */
|
|
j = (i & ((1 << w) - 1)) >> (w - l[h-1]);
|
|
u[h-1][j] = r; /* connect to last table */
|
|
}
|
|
}
|
|
|
|
/* set up table entry in r */
|
|
r.b = (uch)(k - w);
|
|
if (p >= v + n)
|
|
r.e = INVALID_CODE; /* out of values--invalid code */
|
|
else if (*p < s)
|
|
{
|
|
r.e = (uch)(*p < 256 ? 32 : 31); /* 256 is end-of-block code */
|
|
r.v.n = (ush)*p++; /* simple code is just the value */
|
|
}
|
|
else
|
|
{
|
|
r.e = e[*p - s]; /* non-simple--look up in lists */
|
|
r.v.n = d[*p++ - s];
|
|
}
|
|
|
|
/* fill code-like entries with r */
|
|
f = 1 << (k - w);
|
|
for (j = i >> w; j < z; j += f)
|
|
q[j] = r;
|
|
|
|
/* backwards increment the k-bit code i */
|
|
for (j = 1 << (k - 1); i & j; j >>= 1)
|
|
i ^= j;
|
|
i ^= j;
|
|
|
|
/* backup over finished tables */
|
|
while ((i & ((1 << w) - 1)) != x[h])
|
|
w -= l[--h]; /* don't need to update q */
|
|
}
|
|
}
|
|
|
|
|
|
/* return actual size of base table */
|
|
*m = l[0];
|
|
|
|
|
|
/* Return true (1) if we were given an incomplete table */
|
|
return y != 0 && g != 1;
|
|
}
|
|
|
|
|
|
|
|
int huft_free(t)
|
|
struct huft *t; /* table to free */
|
|
/* Free the malloc'ed tables built by huft_build(), which makes a linked
|
|
list of the tables it made, with the links in a dummy first entry of
|
|
each table. */
|
|
{
|
|
register struct huft *p, *q;
|
|
|
|
|
|
/* Go through linked list, freeing from the malloced (t[-1]) address. */
|
|
p = t;
|
|
while (p != (struct huft *)NULL)
|
|
{
|
|
q = (--p)->v.t;
|
|
free((zvoid *)p);
|
|
p = q;
|
|
}
|
|
return 0;
|
|
}
|