Line | Count | Source |
1 | | /* |
2 | | * LZ4 auto-framing library |
3 | | * Copyright (c) Yann Collet. All rights reserved. |
4 | | * |
5 | | * BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php) |
6 | | * |
7 | | * Redistribution and use in source and binary forms, with or without |
8 | | * modification, are permitted provided that the following conditions are |
9 | | * met: |
10 | | * |
11 | | * - Redistributions of source code must retain the above copyright |
12 | | * notice, this list of conditions and the following disclaimer. |
13 | | * - Redistributions in binary form must reproduce the above |
14 | | * copyright notice, this list of conditions and the following disclaimer |
15 | | * in the documentation and/or other materials provided with the |
16 | | * distribution. |
17 | | * |
18 | | * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
19 | | * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
20 | | * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
21 | | * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
22 | | * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
23 | | * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
24 | | * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
25 | | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
26 | | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
27 | | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
28 | | * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
29 | | * |
30 | | * You can contact the author at : |
31 | | * - LZ4 homepage : http://www.lz4.org |
32 | | * - LZ4 source repository : https://github.com/lz4/lz4 |
33 | | */ |
34 | | |
35 | | /* LZ4F is a stand-alone API to create LZ4-compressed Frames |
36 | | * in full conformance with specification v1.6.1 . |
37 | | * This library rely upon memory management capabilities (malloc, free) |
38 | | * provided either by <stdlib.h>, |
39 | | * or redirected towards another library of user's choice |
40 | | * (see Memory Routines below). |
41 | | */ |
42 | | |
43 | | |
44 | | /*-************************************ |
45 | | * Compiler Options |
46 | | **************************************/ |
47 | | #include <limits.h> |
48 | | #ifdef _MSC_VER /* Visual Studio */ |
49 | | # pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */ |
50 | | #endif |
51 | | |
52 | | |
53 | | /*-************************************ |
54 | | * Tuning parameters |
55 | | **************************************/ |
56 | | /* |
57 | | * LZ4F_HEAPMODE : |
58 | | * Control how LZ4F_compressFrame allocates the Compression State, |
59 | | * either on stack (0:default, fastest), or in memory heap (1:requires malloc()). |
60 | | */ |
61 | | #ifndef LZ4F_HEAPMODE |
62 | | # define LZ4F_HEAPMODE 0 |
63 | | #endif |
64 | | |
65 | | |
66 | | /*-************************************ |
67 | | * Library declarations |
68 | | **************************************/ |
69 | | #define LZ4F_STATIC_LINKING_ONLY |
70 | | #include "lz4frame.h" |
71 | | #define LZ4_STATIC_LINKING_ONLY |
72 | | #include "lz4.h" |
73 | | #define LZ4_HC_STATIC_LINKING_ONLY |
74 | | #include "lz4hc.h" |
75 | | #define XXH_STATIC_LINKING_ONLY |
76 | | #include "xxhash.h" |
77 | | |
78 | | |
79 | | /*-************************************ |
80 | | * Memory routines |
81 | | **************************************/ |
82 | | /* |
83 | | * User may redirect invocations of |
84 | | * malloc(), calloc() and free() |
85 | | * towards another library or solution of their choice |
86 | | * by modifying below section. |
87 | | **/ |
88 | | |
89 | | #include <string.h> /* memset, memcpy, memmove */ |
90 | | #ifndef LZ4_SRC_INCLUDED /* avoid redefinition when sources are coalesced */ |
91 | 0 | # define MEM_INIT(p,v,s) memset((p),(v),(s)) |
92 | | #endif |
93 | | |
94 | | #ifndef LZ4_SRC_INCLUDED /* avoid redefinition when sources are coalesced */ |
95 | | # include <stdlib.h> /* malloc, calloc, free */ |
96 | 0 | # define ALLOC(s) malloc(s) |
97 | 0 | # define ALLOC_AND_ZERO(s) calloc(1,(s)) |
98 | 0 | # define FREEMEM(p) free(p) |
99 | | #endif |
100 | | |
101 | | static void* LZ4F_calloc(size_t s, LZ4F_CustomMem cmem) |
102 | 0 | { |
103 | | /* custom calloc defined : use it */ |
104 | 0 | if (cmem.customCalloc != NULL) { |
105 | 0 | return cmem.customCalloc(cmem.opaqueState, s); |
106 | 0 | } |
107 | | /* nothing defined : use default <stdlib.h>'s calloc() */ |
108 | 0 | if (cmem.customAlloc == NULL) { |
109 | 0 | return ALLOC_AND_ZERO(s); |
110 | 0 | } |
111 | | /* only custom alloc defined : use it, and combine it with memset() */ |
112 | 0 | { void* const p = cmem.customAlloc(cmem.opaqueState, s); |
113 | 0 | if (p != NULL) MEM_INIT(p, 0, s); |
114 | 0 | return p; |
115 | 0 | } } |
116 | | |
117 | | static void* LZ4F_malloc(size_t s, LZ4F_CustomMem cmem) |
118 | 0 | { |
119 | | /* custom malloc defined : use it */ |
120 | 0 | if (cmem.customAlloc != NULL) { |
121 | 0 | return cmem.customAlloc(cmem.opaqueState, s); |
122 | 0 | } |
123 | | /* nothing defined : use default <stdlib.h>'s malloc() */ |
124 | 0 | return ALLOC(s); |
125 | 0 | } |
126 | | |
127 | | static void LZ4F_free(void* p, LZ4F_CustomMem cmem) |
128 | 0 | { |
129 | 0 | if (p == NULL) return; |
130 | 0 | if (cmem.customFree != NULL) { |
131 | | /* custom allocation defined : use it */ |
132 | 0 | cmem.customFree(cmem.opaqueState, p); |
133 | 0 | return; |
134 | 0 | } |
135 | | /* nothing defined : use default <stdlib.h>'s free() */ |
136 | 0 | FREEMEM(p); |
137 | 0 | } |
138 | | |
139 | | |
140 | | /*-************************************ |
141 | | * Debug |
142 | | **************************************/ |
143 | | #if defined(LZ4_DEBUG) && (LZ4_DEBUG>=1) |
144 | | # include <assert.h> |
145 | | #else |
146 | | # ifndef assert |
147 | | # define assert(condition) ((void)0) |
148 | | # endif |
149 | | #endif |
150 | | |
151 | 0 | #define LZ4F_STATIC_ASSERT(c) { enum { LZ4F_static_assert = 1/(int)(!!(c)) }; } /* use only *after* variable declarations */ |
152 | | |
153 | | #if defined(LZ4_DEBUG) && (LZ4_DEBUG>=2) && !defined(DEBUGLOG) |
154 | | # include <stdio.h> |
155 | | static int g_debuglog_enable = 1; |
156 | | # define DEBUGLOG(l, ...) { \ |
157 | | if ((g_debuglog_enable) && (l<=LZ4_DEBUG)) { \ |
158 | | fprintf(stderr, __FILE__ " %i: ", __LINE__ ); \ |
159 | | fprintf(stderr, __VA_ARGS__); \ |
160 | | fprintf(stderr, " \n"); \ |
161 | | } } |
162 | | #else |
163 | 0 | # define DEBUGLOG(l, ...) {} /* disabled */ |
164 | | #endif |
165 | | |
166 | | |
167 | | /*-************************************ |
168 | | * Basic Types |
169 | | **************************************/ |
170 | | #ifndef LZ4_SRC_INCLUDED |
171 | | # if !defined (__VMS) && (defined (__cplusplus) || (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */) ) |
172 | | # include <stdint.h> |
173 | | typedef uint8_t BYTE; |
174 | | typedef uint16_t U16; |
175 | | typedef uint32_t U32; |
176 | | typedef int32_t S32; |
177 | | typedef uint64_t U64; |
178 | | # else |
179 | | typedef unsigned char BYTE; |
180 | | typedef unsigned short U16; |
181 | | typedef unsigned int U32; |
182 | | typedef signed int S32; |
183 | | typedef unsigned long long U64; |
184 | | # endif |
185 | | #endif |
186 | | |
187 | | |
188 | | /* unoptimized version; solves endianness & alignment issues */ |
189 | | static U32 LZ4F_readLE32 (const void* src) |
190 | 0 | { |
191 | 0 | const BYTE* const srcPtr = (const BYTE*)src; |
192 | 0 | U32 value32 = srcPtr[0]; |
193 | 0 | value32 |= ((U32)srcPtr[1])<< 8; |
194 | 0 | value32 |= ((U32)srcPtr[2])<<16; |
195 | 0 | value32 |= ((U32)srcPtr[3])<<24; |
196 | 0 | return value32; |
197 | 0 | } |
198 | | |
199 | | static void LZ4F_writeLE32 (void* dst, U32 value32) |
200 | 0 | { |
201 | 0 | BYTE* const dstPtr = (BYTE*)dst; |
202 | 0 | dstPtr[0] = (BYTE)value32; |
203 | 0 | dstPtr[1] = (BYTE)(value32 >> 8); |
204 | 0 | dstPtr[2] = (BYTE)(value32 >> 16); |
205 | 0 | dstPtr[3] = (BYTE)(value32 >> 24); |
206 | 0 | } |
207 | | |
208 | | static U64 LZ4F_readLE64 (const void* src) |
209 | 0 | { |
210 | 0 | const BYTE* const srcPtr = (const BYTE*)src; |
211 | 0 | U64 value64 = srcPtr[0]; |
212 | 0 | value64 |= ((U64)srcPtr[1]<<8); |
213 | 0 | value64 |= ((U64)srcPtr[2]<<16); |
214 | 0 | value64 |= ((U64)srcPtr[3]<<24); |
215 | 0 | value64 |= ((U64)srcPtr[4]<<32); |
216 | 0 | value64 |= ((U64)srcPtr[5]<<40); |
217 | 0 | value64 |= ((U64)srcPtr[6]<<48); |
218 | 0 | value64 |= ((U64)srcPtr[7]<<56); |
219 | 0 | return value64; |
220 | 0 | } |
221 | | |
222 | | static void LZ4F_writeLE64 (void* dst, U64 value64) |
223 | 0 | { |
224 | 0 | BYTE* const dstPtr = (BYTE*)dst; |
225 | 0 | dstPtr[0] = (BYTE)value64; |
226 | 0 | dstPtr[1] = (BYTE)(value64 >> 8); |
227 | 0 | dstPtr[2] = (BYTE)(value64 >> 16); |
228 | 0 | dstPtr[3] = (BYTE)(value64 >> 24); |
229 | 0 | dstPtr[4] = (BYTE)(value64 >> 32); |
230 | 0 | dstPtr[5] = (BYTE)(value64 >> 40); |
231 | 0 | dstPtr[6] = (BYTE)(value64 >> 48); |
232 | 0 | dstPtr[7] = (BYTE)(value64 >> 56); |
233 | 0 | } |
234 | | |
235 | | |
236 | | /*-************************************ |
237 | | * Constants |
238 | | **************************************/ |
239 | | #ifndef LZ4_SRC_INCLUDED /* avoid double definition */ |
240 | 0 | # define KB *(1<<10) |
241 | 0 | # define MB *(1<<20) |
242 | 0 | # define GB *(1<<30) |
243 | | #endif |
244 | | |
245 | 0 | #define _1BIT 0x01 |
246 | 0 | #define _2BITS 0x03 |
247 | 0 | #define _3BITS 0x07 |
248 | 0 | #define _4BITS 0x0F |
249 | | #define _8BITS 0xFF |
250 | | |
251 | 0 | #define LZ4F_BLOCKUNCOMPRESSED_FLAG 0x80000000U |
252 | 0 | #define LZ4F_BLOCKSIZEID_DEFAULT LZ4F_max64KB |
253 | | |
254 | | static const size_t minFHSize = LZ4F_HEADER_SIZE_MIN; /* 7 */ |
255 | | static const size_t maxFHSize = LZ4F_HEADER_SIZE_MAX; /* 19 */ |
256 | | static const size_t BHSize = LZ4F_BLOCK_HEADER_SIZE; /* block header : size, and compress flag */ |
257 | | static const size_t BFSize = LZ4F_BLOCK_CHECKSUM_SIZE; /* block footer : checksum (optional) */ |
258 | | |
259 | | |
260 | | /*-************************************ |
261 | | * Structures and local types |
262 | | **************************************/ |
263 | | |
264 | | typedef enum { LZ4B_COMPRESSED, LZ4B_UNCOMPRESSED} LZ4F_BlockCompressMode_e; |
265 | | typedef enum { ctxNone, ctxFast, ctxHC } LZ4F_CtxType_e; |
266 | | |
267 | | typedef struct LZ4F_cctx_s |
268 | | { |
269 | | LZ4F_CustomMem cmem; |
270 | | LZ4F_preferences_t prefs; |
271 | | U32 version; |
272 | | U32 cStage; /* 0 : compression uninitialized ; 1 : initialized, can compress */ |
273 | | const LZ4F_CDict* cdict; |
274 | | size_t maxBlockSize; |
275 | | size_t maxBufferSize; |
276 | | BYTE* tmpBuff; /* internal buffer, for streaming */ |
277 | | BYTE* tmpIn; /* starting position of data compress within internal buffer (>= tmpBuff) */ |
278 | | size_t tmpInSize; /* amount of data to compress after tmpIn */ |
279 | | U64 totalInSize; |
280 | | XXH32_state_t xxh; |
281 | | void* lz4CtxPtr; |
282 | | U16 lz4CtxAlloc; /* sized for: 0 = none, 1 = lz4 ctx, 2 = lz4hc ctx */ |
283 | | U16 lz4CtxType; /* in use as: 0 = none, 1 = lz4 ctx, 2 = lz4hc ctx */ |
284 | | LZ4F_BlockCompressMode_e blockCompressMode; |
285 | | } LZ4F_cctx_t; |
286 | | |
287 | | |
288 | | /*-************************************ |
289 | | * Error management |
290 | | **************************************/ |
291 | | #define LZ4F_GENERATE_STRING(STRING) #STRING, |
292 | | static const char* LZ4F_errorStrings[] = { LZ4F_LIST_ERRORS(LZ4F_GENERATE_STRING) }; |
293 | | |
294 | | |
295 | | unsigned LZ4F_isError(LZ4F_errorCode_t code) |
296 | 0 | { |
297 | 0 | return (code > (LZ4F_errorCode_t)(-LZ4F_ERROR_maxCode)); |
298 | 0 | } |
299 | | |
300 | | const char* LZ4F_getErrorName(LZ4F_errorCode_t code) |
301 | 0 | { |
302 | 0 | static const char* codeError = "Unspecified error code"; |
303 | 0 | if (LZ4F_isError(code)) return LZ4F_errorStrings[-(int)(code)]; |
304 | 0 | return codeError; |
305 | 0 | } |
306 | | |
307 | | LZ4F_errorCodes LZ4F_getErrorCode(size_t functionResult) |
308 | 0 | { |
309 | 0 | if (!LZ4F_isError(functionResult)) return LZ4F_OK_NoError; |
310 | 0 | return (LZ4F_errorCodes)(-(ptrdiff_t)functionResult); |
311 | 0 | } |
312 | | |
313 | | static LZ4F_errorCode_t LZ4F_returnErrorCode(LZ4F_errorCodes code) |
314 | 0 | { |
315 | | /* A compilation error here means sizeof(ptrdiff_t) is not large enough */ |
316 | 0 | LZ4F_STATIC_ASSERT(sizeof(ptrdiff_t) >= sizeof(size_t)); |
317 | 0 | return (LZ4F_errorCode_t)-(ptrdiff_t)code; |
318 | 0 | } |
319 | | |
320 | 0 | #define RETURN_ERROR(e) return LZ4F_returnErrorCode(LZ4F_ERROR_ ## e) |
321 | | |
322 | 0 | #define RETURN_ERROR_IF(c,e) do { \ |
323 | 0 | if (c) { \ |
324 | 0 | DEBUGLOG(3, "Error: " #c); \ |
325 | 0 | RETURN_ERROR(e); \ |
326 | 0 | } \ |
327 | 0 | } while (0) |
328 | | |
329 | 0 | #define FORWARD_IF_ERROR(r) do { if (LZ4F_isError(r)) return (r); } while (0) |
330 | | |
331 | 0 | unsigned LZ4F_getVersion(void) { return LZ4F_VERSION; } |
332 | | |
333 | 0 | int LZ4F_compressionLevel_max(void) { return LZ4HC_CLEVEL_MAX; } |
334 | | |
335 | | size_t LZ4F_getBlockSize(LZ4F_blockSizeID_t blockSizeID) |
336 | 0 | { |
337 | 0 | static const size_t blockSizes[4] = { 64 KB, 256 KB, 1 MB, 4 MB }; |
338 | |
|
339 | 0 | if (blockSizeID == 0) blockSizeID = LZ4F_BLOCKSIZEID_DEFAULT; |
340 | 0 | if (blockSizeID < LZ4F_max64KB || blockSizeID > LZ4F_max4MB) |
341 | 0 | RETURN_ERROR(maxBlockSize_invalid); |
342 | 0 | { int const blockSizeIdx = (int)blockSizeID - (int)LZ4F_max64KB; |
343 | 0 | return blockSizes[blockSizeIdx]; |
344 | 0 | } } |
345 | | |
346 | | /*-************************************ |
347 | | * Private functions |
348 | | **************************************/ |
349 | 0 | #define MIN(a,b) ( (a) < (b) ? (a) : (b) ) |
350 | | |
351 | | static BYTE LZ4F_headerChecksum (const void* header, size_t length) |
352 | 0 | { |
353 | 0 | U32 const xxh = XXH32(header, length, 0); |
354 | 0 | return (BYTE)(xxh >> 8); |
355 | 0 | } |
356 | | |
357 | | |
358 | | /*-************************************ |
359 | | * Simple-pass compression functions |
360 | | **************************************/ |
361 | | static LZ4F_blockSizeID_t LZ4F_optimalBSID(const LZ4F_blockSizeID_t requestedBSID, |
362 | | const size_t srcSize) |
363 | 0 | { |
364 | 0 | LZ4F_blockSizeID_t proposedBSID = LZ4F_max64KB; |
365 | 0 | size_t maxBlockSize = 64 KB; |
366 | 0 | while (requestedBSID > proposedBSID) { |
367 | 0 | if (srcSize <= maxBlockSize) |
368 | 0 | return proposedBSID; |
369 | 0 | proposedBSID = (LZ4F_blockSizeID_t)((int)proposedBSID + 1); |
370 | 0 | maxBlockSize <<= 2; |
371 | 0 | } |
372 | 0 | return requestedBSID; |
373 | 0 | } |
374 | | |
375 | | /*! LZ4F_compressBound_internal() : |
376 | | * Provides dstCapacity given a srcSize to guarantee operation success in worst case situations. |
377 | | * prefsPtr is optional : if NULL is provided, preferences will be set to cover worst case scenario. |
378 | | * @return is always the same for a srcSize and prefsPtr, so it can be relied upon to size reusable buffers. |
379 | | * When srcSize==0, LZ4F_compressBound() provides an upper bound for LZ4F_flush() and LZ4F_compressEnd() operations. |
380 | | */ |
381 | | static size_t LZ4F_compressBound_internal(size_t srcSize, |
382 | | const LZ4F_preferences_t* preferencesPtr, |
383 | | size_t alreadyBuffered) |
384 | 0 | { |
385 | 0 | LZ4F_preferences_t prefsNull = LZ4F_INIT_PREFERENCES; |
386 | 0 | prefsNull.frameInfo.contentChecksumFlag = LZ4F_contentChecksumEnabled; /* worst case */ |
387 | 0 | prefsNull.frameInfo.blockChecksumFlag = LZ4F_blockChecksumEnabled; /* worst case */ |
388 | 0 | { const LZ4F_preferences_t* const prefsPtr = (preferencesPtr==NULL) ? &prefsNull : preferencesPtr; |
389 | 0 | U32 const flush = prefsPtr->autoFlush | (srcSize==0); |
390 | 0 | LZ4F_blockSizeID_t const blockID = prefsPtr->frameInfo.blockSizeID; |
391 | 0 | size_t const blockSize = LZ4F_getBlockSize(blockID); |
392 | 0 | size_t const maxBuffered = blockSize - 1; |
393 | 0 | size_t const bufferedSize = MIN(alreadyBuffered, maxBuffered); |
394 | 0 | size_t const maxSrcSize = srcSize + bufferedSize; |
395 | 0 | unsigned const nbFullBlocks = (unsigned)(maxSrcSize / blockSize); |
396 | 0 | size_t const partialBlockSize = maxSrcSize & (blockSize-1); |
397 | 0 | size_t const lastBlockSize = flush ? partialBlockSize : 0; |
398 | 0 | unsigned const nbBlocks = nbFullBlocks + (lastBlockSize>0); |
399 | |
|
400 | 0 | size_t const blockCRCSize = BFSize * prefsPtr->frameInfo.blockChecksumFlag; |
401 | 0 | size_t const frameEnd = BHSize + (prefsPtr->frameInfo.contentChecksumFlag*BFSize); |
402 | |
|
403 | 0 | return ((BHSize + blockCRCSize) * nbBlocks) + |
404 | 0 | (blockSize * nbFullBlocks) + lastBlockSize + frameEnd; |
405 | 0 | } |
406 | 0 | } |
407 | | |
408 | | size_t LZ4F_compressFrameBound(size_t srcSize, const LZ4F_preferences_t* preferencesPtr) |
409 | 0 | { |
410 | 0 | LZ4F_preferences_t prefs; |
411 | 0 | size_t const headerSize = maxFHSize; /* max header size, including optional fields */ |
412 | |
|
413 | 0 | if (preferencesPtr!=NULL) prefs = *preferencesPtr; |
414 | 0 | else MEM_INIT(&prefs, 0, sizeof(prefs)); |
415 | 0 | prefs.autoFlush = 1; |
416 | |
|
417 | 0 | return headerSize + LZ4F_compressBound_internal(srcSize, &prefs, 0);; |
418 | 0 | } |
419 | | |
420 | | |
421 | | /*! LZ4F_compressFrame_usingCDict() : |
422 | | * Compress srcBuffer using a dictionary, in a single step. |
423 | | * cdict can be NULL, in which case, no dictionary is used. |
424 | | * dstBuffer MUST be >= LZ4F_compressFrameBound(srcSize, preferencesPtr). |
425 | | * The LZ4F_preferences_t structure is optional : you may provide NULL as argument, |
426 | | * however, it's the only way to provide a dictID, so it's not recommended. |
427 | | * @return : number of bytes written into dstBuffer, |
428 | | * or an error code if it fails (can be tested using LZ4F_isError()) |
429 | | */ |
430 | | size_t LZ4F_compressFrame_usingCDict(LZ4F_cctx* cctx, |
431 | | void* dstBuffer, size_t dstCapacity, |
432 | | const void* srcBuffer, size_t srcSize, |
433 | | const LZ4F_CDict* cdict, |
434 | | const LZ4F_preferences_t* preferencesPtr) |
435 | 0 | { |
436 | 0 | LZ4F_preferences_t prefs; |
437 | 0 | LZ4F_compressOptions_t options; |
438 | 0 | BYTE* const dstStart = (BYTE*) dstBuffer; |
439 | 0 | BYTE* dstPtr = dstStart; |
440 | 0 | BYTE* const dstEnd = dstStart + dstCapacity; |
441 | |
|
442 | 0 | DEBUGLOG(4, "LZ4F_compressFrame_usingCDict (srcSize=%u)", (unsigned)srcSize); |
443 | 0 | if (preferencesPtr!=NULL) |
444 | 0 | prefs = *preferencesPtr; |
445 | 0 | else |
446 | 0 | MEM_INIT(&prefs, 0, sizeof(prefs)); |
447 | 0 | if (prefs.frameInfo.contentSize != 0) |
448 | 0 | prefs.frameInfo.contentSize = (U64)srcSize; /* auto-correct content size if selected (!=0) */ |
449 | |
|
450 | 0 | prefs.frameInfo.blockSizeID = LZ4F_optimalBSID(prefs.frameInfo.blockSizeID, srcSize); |
451 | 0 | prefs.autoFlush = 1; |
452 | 0 | if (srcSize <= LZ4F_getBlockSize(prefs.frameInfo.blockSizeID)) |
453 | 0 | prefs.frameInfo.blockMode = LZ4F_blockIndependent; /* only one block => no need for inter-block link */ |
454 | |
|
455 | 0 | MEM_INIT(&options, 0, sizeof(options)); |
456 | 0 | options.stableSrc = 1; |
457 | |
|
458 | 0 | RETURN_ERROR_IF(dstCapacity < LZ4F_compressFrameBound(srcSize, &prefs), dstMaxSize_tooSmall); |
459 | | |
460 | 0 | { size_t const headerSize = LZ4F_compressBegin_usingCDict(cctx, dstBuffer, dstCapacity, cdict, &prefs); /* write header */ |
461 | 0 | FORWARD_IF_ERROR(headerSize); |
462 | 0 | dstPtr += headerSize; /* header size */ } |
463 | | |
464 | 0 | assert(dstEnd >= dstPtr); |
465 | 0 | { size_t const cSize = LZ4F_compressUpdate(cctx, dstPtr, (size_t)(dstEnd-dstPtr), srcBuffer, srcSize, &options); |
466 | 0 | FORWARD_IF_ERROR(cSize); |
467 | 0 | dstPtr += cSize; } |
468 | | |
469 | 0 | assert(dstEnd >= dstPtr); |
470 | 0 | { size_t const tailSize = LZ4F_compressEnd(cctx, dstPtr, (size_t)(dstEnd-dstPtr), &options); /* flush last block, and generate suffix */ |
471 | 0 | FORWARD_IF_ERROR(tailSize); |
472 | 0 | dstPtr += tailSize; } |
473 | | |
474 | 0 | assert(dstEnd >= dstStart); |
475 | 0 | return (size_t)(dstPtr - dstStart); |
476 | 0 | } |
477 | | |
478 | | |
479 | | /*! LZ4F_compressFrame() : |
480 | | * Compress an entire srcBuffer into a valid LZ4 frame, in a single step. |
481 | | * dstBuffer MUST be >= LZ4F_compressFrameBound(srcSize, preferencesPtr). |
482 | | * The LZ4F_preferences_t structure is optional : you can provide NULL as argument. All preferences will be set to default. |
483 | | * @return : number of bytes written into dstBuffer. |
484 | | * or an error code if it fails (can be tested using LZ4F_isError()) |
485 | | */ |
486 | | size_t LZ4F_compressFrame(void* dstBuffer, size_t dstCapacity, |
487 | | const void* srcBuffer, size_t srcSize, |
488 | | const LZ4F_preferences_t* preferencesPtr) |
489 | 0 | { |
490 | 0 | size_t result; |
491 | | #if (LZ4F_HEAPMODE) |
492 | | LZ4F_cctx_t* cctxPtr; |
493 | | result = LZ4F_createCompressionContext(&cctxPtr, LZ4F_VERSION); |
494 | | FORWARD_IF_ERROR(result); |
495 | | #else |
496 | 0 | LZ4F_cctx_t cctx; |
497 | 0 | LZ4_stream_t lz4ctx; |
498 | 0 | LZ4F_cctx_t* const cctxPtr = &cctx; |
499 | |
|
500 | 0 | MEM_INIT(&cctx, 0, sizeof(cctx)); |
501 | 0 | cctx.version = LZ4F_VERSION; |
502 | 0 | cctx.maxBufferSize = 5 MB; /* mess with real buffer size to prevent dynamic allocation; works only because autoflush==1 & stableSrc==1 */ |
503 | 0 | if ( preferencesPtr == NULL |
504 | 0 | || preferencesPtr->compressionLevel < LZ4HC_CLEVEL_MIN ) { |
505 | 0 | LZ4_initStream(&lz4ctx, sizeof(lz4ctx)); |
506 | 0 | cctxPtr->lz4CtxPtr = &lz4ctx; |
507 | 0 | cctxPtr->lz4CtxAlloc = 1; |
508 | 0 | cctxPtr->lz4CtxType = ctxFast; |
509 | 0 | } |
510 | 0 | #endif |
511 | 0 | DEBUGLOG(4, "LZ4F_compressFrame"); |
512 | |
|
513 | 0 | result = LZ4F_compressFrame_usingCDict(cctxPtr, dstBuffer, dstCapacity, |
514 | 0 | srcBuffer, srcSize, |
515 | 0 | NULL, preferencesPtr); |
516 | |
|
517 | | #if (LZ4F_HEAPMODE) |
518 | | LZ4F_freeCompressionContext(cctxPtr); |
519 | | #else |
520 | 0 | if ( preferencesPtr != NULL |
521 | 0 | && preferencesPtr->compressionLevel >= LZ4HC_CLEVEL_MIN ) { |
522 | 0 | LZ4F_free(cctxPtr->lz4CtxPtr, cctxPtr->cmem); |
523 | 0 | } |
524 | 0 | #endif |
525 | 0 | return result; |
526 | 0 | } |
527 | | |
528 | | |
529 | | /*-*************************************************** |
530 | | * Dictionary compression |
531 | | *****************************************************/ |
532 | | |
533 | | struct LZ4F_CDict_s { |
534 | | LZ4F_CustomMem cmem; |
535 | | void* dictContent; |
536 | | LZ4_stream_t* fastCtx; |
537 | | LZ4_streamHC_t* HCCtx; |
538 | | }; /* typedef'd to LZ4F_CDict within lz4frame_static.h */ |
539 | | |
540 | | LZ4F_CDict* |
541 | | LZ4F_createCDict_advanced(LZ4F_CustomMem cmem, const void* dictBuffer, size_t dictSize) |
542 | 0 | { |
543 | 0 | const char* dictStart = (const char*)dictBuffer; |
544 | 0 | LZ4F_CDict* cdict = NULL; |
545 | |
|
546 | 0 | DEBUGLOG(4, "LZ4F_createCDict_advanced"); |
547 | |
|
548 | 0 | if (!dictStart) |
549 | 0 | return NULL; |
550 | 0 | cdict = (LZ4F_CDict*)LZ4F_malloc(sizeof(*cdict), cmem); |
551 | 0 | if (!cdict) |
552 | 0 | return NULL; |
553 | | |
554 | 0 | cdict->cmem = cmem; |
555 | 0 | if (dictSize > 64 KB) { |
556 | 0 | dictStart += dictSize - 64 KB; |
557 | 0 | dictSize = 64 KB; |
558 | 0 | } |
559 | 0 | cdict->dictContent = LZ4F_malloc(dictSize, cmem); |
560 | | /* note: using @cmem to allocate => can't use default create */ |
561 | 0 | cdict->fastCtx = (LZ4_stream_t*)LZ4F_malloc(sizeof(LZ4_stream_t), cmem); |
562 | 0 | cdict->HCCtx = (LZ4_streamHC_t*)LZ4F_malloc(sizeof(LZ4_streamHC_t), cmem); |
563 | 0 | if (!cdict->dictContent || !cdict->fastCtx || !cdict->HCCtx) { |
564 | 0 | LZ4F_freeCDict(cdict); |
565 | 0 | return NULL; |
566 | 0 | } |
567 | 0 | memcpy(cdict->dictContent, dictStart, dictSize); |
568 | 0 | LZ4_initStream(cdict->fastCtx, sizeof(LZ4_stream_t)); |
569 | 0 | LZ4_loadDictSlow(cdict->fastCtx, (const char*)cdict->dictContent, (int)dictSize); |
570 | 0 | LZ4_initStreamHC(cdict->HCCtx, sizeof(LZ4_streamHC_t)); |
571 | | /* note: we don't know at this point which compression level is going to be used |
572 | | * as a consequence, HCCtx is created for the more common HC mode */ |
573 | 0 | LZ4_setCompressionLevel(cdict->HCCtx, LZ4HC_CLEVEL_DEFAULT); |
574 | 0 | LZ4_loadDictHC(cdict->HCCtx, (const char*)cdict->dictContent, (int)dictSize); |
575 | 0 | return cdict; |
576 | 0 | } |
577 | | |
578 | | /*! LZ4F_createCDict() : |
579 | | * When compressing multiple messages / blocks with the same dictionary, it's recommended to load it just once. |
580 | | * LZ4F_createCDict() will create a digested dictionary, ready to start future compression operations without startup delay. |
581 | | * LZ4F_CDict can be created once and shared by multiple threads concurrently, since its usage is read-only. |
582 | | * @dictBuffer can be released after LZ4F_CDict creation, since its content is copied within CDict |
583 | | * @return : digested dictionary for compression, or NULL if failed */ |
584 | | LZ4F_CDict* LZ4F_createCDict(const void* dictBuffer, size_t dictSize) |
585 | 0 | { |
586 | 0 | DEBUGLOG(4, "LZ4F_createCDict"); |
587 | 0 | return LZ4F_createCDict_advanced(LZ4F_defaultCMem, dictBuffer, dictSize); |
588 | 0 | } |
589 | | |
590 | | void LZ4F_freeCDict(LZ4F_CDict* cdict) |
591 | 0 | { |
592 | 0 | if (cdict==NULL) return; /* support free on NULL */ |
593 | 0 | LZ4F_free(cdict->dictContent, cdict->cmem); |
594 | 0 | LZ4F_free(cdict->fastCtx, cdict->cmem); |
595 | 0 | LZ4F_free(cdict->HCCtx, cdict->cmem); |
596 | 0 | LZ4F_free(cdict, cdict->cmem); |
597 | 0 | } |
598 | | |
599 | | |
600 | | /*-********************************* |
601 | | * Advanced compression functions |
602 | | ***********************************/ |
603 | | |
604 | | LZ4F_cctx* |
605 | | LZ4F_createCompressionContext_advanced(LZ4F_CustomMem customMem, unsigned version) |
606 | 0 | { |
607 | 0 | LZ4F_cctx* const cctxPtr = |
608 | 0 | (LZ4F_cctx*)LZ4F_calloc(sizeof(LZ4F_cctx), customMem); |
609 | 0 | if (cctxPtr==NULL) return NULL; |
610 | | |
611 | 0 | cctxPtr->cmem = customMem; |
612 | 0 | cctxPtr->version = version; |
613 | 0 | cctxPtr->cStage = 0; /* Uninitialized. Next stage : init cctx */ |
614 | |
|
615 | 0 | return cctxPtr; |
616 | 0 | } |
617 | | |
618 | | /*! LZ4F_createCompressionContext() : |
619 | | * The first thing to do is to create a compressionContext object, which will be used in all compression operations. |
620 | | * This is achieved using LZ4F_createCompressionContext(), which takes as argument a version and an LZ4F_preferences_t structure. |
621 | | * The version provided MUST be LZ4F_VERSION. It is intended to track potential incompatible differences between different binaries. |
622 | | * The function will provide a pointer to an allocated LZ4F_compressionContext_t object. |
623 | | * If the result LZ4F_errorCode_t is not OK_NoError, there was an error during context creation. |
624 | | * Object can release its memory using LZ4F_freeCompressionContext(); |
625 | | **/ |
626 | | LZ4F_errorCode_t |
627 | | LZ4F_createCompressionContext(LZ4F_cctx** LZ4F_compressionContextPtr, unsigned version) |
628 | 0 | { |
629 | 0 | assert(LZ4F_compressionContextPtr != NULL); /* considered a violation of narrow contract */ |
630 | | /* in case it nonetheless happen in production */ |
631 | 0 | RETURN_ERROR_IF(LZ4F_compressionContextPtr == NULL, parameter_null); |
632 | | |
633 | 0 | *LZ4F_compressionContextPtr = LZ4F_createCompressionContext_advanced(LZ4F_defaultCMem, version); |
634 | 0 | RETURN_ERROR_IF(*LZ4F_compressionContextPtr==NULL, allocation_failed); |
635 | 0 | return LZ4F_OK_NoError; |
636 | 0 | } |
637 | | |
638 | | LZ4F_errorCode_t LZ4F_freeCompressionContext(LZ4F_cctx* cctxPtr) |
639 | 0 | { |
640 | 0 | if (cctxPtr != NULL) { /* support free on NULL */ |
641 | 0 | LZ4F_free(cctxPtr->lz4CtxPtr, cctxPtr->cmem); /* note: LZ4_streamHC_t and LZ4_stream_t are simple POD types */ |
642 | 0 | LZ4F_free(cctxPtr->tmpBuff, cctxPtr->cmem); |
643 | 0 | LZ4F_free(cctxPtr, cctxPtr->cmem); |
644 | 0 | } |
645 | 0 | return LZ4F_OK_NoError; |
646 | 0 | } |
647 | | |
648 | | |
649 | | /** |
650 | | * This function prepares the internal LZ4(HC) stream for a new compression, |
651 | | * resetting the context and attaching the dictionary, if there is one. |
652 | | * |
653 | | * It needs to be called at the beginning of each independent compression |
654 | | * stream (i.e., at the beginning of a frame in blockLinked mode, or at the |
655 | | * beginning of each block in blockIndependent mode). |
656 | | */ |
657 | | static void LZ4F_initStream(void* ctx, |
658 | | const LZ4F_CDict* cdict, |
659 | | int level, |
660 | 0 | LZ4F_blockMode_t blockMode) { |
661 | 0 | if (level < LZ4HC_CLEVEL_MIN) { |
662 | 0 | if (cdict || blockMode == LZ4F_blockLinked) { |
663 | | /* In these cases, we will call LZ4_compress_fast_continue(), |
664 | | * which needs an already reset context. Otherwise, we'll call a |
665 | | * one-shot API. The non-continued APIs internally perform their own |
666 | | * resets at the beginning of their calls, where they know what |
667 | | * tableType they need the context to be in. So in that case this |
668 | | * would be misguided / wasted work. */ |
669 | 0 | LZ4_resetStream_fast((LZ4_stream_t*)ctx); |
670 | 0 | if (cdict) |
671 | 0 | LZ4_attach_dictionary((LZ4_stream_t*)ctx, cdict->fastCtx); |
672 | 0 | } |
673 | | /* In these cases, we'll call a one-shot API. |
674 | | * The non-continued APIs internally perform their own resets |
675 | | * at the beginning of their calls, where they know |
676 | | * which tableType they need the context to be in. |
677 | | * Therefore, a reset here would be wasted work. */ |
678 | 0 | } else { |
679 | 0 | LZ4_resetStreamHC_fast((LZ4_streamHC_t*)ctx, level); |
680 | 0 | if (cdict) |
681 | 0 | LZ4_attach_HC_dictionary((LZ4_streamHC_t*)ctx, cdict->HCCtx); |
682 | 0 | } |
683 | 0 | } |
684 | | |
685 | 0 | static int ctxTypeID_to_size(int ctxTypeID) { |
686 | 0 | switch(ctxTypeID) { |
687 | 0 | case 1: |
688 | 0 | return LZ4_sizeofState(); |
689 | 0 | case 2: |
690 | 0 | return LZ4_sizeofStateHC(); |
691 | 0 | default: |
692 | 0 | return 0; |
693 | 0 | } |
694 | 0 | } |
695 | | |
696 | 0 | size_t LZ4F_cctx_size(const LZ4F_cctx* cctx) { |
697 | 0 | if (cctx == NULL) { |
698 | 0 | return 0; |
699 | 0 | } |
700 | 0 | return sizeof(*cctx) + cctx->maxBufferSize + ctxTypeID_to_size(cctx->lz4CtxAlloc); |
701 | 0 | } |
702 | | |
703 | | /* LZ4F_compressBegin_internal() |
704 | | * Note: only accepts @cdict _or_ @dictBuffer as non NULL. |
705 | | */ |
706 | | static size_t LZ4F_compressBegin_internal(LZ4F_cctx* cctx, |
707 | | void* dstBuffer, size_t dstCapacity, |
708 | | const void* dictBuffer, size_t dictSize, |
709 | | const LZ4F_CDict* cdict, |
710 | | const LZ4F_preferences_t* preferencesPtr) |
711 | 0 | { |
712 | 0 | LZ4F_preferences_t const prefNull = LZ4F_INIT_PREFERENCES; |
713 | 0 | BYTE* const dstStart = (BYTE*)dstBuffer; |
714 | 0 | BYTE* dstPtr = dstStart; |
715 | |
|
716 | 0 | RETURN_ERROR_IF(dstCapacity < maxFHSize, dstMaxSize_tooSmall); |
717 | 0 | if (preferencesPtr == NULL) preferencesPtr = &prefNull; |
718 | 0 | cctx->prefs = *preferencesPtr; |
719 | 0 | DEBUGLOG(5, "LZ4F_compressBegin_internal: Independent_blocks=%u", cctx->prefs.frameInfo.blockMode); |
720 | | |
721 | | /* cctx Management */ |
722 | 0 | { U16 const ctxTypeID = (cctx->prefs.compressionLevel < LZ4HC_CLEVEL_MIN) ? 1 : 2; |
723 | 0 | int requiredSize = ctxTypeID_to_size(ctxTypeID); |
724 | 0 | int allocatedSize = ctxTypeID_to_size(cctx->lz4CtxAlloc); |
725 | 0 | if (allocatedSize < requiredSize) { |
726 | | /* not enough space allocated */ |
727 | 0 | LZ4F_free(cctx->lz4CtxPtr, cctx->cmem); |
728 | 0 | if (cctx->prefs.compressionLevel < LZ4HC_CLEVEL_MIN) { |
729 | | /* must take ownership of memory allocation, |
730 | | * in order to respect custom allocator contract */ |
731 | 0 | cctx->lz4CtxPtr = LZ4F_malloc(sizeof(LZ4_stream_t), cctx->cmem); |
732 | 0 | if (cctx->lz4CtxPtr) |
733 | 0 | LZ4_initStream(cctx->lz4CtxPtr, sizeof(LZ4_stream_t)); |
734 | 0 | } else { |
735 | 0 | cctx->lz4CtxPtr = LZ4F_malloc(sizeof(LZ4_streamHC_t), cctx->cmem); |
736 | 0 | if (cctx->lz4CtxPtr) |
737 | 0 | LZ4_initStreamHC(cctx->lz4CtxPtr, sizeof(LZ4_streamHC_t)); |
738 | 0 | } |
739 | 0 | RETURN_ERROR_IF(cctx->lz4CtxPtr == NULL, allocation_failed); |
740 | 0 | cctx->lz4CtxAlloc = ctxTypeID; |
741 | 0 | cctx->lz4CtxType = ctxTypeID; |
742 | 0 | } else if (cctx->lz4CtxType != ctxTypeID) { |
743 | | /* otherwise, a sufficient buffer is already allocated, |
744 | | * but we need to reset it to the correct context type */ |
745 | 0 | if (cctx->prefs.compressionLevel < LZ4HC_CLEVEL_MIN) { |
746 | 0 | LZ4_initStream((LZ4_stream_t*)cctx->lz4CtxPtr, sizeof(LZ4_stream_t)); |
747 | 0 | } else { |
748 | 0 | LZ4_initStreamHC((LZ4_streamHC_t*)cctx->lz4CtxPtr, sizeof(LZ4_streamHC_t)); |
749 | 0 | LZ4_setCompressionLevel((LZ4_streamHC_t*)cctx->lz4CtxPtr, cctx->prefs.compressionLevel); |
750 | 0 | } |
751 | 0 | cctx->lz4CtxType = ctxTypeID; |
752 | 0 | } } |
753 | | |
754 | | /* Buffer Management */ |
755 | 0 | if (cctx->prefs.frameInfo.blockSizeID == 0) |
756 | 0 | cctx->prefs.frameInfo.blockSizeID = LZ4F_BLOCKSIZEID_DEFAULT; |
757 | 0 | cctx->maxBlockSize = LZ4F_getBlockSize(cctx->prefs.frameInfo.blockSizeID); |
758 | |
|
759 | 0 | { size_t const requiredBuffSize = preferencesPtr->autoFlush ? |
760 | 0 | ((cctx->prefs.frameInfo.blockMode == LZ4F_blockLinked) ? 64 KB : 0) : /* only needs past data up to window size */ |
761 | 0 | cctx->maxBlockSize + ((cctx->prefs.frameInfo.blockMode == LZ4F_blockLinked) ? 128 KB : 0); |
762 | |
|
763 | 0 | if (cctx->maxBufferSize < requiredBuffSize) { |
764 | 0 | cctx->maxBufferSize = 0; |
765 | 0 | LZ4F_free(cctx->tmpBuff, cctx->cmem); |
766 | 0 | cctx->tmpBuff = (BYTE*)LZ4F_malloc(requiredBuffSize, cctx->cmem); |
767 | 0 | RETURN_ERROR_IF(cctx->tmpBuff == NULL, allocation_failed); |
768 | 0 | cctx->maxBufferSize = requiredBuffSize; |
769 | 0 | } } |
770 | 0 | cctx->tmpIn = cctx->tmpBuff; |
771 | 0 | cctx->tmpInSize = 0; |
772 | 0 | (void)XXH32_reset(&(cctx->xxh), 0); |
773 | | |
774 | | /* context init */ |
775 | 0 | cctx->cdict = cdict; |
776 | 0 | if (cctx->prefs.frameInfo.blockMode == LZ4F_blockLinked) { |
777 | | /* frame init only for blockLinked : blockIndependent will be init at each block */ |
778 | 0 | LZ4F_initStream(cctx->lz4CtxPtr, cdict, cctx->prefs.compressionLevel, LZ4F_blockLinked); |
779 | 0 | } |
780 | 0 | if (preferencesPtr->compressionLevel >= LZ4HC_CLEVEL_MIN) { |
781 | 0 | LZ4_favorDecompressionSpeed((LZ4_streamHC_t*)cctx->lz4CtxPtr, (int)preferencesPtr->favorDecSpeed); |
782 | 0 | } |
783 | 0 | if (dictBuffer) { |
784 | 0 | assert(cdict == NULL); |
785 | 0 | RETURN_ERROR_IF(dictSize > INT_MAX, parameter_invalid); |
786 | 0 | if (cctx->lz4CtxType == ctxFast) { |
787 | | /* lz4 fast*/ |
788 | 0 | LZ4_loadDict((LZ4_stream_t*)cctx->lz4CtxPtr, (const char*)dictBuffer, (int)dictSize); |
789 | 0 | } else { |
790 | | /* lz4hc */ |
791 | 0 | assert(cctx->lz4CtxType == ctxHC); |
792 | 0 | LZ4_loadDictHC((LZ4_streamHC_t*)cctx->lz4CtxPtr, (const char*)dictBuffer, (int)dictSize); |
793 | 0 | } |
794 | 0 | } |
795 | | |
796 | | /* Stage 2 : Write Frame Header */ |
797 | | |
798 | | /* Magic Number */ |
799 | 0 | LZ4F_writeLE32(dstPtr, LZ4F_MAGICNUMBER); |
800 | 0 | dstPtr += 4; |
801 | 0 | { BYTE* const headerStart = dstPtr; |
802 | | |
803 | | /* FLG Byte */ |
804 | 0 | *dstPtr++ = (BYTE)(((1 & _2BITS) << 6) /* Version('01') */ |
805 | 0 | + ((cctx->prefs.frameInfo.blockMode & _1BIT ) << 5) |
806 | 0 | + ((cctx->prefs.frameInfo.blockChecksumFlag & _1BIT ) << 4) |
807 | 0 | + ((unsigned)(cctx->prefs.frameInfo.contentSize > 0) << 3) |
808 | 0 | + ((cctx->prefs.frameInfo.contentChecksumFlag & _1BIT ) << 2) |
809 | 0 | + (cctx->prefs.frameInfo.dictID > 0) ); |
810 | | /* BD Byte */ |
811 | 0 | *dstPtr++ = (BYTE)((cctx->prefs.frameInfo.blockSizeID & _3BITS) << 4); |
812 | | /* Optional Frame content size field */ |
813 | 0 | if (cctx->prefs.frameInfo.contentSize) { |
814 | 0 | LZ4F_writeLE64(dstPtr, cctx->prefs.frameInfo.contentSize); |
815 | 0 | dstPtr += 8; |
816 | 0 | cctx->totalInSize = 0; |
817 | 0 | } |
818 | | /* Optional dictionary ID field */ |
819 | 0 | if (cctx->prefs.frameInfo.dictID) { |
820 | 0 | LZ4F_writeLE32(dstPtr, cctx->prefs.frameInfo.dictID); |
821 | 0 | dstPtr += 4; |
822 | 0 | } |
823 | | /* Header CRC Byte */ |
824 | 0 | *dstPtr = LZ4F_headerChecksum(headerStart, (size_t)(dstPtr - headerStart)); |
825 | 0 | dstPtr++; |
826 | 0 | } |
827 | |
|
828 | 0 | cctx->cStage = 1; /* header written, now request input data block */ |
829 | 0 | return (size_t)(dstPtr - dstStart); |
830 | 0 | } |
831 | | |
832 | | size_t LZ4F_compressBegin(LZ4F_cctx* cctx, |
833 | | void* dstBuffer, size_t dstCapacity, |
834 | | const LZ4F_preferences_t* preferencesPtr) |
835 | 0 | { |
836 | 0 | return LZ4F_compressBegin_internal(cctx, dstBuffer, dstCapacity, |
837 | 0 | NULL, 0, |
838 | 0 | NULL, preferencesPtr); |
839 | 0 | } |
840 | | |
841 | | /* LZ4F_compressBegin_usingDictOnce: |
842 | | * Hidden implementation, |
843 | | * employed for multi-threaded compression |
844 | | * when frame defines linked blocks */ |
845 | | static size_t LZ4F_compressBegin_usingDictOnce(LZ4F_cctx* cctx, |
846 | | void* dstBuffer, size_t dstCapacity, |
847 | | const void* dict, size_t dictSize, |
848 | | const LZ4F_preferences_t* preferencesPtr) |
849 | 0 | { |
850 | 0 | return LZ4F_compressBegin_internal(cctx, dstBuffer, dstCapacity, |
851 | 0 | dict, dictSize, |
852 | 0 | NULL, preferencesPtr); |
853 | 0 | } |
854 | | |
855 | | size_t LZ4F_compressBegin_usingDict(LZ4F_cctx* cctx, |
856 | | void* dstBuffer, size_t dstCapacity, |
857 | | const void* dict, size_t dictSize, |
858 | | const LZ4F_preferences_t* preferencesPtr) |
859 | 0 | { |
860 | | /* note : incorrect implementation : |
861 | | * this will only use the dictionary once, |
862 | | * instead of once *per* block when frames defines independent blocks */ |
863 | 0 | return LZ4F_compressBegin_usingDictOnce(cctx, dstBuffer, dstCapacity, |
864 | 0 | dict, dictSize, |
865 | 0 | preferencesPtr); |
866 | 0 | } |
867 | | |
868 | | size_t LZ4F_compressBegin_usingCDict(LZ4F_cctx* cctx, |
869 | | void* dstBuffer, size_t dstCapacity, |
870 | | const LZ4F_CDict* cdict, |
871 | | const LZ4F_preferences_t* preferencesPtr) |
872 | 0 | { |
873 | 0 | return LZ4F_compressBegin_internal(cctx, dstBuffer, dstCapacity, |
874 | 0 | NULL, 0, |
875 | 0 | cdict, preferencesPtr); |
876 | 0 | } |
877 | | |
878 | | |
879 | | /* LZ4F_compressBound() : |
880 | | * @return minimum capacity of dstBuffer for a given srcSize to handle worst case scenario. |
881 | | * LZ4F_preferences_t structure is optional : if NULL, preferences will be set to cover worst case scenario. |
882 | | * This function cannot fail. |
883 | | */ |
884 | | size_t LZ4F_compressBound(size_t srcSize, const LZ4F_preferences_t* preferencesPtr) |
885 | 0 | { |
886 | 0 | if (preferencesPtr && preferencesPtr->autoFlush) { |
887 | 0 | return LZ4F_compressBound_internal(srcSize, preferencesPtr, 0); |
888 | 0 | } |
889 | 0 | return LZ4F_compressBound_internal(srcSize, preferencesPtr, (size_t)-1); |
890 | 0 | } |
891 | | |
892 | | |
893 | | typedef int (*compressFunc_t)(void* ctx, const char* src, char* dst, int srcSize, int dstSize, int level, const LZ4F_CDict* cdict); |
894 | | |
895 | | |
896 | | /*! LZ4F_makeBlock(): |
897 | | * compress a single block, add header and optional checksum. |
898 | | * assumption : dst buffer capacity is >= BHSize + srcSize + crcSize |
899 | | */ |
900 | | static size_t LZ4F_makeBlock(void* dst, |
901 | | const void* src, size_t srcSize, |
902 | | compressFunc_t compress, void* lz4ctx, int level, |
903 | | const LZ4F_CDict* cdict, |
904 | | LZ4F_blockChecksum_t crcFlag) |
905 | 0 | { |
906 | 0 | BYTE* const cSizePtr = (BYTE*)dst; |
907 | 0 | int dstCapacity = (srcSize > 1) ? (int)srcSize - 1 : 1; |
908 | 0 | U32 cSize; |
909 | 0 | assert(compress != NULL); |
910 | 0 | cSize = (U32)compress(lz4ctx, (const char*)src, (char*)(cSizePtr+BHSize), |
911 | 0 | (int)srcSize, dstCapacity, |
912 | 0 | level, cdict); |
913 | |
|
914 | 0 | if (cSize == 0 || cSize >= srcSize) { |
915 | 0 | cSize = (U32)srcSize; |
916 | 0 | LZ4F_writeLE32(cSizePtr, cSize | LZ4F_BLOCKUNCOMPRESSED_FLAG); |
917 | 0 | memcpy(cSizePtr+BHSize, src, srcSize); |
918 | 0 | } else { |
919 | 0 | LZ4F_writeLE32(cSizePtr, cSize); |
920 | 0 | } |
921 | 0 | if (crcFlag) { |
922 | 0 | U32 const crc32 = XXH32(cSizePtr+BHSize, cSize, 0); /* checksum of compressed data */ |
923 | 0 | LZ4F_writeLE32(cSizePtr+BHSize+cSize, crc32); |
924 | 0 | } |
925 | 0 | return BHSize + cSize + ((U32)crcFlag)*BFSize; |
926 | 0 | } |
927 | | |
928 | | |
929 | | static int LZ4F_compressBlock(void* ctx, const char* src, char* dst, int srcSize, int dstCapacity, int level, const LZ4F_CDict* cdict) |
930 | 0 | { |
931 | 0 | int const acceleration = (level < 0) ? -level + 1 : 1; |
932 | 0 | DEBUGLOG(5, "LZ4F_compressBlock (srcSize=%i)", srcSize); |
933 | 0 | LZ4F_initStream(ctx, cdict, level, LZ4F_blockIndependent); |
934 | 0 | if (cdict) { |
935 | 0 | return LZ4_compress_fast_continue((LZ4_stream_t*)ctx, src, dst, srcSize, dstCapacity, acceleration); |
936 | 0 | } else { |
937 | 0 | return LZ4_compress_fast_extState_fastReset(ctx, src, dst, srcSize, dstCapacity, acceleration); |
938 | 0 | } |
939 | 0 | } |
940 | | |
941 | | static int LZ4F_compressBlock_continue(void* ctx, const char* src, char* dst, int srcSize, int dstCapacity, int level, const LZ4F_CDict* cdict) |
942 | 0 | { |
943 | 0 | int const acceleration = (level < 0) ? -level + 1 : 1; |
944 | 0 | (void)cdict; /* init once at beginning of frame */ |
945 | 0 | DEBUGLOG(5, "LZ4F_compressBlock_continue (srcSize=%i)", srcSize); |
946 | 0 | return LZ4_compress_fast_continue((LZ4_stream_t*)ctx, src, dst, srcSize, dstCapacity, acceleration); |
947 | 0 | } |
948 | | |
949 | | static int LZ4F_compressBlockHC(void* ctx, const char* src, char* dst, int srcSize, int dstCapacity, int level, const LZ4F_CDict* cdict) |
950 | 0 | { |
951 | 0 | LZ4F_initStream(ctx, cdict, level, LZ4F_blockIndependent); |
952 | 0 | if (cdict) { |
953 | 0 | return LZ4_compress_HC_continue((LZ4_streamHC_t*)ctx, src, dst, srcSize, dstCapacity); |
954 | 0 | } |
955 | 0 | return LZ4_compress_HC_extStateHC_fastReset(ctx, src, dst, srcSize, dstCapacity, level); |
956 | 0 | } |
957 | | |
958 | | static int LZ4F_compressBlockHC_continue(void* ctx, const char* src, char* dst, int srcSize, int dstCapacity, int level, const LZ4F_CDict* cdict) |
959 | 0 | { |
960 | 0 | (void)level; (void)cdict; /* init once at beginning of frame */ |
961 | 0 | return LZ4_compress_HC_continue((LZ4_streamHC_t*)ctx, src, dst, srcSize, dstCapacity); |
962 | 0 | } |
963 | | |
964 | | static int LZ4F_doNotCompressBlock(void* ctx, const char* src, char* dst, int srcSize, int dstCapacity, int level, const LZ4F_CDict* cdict) |
965 | 0 | { |
966 | 0 | (void)ctx; (void)src; (void)dst; (void)srcSize; (void)dstCapacity; (void)level; (void)cdict; |
967 | 0 | return 0; |
968 | 0 | } |
969 | | |
970 | | static compressFunc_t LZ4F_selectCompression(LZ4F_blockMode_t blockMode, int level, LZ4F_BlockCompressMode_e compressMode) |
971 | 0 | { |
972 | 0 | if (compressMode == LZ4B_UNCOMPRESSED) |
973 | 0 | return LZ4F_doNotCompressBlock; |
974 | 0 | if (level < LZ4HC_CLEVEL_MIN) { |
975 | 0 | if (blockMode == LZ4F_blockIndependent) return LZ4F_compressBlock; |
976 | 0 | return LZ4F_compressBlock_continue; |
977 | 0 | } |
978 | 0 | if (blockMode == LZ4F_blockIndependent) return LZ4F_compressBlockHC; |
979 | 0 | return LZ4F_compressBlockHC_continue; |
980 | 0 | } |
981 | | |
982 | | /* Save or shorten history (up to 64KB) into @tmpBuff */ |
983 | | static void LZ4F_localSaveDict(LZ4F_cctx_t* cctxPtr) |
984 | 0 | { |
985 | 0 | int const dictSize = (cctxPtr->prefs.compressionLevel < LZ4HC_CLEVEL_MIN) ? |
986 | 0 | LZ4_saveDict ((LZ4_stream_t*)(cctxPtr->lz4CtxPtr), (char*)(cctxPtr->tmpBuff), 64 KB) : |
987 | 0 | LZ4_saveDictHC ((LZ4_streamHC_t*)(cctxPtr->lz4CtxPtr), (char*)(cctxPtr->tmpBuff), 64 KB); |
988 | 0 | cctxPtr->tmpIn = cctxPtr->tmpBuff + dictSize; |
989 | 0 | } |
990 | | |
991 | | typedef enum { notDone, fromTmpBuffer, fromSrcBuffer } LZ4F_lastBlockStatus; |
992 | | |
993 | | static const LZ4F_compressOptions_t k_cOptionsNull = { 0, { 0, 0, 0 } }; |
994 | | |
995 | | |
996 | | /*! LZ4F_compressUpdateImpl() : |
997 | | * LZ4F_compressUpdate() can be called repetitively to compress as much data as necessary. |
998 | | * When successful, the function always entirely consumes @srcBuffer. |
999 | | * src data is either buffered or compressed into @dstBuffer. |
1000 | | * If the block compression does not match the compression of the previous block, the old data is flushed |
1001 | | * and operations continue with the new compression mode. |
1002 | | * @dstCapacity MUST be >= LZ4F_compressBound(srcSize, preferencesPtr) when block compression is turned on. |
1003 | | * @compressOptionsPtr is optional : provide NULL to mean "default". |
1004 | | * @return : the number of bytes written into dstBuffer. It can be zero, meaning input data was just buffered. |
1005 | | * or an error code if it fails (which can be tested using LZ4F_isError()) |
1006 | | * After an error, the state is left in a UB state, and must be re-initialized. |
1007 | | */ |
1008 | | static size_t LZ4F_compressUpdateImpl(LZ4F_cctx* cctxPtr, |
1009 | | void* dstBuffer, size_t dstCapacity, |
1010 | | const void* srcBuffer, size_t srcSize, |
1011 | | const LZ4F_compressOptions_t* compressOptionsPtr, |
1012 | | LZ4F_BlockCompressMode_e blockCompression) |
1013 | 0 | { |
1014 | 0 | size_t const blockSize = cctxPtr->maxBlockSize; |
1015 | 0 | const BYTE* srcPtr = (const BYTE*)srcBuffer; |
1016 | 0 | const BYTE* const srcEnd = srcSize ? (assert(srcPtr!=NULL), srcPtr + srcSize) : srcPtr; |
1017 | 0 | BYTE* const dstStart = (BYTE*)dstBuffer; |
1018 | 0 | BYTE* dstPtr = dstStart; |
1019 | 0 | LZ4F_lastBlockStatus lastBlockCompressed = notDone; |
1020 | 0 | compressFunc_t const compress = LZ4F_selectCompression(cctxPtr->prefs.frameInfo.blockMode, cctxPtr->prefs.compressionLevel, blockCompression); |
1021 | 0 | size_t bytesWritten; |
1022 | 0 | DEBUGLOG(4, "LZ4F_compressUpdate (srcSize=%zu)", srcSize); |
1023 | |
|
1024 | 0 | RETURN_ERROR_IF(cctxPtr->cStage != 1, compressionState_uninitialized); /* state must be initialized and waiting for next block */ |
1025 | 0 | if (dstCapacity < LZ4F_compressBound_internal(srcSize, &(cctxPtr->prefs), cctxPtr->tmpInSize)) |
1026 | 0 | RETURN_ERROR(dstMaxSize_tooSmall); |
1027 | | |
1028 | 0 | if (blockCompression == LZ4B_UNCOMPRESSED && dstCapacity < srcSize) |
1029 | 0 | RETURN_ERROR(dstMaxSize_tooSmall); |
1030 | | |
1031 | | /* flush currently written block, to continue with new block compression */ |
1032 | 0 | if (cctxPtr->blockCompressMode != blockCompression) { |
1033 | 0 | bytesWritten = LZ4F_flush(cctxPtr, dstBuffer, dstCapacity, compressOptionsPtr); |
1034 | 0 | dstPtr += bytesWritten; |
1035 | 0 | cctxPtr->blockCompressMode = blockCompression; |
1036 | 0 | } |
1037 | |
|
1038 | 0 | if (compressOptionsPtr == NULL) compressOptionsPtr = &k_cOptionsNull; |
1039 | | |
1040 | | /* complete tmp buffer */ |
1041 | 0 | if (cctxPtr->tmpInSize > 0) { /* some data already within tmp buffer */ |
1042 | 0 | size_t const sizeToCopy = blockSize - cctxPtr->tmpInSize; |
1043 | 0 | assert(blockSize > cctxPtr->tmpInSize); |
1044 | 0 | if (sizeToCopy > srcSize) { |
1045 | | /* add src to tmpIn buffer */ |
1046 | 0 | memcpy(cctxPtr->tmpIn + cctxPtr->tmpInSize, srcBuffer, srcSize); |
1047 | 0 | srcPtr = srcEnd; |
1048 | 0 | cctxPtr->tmpInSize += srcSize; |
1049 | | /* still needs some CRC */ |
1050 | 0 | } else { |
1051 | | /* complete tmpIn block and then compress it */ |
1052 | 0 | lastBlockCompressed = fromTmpBuffer; |
1053 | 0 | memcpy(cctxPtr->tmpIn + cctxPtr->tmpInSize, srcBuffer, sizeToCopy); |
1054 | 0 | srcPtr += sizeToCopy; |
1055 | |
|
1056 | 0 | dstPtr += LZ4F_makeBlock(dstPtr, |
1057 | 0 | cctxPtr->tmpIn, blockSize, |
1058 | 0 | compress, cctxPtr->lz4CtxPtr, cctxPtr->prefs.compressionLevel, |
1059 | 0 | cctxPtr->cdict, |
1060 | 0 | cctxPtr->prefs.frameInfo.blockChecksumFlag); |
1061 | 0 | if (cctxPtr->prefs.frameInfo.blockMode==LZ4F_blockLinked) cctxPtr->tmpIn += blockSize; |
1062 | 0 | cctxPtr->tmpInSize = 0; |
1063 | 0 | } } |
1064 | |
|
1065 | 0 | while ((size_t)(srcEnd - srcPtr) >= blockSize) { |
1066 | | /* compress full blocks */ |
1067 | 0 | lastBlockCompressed = fromSrcBuffer; |
1068 | 0 | dstPtr += LZ4F_makeBlock(dstPtr, |
1069 | 0 | srcPtr, blockSize, |
1070 | 0 | compress, cctxPtr->lz4CtxPtr, cctxPtr->prefs.compressionLevel, |
1071 | 0 | cctxPtr->cdict, |
1072 | 0 | cctxPtr->prefs.frameInfo.blockChecksumFlag); |
1073 | 0 | srcPtr += blockSize; |
1074 | 0 | } |
1075 | |
|
1076 | 0 | if ((cctxPtr->prefs.autoFlush) && (srcPtr < srcEnd)) { |
1077 | | /* autoFlush : remaining input (< blockSize) is compressed */ |
1078 | 0 | lastBlockCompressed = fromSrcBuffer; |
1079 | 0 | dstPtr += LZ4F_makeBlock(dstPtr, |
1080 | 0 | srcPtr, (size_t)(srcEnd - srcPtr), |
1081 | 0 | compress, cctxPtr->lz4CtxPtr, cctxPtr->prefs.compressionLevel, |
1082 | 0 | cctxPtr->cdict, |
1083 | 0 | cctxPtr->prefs.frameInfo.blockChecksumFlag); |
1084 | 0 | srcPtr = srcEnd; |
1085 | 0 | } |
1086 | | |
1087 | | /* preserve dictionary within @tmpBuff whenever necessary */ |
1088 | 0 | if ((cctxPtr->prefs.frameInfo.blockMode==LZ4F_blockLinked) && (lastBlockCompressed==fromSrcBuffer)) { |
1089 | | /* linked blocks are only supported in compressed mode, see LZ4F_uncompressedUpdate */ |
1090 | 0 | assert(blockCompression == LZ4B_COMPRESSED); |
1091 | 0 | if (compressOptionsPtr->stableSrc) { |
1092 | 0 | cctxPtr->tmpIn = cctxPtr->tmpBuff; /* src is stable : dictionary remains in src across invocations */ |
1093 | 0 | } else { |
1094 | 0 | LZ4F_localSaveDict(cctxPtr); |
1095 | 0 | } |
1096 | 0 | } |
1097 | | |
1098 | | /* keep tmpIn within limits */ |
1099 | 0 | if (!(cctxPtr->prefs.autoFlush) /* no autoflush : there may be some data left within internal buffer */ |
1100 | 0 | && (cctxPtr->tmpIn + blockSize) > (cctxPtr->tmpBuff + cctxPtr->maxBufferSize) ) /* not enough room to store next block */ |
1101 | 0 | { |
1102 | | /* only preserve 64KB within internal buffer. Ensures there is enough room for next block. |
1103 | | * note: this situation necessarily implies lastBlockCompressed==fromTmpBuffer */ |
1104 | 0 | LZ4F_localSaveDict(cctxPtr); |
1105 | 0 | assert((cctxPtr->tmpIn + blockSize) <= (cctxPtr->tmpBuff + cctxPtr->maxBufferSize)); |
1106 | 0 | } |
1107 | | |
1108 | | /* some input data left, necessarily < blockSize */ |
1109 | 0 | if (srcPtr < srcEnd) { |
1110 | | /* fill tmp buffer */ |
1111 | 0 | size_t const sizeToCopy = (size_t)(srcEnd - srcPtr); |
1112 | 0 | memcpy(cctxPtr->tmpIn, srcPtr, sizeToCopy); |
1113 | 0 | cctxPtr->tmpInSize = sizeToCopy; |
1114 | 0 | } |
1115 | |
|
1116 | 0 | if (cctxPtr->prefs.frameInfo.contentChecksumFlag == LZ4F_contentChecksumEnabled) |
1117 | 0 | (void)XXH32_update(&(cctxPtr->xxh), srcBuffer, srcSize); |
1118 | |
|
1119 | 0 | cctxPtr->totalInSize += srcSize; |
1120 | 0 | return (size_t)(dstPtr - dstStart); |
1121 | 0 | } |
1122 | | |
1123 | | /*! LZ4F_compressUpdate() : |
1124 | | * LZ4F_compressUpdate() can be called repetitively to compress as much data as necessary. |
1125 | | * When successful, the function always entirely consumes @srcBuffer. |
1126 | | * src data is either buffered or compressed into @dstBuffer. |
1127 | | * If previously an uncompressed block was written, buffered data is flushed |
1128 | | * before appending compressed data is continued. |
1129 | | * @dstCapacity MUST be >= LZ4F_compressBound(srcSize, preferencesPtr). |
1130 | | * @compressOptionsPtr is optional : provide NULL to mean "default". |
1131 | | * @return : the number of bytes written into dstBuffer. It can be zero, meaning input data was just buffered. |
1132 | | * or an error code if it fails (which can be tested using LZ4F_isError()) |
1133 | | * After an error, the state is left in a UB state, and must be re-initialized. |
1134 | | */ |
1135 | | size_t LZ4F_compressUpdate(LZ4F_cctx* cctxPtr, |
1136 | | void* dstBuffer, size_t dstCapacity, |
1137 | | const void* srcBuffer, size_t srcSize, |
1138 | | const LZ4F_compressOptions_t* compressOptionsPtr) |
1139 | 0 | { |
1140 | 0 | return LZ4F_compressUpdateImpl(cctxPtr, |
1141 | 0 | dstBuffer, dstCapacity, |
1142 | 0 | srcBuffer, srcSize, |
1143 | 0 | compressOptionsPtr, LZ4B_COMPRESSED); |
1144 | 0 | } |
1145 | | |
1146 | | /*! LZ4F_uncompressedUpdate() : |
1147 | | * Same as LZ4F_compressUpdate(), but requests blocks to be sent uncompressed. |
1148 | | * This symbol is only supported when LZ4F_blockIndependent is used |
1149 | | * @dstCapacity MUST be >= LZ4F_compressBound(srcSize, preferencesPtr). |
1150 | | * @compressOptionsPtr is optional : provide NULL to mean "default". |
1151 | | * @return : the number of bytes written into dstBuffer. It can be zero, meaning input data was just buffered. |
1152 | | * or an error code if it fails (which can be tested using LZ4F_isError()) |
1153 | | * After an error, the state is left in a UB state, and must be re-initialized. |
1154 | | */ |
1155 | | size_t LZ4F_uncompressedUpdate(LZ4F_cctx* cctxPtr, |
1156 | | void* dstBuffer, size_t dstCapacity, |
1157 | | const void* srcBuffer, size_t srcSize, |
1158 | | const LZ4F_compressOptions_t* compressOptionsPtr) |
1159 | 0 | { |
1160 | 0 | return LZ4F_compressUpdateImpl(cctxPtr, |
1161 | 0 | dstBuffer, dstCapacity, |
1162 | 0 | srcBuffer, srcSize, |
1163 | 0 | compressOptionsPtr, LZ4B_UNCOMPRESSED); |
1164 | 0 | } |
1165 | | |
1166 | | |
1167 | | /*! LZ4F_flush() : |
1168 | | * When compressed data must be sent immediately, without waiting for a block to be filled, |
1169 | | * invoke LZ4F_flush(), which will immediately compress any remaining data stored within LZ4F_cctx. |
1170 | | * The result of the function is the number of bytes written into dstBuffer. |
1171 | | * It can be zero, this means there was no data left within LZ4F_cctx. |
1172 | | * The function outputs an error code if it fails (can be tested using LZ4F_isError()) |
1173 | | * LZ4F_compressOptions_t* is optional. NULL is a valid argument. |
1174 | | */ |
1175 | | size_t LZ4F_flush(LZ4F_cctx* cctxPtr, |
1176 | | void* dstBuffer, size_t dstCapacity, |
1177 | | const LZ4F_compressOptions_t* compressOptionsPtr) |
1178 | 0 | { |
1179 | 0 | BYTE* const dstStart = (BYTE*)dstBuffer; |
1180 | 0 | BYTE* dstPtr = dstStart; |
1181 | 0 | compressFunc_t compress; |
1182 | |
|
1183 | 0 | DEBUGLOG(5, "LZ4F_flush: %zu buffered bytes (saved dict size = %i) (dstCapacity=%u)", |
1184 | 0 | cctxPtr->tmpInSize, (int)(cctxPtr->tmpIn - cctxPtr->tmpBuff), (unsigned)dstCapacity); |
1185 | 0 | if (cctxPtr->tmpInSize == 0) return 0; /* nothing to flush */ |
1186 | 0 | RETURN_ERROR_IF(cctxPtr->cStage != 1, compressionState_uninitialized); |
1187 | 0 | RETURN_ERROR_IF(dstCapacity < (cctxPtr->tmpInSize + BHSize + BFSize), dstMaxSize_tooSmall); |
1188 | 0 | (void)compressOptionsPtr; /* not useful (yet) */ |
1189 | | |
1190 | | /* select compression function */ |
1191 | 0 | compress = LZ4F_selectCompression(cctxPtr->prefs.frameInfo.blockMode, cctxPtr->prefs.compressionLevel, cctxPtr->blockCompressMode); |
1192 | | |
1193 | | /* compress tmp buffer */ |
1194 | 0 | dstPtr += LZ4F_makeBlock(dstPtr, |
1195 | 0 | cctxPtr->tmpIn, cctxPtr->tmpInSize, |
1196 | 0 | compress, cctxPtr->lz4CtxPtr, cctxPtr->prefs.compressionLevel, |
1197 | 0 | cctxPtr->cdict, |
1198 | 0 | cctxPtr->prefs.frameInfo.blockChecksumFlag); |
1199 | 0 | assert(((void)"flush overflows dstBuffer!", (size_t)(dstPtr - dstStart) <= dstCapacity)); |
1200 | |
|
1201 | 0 | if (cctxPtr->prefs.frameInfo.blockMode == LZ4F_blockLinked) |
1202 | 0 | cctxPtr->tmpIn += cctxPtr->tmpInSize; |
1203 | 0 | cctxPtr->tmpInSize = 0; |
1204 | | |
1205 | | /* keep tmpIn within limits */ |
1206 | 0 | if ((cctxPtr->tmpIn + cctxPtr->maxBlockSize) > (cctxPtr->tmpBuff + cctxPtr->maxBufferSize)) { |
1207 | 0 | assert(cctxPtr->prefs.frameInfo.blockMode == LZ4F_blockLinked); |
1208 | 0 | LZ4F_localSaveDict(cctxPtr); |
1209 | 0 | } |
1210 | |
|
1211 | 0 | return (size_t)(dstPtr - dstStart); |
1212 | 0 | } |
1213 | | |
1214 | | |
1215 | | /*! LZ4F_compressEnd() : |
1216 | | * When you want to properly finish the compressed frame, just call LZ4F_compressEnd(). |
1217 | | * It will flush whatever data remained within compressionContext (like LZ4_flush()) |
1218 | | * but also properly finalize the frame, with an endMark and an (optional) checksum. |
1219 | | * LZ4F_compressOptions_t structure is optional : you can provide NULL as argument. |
1220 | | * @return: the number of bytes written into dstBuffer (necessarily >= 4 (endMark size)) |
1221 | | * or an error code if it fails (can be tested using LZ4F_isError()) |
1222 | | * The context can then be used again to compress a new frame, starting with LZ4F_compressBegin(). |
1223 | | */ |
1224 | | size_t LZ4F_compressEnd(LZ4F_cctx* cctxPtr, |
1225 | | void* dstBuffer, size_t dstCapacity, |
1226 | | const LZ4F_compressOptions_t* compressOptionsPtr) |
1227 | 0 | { |
1228 | 0 | BYTE* const dstStart = (BYTE*)dstBuffer; |
1229 | 0 | BYTE* dstPtr = dstStart; |
1230 | |
|
1231 | 0 | size_t const flushSize = LZ4F_flush(cctxPtr, dstBuffer, dstCapacity, compressOptionsPtr); |
1232 | 0 | DEBUGLOG(5,"LZ4F_compressEnd: dstCapacity=%u", (unsigned)dstCapacity); |
1233 | 0 | FORWARD_IF_ERROR(flushSize); |
1234 | 0 | dstPtr += flushSize; |
1235 | |
|
1236 | 0 | assert(flushSize <= dstCapacity); |
1237 | 0 | dstCapacity -= flushSize; |
1238 | |
|
1239 | 0 | RETURN_ERROR_IF(dstCapacity < 4, dstMaxSize_tooSmall); |
1240 | 0 | LZ4F_writeLE32(dstPtr, 0); |
1241 | 0 | dstPtr += 4; /* endMark */ |
1242 | |
|
1243 | 0 | if (cctxPtr->prefs.frameInfo.contentChecksumFlag == LZ4F_contentChecksumEnabled) { |
1244 | 0 | U32 const xxh = XXH32_digest(&(cctxPtr->xxh)); |
1245 | 0 | RETURN_ERROR_IF(dstCapacity < 8, dstMaxSize_tooSmall); |
1246 | 0 | DEBUGLOG(5,"Writing 32-bit content checksum (0x%0X)", xxh); |
1247 | 0 | LZ4F_writeLE32(dstPtr, xxh); |
1248 | 0 | dstPtr+=4; /* content Checksum */ |
1249 | 0 | } |
1250 | | |
1251 | 0 | cctxPtr->cStage = 0; /* state is now re-usable (with identical preferences) */ |
1252 | |
|
1253 | 0 | if (cctxPtr->prefs.frameInfo.contentSize) { |
1254 | 0 | if (cctxPtr->prefs.frameInfo.contentSize != cctxPtr->totalInSize) |
1255 | 0 | RETURN_ERROR(frameSize_wrong); |
1256 | 0 | } |
1257 | | |
1258 | 0 | return (size_t)(dstPtr - dstStart); |
1259 | 0 | } |
1260 | | |
1261 | | |
1262 | | /*-*************************************************** |
1263 | | * Frame Decompression |
1264 | | *****************************************************/ |
1265 | | |
1266 | | typedef enum { |
1267 | | dstage_getFrameHeader=0, dstage_storeFrameHeader, |
1268 | | dstage_init, |
1269 | | dstage_getBlockHeader, dstage_storeBlockHeader, |
1270 | | dstage_copyDirect, dstage_getBlockChecksum, |
1271 | | dstage_getCBlock, dstage_storeCBlock, |
1272 | | dstage_flushOut, |
1273 | | dstage_getSuffix, dstage_storeSuffix, |
1274 | | dstage_getSFrameSize, dstage_storeSFrameSize, |
1275 | | dstage_skipSkippable |
1276 | | } dStage_t; |
1277 | | |
1278 | | struct LZ4F_dctx_s { |
1279 | | LZ4F_CustomMem cmem; |
1280 | | LZ4F_frameInfo_t frameInfo; |
1281 | | U32 version; |
1282 | | dStage_t dStage; |
1283 | | U64 frameRemainingSize; |
1284 | | size_t maxBlockSize; |
1285 | | size_t maxBufferSize; |
1286 | | BYTE* tmpIn; |
1287 | | size_t tmpInSize; |
1288 | | size_t tmpInTarget; |
1289 | | BYTE* tmpOutBuffer; |
1290 | | const BYTE* dict; |
1291 | | size_t dictSize; |
1292 | | BYTE* tmpOut; |
1293 | | size_t tmpOutSize; |
1294 | | size_t tmpOutStart; |
1295 | | XXH32_state_t xxh; |
1296 | | XXH32_state_t blockChecksum; |
1297 | | int skipChecksum; |
1298 | | BYTE header[LZ4F_HEADER_SIZE_MAX]; |
1299 | | }; /* typedef'd to LZ4F_dctx in lz4frame.h */ |
1300 | | |
1301 | | |
1302 | | LZ4F_dctx* LZ4F_createDecompressionContext_advanced(LZ4F_CustomMem customMem, unsigned version) |
1303 | 0 | { |
1304 | 0 | LZ4F_dctx* const dctx = (LZ4F_dctx*)LZ4F_calloc(sizeof(LZ4F_dctx), customMem); |
1305 | 0 | if (dctx == NULL) return NULL; |
1306 | | |
1307 | 0 | dctx->cmem = customMem; |
1308 | 0 | dctx->version = version; |
1309 | 0 | return dctx; |
1310 | 0 | } |
1311 | | |
1312 | | /*! LZ4F_createDecompressionContext() : |
1313 | | * Create a decompressionContext object, which will track all decompression operations. |
1314 | | * Provides a pointer to a fully allocated and initialized LZ4F_decompressionContext object. |
1315 | | * Object can later be released using LZ4F_freeDecompressionContext(). |
1316 | | * @return : if != 0, there was an error during context creation. |
1317 | | */ |
1318 | | LZ4F_errorCode_t |
1319 | | LZ4F_createDecompressionContext(LZ4F_dctx** LZ4F_decompressionContextPtr, unsigned versionNumber) |
1320 | 0 | { |
1321 | 0 | assert(LZ4F_decompressionContextPtr != NULL); /* violation of narrow contract */ |
1322 | 0 | RETURN_ERROR_IF(LZ4F_decompressionContextPtr == NULL, parameter_null); /* in case it nonetheless happen in production */ |
1323 | | |
1324 | 0 | *LZ4F_decompressionContextPtr = LZ4F_createDecompressionContext_advanced(LZ4F_defaultCMem, versionNumber); |
1325 | 0 | if (*LZ4F_decompressionContextPtr == NULL) { /* failed allocation */ |
1326 | 0 | RETURN_ERROR(allocation_failed); |
1327 | 0 | } |
1328 | 0 | return LZ4F_OK_NoError; |
1329 | 0 | } |
1330 | | |
1331 | | LZ4F_errorCode_t LZ4F_freeDecompressionContext(LZ4F_dctx* dctx) |
1332 | 0 | { |
1333 | 0 | LZ4F_errorCode_t result = LZ4F_OK_NoError; |
1334 | 0 | if (dctx != NULL) { /* can accept NULL input, like free() */ |
1335 | 0 | result = (LZ4F_errorCode_t)dctx->dStage; |
1336 | 0 | LZ4F_free(dctx->tmpIn, dctx->cmem); |
1337 | 0 | LZ4F_free(dctx->tmpOutBuffer, dctx->cmem); |
1338 | 0 | LZ4F_free(dctx, dctx->cmem); |
1339 | 0 | } |
1340 | 0 | return result; |
1341 | 0 | } |
1342 | | |
1343 | 0 | size_t LZ4F_dctx_size(const LZ4F_dctx* dctx) { |
1344 | 0 | if (dctx == NULL) { |
1345 | 0 | return 0; |
1346 | 0 | } |
1347 | 0 | return sizeof(*dctx) |
1348 | 0 | + (dctx->tmpIn != NULL ? dctx->maxBlockSize + BFSize : 0) |
1349 | 0 | + (dctx->tmpOutBuffer != NULL ? dctx->maxBufferSize : 0); |
1350 | 0 | } |
1351 | | |
1352 | | |
1353 | | /*==--- Streaming Decompression operations ---==*/ |
1354 | | void LZ4F_resetDecompressionContext(LZ4F_dctx* dctx) |
1355 | 0 | { |
1356 | 0 | DEBUGLOG(5, "LZ4F_resetDecompressionContext"); |
1357 | 0 | dctx->dStage = dstage_getFrameHeader; |
1358 | 0 | dctx->dict = NULL; |
1359 | 0 | dctx->dictSize = 0; |
1360 | 0 | dctx->skipChecksum = 0; |
1361 | 0 | dctx->frameRemainingSize = 0; |
1362 | 0 | } |
1363 | | |
1364 | | |
1365 | | /*! LZ4F_decodeHeader() : |
1366 | | * input : `src` points at the **beginning of the frame** |
1367 | | * output : set internal values of dctx, such as |
1368 | | * dctx->frameInfo and dctx->dStage. |
1369 | | * Also allocates internal buffers. |
1370 | | * @return : nb Bytes read from src (necessarily <= srcSize) |
1371 | | * or an error code (testable with LZ4F_isError()) |
1372 | | */ |
1373 | | static size_t LZ4F_decodeHeader(LZ4F_dctx* dctx, const void* src, size_t srcSize) |
1374 | 0 | { |
1375 | 0 | unsigned blockMode, blockChecksumFlag, contentSizeFlag, contentChecksumFlag, dictIDFlag, blockSizeID; |
1376 | 0 | size_t frameHeaderSize; |
1377 | 0 | const BYTE* srcPtr = (const BYTE*)src; |
1378 | |
|
1379 | 0 | DEBUGLOG(5, "LZ4F_decodeHeader"); |
1380 | | /* need to decode header to get frameInfo */ |
1381 | 0 | RETURN_ERROR_IF(srcSize < minFHSize, frameHeader_incomplete); /* minimal frame header size */ |
1382 | 0 | MEM_INIT(&(dctx->frameInfo), 0, sizeof(dctx->frameInfo)); |
1383 | | |
1384 | | /* special case : skippable frames */ |
1385 | 0 | if ((LZ4F_readLE32(srcPtr) & 0xFFFFFFF0U) == LZ4F_MAGIC_SKIPPABLE_START) { |
1386 | 0 | dctx->frameInfo.frameType = LZ4F_skippableFrame; |
1387 | 0 | if (src == (void*)(dctx->header)) { |
1388 | 0 | dctx->tmpInSize = srcSize; |
1389 | 0 | dctx->tmpInTarget = 8; |
1390 | 0 | dctx->dStage = dstage_storeSFrameSize; |
1391 | 0 | return srcSize; |
1392 | 0 | } else { |
1393 | 0 | dctx->dStage = dstage_getSFrameSize; |
1394 | 0 | return 4; |
1395 | 0 | } } |
1396 | | |
1397 | | /* control magic number */ |
1398 | | #ifndef FUZZING_BUILD_MODE_UNSAFE_FOR_PRODUCTION |
1399 | | if (LZ4F_readLE32(srcPtr) != LZ4F_MAGICNUMBER) { |
1400 | | DEBUGLOG(4, "frame header error : unknown magic number"); |
1401 | | RETURN_ERROR(frameType_unknown); |
1402 | | } |
1403 | | #endif |
1404 | 0 | dctx->frameInfo.frameType = LZ4F_frame; |
1405 | | |
1406 | | /* Flags */ |
1407 | 0 | { U32 const FLG = srcPtr[4]; |
1408 | 0 | U32 const version = (FLG>>6) & _2BITS; |
1409 | 0 | blockChecksumFlag = (FLG>>4) & _1BIT; |
1410 | 0 | blockMode = (FLG>>5) & _1BIT; |
1411 | 0 | contentSizeFlag = (FLG>>3) & _1BIT; |
1412 | 0 | contentChecksumFlag = (FLG>>2) & _1BIT; |
1413 | 0 | dictIDFlag = FLG & _1BIT; |
1414 | | /* validate */ |
1415 | 0 | if (((FLG>>1)&_1BIT) != 0) RETURN_ERROR(reservedFlag_set); /* Reserved bit */ |
1416 | 0 | if (version != 1) RETURN_ERROR(headerVersion_wrong); /* Version Number, only supported value */ |
1417 | 0 | } |
1418 | 0 | DEBUGLOG(6, "contentSizeFlag: %u", contentSizeFlag); |
1419 | | |
1420 | | /* Frame Header Size */ |
1421 | 0 | frameHeaderSize = minFHSize + (contentSizeFlag?8:0) + (dictIDFlag?4:0); |
1422 | |
|
1423 | 0 | if (srcSize < frameHeaderSize) { |
1424 | | /* not enough input to fully decode frame header */ |
1425 | 0 | if (srcPtr != dctx->header) |
1426 | 0 | memcpy(dctx->header, srcPtr, srcSize); |
1427 | 0 | dctx->tmpInSize = srcSize; |
1428 | 0 | dctx->tmpInTarget = frameHeaderSize; |
1429 | 0 | dctx->dStage = dstage_storeFrameHeader; |
1430 | 0 | return srcSize; |
1431 | 0 | } |
1432 | | |
1433 | 0 | { U32 const BD = srcPtr[5]; |
1434 | 0 | blockSizeID = (BD>>4) & _3BITS; |
1435 | | /* validate */ |
1436 | 0 | if (((BD>>7)&_1BIT) != 0) RETURN_ERROR(reservedFlag_set); /* Reserved bit */ |
1437 | 0 | if (blockSizeID < 4) RETURN_ERROR(maxBlockSize_invalid); /* 4-7 only supported values for the time being */ |
1438 | 0 | if (((BD>>0)&_4BITS) != 0) RETURN_ERROR(reservedFlag_set); /* Reserved bits */ |
1439 | 0 | } |
1440 | | |
1441 | | /* check header */ |
1442 | 0 | assert(frameHeaderSize > 5); |
1443 | | #ifndef FUZZING_BUILD_MODE_UNSAFE_FOR_PRODUCTION |
1444 | | { BYTE const HC = LZ4F_headerChecksum(srcPtr+4, frameHeaderSize-5); |
1445 | | RETURN_ERROR_IF(HC != srcPtr[frameHeaderSize-1], headerChecksum_invalid); |
1446 | | } |
1447 | | #endif |
1448 | | |
1449 | | /* save */ |
1450 | 0 | dctx->frameInfo.blockMode = (LZ4F_blockMode_t)blockMode; |
1451 | 0 | dctx->frameInfo.blockChecksumFlag = (LZ4F_blockChecksum_t)blockChecksumFlag; |
1452 | 0 | dctx->frameInfo.contentChecksumFlag = (LZ4F_contentChecksum_t)contentChecksumFlag; |
1453 | 0 | dctx->frameInfo.blockSizeID = (LZ4F_blockSizeID_t)blockSizeID; |
1454 | 0 | dctx->maxBlockSize = LZ4F_getBlockSize((LZ4F_blockSizeID_t)blockSizeID); |
1455 | 0 | if (contentSizeFlag) { |
1456 | 0 | dctx->frameRemainingSize = dctx->frameInfo.contentSize = LZ4F_readLE64(srcPtr+6); |
1457 | 0 | } |
1458 | 0 | if (dictIDFlag) |
1459 | 0 | dctx->frameInfo.dictID = LZ4F_readLE32(srcPtr + frameHeaderSize - 5); |
1460 | |
|
1461 | 0 | dctx->dStage = dstage_init; |
1462 | |
|
1463 | 0 | return frameHeaderSize; |
1464 | 0 | } |
1465 | | |
1466 | | |
1467 | | /*! LZ4F_headerSize() : |
1468 | | * @return : size of frame header |
1469 | | * or an error code, which can be tested using LZ4F_isError() |
1470 | | */ |
1471 | | size_t LZ4F_headerSize(const void* src, size_t srcSize) |
1472 | 0 | { |
1473 | 0 | RETURN_ERROR_IF(src == NULL, srcPtr_wrong); |
1474 | | |
1475 | | /* minimal srcSize to determine header size */ |
1476 | 0 | if (srcSize < LZ4F_MIN_SIZE_TO_KNOW_HEADER_LENGTH) |
1477 | 0 | RETURN_ERROR(frameHeader_incomplete); |
1478 | | |
1479 | | /* special case : skippable frames */ |
1480 | 0 | if ((LZ4F_readLE32(src) & 0xFFFFFFF0U) == LZ4F_MAGIC_SKIPPABLE_START) |
1481 | 0 | return 8; |
1482 | | |
1483 | | /* control magic number */ |
1484 | | #ifndef FUZZING_BUILD_MODE_UNSAFE_FOR_PRODUCTION |
1485 | | if (LZ4F_readLE32(src) != LZ4F_MAGICNUMBER) |
1486 | | RETURN_ERROR(frameType_unknown); |
1487 | | #endif |
1488 | | |
1489 | | /* Frame Header Size */ |
1490 | 0 | { BYTE const FLG = ((const BYTE*)src)[4]; |
1491 | 0 | U32 const contentSizeFlag = (FLG>>3) & _1BIT; |
1492 | 0 | U32 const dictIDFlag = FLG & _1BIT; |
1493 | 0 | return minFHSize + (contentSizeFlag?8:0) + (dictIDFlag?4:0); |
1494 | 0 | } |
1495 | 0 | } |
1496 | | |
1497 | | /*! LZ4F_getFrameInfo() : |
1498 | | * This function extracts frame parameters (max blockSize, frame checksum, etc.). |
1499 | | * Usage is optional. Objective is to provide relevant information for allocation purposes. |
1500 | | * This function works in 2 situations : |
1501 | | * - At the beginning of a new frame, in which case it will decode this information from `srcBuffer`, and start the decoding process. |
1502 | | * Amount of input data provided must be large enough to successfully decode the frame header. |
1503 | | * A header size is variable, but is guaranteed to be <= LZ4F_HEADER_SIZE_MAX bytes. It's possible to provide more input data than this minimum. |
1504 | | * - After decoding has been started. In which case, no input is read, frame parameters are extracted from dctx. |
1505 | | * The number of bytes consumed from srcBuffer will be updated within *srcSizePtr (necessarily <= original value). |
1506 | | * Decompression must resume from (srcBuffer + *srcSizePtr). |
1507 | | * @return : an hint about how many srcSize bytes LZ4F_decompress() expects for next call, |
1508 | | * or an error code which can be tested using LZ4F_isError() |
1509 | | * note 1 : in case of error, dctx is not modified. Decoding operations can resume from where they stopped. |
1510 | | * note 2 : frame parameters are *copied into* an already allocated LZ4F_frameInfo_t structure. |
1511 | | */ |
1512 | | LZ4F_errorCode_t LZ4F_getFrameInfo(LZ4F_dctx* dctx, |
1513 | | LZ4F_frameInfo_t* frameInfoPtr, |
1514 | | const void* srcBuffer, size_t* srcSizePtr) |
1515 | 0 | { |
1516 | 0 | assert(dctx != NULL); |
1517 | 0 | RETURN_ERROR_IF(frameInfoPtr == NULL, parameter_null); |
1518 | 0 | RETURN_ERROR_IF(srcSizePtr == NULL, parameter_null); |
1519 | | |
1520 | 0 | LZ4F_STATIC_ASSERT(dstage_getFrameHeader < dstage_storeFrameHeader); |
1521 | 0 | if (dctx->dStage > dstage_storeFrameHeader) { |
1522 | | /* frameInfo already decoded */ |
1523 | 0 | size_t o=0, i=0; |
1524 | 0 | *srcSizePtr = 0; |
1525 | 0 | *frameInfoPtr = dctx->frameInfo; |
1526 | | /* returns : recommended nb of bytes for LZ4F_decompress() */ |
1527 | 0 | return LZ4F_decompress(dctx, NULL, &o, NULL, &i, NULL); |
1528 | 0 | } else { |
1529 | 0 | if (dctx->dStage == dstage_storeFrameHeader) { |
1530 | | /* frame decoding already started, in the middle of header => automatic fail */ |
1531 | 0 | *srcSizePtr = 0; |
1532 | 0 | RETURN_ERROR(frameDecoding_alreadyStarted); |
1533 | 0 | } else { |
1534 | 0 | size_t const hSize = LZ4F_headerSize(srcBuffer, *srcSizePtr); |
1535 | 0 | if (LZ4F_isError(hSize)) { *srcSizePtr=0; return hSize; } |
1536 | 0 | if (*srcSizePtr < hSize) { |
1537 | 0 | *srcSizePtr=0; |
1538 | 0 | RETURN_ERROR(frameHeader_incomplete); |
1539 | 0 | } |
1540 | | |
1541 | 0 | { size_t decodeResult = LZ4F_decodeHeader(dctx, srcBuffer, hSize); |
1542 | 0 | if (LZ4F_isError(decodeResult)) { |
1543 | 0 | *srcSizePtr = 0; |
1544 | 0 | } else { |
1545 | 0 | *srcSizePtr = decodeResult; |
1546 | 0 | decodeResult = BHSize; /* block header size */ |
1547 | 0 | } |
1548 | 0 | *frameInfoPtr = dctx->frameInfo; |
1549 | 0 | return decodeResult; |
1550 | 0 | } } } |
1551 | 0 | } |
1552 | | |
1553 | | |
1554 | | /* LZ4F_updateDict() : |
1555 | | * only used for LZ4F_blockLinked mode |
1556 | | * Condition : @dstPtr != NULL |
1557 | | */ |
1558 | | static void LZ4F_updateDict(LZ4F_dctx* dctx, |
1559 | | const BYTE* dstPtr, size_t dstSize, const BYTE* dstBufferStart, |
1560 | | unsigned withinTmp) |
1561 | 0 | { |
1562 | 0 | assert(dstPtr != NULL); |
1563 | 0 | if (dctx->dictSize==0) dctx->dict = (const BYTE*)dstPtr; /* will lead to prefix mode */ |
1564 | 0 | assert(dctx->dict != NULL); |
1565 | |
|
1566 | 0 | if (dctx->dict + dctx->dictSize == dstPtr) { /* prefix mode, everything within dstBuffer */ |
1567 | 0 | dctx->dictSize += dstSize; |
1568 | 0 | return; |
1569 | 0 | } |
1570 | | |
1571 | 0 | assert(dstPtr >= dstBufferStart); |
1572 | 0 | if ((size_t)(dstPtr - dstBufferStart) + dstSize >= 64 KB) { /* history in dstBuffer becomes large enough to become dictionary */ |
1573 | 0 | dctx->dict = (const BYTE*)dstBufferStart; |
1574 | 0 | dctx->dictSize = (size_t)(dstPtr - dstBufferStart) + dstSize; |
1575 | 0 | return; |
1576 | 0 | } |
1577 | | |
1578 | 0 | assert(dstSize < 64 KB); /* if dstSize >= 64 KB, dictionary would be set into dstBuffer directly */ |
1579 | | |
1580 | | /* dstBuffer does not contain whole useful history (64 KB), so it must be saved within tmpOutBuffer */ |
1581 | 0 | assert(dctx->tmpOutBuffer != NULL); |
1582 | |
|
1583 | 0 | if (withinTmp && (dctx->dict == dctx->tmpOutBuffer)) { /* continue history within tmpOutBuffer */ |
1584 | | /* withinTmp expectation : content of [dstPtr,dstSize] is same as [dict+dictSize,dstSize], so we just extend it */ |
1585 | 0 | assert(dctx->dict + dctx->dictSize == dctx->tmpOut + dctx->tmpOutStart); |
1586 | 0 | dctx->dictSize += dstSize; |
1587 | 0 | return; |
1588 | 0 | } |
1589 | | |
1590 | 0 | if (withinTmp) { /* copy relevant dict portion in front of tmpOut within tmpOutBuffer */ |
1591 | 0 | size_t const preserveSize = (size_t)(dctx->tmpOut - dctx->tmpOutBuffer); |
1592 | 0 | size_t copySize = 64 KB - dctx->tmpOutSize; |
1593 | 0 | const BYTE* const oldDictEnd = dctx->dict + dctx->dictSize - dctx->tmpOutStart; |
1594 | 0 | if (dctx->tmpOutSize > 64 KB) copySize = 0; |
1595 | 0 | if (copySize > preserveSize) copySize = preserveSize; |
1596 | |
|
1597 | 0 | memcpy(dctx->tmpOutBuffer + preserveSize - copySize, oldDictEnd - copySize, copySize); |
1598 | |
|
1599 | 0 | dctx->dict = dctx->tmpOutBuffer; |
1600 | 0 | dctx->dictSize = preserveSize + dctx->tmpOutStart + dstSize; |
1601 | 0 | return; |
1602 | 0 | } |
1603 | | |
1604 | 0 | if (dctx->dict == dctx->tmpOutBuffer) { /* copy dst into tmp to complete dict */ |
1605 | 0 | if (dctx->dictSize + dstSize > dctx->maxBufferSize) { /* tmp buffer not large enough */ |
1606 | 0 | size_t const preserveSize = 64 KB - dstSize; |
1607 | 0 | memcpy(dctx->tmpOutBuffer, dctx->dict + dctx->dictSize - preserveSize, preserveSize); |
1608 | 0 | dctx->dictSize = preserveSize; |
1609 | 0 | } |
1610 | 0 | memcpy(dctx->tmpOutBuffer + dctx->dictSize, dstPtr, dstSize); |
1611 | 0 | dctx->dictSize += dstSize; |
1612 | 0 | return; |
1613 | 0 | } |
1614 | | |
1615 | | /* join dict & dest into tmp */ |
1616 | 0 | { size_t preserveSize = 64 KB - dstSize; |
1617 | 0 | if (preserveSize > dctx->dictSize) preserveSize = dctx->dictSize; |
1618 | 0 | memcpy(dctx->tmpOutBuffer, dctx->dict + dctx->dictSize - preserveSize, preserveSize); |
1619 | 0 | memcpy(dctx->tmpOutBuffer + preserveSize, dstPtr, dstSize); |
1620 | 0 | dctx->dict = dctx->tmpOutBuffer; |
1621 | 0 | dctx->dictSize = preserveSize + dstSize; |
1622 | 0 | } |
1623 | 0 | } |
1624 | | |
1625 | | |
1626 | | /*! LZ4F_decompress() : |
1627 | | * Call this function repetitively to regenerate compressed data in srcBuffer. |
1628 | | * The function will attempt to decode up to *srcSizePtr bytes from srcBuffer |
1629 | | * into dstBuffer of capacity *dstSizePtr. |
1630 | | * |
1631 | | * The number of bytes regenerated into dstBuffer will be provided within *dstSizePtr (necessarily <= original value). |
1632 | | * |
1633 | | * The number of bytes effectively read from srcBuffer will be provided within *srcSizePtr (necessarily <= original value). |
1634 | | * If number of bytes read is < number of bytes provided, then decompression operation is not complete. |
1635 | | * Remaining data will have to be presented again in a subsequent invocation. |
1636 | | * |
1637 | | * The function result is an hint of the better srcSize to use for next call to LZ4F_decompress. |
1638 | | * Schematically, it's the size of the current (or remaining) compressed block + header of next block. |
1639 | | * Respecting the hint provides a small boost to performance, since it allows less buffer shuffling. |
1640 | | * Note that this is just a hint, and it's always possible to any srcSize value. |
1641 | | * When a frame is fully decoded, @return will be 0. |
1642 | | * If decompression failed, @return is an error code which can be tested using LZ4F_isError(). |
1643 | | */ |
1644 | | size_t LZ4F_decompress(LZ4F_dctx* dctx, |
1645 | | void* dstBuffer, size_t* dstSizePtr, |
1646 | | const void* srcBuffer, size_t* srcSizePtr, |
1647 | | const LZ4F_decompressOptions_t* decompressOptionsPtr) |
1648 | 0 | { |
1649 | 0 | LZ4F_decompressOptions_t optionsNull; |
1650 | 0 | const BYTE* const srcStart = (const BYTE*)srcBuffer; |
1651 | 0 | const BYTE* const srcEnd = srcStart + *srcSizePtr; |
1652 | 0 | const BYTE* srcPtr = srcStart; |
1653 | 0 | BYTE* const dstStart = (BYTE*)dstBuffer; |
1654 | 0 | BYTE* const dstEnd = dstStart ? dstStart + *dstSizePtr : NULL; |
1655 | 0 | BYTE* dstPtr = dstStart; |
1656 | 0 | const BYTE* selectedIn = NULL; |
1657 | 0 | unsigned doAnotherStage = 1; |
1658 | 0 | size_t nextSrcSizeHint = 1; |
1659 | | |
1660 | |
|
1661 | 0 | DEBUGLOG(5, "LZ4F_decompress: src[%p](%u) => dst[%p](%u)", |
1662 | 0 | srcBuffer, (unsigned)*srcSizePtr, dstBuffer, (unsigned)*dstSizePtr); |
1663 | 0 | if (dstBuffer == NULL) assert(*dstSizePtr == 0); |
1664 | 0 | MEM_INIT(&optionsNull, 0, sizeof(optionsNull)); |
1665 | 0 | if (decompressOptionsPtr==NULL) decompressOptionsPtr = &optionsNull; |
1666 | 0 | *srcSizePtr = 0; |
1667 | 0 | *dstSizePtr = 0; |
1668 | 0 | assert(dctx != NULL); |
1669 | 0 | dctx->skipChecksum |= (decompressOptionsPtr->skipChecksums != 0); /* once set, disable for the remainder of the frame */ |
1670 | | |
1671 | | /* behaves as a state machine */ |
1672 | |
|
1673 | 0 | while (doAnotherStage) { |
1674 | |
|
1675 | 0 | switch(dctx->dStage) |
1676 | 0 | { |
1677 | | |
1678 | 0 | case dstage_getFrameHeader: |
1679 | 0 | DEBUGLOG(6, "dstage_getFrameHeader"); |
1680 | 0 | if ((size_t)(srcEnd-srcPtr) >= maxFHSize) { /* enough to decode - shortcut */ |
1681 | 0 | size_t const hSize = LZ4F_decodeHeader(dctx, srcPtr, (size_t)(srcEnd-srcPtr)); /* will update dStage appropriately */ |
1682 | 0 | FORWARD_IF_ERROR(hSize); |
1683 | 0 | srcPtr += hSize; |
1684 | 0 | break; |
1685 | 0 | } |
1686 | 0 | dctx->tmpInSize = 0; |
1687 | 0 | if (srcEnd-srcPtr == 0) return minFHSize; /* 0-size input */ |
1688 | 0 | dctx->tmpInTarget = minFHSize; /* minimum size to decode header */ |
1689 | 0 | dctx->dStage = dstage_storeFrameHeader; |
1690 | | /* fall-through */ |
1691 | |
|
1692 | 0 | case dstage_storeFrameHeader: |
1693 | 0 | DEBUGLOG(6, "dstage_storeFrameHeader"); |
1694 | 0 | { size_t const sizeToCopy = MIN(dctx->tmpInTarget - dctx->tmpInSize, (size_t)(srcEnd - srcPtr)); |
1695 | 0 | memcpy(dctx->header + dctx->tmpInSize, srcPtr, sizeToCopy); |
1696 | 0 | dctx->tmpInSize += sizeToCopy; |
1697 | 0 | srcPtr += sizeToCopy; |
1698 | 0 | } |
1699 | 0 | if (dctx->tmpInSize < dctx->tmpInTarget) { |
1700 | 0 | nextSrcSizeHint = (dctx->tmpInTarget - dctx->tmpInSize) + BHSize; /* rest of header + nextBlockHeader */ |
1701 | 0 | doAnotherStage = 0; /* not enough src data, ask for some more */ |
1702 | 0 | break; |
1703 | 0 | } |
1704 | 0 | FORWARD_IF_ERROR( LZ4F_decodeHeader(dctx, dctx->header, dctx->tmpInTarget) ); /* will update dStage appropriately */ |
1705 | 0 | break; |
1706 | | |
1707 | 0 | case dstage_init: |
1708 | 0 | DEBUGLOG(6, "dstage_init"); |
1709 | 0 | if (dctx->frameInfo.contentChecksumFlag) (void)XXH32_reset(&(dctx->xxh), 0); |
1710 | | /* internal buffers allocation */ |
1711 | 0 | { size_t const bufferNeeded = dctx->maxBlockSize |
1712 | 0 | + ((dctx->frameInfo.blockMode==LZ4F_blockLinked) ? 128 KB : 0); |
1713 | 0 | if (bufferNeeded > dctx->maxBufferSize) { /* tmp buffers too small */ |
1714 | 0 | dctx->maxBufferSize = 0; /* ensure allocation will be re-attempted on next entry*/ |
1715 | 0 | LZ4F_free(dctx->tmpIn, dctx->cmem); |
1716 | 0 | dctx->tmpIn = (BYTE*)LZ4F_malloc(dctx->maxBlockSize + BFSize /* block checksum */, dctx->cmem); |
1717 | 0 | RETURN_ERROR_IF(dctx->tmpIn == NULL, allocation_failed); |
1718 | 0 | LZ4F_free(dctx->tmpOutBuffer, dctx->cmem); |
1719 | 0 | dctx->tmpOutBuffer= (BYTE*)LZ4F_malloc(bufferNeeded, dctx->cmem); |
1720 | 0 | RETURN_ERROR_IF(dctx->tmpOutBuffer== NULL, allocation_failed); |
1721 | 0 | dctx->maxBufferSize = bufferNeeded; |
1722 | 0 | } } |
1723 | 0 | dctx->tmpInSize = 0; |
1724 | 0 | dctx->tmpInTarget = 0; |
1725 | 0 | dctx->tmpOut = dctx->tmpOutBuffer; |
1726 | 0 | dctx->tmpOutStart = 0; |
1727 | 0 | dctx->tmpOutSize = 0; |
1728 | |
|
1729 | 0 | dctx->dStage = dstage_getBlockHeader; |
1730 | | /* fall-through */ |
1731 | |
|
1732 | 0 | case dstage_getBlockHeader: |
1733 | 0 | if ((size_t)(srcEnd - srcPtr) >= BHSize) { |
1734 | 0 | selectedIn = srcPtr; |
1735 | 0 | srcPtr += BHSize; |
1736 | 0 | } else { |
1737 | | /* not enough input to read cBlockSize field */ |
1738 | 0 | dctx->tmpInSize = 0; |
1739 | 0 | dctx->dStage = dstage_storeBlockHeader; |
1740 | 0 | } |
1741 | |
|
1742 | 0 | if (dctx->dStage == dstage_storeBlockHeader) /* can be skipped */ |
1743 | 0 | case dstage_storeBlockHeader: |
1744 | 0 | { size_t const remainingInput = (size_t)(srcEnd - srcPtr); |
1745 | 0 | size_t const wantedData = BHSize - dctx->tmpInSize; |
1746 | 0 | size_t const sizeToCopy = MIN(wantedData, remainingInput); |
1747 | 0 | memcpy(dctx->tmpIn + dctx->tmpInSize, srcPtr, sizeToCopy); |
1748 | 0 | srcPtr += sizeToCopy; |
1749 | 0 | dctx->tmpInSize += sizeToCopy; |
1750 | |
|
1751 | 0 | if (dctx->tmpInSize < BHSize) { /* not enough input for cBlockSize */ |
1752 | 0 | nextSrcSizeHint = BHSize - dctx->tmpInSize; |
1753 | 0 | doAnotherStage = 0; |
1754 | 0 | break; |
1755 | 0 | } |
1756 | 0 | selectedIn = dctx->tmpIn; |
1757 | 0 | } /* if (dctx->dStage == dstage_storeBlockHeader) */ |
1758 | | |
1759 | | /* decode block header */ |
1760 | 0 | { U32 const blockHeader = LZ4F_readLE32(selectedIn); |
1761 | 0 | size_t const nextCBlockSize = blockHeader & 0x7FFFFFFFU; |
1762 | 0 | size_t const crcSize = dctx->frameInfo.blockChecksumFlag * BFSize; |
1763 | 0 | if (blockHeader==0) { /* frameEnd signal, no more block */ |
1764 | 0 | DEBUGLOG(5, "end of frame"); |
1765 | 0 | dctx->dStage = dstage_getSuffix; |
1766 | 0 | break; |
1767 | 0 | } |
1768 | 0 | if (nextCBlockSize > dctx->maxBlockSize) { |
1769 | 0 | RETURN_ERROR(maxBlockSize_invalid); |
1770 | 0 | } |
1771 | 0 | if (blockHeader & LZ4F_BLOCKUNCOMPRESSED_FLAG) { |
1772 | | /* next block is uncompressed */ |
1773 | 0 | dctx->tmpInTarget = nextCBlockSize; |
1774 | 0 | DEBUGLOG(5, "next block is uncompressed (size %u)", (U32)nextCBlockSize); |
1775 | 0 | if (dctx->frameInfo.blockChecksumFlag) { |
1776 | 0 | (void)XXH32_reset(&dctx->blockChecksum, 0); |
1777 | 0 | } |
1778 | 0 | dctx->dStage = dstage_copyDirect; |
1779 | 0 | break; |
1780 | 0 | } |
1781 | | /* next block is a compressed block */ |
1782 | 0 | dctx->tmpInTarget = nextCBlockSize + crcSize; |
1783 | 0 | dctx->dStage = dstage_getCBlock; |
1784 | 0 | if (dstPtr==dstEnd || srcPtr==srcEnd) { |
1785 | 0 | nextSrcSizeHint = BHSize + nextCBlockSize + crcSize; |
1786 | 0 | doAnotherStage = 0; |
1787 | 0 | } |
1788 | 0 | break; |
1789 | 0 | } |
1790 | | |
1791 | 0 | case dstage_copyDirect: /* uncompressed block */ |
1792 | 0 | DEBUGLOG(6, "dstage_copyDirect"); |
1793 | 0 | { size_t sizeToCopy; |
1794 | 0 | if (dstPtr == NULL) { |
1795 | 0 | sizeToCopy = 0; |
1796 | 0 | } else { |
1797 | 0 | size_t const minBuffSize = MIN((size_t)(srcEnd-srcPtr), (size_t)(dstEnd-dstPtr)); |
1798 | 0 | sizeToCopy = MIN(dctx->tmpInTarget, minBuffSize); |
1799 | 0 | memcpy(dstPtr, srcPtr, sizeToCopy); |
1800 | 0 | if (!dctx->skipChecksum) { |
1801 | 0 | if (dctx->frameInfo.blockChecksumFlag) { |
1802 | 0 | (void)XXH32_update(&dctx->blockChecksum, srcPtr, sizeToCopy); |
1803 | 0 | } |
1804 | 0 | if (dctx->frameInfo.contentChecksumFlag) |
1805 | 0 | (void)XXH32_update(&dctx->xxh, srcPtr, sizeToCopy); |
1806 | 0 | } |
1807 | 0 | if (dctx->frameInfo.contentSize) |
1808 | 0 | dctx->frameRemainingSize -= sizeToCopy; |
1809 | | |
1810 | | /* history management (linked blocks only)*/ |
1811 | 0 | if (dctx->frameInfo.blockMode == LZ4F_blockLinked) { |
1812 | 0 | LZ4F_updateDict(dctx, dstPtr, sizeToCopy, dstStart, 0); |
1813 | 0 | } |
1814 | 0 | srcPtr += sizeToCopy; |
1815 | 0 | dstPtr += sizeToCopy; |
1816 | 0 | } |
1817 | 0 | if (sizeToCopy == dctx->tmpInTarget) { /* all done */ |
1818 | 0 | if (dctx->frameInfo.blockChecksumFlag) { |
1819 | 0 | dctx->tmpInSize = 0; |
1820 | 0 | dctx->dStage = dstage_getBlockChecksum; |
1821 | 0 | } else |
1822 | 0 | dctx->dStage = dstage_getBlockHeader; /* new block */ |
1823 | 0 | break; |
1824 | 0 | } |
1825 | 0 | dctx->tmpInTarget -= sizeToCopy; /* need to copy more */ |
1826 | 0 | } |
1827 | 0 | nextSrcSizeHint = dctx->tmpInTarget + |
1828 | 0 | +(dctx->frameInfo.blockChecksumFlag ? BFSize : 0) |
1829 | 0 | + BHSize /* next header size */; |
1830 | 0 | doAnotherStage = 0; |
1831 | 0 | break; |
1832 | | |
1833 | | /* check block checksum for recently transferred uncompressed block */ |
1834 | 0 | case dstage_getBlockChecksum: |
1835 | 0 | DEBUGLOG(6, "dstage_getBlockChecksum"); |
1836 | 0 | { const void* crcSrc; |
1837 | 0 | if ((srcEnd-srcPtr >= 4) && (dctx->tmpInSize==0)) { |
1838 | 0 | crcSrc = srcPtr; |
1839 | 0 | srcPtr += 4; |
1840 | 0 | } else { |
1841 | 0 | size_t const stillToCopy = 4 - dctx->tmpInSize; |
1842 | 0 | size_t const sizeToCopy = MIN(stillToCopy, (size_t)(srcEnd-srcPtr)); |
1843 | 0 | memcpy(dctx->header + dctx->tmpInSize, srcPtr, sizeToCopy); |
1844 | 0 | dctx->tmpInSize += sizeToCopy; |
1845 | 0 | srcPtr += sizeToCopy; |
1846 | 0 | if (dctx->tmpInSize < 4) { /* all input consumed */ |
1847 | 0 | doAnotherStage = 0; |
1848 | 0 | break; |
1849 | 0 | } |
1850 | 0 | crcSrc = dctx->header; |
1851 | 0 | } |
1852 | 0 | if (!dctx->skipChecksum) { |
1853 | 0 | U32 const readCRC = LZ4F_readLE32(crcSrc); |
1854 | 0 | U32 const calcCRC = XXH32_digest(&dctx->blockChecksum); |
1855 | | #ifndef FUZZING_BUILD_MODE_UNSAFE_FOR_PRODUCTION |
1856 | | DEBUGLOG(6, "compare block checksum"); |
1857 | | if (readCRC != calcCRC) { |
1858 | | DEBUGLOG(4, "incorrect block checksum: %08X != %08X", |
1859 | | readCRC, calcCRC); |
1860 | | RETURN_ERROR(blockChecksum_invalid); |
1861 | | } |
1862 | | #else |
1863 | 0 | (void)readCRC; |
1864 | 0 | (void)calcCRC; |
1865 | 0 | #endif |
1866 | 0 | } } |
1867 | 0 | dctx->dStage = dstage_getBlockHeader; /* new block */ |
1868 | 0 | break; |
1869 | | |
1870 | 0 | case dstage_getCBlock: |
1871 | 0 | DEBUGLOG(6, "dstage_getCBlock"); |
1872 | 0 | if ((size_t)(srcEnd-srcPtr) < dctx->tmpInTarget) { |
1873 | 0 | dctx->tmpInSize = 0; |
1874 | 0 | dctx->dStage = dstage_storeCBlock; |
1875 | 0 | break; |
1876 | 0 | } |
1877 | | /* input large enough to read full block directly */ |
1878 | 0 | selectedIn = srcPtr; |
1879 | 0 | srcPtr += dctx->tmpInTarget; |
1880 | |
|
1881 | 0 | if (0) /* always jump over next block */ |
1882 | 0 | case dstage_storeCBlock: |
1883 | 0 | { size_t const wantedData = dctx->tmpInTarget - dctx->tmpInSize; |
1884 | 0 | size_t const inputLeft = (size_t)(srcEnd-srcPtr); |
1885 | 0 | size_t const sizeToCopy = MIN(wantedData, inputLeft); |
1886 | 0 | memcpy(dctx->tmpIn + dctx->tmpInSize, srcPtr, sizeToCopy); |
1887 | 0 | dctx->tmpInSize += sizeToCopy; |
1888 | 0 | srcPtr += sizeToCopy; |
1889 | 0 | if (dctx->tmpInSize < dctx->tmpInTarget) { /* need more input */ |
1890 | 0 | nextSrcSizeHint = (dctx->tmpInTarget - dctx->tmpInSize) |
1891 | 0 | + (dctx->frameInfo.blockChecksumFlag ? BFSize : 0) |
1892 | 0 | + BHSize /* next header size */; |
1893 | 0 | doAnotherStage = 0; |
1894 | 0 | break; |
1895 | 0 | } |
1896 | 0 | selectedIn = dctx->tmpIn; |
1897 | 0 | } |
1898 | | |
1899 | | /* At this stage, input is large enough to decode a block */ |
1900 | | |
1901 | | /* First, decode and control block checksum if it exists */ |
1902 | 0 | if (dctx->frameInfo.blockChecksumFlag) { |
1903 | 0 | assert(dctx->tmpInTarget >= 4); |
1904 | 0 | dctx->tmpInTarget -= 4; |
1905 | 0 | assert(selectedIn != NULL); /* selectedIn is defined at this stage (either srcPtr, or dctx->tmpIn) */ |
1906 | 0 | { U32 const readBlockCrc = LZ4F_readLE32(selectedIn + dctx->tmpInTarget); |
1907 | 0 | U32 const calcBlockCrc = XXH32(selectedIn, dctx->tmpInTarget, 0); |
1908 | | #ifndef FUZZING_BUILD_MODE_UNSAFE_FOR_PRODUCTION |
1909 | | RETURN_ERROR_IF(readBlockCrc != calcBlockCrc, blockChecksum_invalid); |
1910 | | #else |
1911 | 0 | (void)readBlockCrc; |
1912 | 0 | (void)calcBlockCrc; |
1913 | 0 | #endif |
1914 | 0 | } } |
1915 | | |
1916 | | /* decode directly into destination buffer if there is enough room */ |
1917 | 0 | if ( ((size_t)(dstEnd-dstPtr) >= dctx->maxBlockSize) |
1918 | | /* unless the dictionary is stored in tmpOut: |
1919 | | * in which case it's faster to decode within tmpOut |
1920 | | * to benefit from prefix speedup */ |
1921 | 0 | && !(dctx->dict!= NULL && (const BYTE*)dctx->dict + dctx->dictSize == dctx->tmpOut) ) |
1922 | 0 | { |
1923 | 0 | const char* dict = (const char*)dctx->dict; |
1924 | 0 | size_t dictSize = dctx->dictSize; |
1925 | 0 | int decodedSize; |
1926 | 0 | assert(dstPtr != NULL); |
1927 | 0 | if (dict && dictSize > 1 GB) { |
1928 | | /* overflow control : dctx->dictSize is an int, avoid truncation / sign issues */ |
1929 | 0 | dict += dictSize - 64 KB; |
1930 | 0 | dictSize = 64 KB; |
1931 | 0 | } |
1932 | 0 | decodedSize = LZ4_decompress_safe_usingDict( |
1933 | 0 | (const char*)selectedIn, (char*)dstPtr, |
1934 | 0 | (int)dctx->tmpInTarget, (int)dctx->maxBlockSize, |
1935 | 0 | dict, (int)dictSize); |
1936 | 0 | RETURN_ERROR_IF(decodedSize < 0, decompressionFailed); |
1937 | 0 | if ((dctx->frameInfo.contentChecksumFlag) && (!dctx->skipChecksum)) |
1938 | 0 | XXH32_update(&(dctx->xxh), dstPtr, (size_t)decodedSize); |
1939 | 0 | if (dctx->frameInfo.contentSize) |
1940 | 0 | dctx->frameRemainingSize -= (size_t)decodedSize; |
1941 | | |
1942 | | /* dictionary management */ |
1943 | 0 | if (dctx->frameInfo.blockMode==LZ4F_blockLinked) { |
1944 | 0 | LZ4F_updateDict(dctx, dstPtr, (size_t)decodedSize, dstStart, 0); |
1945 | 0 | } |
1946 | |
|
1947 | 0 | dstPtr += decodedSize; |
1948 | 0 | dctx->dStage = dstage_getBlockHeader; /* end of block, let's get another one */ |
1949 | 0 | break; |
1950 | 0 | } |
1951 | | |
1952 | | /* not enough place into dst : decode into tmpOut */ |
1953 | | |
1954 | | /* manage dictionary */ |
1955 | 0 | if (dctx->frameInfo.blockMode == LZ4F_blockLinked) { |
1956 | 0 | if (dctx->dict == dctx->tmpOutBuffer) { |
1957 | | /* truncate dictionary to 64 KB if too big */ |
1958 | 0 | if (dctx->dictSize > 128 KB) { |
1959 | 0 | memcpy(dctx->tmpOutBuffer, dctx->dict + dctx->dictSize - 64 KB, 64 KB); |
1960 | 0 | dctx->dictSize = 64 KB; |
1961 | 0 | } |
1962 | 0 | dctx->tmpOut = dctx->tmpOutBuffer + dctx->dictSize; |
1963 | 0 | } else { /* dict not within tmpOut */ |
1964 | 0 | size_t const reservedDictSpace = MIN(dctx->dictSize, 64 KB); |
1965 | 0 | dctx->tmpOut = dctx->tmpOutBuffer + reservedDictSpace; |
1966 | 0 | } } |
1967 | | |
1968 | | /* Decode block into tmpOut */ |
1969 | 0 | { const char* dict = (const char*)dctx->dict; |
1970 | 0 | size_t dictSize = dctx->dictSize; |
1971 | 0 | int decodedSize; |
1972 | 0 | if (dict && dictSize > 1 GB) { |
1973 | | /* the dictSize param is an int, avoid truncation / sign issues */ |
1974 | 0 | dict += dictSize - 64 KB; |
1975 | 0 | dictSize = 64 KB; |
1976 | 0 | } |
1977 | 0 | decodedSize = LZ4_decompress_safe_usingDict( |
1978 | 0 | (const char*)selectedIn, (char*)dctx->tmpOut, |
1979 | 0 | (int)dctx->tmpInTarget, (int)dctx->maxBlockSize, |
1980 | 0 | dict, (int)dictSize); |
1981 | 0 | RETURN_ERROR_IF(decodedSize < 0, decompressionFailed); |
1982 | 0 | if (dctx->frameInfo.contentChecksumFlag && !dctx->skipChecksum) |
1983 | 0 | XXH32_update(&(dctx->xxh), dctx->tmpOut, (size_t)decodedSize); |
1984 | 0 | if (dctx->frameInfo.contentSize) |
1985 | 0 | dctx->frameRemainingSize -= (size_t)decodedSize; |
1986 | 0 | dctx->tmpOutSize = (size_t)decodedSize; |
1987 | 0 | dctx->tmpOutStart = 0; |
1988 | 0 | dctx->dStage = dstage_flushOut; |
1989 | 0 | } |
1990 | | /* fall-through */ |
1991 | | |
1992 | 0 | case dstage_flushOut: /* flush decoded data from tmpOut to dstBuffer */ |
1993 | 0 | DEBUGLOG(6, "dstage_flushOut"); |
1994 | 0 | if (dstPtr != NULL) { |
1995 | 0 | size_t const sizeToCopy = MIN(dctx->tmpOutSize - dctx->tmpOutStart, (size_t)(dstEnd-dstPtr)); |
1996 | 0 | memcpy(dstPtr, dctx->tmpOut + dctx->tmpOutStart, sizeToCopy); |
1997 | | |
1998 | | /* dictionary management */ |
1999 | 0 | if (dctx->frameInfo.blockMode == LZ4F_blockLinked) |
2000 | 0 | LZ4F_updateDict(dctx, dstPtr, sizeToCopy, dstStart, 1 /*withinTmp*/); |
2001 | |
|
2002 | 0 | dctx->tmpOutStart += sizeToCopy; |
2003 | 0 | dstPtr += sizeToCopy; |
2004 | 0 | } |
2005 | 0 | if (dctx->tmpOutStart == dctx->tmpOutSize) { /* all flushed */ |
2006 | 0 | dctx->dStage = dstage_getBlockHeader; /* get next block */ |
2007 | 0 | break; |
2008 | 0 | } |
2009 | | /* could not flush everything : stop there, just request a block header */ |
2010 | 0 | doAnotherStage = 0; |
2011 | 0 | nextSrcSizeHint = BHSize; |
2012 | 0 | break; |
2013 | | |
2014 | 0 | case dstage_getSuffix: |
2015 | 0 | RETURN_ERROR_IF(dctx->frameRemainingSize, frameSize_wrong); /* incorrect frame size decoded */ |
2016 | 0 | if (!dctx->frameInfo.contentChecksumFlag) { /* no checksum, frame is completed */ |
2017 | 0 | nextSrcSizeHint = 0; |
2018 | 0 | LZ4F_resetDecompressionContext(dctx); |
2019 | 0 | doAnotherStage = 0; |
2020 | 0 | break; |
2021 | 0 | } |
2022 | 0 | if ((srcEnd - srcPtr) < 4) { /* not enough size for entire CRC */ |
2023 | 0 | dctx->tmpInSize = 0; |
2024 | 0 | dctx->dStage = dstage_storeSuffix; |
2025 | 0 | } else { |
2026 | 0 | selectedIn = srcPtr; |
2027 | 0 | srcPtr += 4; |
2028 | 0 | } |
2029 | |
|
2030 | 0 | if (dctx->dStage == dstage_storeSuffix) /* can be skipped */ |
2031 | 0 | case dstage_storeSuffix: |
2032 | 0 | { size_t const remainingInput = (size_t)(srcEnd - srcPtr); |
2033 | 0 | size_t const wantedData = 4 - dctx->tmpInSize; |
2034 | 0 | size_t const sizeToCopy = MIN(wantedData, remainingInput); |
2035 | 0 | memcpy(dctx->tmpIn + dctx->tmpInSize, srcPtr, sizeToCopy); |
2036 | 0 | srcPtr += sizeToCopy; |
2037 | 0 | dctx->tmpInSize += sizeToCopy; |
2038 | 0 | if (dctx->tmpInSize < 4) { /* not enough input to read complete suffix */ |
2039 | 0 | nextSrcSizeHint = 4 - dctx->tmpInSize; |
2040 | 0 | doAnotherStage=0; |
2041 | 0 | break; |
2042 | 0 | } |
2043 | 0 | selectedIn = dctx->tmpIn; |
2044 | 0 | } /* if (dctx->dStage == dstage_storeSuffix) */ |
2045 | | |
2046 | | /* case dstage_checkSuffix: */ /* no direct entry, avoid initialization risks */ |
2047 | 0 | if (!dctx->skipChecksum) { |
2048 | 0 | U32 const readCRC = LZ4F_readLE32(selectedIn); |
2049 | 0 | U32 const resultCRC = XXH32_digest(&(dctx->xxh)); |
2050 | 0 | DEBUGLOG(4, "frame checksum: stored 0x%0X vs 0x%0X processed", readCRC, resultCRC); |
2051 | | #ifndef FUZZING_BUILD_MODE_UNSAFE_FOR_PRODUCTION |
2052 | | RETURN_ERROR_IF(readCRC != resultCRC, contentChecksum_invalid); |
2053 | | #else |
2054 | 0 | (void)readCRC; |
2055 | 0 | (void)resultCRC; |
2056 | 0 | #endif |
2057 | 0 | } |
2058 | 0 | nextSrcSizeHint = 0; |
2059 | 0 | LZ4F_resetDecompressionContext(dctx); |
2060 | 0 | doAnotherStage = 0; |
2061 | 0 | break; |
2062 | | |
2063 | 0 | case dstage_getSFrameSize: |
2064 | 0 | if ((srcEnd - srcPtr) >= 4) { |
2065 | 0 | selectedIn = srcPtr; |
2066 | 0 | srcPtr += 4; |
2067 | 0 | } else { |
2068 | | /* not enough input to read cBlockSize field */ |
2069 | 0 | dctx->tmpInSize = 4; |
2070 | 0 | dctx->tmpInTarget = 8; |
2071 | 0 | dctx->dStage = dstage_storeSFrameSize; |
2072 | 0 | } |
2073 | |
|
2074 | 0 | if (dctx->dStage == dstage_storeSFrameSize) |
2075 | 0 | case dstage_storeSFrameSize: |
2076 | 0 | { size_t const sizeToCopy = MIN(dctx->tmpInTarget - dctx->tmpInSize, |
2077 | 0 | (size_t)(srcEnd - srcPtr) ); |
2078 | 0 | memcpy(dctx->header + dctx->tmpInSize, srcPtr, sizeToCopy); |
2079 | 0 | srcPtr += sizeToCopy; |
2080 | 0 | dctx->tmpInSize += sizeToCopy; |
2081 | 0 | if (dctx->tmpInSize < dctx->tmpInTarget) { |
2082 | | /* not enough input to get full sBlockSize; wait for more */ |
2083 | 0 | nextSrcSizeHint = dctx->tmpInTarget - dctx->tmpInSize; |
2084 | 0 | doAnotherStage = 0; |
2085 | 0 | break; |
2086 | 0 | } |
2087 | 0 | selectedIn = dctx->header + 4; |
2088 | 0 | } /* if (dctx->dStage == dstage_storeSFrameSize) */ |
2089 | | |
2090 | | /* case dstage_decodeSFrameSize: */ /* no direct entry */ |
2091 | 0 | { size_t const SFrameSize = LZ4F_readLE32(selectedIn); |
2092 | 0 | dctx->frameInfo.contentSize = SFrameSize; |
2093 | 0 | dctx->tmpInTarget = SFrameSize; |
2094 | 0 | dctx->dStage = dstage_skipSkippable; |
2095 | 0 | break; |
2096 | 0 | } |
2097 | | |
2098 | 0 | case dstage_skipSkippable: |
2099 | 0 | { size_t const skipSize = MIN(dctx->tmpInTarget, (size_t)(srcEnd-srcPtr)); |
2100 | 0 | srcPtr += skipSize; |
2101 | 0 | dctx->tmpInTarget -= skipSize; |
2102 | 0 | doAnotherStage = 0; |
2103 | 0 | nextSrcSizeHint = dctx->tmpInTarget; |
2104 | 0 | if (nextSrcSizeHint) break; /* still more to skip */ |
2105 | | /* frame fully skipped : prepare context for a new frame */ |
2106 | 0 | LZ4F_resetDecompressionContext(dctx); |
2107 | 0 | break; |
2108 | 0 | } |
2109 | 0 | } /* switch (dctx->dStage) */ |
2110 | 0 | } /* while (doAnotherStage) */ |
2111 | | |
2112 | | /* preserve history within tmpOut whenever necessary */ |
2113 | 0 | LZ4F_STATIC_ASSERT((unsigned)dstage_init == 2); |
2114 | 0 | if ( (dctx->frameInfo.blockMode==LZ4F_blockLinked) /* next block will use up to 64KB from previous ones */ |
2115 | 0 | && (dctx->dict != dctx->tmpOutBuffer) /* dictionary is not already within tmp */ |
2116 | 0 | && (dctx->dict != NULL) /* dictionary exists */ |
2117 | 0 | && (!decompressOptionsPtr->stableDst) /* cannot rely on dst data to remain there for next call */ |
2118 | 0 | && ((unsigned)(dctx->dStage)-2 < (unsigned)(dstage_getSuffix)-2) ) /* valid stages : [init ... getSuffix[ */ |
2119 | 0 | { |
2120 | 0 | if (dctx->dStage == dstage_flushOut) { |
2121 | 0 | size_t const preserveSize = (size_t)(dctx->tmpOut - dctx->tmpOutBuffer); |
2122 | 0 | size_t copySize = 64 KB - dctx->tmpOutSize; |
2123 | 0 | const BYTE* oldDictEnd = dctx->dict + dctx->dictSize - dctx->tmpOutStart; |
2124 | 0 | if (dctx->tmpOutSize > 64 KB) copySize = 0; |
2125 | 0 | if (copySize > preserveSize) copySize = preserveSize; |
2126 | 0 | assert(dctx->tmpOutBuffer != NULL); |
2127 | |
|
2128 | 0 | memcpy(dctx->tmpOutBuffer + preserveSize - copySize, oldDictEnd - copySize, copySize); |
2129 | |
|
2130 | 0 | dctx->dict = dctx->tmpOutBuffer; |
2131 | 0 | dctx->dictSize = preserveSize + dctx->tmpOutStart; |
2132 | 0 | } else { |
2133 | 0 | const BYTE* const oldDictEnd = dctx->dict + dctx->dictSize; |
2134 | 0 | size_t const newDictSize = MIN(dctx->dictSize, 64 KB); |
2135 | |
|
2136 | 0 | memcpy(dctx->tmpOutBuffer, oldDictEnd - newDictSize, newDictSize); |
2137 | |
|
2138 | 0 | dctx->dict = dctx->tmpOutBuffer; |
2139 | 0 | dctx->dictSize = newDictSize; |
2140 | 0 | dctx->tmpOut = dctx->tmpOutBuffer + newDictSize; |
2141 | 0 | } |
2142 | 0 | } |
2143 | |
|
2144 | 0 | *srcSizePtr = (size_t)(srcPtr - srcStart); |
2145 | 0 | *dstSizePtr = (size_t)(dstPtr - dstStart); |
2146 | 0 | return nextSrcSizeHint; |
2147 | 0 | } |
2148 | | |
2149 | | /*! LZ4F_decompress_usingDict() : |
2150 | | * Same as LZ4F_decompress(), using a predefined dictionary. |
2151 | | * Dictionary is used "in place", without any preprocessing. |
2152 | | * It must remain accessible throughout the entire frame decoding. |
2153 | | */ |
2154 | | size_t LZ4F_decompress_usingDict(LZ4F_dctx* dctx, |
2155 | | void* dstBuffer, size_t* dstSizePtr, |
2156 | | const void* srcBuffer, size_t* srcSizePtr, |
2157 | | const void* dict, size_t dictSize, |
2158 | | const LZ4F_decompressOptions_t* decompressOptionsPtr) |
2159 | 0 | { |
2160 | 0 | if (dctx->dStage <= dstage_init) { |
2161 | 0 | dctx->dict = (const BYTE*)dict; |
2162 | 0 | dctx->dictSize = dictSize; |
2163 | 0 | } |
2164 | 0 | return LZ4F_decompress(dctx, dstBuffer, dstSizePtr, |
2165 | 0 | srcBuffer, srcSizePtr, |
2166 | 0 | decompressOptionsPtr); |
2167 | 0 | } |