git subrepo pull (merge) --force deps/libchdr
[pcsx_rearmed.git] / deps / libchdr / deps / zstd-1.5.5 / lib / common / zstd_internal.h
CommitLineData
648db22b 1/*
2 * Copyright (c) Meta Platforms, Inc. and affiliates.
3 * All rights reserved.
4 *
5 * This source code is licensed under both the BSD-style license (found in the
6 * LICENSE file in the root directory of this source tree) and the GPLv2 (found
7 * in the COPYING file in the root directory of this source tree).
8 * You may select, at your option, one of the above-listed licenses.
9 */
10
11#ifndef ZSTD_CCOMMON_H_MODULE
12#define ZSTD_CCOMMON_H_MODULE
13
14/* this module contains definitions which must be identical
15 * across compression, decompression and dictBuilder.
16 * It also contains a few functions useful to at least 2 of them
17 * and which benefit from being inlined */
18
19/*-*************************************
20* Dependencies
21***************************************/
22#include "compiler.h"
23#include "cpu.h"
24#include "mem.h"
25#include "debug.h" /* assert, DEBUGLOG, RAWLOG, g_debuglevel */
26#include "error_private.h"
27#define ZSTD_STATIC_LINKING_ONLY
28#include "../zstd.h"
29#define FSE_STATIC_LINKING_ONLY
30#include "fse.h"
31#include "huf.h"
32#ifndef XXH_STATIC_LINKING_ONLY
33# define XXH_STATIC_LINKING_ONLY /* XXH64_state_t */
34#endif
35#include "xxhash.h" /* XXH_reset, update, digest */
36#ifndef ZSTD_NO_TRACE
37# include "zstd_trace.h"
38#else
39# define ZSTD_TRACE 0
40#endif
41
42#if defined (__cplusplus)
43extern "C" {
44#endif
45
46/* ---- static assert (debug) --- */
47#define ZSTD_STATIC_ASSERT(c) DEBUG_STATIC_ASSERT(c)
48#define ZSTD_isError ERR_isError /* for inlining */
49#define FSE_isError ERR_isError
50#define HUF_isError ERR_isError
51
52
53/*-*************************************
54* shared macros
55***************************************/
56#undef MIN
57#undef MAX
58#define MIN(a,b) ((a)<(b) ? (a) : (b))
59#define MAX(a,b) ((a)>(b) ? (a) : (b))
60#define BOUNDED(min,val,max) (MAX(min,MIN(val,max)))
61
62
63/*-*************************************
64* Common constants
65***************************************/
66#define ZSTD_OPT_NUM (1<<12)
67
68#define ZSTD_REP_NUM 3 /* number of repcodes */
69static UNUSED_ATTR const U32 repStartValue[ZSTD_REP_NUM] = { 1, 4, 8 };
70
71#define KB *(1 <<10)
72#define MB *(1 <<20)
73#define GB *(1U<<30)
74
75#define BIT7 128
76#define BIT6 64
77#define BIT5 32
78#define BIT4 16
79#define BIT1 2
80#define BIT0 1
81
82#define ZSTD_WINDOWLOG_ABSOLUTEMIN 10
83static UNUSED_ATTR const size_t ZSTD_fcs_fieldSize[4] = { 0, 2, 4, 8 };
84static UNUSED_ATTR const size_t ZSTD_did_fieldSize[4] = { 0, 1, 2, 4 };
85
86#define ZSTD_FRAMEIDSIZE 4 /* magic number size */
87
88#define ZSTD_BLOCKHEADERSIZE 3 /* C standard doesn't allow `static const` variable to be init using another `static const` variable */
89static UNUSED_ATTR const size_t ZSTD_blockHeaderSize = ZSTD_BLOCKHEADERSIZE;
90typedef enum { bt_raw, bt_rle, bt_compressed, bt_reserved } blockType_e;
91
92#define ZSTD_FRAMECHECKSUMSIZE 4
93
94#define MIN_SEQUENCES_SIZE 1 /* nbSeq==0 */
95#define MIN_CBLOCK_SIZE (1 /*litCSize*/ + 1 /* RLE or RAW */) /* for a non-null block */
96#define MIN_LITERALS_FOR_4_STREAMS 6
97
98typedef enum { set_basic, set_rle, set_compressed, set_repeat } symbolEncodingType_e;
99
100#define LONGNBSEQ 0x7F00
101
102#define MINMATCH 3
103
104#define Litbits 8
105#define LitHufLog 11
106#define MaxLit ((1<<Litbits) - 1)
107#define MaxML 52
108#define MaxLL 35
109#define DefaultMaxOff 28
110#define MaxOff 31
111#define MaxSeq MAX(MaxLL, MaxML) /* Assumption : MaxOff < MaxLL,MaxML */
112#define MLFSELog 9
113#define LLFSELog 9
114#define OffFSELog 8
115#define MaxFSELog MAX(MAX(MLFSELog, LLFSELog), OffFSELog)
116#define MaxMLBits 16
117#define MaxLLBits 16
118
119#define ZSTD_MAX_HUF_HEADER_SIZE 128 /* header + <= 127 byte tree description */
120/* Each table cannot take more than #symbols * FSELog bits */
121#define ZSTD_MAX_FSE_HEADERS_SIZE (((MaxML + 1) * MLFSELog + (MaxLL + 1) * LLFSELog + (MaxOff + 1) * OffFSELog + 7) / 8)
122
123static UNUSED_ATTR const U8 LL_bits[MaxLL+1] = {
124 0, 0, 0, 0, 0, 0, 0, 0,
125 0, 0, 0, 0, 0, 0, 0, 0,
126 1, 1, 1, 1, 2, 2, 3, 3,
127 4, 6, 7, 8, 9,10,11,12,
128 13,14,15,16
129};
130static UNUSED_ATTR const S16 LL_defaultNorm[MaxLL+1] = {
131 4, 3, 2, 2, 2, 2, 2, 2,
132 2, 2, 2, 2, 2, 1, 1, 1,
133 2, 2, 2, 2, 2, 2, 2, 2,
134 2, 3, 2, 1, 1, 1, 1, 1,
135 -1,-1,-1,-1
136};
137#define LL_DEFAULTNORMLOG 6 /* for static allocation */
138static UNUSED_ATTR const U32 LL_defaultNormLog = LL_DEFAULTNORMLOG;
139
140static UNUSED_ATTR const U8 ML_bits[MaxML+1] = {
141 0, 0, 0, 0, 0, 0, 0, 0,
142 0, 0, 0, 0, 0, 0, 0, 0,
143 0, 0, 0, 0, 0, 0, 0, 0,
144 0, 0, 0, 0, 0, 0, 0, 0,
145 1, 1, 1, 1, 2, 2, 3, 3,
146 4, 4, 5, 7, 8, 9,10,11,
147 12,13,14,15,16
148};
149static UNUSED_ATTR const S16 ML_defaultNorm[MaxML+1] = {
150 1, 4, 3, 2, 2, 2, 2, 2,
151 2, 1, 1, 1, 1, 1, 1, 1,
152 1, 1, 1, 1, 1, 1, 1, 1,
153 1, 1, 1, 1, 1, 1, 1, 1,
154 1, 1, 1, 1, 1, 1, 1, 1,
155 1, 1, 1, 1, 1, 1,-1,-1,
156 -1,-1,-1,-1,-1
157};
158#define ML_DEFAULTNORMLOG 6 /* for static allocation */
159static UNUSED_ATTR const U32 ML_defaultNormLog = ML_DEFAULTNORMLOG;
160
161static UNUSED_ATTR const S16 OF_defaultNorm[DefaultMaxOff+1] = {
162 1, 1, 1, 1, 1, 1, 2, 2,
163 2, 1, 1, 1, 1, 1, 1, 1,
164 1, 1, 1, 1, 1, 1, 1, 1,
165 -1,-1,-1,-1,-1
166};
167#define OF_DEFAULTNORMLOG 5 /* for static allocation */
168static UNUSED_ATTR const U32 OF_defaultNormLog = OF_DEFAULTNORMLOG;
169
170
171/*-*******************************************
172* Shared functions to include for inlining
173*********************************************/
174static void ZSTD_copy8(void* dst, const void* src) {
175#if defined(ZSTD_ARCH_ARM_NEON)
176 vst1_u8((uint8_t*)dst, vld1_u8((const uint8_t*)src));
177#else
178 ZSTD_memcpy(dst, src, 8);
179#endif
180}
181#define COPY8(d,s) { ZSTD_copy8(d,s); d+=8; s+=8; }
182
183/* Need to use memmove here since the literal buffer can now be located within
184 the dst buffer. In circumstances where the op "catches up" to where the
185 literal buffer is, there can be partial overlaps in this call on the final
186 copy if the literal is being shifted by less than 16 bytes. */
187static void ZSTD_copy16(void* dst, const void* src) {
188#if defined(ZSTD_ARCH_ARM_NEON)
189 vst1q_u8((uint8_t*)dst, vld1q_u8((const uint8_t*)src));
190#elif defined(ZSTD_ARCH_X86_SSE2)
191 _mm_storeu_si128((__m128i*)dst, _mm_loadu_si128((const __m128i*)src));
192#elif defined(__clang__)
193 ZSTD_memmove(dst, src, 16);
194#else
195 /* ZSTD_memmove is not inlined properly by gcc */
196 BYTE copy16_buf[16];
197 ZSTD_memcpy(copy16_buf, src, 16);
198 ZSTD_memcpy(dst, copy16_buf, 16);
199#endif
200}
201#define COPY16(d,s) { ZSTD_copy16(d,s); d+=16; s+=16; }
202
203#define WILDCOPY_OVERLENGTH 32
204#define WILDCOPY_VECLEN 16
205
206typedef enum {
207 ZSTD_no_overlap,
208 ZSTD_overlap_src_before_dst
209 /* ZSTD_overlap_dst_before_src, */
210} ZSTD_overlap_e;
211
212/*! ZSTD_wildcopy() :
213 * Custom version of ZSTD_memcpy(), can over read/write up to WILDCOPY_OVERLENGTH bytes (if length==0)
214 * @param ovtype controls the overlap detection
215 * - ZSTD_no_overlap: The source and destination are guaranteed to be at least WILDCOPY_VECLEN bytes apart.
216 * - ZSTD_overlap_src_before_dst: The src and dst may overlap, but they MUST be at least 8 bytes apart.
217 * The src buffer must be before the dst buffer.
218 */
219MEM_STATIC FORCE_INLINE_ATTR
220void ZSTD_wildcopy(void* dst, const void* src, ptrdiff_t length, ZSTD_overlap_e const ovtype)
221{
222 ptrdiff_t diff = (BYTE*)dst - (const BYTE*)src;
223 const BYTE* ip = (const BYTE*)src;
224 BYTE* op = (BYTE*)dst;
225 BYTE* const oend = op + length;
226
227 if (ovtype == ZSTD_overlap_src_before_dst && diff < WILDCOPY_VECLEN) {
228 /* Handle short offset copies. */
229 do {
230 COPY8(op, ip)
231 } while (op < oend);
232 } else {
233 assert(diff >= WILDCOPY_VECLEN || diff <= -WILDCOPY_VECLEN);
234 /* Separate out the first COPY16() call because the copy length is
235 * almost certain to be short, so the branches have different
236 * probabilities. Since it is almost certain to be short, only do
237 * one COPY16() in the first call. Then, do two calls per loop since
238 * at that point it is more likely to have a high trip count.
239 */
240 ZSTD_copy16(op, ip);
241 if (16 >= length) return;
242 op += 16;
243 ip += 16;
244 do {
245 COPY16(op, ip);
246 COPY16(op, ip);
247 }
248 while (op < oend);
249 }
250}
251
252MEM_STATIC size_t ZSTD_limitCopy(void* dst, size_t dstCapacity, const void* src, size_t srcSize)
253{
254 size_t const length = MIN(dstCapacity, srcSize);
255 if (length > 0) {
256 ZSTD_memcpy(dst, src, length);
257 }
258 return length;
259}
260
261/* define "workspace is too large" as this number of times larger than needed */
262#define ZSTD_WORKSPACETOOLARGE_FACTOR 3
263
264/* when workspace is continuously too large
265 * during at least this number of times,
266 * context's memory usage is considered wasteful,
267 * because it's sized to handle a worst case scenario which rarely happens.
268 * In which case, resize it down to free some memory */
269#define ZSTD_WORKSPACETOOLARGE_MAXDURATION 128
270
271/* Controls whether the input/output buffer is buffered or stable. */
272typedef enum {
273 ZSTD_bm_buffered = 0, /* Buffer the input/output */
274 ZSTD_bm_stable = 1 /* ZSTD_inBuffer/ZSTD_outBuffer is stable */
275} ZSTD_bufferMode_e;
276
277
278/*-*******************************************
279* Private declarations
280*********************************************/
281typedef struct seqDef_s {
282 U32 offBase; /* offBase == Offset + ZSTD_REP_NUM, or repcode 1,2,3 */
283 U16 litLength;
284 U16 mlBase; /* mlBase == matchLength - MINMATCH */
285} seqDef;
286
287/* Controls whether seqStore has a single "long" litLength or matchLength. See seqStore_t. */
288typedef enum {
289 ZSTD_llt_none = 0, /* no longLengthType */
290 ZSTD_llt_literalLength = 1, /* represents a long literal */
291 ZSTD_llt_matchLength = 2 /* represents a long match */
292} ZSTD_longLengthType_e;
293
294typedef struct {
295 seqDef* sequencesStart;
296 seqDef* sequences; /* ptr to end of sequences */
297 BYTE* litStart;
298 BYTE* lit; /* ptr to end of literals */
299 BYTE* llCode;
300 BYTE* mlCode;
301 BYTE* ofCode;
302 size_t maxNbSeq;
303 size_t maxNbLit;
304
305 /* longLengthPos and longLengthType to allow us to represent either a single litLength or matchLength
306 * in the seqStore that has a value larger than U16 (if it exists). To do so, we increment
307 * the existing value of the litLength or matchLength by 0x10000.
308 */
309 ZSTD_longLengthType_e longLengthType;
310 U32 longLengthPos; /* Index of the sequence to apply long length modification to */
311} seqStore_t;
312
313typedef struct {
314 U32 litLength;
315 U32 matchLength;
316} ZSTD_sequenceLength;
317
318/**
319 * Returns the ZSTD_sequenceLength for the given sequences. It handles the decoding of long sequences
320 * indicated by longLengthPos and longLengthType, and adds MINMATCH back to matchLength.
321 */
322MEM_STATIC ZSTD_sequenceLength ZSTD_getSequenceLength(seqStore_t const* seqStore, seqDef const* seq)
323{
324 ZSTD_sequenceLength seqLen;
325 seqLen.litLength = seq->litLength;
326 seqLen.matchLength = seq->mlBase + MINMATCH;
327 if (seqStore->longLengthPos == (U32)(seq - seqStore->sequencesStart)) {
328 if (seqStore->longLengthType == ZSTD_llt_literalLength) {
329 seqLen.litLength += 0x10000;
330 }
331 if (seqStore->longLengthType == ZSTD_llt_matchLength) {
332 seqLen.matchLength += 0x10000;
333 }
334 }
335 return seqLen;
336}
337
338/**
339 * Contains the compressed frame size and an upper-bound for the decompressed frame size.
340 * Note: before using `compressedSize`, check for errors using ZSTD_isError().
341 * similarly, before using `decompressedBound`, check for errors using:
342 * `decompressedBound != ZSTD_CONTENTSIZE_ERROR`
343 */
344typedef struct {
345 size_t nbBlocks;
346 size_t compressedSize;
347 unsigned long long decompressedBound;
348} ZSTD_frameSizeInfo; /* decompress & legacy */
349
350const seqStore_t* ZSTD_getSeqStore(const ZSTD_CCtx* ctx); /* compress & dictBuilder */
351int ZSTD_seqToCodes(const seqStore_t* seqStorePtr); /* compress, dictBuilder, decodeCorpus (shouldn't get its definition from here) */
352
353
354/* ZSTD_invalidateRepCodes() :
355 * ensures next compression will not use repcodes from previous block.
356 * Note : only works with regular variant;
357 * do not use with extDict variant ! */
358void ZSTD_invalidateRepCodes(ZSTD_CCtx* cctx); /* zstdmt, adaptive_compression (shouldn't get this definition from here) */
359
360
361typedef struct {
362 blockType_e blockType;
363 U32 lastBlock;
364 U32 origSize;
365} blockProperties_t; /* declared here for decompress and fullbench */
366
367/*! ZSTD_getcBlockSize() :
368 * Provides the size of compressed block from block header `src` */
369/* Used by: decompress, fullbench (does not get its definition from here) */
370size_t ZSTD_getcBlockSize(const void* src, size_t srcSize,
371 blockProperties_t* bpPtr);
372
373/*! ZSTD_decodeSeqHeaders() :
374 * decode sequence header from src */
375/* Used by: decompress, fullbench (does not get its definition from here) */
376size_t ZSTD_decodeSeqHeaders(ZSTD_DCtx* dctx, int* nbSeqPtr,
377 const void* src, size_t srcSize);
378
379/**
380 * @returns true iff the CPU supports dynamic BMI2 dispatch.
381 */
382MEM_STATIC int ZSTD_cpuSupportsBmi2(void)
383{
384 ZSTD_cpuid_t cpuid = ZSTD_cpuid();
385 return ZSTD_cpuid_bmi1(cpuid) && ZSTD_cpuid_bmi2(cpuid);
386}
387
388#if defined (__cplusplus)
389}
390#endif
391
392#endif /* ZSTD_CCOMMON_H_MODULE */