| 1 | /* |
| 2 | * Copyright (c) Meta Platforms, Inc. and affiliates. |
| 3 | * All rights reserved. |
| 4 | * |
| 5 | * This source code is licensed under both the BSD-style license (found in the |
| 6 | * LICENSE file in the root directory of this source tree) and the GPLv2 (found |
| 7 | * in the COPYING file in the root directory of this source tree). |
| 8 | * You may select, at your option, one of the above-listed licenses. |
| 9 | */ |
| 10 | |
| 11 | #ifndef ZSTD_COMPILER_H |
| 12 | #define ZSTD_COMPILER_H |
| 13 | |
| 14 | #include <stddef.h> |
| 15 | |
| 16 | #include "portability_macros.h" |
| 17 | |
| 18 | /*-******************************************************* |
| 19 | * Compiler specifics |
| 20 | *********************************************************/ |
| 21 | /* force inlining */ |
| 22 | |
| 23 | #if !defined(ZSTD_NO_INLINE) |
| 24 | #if (defined(__GNUC__) && !defined(__STRICT_ANSI__)) || defined(__cplusplus) || defined(__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */ |
| 25 | # define INLINE_KEYWORD inline |
| 26 | #else |
| 27 | # define INLINE_KEYWORD |
| 28 | #endif |
| 29 | |
| 30 | #if defined(__GNUC__) || defined(__ICCARM__) |
| 31 | # define FORCE_INLINE_ATTR __attribute__((always_inline)) |
| 32 | #elif defined(_MSC_VER) |
| 33 | # define FORCE_INLINE_ATTR __forceinline |
| 34 | #else |
| 35 | # define FORCE_INLINE_ATTR |
| 36 | #endif |
| 37 | |
| 38 | #else |
| 39 | |
| 40 | #define INLINE_KEYWORD |
| 41 | #define FORCE_INLINE_ATTR |
| 42 | |
| 43 | #endif |
| 44 | |
| 45 | /** |
| 46 | On MSVC qsort requires that functions passed into it use the __cdecl calling conversion(CC). |
| 47 | This explicitly marks such functions as __cdecl so that the code will still compile |
| 48 | if a CC other than __cdecl has been made the default. |
| 49 | */ |
| 50 | #if defined(_MSC_VER) |
| 51 | # define WIN_CDECL __cdecl |
| 52 | #else |
| 53 | # define WIN_CDECL |
| 54 | #endif |
| 55 | |
| 56 | /* UNUSED_ATTR tells the compiler it is okay if the function is unused. */ |
| 57 | #if defined(__GNUC__) |
| 58 | # define UNUSED_ATTR __attribute__((unused)) |
| 59 | #else |
| 60 | # define UNUSED_ATTR |
| 61 | #endif |
| 62 | |
| 63 | /** |
| 64 | * FORCE_INLINE_TEMPLATE is used to define C "templates", which take constant |
| 65 | * parameters. They must be inlined for the compiler to eliminate the constant |
| 66 | * branches. |
| 67 | */ |
| 68 | #define FORCE_INLINE_TEMPLATE static INLINE_KEYWORD FORCE_INLINE_ATTR UNUSED_ATTR |
| 69 | /** |
| 70 | * HINT_INLINE is used to help the compiler generate better code. It is *not* |
| 71 | * used for "templates", so it can be tweaked based on the compilers |
| 72 | * performance. |
| 73 | * |
| 74 | * gcc-4.8 and gcc-4.9 have been shown to benefit from leaving off the |
| 75 | * always_inline attribute. |
| 76 | * |
| 77 | * clang up to 5.0.0 (trunk) benefit tremendously from the always_inline |
| 78 | * attribute. |
| 79 | */ |
| 80 | #if !defined(__clang__) && defined(__GNUC__) && __GNUC__ >= 4 && __GNUC_MINOR__ >= 8 && __GNUC__ < 5 |
| 81 | # define HINT_INLINE static INLINE_KEYWORD |
| 82 | #else |
| 83 | # define HINT_INLINE FORCE_INLINE_TEMPLATE |
| 84 | #endif |
| 85 | |
| 86 | /* "soft" inline : |
| 87 | * The compiler is free to select if it's a good idea to inline or not. |
| 88 | * The main objective is to silence compiler warnings |
| 89 | * when a defined function in included but not used. |
| 90 | * |
| 91 | * Note : this macro is prefixed `MEM_` because it used to be provided by `mem.h` unit. |
| 92 | * Updating the prefix is probably preferable, but requires a fairly large codemod, |
| 93 | * since this name is used everywhere. |
| 94 | */ |
| 95 | #ifndef MEM_STATIC /* already defined in Linux Kernel mem.h */ |
| 96 | #if defined(__GNUC__) |
| 97 | # define MEM_STATIC static __inline UNUSED_ATTR |
| 98 | #elif defined (__cplusplus) || (defined (__STDC_VERSION__) && (__STDC_VERSION__ >= 199901L) /* C99 */) |
| 99 | # define MEM_STATIC static inline |
| 100 | #elif defined(_MSC_VER) |
| 101 | # define MEM_STATIC static __inline |
| 102 | #else |
| 103 | # define MEM_STATIC static /* this version may generate warnings for unused static functions; disable the relevant warning */ |
| 104 | #endif |
| 105 | #endif |
| 106 | |
| 107 | /* force no inlining */ |
| 108 | #ifdef _MSC_VER |
| 109 | # define FORCE_NOINLINE static __declspec(noinline) |
| 110 | #else |
| 111 | # if defined(__GNUC__) || defined(__ICCARM__) |
| 112 | # define FORCE_NOINLINE static __attribute__((__noinline__)) |
| 113 | # else |
| 114 | # define FORCE_NOINLINE static |
| 115 | # endif |
| 116 | #endif |
| 117 | |
| 118 | |
| 119 | /* target attribute */ |
| 120 | #if defined(__GNUC__) || defined(__ICCARM__) |
| 121 | # define TARGET_ATTRIBUTE(target) __attribute__((__target__(target))) |
| 122 | #else |
| 123 | # define TARGET_ATTRIBUTE(target) |
| 124 | #endif |
| 125 | |
| 126 | /* Target attribute for BMI2 dynamic dispatch. |
| 127 | * Enable lzcnt, bmi, and bmi2. |
| 128 | * We test for bmi1 & bmi2. lzcnt is included in bmi1. |
| 129 | */ |
| 130 | #define BMI2_TARGET_ATTRIBUTE TARGET_ATTRIBUTE("lzcnt,bmi,bmi2") |
| 131 | |
| 132 | /* prefetch |
| 133 | * can be disabled, by declaring NO_PREFETCH build macro */ |
| 134 | #if defined(NO_PREFETCH) |
| 135 | # define PREFETCH_L1(ptr) do { (void)(ptr); } while (0) /* disabled */ |
| 136 | # define PREFETCH_L2(ptr) do { (void)(ptr); } while (0) /* disabled */ |
| 137 | #else |
| 138 | # if defined(_MSC_VER) && (defined(_M_X64) || defined(_M_I86)) && !defined(_M_ARM64EC) /* _mm_prefetch() is not defined outside of x86/x64 */ |
| 139 | # include <mmintrin.h> /* https://msdn.microsoft.com/fr-fr/library/84szxsww(v=vs.90).aspx */ |
| 140 | # define PREFETCH_L1(ptr) _mm_prefetch((const char*)(ptr), _MM_HINT_T0) |
| 141 | # define PREFETCH_L2(ptr) _mm_prefetch((const char*)(ptr), _MM_HINT_T1) |
| 142 | # elif defined(__GNUC__) && ( (__GNUC__ >= 4) || ( (__GNUC__ == 3) && (__GNUC_MINOR__ >= 1) ) ) |
| 143 | # define PREFETCH_L1(ptr) __builtin_prefetch((ptr), 0 /* rw==read */, 3 /* locality */) |
| 144 | # define PREFETCH_L2(ptr) __builtin_prefetch((ptr), 0 /* rw==read */, 2 /* locality */) |
| 145 | # elif defined(__aarch64__) |
| 146 | # define PREFETCH_L1(ptr) do { __asm__ __volatile__("prfm pldl1keep, %0" ::"Q"(*(ptr))); } while (0) |
| 147 | # define PREFETCH_L2(ptr) do { __asm__ __volatile__("prfm pldl2keep, %0" ::"Q"(*(ptr))); } while (0) |
| 148 | # else |
| 149 | # define PREFETCH_L1(ptr) do { (void)(ptr); } while (0) /* disabled */ |
| 150 | # define PREFETCH_L2(ptr) do { (void)(ptr); } while (0) /* disabled */ |
| 151 | # endif |
| 152 | #endif /* NO_PREFETCH */ |
| 153 | |
| 154 | #define CACHELINE_SIZE 64 |
| 155 | |
| 156 | #define PREFETCH_AREA(p, s) \ |
| 157 | do { \ |
| 158 | const char* const _ptr = (const char*)(p); \ |
| 159 | size_t const _size = (size_t)(s); \ |
| 160 | size_t _pos; \ |
| 161 | for (_pos=0; _pos<_size; _pos+=CACHELINE_SIZE) { \ |
| 162 | PREFETCH_L2(_ptr + _pos); \ |
| 163 | } \ |
| 164 | } while (0) |
| 165 | |
| 166 | /* vectorization |
| 167 | * older GCC (pre gcc-4.3 picked as the cutoff) uses a different syntax, |
| 168 | * and some compilers, like Intel ICC and MCST LCC, do not support it at all. */ |
| 169 | #if !defined(__INTEL_COMPILER) && !defined(__clang__) && defined(__GNUC__) && !defined(__LCC__) |
| 170 | # if (__GNUC__ == 4 && __GNUC_MINOR__ > 3) || (__GNUC__ >= 5) |
| 171 | # define DONT_VECTORIZE __attribute__((optimize("no-tree-vectorize"))) |
| 172 | # else |
| 173 | # define DONT_VECTORIZE _Pragma("GCC optimize(\"no-tree-vectorize\")") |
| 174 | # endif |
| 175 | #else |
| 176 | # define DONT_VECTORIZE |
| 177 | #endif |
| 178 | |
| 179 | /* Tell the compiler that a branch is likely or unlikely. |
| 180 | * Only use these macros if it causes the compiler to generate better code. |
| 181 | * If you can remove a LIKELY/UNLIKELY annotation without speed changes in gcc |
| 182 | * and clang, please do. |
| 183 | */ |
| 184 | #if defined(__GNUC__) |
| 185 | #define LIKELY(x) (__builtin_expect((x), 1)) |
| 186 | #define UNLIKELY(x) (__builtin_expect((x), 0)) |
| 187 | #else |
| 188 | #define LIKELY(x) (x) |
| 189 | #define UNLIKELY(x) (x) |
| 190 | #endif |
| 191 | |
| 192 | #if __has_builtin(__builtin_unreachable) || (defined(__GNUC__) && (__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 5))) |
| 193 | # define ZSTD_UNREACHABLE do { assert(0), __builtin_unreachable(); } while (0) |
| 194 | #else |
| 195 | # define ZSTD_UNREACHABLE do { assert(0); } while (0) |
| 196 | #endif |
| 197 | |
| 198 | /* disable warnings */ |
| 199 | #ifdef _MSC_VER /* Visual Studio */ |
| 200 | # include <intrin.h> /* For Visual 2005 */ |
| 201 | # pragma warning(disable : 4100) /* disable: C4100: unreferenced formal parameter */ |
| 202 | # pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */ |
| 203 | # pragma warning(disable : 4204) /* disable: C4204: non-constant aggregate initializer */ |
| 204 | # pragma warning(disable : 4214) /* disable: C4214: non-int bitfields */ |
| 205 | # pragma warning(disable : 4324) /* disable: C4324: padded structure */ |
| 206 | #endif |
| 207 | |
| 208 | /*Like DYNAMIC_BMI2 but for compile time determination of BMI2 support*/ |
| 209 | #ifndef STATIC_BMI2 |
| 210 | # if defined(_MSC_VER) && (defined(_M_X64) || defined(_M_I86)) |
| 211 | # ifdef __AVX2__ //MSVC does not have a BMI2 specific flag, but every CPU that supports AVX2 also supports BMI2 |
| 212 | # define STATIC_BMI2 1 |
| 213 | # endif |
| 214 | # elif defined(__BMI2__) && defined(__x86_64__) && defined(__GNUC__) |
| 215 | # define STATIC_BMI2 1 |
| 216 | # endif |
| 217 | #endif |
| 218 | |
| 219 | #ifndef STATIC_BMI2 |
| 220 | #define STATIC_BMI2 0 |
| 221 | #endif |
| 222 | |
| 223 | /* compile time determination of SIMD support */ |
| 224 | #if !defined(ZSTD_NO_INTRINSICS) |
| 225 | # if defined(__SSE2__) || defined(_M_AMD64) || (defined (_M_IX86) && defined(_M_IX86_FP) && (_M_IX86_FP >= 2)) |
| 226 | # define ZSTD_ARCH_X86_SSE2 |
| 227 | # endif |
| 228 | # if defined(__ARM_NEON) || defined(_M_ARM64) |
| 229 | # define ZSTD_ARCH_ARM_NEON |
| 230 | # endif |
| 231 | # |
| 232 | # if defined(ZSTD_ARCH_X86_SSE2) |
| 233 | # include <emmintrin.h> |
| 234 | # elif defined(ZSTD_ARCH_ARM_NEON) |
| 235 | # include <arm_neon.h> |
| 236 | # endif |
| 237 | #endif |
| 238 | |
| 239 | /* C-language Attributes are added in C23. */ |
| 240 | #if defined(__STDC_VERSION__) && (__STDC_VERSION__ > 201710L) && defined(__has_c_attribute) |
| 241 | # define ZSTD_HAS_C_ATTRIBUTE(x) __has_c_attribute(x) |
| 242 | #else |
| 243 | # define ZSTD_HAS_C_ATTRIBUTE(x) 0 |
| 244 | #endif |
| 245 | |
| 246 | /* Only use C++ attributes in C++. Some compilers report support for C++ |
| 247 | * attributes when compiling with C. |
| 248 | */ |
| 249 | #if defined(__cplusplus) && defined(__has_cpp_attribute) |
| 250 | # define ZSTD_HAS_CPP_ATTRIBUTE(x) __has_cpp_attribute(x) |
| 251 | #else |
| 252 | # define ZSTD_HAS_CPP_ATTRIBUTE(x) 0 |
| 253 | #endif |
| 254 | |
| 255 | /* Define ZSTD_FALLTHROUGH macro for annotating switch case with the 'fallthrough' attribute. |
| 256 | * - C23: https://en.cppreference.com/w/c/language/attributes/fallthrough |
| 257 | * - CPP17: https://en.cppreference.com/w/cpp/language/attributes/fallthrough |
| 258 | * - Else: __attribute__((__fallthrough__)) |
| 259 | */ |
| 260 | #ifndef ZSTD_FALLTHROUGH |
| 261 | # if ZSTD_HAS_C_ATTRIBUTE(fallthrough) |
| 262 | # define ZSTD_FALLTHROUGH [[fallthrough]] |
| 263 | # elif ZSTD_HAS_CPP_ATTRIBUTE(fallthrough) |
| 264 | # define ZSTD_FALLTHROUGH [[fallthrough]] |
| 265 | # elif __has_attribute(__fallthrough__) |
| 266 | /* Leading semicolon is to satisfy gcc-11 with -pedantic. Without the semicolon |
| 267 | * gcc complains about: a label can only be part of a statement and a declaration is not a statement. |
| 268 | */ |
| 269 | # define ZSTD_FALLTHROUGH ; __attribute__((__fallthrough__)) |
| 270 | # else |
| 271 | # define ZSTD_FALLTHROUGH |
| 272 | # endif |
| 273 | #endif |
| 274 | |
| 275 | /*-************************************************************** |
| 276 | * Alignment check |
| 277 | *****************************************************************/ |
| 278 | |
| 279 | /* this test was initially positioned in mem.h, |
| 280 | * but this file is removed (or replaced) for linux kernel |
| 281 | * so it's now hosted in compiler.h, |
| 282 | * which remains valid for both user & kernel spaces. |
| 283 | */ |
| 284 | |
| 285 | #ifndef ZSTD_ALIGNOF |
| 286 | # if defined(__GNUC__) || defined(_MSC_VER) |
| 287 | /* covers gcc, clang & MSVC */ |
| 288 | /* note : this section must come first, before C11, |
| 289 | * due to a limitation in the kernel source generator */ |
| 290 | # define ZSTD_ALIGNOF(T) __alignof(T) |
| 291 | |
| 292 | # elif defined(__STDC_VERSION__) && (__STDC_VERSION__ >= 201112L) |
| 293 | /* C11 support */ |
| 294 | # include <stdalign.h> |
| 295 | # define ZSTD_ALIGNOF(T) alignof(T) |
| 296 | |
| 297 | # else |
| 298 | /* No known support for alignof() - imperfect backup */ |
| 299 | # define ZSTD_ALIGNOF(T) (sizeof(void*) < sizeof(T) ? sizeof(void*) : sizeof(T)) |
| 300 | |
| 301 | # endif |
| 302 | #endif /* ZSTD_ALIGNOF */ |
| 303 | |
| 304 | /*-************************************************************** |
| 305 | * Sanitizer |
| 306 | *****************************************************************/ |
| 307 | |
| 308 | /** |
| 309 | * Zstd relies on pointer overflow in its decompressor. |
| 310 | * We add this attribute to functions that rely on pointer overflow. |
| 311 | */ |
| 312 | #ifndef ZSTD_ALLOW_POINTER_OVERFLOW_ATTR |
| 313 | # if __has_attribute(no_sanitize) |
| 314 | # if !defined(__clang__) && defined(__GNUC__) && __GNUC__ < 8 |
| 315 | /* gcc < 8 only has signed-integer-overlow which triggers on pointer overflow */ |
| 316 | # define ZSTD_ALLOW_POINTER_OVERFLOW_ATTR __attribute__((no_sanitize("signed-integer-overflow"))) |
| 317 | # else |
| 318 | /* older versions of clang [3.7, 5.0) will warn that pointer-overflow is ignored. */ |
| 319 | # define ZSTD_ALLOW_POINTER_OVERFLOW_ATTR __attribute__((no_sanitize("pointer-overflow"))) |
| 320 | # endif |
| 321 | # else |
| 322 | # define ZSTD_ALLOW_POINTER_OVERFLOW_ATTR |
| 323 | # endif |
| 324 | #endif |
| 325 | |
| 326 | /** |
| 327 | * Helper function to perform a wrapped pointer difference without trigging |
| 328 | * UBSAN. |
| 329 | * |
| 330 | * @returns lhs - rhs with wrapping |
| 331 | */ |
| 332 | MEM_STATIC |
| 333 | ZSTD_ALLOW_POINTER_OVERFLOW_ATTR |
| 334 | ptrdiff_t ZSTD_wrappedPtrDiff(unsigned char const* lhs, unsigned char const* rhs) |
| 335 | { |
| 336 | return lhs - rhs; |
| 337 | } |
| 338 | |
| 339 | /** |
| 340 | * Helper function to perform a wrapped pointer add without triggering UBSAN. |
| 341 | * |
| 342 | * @return ptr + add with wrapping |
| 343 | */ |
| 344 | MEM_STATIC |
| 345 | ZSTD_ALLOW_POINTER_OVERFLOW_ATTR |
| 346 | unsigned char const* ZSTD_wrappedPtrAdd(unsigned char const* ptr, ptrdiff_t add) |
| 347 | { |
| 348 | return ptr + add; |
| 349 | } |
| 350 | |
| 351 | /** |
| 352 | * Helper function to perform a wrapped pointer subtraction without triggering |
| 353 | * UBSAN. |
| 354 | * |
| 355 | * @return ptr - sub with wrapping |
| 356 | */ |
| 357 | MEM_STATIC |
| 358 | ZSTD_ALLOW_POINTER_OVERFLOW_ATTR |
| 359 | unsigned char const* ZSTD_wrappedPtrSub(unsigned char const* ptr, ptrdiff_t sub) |
| 360 | { |
| 361 | return ptr - sub; |
| 362 | } |
| 363 | |
| 364 | /** |
| 365 | * Helper function to add to a pointer that works around C's undefined behavior |
| 366 | * of adding 0 to NULL. |
| 367 | * |
| 368 | * @returns `ptr + add` except it defines `NULL + 0 == NULL`. |
| 369 | */ |
| 370 | MEM_STATIC |
| 371 | unsigned char* ZSTD_maybeNullPtrAdd(unsigned char* ptr, ptrdiff_t add) |
| 372 | { |
| 373 | return add > 0 ? ptr + add : ptr; |
| 374 | } |
| 375 | |
| 376 | /* Issue #3240 reports an ASAN failure on an llvm-mingw build. Out of an |
| 377 | * abundance of caution, disable our custom poisoning on mingw. */ |
| 378 | #ifdef __MINGW32__ |
| 379 | #ifndef ZSTD_ASAN_DONT_POISON_WORKSPACE |
| 380 | #define ZSTD_ASAN_DONT_POISON_WORKSPACE 1 |
| 381 | #endif |
| 382 | #ifndef ZSTD_MSAN_DONT_POISON_WORKSPACE |
| 383 | #define ZSTD_MSAN_DONT_POISON_WORKSPACE 1 |
| 384 | #endif |
| 385 | #endif |
| 386 | |
| 387 | #if ZSTD_MEMORY_SANITIZER && !defined(ZSTD_MSAN_DONT_POISON_WORKSPACE) |
| 388 | /* Not all platforms that support msan provide sanitizers/msan_interface.h. |
| 389 | * We therefore declare the functions we need ourselves, rather than trying to |
| 390 | * include the header file... */ |
| 391 | #include <stddef.h> /* size_t */ |
| 392 | #define ZSTD_DEPS_NEED_STDINT |
| 393 | #include "zstd_deps.h" /* intptr_t */ |
| 394 | |
| 395 | /* Make memory region fully initialized (without changing its contents). */ |
| 396 | void __msan_unpoison(const volatile void *a, size_t size); |
| 397 | |
| 398 | /* Make memory region fully uninitialized (without changing its contents). |
| 399 | This is a legacy interface that does not update origin information. Use |
| 400 | __msan_allocated_memory() instead. */ |
| 401 | void __msan_poison(const volatile void *a, size_t size); |
| 402 | |
| 403 | /* Returns the offset of the first (at least partially) poisoned byte in the |
| 404 | memory range, or -1 if the whole range is good. */ |
| 405 | intptr_t __msan_test_shadow(const volatile void *x, size_t size); |
| 406 | |
| 407 | /* Print shadow and origin for the memory range to stderr in a human-readable |
| 408 | format. */ |
| 409 | void __msan_print_shadow(const volatile void *x, size_t size); |
| 410 | #endif |
| 411 | |
| 412 | #if ZSTD_ADDRESS_SANITIZER && !defined(ZSTD_ASAN_DONT_POISON_WORKSPACE) |
| 413 | /* Not all platforms that support asan provide sanitizers/asan_interface.h. |
| 414 | * We therefore declare the functions we need ourselves, rather than trying to |
| 415 | * include the header file... */ |
| 416 | #include <stddef.h> /* size_t */ |
| 417 | |
| 418 | /** |
| 419 | * Marks a memory region (<c>[addr, addr+size)</c>) as unaddressable. |
| 420 | * |
| 421 | * This memory must be previously allocated by your program. Instrumented |
| 422 | * code is forbidden from accessing addresses in this region until it is |
| 423 | * unpoisoned. This function is not guaranteed to poison the entire region - |
| 424 | * it could poison only a subregion of <c>[addr, addr+size)</c> due to ASan |
| 425 | * alignment restrictions. |
| 426 | * |
| 427 | * \note This function is not thread-safe because no two threads can poison or |
| 428 | * unpoison memory in the same memory region simultaneously. |
| 429 | * |
| 430 | * \param addr Start of memory region. |
| 431 | * \param size Size of memory region. */ |
| 432 | void __asan_poison_memory_region(void const volatile *addr, size_t size); |
| 433 | |
| 434 | /** |
| 435 | * Marks a memory region (<c>[addr, addr+size)</c>) as addressable. |
| 436 | * |
| 437 | * This memory must be previously allocated by your program. Accessing |
| 438 | * addresses in this region is allowed until this region is poisoned again. |
| 439 | * This function could unpoison a super-region of <c>[addr, addr+size)</c> due |
| 440 | * to ASan alignment restrictions. |
| 441 | * |
| 442 | * \note This function is not thread-safe because no two threads can |
| 443 | * poison or unpoison memory in the same memory region simultaneously. |
| 444 | * |
| 445 | * \param addr Start of memory region. |
| 446 | * \param size Size of memory region. */ |
| 447 | void __asan_unpoison_memory_region(void const volatile *addr, size_t size); |
| 448 | #endif |
| 449 | |
| 450 | #endif /* ZSTD_COMPILER_H */ |