2017-08-27 12:05:17 +02:00
|
|
|
/*
|
2020-09-18 21:38:36 +02:00
|
|
|
* Copyright (c) 2016-2020, Yann Collet, Facebook, Inc.
|
2017-08-27 12:05:17 +02:00
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* This source code is licensed under both the BSD-style license (found in the
|
|
|
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
|
|
|
* in the COPYING file in the root directory of this source tree).
|
2017-10-26 22:41:47 +02:00
|
|
|
* You may select, at your option, one of the above-listed licenses.
|
2017-08-27 12:05:17 +02:00
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef ZSTD_COMPILER_H
|
|
|
|
#define ZSTD_COMPILER_H
|
|
|
|
|
|
|
|
/*-*******************************************************
|
|
|
|
* Compiler specifics
|
|
|
|
*********************************************************/
|
|
|
|
/* force inlining */
|
2019-01-04 01:30:03 +01:00
|
|
|
|
|
|
|
#if !defined(ZSTD_NO_INLINE)
|
2020-09-18 21:38:36 +02:00
|
|
|
#if (defined(__GNUC__) && !defined(__STRICT_ANSI__)) || defined(__cplusplus) || defined(__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */
|
2017-08-27 12:05:17 +02:00
|
|
|
# define INLINE_KEYWORD inline
|
|
|
|
#else
|
|
|
|
# define INLINE_KEYWORD
|
|
|
|
#endif
|
|
|
|
|
2019-11-02 03:36:06 +01:00
|
|
|
#if defined(__GNUC__) || defined(__ICCARM__)
|
2017-08-27 12:05:17 +02:00
|
|
|
# define FORCE_INLINE_ATTR __attribute__((always_inline))
|
|
|
|
#elif defined(_MSC_VER)
|
|
|
|
# define FORCE_INLINE_ATTR __forceinline
|
|
|
|
#else
|
|
|
|
# define FORCE_INLINE_ATTR
|
|
|
|
#endif
|
|
|
|
|
2019-01-04 01:30:03 +01:00
|
|
|
#else
|
|
|
|
|
|
|
|
#define INLINE_KEYWORD
|
|
|
|
#define FORCE_INLINE_ATTR
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2021-01-08 11:21:43 +01:00
|
|
|
/**
|
|
|
|
On MSVC qsort requires that functions passed into it use the __cdecl calling conversion(CC).
|
|
|
|
This explictly marks such functions as __cdecl so that the code will still compile
|
|
|
|
if a CC other than __cdecl has been made the default.
|
|
|
|
*/
|
|
|
|
#if defined(_MSC_VER)
|
|
|
|
# define WIN_CDECL __cdecl
|
|
|
|
#else
|
|
|
|
# define WIN_CDECL
|
|
|
|
#endif
|
|
|
|
|
2017-08-27 12:05:17 +02:00
|
|
|
/**
|
|
|
|
* FORCE_INLINE_TEMPLATE is used to define C "templates", which take constant
|
2019-04-18 11:53:29 +02:00
|
|
|
* parameters. They must be inlined for the compiler to eliminate the constant
|
2017-08-27 12:05:17 +02:00
|
|
|
* branches.
|
|
|
|
*/
|
|
|
|
#define FORCE_INLINE_TEMPLATE static INLINE_KEYWORD FORCE_INLINE_ATTR
|
|
|
|
/**
|
|
|
|
* HINT_INLINE is used to help the compiler generate better code. It is *not*
|
|
|
|
* used for "templates", so it can be tweaked based on the compilers
|
|
|
|
* performance.
|
|
|
|
*
|
|
|
|
* gcc-4.8 and gcc-4.9 have been shown to benefit from leaving off the
|
|
|
|
* always_inline attribute.
|
|
|
|
*
|
|
|
|
* clang up to 5.0.0 (trunk) benefit tremendously from the always_inline
|
|
|
|
* attribute.
|
|
|
|
*/
|
|
|
|
#if !defined(__clang__) && defined(__GNUC__) && __GNUC__ >= 4 && __GNUC_MINOR__ >= 8 && __GNUC__ < 5
|
|
|
|
# define HINT_INLINE static INLINE_KEYWORD
|
|
|
|
#else
|
|
|
|
# define HINT_INLINE static INLINE_KEYWORD FORCE_INLINE_ATTR
|
|
|
|
#endif
|
|
|
|
|
2019-11-10 05:31:00 +01:00
|
|
|
/* UNUSED_ATTR tells the compiler it is okay if the function is unused. */
|
|
|
|
#if defined(__GNUC__)
|
|
|
|
# define UNUSED_ATTR __attribute__((unused))
|
|
|
|
#else
|
|
|
|
# define UNUSED_ATTR
|
|
|
|
#endif
|
|
|
|
|
2017-08-27 12:05:17 +02:00
|
|
|
/* force no inlining */
|
|
|
|
#ifdef _MSC_VER
|
|
|
|
# define FORCE_NOINLINE static __declspec(noinline)
|
|
|
|
#else
|
2019-11-02 03:36:06 +01:00
|
|
|
# if defined(__GNUC__) || defined(__ICCARM__)
|
2017-08-27 12:05:17 +02:00
|
|
|
# define FORCE_NOINLINE static __attribute__((__noinline__))
|
|
|
|
# else
|
|
|
|
# define FORCE_NOINLINE static
|
|
|
|
# endif
|
|
|
|
#endif
|
|
|
|
|
2018-05-15 19:45:22 +02:00
|
|
|
/* target attribute */
|
|
|
|
#ifndef __has_attribute
|
|
|
|
#define __has_attribute(x) 0 /* Compatibility with non-clang compilers. */
|
|
|
|
#endif
|
2019-11-02 03:36:06 +01:00
|
|
|
#if defined(__GNUC__) || defined(__ICCARM__)
|
2018-05-15 19:45:22 +02:00
|
|
|
# define TARGET_ATTRIBUTE(target) __attribute__((__target__(target)))
|
|
|
|
#else
|
|
|
|
# define TARGET_ATTRIBUTE(target)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* Enable runtime BMI2 dispatch based on the CPU.
|
|
|
|
* Enabled for clang & gcc >=4.8 on x86 when BMI2 isn't enabled by default.
|
|
|
|
*/
|
|
|
|
#ifndef DYNAMIC_BMI2
|
2019-01-04 01:30:03 +01:00
|
|
|
#if ((defined(__clang__) && __has_attribute(__target__)) \
|
2018-05-15 19:45:22 +02:00
|
|
|
|| (defined(__GNUC__) \
|
2019-01-04 01:30:03 +01:00
|
|
|
&& (__GNUC__ >= 5 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8)))) \
|
2018-05-15 19:45:22 +02:00
|
|
|
&& (defined(__x86_64__) || defined(_M_X86)) \
|
|
|
|
&& !defined(__BMI2__)
|
|
|
|
# define DYNAMIC_BMI2 1
|
|
|
|
#else
|
|
|
|
# define DYNAMIC_BMI2 0
|
|
|
|
#endif
|
|
|
|
#endif
|
|
|
|
|
2019-01-04 01:30:03 +01:00
|
|
|
/* prefetch
|
|
|
|
* can be disabled, by declaring NO_PREFETCH build macro */
|
|
|
|
#if defined(NO_PREFETCH)
|
|
|
|
# define PREFETCH_L1(ptr) (void)(ptr) /* disabled */
|
|
|
|
# define PREFETCH_L2(ptr) (void)(ptr) /* disabled */
|
2017-08-27 12:05:17 +02:00
|
|
|
#else
|
2019-01-04 01:30:03 +01:00
|
|
|
# if defined(_MSC_VER) && (defined(_M_X64) || defined(_M_I86)) /* _mm_prefetch() is not defined outside of x86/x64 */
|
|
|
|
# include <mmintrin.h> /* https://msdn.microsoft.com/fr-fr/library/84szxsww(v=vs.90).aspx */
|
|
|
|
# define PREFETCH_L1(ptr) _mm_prefetch((const char*)(ptr), _MM_HINT_T0)
|
|
|
|
# define PREFETCH_L2(ptr) _mm_prefetch((const char*)(ptr), _MM_HINT_T1)
|
|
|
|
# elif defined(__GNUC__) && ( (__GNUC__ >= 4) || ( (__GNUC__ == 3) && (__GNUC_MINOR__ >= 1) ) )
|
|
|
|
# define PREFETCH_L1(ptr) __builtin_prefetch((ptr), 0 /* rw==read */, 3 /* locality */)
|
|
|
|
# define PREFETCH_L2(ptr) __builtin_prefetch((ptr), 0 /* rw==read */, 2 /* locality */)
|
2021-01-08 11:21:43 +01:00
|
|
|
# elif defined(__aarch64__)
|
|
|
|
# define PREFETCH_L1(ptr) __asm__ __volatile__("prfm pldl1keep, %0" ::"Q"(*(ptr)))
|
|
|
|
# define PREFETCH_L2(ptr) __asm__ __volatile__("prfm pldl2keep, %0" ::"Q"(*(ptr)))
|
2019-01-04 01:30:03 +01:00
|
|
|
# else
|
|
|
|
# define PREFETCH_L1(ptr) (void)(ptr) /* disabled */
|
|
|
|
# define PREFETCH_L2(ptr) (void)(ptr) /* disabled */
|
|
|
|
# endif
|
|
|
|
#endif /* NO_PREFETCH */
|
|
|
|
|
|
|
|
#define CACHELINE_SIZE 64
|
|
|
|
|
|
|
|
#define PREFETCH_AREA(p, s) { \
|
|
|
|
const char* const _ptr = (const char*)(p); \
|
|
|
|
size_t const _size = (size_t)(s); \
|
|
|
|
size_t _pos; \
|
|
|
|
for (_pos=0; _pos<_size; _pos+=CACHELINE_SIZE) { \
|
|
|
|
PREFETCH_L2(_ptr + _pos); \
|
|
|
|
} \
|
|
|
|
}
|
2017-08-27 12:05:17 +02:00
|
|
|
|
2019-11-10 05:31:00 +01:00
|
|
|
/* vectorization
|
|
|
|
* older GCC (pre gcc-4.3 picked as the cutoff) uses a different syntax */
|
2020-09-18 21:38:36 +02:00
|
|
|
#if !defined(__INTEL_COMPILER) && !defined(__clang__) && defined(__GNUC__)
|
2019-11-10 05:31:00 +01:00
|
|
|
# if (__GNUC__ == 4 && __GNUC_MINOR__ > 3) || (__GNUC__ >= 5)
|
|
|
|
# define DONT_VECTORIZE __attribute__((optimize("no-tree-vectorize")))
|
|
|
|
# else
|
|
|
|
# define DONT_VECTORIZE _Pragma("GCC optimize(\"no-tree-vectorize\")")
|
|
|
|
# endif
|
2019-07-20 20:47:07 +02:00
|
|
|
#else
|
|
|
|
# define DONT_VECTORIZE
|
|
|
|
#endif
|
|
|
|
|
2020-09-18 21:38:36 +02:00
|
|
|
/* Tell the compiler that a branch is likely or unlikely.
|
|
|
|
* Only use these macros if it causes the compiler to generate better code.
|
|
|
|
* If you can remove a LIKELY/UNLIKELY annotation without speed changes in gcc
|
|
|
|
* and clang, please do.
|
|
|
|
*/
|
|
|
|
#if defined(__GNUC__)
|
|
|
|
#define LIKELY(x) (__builtin_expect((x), 1))
|
|
|
|
#define UNLIKELY(x) (__builtin_expect((x), 0))
|
|
|
|
#else
|
|
|
|
#define LIKELY(x) (x)
|
|
|
|
#define UNLIKELY(x) (x)
|
|
|
|
#endif
|
|
|
|
|
2017-08-27 12:05:17 +02:00
|
|
|
/* disable warnings */
|
|
|
|
#ifdef _MSC_VER /* Visual Studio */
|
|
|
|
# include <intrin.h> /* For Visual 2005 */
|
|
|
|
# pragma warning(disable : 4100) /* disable: C4100: unreferenced formal parameter */
|
|
|
|
# pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */
|
|
|
|
# pragma warning(disable : 4204) /* disable: C4204: non-constant aggregate initializer */
|
|
|
|
# pragma warning(disable : 4214) /* disable: C4214: non-int bitfields */
|
|
|
|
# pragma warning(disable : 4324) /* disable: C4324: padded structure */
|
|
|
|
#endif
|
|
|
|
|
2021-01-08 11:21:43 +01:00
|
|
|
/*Like DYNAMIC_BMI2 but for compile time determination of BMI2 support*/
|
|
|
|
#ifndef STATIC_BMI2
|
|
|
|
# if defined(_MSC_VER) && (defined(_M_X64) || defined(_M_I86))
|
|
|
|
# ifdef __AVX2__ //MSVC does not have a BMI2 specific flag, but every CPU that supports AVX2 also supports BMI2
|
|
|
|
# define STATIC_BMI2 1
|
|
|
|
# endif
|
|
|
|
# endif
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifndef STATIC_BMI2
|
|
|
|
#define STATIC_BMI2 0
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* compat. with non-clang compilers */
|
|
|
|
#ifndef __has_builtin
|
|
|
|
# define __has_builtin(x) 0
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* compat. with non-clang compilers */
|
|
|
|
#ifndef __has_feature
|
|
|
|
# define __has_feature(x) 0
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* detects whether we are being compiled under msan */
|
|
|
|
#ifndef ZSTD_MEMORY_SANITIZER
|
|
|
|
# if __has_feature(memory_sanitizer)
|
|
|
|
# define ZSTD_MEMORY_SANITIZER 1
|
|
|
|
# else
|
|
|
|
# define ZSTD_MEMORY_SANITIZER 0
|
|
|
|
# endif
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if ZSTD_MEMORY_SANITIZER
|
|
|
|
/* Not all platforms that support msan provide sanitizers/msan_interface.h.
|
|
|
|
* We therefore declare the functions we need ourselves, rather than trying to
|
|
|
|
* include the header file... */
|
|
|
|
#include <stddef.h> /* size_t */
|
|
|
|
#define ZSTD_DEPS_NEED_STDINT
|
|
|
|
#include "zstd_deps.h" /* intptr_t */
|
|
|
|
|
|
|
|
/* Make memory region fully initialized (without changing its contents). */
|
|
|
|
void __msan_unpoison(const volatile void *a, size_t size);
|
|
|
|
|
|
|
|
/* Make memory region fully uninitialized (without changing its contents).
|
|
|
|
This is a legacy interface that does not update origin information. Use
|
|
|
|
__msan_allocated_memory() instead. */
|
|
|
|
void __msan_poison(const volatile void *a, size_t size);
|
|
|
|
|
|
|
|
/* Returns the offset of the first (at least partially) poisoned byte in the
|
|
|
|
memory range, or -1 if the whole range is good. */
|
|
|
|
intptr_t __msan_test_shadow(const volatile void *x, size_t size);
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/* detects whether we are being compiled under asan */
|
|
|
|
#ifndef ZSTD_ADDRESS_SANITIZER
|
|
|
|
# if __has_feature(address_sanitizer)
|
|
|
|
# define ZSTD_ADDRESS_SANITIZER 1
|
|
|
|
# elif defined(__SANITIZE_ADDRESS__)
|
|
|
|
# define ZSTD_ADDRESS_SANITIZER 1
|
|
|
|
# else
|
|
|
|
# define ZSTD_ADDRESS_SANITIZER 0
|
|
|
|
# endif
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#if ZSTD_ADDRESS_SANITIZER
|
|
|
|
/* Not all platforms that support asan provide sanitizers/asan_interface.h.
|
|
|
|
* We therefore declare the functions we need ourselves, rather than trying to
|
|
|
|
* include the header file... */
|
|
|
|
#include <stddef.h> /* size_t */
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Marks a memory region (<c>[addr, addr+size)</c>) as unaddressable.
|
|
|
|
*
|
|
|
|
* This memory must be previously allocated by your program. Instrumented
|
|
|
|
* code is forbidden from accessing addresses in this region until it is
|
|
|
|
* unpoisoned. This function is not guaranteed to poison the entire region -
|
|
|
|
* it could poison only a subregion of <c>[addr, addr+size)</c> due to ASan
|
|
|
|
* alignment restrictions.
|
|
|
|
*
|
|
|
|
* \note This function is not thread-safe because no two threads can poison or
|
|
|
|
* unpoison memory in the same memory region simultaneously.
|
|
|
|
*
|
|
|
|
* \param addr Start of memory region.
|
|
|
|
* \param size Size of memory region. */
|
|
|
|
void __asan_poison_memory_region(void const volatile *addr, size_t size);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Marks a memory region (<c>[addr, addr+size)</c>) as addressable.
|
|
|
|
*
|
|
|
|
* This memory must be previously allocated by your program. Accessing
|
|
|
|
* addresses in this region is allowed until this region is poisoned again.
|
|
|
|
* This function could unpoison a super-region of <c>[addr, addr+size)</c> due
|
|
|
|
* to ASan alignment restrictions.
|
|
|
|
*
|
|
|
|
* \note This function is not thread-safe because no two threads can
|
|
|
|
* poison or unpoison memory in the same memory region simultaneously.
|
|
|
|
*
|
|
|
|
* \param addr Start of memory region.
|
|
|
|
* \param size Size of memory region. */
|
|
|
|
void __asan_unpoison_memory_region(void const volatile *addr, size_t size);
|
|
|
|
#endif
|
|
|
|
|
2017-08-27 12:05:17 +02:00
|
|
|
#endif /* ZSTD_COMPILER_H */
|