summaryrefslogtreecommitdiff
path: root/neozip/zbuild.h
diff options
context:
space:
mode:
Diffstat (limited to 'neozip/zbuild.h')
-rw-r--r--neozip/zbuild.h352
1 files changed, 352 insertions, 0 deletions
diff --git a/neozip/zbuild.h b/neozip/zbuild.h
new file mode 100644
index 0000000000..16a9d0d774
--- /dev/null
+++ b/neozip/zbuild.h
@@ -0,0 +1,352 @@
+#ifndef _ZBUILD_H
+#define _ZBUILD_H
+
+#define _POSIX_SOURCE 1 /* fileno */
+#ifndef _POSIX_C_SOURCE
+# define _POSIX_C_SOURCE 200809L /* snprintf, posix_memalign, strdup */
+#endif
+#ifndef _ISOC11_SOURCE
+# define _ISOC11_SOURCE 1 /* aligned_alloc */
+#endif
+#ifdef __OpenBSD__
+# define _BSD_SOURCE 1
+#endif
+
+#include <stddef.h>
+#include <string.h>
+#include <stdlib.h>
+#include <stdint.h>
+#include <stdio.h>
+
+#include "zarch.h"
+
+/* Determine compiler version of C Standard */
+#ifdef __STDC_VERSION__
+# if __STDC_VERSION__ >= 199901L
+# ifndef STDC99
+# define STDC99
+# endif
+# endif
+# if __STDC_VERSION__ >= 201112L
+# ifndef STDC11
+# define STDC11
+# endif
+# endif
+#endif
+
+#ifndef Z_HAS_ATTRIBUTE
+# if defined(__has_attribute)
+# define Z_HAS_ATTRIBUTE(a) __has_attribute(a)
+# else
+# define Z_HAS_ATTRIBUTE(a) 0
+# endif
+#endif
+
+#ifndef Z_FALLTHROUGH
+# if Z_HAS_ATTRIBUTE(__fallthrough__) || (defined(__GNUC__) && (__GNUC__ >= 7))
+# define Z_FALLTHROUGH __attribute__((__fallthrough__))
+# else
+# define Z_FALLTHROUGH do {} while(0) /* fallthrough */
+# endif
+#endif
+
+/* Hint to compiler that a block of code is unreachable, typically in a switch default condition */
+#ifndef Z_UNREACHABLE
+# if defined(__STDC_VERSION__) && __STDC_VERSION__ >= 202311L
+# if !defined(unreachable) && defined(_MSC_VER)
+# define Z_UNREACHABLE() __assume(0)
+# else
+# define Z_UNREACHABLE() unreachable() // C23 approach
+# endif
+# elif (defined(__GNUC__) && (__GNUC__ >= 5)) || defined(__clang__)
+# define Z_UNREACHABLE() __builtin_unreachable()
+# else
+# define Z_UNREACHABLE()
+# endif
+#endif
+
+#ifndef Z_TARGET
+# if Z_HAS_ATTRIBUTE(__target__)
+# define Z_TARGET(x) __attribute__((__target__(x)))
+# else
+# define Z_TARGET(x)
+# endif
+#endif
+
+/* This has to be first include that defines any types */
+#if defined(_MSC_VER)
+# if defined(_WIN64)
+ typedef __int64 ssize_t;
+# else
+ typedef long ssize_t;
+# endif
+
+# if defined(_WIN64)
+ #define SSIZE_MAX _I64_MAX
+# else
+ #define SSIZE_MAX LONG_MAX
+# endif
+#endif
+
+/* A forced inline decorator */
+#if defined(_MSC_VER)
+# define Z_FORCEINLINE __forceinline
+#elif defined(__GNUC__)
+# define Z_FORCEINLINE inline __attribute__((always_inline))
+#else
+ /* It won't actually force inlining but it will suggest it */
+# define Z_FORCEINLINE inline
+#endif
+
+/* MS Visual Studio does not allow inline in C, only C++.
+ But it provides __inline instead, so use that. */
+#if defined(_MSC_VER) && !defined(inline) && !defined(__cplusplus)
+# define inline __inline
+#endif
+
+#if defined(ZLIB_COMPAT)
+# define PREFIX(x) x
+# define PREFIX2(x) ZLIB_ ## x
+# define PREFIX3(x) z_ ## x
+# define PREFIX4(x) x ## 64
+# define zVersion zlibVersion
+#else
+# define PREFIX(x) zng_ ## x
+# define PREFIX2(x) ZLIBNG_ ## x
+# define PREFIX3(x) zng_ ## x
+# define PREFIX4(x) zng_ ## x
+# define zVersion zlibng_version
+# define z_size_t size_t
+#endif
+
+/* In zlib-compat some functions and types use unsigned long, but zlib-ng use size_t */
+#if defined(ZLIB_COMPAT)
+# define z_uintmax_t unsigned long
+#else
+# define z_uintmax_t size_t
+#endif
+
+/* In zlib-compat headers some function return values and parameter types use int or unsigned, but zlib-ng headers use
+ int32_t and uint32_t, which will cause type mismatch when compiling zlib-ng if int32_t is long and uint32_t is
+ unsigned long */
+#if defined(ZLIB_COMPAT)
+# define z_int32_t int
+# define z_uint32_t unsigned int
+#else
+# define z_int32_t int32_t
+# define z_uint32_t uint32_t
+#endif
+
+/* Minimum of a and b. */
+#define MIN(a, b) ((a) > (b) ? (b) : (a))
+/* Maximum of a and b. */
+#define MAX(a, b) ((a) < (b) ? (b) : (a))
+/* Absolute value of a. */
+#define ABS(a) ((a) < 0 ? -(a) : (a))
+/* Ignore unused variable warning */
+#define Z_UNUSED(var) (void)(var)
+
+/* Force the compiler to treat variable as modified. Empty asm statement with a "+r" constraint prevents
+ the compiler from reordering or eliminating loads into the variable. This can help keep critical latency
+ chains in the hot path from being shortened or optimized away. */
+#if (defined(__GNUC__) || defined(__clang__)) && \
+ (defined(ARCH_X86) || (defined(ARCH_ARM) && defined(ARCH_64BIT)))
+# define Z_TOUCH(var) __asm__ ("" : "+r"(var))
+#else
+# define Z_TOUCH(var) (void)(var)
+#endif
+
+#if defined(HAVE_VISIBILITY_INTERNAL)
+# define Z_INTERNAL __attribute__((visibility ("internal")))
+#elif defined(HAVE_VISIBILITY_HIDDEN)
+# define Z_INTERNAL __attribute__((visibility ("hidden")))
+#else
+# define Z_INTERNAL
+#endif
+
+/* Symbol versioning helpers, allowing multiple versions of a function to exist.
+ * Functions using this must also be added to zlib-ng.map for each version.
+ * Double @@ means this is the default for newly compiled applications to link against.
+ * Single @ means this is kept for backwards compatibility.
+ * This is only used for Zlib-ng native API, and only on platforms supporting this.
+ */
+#if defined(HAVE_SYMVER)
+# define ZSYMVER(func,alias,ver) __asm__(".symver " func ", " alias "@ZLIB_NG_" ver);
+# define ZSYMVER_DEF(func,alias,ver) __asm__(".symver " func ", " alias "@@ZLIB_NG_" ver);
+#else
+# define ZSYMVER(func,alias,ver)
+# define ZSYMVER_DEF(func,alias,ver)
+#endif
+
+#ifndef __cplusplus
+# define Z_REGISTER register
+#else
+# define Z_REGISTER
+#endif
+
+#if defined(_MSC_VER)
+# define Z_RESTRICT __restrict
+#elif defined(__STDC_VERSION__) && __STDC_VERSION__ >= 199901L
+# define Z_RESTRICT restrict
+#else
+# define Z_RESTRICT __restrict__
+#endif
+
+/* Reverse the bytes in a value. Use compiler intrinsics when
+ possible to take advantage of hardware implementations. */
+#if defined(_MSC_VER) && (_MSC_VER >= 1300)
+# include <stdlib.h>
+# pragma intrinsic(_byteswap_ulong)
+# define ZSWAP16(q) _byteswap_ushort(q)
+# define ZSWAP32(q) _byteswap_ulong(q)
+# define ZSWAP64(q) _byteswap_uint64(q)
+
+#elif defined(__clang__) || (defined(__GNUC__) && \
+ (__GNUC__ > 4 || (__GNUC__ == 4 && __GNUC_MINOR__ >= 8)))
+# define ZSWAP16(q) __builtin_bswap16(q)
+# define ZSWAP32(q) __builtin_bswap32(q)
+# define ZSWAP64(q) __builtin_bswap64(q)
+
+#elif defined(__GNUC__) && (__GNUC__ >= 2) && defined(__linux__)
+# include <byteswap.h>
+# define ZSWAP16(q) bswap_16(q)
+# define ZSWAP32(q) bswap_32(q)
+# define ZSWAP64(q) bswap_64(q)
+
+#elif defined(__FreeBSD__) || defined(__NetBSD__) || defined(__DragonFly__)
+# include <sys/endian.h>
+# define ZSWAP16(q) bswap16(q)
+# define ZSWAP32(q) bswap32(q)
+# define ZSWAP64(q) bswap64(q)
+#elif defined(__OpenBSD__)
+# include <sys/endian.h>
+# define ZSWAP16(q) swap16(q)
+# define ZSWAP32(q) swap32(q)
+# define ZSWAP64(q) swap64(q)
+#elif defined(__INTEL_COMPILER)
+/* ICC does not provide a two byte swap. */
+# define ZSWAP16(q) ((((q) & 0xff) << 8) | (((q) & 0xff00) >> 8))
+# define ZSWAP32(q) _bswap(q)
+# define ZSWAP64(q) _bswap64(q)
+
+#else
+# define ZSWAP16(q) ((((q) & 0xff) << 8) | (((q) & 0xff00) >> 8))
+# define ZSWAP32(q) ((((q) >> 24) & 0xff) + (((q) >> 8) & 0xff00) + \
+ (((q) & 0xff00) << 8) + (((q) & 0xff) << 24))
+# define ZSWAP64(q) \
+ (((q & 0xFF00000000000000u) >> 56u) | \
+ ((q & 0x00FF000000000000u) >> 40u) | \
+ ((q & 0x0000FF0000000000u) >> 24u) | \
+ ((q & 0x000000FF00000000u) >> 8u) | \
+ ((q & 0x00000000FF000000u) << 8u) | \
+ ((q & 0x0000000000FF0000u) << 24u) | \
+ ((q & 0x000000000000FF00u) << 40u) | \
+ ((q & 0x00000000000000FFu) << 56u))
+#endif
+
+/* Only enable likely/unlikely if the compiler is known to support it */
+#if (defined(__GNUC__) && (__GNUC__ >= 3)) || defined(__INTEL_COMPILER) || defined(__clang__)
+# define LIKELY_NULL(x) __builtin_expect((x) != 0, 0)
+# define LIKELY(x) __builtin_expect(!!(x), 1)
+# define UNLIKELY(x) __builtin_expect(!!(x), 0)
+#else
+# define LIKELY_NULL(x) x
+# define LIKELY(x) x
+# define UNLIKELY(x) x
+#endif /* (un)likely */
+
+#if defined(HAVE_ATTRIBUTE_ALIGNED)
+# define ALIGNED_(x) __attribute__ ((aligned(x)))
+#elif defined(_MSC_VER)
+# define ALIGNED_(x) __declspec(align(x))
+#else
+/* TODO: Define ALIGNED_ for your compiler */
+# define ALIGNED_(x)
+#endif
+
+#ifdef HAVE_BUILTIN_ASSUME_ALIGNED
+# define HINT_ALIGNED(p,n) __builtin_assume_aligned((void *)(p),(n))
+#else
+# define HINT_ALIGNED(p,n) (p)
+#endif
+#define HINT_ALIGNED_16(p) HINT_ALIGNED((p),16)
+#define HINT_ALIGNED_64(p) HINT_ALIGNED((p),64)
+#define HINT_ALIGNED_4096(p) HINT_ALIGNED((p),4096)
+
+/* Number of bytes needed to align ptr to the next alignment boundary */
+#define ALIGN_DIFF(ptr, align) \
+ (((uintptr_t)(align) - ((uintptr_t)(ptr) & ((align) - 1))) & ((align) - 1))
+
+/* Round up value to the nearest multiple of align (align must be power of 2) */
+#define ALIGN_UP(value, align) \
+ (((value) + ((align) - 1)) & ~((align) - 1))
+
+/* Round down value to the nearest multiple of align (align must be power of 2) */
+#define ALIGN_DOWN(value, align) \
+ ((value) & ~((align) - 1))
+
+/* PADSZ returns needed bytes to pad bpos to pad size
+ * PAD_NN calculates pad size and adds it to bpos, returning the result.
+ * All take an integer or a pointer as bpos input.
+ */
+#define PADSZ(bpos, pad) (((pad) - ((uintptr_t)(bpos) % (pad))) % (pad))
+#define PAD_16(bpos) ((bpos) + PADSZ((bpos),16))
+#define PAD_64(bpos) ((bpos) + PADSZ((bpos),64))
+#define PAD_4096(bpos) ((bpos) + PADSZ((bpos),4096))
+
+/* Diagnostic functions */
+#ifdef ZLIB_DEBUG
+ extern int Z_INTERNAL z_verbose;
+ extern void Z_INTERNAL z_error(const char *m);
+# define Assert(cond, msg) {int _cond = (cond); if (!(_cond)) z_error(msg);}
+# define Trace(x) {if (z_verbose >= 0) fprintf x;}
+# define Tracev(x) {if (z_verbose > 0) fprintf x;}
+# define Tracevv(x) {if (z_verbose > 1) fprintf x;}
+# define Tracec(c, x) {if (z_verbose > 0 && (c)) fprintf x;}
+# define Tracecv(c, x) {if (z_verbose > 1 && (c)) fprintf x;}
+#else
+# define Assert(cond, msg)
+# define Trace(x)
+# define Tracev(x)
+# define Tracevv(x)
+# define Tracec(c, x)
+# define Tracecv(c, x)
+#endif
+
+/* OPTIMAL_CMP values determine the comparison width:
+ * 64: Best for 64-bit architectures with unaligned access
+ * 32: Best for 32-bit architectures with unaligned access
+ * 16: Safe default for unknown architectures
+ * 8: Safe fallback for architectures without unaligned access
+ * Note: The unaligned access mentioned is cpu-support, this allows compiler or
+ * separate unaligned intrinsics to utilize safe unaligned access, without
+ * utilizing unaligned C pointers that are known to have undefined behavior.
+ */
+#if !defined(OPTIMAL_CMP)
+# ifdef ARCH_64BIT
+# ifdef ARCH_ARM
+# if defined(__ARM_FEATURE_UNALIGNED) || defined(_WIN32)
+# define OPTIMAL_CMP 64
+# else
+# define OPTIMAL_CMP 8
+# endif
+# else
+# define OPTIMAL_CMP 64
+# endif
+# elif defined(ARCH_32BIT)
+# ifdef ARCH_ARM
+# if defined(__ARM_FEATURE_UNALIGNED) || defined(_WIN32)
+# define OPTIMAL_CMP 32
+# else
+# define OPTIMAL_CMP 8
+# endif
+# else
+# define OPTIMAL_CMP 32
+# endif
+# endif
+#endif
+#if !defined(OPTIMAL_CMP)
+# define OPTIMAL_CMP 16
+#endif
+
+#endif