diff options
| author | Mehmet Samet Duman <yongdohyun@projecttick.org> | 2026-04-02 19:56:09 +0300 |
|---|---|---|
| committer | Mehmet Samet Duman <yongdohyun@projecttick.org> | 2026-04-02 19:56:09 +0300 |
| commit | 7fb132859fda54aa96bc9dd46d302b343eeb5a02 (patch) | |
| tree | b43ae77d7451fb470a260c03349a1caf2846c5e5 /neozip/test/benchmarks/benchmark_crc32_copy.cc | |
| parent | b1e34e861b5d732afe828d58aad2c638135061fd (diff) | |
| parent | c2712b8a345191f6ed79558c089777df94590087 (diff) | |
| download | Project-Tick-7fb132859fda54aa96bc9dd46d302b343eeb5a02.tar.gz Project-Tick-7fb132859fda54aa96bc9dd46d302b343eeb5a02.zip | |
Add 'neozip/' from commit 'c2712b8a345191f6ed79558c089777df94590087'
git-subtree-dir: neozip
git-subtree-mainline: b1e34e861b5d732afe828d58aad2c638135061fd
git-subtree-split: c2712b8a345191f6ed79558c089777df94590087
Diffstat (limited to 'neozip/test/benchmarks/benchmark_crc32_copy.cc')
| -rw-r--r-- | neozip/test/benchmarks/benchmark_crc32_copy.cc | 177 |
1 files changed, 177 insertions, 0 deletions
diff --git a/neozip/test/benchmarks/benchmark_crc32_copy.cc b/neozip/test/benchmarks/benchmark_crc32_copy.cc new file mode 100644 index 0000000000..eafa5aee44 --- /dev/null +++ b/neozip/test/benchmarks/benchmark_crc32_copy.cc @@ -0,0 +1,177 @@ +/* benchmark_crc32_copy.cc -- benchmark for crc32 implementations with copying + * Copyright (C) 2025 Hans Kristian Rosbach + * For conditions of distribution and use, see copyright notice in zlib.h + */ + +#include <benchmark/benchmark.h> + +extern "C" { +# include "zbuild.h" +# include "arch_functions.h" +# include "../test_cpu_features.h" +} + +// Hash copy functions are used on strm->next_in buffers, we process +// 512-32k sizes (x2 for initial fill) at a time if enough data is available. +#define BUFSIZE (65536 + 64) + +class crc32_copy: public benchmark::Fixture { +protected: + uint32_t *testdata; + uint8_t *dstbuf; + +public: + void SetUp(::benchmark::State& state) { + testdata = (uint32_t *)zng_alloc_aligned(BUFSIZE, 64); + dstbuf = (uint8_t *)zng_alloc_aligned(BUFSIZE, 64); + if (testdata == NULL || dstbuf == NULL) { + state.SkipWithError("malloc failed"); + return; + } + + for (uint32_t i = 0; i < BUFSIZE/sizeof(uint32_t); i++) { + testdata[i] = rand(); + } + } + + // Benchmark CRC32_copy, with rolling buffer misalignment for consistent results + void Bench(benchmark::State& state, crc32_copy_func crc32_copy, const int DO_ALIGNED) { + int misalign = 0; + uint32_t hash = 0; + + for (auto _ : state) { + hash = crc32_copy(hash, dstbuf + misalign, (const unsigned char*)testdata + misalign, (size_t)state.range(0)); + if (misalign >= 63) + misalign = 0; + else + misalign += (DO_ALIGNED) ? 16 : 1; + + // Prevent the result from being optimized away + benchmark::DoNotOptimize(hash); + } + } + + void TearDown(const ::benchmark::State&) { + zng_free_aligned(testdata); + zng_free_aligned(dstbuf); + } +}; + +// Misaligned +#define BENCHMARK_CRC32_COPY_MISALIGNED(name, copyfunc, support_flag) \ + BENCHMARK_DEFINE_F(crc32_copy, name)(benchmark::State& state) { \ + if (!(support_flag)) { \ + state.SkipWithError("CPU does not support " #name); \ + } \ + Bench(state, copyfunc, 0); \ + } \ + BENCHMARK_REGISTER_F(crc32_copy, name)->Arg(32)->Arg(512)->Arg(8<<10)->Arg(32<<10)->Arg(64<<10); + +// Aligned +#define ALIGNED_NAME(name) name##_aligned +#define BENCHMARK_CRC32_COPY_ALIGNED(name, copyfunc, support_flag) \ + BENCHMARK_DEFINE_F(crc32_copy, ALIGNED_NAME(name))(benchmark::State& state) { \ + if (!(support_flag)) { \ + state.SkipWithError("CPU does not support " #name); \ + } \ + Bench(state, copyfunc, 1); \ + } \ + BENCHMARK_REGISTER_F(crc32_copy, ALIGNED_NAME(name))->Arg(32)->Arg(512)->Arg(8<<10)->Arg(32<<10)->Arg(64<<10); + +// CRC32 + memcpy benchmarks for reference +#ifdef HASH_BASELINE +#define MEMCPY_NAME(name) name##_memcpy +#define BENCHMARK_CRC32_MEMCPY_MISALIGNED(name, hashfunc, support_flag) \ + BENCHMARK_DEFINE_F(crc32_copy, MEMCPY_NAME(name))(benchmark::State& state) { \ + if (!(support_flag)) { \ + state.SkipWithError("CPU does not support " #name); \ + } \ + Bench(state, [](uint32_t init_sum, unsigned char *dst, \ + const uint8_t *buf, size_t len) -> uint32_t { \ + memcpy(dst, buf, (size_t)len); \ + return hashfunc(init_sum, buf, len); \ + }, 0); \ + } \ + BENCHMARK_REGISTER_F(crc32_copy, MEMCPY_NAME(name))->Arg(32)->Arg(512)->Arg(8<<10)->Arg(32<<10)->Arg(64<<10); + +#define MEMCPY_ALIGNED_NAME(name) name##_memcpy_aligned +#define BENCHMARK_CRC32_MEMCPY_ALIGNED(name, hashfunc, support_flag) \ + BENCHMARK_DEFINE_F(crc32_copy, MEMCPY_ALIGNED_NAME(name))(benchmark::State& state) { \ + if (!(support_flag)) { \ + state.SkipWithError("CPU does not support " #name); \ + } \ + Bench(state, [](uint32_t init_sum, unsigned char *dst, \ + const uint8_t *buf, size_t len) -> uint32_t { \ + memcpy(dst, buf, (size_t)len); \ + return hashfunc(init_sum, buf, len); \ + }, 1); \ + } \ + BENCHMARK_REGISTER_F(crc32_copy, MEMCPY_ALIGNED_NAME(name))->Arg(32)->Arg(512)->Arg(8<<10)->Arg(32<<10)->Arg(64<<10); +#endif + + +// Queue both misaligned and aligned for each benchmark +#define BENCHMARK_CRC32_COPY_ONLY(name, copyfunc, support_flag) \ + BENCHMARK_CRC32_COPY_MISALIGNED(name, copyfunc, support_flag); \ + BENCHMARK_CRC32_COPY_ALIGNED(name, copyfunc, support_flag); + +// Optionally also benchmark using memcpy with normal hash function for baseline +#ifdef HASH_BASELINE +#define BENCHMARK_CRC32_COPY(name, hashfunc, copyfunc, support_flag) \ + BENCHMARK_CRC32_COPY_MISALIGNED(name, copyfunc, support_flag); \ + BENCHMARK_CRC32_COPY_ALIGNED(name, copyfunc, support_flag); \ + BENCHMARK_CRC32_MEMCPY_MISALIGNED(name, copyfunc, support_flag); \ + BENCHMARK_CRC32_MEMCPY_ALIGNED(name, copyfunc, support_flag); +#else +#define BENCHMARK_CRC32_COPY(name, hashfunc, copyfunc, support_flag) \ + BENCHMARK_CRC32_COPY_ONLY(name, copyfunc, support_flag) +#endif + +// Base test +BENCHMARK_CRC32_COPY(braid, crc32_braid, crc32_copy_braid, 1); + +#ifdef DISABLE_RUNTIME_CPU_DETECTION + // Native + BENCHMARK_CRC32_COPY(native, native_crc32, native_crc32_copy, 1) +#else + // Optimized functions +# ifndef WITHOUT_CHORBA + BENCHMARK_CRC32_COPY(chorba, crc32_chorba, crc32_copy_chorba, 1) +# endif +# ifndef WITHOUT_CHORBA_SSE +# ifdef X86_SSE2 + BENCHMARK_CRC32_COPY(chorba_sse2, crc32_chorba_sse2, crc32_copy_chorba_sse2, test_cpu_features.x86.has_sse2); +# endif +# ifdef X86_SSE41 + BENCHMARK_CRC32_COPY(chorba_sse41, crc32_chorba_sse41, crc32_copy_chorba_sse41, test_cpu_features.x86.has_sse41); +# endif +# endif +# ifdef ARM_CRC32 + BENCHMARK_CRC32_COPY(armv8, crc32_armv8, crc32_copy_armv8, test_cpu_features.arm.has_crc32) +# endif +# ifdef ARM_PMULL_EOR3 + BENCHMARK_CRC32_COPY(armv8_pmull_eor3, crc32_armv8_pmull_eor3, crc32_copy_armv8_pmull_eor3, test_cpu_features.arm.has_crc32 && test_cpu_features.arm.has_pmull && test_cpu_features.arm.has_eor3) +# endif +# ifdef LOONGARCH_CRC + BENCHMARK_CRC32_COPY(loongarch, crc32_loongarch64, crc32_copy_loongarch64, test_cpu_features.loongarch.has_crc) +# endif +# ifdef POWER8_VSX_CRC32 + BENCHMARK_CRC32_COPY(power8, crc32_power8, crc32_copy_power8, test_cpu_features.power.has_arch_2_07) +# endif +# ifdef RISCV_CRC32_ZBC + BENCHMARK_CRC32_COPY(riscv, crc32_riscv, crc32_copy_riscv64_zbc, test_cpu_features.riscv.has_zbc) +# endif +# ifdef S390_CRC32_VX + BENCHMARK_CRC32_COPY(vx, crc32_s390_vx, crc32_copy_s390_vx, test_cpu_features.s390.has_vx) +# endif +# ifdef X86_PCLMULQDQ_CRC + BENCHMARK_CRC32_COPY(pclmulqdq, crc32_pclmulqdq, crc32_copy_pclmulqdq, test_cpu_features.x86.has_pclmulqdq) +# endif +# ifdef X86_VPCLMULQDQ_AVX2 + BENCHMARK_CRC32_COPY(vpclmulqdq_avx2, crc32_vpclmulqdq_avx2, crc32_copy_vpclmulqdq_avx2, (test_cpu_features.x86.has_pclmulqdq && test_cpu_features.x86.has_avx2 && test_cpu_features.x86.has_vpclmulqdq)) +# endif +# ifdef X86_VPCLMULQDQ_AVX512 + BENCHMARK_CRC32_COPY(vpclmulqdq_avx512, crc32_vpclmulqdq_avx512, crc32_copy_vpclmulqdq_avx512, (test_cpu_features.x86.has_pclmulqdq && test_cpu_features.x86.has_avx512_common && test_cpu_features.x86.has_vpclmulqdq)) +# endif + +#endif |
