[libc-commits] [libc] [libc] Optimize mempcy size thresholds (PR #70049)
via libc-commits
libc-commits at lists.llvm.org
Tue Oct 24 07:28:31 PDT 2023
llvmbot wrote:
<!--LLVM PR SUMMARY COMMENT-->
@llvm/pr-subscribers-libc
Author: Dmitry Vyukov (dvyukov)
<details>
<summary>Changes</summary>
Adjust boundary conditions for sizes = 16/32/64.
See the added comment for explanations.
Results on a machine with AVX2, so sizes 64/128 affected:
```
│ baseline │ pow2-sizes │
│ sec/op │ sec/op vs base │
memcpy/Google_B 3.775n ± 1% 3.739n ± 1% -0.96% (p=0.000 n=100)
memcpy/Google_D 11.32n ± 1% 11.21n ± 1% -0.97% (p=0.000 n=100)
memcpy/Google_A 5.450n ± 1% 5.428n ± 1% ~ (p=0.058 n=100)
memcpy/Google_U 3.856n ± 1% 3.826n ± 1% -0.79% (p=0.015 n=100)
memcpy/64 3.544n ± 0% 3.258n ± 0% -8.09% (n=100)
memcpy/128 4.747n ± 0% 3.556n ± 0% -25.10% (n=100)
```
---
Full diff: https://github.com/llvm/llvm-project/pull/70049.diff
1 Files Affected:
- (modified) libc/src/string/memory_utils/x86_64/inline_memcpy.h (+25-7)
``````````diff
diff --git a/libc/src/string/memory_utils/x86_64/inline_memcpy.h b/libc/src/string/memory_utils/x86_64/inline_memcpy.h
index f43230ffd8ad125..81c995fff51d6d6 100644
--- a/libc/src/string/memory_utils/x86_64/inline_memcpy.h
+++ b/libc/src/string/memory_utils/x86_64/inline_memcpy.h
@@ -55,7 +55,7 @@ LIBC_INLINE_VAR constexpr size_t kRepMovsbThreshold =
[[maybe_unused]] LIBC_INLINE void
inline_memcpy_x86_sse2_ge64(Ptr __restrict dst, CPtr __restrict src,
size_t count) {
- if (count < 128)
+ if (count <= 128)
return builtin::Memcpy<64>::head_tail(dst, src, count);
builtin::Memcpy<32>::block(dst, src);
align_to_next_boundary<32, Arg::Dst>(dst, src, count);
@@ -65,7 +65,7 @@ inline_memcpy_x86_sse2_ge64(Ptr __restrict dst, CPtr __restrict src,
[[maybe_unused]] LIBC_INLINE void
inline_memcpy_x86_avx_ge64(Ptr __restrict dst, CPtr __restrict src,
size_t count) {
- if (count < 128)
+ if (count <= 128)
return builtin::Memcpy<64>::head_tail(dst, src, count);
if (count < 256)
return builtin::Memcpy<128>::head_tail(dst, src, count);
@@ -79,7 +79,7 @@ inline_memcpy_x86_sse2_ge64_sw_prefetching(Ptr __restrict dst,
CPtr __restrict src, size_t count) {
using namespace LIBC_NAMESPACE::x86;
prefetch_to_local_cache(src + kOneCacheline);
- if (count < 128)
+ if (count <= 128)
return builtin::Memcpy<64>::head_tail(dst, src, count);
prefetch_to_local_cache(src + kTwoCachelines);
// Aligning 'dst' on a 32B boundary.
@@ -120,7 +120,7 @@ inline_memcpy_x86_avx_ge64_sw_prefetching(Ptr __restrict dst,
CPtr __restrict src, size_t count) {
using namespace LIBC_NAMESPACE::x86;
prefetch_to_local_cache(src + kOneCacheline);
- if (count < 128)
+ if (count <= 128)
return builtin::Memcpy<64>::head_tail(dst, src, count);
prefetch_to_local_cache(src + kTwoCachelines);
prefetch_to_local_cache(src + kThreeCachelines);
@@ -149,6 +149,15 @@ inline_memcpy_x86_avx_ge64_sw_prefetching(Ptr __restrict dst,
[[maybe_unused]] LIBC_INLINE void
inline_memcpy_x86(Ptr __restrict dst, CPtr __restrict src, size_t count) {
+#if defined(__AVX512F__)
+ constexpr size_t vector_size = 64;
+#elif defined(__AVX__)
+ constexpr size_t vector_size = 32;
+#elif defined(__SSE2__)
+ constexpr size_t vector_size = 16;
+#else
+ constexpr size_t vector_size = 8;
+#endif
if (count == 0)
return;
if (count == 1)
@@ -161,11 +170,20 @@ inline_memcpy_x86(Ptr __restrict dst, CPtr __restrict src, size_t count) {
return builtin::Memcpy<4>::block(dst, src);
if (count < 8)
return builtin::Memcpy<4>::head_tail(dst, src, count);
- if (count < 16)
+ // If count is equal to a power of 2, we can handle it as head-tail
+ // of both smaller size and larger size (head-tail are either
+ // non-overlapping for smaller size, or completely collapsed
+ // for larger size). It seems to be more profitable to do the copy
+ // with the larger size, if it's natively supported (e.g. doing
+ // 2 collapsed 32-byte moves for count=64 if AVX2 is supported).
+ // But it's not profitable to use larger size if it's not natively
+ // supported: we will both use more instructions and handle fewer
+ // sizes in earlier branches.
+ if (count < 16 + (vector_size <= 8))
return builtin::Memcpy<8>::head_tail(dst, src, count);
- if (count < 32)
+ if (count < 32 + (vector_size <= 16))
return builtin::Memcpy<16>::head_tail(dst, src, count);
- if (count < 64)
+ if (count < 64 + (vector_size <= 32))
return builtin::Memcpy<32>::head_tail(dst, src, count);
if constexpr (x86::kAvx) {
if constexpr (x86::kUseSoftwarePrefetching) {
``````````
</details>
https://github.com/llvm/llvm-project/pull/70049
More information about the libc-commits
mailing list