[PATCH v1] x86: Move wcslen SSE2 implementation to multiarch/wcslen-sse2.S
Noah Goldstein
goldstein.w.n@gmail.com
Tue Jul 12 19:29:08 GMT 2022
This commit doesn't affect libc.so.6, its just housekeeping to prepare
for adding explicit ISA level support.
Tested build on x86_64 and x86_32 with/without multiarch.
---
sysdeps/x86_64/multiarch/wcslen-sse2.S | 221 ++++++++++++++++++++++++-
sysdeps/x86_64/wcslen.S | 216 +-----------------------
2 files changed, 218 insertions(+), 219 deletions(-)
diff --git a/sysdeps/x86_64/multiarch/wcslen-sse2.S b/sysdeps/x86_64/multiarch/wcslen-sse2.S
index 2b3a9efd64..944c3bd9c6 100644
--- a/sysdeps/x86_64/multiarch/wcslen-sse2.S
+++ b/sysdeps/x86_64/multiarch/wcslen-sse2.S
@@ -17,10 +17,221 @@
<https://www.gnu.org/licenses/>. */
#if IS_IN (libc)
-# define __wcslen __wcslen_sse2
-
-# undef weak_alias
-# define weak_alias(__wcslen, wcslen)
+# ifndef WCSLEN
+# define WCSLEN __wcslen_sse2
+# endif
#endif
-#include "../wcslen.S"
+#include <sysdep.h>
+
+ .text
+ENTRY (WCSLEN)
+ cmpl $0, (%rdi)
+ jz L(exit_tail0)
+ cmpl $0, 4(%rdi)
+ jz L(exit_tail1)
+ cmpl $0, 8(%rdi)
+ jz L(exit_tail2)
+ cmpl $0, 12(%rdi)
+ jz L(exit_tail3)
+ cmpl $0, 16(%rdi)
+ jz L(exit_tail4)
+ cmpl $0, 20(%rdi)
+ jz L(exit_tail5)
+ cmpl $0, 24(%rdi)
+ jz L(exit_tail6)
+ cmpl $0, 28(%rdi)
+ jz L(exit_tail7)
+
+ pxor %xmm0, %xmm0
+
+ lea 32(%rdi), %rax
+ addq $16, %rdi
+ and $-16, %rax
+
+ pcmpeqd (%rax), %xmm0
+ pmovmskb %xmm0, %edx
+ pxor %xmm1, %xmm1
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm1
+ pmovmskb %xmm1, %edx
+ pxor %xmm2, %xmm2
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm2
+ pmovmskb %xmm2, %edx
+ pxor %xmm3, %xmm3
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm3
+ pmovmskb %xmm3, %edx
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm0
+ pmovmskb %xmm0, %edx
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm1
+ pmovmskb %xmm1, %edx
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm2
+ pmovmskb %xmm2, %edx
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm3
+ pmovmskb %xmm3, %edx
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm0
+ pmovmskb %xmm0, %edx
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm1
+ pmovmskb %xmm1, %edx
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm2
+ pmovmskb %xmm2, %edx
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd (%rax), %xmm3
+ pmovmskb %xmm3, %edx
+ addq $16, %rax
+ test %edx, %edx
+ jnz L(exit)
+
+ and $-0x40, %rax
+
+ .p2align 4
+L(aligned_64_loop):
+ movaps (%rax), %xmm0
+ movaps 16(%rax), %xmm1
+ movaps 32(%rax), %xmm2
+ movaps 48(%rax), %xmm6
+
+ pminub %xmm1, %xmm0
+ pminub %xmm6, %xmm2
+ pminub %xmm0, %xmm2
+ pcmpeqd %xmm3, %xmm2
+ pmovmskb %xmm2, %edx
+ addq $64, %rax
+ test %edx, %edx
+ jz L(aligned_64_loop)
+
+ pcmpeqd -64(%rax), %xmm3
+ pmovmskb %xmm3, %edx
+ addq $48, %rdi
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd %xmm1, %xmm3
+ pmovmskb %xmm3, %edx
+ addq $-16, %rdi
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd -32(%rax), %xmm3
+ pmovmskb %xmm3, %edx
+ addq $-16, %rdi
+ test %edx, %edx
+ jnz L(exit)
+
+ pcmpeqd %xmm6, %xmm3
+ pmovmskb %xmm3, %edx
+ addq $-16, %rdi
+ test %edx, %edx
+ jz L(aligned_64_loop)
+
+ .p2align 4
+L(exit):
+ sub %rdi, %rax
+ shr $2, %rax
+ test %dl, %dl
+ jz L(exit_high)
+
+ andl $15, %edx
+ jz L(exit_1)
+ ret
+
+ /* No align here. Naturally aligned % 16 == 1. */
+L(exit_high):
+ andl $(15 << 8), %edx
+ jz L(exit_3)
+ add $2, %rax
+ ret
+
+ .p2align 3
+L(exit_1):
+ add $1, %rax
+ ret
+
+ .p2align 3
+L(exit_3):
+ add $3, %rax
+ ret
+
+ .p2align 3
+L(exit_tail0):
+ xorl %eax, %eax
+ ret
+
+ .p2align 3
+L(exit_tail1):
+ movl $1, %eax
+ ret
+
+ .p2align 3
+L(exit_tail2):
+ movl $2, %eax
+ ret
+
+ .p2align 3
+L(exit_tail3):
+ movl $3, %eax
+ ret
+
+ .p2align 3
+L(exit_tail4):
+ movl $4, %eax
+ ret
+
+ .p2align 3
+L(exit_tail5):
+ movl $5, %eax
+ ret
+
+ .p2align 3
+L(exit_tail6):
+ movl $6, %eax
+ ret
+
+ .p2align 3
+L(exit_tail7):
+ movl $7, %eax
+ ret
+
+END (WCSLEN)
diff --git a/sysdeps/x86_64/wcslen.S b/sysdeps/x86_64/wcslen.S
index d641141d75..588a0fbe01 100644
--- a/sysdeps/x86_64/wcslen.S
+++ b/sysdeps/x86_64/wcslen.S
@@ -16,218 +16,6 @@
License along with the GNU C Library; if not, see
<https://www.gnu.org/licenses/>. */
-#include <sysdep.h>
-
- .text
-ENTRY (__wcslen)
- cmpl $0, (%rdi)
- jz L(exit_tail0)
- cmpl $0, 4(%rdi)
- jz L(exit_tail1)
- cmpl $0, 8(%rdi)
- jz L(exit_tail2)
- cmpl $0, 12(%rdi)
- jz L(exit_tail3)
- cmpl $0, 16(%rdi)
- jz L(exit_tail4)
- cmpl $0, 20(%rdi)
- jz L(exit_tail5)
- cmpl $0, 24(%rdi)
- jz L(exit_tail6)
- cmpl $0, 28(%rdi)
- jz L(exit_tail7)
-
- pxor %xmm0, %xmm0
-
- lea 32(%rdi), %rax
- addq $16, %rdi
- and $-16, %rax
-
- pcmpeqd (%rax), %xmm0
- pmovmskb %xmm0, %edx
- pxor %xmm1, %xmm1
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm1
- pmovmskb %xmm1, %edx
- pxor %xmm2, %xmm2
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm2
- pmovmskb %xmm2, %edx
- pxor %xmm3, %xmm3
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm3
- pmovmskb %xmm3, %edx
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm0
- pmovmskb %xmm0, %edx
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm1
- pmovmskb %xmm1, %edx
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm2
- pmovmskb %xmm2, %edx
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm3
- pmovmskb %xmm3, %edx
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm0
- pmovmskb %xmm0, %edx
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm1
- pmovmskb %xmm1, %edx
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm2
- pmovmskb %xmm2, %edx
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd (%rax), %xmm3
- pmovmskb %xmm3, %edx
- addq $16, %rax
- test %edx, %edx
- jnz L(exit)
-
- and $-0x40, %rax
-
- .p2align 4
-L(aligned_64_loop):
- movaps (%rax), %xmm0
- movaps 16(%rax), %xmm1
- movaps 32(%rax), %xmm2
- movaps 48(%rax), %xmm6
-
- pminub %xmm1, %xmm0
- pminub %xmm6, %xmm2
- pminub %xmm0, %xmm2
- pcmpeqd %xmm3, %xmm2
- pmovmskb %xmm2, %edx
- addq $64, %rax
- test %edx, %edx
- jz L(aligned_64_loop)
-
- pcmpeqd -64(%rax), %xmm3
- pmovmskb %xmm3, %edx
- addq $48, %rdi
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd %xmm1, %xmm3
- pmovmskb %xmm3, %edx
- addq $-16, %rdi
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd -32(%rax), %xmm3
- pmovmskb %xmm3, %edx
- addq $-16, %rdi
- test %edx, %edx
- jnz L(exit)
-
- pcmpeqd %xmm6, %xmm3
- pmovmskb %xmm3, %edx
- addq $-16, %rdi
- test %edx, %edx
- jz L(aligned_64_loop)
-
- .p2align 4
-L(exit):
- sub %rdi, %rax
- shr $2, %rax
- test %dl, %dl
- jz L(exit_high)
-
- andl $15, %edx
- jz L(exit_1)
- ret
-
- /* No align here. Naturally aligned % 16 == 1. */
-L(exit_high):
- andl $(15 << 8), %edx
- jz L(exit_3)
- add $2, %rax
- ret
-
- .p2align 3
-L(exit_1):
- add $1, %rax
- ret
-
- .p2align 3
-L(exit_3):
- add $3, %rax
- ret
-
- .p2align 3
-L(exit_tail0):
- xorl %eax, %eax
- ret
-
- .p2align 3
-L(exit_tail1):
- movl $1, %eax
- ret
-
- .p2align 3
-L(exit_tail2):
- movl $2, %eax
- ret
-
- .p2align 3
-L(exit_tail3):
- movl $3, %eax
- ret
-
- .p2align 3
-L(exit_tail4):
- movl $4, %eax
- ret
-
- .p2align 3
-L(exit_tail5):
- movl $5, %eax
- ret
-
- .p2align 3
-L(exit_tail6):
- movl $6, %eax
- ret
-
- .p2align 3
-L(exit_tail7):
- movl $7, %eax
- ret
-
-END (__wcslen)
-
+#define WCSLEN __wcslen
+#include "multiarch/wcslen-sse2.S"
weak_alias(__wcslen, wcslen)
--
2.34.1
More information about the Libc-alpha
mailing list