This is the mail archive of the
libc-alpha@sourceware.org
mailing list for the glibc project.
Re: [PATCH] x86_32: memcpy, mempcpy, memmove, memcpy_chk, mempcpy_chk, memmove_chk optimized with SSE2 unaligned loads/stores
- From: Allan McRae <allan at archlinux dot org>
- To: Andrew Senkevich <andrew dot n dot senkevich at gmail dot com>, GNU C Library <libc-alpha at sourceware dot org>
- Date: Sat, 05 Jul 2014 08:22:41 +1000
- Subject: Re: [PATCH] x86_32: memcpy, mempcpy, memmove, memcpy_chk, mempcpy_chk, memmove_chk optimized with SSE2 unaligned loads/stores
- Authentication-results: sourceware.org; auth=none
- References: <CAMXFM3t+TwhkeJbDXz0TSt-MZH3KOXTDWoT1nTiWjEyw1VSgcg at mail dot gmail dot com>
On 05/07/14 03:11, Andrew Senkevich wrote:
> Hi,
>
> this new functions based on new memcpy which is 32bit analogue of
> x86_64 sse2 unaligned memcpy version.
> Benchmarked on Silvermont, Haswell, Ivy Bridge, Sandy Bridge and
> Westmere, performance results attached.
> Testsuite was run on x86_64 with no new regressions.
>
Am I missing something? Why test on x86_64?
> Change log:
>
> 2014-07-04 Andrew Senkevich <andrew.n.senkevich@gmail.com>
>
> * sysdeps/i386/i686/multiarch/memcpy-sse2-unaligned.S: New file,
> contains implementation optimized with sse2 unaligned loads/stores.
> * sysdeps/i386/i686/multiarch/memmove-sse2-unaligned.S: Likewise.
> * sysdeps/i386/i686/multiarch/mempcpy-sse2-unaligned.S: Likewise.
> * sysdeps/i386/i686/multiarch/memcpy.S: Selection of new function if
> bit_Fast_Unaligned_Load is set.
> * sysdeps/i386/i686/multiarch/memcpy_chk.S: Likewise.
> * sysdeps/i386/i686/multiarch/memmove.S: Likewise.
> * sysdeps/i386/i686/multiarch/memmove_chk.S: Likewise.
> * sysdeps/i386/i686/multiarch/mempcpy.S: Likewise.
> * sysdeps/i386/i686/multiarch/mempcpy_chk.S: Likewise.
> * sysdeps/i386/i686/multiarch/Makefile: Added new files to build.
> * sysdeps/i386/i686/multiarch/ifunc-impl-list.c
> (__libc_ifunc_impl_list):
> Added testing of new functions.
>