diff --git a/sysdeps/x86_64/multiarch/strcat-sse2-unaligned.S b/sysdeps/x86_64/multiarch/strcat-sse2-unaligned.S index 1150281fe4..c02c6f0984 100644 --- a/sysdeps/x86_64/multiarch/strcat-sse2-unaligned.S +++ b/sysdeps/x86_64/multiarch/strcat-sse2-unaligned.S @@ -52,4 +52,3 @@ L(StartStrcpyPart): # include "strcpy-sse2-unaligned.S" #endif - diff --git a/sysdeps/x86_64/multiarch/strcat-ssse3.S b/sysdeps/x86_64/multiarch/strcat-ssse3.S index 66736a7087..34b61b8468 100644 --- a/sysdeps/x86_64/multiarch/strcat-ssse3.S +++ b/sysdeps/x86_64/multiarch/strcat-ssse3.S @@ -556,4 +556,3 @@ L(StrncatExit8Bytes): # endif END (STRCAT) #endif - diff --git a/sysdeps/x86_64/multiarch/strcpy-sse2-unaligned.S b/sysdeps/x86_64/multiarch/strcpy-sse2-unaligned.S index 6de8c47ad4..e73778ae1a 100644 --- a/sysdeps/x86_64/multiarch/strcpy-sse2-unaligned.S +++ b/sysdeps/x86_64/multiarch/strcpy-sse2-unaligned.S @@ -1742,7 +1742,7 @@ L(Unaligned64LeaveCase2): # ifndef USE_AS_STRCAT jnz L(CopyFrom1To16BytesUnalignedXmm4) # else - jnz L(CopyFrom1To16Bytes) + jnz L(CopyFrom1To16Bytes) # endif pcmpeqb %xmm5, %xmm0 pmovmskb %xmm0, %rdx @@ -1754,7 +1754,7 @@ L(Unaligned64LeaveCase2): # ifndef USE_AS_STRCAT jnz L(CopyFrom1To16BytesUnalignedXmm5) # else - jnz L(CopyFrom1To16Bytes) + jnz L(CopyFrom1To16Bytes) # endif pcmpeqb %xmm6, %xmm0 @@ -1767,7 +1767,7 @@ L(Unaligned64LeaveCase2): # ifndef USE_AS_STRCAT jnz L(CopyFrom1To16BytesUnalignedXmm6) # else - jnz L(CopyFrom1To16Bytes) + jnz L(CopyFrom1To16Bytes) # endif pcmpeqb %xmm7, %xmm0 @@ -1888,4 +1888,3 @@ L(FillTable): # endif # endif #endif - diff --git a/sysdeps/x86_64/multiarch/strlen-sse2-pminub.S b/sysdeps/x86_64/multiarch/strlen-sse2-pminub.S index 57778cf2d1..6048072315 100644 --- a/sysdeps/x86_64/multiarch/strlen-sse2-pminub.S +++ b/sysdeps/x86_64/multiarch/strlen-sse2-pminub.S @@ -141,7 +141,7 @@ L(align16_start): pmovmskb %xmm3, %edx test %edx, %edx jnz L(exit64) - + test $0x3f, %rax jz L(align64_loop) @@ -183,10 +183,10 @@ L(align16_start): .p2align 4 L(align64_loop): movaps (%rax), %xmm4 - pminub 16(%rax), %xmm4 - movaps 32(%rax), %xmm5 - pminub 48(%rax), %xmm5 - add $64, %rax + pminub 16(%rax), %xmm4 + movaps 32(%rax), %xmm5 + pminub 48(%rax), %xmm5 + add $64, %rax pminub %xmm4, %xmm5 pcmpeqb %xmm0, %xmm5 pmovmskb %xmm5, %edx @@ -195,7 +195,7 @@ L(align16_start): pcmpeqb -64(%rax), %xmm0 - sub $80, %rax + sub $80, %rax pmovmskb %xmm0, %edx test %edx, %edx jnz L(exit16)