diff options
author | Paul Pluzhnikov <ppluzhnikov@google.com> | 2023-05-30 23:02:29 +0000 |
---|---|---|
committer | Paul Pluzhnikov <ppluzhnikov@google.com> | 2023-05-30 23:02:29 +0000 |
commit | 65cc53fe7c6556b90159b8c3da1eb283792387db (patch) | |
tree | 879c71521e54f86942bf3ba2a39f9b4bc2dfb915 /sysdeps/mips/memcpy.S | |
parent | 4d0fe291aed3a476a3b59c4ecfae9d35ac0f15e8 (diff) | |
download | glibc-65cc53fe7c6556b90159b8c3da1eb283792387db.tar.gz glibc-65cc53fe7c6556b90159b8c3da1eb283792387db.tar.xz glibc-65cc53fe7c6556b90159b8c3da1eb283792387db.zip |
Fix misspellings in sysdeps/ -- BZ 25337
Diffstat (limited to 'sysdeps/mips/memcpy.S')
-rw-r--r-- | sysdeps/mips/memcpy.S | 6 |
1 files changed, 3 insertions, 3 deletions
diff --git a/sysdeps/mips/memcpy.S b/sysdeps/mips/memcpy.S index b1e1c3dc94..91a7bf0a68 100644 --- a/sysdeps/mips/memcpy.S +++ b/sysdeps/mips/memcpy.S @@ -88,7 +88,7 @@ /* * Using PREFETCH_HINT_LOAD_STREAMED instead of PREFETCH_LOAD on load - * prefetches appears to offer a slight preformance advantage. + * prefetches appear to offer a slight performance advantage. * * Using PREFETCH_HINT_PREPAREFORSTORE instead of PREFETCH_STORE * or PREFETCH_STORE_STREAMED offers a large performance advantage @@ -317,7 +317,7 @@ L(memcpy): andi a3,a3,(NSIZE-1) /* copy a3 bytes to align a0/a1 */ beq a3,zero,L(aligned) /* if a3=0, it is already aligned */ - PTR_SUBU a2,a2,a3 /* a2 is the remining bytes count */ + PTR_SUBU a2,a2,a3 /* a2 is the remaining bytes count */ C_LDHI t8,0(a1) PTR_ADDU a1,a1,a3 @@ -626,7 +626,7 @@ L(wcopy_loop): L(unaligned): andi a3,a3,(NSIZE-1) /* copy a3 bytes to align a0/a1 */ beqz a3,L(ua_chk16w) /* if a3=0, it is already aligned */ - PTR_SUBU a2,a2,a3 /* a2 is the remining bytes count */ + PTR_SUBU a2,a2,a3 /* a2 is the remaining bytes count */ C_LDHI v1,UNIT(0)(a1) C_LDLO v1,UNITM1(1)(a1) |