/* Optimized memmove implementation using LoongArch LASX instructions.
Copyright (C) 2023-2024 Free Software Foundation, Inc.
This file is part of the GNU C Library.
The GNU C Library is free software; you can redistribute it and/or
modify it under the terms of the GNU Lesser General Public
License as published by the Free Software Foundation; either
version 2.1 of the License, or (at your option) any later version.
The GNU C Library is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
License along with the GNU C Library. If not, see
. */
#include
#include
#include
#if IS_IN (libc) && !defined __loongarch_soft_float
#ifndef MEMCPY_NAME
# define MEMCPY_NAME __memcpy_lasx
#endif
#ifndef MEMMOVE_NAME
# define MEMMOVE_NAME __memmove_lasx
#endif
LEAF(MEMCPY_NAME, 6)
li.d t0, 32
add.d a3, a0, a2
add.d a4, a1, a2
bgeu t0, a2, L(less_32bytes)
li.d t1, 64
bltu t1, a2, L(copy_long)
xvld xr0, a1, 0
xvld xr1, a4, -32
xvst xr0, a0, 0
xvst xr1, a3, -32
jr ra
L(less_32bytes):
srli.d t0, a2, 4
beqz t0, L(less_16bytes)
vld vr0, a1, 0
vld vr1, a4, -16
vst vr0, a0, 0
vst vr1, a3, -16
jr ra
L(less_16bytes):
srli.d t0, a2, 3
beqz t0, L(less_8bytes)
ld.d t0, a1, 0
ld.d t1, a4, -8
st.d t0, a0, 0
st.d t1, a3, -8
jr ra
L(less_8bytes):
srli.d t0, a2, 2
beqz t0, L(less_4bytes)
ld.w t0, a1, 0
ld.w t1, a4, -4
st.w t0, a0, 0
st.w t1, a3, -4
jr ra
L(less_4bytes):
srli.d t0, a2, 1
beqz t0, L(less_2bytes)
ld.h t0, a1, 0
ld.h t1, a4, -2
st.h t0, a0, 0
st.h t1, a3, -2
jr ra
L(less_2bytes):
beqz a2, L(less_1bytes)
ld.b t0, a1, 0
st.b t0, a0, 0
L(less_1bytes):
jr ra
END(MEMCPY_NAME)
LEAF(MEMMOVE_NAME, 6)
li.d t0, 32
add.d a3, a0, a2
add.d a4, a1, a2
bgeu t0, a2, L(less_32bytes)
li.d t1, 64
bltu t1, a2, L(move_long)
xvld xr0, a1, 0
xvld xr1, a4, -32
xvst xr0, a0, 0
xvst xr1, a3, -32
jr ra
L(move_long):
sub.d t2, a0, a1
bltu t2, a2, L(copy_back)
L(copy_long):
andi t2, a0, 0x1f
addi.d a2, a2, -1
sub.d t2, t0, t2
xvld xr8, a1, 0
xvld xr9, a4, -32
sub.d t3, a2, t2
add.d a5, a0, t2
andi a2, t3, 0xff
add.d a1, a1, t2
beq a2, t3, L(lt256)
sub.d a6, a4, a2
addi.d a6, a6, -1
L(loop_256):
xvld xr0, a1, 0
xvld xr1, a1, 32
xvld xr2, a1, 64
xvld xr3, a1, 96
xvld xr4, a1, 128
xvld xr5, a1, 160
xvld xr6, a1, 192
xvld xr7, a1, 224
addi.d a1, a1, 256
xvst xr0, a5, 0
xvst xr1, a5, 32
xvst xr2, a5, 64
xvst xr3, a5, 96
xvst xr4, a5, 128
xvst xr5, a5, 160
xvst xr6, a5, 192
xvst xr7, a5, 224
addi.d a5, a5, 256
bne a1, a6, L(loop_256)
L(lt256):
srli.d t2, a2, 7
beqz t2, L(lt128)
xvld xr0, a1, 0
xvld xr1, a1, 32
xvld xr2, a1, 64
xvld xr3, a1, 96
addi.d a1, a1, 128
addi.d a2, a2, -128
xvst xr0, a5, 0
xvst xr1, a5, 32
xvst xr2, a5, 64
xvst xr3, a5, 96
addi.d a5, a5, 128
L(lt128):
bltu a2, t1, L(lt64)
xvld xr0, a1, 0
xvld xr1, a1, 32
addi.d a1, a1, 64
addi.d a2, a2, -64
xvst xr0, a5, 0
xvst xr1, a5, 32
addi.d a5, a5, 64
L(lt64):
bltu a2, t0, L(lt32)
xvld xr0, a1, 0
xvst xr0, a5, 0
L(lt32):
xvst xr8, a0, 0
xvst xr9, a3, -32
jr ra
nop
L(copy_back):
addi.d a3, a3, -1
addi.d a2, a2, -2
andi t2, a3, 0x1f
xvld xr8, a1, 0
xvld xr9, a4, -32
sub.d t3, a2, t2
sub.d a5, a3, t2
sub.d a4, a4, t2
andi a2, t3, 0xff
beq a2, t3, L(back_lt256)
add.d a6, a1, a2
addi.d a6, a6, 2
L(back_loop_256):
xvld xr0, a4, -33
xvld xr1, a4, -65
xvld xr2, a4, -97
xvld xr3, a4, -129
xvld xr4, a4, -161
xvld xr5, a4, -193
xvld xr6, a4, -225
xvld xr7, a4, -257
addi.d a4, a4, -256
xvst xr0, a5, -32
xvst xr1, a5, -64
xvst xr2, a5, -96
xvst xr3, a5, -128
xvst xr4, a5, -160
xvst xr5, a5, -192
xvst xr6, a5, -224
xvst xr7, a5, -256
addi.d a5, a5, -256
bne a4, a6, L(back_loop_256)
L(back_lt256):
srli.d t2, a2, 7
beqz t2, L(back_lt128)
xvld xr0, a4, -33
xvld xr1, a4, -65
xvld xr2, a4, -97
xvld xr3, a4, -129
addi.d a2, a2, -128
addi.d a4, a4, -128
xvst xr0, a5, -32
xvst xr1, a5, -64
xvst xr2, a5, -96
xvst xr3, a5, -128
addi.d a5, a5, -128
L(back_lt128):
blt a2, t1, L(back_lt64)
xvld xr0, a4, -33
xvld xr1, a4, -65
addi.d a2, a2, -64
addi.d a4, a4, -64
xvst xr0, a5, -32
xvst xr1, a5, -64
addi.d a5, a5, -64
L(back_lt64):
bltu a2, t0, L(back_lt32)
xvld xr0, a4, -33
xvst xr0, a5, -32
L(back_lt32):
xvst xr8, a0, 0
xvst xr9, a3, -31
jr ra
END(MEMMOVE_NAME)
libc_hidden_builtin_def (MEMCPY_NAME)
libc_hidden_builtin_def (MEMMOVE_NAME)
#endif