Welcome to mirror list, hosted at ThFree Co, Russian Federation.

cygwin.com/git/newlib-cygwin.git - Unnamed repository; edit this file 'description' to name the repository.
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
Diffstat (limited to 'newlib/libc/machine/sh/memcpy.S')
-rw-r--r--newlib/libc/machine/sh/memcpy.S378
1 files changed, 102 insertions, 276 deletions
diff --git a/newlib/libc/machine/sh/memcpy.S b/newlib/libc/machine/sh/memcpy.S
index 56e6c6f43..4df72e371 100644
--- a/newlib/libc/machine/sh/memcpy.S
+++ b/newlib/libc/machine/sh/memcpy.S
@@ -2,15 +2,14 @@
! Fast SH memcpy
!
! by Toshiyasu Morita (tm@netcom.com)
-! hacked by J"orn Rernnecke (joern.rennecke@superh.com) ("o for o-umlaut)
-! SH5 code Copyright 2002 SuperH Ltd.
+! hacked by J"orn Rernnecke (amylaar@cygnus.co.uk) ("o for o-umlaut)
!
-! Entry: ARG0: destination pointer
-! ARG1: source pointer
-! ARG3: byte count
+! Entry: r4: destination pointer
+! r5: source pointer
+! r6: byte count
!
-! Exit: RESULT: destination pointer
-! any other registers in the range r0-r7: trashed
+! Exit: r0: destination pointer
+! r1-r7: trashed
!
! Notes: Usually one wants to do small reads and write a longword, but
! unfortunately it is difficult in some cases to concatanate bytes
@@ -24,372 +23,199 @@
! be copied. This could be easily swapped for a signed comparison,
! but the algorithm used needs some comparison.
!
-! 2.: When there are two or three bytes in the last word of an 11-or-more
+! 2.: When there are two or three bytes in the last word of an 11-or-bore
! bytes memory chunk to b copied, the rest of the word can be read
-! without side effects.
+! without size effects.
! This could be easily changed by increasing the minumum size of
! a fast memcpy and the amount subtracted from r7 before L_2l_loop be 2,
! however, this would cost a few extra cyles on average.
-! For SHmedia, the assumption is that any quadword can be read in its
-! enirety if at least one byte is included in the copy.
!
#include "asm.h"
ENTRY(memcpy)
-
-#if __SHMEDIA__
-
-#define LDUAQ(P,O,D0,D1) ldlo.q P,O,D0; ldhi.q P,O+7,D1
-#define STUAQ(P,O,D0,D1) stlo.q P,O,D0; sthi.q P,O+7,D1
-#define LDUAL(P,O,D0,D1) ldlo.l P,O,D0; ldhi.l P,O+3,D1
-#define STUAL(P,O,D0,D1) stlo.l P,O,D0; sthi.l P,O+3,D1
-
- ld.b r3,0,r63
- pta/l Large,tr0
- movi 25,r0
- bgeu/u r4,r0,tr0
- nsb r4,r0
- shlli r0,5,r0
- movi (L1-L0+63*32 + 1) & 0xffff,r1
- sub r1, r0, r0
-L0: ptrel r0,tr0
- add r2,r4,r5
- ptabs r18,tr1
- add r3,r4,r6
- blink tr0,r63
-
- .balign 8
-L1:
- /* 0 byte memcpy */
- blink tr1,r63
-
-L4_7: /* 4..7 byte memcpy cntd. */
- stlo.l r2, 0, r0
- or r6, r7, r6
- sthi.l r5, -1, r6
- stlo.l r5, -4, r6
- blink tr1,r63
-
-L2_3: /* 2 or 3 byte memcpy cntd. */
- st.b r5,-1,r6
- blink tr1,r63
-
- /* 1 byte memcpy */
- ld.b r3,0,r0
- st.b r2,0,r0
- blink tr1,r63
-
-L8_15: /* 8..15 byte memcpy cntd. */
- stlo.q r2, 0, r0
- or r6, r7, r6
- sthi.q r5, -1, r6
- stlo.q r5, -8, r6
- blink tr1,r63
-
- /* 2 or 3 byte memcpy */
- ld.b r3,0,r0
- ld.b r2,0,r63
- ld.b r3,1,r1
- st.b r2,0,r0
- pta/l L2_3,tr0
- ld.b r6,-1,r6
- st.b r2,1,r1
- blink tr0, r63
-
- /* 4 .. 7 byte memcpy */
- LDUAL (r3, 0, r0, r1)
- pta L4_7, tr0
- ldlo.l r6, -4, r7
- or r0, r1, r0
- sthi.l r2, 3, r0
- ldhi.l r6, -1, r6
- blink tr0, r63
-
- /* 8 .. 15 byte memcpy */
- LDUAQ (r3, 0, r0, r1)
- pta L8_15, tr0
- ldlo.q r6, -8, r7
- or r0, r1, r0
- sthi.q r2, 7, r0
- ldhi.q r6, -1, r6
- blink tr0, r63
-
- /* 16 .. 24 byte memcpy */
- LDUAQ (r3, 0, r0, r1)
- LDUAQ (r3, 8, r8, r9)
- or r0, r1, r0
- sthi.q r2, 7, r0
- or r8, r9, r8
- sthi.q r2, 15, r8
- ldlo.q r6, -8, r7
- ldhi.q r6, -1, r6
- stlo.q r2, 8, r8
- stlo.q r2, 0, r0
- or r6, r7, r6
- sthi.q r5, -1, r6
- stlo.q r5, -8, r6
- blink tr1,r63
-
-Large:
- ld.b r2, 0, r63
- pta/l Loop_ua, tr1
- ori r3, -8, r7
- sub r2, r7, r22
- sub r3, r2, r6
- add r2, r4, r5
- ldlo.q r3, 0, r0
- addi r5, -16, r5
- movi 64+8, r27 // could subtract r7 from that.
- stlo.q r2, 0, r0
- sthi.q r2, 7, r0
- ldx.q r22, r6, r0
- bgtu/l r27, r4, tr1
-
- addi r5, -48, r27
- pta/l Loop_line, tr0
- addi r6, 64, r36
- addi r6, -24, r19
- addi r6, -16, r20
- addi r6, -8, r21
-
-Loop_line:
- ldx.q r22, r36, r63
- alloco r22, 32
- addi r22, 32, r22
- ldx.q r22, r19, r23
- sthi.q r22, -25, r0
- ldx.q r22, r20, r24
- ldx.q r22, r21, r25
- stlo.q r22, -32, r0
- ldx.q r22, r6, r0
- sthi.q r22, -17, r23
- sthi.q r22, -9, r24
- sthi.q r22, -1, r25
- stlo.q r22, -24, r23
- stlo.q r22, -16, r24
- stlo.q r22, -8, r25
- bgeu r27, r22, tr0
-
-Loop_ua:
- addi r22, 8, r22
- sthi.q r22, -1, r0
- stlo.q r22, -8, r0
- ldx.q r22, r6, r0
- bgtu/l r5, r22, tr1
-
- add r3, r4, r7
- ldlo.q r7, -8, r1
- sthi.q r22, 7, r0
- ldhi.q r7, -1, r7
- ptabs r18,tr1
- stlo.q r22, 0, r0
- or r1, r7, r1
- sthi.q r5, 15, r1
- stlo.q r5, 8, r1
- blink tr1, r63
-
-#else /* ! SHMEDIA, i.e. SH1 .. SH4 / SHcompact */
-
-#ifdef __SH5__
-#define DST r2
-#define SRC r3
-#define COUNT r4
-#define TMP0 r5
-#define TMP1 r6
-#define RESULT r2
-#else
-#define DST r4
-#define SRC r5
-#define COUNT r6
-#define TMP0 r2
-#define TMP1 r3
-#define RESULT r0
-#endif
-
#ifdef __LITTLE_ENDIAN__
! Little endian version copies with increasing addresses.
- mov DST,TMP1 ! Save return value
+ mov r4,r3 ! Save return value
mov #11,r0 ! Check if small number of bytes
- cmp/hs r0,COUNT
- ! COUNT becomes src end address
- SL(bf, L_small, add SRC,COUNT)
+ cmp/hs r0,r6
+ ! r6 becomes src end address
+ SL(bf, L_small, add r5,r6)
mov #1,r1
- tst r1,SRC ! check if source even
- SL(bt, L_even, mov COUNT,r7)
- mov.b @SRC+,r0 ! no, make it even.
- mov.b r0,@DST
- add #1,DST
-L_even: tst r1,DST ! check if destination is even
+ tst r1,r5 ! check if source even
+ SL(bt, L_even, mov r6,r7)
+ mov.b @r5+,r0 ! no, make it even.
+ mov.b r0,@r4
+ add #1,r4
+L_even: tst r1,r4 ! check if destination is even
add #-3,r7
SL(bf, L_odddst, mov #2,r1)
- tst r1,DST ! check if destination is 4-byte aligned
- mov DST,r0
- SL(bt, L_al4dst, sub SRC,r0)
- mov.w @SRC+,TMP0
- mov.w TMP0,@DST
- ! add #2,DST DST is dead here.
+ tst r1,r4 ! check if destination is 4-byte aligned
+ mov r4,r0
+ SL(bt, L_al4dst, sub r5,r0)
+ mov.w @r5+,r2
+ mov.w r2,@r4
+ ! add #2,r4 r4 is dead here.
L_al4dst:
- tst r1,SRC
+ tst r1,r5
bt L_al4both
- mov.w @SRC+,r1
+ mov.w @r5+,r1
swap.w r1,r1
add #-6,r0
add #-6,r7 ! r7 := src end address minus 9.
.align 2
L_2l_loop:
- mov.l @SRC+,TMP0 ! Read & write two longwords per iteration
- xtrct TMP0,r1
- mov.l r1,@(r0,SRC)
- cmp/hs r7,SRC
- mov.l @SRC+,r1
- xtrct r1,TMP0
- mov.l TMP0,@(r0,SRC)
+ mov.l @r5+,r2 ! Read & write two longwords per iteration
+ xtrct r2,r1
+ mov.l r1,@(r0,r5)
+ cmp/hs r7,r5
+ mov.l @r5+,r1
+ xtrct r1,r2
+ mov.l r2,@(r0,r5)
bf L_2l_loop
- add #-2,SRC
+ add #-2,r5
bra L_cleanup
add #5,r0
L_al4both:
add #-4,r0
.align 2
L_al4both_loop:
- mov.l @SRC+,DST ! Read longword, write longword per iteration
- cmp/hs r7,SRC
- SL(bf, L_al4both_loop, mov.l DST,@(r0,SRC))
+ mov.l @r5+,r4 ! Read longword, write longword per iteration
+ cmp/hs r7,r5
+ SL(bf, L_al4both_loop, mov.l r4,@(r0,r5))
bra L_cleanup
add #3,r0
L_odddst:
- tst r1,SRC
- SL(bt, L_al4src, add #-1,DST)
- mov.w @SRC+,r0
- mov.b r0,@(1,DST)
+ tst r1,r5
+ SL(bt, L_al4src, add #-1,r4)
+ mov.w @r5+,r0
+ mov.b r0,@(1,r4)
shlr8 r0
- mov.b r0,@(2,DST)
- add #2,DST
+ mov.b r0,@(2,r4)
+ add #2,r4
L_al4src:
.align 2
L_odd_loop:
- mov.l @SRC+,r0 ! Read longword, write byte, word, byte per iteration
- cmp/hs r7,SRC
- mov.b r0,@(1,DST)
+ mov.l @r5+,r0 ! Read longword, write byte, word, byte per iteration
+ cmp/hs r7,r5
+ mov.b r0,@(1,r4)
shlr8 r0
- mov.w r0,@(2,DST)
+ mov.w r0,@(2,r4)
shlr16 r0
- mov.b r0,@(4,DST)
- SL(bf, L_odd_loop, add #4,DST)
+ mov.b r0,@(4,r4)
+ SL(bf, L_odd_loop, add #4,r4)
.align 2 ! avoid nop in more frequently executed code.
L_cleanup2:
- mov DST,r0
- sub SRC,r0
+ mov r4,r0
+ sub r5,r0
L_cleanup:
- cmp/eq COUNT,SRC
+ cmp/eq r6,r5
bt L_ready
.align 2
L_cleanup_loop:
- mov.b @SRC+,r1
- cmp/eq COUNT,SRC
- mov.b r1,@(r0,SRC)
+ mov.b @r5+,r1
+ cmp/eq r6,r5
+ mov.b r1,@(r0,r5)
bf L_cleanup_loop
L_ready:
rts
- mov TMP1,RESULT
+ mov r3,r0
L_small:
bra L_cleanup2
- add #-1,DST
-#else /* ! __LITTLE_ENDIAN__ */
+ add #-1,r4
+#else
! Big endian version copies with decreasing addresses.
- mov DST,r0
- add COUNT,r0
- sub DST,SRC
+ mov r4,r0
+ add r6,r0
+ sub r4,r5
mov #11,r1
- cmp/hs r1,COUNT
- SL(bf, L_small, add #-1,SRC)
- mov SRC,TMP1
- add r0,TMP1
- shlr TMP1
+ cmp/hs r1,r6
+ SL(bf, L_small, add #-1,r5)
+ mov r5,r3
+ add r0,r3
+ shlr r3
SL(bt, L_even,
- mov DST,r7)
- mov.b @(r0,SRC),TMP0
- add #-1,TMP1
- mov.b TMP0,@-r0
+ mov r4,r7)
+ mov.b @(r0,r5),r2
+ add #-1,r3
+ mov.b r2,@-r0
L_even:
tst #1,r0
- add #-1,SRC
+ add #-1,r5
SL(bf, L_odddst, add #8,r7)
tst #2,r0
bt L_al4dst
- add #-1,TMP1
- mov.w @(r0,SRC),r1
+ add #-1,r3
+ mov.w @(r0,r5),r1
mov.w r1,@-r0
L_al4dst:
- shlr TMP1
+ shlr r3
bt L_al4both
- mov.w @(r0,SRC),r1
+ mov.w @(r0,r5),r1
swap.w r1,r1
add #4,r7
- add #-4,SRC
+ add #-4,r5
.align 2
L_2l_loop:
- mov.l @(r0,SRC),TMP0
- xtrct TMP0,r1
+ mov.l @(r0,r5),r2
+ xtrct r2,r1
mov.l r1,@-r0
cmp/hs r7,r0
- mov.l @(r0,SRC),r1
- xtrct r1,TMP0
- mov.l TMP0,@-r0
+ mov.l @(r0,r5),r1
+ xtrct r1,r2
+ mov.l r2,@-r0
bt L_2l_loop
bra L_cleanup
- add #5,SRC
+ add #5,r5
nop ! avoid nop in executed code.
L_al4both:
- add #-2,SRC
+ add #-2,r5
.align 2
L_al4both_loop:
- mov.l @(r0,SRC),r1
+ mov.l @(r0,r5),r1
cmp/hs r7,r0
SL(bt, L_al4both_loop,
mov.l r1,@-r0)
bra L_cleanup
- add #3,SRC
+ add #3,r5
nop ! avoid nop in executed code.
L_odddst:
- shlr TMP1
+ shlr r3
bt L_al4src
- mov.w @(r0,SRC),r1
+ mov.w @(r0,r5),r1
mov.b r1,@-r0
shlr8 r1
mov.b r1,@-r0
L_al4src:
- add #-2,SRC
+ add #-2,r5
.align 2
L_odd_loop:
- mov.l @(r0,SRC),TMP0
+ mov.l @(r0,r5),r2
cmp/hs r7,r0
- mov.b TMP0,@-r0
- shlr8 TMP0
- mov.w TMP0,@-r0
- shlr16 TMP0
- mov.b TMP0,@-r0
+ mov.b r2,@-r0
+ shlr8 r2
+ mov.w r2,@-r0
+ shlr16 r2
+ mov.b r2,@-r0
bt L_odd_loop
- add #3,SRC
+ add #3,r5
L_cleanup:
L_small:
- cmp/eq DST,r0
+ cmp/eq r4,r0
bt L_ready
- add #1,DST
+ add #1,r4
.align 2
L_cleanup_loop:
- mov.b @(r0,SRC),TMP0
- cmp/eq DST,r0
- mov.b TMP0,@-r0
+ mov.b @(r0,r5),r2
+ cmp/eq r4,r0
+ mov.b r2,@-r0
bf L_cleanup_loop
L_ready:
rts
- mov r0,RESULT
-#endif /* ! __LITTLE_ENDIAN__ */
-#endif /* ! SHMEDIA */
+ nop
+#endif