about summary refs log tree commit diff
path: root/sysdeps/powerpc
diff options
context:
space:
mode:
Diffstat (limited to 'sysdeps/powerpc')
-rw-r--r--sysdeps/powerpc/powerpc32/cell/memcpy.S24
-rw-r--r--sysdeps/powerpc/powerpc64/cell/memcpy.S24
2 files changed, 24 insertions, 24 deletions
diff --git a/sysdeps/powerpc/powerpc32/cell/memcpy.S b/sysdeps/powerpc/powerpc32/cell/memcpy.S
index e6c076cbe1..cc1da99fd9 100644
--- a/sysdeps/powerpc/powerpc32/cell/memcpy.S
+++ b/sysdeps/powerpc/powerpc32/cell/memcpy.S
@@ -43,16 +43,16 @@
 .align  7
 
 EALIGN (BP_SYM (memcpy), 5, 0)
-        CALL_MCOUNT
+	CALL_MCOUNT
 
 	dcbt	0,r4		/* Prefetch ONE SRC cacheline  */
 	cmplwi	cr1,r5,16	/* is size < 16 ?  */
-	mr	r6,r3		
+	mr	r6,r3
 	blt+	cr1,.Lshortcopy
 
 .Lbigcopy:
 	neg	r8,r3		/* LS 3 bits = # bytes to 8-byte dest bdry  */
-        clrlwi  r8,r8,32-4	/* aling to 16byte boundary  */
+	clrlwi  r8,r8,32-4	/* aling to 16byte boundary  */
 	sub     r7,r4,r3
 	cmplwi	cr0,r8,0
 	beq+	.Ldst_aligned
@@ -112,8 +112,8 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 
 .LprefetchSRC:
 	dcbt    r12,r4
-        addi    r12,r12,128
-        bdnz    .LprefetchSRC
+	addi    r12,r12,128
+	bdnz    .LprefetchSRC
 
 .Lnocacheprefetch:
 	mtctr	r7
@@ -122,7 +122,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 	beq	cr6,.Lcachelinealigned
 
 .Laligntocacheline:
-	lfd 	fp9,0x08(r4)
+	lfd	fp9,0x08(r4)
 	lfdu	fp10,0x10(r4)
 	stfd	fp9,0x08(r6)
 	stfdu	fp10,0x10(r6)
@@ -131,10 +131,10 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 
 .Lcachelinealigned:		/* copy while cache lines  */
 
-	blt- 	cr1,.Llessthancacheline	/* size <128  */
+	blt-	cr1,.Llessthancacheline	/* size <128  */
 
 .Louterloop:
-        cmpwi   r11,0
+	cmpwi   r11,0
 	mtctr	r11
 	beq-	.Lendloop
 
@@ -142,7 +142,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 
 .align	4
 	/* Copy whole cachelines, optimized by prefetching SRC cacheline  */
-.Lloop: 			/* Copy aligned body  */
+.Lloop:				/* Copy aligned body  */
 	dcbt	r12,r4		/* PREFETCH SOURCE some cache lines ahead  */
 	lfd	fp9, 0x08(r4)
 	dcbz	r11,r6
@@ -186,7 +186,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 	beq-	.Lendloop2
 	mtctr	r10
 
-.Lloop2: 			/* Copy aligned body  */
+.Lloop2:			/* Copy aligned body  */
 	lfd	fp9, 0x08(r4)
 	lfd	fp10, 0x10(r4)
 	lfd	fp11, 0x18(r4)
@@ -206,7 +206,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 	mtctr	r7
 
 .Lcopy_remaining:
-	lfd 	fp9,0x08(r4)
+	lfd	fp9,0x08(r4)
 	lfdu	fp10,0x10(r4)
 	stfd	fp9,0x08(r6)
 	stfdu	fp10,0x10(r6)
@@ -214,7 +214,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 
 .Ldo_lt16:			/* less than 16 ?  */
 	cmplwi	cr0,r5,0	/* copy remaining bytes (0-15)  */
-	beqlr+			/* no rest to copy  */	
+	beqlr+			/* no rest to copy  */
 	addi	r4,r4,8
 	addi	r6,r6,8
 
diff --git a/sysdeps/powerpc/powerpc64/cell/memcpy.S b/sysdeps/powerpc/powerpc64/cell/memcpy.S
index 2a00a6ed52..c6ee730e4e 100644
--- a/sysdeps/powerpc/powerpc64/cell/memcpy.S
+++ b/sysdeps/powerpc/powerpc64/cell/memcpy.S
@@ -43,16 +43,16 @@
 .align  7
 
 EALIGN (BP_SYM (memcpy), 5, 0)
-        CALL_MCOUNT 3
+	CALL_MCOUNT 3
 
 	dcbt	0,r4		/* Prefetch ONE SRC cacheline  */
 	cmpldi	cr1,r5,16	/* is size < 16 ?  */
-	mr	r6,r3		
+	mr	r6,r3
 	blt+	cr1,.Lshortcopy
 
 .Lbigcopy:
 	neg	r8,r3		/* LS 3 bits = # bytes to 8-byte dest bdry  */
-        clrldi  r8,r8,64-4	/* aling to 16byte boundary  */
+	clrldi  r8,r8,64-4	/* aling to 16byte boundary  */
 	sub     r7,r4,r3
 	cmpldi	cr0,r8,0
 	beq+	.Ldst_aligned
@@ -112,8 +112,8 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 
 .LprefetchSRC:
 	dcbt    r12,r4
-        addi    r12,r12,128
-        bdnz    .LprefetchSRC
+	addi    r12,r12,128
+	bdnz    .LprefetchSRC
 
 .Lnocacheprefetch:
 	mtctr	r7
@@ -122,7 +122,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 	beq	cr6,.Lcachelinealigned
 
 .Laligntocacheline:
-	ld 	r9,0x08(r4)
+	ld	r9,0x08(r4)
 	ldu	r7,0x10(r4)
 	std	r9,0x08(r6)
 	stdu	r7,0x10(r6)
@@ -131,10 +131,10 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 
 .Lcachelinealigned:		/* copy while cache lines  */
 
-	blt- 	cr1,.Llessthancacheline	/* size <128  */
+	blt-	cr1,.Llessthancacheline	/* size <128  */
 
 .Louterloop:
-        cmpdi   r11,0
+	cmpdi   r11,0
 	mtctr	r11
 	beq-	.Lendloop
 
@@ -142,7 +142,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 
 .align	4
 	/* Copy whole cachelines, optimized by prefetching SRC cacheline  */
-.Lloop: 			/* Copy aligned body  */
+.Lloop:				/* Copy aligned body  */
 	dcbt	r12,r4		/* PREFETCH SOURCE some cache lines ahead  */
 	ld	r9, 0x08(r4)
 	dcbz	r11,r6
@@ -186,7 +186,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 	beq-	.Lendloop2
 	mtctr	r10
 
-.Lloop2: 			/* Copy aligned body  */
+.Lloop2:			/* Copy aligned body  */
 	ld	r9, 0x08(r4)
 	ld	r7, 0x10(r4)
 	ld	r8, 0x18(r4)
@@ -206,7 +206,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 	mtctr	r7
 
 .Lcopy_remaining:
-	ld 	r8,0x08(r4)
+	ld	r8,0x08(r4)
 	ldu	r7,0x10(r4)
 	std	r8,0x08(r6)
 	stdu	r7,0x10(r6)
@@ -214,7 +214,7 @@ EALIGN (BP_SYM (memcpy), 5, 0)
 
 .Ldo_lt16:			/* less than 16 ?  */
 	cmpldi	cr0,r5,0	/* copy remaining bytes (0-15)  */
-	beqlr+			/* no rest to copy  */	
+	beqlr+			/* no rest to copy  */
 	addi	r4,r4,8
 	addi	r6,r6,8