|
@@ -74,7 +74,7 @@
|
|
|
.align 5
|
|
|
LEAF(memset)
|
|
|
beqz a1, 1f
|
|
|
- move v0, a0 /* result */
|
|
|
+ move v0, a0 /* result */
|
|
|
|
|
|
andi a1, 0xff /* spread fillword */
|
|
|
LONG_SLL t1, a1, 8
|
|
@@ -90,7 +90,7 @@ LEAF(memset)
|
|
|
FEXPORT(__bzero)
|
|
|
sltiu t0, a2, STORSIZE /* very small region? */
|
|
|
bnez t0, .Lsmall_memset
|
|
|
- andi t0, a0, STORMASK /* aligned? */
|
|
|
+ andi t0, a0, STORMASK /* aligned? */
|
|
|
|
|
|
#ifdef CONFIG_CPU_MICROMIPS
|
|
|
move t8, a1 /* used by 'swp' instruction */
|
|
@@ -98,12 +98,12 @@ FEXPORT(__bzero)
|
|
|
#endif
|
|
|
#ifndef CONFIG_CPU_DADDI_WORKAROUNDS
|
|
|
beqz t0, 1f
|
|
|
- PTR_SUBU t0, STORSIZE /* alignment in bytes */
|
|
|
+ PTR_SUBU t0, STORSIZE /* alignment in bytes */
|
|
|
#else
|
|
|
.set noat
|
|
|
li AT, STORSIZE
|
|
|
beqz t0, 1f
|
|
|
- PTR_SUBU t0, AT /* alignment in bytes */
|
|
|
+ PTR_SUBU t0, AT /* alignment in bytes */
|
|
|
.set at
|
|
|
#endif
|
|
|
|
|
@@ -120,7 +120,7 @@ FEXPORT(__bzero)
|
|
|
1: ori t1, a2, 0x3f /* # of full blocks */
|
|
|
xori t1, 0x3f
|
|
|
beqz t1, .Lmemset_partial /* no block to fill */
|
|
|
- andi t0, a2, 0x40-STORSIZE
|
|
|
+ andi t0, a2, 0x40-STORSIZE
|
|
|
|
|
|
PTR_ADDU t1, a0 /* end address */
|
|
|
.set reorder
|
|
@@ -145,7 +145,7 @@ FEXPORT(__bzero)
|
|
|
.set at
|
|
|
#endif
|
|
|
jr t1
|
|
|
- PTR_ADDU a0, t0 /* dest ptr */
|
|
|
+ PTR_ADDU a0, t0 /* dest ptr */
|
|
|
|
|
|
.set push
|
|
|
.set noreorder
|
|
@@ -155,7 +155,7 @@ FEXPORT(__bzero)
|
|
|
andi a2, STORMASK /* At most one long to go */
|
|
|
|
|
|
beqz a2, 1f
|
|
|
- PTR_ADDU a0, a2 /* What's left */
|
|
|
+ PTR_ADDU a0, a2 /* What's left */
|
|
|
R10KCBARRIER(0(ra))
|
|
|
#ifdef __MIPSEB__
|
|
|
EX(LONG_S_R, a1, -1(a0), .Llast_fixup)
|
|
@@ -164,24 +164,24 @@ FEXPORT(__bzero)
|
|
|
EX(LONG_S_L, a1, -1(a0), .Llast_fixup)
|
|
|
#endif
|
|
|
1: jr ra
|
|
|
- move a2, zero
|
|
|
+ move a2, zero
|
|
|
|
|
|
.Lsmall_memset:
|
|
|
beqz a2, 2f
|
|
|
- PTR_ADDU t1, a0, a2
|
|
|
+ PTR_ADDU t1, a0, a2
|
|
|
|
|
|
1: PTR_ADDIU a0, 1 /* fill bytewise */
|
|
|
R10KCBARRIER(0(ra))
|
|
|
bne t1, a0, 1b
|
|
|
- sb a1, -1(a0)
|
|
|
+ sb a1, -1(a0)
|
|
|
|
|
|
2: jr ra /* done */
|
|
|
- move a2, zero
|
|
|
+ move a2, zero
|
|
|
END(memset)
|
|
|
|
|
|
.Lfirst_fixup:
|
|
|
jr ra
|
|
|
- nop
|
|
|
+ nop
|
|
|
|
|
|
.Lfwd_fixup:
|
|
|
PTR_L t0, TI_TASK($28)
|
|
@@ -189,7 +189,7 @@ FEXPORT(__bzero)
|
|
|
LONG_L t0, THREAD_BUADDR(t0)
|
|
|
LONG_ADDU a2, t1
|
|
|
jr ra
|
|
|
- LONG_SUBU a2, t0
|
|
|
+ LONG_SUBU a2, t0
|
|
|
|
|
|
.Lpartial_fixup:
|
|
|
PTR_L t0, TI_TASK($28)
|
|
@@ -197,8 +197,8 @@ FEXPORT(__bzero)
|
|
|
LONG_L t0, THREAD_BUADDR(t0)
|
|
|
LONG_ADDU a2, t1
|
|
|
jr ra
|
|
|
- LONG_SUBU a2, t0
|
|
|
+ LONG_SUBU a2, t0
|
|
|
|
|
|
.Llast_fixup:
|
|
|
jr ra
|
|
|
- andi v1, a2, STORMASK
|
|
|
+ andi v1, a2, STORMASK
|