memset.S
上传用户:lgb322
上传日期:2013-02-24
资源大小:30529k
文件大小:4k
- /* linux/arch/sparc/lib/memset.S: Sparc optimized memset, bzero and clear_user code
- * Copyright (C) 1991,1996 Free Software Foundation
- * Copyright (C) 1996,1997 Jakub Jelinek (jj@sunsite.mff.cuni.cz)
- * Copyright (C) 1996 David S. Miller (davem@caip.rutgers.edu)
- *
- * Returns 0, if ok, and number of bytes not yet set if exception
- * occurs and we were called as clear_user.
- */
- #include <asm/cprefix.h>
- #include <asm/ptrace.h>
- #define EX(x,y,a,b,z)
- 98: x,y;
- .section .fixup,z##alloc,z##execinstr;
- .align 4;
- 99: ba 30f;
- a, b, %o0;
- .section __ex_table,z##alloc;
- .align 4;
- .word 98b, 99b;
- .text;
- .align 4
- #define EXT(start,end,handler,z)
- .section __ex_table,z##alloc;
- .align 4;
- .word start, 0, end, handler;
- .text;
- .align 4
- /* Please don't change these macros, unless you change the logic
- * in the .fixup section below as well.
- * Store 64 bytes at (BASE + OFFSET) using value SOURCE. */
- #define ZERO_BIG_BLOCK(base, offset, source)
- std source, [base + offset + 0x00];
- std source, [base + offset + 0x08];
- std source, [base + offset + 0x10];
- std source, [base + offset + 0x18];
- std source, [base + offset + 0x20];
- std source, [base + offset + 0x28];
- std source, [base + offset + 0x30];
- std source, [base + offset + 0x38];
- #define ZERO_LAST_BLOCKS(base, offset, source)
- std source, [base - offset - 0x38];
- std source, [base - offset - 0x30];
- std source, [base - offset - 0x28];
- std source, [base - offset - 0x20];
- std source, [base - offset - 0x18];
- std source, [base - offset - 0x10];
- std source, [base - offset - 0x08];
- std source, [base - offset - 0x00];
- .text
- .align 4
- .globl __bzero_begin
- __bzero_begin:
- .globl C_LABEL(__bzero), C_LABEL(__memset),
- .globl C_LABEL(memset)
- .globl C_LABEL(__memset_start), C_LABEL(__memset_end)
- C_LABEL(__memset_start):
- C_LABEL(__memset):
- C_LABEL(memset):
- and %o1, 0xff, %g3
- sll %g3, 8, %g2
- or %g3, %g2, %g3
- sll %g3, 16, %g2
- or %g3, %g2, %g3
- b 1f
- mov %o2, %o1
- 3:
- cmp %o2, 3
- be 2f
- EX(stb %g3, [%o0], sub %o1, 0,#)
- cmp %o2, 2
- be 2f
- EX(stb %g3, [%o0 + 0x01], sub %o1, 1,#)
- EX(stb %g3, [%o0 + 0x02], sub %o1, 2,#)
- 2:
- sub %o2, 4, %o2
- add %o1, %o2, %o1
- b 4f
- sub %o0, %o2, %o0
- C_LABEL(__bzero):
- mov %g0, %g3
- 1:
- cmp %o1, 7
- bleu 7f
- andcc %o0, 3, %o2
- bne 3b
- 4:
- andcc %o0, 4, %g0
- be 2f
- mov %g3, %g2
- EX(st %g3, [%o0], sub %o1, 0,#)
- sub %o1, 4, %o1
- add %o0, 4, %o0
- 2:
- andcc %o1, 0xffffff80, %o3 ! Now everything is 8 aligned and o1 is len to run
- be 9f
- andcc %o1, 0x78, %o2
- 10:
- ZERO_BIG_BLOCK(%o0, 0x00, %g2)
- subcc %o3, 128, %o3
- ZERO_BIG_BLOCK(%o0, 0x40, %g2)
- 11:
- EXT(10b, 11b, 20f,#)
- bne 10b
- add %o0, 128, %o0
- orcc %o2, %g0, %g0
- 9:
- be 13f
- andcc %o1, 7, %o1
- srl %o2, 1, %o3
- set 13f, %o4
- sub %o4, %o3, %o4
- jmp %o4
- add %o0, %o2, %o0
- 12:
- ZERO_LAST_BLOCKS(%o0, 0x48, %g2)
- ZERO_LAST_BLOCKS(%o0, 0x08, %g2)
- 13:
- be 8f
- andcc %o1, 4, %g0
- be 1f
- andcc %o1, 2, %g0
- EX(st %g3, [%o0], and %o1, 7,#)
- add %o0, 4, %o0
- 1:
- be 1f
- andcc %o1, 1, %g0
- EX(sth %g3, [%o0], and %o1, 3,#)
- add %o0, 2, %o0
- 1:
- bne,a 8f
- EX(stb %g3, [%o0], and %o1, 1,#)
- 8:
- retl
- clr %o0
- 7:
- be 13b
- orcc %o1, 0, %g0
- be 0f
- 8:
- add %o0, 1, %o0
- subcc %o1, 1, %o1
- bne,a 8b
- EX(stb %g3, [%o0 - 1], add %o1, 1,#)
- 0:
- retl
- clr %o0
- C_LABEL(__memset_end):
- .section .fixup,#alloc,#execinstr
- .align 4
- 20:
- cmp %g2, 8
- bleu 1f
- and %o1, 0x7f, %o1
- sub %g2, 9, %g2
- add %o3, 64, %o3
- 1:
- sll %g2, 3, %g2
- add %o3, %o1, %o0
- b 30f
- sub %o0, %g2, %o0
- 21:
- mov 8, %o0
- and %o1, 7, %o1
- sub %o0, %g2, %o0
- sll %o0, 3, %o0
- b 30f
- add %o0, %o1, %o0
- 30:
- /* %o4 is faulting address, %o5 is %pc where fault occurred */
- save %sp, -104, %sp
- mov %i5, %o0
- mov %i7, %o1
- call C_LABEL(lookup_fault)
- mov %i4, %o2
- ret
- restore
- .globl __bzero_end
- __bzero_end: