| /linux/arch/xtensa/mm/ |
| H A D | misc.S | 36 __loopi a2, a7, PAGE_SIZE, 32 37 s32i a3, a2, 0 38 s32i a3, a2, 4 39 s32i a3, a2, 8 40 s32i a3, a2, 12 41 s32i a3, a2, 16 42 s32i a3, a2, 20 43 s32i a3, a2, 24 44 s32i a3, a2, 28 45 __endla a2, a7, 32 [all …]
|
| /linux/arch/xtensa/lib/ |
| H A D | checksum.S | 47 extui a5, a2, 0, 2 57 add a5, a5, a2 /* a5 = end of last 32-byte chunk */ 60 l32i a6, a2, 0 61 l32i a7, a2, 4 64 l32i a6, a2, 8 65 l32i a7, a2, 12 68 l32i a6, a2, 16 69 l32i a7, a2, 20 72 l32i a6, a2, 24 73 l32i a7, a2, 28 [all …]
|
| H A D | mulsi3.S | 38 mull a2, a2, a3 41 or a4, a2, a3 44 mul16u a2, a2, a3 47 srai a4, a2, 16 50 mul16u a6, a5, a2 51 mul16u a4, a2, a3 54 add a2, a7, a4 57 mul.aa.hl a2, a3 58 mula.aa.lh a2, a3 60 umul.aa.ll a2, a3 [all …]
|
| H A D | divsi3.S | 10 quos a2, a2, a3 12 xor a7, a2, a3 /* sign = dividend ^ divisor */ 13 do_abs a6, a2, a4 /* udividend = abs (dividend) */ 16 do_nsau a5, a6, a2, a8 /* udividend_shift = nsau (udividend) */ 17 do_nsau a4, a3, a2, a8 /* udivisor_shift = nsau (udivisor) */ 23 movi a2, 0 /* quotient = 0 */ 32 addi a2, a2, 1 34 slli a2, a2, 1 43 addi a2, a2, 1 /* increment if udividend >= udivisor */ 45 neg a5, a2 [all …]
|
| H A D | udivsi3.S | 10 quou a2, a2, a3 14 mov a6, a2 /* keep dividend in a6 */ 15 do_nsau a5, a6, a2, a7 /* dividend_shift = nsau (dividend) */ 16 do_nsau a4, a3, a2, a7 /* divisor_shift = nsau (divisor) */ 22 movi a2, 0 /* quotient = 0 */ 31 addi a2, a2, 1 33 slli a2, a2, 1 42 addi a2, a2, 1 /* increment quotient if dividend >= divisor */ 53 movi a2, 1 64 movi a2, 0
|
| H A D | modsi3.S | 10 rems a2, a2, a3 12 mov a7, a2 /* save original (signed) dividend */ 13 do_abs a2, a2, a4 /* udividend = abs (dividend) */ 16 do_nsau a5, a2, a6, a8 /* udividend_shift = nsau (udividend) */ 29 bltu a2, a3, .Lzerobit 30 sub a2, a2, a3 40 bltu a2, a3, .Lreturn 41 sub a2, a2, a3 /* subtract again if udividend >= udivisor */ 44 neg a2, a2 /* if (dividend < 0), return -udividend */ 58 movi a2, 0
|
| H A D | bswapdi2.S | 10 srli a4, a2, 16 11 src a4, a4, a2 13 src a4, a2, a4 14 srli a2, a3, 16 15 src a2, a2, a3 16 src a2, a2, a2 17 src a2, a3, a2
|
| H A D | umodsi3.S | 10 remu a2, a2, a3 14 do_nsau a5, a2, a6, a7 /* dividend_shift = nsau (dividend) */ 27 bltu a2, a3, .Lzerobit 28 sub a2, a2, a3 38 bltu a2, a3, .Lreturn 39 sub a2, a2, a3 /* subtract once more if dividend >= divisor */ 53 movi a2, 0
|
| /linux/arch/csky/abiv2/ |
| H A D | strcpy.S | 14 ldw a2, (a1) 15 tstnbz a2 17 stw a2, (a3) 19 ldw a2, (a1, 4) 20 tstnbz a2 22 stw a2, (a3, 4) 24 ldw a2, (a1, 8) 25 tstnbz a2 27 stw a2, (a3, 8) 29 ldw a2, (a1, 12) [all …]
|
| H A D | strcmp.S | 10 xor a2, a3, a1 11 andi a2, 0x3 12 bnez a2, 7f 86 xtrb0 a2, t1 87 subu a0, a2 88 bez a2, 4f 93 xtrb1 a2, t1 94 subu a0, a2 95 bez a2, 4f 100 xtrb2 a2, t1 [all …]
|
| H A D | mcount.S | 31 stw a2, (sp, 8) 40 ldw a2, (sp, 8) 71 stw a2, (sp, 8) 79 ldw a2, (sp, 8) 108 lrw a2, function_trace_op 109 ldw a2, (a2, 0) 147 lrw a2, function_trace_op 148 ldw a2, (a2, 0) 170 mov a2, r8 195 lrw a2, function_trace_op [all …]
|
| /linux/arch/riscv/lib/ |
| H A D | tishift.S | 10 beqz a2, .L1 12 sub a5,a5,a2 15 sext.w a2,a2 16 srl a0,a0,a2 18 srl a2,a1,a2 20 mv a1,a2 25 li a2,0 27 mv a1,a2 33 beqz a2, .L3 35 sub a5,a5,a2 [all …]
|
| H A D | xor.S | 12 vle8.v v8, (a2) 15 add a2, a2, a3 26 vle8.v v8, (a2) 30 add a2, a2, a4 43 vle8.v v8, (a2) 47 add a2, a2, a5 63 vle8.v v8, (a2) 67 add a2, a2, a6
|
| /linux/arch/xtensa/kernel/ |
| H A D | head.S | 58 wsr a2, excsave1 89 rsr a2, excsave1 91 bltu a2, a3, 1f 92 sub a2, a2, a3 94 bgeu a2, a3, 1f 96 add a2, a2, a3 97 wsr a2, excsave1 118 movi a2, VECBASE_VADDR 119 wsr a2, vecbase 159 movi a2, XCHAL_INTTYPE_MASK_SOFTWARE | XCHAL_INTTYPE_MASK_EXTERN_EDGE [all …]
|
| H A D | entry.S | 127 s32i a1, a2, PT_AREG1 128 s32i a0, a2, PT_AREG2 129 s32i a3, a2, PT_AREG3 130 mov a1, a2 137 movi a2, 0 138 wsr a2, depc # terminate user stack trace with 0 140 xsr a2, icountlevel 142 s32i a2, a1, PT_ICOUNTLEVEL 145 rur a2, threadptr 146 s32i a2, a1, PT_THREADPTR [all …]
|
| H A D | vectors.S | 75 wsr a2, depc # save a2 76 l32i a2, a3, EXC_TABLE_KSTK # load kernel stack to a2 77 s32i a0, a2, PT_AREG0 # save a0 to ESF 79 s32i a0, a2, PT_DEPC # mark it as a regular exception 102 wsr a2, depc # save a2 103 addi a2, a1, -16 - PT_KERNEL_SIZE # adjust stack pointer 104 s32i a0, a2, PT_AREG0 # save a0 to ESF 106 s32i a0, a2, PT_DEPC # mark it as a regular exception 213 s32i a2, a3, EXC_TABLE_DOUBLE_SAVE 217 rsr a2, ps [all …]
|
| H A D | coprocessor.S | 52 xchal_cp##x##_store a2 a3 a4 a5 a6; \ 60 xchal_cp##x##_load a2 a3 a4 a5 a6; \ 124 s32i a3, a2, PT_AREG3 132 GET_THREAD_INFO (a0, a2) 146 l32i a3, a2, PT_AREG3 156 s32i a1, a2, PT_AREG1 157 s32i a3, a2, PT_SAR 158 mov a1, a2 159 rsr a2, depc 160 s32i a2, a1, PT_AREG2 [all …]
|
| H A D | align.S | 309 s32i a3, a2, PT_AREG0; _j .Lexit; .align 8 311 s32i a3, a2, PT_AREG2; _j .Lexit; .align 8 312 s32i a3, a2, PT_AREG3; _j .Lexit; .align 8 313 s32i a3, a2, PT_AREG4; _j .Lexit; .align 8 314 s32i a3, a2, PT_AREG5; _j .Lexit; .align 8 315 s32i a3, a2, PT_AREG6; _j .Lexit; .align 8 316 s32i a3, a2, PT_AREG7; _j .Lexit; .align 8 317 s32i a3, a2, PT_AREG8; _j .Lexit; .align 8 328 l32i a3, a2, PT_AREG0; _j .Lstore_w; .align 8 330 l32i a3, a2, PT_AREG2; _j .Lstore_w; .align 8 [all …]
|
| /linux/arch/loongarch/kvm/ |
| H A D | switch.S | 56 ld.d t0, a2, KVM_ARCH_GEENTRY 60 ld.d t0, a2, KVM_ARCH_GPC 64 ld.d t0, a2, KVM_ARCH_KVMPGD 87 kvm_restore_guest_gprs a2 89 ld.d a2, a2, (KVM_ARCH_GGPR + 8 * REG_A2) 108 csrwr a2, KVM_TEMP_KS 109 csrrd a2, KVM_VCPU_KS 110 addi.d a2, a2, KVM_VCPU_ARCH 113 kvm_save_guest_gprs a2 116 st.d t0, a2, (KVM_ARCH_GGPR + 8 * REG_A2) [all …]
|
| /linux/arch/loongarch/lib/ |
| H A D | unaligned.S | 27 beqz a2, 5f 30 LONG_ADDI t0, a2, -1 43 PTR_ADDI a2, a2, -1 44 bgtz a2, 2b 47 move a0, a2 66 beqz a2, 3f 72 PTR_ADDI a2, a2, -1 74 bgtz a2, 1b 76 move a0, a2
|
| H A D | memset.S | 46 beqz a2, 2f 50 addi.d a2, a2, -1 51 bgt a2, zero, 1b 69 sltui t0, a2, 9 72 add.d a2, a0, a2 79 addi.d a4, a2, -64 97 addi.d a4, a2, -32 106 addi.d a4, a2, -16 113 addi.d a4, a2, -8 118 st.d a1, a2, -8 [all …]
|
| H A D | memmove.S | 46 beqz a2, 2f 48 add.d a0, a0, a2 49 add.d a1, a1, a2 55 addi.d a2, a2, -1 56 bgt a2, zero, 1b 71 sltui t0, a2, 9 74 add.d a3, a1, a2 75 add.d a2, a0, a2 80 andi t1, a2, 7 82 sub.d a5, a2, t1 [all …]
|
| /linux/arch/mips/lib/ |
| H A D | memset.S | 95 sltiu t0, a2, STORSIZE /* very small region? */ 126 PTR_ADDU a2, t0 /* correct size */ 136 PTR_ADDU a2, t0 /* correct size */ 154 1: ori t1, a2, 0x3f /* # of full blocks */ 156 andi t0, a2, 0x40-STORSIZE 184 2: andi a2, STORMASK /* At most one long to go */ 187 beqz a2, 1f 189 PTR_ADDU a0, a2 /* What's left */ 198 PTR_SUBU t0, $0, a2 200 move a2, zero /* No remaining longs */ [all …]
|
| /linux/arch/riscv/kernel/ |
| H A D | head.S | 79 la a2, _start 80 sub a1, a1, a2 84 la a2, 1f 85 add a2, a2, a1 86 csrw CSR_TVEC, a2 89 srl a2, a0, PAGE_SHIFT 93 or a2, a2, a1 122 csrw CSR_SATP, a2 155 li a2, SBI_HART_BOOT_TASK_PTR_OFFSET 156 XIP_FIXUP_OFFSET a2 [all …]
|
| /linux/arch/mips/kernel/ |
| H A D | linux32.c | 55 unsigned long, __dummy, unsigned long, a2, unsigned long, a3) 57 return ksys_truncate(path, merge_64(a2, a3)); 61 unsigned long, a2, unsigned long, a3) 63 return ksys_ftruncate(fd, merge_64(a2, a3)); 103 asmlinkage ssize_t sys32_readahead(int fd, u32 pad0, u64 a2, u64 a3, in sys32_readahead() argument 106 return ksys_readahead(fd, merge_64(a2, a3), count); in sys32_readahead() 110 unsigned long a2, unsigned long a3, in sys32_sync_file_range() argument 115 merge_64(a2, a3), merge_64(a4, a5), in sys32_sync_file_range() 120 unsigned long a2, unsigned long a3, in sys32_fadvise64_64() argument 125 merge_64(a2, a3), merge_64(a4, a5), in sys32_fadvise64_64()
|