/linux/arch/powerpc/platforms/ps3/ |
H A D | hvcall.S | 20 std r0, LRSAVE(r1); \ 22 stdu r1, -STACK_FRAME_MIN_SIZE(r1); \ 25 addi r1, r1, STACK_FRAME_MIN_SIZE; \ 27 ld r0, LRSAVE(r1); \ 44 std r0, LRSAVE(r1); \ 46 std r3, -8(r1); \ 47 stdu r1, -STACK_FRAME_MIN_SIZE-8(r1); \ 52 addi r1, r1, STACK_FRAME_MIN_SIZE+8; \ 53 ld r11, -8(r1); \ 56 ld r0, LRSAVE(r1); \ [all …]
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | verifier_xdp_direct_packet_access.c | 14 r2 = *(u32*)(r1 + %[xdp_md_data]); \ in end_mangling_bad_access_1() 15 r3 = *(u32*)(r1 + %[xdp_md_data_end]); \ in end_mangling_bad_access_1() 16 r1 = r2; \ in end_mangling_bad_access_1() 17 r1 += 8; \ in end_mangling_bad_access_1() 19 if r1 > r3 goto l0_%=; \ in end_mangling_bad_access_1() 20 r0 = *(u64*)(r1 - 8); \ in end_mangling_bad_access_1() 35 r2 = *(u32*)(r1 + %[xdp_md_data]); \ in end_mangling_bad_access_2() 36 r3 = *(u32*)(r1 + %[xdp_md_data_end]); \ in end_mangling_bad_access_2() 37 r1 = r2; \ in end_mangling_bad_access_2() 38 r1 += 8; \ in end_mangling_bad_access_2() [all …]
|
H A D | verifier_value_ptr_arith.c | 45 __msg_unpriv("R1 tried to add from different maps, paths or scalars") 50 r0 = *(u32*)(r1 + %[__sk_buff_len]); \ in value_ptr_unknown_vs_const() 51 r1 = 0; \ in value_ptr_unknown_vs_const() 52 *(u64*)(r10 - 8) = r1; \ in value_ptr_unknown_vs_const() 56 r1 = %[map_hash_16b] ll; \ in value_ptr_unknown_vs_const() 58 l0_%=: r1 = %[map_array_48b] ll; \ in value_ptr_unknown_vs_const() 63 r1 = 6; \ in value_ptr_unknown_vs_const() 64 r1 = -r1; \ in value_ptr_unknown_vs_const() 65 r1 &= 0x7; \ in value_ptr_unknown_vs_const() 67 l3_%=: r1 = 3; \ in value_ptr_unknown_vs_const() [all …]
|
H A D | verifier_stack_ptr.c | 29 r1 = r10; \ in ptr_to_stack_store_load() 30 r1 += -10; \ in ptr_to_stack_store_load() 32 *(u64*)(r1 + 2) = r0; \ in ptr_to_stack_store_load() 33 r0 = *(u64*)(r1 + 2); \ in ptr_to_stack_store_load() 45 r1 = r10; \ in load_bad_alignment_on_off() 46 r1 += -8; \ in load_bad_alignment_on_off() 48 *(u64*)(r1 + 2) = r0; \ in load_bad_alignment_on_off() 49 r0 = *(u64*)(r1 + 2); \ in load_bad_alignment_on_off() 61 r1 = r10; \ in load_bad_alignment_on_reg() 62 r1 += -10; \ in load_bad_alignment_on_reg() [all …]
|
H A D | verifier_bounds_mix_sign_unsign.c | 24 r1 = 0; \ in signed_and_unsigned_positive_bounds() 25 *(u64*)(r10 - 8) = r1; \ in signed_and_unsigned_positive_bounds() 28 r1 = %[map_hash_8b] ll; \ in signed_and_unsigned_positive_bounds() 31 r1 = *(u64*)(r10 - 16); \ in signed_and_unsigned_positive_bounds() 33 if r2 >= r1 goto l0_%=; \ in signed_and_unsigned_positive_bounds() 34 if r1 s> 4 goto l0_%=; \ in signed_and_unsigned_positive_bounds() 35 r0 += r1; \ in signed_and_unsigned_positive_bounds() 36 r1 = 0; \ in signed_and_unsigned_positive_bounds() 37 *(u8*)(r0 + 0) = r1; \ in signed_and_unsigned_positive_bounds() 56 r1 = 0; \ in checks_mixing_signed_and_unsigned() [all …]
|
H A D | verifier_bounds.c | 22 r1 = 0; \ in bounds_map_value_variant_1() 23 *(u64*)(r10 - 8) = r1; \ in bounds_map_value_variant_1() 26 r1 = %[map_hash_8b] ll; \ in bounds_map_value_variant_1() 29 r1 = *(u8*)(r0 + 0); \ in bounds_map_value_variant_1() 30 if r1 > 0xff goto l0_%=; \ in bounds_map_value_variant_1() 33 r1 -= r3; \ in bounds_map_value_variant_1() 34 r1 >>= 56; \ in bounds_map_value_variant_1() 35 r0 += r1; \ in bounds_map_value_variant_1() 50 __msg_unpriv("R1 has unknown scalar with mixed signed bounds") 54 r1 = 0; \ in bounds_map_value_variant_2() [all …]
|
H A D | verifier_masking.c | 16 r2 -= r1; \ in test_out_of_bounds_1() 17 r2 |= r1; \ in test_out_of_bounds_1() 20 r1 &= r2; \ in test_out_of_bounds_1() 21 r0 = r1; \ in test_out_of_bounds_1() 36 r2 -= r1; \ in test_out_of_bounds_2() 37 r2 |= r1; \ in test_out_of_bounds_2() 40 r1 &= r2; \ in test_out_of_bounds_2() 41 r0 = r1; \ in test_out_of_bounds_2() 56 r2 -= r1; \ in test_out_of_bounds_3() 57 r2 |= r1; \ in test_out_of_bounds_3() [all …]
|
H A D | verifier_helper_value_access.c | 49 r1 = 0; \ in access_to_map_full_range() 50 *(u64*)(r2 + 0) = r1; \ in access_to_map_full_range() 51 r1 = %[map_hash_48b] ll; \ in access_to_map_full_range() 54 r1 = r0; \ in access_to_map_full_range() 75 r1 = 0; \ in access_to_map_partial_range() 76 *(u64*)(r2 + 0) = r1; \ in access_to_map_partial_range() 77 r1 = %[map_hash_48b] ll; \ in access_to_map_partial_range() 80 r1 = r0; \ in access_to_map_partial_range() 105 r1 = 0; \ in access_to_map_empty_range() 106 *(u64*)(r2 + 0) = r1; \ in access_to_map_empty_range() [all …]
|
H A D | verifier_spin_lock.c | 27 r1 = 0; \ in spin_lock_test1_success() 28 *(u32*)(r10 - 4) = r1; \ in spin_lock_test1_success() 31 r1 = %[map_spin_lock] ll; \ in spin_lock_test1_success() 36 r1 = r0; \ in spin_lock_test1_success() 37 r1 += 4; \ in spin_lock_test1_success() 39 r1 = r6; \ in spin_lock_test1_success() 40 r1 += 4; \ in spin_lock_test1_success() 60 r1 = 0; \ in lock_test2_direct_ld_st() 61 *(u32*)(r10 - 4) = r1; \ in lock_test2_direct_ld_st() 64 r1 in lock_test2_direct_ld_st() [all...] |
H A D | verifier_array_access.c | 45 r1 = 0; \ in an_array_with_a_constant_1() 46 *(u64*)(r10 - 8) = r1; \ in an_array_with_a_constant_1() 49 r1 = %[map_hash_48b] ll; \ in an_array_with_a_constant_1() 52 r1 = %[test_val_foo]; \ in an_array_with_a_constant_1() 53 *(u64*)(r0 + 0) = r1; \ in an_array_with_a_constant_1() 69 r1 = 0; \ in __flag() 70 *(u64*)(r10 - 8) = r1; \ in __flag() 73 r1 = %[map_hash_48b] ll; \ in __flag() 76 r1 = 4; \ in __flag() 77 r1 << in __flag() [all...] |
H A D | verifier_bpf_fastcall.c | 23 "r1 = 1;" in simple() 28 "*(u64 *)(r10 - 16) = r1;" in simple() 38 "r1 = *(u64 *)(r10 - 16);" in simple() 64 __xlated("0: r1 = 1") 71 "r1 = 1;" in canary_arm64_riscv64() 72 "*(u64 *)(r10 - 16) = r1;" in canary_arm64_riscv64() 74 "r1 = *(u64 *)(r10 - 16);" in canary_arm64_riscv64() 100 __xlated("1: *(u64 *)(r10 -16) = r1") 109 "r1 = 1;" in wrong_reg_in_pattern1() 110 "*(u64 *)(r10 - 16) = r1;" in wrong_reg_in_pattern1() [all …]
|
H A D | verifier_value.c | 24 __failure __msg("R1 !read_ok") 25 __failure_unpriv __msg_unpriv("R1 !read_ok") 31 r1 = 0; \ in store_of_cleared_call_register() 32 *(u64*)(r2 + 0) = r1; \ in store_of_cleared_call_register() 33 r1 = %[map_hash_48b] ll; \ in store_of_cleared_call_register() 36 *(u64*)(r0 + 0) = r1; \ in store_of_cleared_call_register() 53 r1 = 0; \ in __flag() 54 *(u64*)(r2 + 0) = r1; \ in __flag() 55 r1 = %[map_hash_48b] ll; \ in __flag() 59 r1 = 42; \ in __flag() [all …]
|
/linux/drivers/memory/ |
H A D | ti-emif-sram-pm.S | 50 ldr r1, [r0, #EMIF_SDRAM_CONFIG] 51 str r1, [r2, #EMIF_SDCFG_VAL_OFFSET] 53 ldr r1, [r0, #EMIF_SDRAM_REFRESH_CONTROL] 54 str r1, [r2, #EMIF_REF_CTRL_VAL_OFFSET] 56 ldr r1, [r0, #EMIF_SDRAM_TIMING_1] 57 str r1, [r2, #EMIF_TIMING1_VAL_OFFSET] 59 ldr r1, [r0, #EMIF_SDRAM_TIMING_2] 60 str r1, [r2, #EMIF_TIMING2_VAL_OFFSET] 62 ldr r1, [r0, #EMIF_SDRAM_TIMING_3] 63 str r1, [r2, #EMIF_TIMING3_VAL_OFFSET] [all …]
|
/linux/tools/testing/selftests/powerpc/pmu/ebb/ |
H A D | busy_loop.S | 11 stdu %r1, -168(%r1) 12 std r14, 160(%r1) 13 std r15, 152(%r1) 14 std r16, 144(%r1) 15 std r17, 136(%r1) 16 std r18, 128(%r1) 17 std r19, 120(%r1) 18 std r20, 112(%r1) 19 std r21, 104(%r1) 20 std r22, 96(%r1) [all …]
|
/linux/arch/microblaze/kernel/ |
H A D | entry.S | 180 swi r2, r1, PT_R2; /* Save SDA */ \ 181 swi r3, r1, PT_R3; \ 182 swi r4, r1, PT_R4; \ 183 swi r5, r1, PT_R5; \ 184 swi r6, r1, PT_R6; \ 185 swi r7, r1, PT_R7; \ 186 swi r8, r1, PT_R8; \ 187 swi r9, r1, PT_R9; \ 188 swi r10, r1, PT_R10; \ 189 swi r11, r1, PT_R11; /* save clobbered regs after rval */\ [all …]
|
H A D | mcount.S | 18 addik r1, r1, -120; \ 19 swi r2, r1, 4; \ 20 swi r3, r1, 8; \ 21 swi r4, r1, 12; \ 22 swi r5, r1, 116; \ 23 swi r6, r1, 16; \ 24 swi r7, r1, 20; \ 25 swi r8, r1, 24; \ 26 swi r9, r1, 28; \ 27 swi r10, r1, 32; \ [all …]
|
/linux/crypto/ |
H A D | serpent_generic.c | 228 static noinline void __serpent_setkey_sbox(u32 r0, u32 r1, u32 r2, in __serpent_setkey_sbox() argument 232 S3(r3, r4, r0, r1, r2); store_and_load_keys(r1, r2, r4, r3, 28, 24); in __serpent_setkey_sbox() 233 S4(r1, r2, r4, r3, r0); store_and_load_keys(r2, r4, r3, r0, 24, 20); in __serpent_setkey_sbox() 234 S5(r2, r4, r3, r0, r1); store_and_load_keys(r1, r2, r4, r0, 20, 16); in __serpent_setkey_sbox() 235 S6(r1, r2, r4, r0, r3); store_and_load_keys(r4, r3, r2, r0, 16, 12); in __serpent_setkey_sbox() 236 S7(r4, r3, r2, r0, r1); store_and_load_keys(r1, r2, r0, r4, 12, 8); in __serpent_setkey_sbox() 237 S0(r1, r2, r0, r4, r3); store_and_load_keys(r0, r2, r4, r1, 8, 4); in __serpent_setkey_sbox() 238 S1(r0, r2, r4, r1, r3); store_and_load_keys(r3, r4, r1, r0, 4, 0); in __serpent_setkey_sbox() 239 S2(r3, r4, r1, r0, r2); store_and_load_keys(r2, r4, r3, r0, 0, -4); in __serpent_setkey_sbox() 240 S3(r2, r4, r3, r0, r1); store_and_load_keys(r0, r1, r4, r2, -4, -8); in __serpent_setkey_sbox() [all …]
|
/linux/arch/arm/mach-tegra/ |
H A D | sleep-tegra30.S | 209 cpu_to_csr_reg r1, r3 210 add r1, r1, r12 @ virtual CSR address for this CPU 227 str r12, [r1] 235 ldr r3, [r1] @ read CSR 236 str r3, [r1] @ clear CSR 308 mov32 r1, tegra30_iram_start 309 sub r0, r0, r1 310 mov32 r1, TEGRA_IRAM_LPx_RESUME_AREA 311 add r0, r0, r1 372 mov r1, #(1 << 28) [all …]
|
/linux/arch/powerpc/kernel/ |
H A D | interrupt_64.S | 22 ld r12,_NIP(r1) 28 ld r12,_MSR(r1) 33 ld r12,_NIP(r1) 39 ld r12,_MSR(r1) 52 mr r10,r1 53 ld r1,PACAKSAVE(r13) 54 std r10,0(r1) 55 std r11,_LINK(r1) 56 std r11,_NIP(r1) /* Saved LR is also the next instruction */ 57 std r12,_MSR(r1) [all …]
|
H A D | idle_book3s.S | 52 std r1,PACAR1(r13) 60 std r2,-8*1(r1) 61 std r14,-8*2(r1) 62 std r15,-8*3(r1) 63 std r16,-8*4(r1) 64 std r17,-8*5(r1) 65 std r18,-8*6(r1) 66 std r19,-8*7(r1) 67 std r20,-8*8(r1) 68 std r21,-8*9(r1) [all …]
|
H A D | entry_32.S | 106 stw r3, ORIG_GPR3(r1) 107 stw r11, GPR1(r1) 108 stw r11, 0(r1) 110 stw r12, _LINK(r1) 115 SAVE_GPR(2, r1) 117 stw r9,_MSR(r1) 119 stw r12,STACK_INT_FRAME_MARKER(r1) 120 stw r2,_TRAP(r1) 121 SAVE_GPR(0, r1) 122 SAVE_GPRS(3, 8, r1) [all …]
|
/linux/arch/powerpc/kernel/vdso/ |
H A D | vgetrandom-chacha.S | 127 std counter, -216(r1) 129 std r14, -144(r1) 130 std r15, -136(r1) 131 std r16, -128(r1) 132 std r17, -120(r1) 133 std r18, -112(r1) 134 std r19, -104(r1) 135 std r20, -96(r1) 136 std r21, -88(r1) 137 std r22, -80(r1) [all …]
|
/linux/arch/arc/lib/ |
H A D | strlen.S | 15 asl_s r1,r0,3 17 asl r7,r4,r1 19 sub r1,r2,r7 20 bic_s r1,r1,r2 24 or.eq r12,r12,r1 30 mov_s r1,31 31 sub3 r7,r1,r0 32 sub r1,r2,r4 33 bic_s r1,r1,r2 34 bmsk r1,r1,r7 [all …]
|
/linux/arch/nios2/boot/compressed/ |
H A D | head.S | 25 movia r1, NIOS2_ICACHE_SIZE 27 1: initi r1 28 sub r1, r1, r2 29 bgt r1, r0, 1b 31 movia r1, NIOS2_DCACHE_SIZE 33 1: initd 0(r1) 34 sub r1, r1, r2 35 bgt r1, r0, 1b 37 nextpc r1 /* Find out where we are */ 40 beq r1, r2, finish_move /* We are running in correct address, [all …]
|
/linux/arch/openrisc/kernel/ |
H A D | entry.S | 46 l.sw -8(r1),r2 /* store frame pointer */ ;\ 47 l.sw -4(r1),r9 /* store return address */ ;\ 48 l.addi r2,r1,0 /* move sp to fp */ ;\ 50 l.addi r1,r1,-8 ;\ 51 l.ori r1,r2,0 /* restore sp */ ;\ 52 l.lwz r9,-4(r1) /* restore return address */ ;\ 53 l.lwz r2,-8(r1) /* restore fp */ ;\ 59 l.sw -12(r1),t1 /* save extra reg */ ;\ 60 l.sw -8(r1),r2 /* store frame pointer */ ;\ 61 l.sw -4(r1),r9 /* store return address */ ;\ [all …]
|