| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | verifier_raw_stack.c | 11 __failure_unpriv __msg_unpriv("invalid read from stack R6 off=-8 size=8") 16 r6 = r10; \ in stack_no_skb_load_bytes() 17 r6 += -8; \ in stack_no_skb_load_bytes() 18 r3 = r6; \ in stack_no_skb_load_bytes() 21 r0 = *(u64*)(r6 + 0); \ in stack_no_skb_load_bytes() 33 r6 = r10; \ in skb_load_bytes_negative_len() 34 r6 += -8; \ in skb_load_bytes_negative_len() 35 r3 = r6; \ in skb_load_bytes_negative_len() 38 r0 = *(u64*)(r6 + 0); \ in skb_load_bytes_negative_len() 52 r6 = r10; \ in load_bytes_negative_len_2() [all …]
|
| H A D | verifier_subprog_precision.c | 48 __msg("mark_precise: frame0: regs=r1 stack= before 3: (bf) r1 = r6") 49 __msg("mark_precise: frame0: regs=r6 stack= before 2: (b7) r6 = 3") 53 "r6 = 3;" in subprog_result_precise() 54 /* pass r6 through r1 into subprog to get it back as r0; in subprog_result_precise() 57 "r1 = r6;" in subprog_result_precise() 60 * r6 -> r1 -> r0 chain), as index into vals array, forcing in subprog_result_precise() 65 /* here r0->r1->r6 chain is forced to be precise and has to be in subprog_result_precise() 74 : __clobber_common, "r6" in subprog_result_precise() 103 __msg("mark_precise: frame0: regs=r0 stack= before 5: (bf) r1 = r6") [all...] |
| H A D | pro_epilogue.c | 25 __xlated("0: r6 = *(u64 *)(r1 +0)") 26 __xlated("1: r7 = *(u64 *)(r6 +0)") 28 __xlated("3: *(u64 *)(r6 +0) = r7") 31 __xlated("5: r6 = r1") 33 __xlated("7: r1 = r6") 41 "r6 = r1;" in test_prologue() 43 "r1 = r6;" in test_prologue() 56 __xlated("2: r6 = r1") 58 __xlated("4: r1 = r6") 63 __xlated("8: r6 = *(u64 *)(r1 +0)") [all …]
|
| H A D | verifier_jeq_infer_not_null.c | 17 * r6 = skb->sk; 18 * r7 = sk_fullsock(r6); 19 * r0 = sk_fullsock(r6); 35 /* r6 = skb->sk; */ \ in socket_for_jne_false_branch() 36 r6 = *(u64*)(r1 + %[__sk_buff_sk]); \ in socket_for_jne_false_branch() 37 /* if (r6 == 0) return 0; */ \ in socket_for_jne_false_branch() 38 if r6 == 0 goto l0_%=; \ in socket_for_jne_false_branch() 40 r1 = r6; \ in socket_for_jne_false_branch() 44 r1 = r6; \ in socket_for_jne_false_branch() 71 /* r6 = skb->sk */ \ in unchanged_for_jne_true_branch() [all …]
|
| H A D | verifier_subreg.c | 55 r6 = r0; \ in add32_imm_zero_extend_check() 61 r0 |= r6; \ in add32_imm_zero_extend_check() 96 r6 = r0; \ in sub32_imm_zero_extend_check() 102 r0 |= r6; \ in sub32_imm_zero_extend_check() 137 r6 = r0; \ in mul32_imm_zero_extend_check() 143 r0 |= r6; \ in mul32_imm_zero_extend_check() 178 r6 = r0; \ in div32_imm_zero_extend_check() 184 r0 |= r6; \ in div32_imm_zero_extend_check() 219 r6 = r0; \ in or32_imm_zero_extend_check() 225 r0 |= r6; \ in or32_imm_zero_extend_check() [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | 85xx_entry_mapping.S | 5 invstr: mflr r6 /* Make it accessible */ 12 tlbsx 0,r6 /* search MSR[IS], SPID=PID0 */ 26 tlbsx 0,r6 /* search MSR[IS], SPID=PID1 */ 34 tlbsx 0,r6 /* Fall through, we had to match */ 48 li r6,0 /* Set Entry counter to 0 */ 50 rlwimi r7,r6,16,4,15 /* Setup MAS0 = TLBSEL | ESEL(r6) */ 55 cmpw r3,r6 60 skpinv: addi r6,r6,1 /* Increment */ 61 cmpw r6,r9 /* Are we done? */ 65 li r6,0x04 [all …]
|
| /linux/arch/powerpc/lib/ |
| H A D | mem_64.S | 40 mr r6,r3 45 stb r4,0(r6) 46 addi r6,r6,1 48 sth r4,0(r6) 49 addi r6,r6,2 51 stw r4,0(r6) 52 addi r6,r6,4 58 4: std r4,0(r6) 59 std r4,8(r6) 60 std r4,16(r6) [all …]
|
| H A D | checksum_64.S | 26 srdi. r6,r4,3 /* less than 8 bytes? */ 35 rldicl. r6,r3,64-1,64-2 /* r6 = (r3 >> 1) & 0x3 */ 39 sub r6,r7,r6 40 mtctr r6 43 lhz r6,0(r3) /* align to doubleword */ 46 adde r0,r0,r6 55 srdi. r6,r4,7 58 srdi r6,r4,6 59 subi r6,r6,1 60 mtctr r6 [all …]
|
| H A D | copy_32.S | 20 stw r7,4(r6); \ 21 stw r8,8(r6); \ 22 stw r9,12(r6); \ 23 stwu r10,16(r6) 35 stw r7,4(r6); \ 37 stw r8,8(r6); \ 39 stw r9,12(r6); \ 41 stwu r10,16(r6) 68 addi r6, r3, -4 72 1: stwu r4, 4(r6) [all …]
|
| /linux/tools/testing/selftests/powerpc/copyloops/ |
| H A D | mem_64.S | 40 mr r6,r3 45 stb r4,0(r6) 46 addi r6,r6,1 48 sth r4,0(r6) 49 addi r6,r6,2 51 stw r4,0(r6) 52 addi r6,r6,4 58 4: std r4,0(r6) 59 std r4,8(r6) 60 std r4,16(r6) [all …]
|
| /linux/arch/arm/mm/ |
| H A D | abort-lv4t.S | 67 and r6, r8, r7 69 add r6, r6, r9, lsr #1 71 add r6, r6, r9, lsr #2 73 add r6, r6, r9, lsr #3 74 add r6, r6, r6, lsr #8 75 add r6, r6, r6, lsr #4 76 and r6, r6, #15 @ r6 = no. of registers to transfer. 80 subne r7, r7, r6, lsl #2 @ Undo increment 81 addeq r7, r7, r6, lsl #2 @ Undo decrement 93 andne r6, r8, #0xf00 @ { immediate high nibble [all …]
|
| /linux/arch/arm/kernel/ |
| H A D | head-nommu.S | 246 * Region 2: Normal, Shared, cacheable for RAM. From PHYS_OFFSET, size from r6 249 * r6: Value to be written to DRSR (and IRSR if required) for PMSAv7_RAM_REGION 273 ldr r6, =(_end) @ Cover whole kernel 274 sub r6, r6, r5 @ Minimum size of region to map 275 clz r6, r6 @ Region size must be 2^N... 276 rsb r6, r6, #31 @ ...so round up region size 277 lsl r6, r6, #PMSAv7_RSR_SZ @ Put size in right field 278 orr r6, r6, #(1 << PMSAv7_RSR_EN) @ Set region enabled bit 288 /* Setup second region first to free up r6 */ 295 setup_region r0, r5, r6, PMSAv7_DATA_SIDE, r12 @ PHYS_OFFSET, shared, enabled [all …]
|
| /linux/arch/hexagon/include/asm/ |
| H A D | spinlock.h | 31 "1: R6 = memw_locked(%0);\n" in arch_read_lock() 32 " { P3 = cmp.ge(R6,#0); R6 = add(R6,#1);}\n" in arch_read_lock() 34 " memw_locked(%0,P3) = R6;\n" in arch_read_lock() 38 : "memory", "r6", "p3" in arch_read_lock() 46 "1: R6 = memw_locked(%0);\n" in arch_read_unlock() 47 " R6 = add(R6,#-1);\n" in arch_read_unlock() 48 " memw_locked(%0,P3) = R6\n" in arch_read_unlock() 52 : "memory", "r6", "p3" in arch_read_unlock() 62 " R6 = memw_locked(%1);\n" in arch_read_trylock() 63 " { %0 = #0; P3 = cmp.ge(R6,#0); R6 = add(R6,#1);}\n" in arch_read_trylock() [all …]
|
| /linux/arch/powerpc/platforms/83xx/ |
| H A D | suspend-asm.S | 64 lwz r6, 4(r4) 67 stw r6, SS_MEMSAVE+4(r3) 70 mfspr r6, SPRN_HID1 75 stw r6, SS_HID+4(r3) 80 mfspr r6, SPRN_IBCR 87 stw r6, SS_IBCR(r3) 94 mfspr r6, SPRN_SPRG2 100 stw r6, SS_SPRG+8(r3) 106 mfspr r6, SPRN_SPRG6 111 stw r6, SS_SPRG+24(r3) [all …]
|
| /linux/arch/nios2/mm/ |
| H A D | uaccess.c | 18 " bge r2,r6,1f\n" 23 "1: addi r6,r6,-1\n" 25 " beq r6,r2,3f\n" 28 " addi r6,r6,-1\n" 32 " bne r6,r7,2b\n" 34 " addi r2,r6,1\n" 36 "13:mov r2,r6\n" 44 " addi r6,r6,-2\n" 48 "6: bge r7,r6,1b\n" 50 " addi r6,r6,-4\n" [all …]
|
| /linux/arch/sh/lib/ |
| H A D | checksum.S | 64 addc r0, r6 66 add #1, r6 78 addc r0, r6 80 addc r2, r6 82 addc r3, r6 84 addc r0, r6 86 addc r2, r6 88 addc r3, r6 89 addc r0, r6 90 addc r2, r6 [all …]
|
| /linux/arch/powerpc/crypto/ |
| H A D | aes-spe-keys.S | 30 xor r6,r6,r6; \ 77 LOAD_KEY(r6,r4,4) 81 stw r6,4(r3) 93 xor r6,r6,r5 94 xor r7,r7,r6 97 stw r6,4(r3) 119 LOAD_KEY(r6,r4,4) 125 stw r6,4(r3) 139 xor r6,r6,r5 140 xor r7,r7,r6 [all …]
|
| /linux/drivers/scsi/arm/ |
| H A D | acornscsi-io.S | 29 ldmia r0!, {r3, r4, r5, r6} 33 orr r4, r4, r6, lsl #16 34 ldmia r0!, {r5, r6, r7, ip} 36 orr r5, r5, r6, lsl #16 37 and r6, r7, lr 38 orr r6, r6, ip, lsl #16 39 stmia r1!, {r3 - r6} 45 ldmia r0!, {r3, r4, r5, r6} 49 orr r4, r4, r6, lsl #16 76 stmfd sp!, {r4 - r6, lr} [all …]
|
| /linux/arch/arm/mach-imx/ |
| H A D | suspend-imx6.S | 81 mov r6, #0x0 82 str r6, [r11, #L2X0_CACHE_SYNC] 84 ldr r6, [r11, #L2X0_CACHE_SYNC] 85 ands r6, r6, #0x1 99 ldr r6, [r0, #PM_INFO_MMDC_IO_NUM_OFFSET] 106 subs r6, r6, #0x1 118 ldr r6, [r11, r7] 119 orr r6, r6, #(1 << 31) 120 str r6, [r11, r7] 122 ldr r6, [r11, r7] [all …]
|
| /linux/arch/arm/lib/ |
| H A D | io-readsb.S | 29 .Linsb_aligned: stmfd sp!, {r4 - r6, lr} 38 ldrb r6, [r0] 43 orr r3, r3, r6, put_byte_3 44 ldrb r6, [r0] 49 orr r4, r4, r6, put_byte_2 50 ldrb r6, [r0] 55 orr r5, r5, r6, put_byte_1 56 ldrb r6, [r0] 61 mov r6, r6, put_byte_0 62 orr r6, r6, ip, put_byte_1 [all …]
|
| H A D | csumpartialcopygeneric.S | 112 1: load4l r4, r5, r6, r7 113 stmia dst!, {r4, r5, r6, r7} 116 adcs sum, sum, r6 176 1: load4l r5, r6, r7, r8 179 orr r5, r5, r6, lspush #24 180 mov r6, r6, lspull #8 181 orr r6, r6, r7, lspush #24 184 stmia dst!, {r4, r5, r6, r7} 187 adcs sum, sum, r6 197 load2l r5, r6 [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/fuc/ |
| H A D | memx.fuc | 87 imm32($r6, 0xfffffffc) 88 and $r7 $r6 89 mov $r6 0x2 90 or $r7 $r6 93 mov $r6 0x001620 95 nv_rd32($r8, $r6) 97 nv_wr32($r6, $r8) 100 nv_rd32($r8, $r6) 102 nv_wr32($r6, $r8) 104 mov $r6 0x0026f0 [all …]
|
| /linux/arch/powerpc/boot/ |
| H A D | string.S | 26 addi r6,r3,-1 30 stbu r0,1(r6) 97 addi r6,r3,-4 100 stwu r4,4(r6) 102 andi. r0,r6,3 104 subf r6,r0,r6 108 1: stwu r4,4(r6) 114 addi r6,r6,3 115 8: stbu r4,1(r6) 128 addi r6,r3,-4 [all …]
|
| /linux/arch/powerpc/mm/nohash/ |
| H A D | tlb_low.S | 67 tlbsx. r6,0,r3 75 tlbwe r6,r6,PPC44x_TLB_PAGEID 77 oris r7,r6,0x8000 /* specify way explicitly */ 118 li r6,0 /* Default entry value 0 */ 126 tlbre r6,r5,0 /* Read entry */ 128 andi. r0,r6,PPC47x_TLB0_VALID /* Valid entry ? */ 131 rlwinm r6,r6,0,21,19 /* Clear V */ 132 tlbwe r6,r7,0 /* Write it */ 178 2: mflr r6 180 PPC_ICBT(0,R6,R7) /* touch next cache line */ [all …]
|
| /linux/arch/csky/kernel/ |
| H A D | head.S | 13 lrw r6, init_thread_union + THREAD_SIZE 14 mov sp, r6 25 lrw r6, secondary_msa1 26 ld.w r6, (r6, 0) 27 mtcr r6, cr<31, 15> 30 lrw r6, secondary_pgd 31 ld.w r6, (r6, 0) 32 mtcr r6, cr<28, 15> 33 mtcr r6, cr<29, 15> 36 lrw r6, secondary_stack [all …]
|