/linux/drivers/net/ethernet/brocade/bna/ |
H A D | bfa_ioc_ct.c | 49 static enum bfa_status bfa_ioc_ct_pll_init(void __iomem *rb, 51 static enum bfa_status bfa_ioc_ct2_pll_init(void __iomem *rb, 251 void __iomem *rb; in bfa_ioc_ct_reg_init() local 254 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init() 256 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init() 257 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init() 258 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init() 261 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init() 262 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init() 263 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init() [all …]
|
/linux/drivers/scsi/bfa/ |
H A D | bfa_ioc_ct.c | 185 void __iomem *rb; in bfa_ioc_ct_reg_init() local 188 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init() 190 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init() 191 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init() 192 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init() 195 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init() 196 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init() 197 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init() 198 ioc->ioc_regs.hfn_mbox_cmd = rb + ct_p0reg[pcifn].hfn; in bfa_ioc_ct_reg_init() 199 ioc->ioc_regs.lpu_mbox_cmd = rb + ct_p0reg[pcifn].lpu; in bfa_ioc_ct_reg_init() [all …]
|
H A D | bfa_ioc_cb.c | 138 void __iomem *rb; in bfa_ioc_cb_reg_init() local 141 rb = bfa_ioc_bar0(ioc); in bfa_ioc_cb_reg_init() 143 ioc->ioc_regs.hfn_mbox = rb + iocreg_fnreg[pcifn].hfn_mbox; in bfa_ioc_cb_reg_init() 144 ioc->ioc_regs.lpu_mbox = rb + iocreg_fnreg[pcifn].lpu_mbox; in bfa_ioc_cb_reg_init() 145 ioc->ioc_regs.host_page_num_fn = rb + iocreg_fnreg[pcifn].hfn_pgn; in bfa_ioc_cb_reg_init() 148 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_cb_reg_init() 149 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_cb_reg_init() 150 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_cb_reg_init() 152 ioc->ioc_regs.heartbeat = (rb + BFA_IOC1_HBEAT_REG); in bfa_ioc_cb_reg_init() 153 ioc->ioc_regs.ioc_fwstate = (rb + BFA_IOC1_STATE_REG); in bfa_ioc_cb_reg_init() [all …]
|
/linux/tools/testing/selftests/bpf/benchs/ |
H A D | run_bench_ringbufs.sh | 10 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 15 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 16 summarize $b "$($RUN_RB_BENCH --rb-sampled $b)" 20 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 21 summarize $b "$($RUN_RB_BENCH --rb-b2b $b)" 22 summarize $b-sampled "$($RUN_RB_BENCH --rb-sampled --rb-b2b $b)" 27 …summarize "rb-sampled-$b" "$($RUN_RB_BENCH --rb-b2b --rb-batch-cnt $b --rb-sampled --rb-sample-rat… 31 …summarize "pb-sampled-$b" "$($RUN_RB_BENCH --rb-b2b --rb-batch-cnt $b --rb-sampled --rb-sample-rat… 35 summarize "reserve" "$($RUN_RB_BENCH --rb-b2b rb-custom)" 36 summarize "output" "$($RUN_RB_BENCH --rb-b2b --rb-use-output rb-custom)" [all …]
|
/linux/arch/x86/crypto/ |
H A D | serpent-sse2-i586-asm_32.S | 28 #define RB %xmm1 macro 513 read_blocks(%eax, RA, RB, RC, RD, RT0, RT1, RE); 515 K(RA, RB, RC, RD, RE, 0); 516 S0(RA, RB, RC, RD, RE); LK(RC, RB, RD, RA, RE, 1); 517 S1(RC, RB, RD, RA, RE); LK(RE, RD, RA, RC, RB, 2); 518 S2(RE, RD, RA, RC, RB); LK(RB, RD, RE, RC, RA, 3); 519 S3(RB, RD, RE, RC, RA); LK(RC, RA, RD, RB, RE, 4); 520 S4(RC, RA, RD, RB, RE); LK(RA, RD, RB, RE, RC, 5); 521 S5(RA, RD, RB, RE, RC); LK(RC, RA, RD, RE, RB, 6); 522 S6(RC, RA, RD, RE, RB); LK(RD, RB, RA, RE, RC, 7); [all …]
|
H A D | serpent-sse2-x86_64-asm_64.S | 636 K2(RA, RB, RC, RD, RE, 0); 637 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1); 638 S(S1, RC, RB, RD, RA, RE); LK2(RE, RD, RA, RC, RB, 2); 639 S(S2, RE, RD, RA, RC, RB); LK2(RB, RD, RE, RC, RA, 3); 640 S(S3, RB, RD, RE, RC, RA); LK2(RC, RA, RD, RB, RE, 4); 641 S(S4, RC, RA, RD, RB, RE); LK2(RA, RD, RB, RE, RC, 5); 642 S(S5, RA, RD, RB, RE, RC); LK2(RC, RA, RD, RE, RB, 6); 643 S(S6, RC, RA, RD, RE, RB); LK2(RD, RB, RA, RE, RC, 7); 644 S(S7, RD, RB, RA, RE, RC); LK2(RC, RA, RE, RD, RB, 8); 645 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9); [all …]
|
H A D | serpent-avx2-asm_64.S | 566 K2(RA, RB, RC, RD, RE, 0); 567 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1); 568 S(S1, RC, RB, RD, RA, RE); LK2(RE, RD, RA, RC, RB, 2); 569 S(S2, RE, RD, RA, RC, RB); LK2(RB, RD, RE, RC, RA, 3); 570 S(S3, RB, RD, RE, RC, RA); LK2(RC, RA, RD, RB, RE, 4); 571 S(S4, RC, RA, RD, RB, RE); LK2(RA, RD, RB, RE, RC, 5); 572 S(S5, RA, RD, RB, RE, RC); LK2(RC, RA, RD, RE, RB, 6); 573 S(S6, RC, RA, RD, RE, RB); LK2(RD, RB, RA, RE, RC, 7); 574 S(S7, RD, RB, RA, RE, RC); LK2(RC, RA, RE, RD, RB, 8); 575 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9); [all …]
|
/linux/kernel/bpf/ |
H A D | ringbuf.c | 81 struct bpf_ringbuf *rb; member 98 struct bpf_ringbuf *rb; in bpf_ringbuf_area_alloc() local 135 rb = vmap(pages, nr_meta_pages + 2 * nr_data_pages, in bpf_ringbuf_area_alloc() 137 if (rb) { in bpf_ringbuf_area_alloc() 139 rb->pages = pages; in bpf_ringbuf_area_alloc() 140 rb->nr_pages = nr_pages; in bpf_ringbuf_area_alloc() 141 return rb; in bpf_ringbuf_area_alloc() 153 struct bpf_ringbuf *rb = container_of(work, struct bpf_ringbuf, work); in bpf_ringbuf_notify() local 155 wake_up_all(&rb->waitq); in bpf_ringbuf_notify() 171 struct bpf_ringbuf *rb; in bpf_ringbuf_alloc() local [all …]
|
/linux/fs/xfs/scrub/ |
H A D | bmap_repair.c | 99 struct xrep_bmap *rb, in xrep_bmap_discover_shared() argument 103 struct xfs_scrub *sc = rb->sc; in xrep_bmap_discover_shared() 123 rb->reflink_scan = RLS_SET_IFLAG; in xrep_bmap_discover_shared() 131 struct xrep_bmap *rb, in xrep_bmap_from_rmap() argument 143 struct xfs_scrub *sc = rb->sc; in xrep_bmap_from_rmap() 151 if (rb->reflink_scan == RLS_UNKNOWN && !unwritten) { in xrep_bmap_from_rmap() 152 error = xrep_bmap_discover_shared(rb, startblock, blockcount); in xrep_bmap_from_rmap() 163 fa = xfs_bmap_validate_extent(sc->ip, rb->whichfork, &irec); in xrep_bmap_from_rmap() 169 trace_xrep_bmap_found(sc->ip, rb->whichfork, &irec); in xrep_bmap_from_rmap() 174 error = xfarray_append(rb->bmap_records, &rbe); in xrep_bmap_from_rmap() [all …]
|
/linux/kernel/events/ |
H A D | internal.h | 63 extern void rb_free(struct perf_buffer *rb); 67 struct perf_buffer *rb; in rb_free_rcu() local 69 rb = container_of(rcu_head, struct perf_buffer, rcu_head); in rb_free_rcu() 70 rb_free(rb); in rb_free_rcu() 73 static inline void rb_toggle_paused(struct perf_buffer *rb, bool pause) in rb_toggle_paused() argument 75 if (!pause && rb->nr_pages) in rb_toggle_paused() 76 rb->paused = 0; in rb_toggle_paused() 78 rb->paused = 1; in rb_toggle_paused() 84 extern int rb_alloc_aux(struct perf_buffer *rb, struct perf_event *event, 86 extern void rb_free_aux(struct perf_buffer *rb); [all …]
|
/linux/Documentation/translations/zh_CN/core-api/ |
H A D | rbtree.rst | 271 node = rb_entry(root->rb_node, struct interval_tree_node, rb); 274 if (node->rb.rb_left) { 276 rb_entry(node->rb.rb_left, 277 struct interval_tree_node, rb); 294 if (node->rb.rb_right) { 295 node = rb_entry(node->rb.rb_right, 296 struct interval_tree_node, rb); 311 if (node->rb.rb_left) { 312 subtree_last = rb_entry(node->rb.rb_left, 313 struct interval_tree_node, rb)->__subtree_last; [all …]
|
/linux/tools/testing/selftests/powerpc/stringloops/ |
H A D | memcmp_64.S | 15 #define rB r10 macro 121 lbz rB,0(r4) 122 subf. rC,rB,rA 127 lbz rB,1(r4) 128 subf. rC,rB,rA 133 lbz rB,2(r4) 134 subf. rC,rB,rA 139 lbz rB,3(r4) 140 subf. rC,rB,rA 173 LD rB,0,r4 [all …]
|
/linux/arch/powerpc/lib/ |
H A D | memcmp_64.S | 15 #define rB r10 macro 121 lbz rB,0(r4) 122 subf. rC,rB,rA 127 lbz rB,1(r4) 128 subf. rC,rB,rA 133 lbz rB,2(r4) 134 subf. rC,rB,rA 139 lbz rB,3(r4) 140 subf. rC,rB,rA 173 LD rB,0,r4 [all …]
|
/linux/drivers/misc/mchp_pci1xxxx/ |
H A D | mchp_pci1xxxx_otpe2p.c | 99 void __iomem *rb = priv->reg_base; in is_eeprom_responsive() local 104 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive() 106 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive() 111 true, rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive() 124 void __iomem *rb = priv->reg_base; in pci1xxxx_eeprom_read() local 141 writel(EEPROM_CMD_EPC_BUSY_BIT | (off + byte), rb + in pci1xxxx_eeprom_read() 148 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in pci1xxxx_eeprom_read() 154 buf[byte] = readl(rb + MMAP_EEPROM_OFFSET(EEPROM_DATA_REG)); in pci1xxxx_eeprom_read() 165 void __iomem *rb = priv->reg_base; in pci1xxxx_eeprom_write() local 182 writel(*(value + byte), rb + MMAP_EEPROM_OFFSET(EEPROM_DATA_REG)); in pci1xxxx_eeprom_write() [all …]
|
/linux/arch/arm64/crypto/ |
H A D | sm3-neon-core.S | 42 #define rb w4 macro 356 ldp ra, rb, [RSTATE, #0] 401 R1(ra, rb, rc, rd, re, rf, rg, rh, k_even, KL, 0, 0, IW, _, 0) 402 R1(rd, ra, rb, rc, rh, re, rf, rg, k_odd, _, 1, 1, IW, _, 0) 403 R1(rc, rd, ra, rb, rg, rh, re, rf, k_even, KL, 2, 2, IW, _, 0) 404 R1(rb, rc, rd, ra, rf, rg, rh, re, k_odd, _, 3, 3, IW, _, 0) 407 R1(ra, rb, rc, rd, re, rf, rg, rh, k_even, KL, 4, 0, IW, _, 0) 408 R1(rd, ra, rb, rc, rh, re, rf, rg, k_odd, _, 5, 1, IW, _, 0) 409 R1(rc, rd, ra, rb, rg, rh, re, rf, k_even, KL, 6, 2, IW, SCHED_W_W0W1W2W3W4W5_1, 12) 410 R1(rb, rc, rd, ra, rf, rg, rh, re, k_odd, _, 7, 3, IW, SCHED_W_W0W1W2W3W4W5_2, 12) [all …]
|
/linux/arch/powerpc/crypto/ |
H A D | sha1-powerpc-asm.S | 28 #define RB(t) ((((t)+3)%6)+7) macro 40 andc r0,RD(t),RB(t); \ 41 and r6,RB(t),RC(t); \ 48 rotlwi RB(t),RB(t),30; \ 52 and r6,RB(t),RC(t); \ 53 andc r0,RD(t),RB(t); \ 55 rotlwi RB(t),RB(t),30; \ 67 xor r6,RB(t),RC(t); \ 69 rotlwi RB(t),RB(t),30; \ 77 xor r6,RB(t),RC(t); \ [all …]
|
/linux/mm/ |
H A D | interval_tree.c | 23 INTERVAL_TREE_DEFINE(struct vm_area_struct, shared.rb, 38 if (!prev->shared.rb.rb_right) { in vma_interval_tree_insert_after() 40 link = &prev->shared.rb.rb_right; in vma_interval_tree_insert_after() 42 parent = rb_entry(prev->shared.rb.rb_right, in vma_interval_tree_insert_after() 43 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after() 46 while (parent->shared.rb.rb_left) { in vma_interval_tree_insert_after() 47 parent = rb_entry(parent->shared.rb.rb_left, in vma_interval_tree_insert_after() 48 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after() 52 link = &parent->shared.rb.rb_left; in vma_interval_tree_insert_after() 56 rb_link_node(&node->shared.rb, &parent->shared.rb, link); in vma_interval_tree_insert_after() [all …]
|
/linux/drivers/target/iscsi/ |
H A D | iscsi_target_configfs.c | 44 ssize_t rb; in lio_target_np_driver_show() local 48 rb = sysfs_emit(page, "1\n"); in lio_target_np_driver_show() 50 rb = sysfs_emit(page, "0\n"); in lio_target_np_driver_show() 52 return rb; in lio_target_np_driver_show() 474 ssize_t rb; \ 479 rb = snprintf(page, PAGE_SIZE, \ 483 rb = snprintf(page, PAGE_SIZE, "%u\n", \ 488 return rb; \ 530 ssize_t rb = 0; in lio_target_nacl_info_show() local 536 rb += sysfs_emit_at(page, rb, "No active iSCSI Session for Initiator" in lio_target_nacl_info_show() [all …]
|
/linux/arch/powerpc/xmon/ |
H A D | ppc-opc.c | 538 /* The RB field in an X, XO, M, or MDS form instruction. */ 539 #define RB RAOPT + 1 macro 543 /* The RB field in an X form instruction when it must be the same as 546 #define RBS RB + 1 549 /* The RB field in an lswx instruction, which has special value 554 /* The RB field of the dccci and iccci instructions, which are optional. */ 1783 /* The RB field in an X form instruction when it must be the same as 1808 /* The RB field in an lswx instruction, which has special value 2629 /* An XX1_MASK with the RB field fixed. */ 2681 /* An X_MASK with the RB field fixed. */ [all …]
|
/linux/Documentation/core-api/ |
H A D | rbtree.rst | 283 Interval tree is an example of augmented rb tree. Reference - 309 node = rb_entry(root->rb_node, struct interval_tree_node, rb); 312 if (node->rb.rb_left) { 314 rb_entry(node->rb.rb_left, 315 struct interval_tree_node, rb); 332 if (node->rb.rb_right) { 333 node = rb_entry(node->rb.rb_right, 334 struct interval_tree_node, rb); 349 if (node->rb.rb_left) { 350 subtree_last = rb_entry(node->rb.rb_left, [all …]
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | pin_system.c | 17 struct mmu_rb_node rb; member 74 return node->rb.handler->mn.mm; in mm_from_sdma_node() 111 return container_of(rb_node, struct sdma_mmu_node, rb); in find_system_node() 151 node->rb.addr = start_address; in pin_system_pages() 152 node->rb.len = length; in pin_system_pages() 181 kref_init(&node->rb.refcount); in add_system_pinning() 184 kref_get(&node->rb.refcount); in add_system_pinning() 189 ret = hfi1_mmu_rb_insert(pq->handler, &node->rb); in add_system_pinning() 239 if (node->rb.addr <= start) { in get_system_cache_entry() 248 SDMA_DBG(req, "prepend: node->rb.addr %lx, node->rb.refcount %d", in get_system_cache_entry() [all …]
|
/linux/fs/jffs2/ |
H A D | nodelist.h | 230 struct rb_node rb; member 271 struct rb_node rb; member 334 return rb_entry(node, struct jffs2_node_frag, rb); in frag_first() 344 return rb_entry(node, struct jffs2_node_frag, rb); in frag_last() 347 #define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb) 348 #define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb) 349 #define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb) 350 #define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb) 351 #define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb) 352 #define frag_erase(frag, list) rb_erase(&frag->rb, list) [all …]
|
/linux/arch/arm/lib/ |
H A D | getuser.S | 49 rb .req ip label 51 3: ldrbt rb, [r0], #0 53 rb .req r0 label 55 3: ldrb rb, [r0, #1] 58 orr r2, r2, rb, lsl #8 60 orr r2, rb, r2, lsl #8 117 rb .req ip label 119 10: ldrbt rb, [r0], #0 121 rb .req r0 label 123 10: ldrb rb, [r0, #1] [all …]
|
/linux/tools/testing/selftests/bpf/progs/ |
H A D | refcounted_kptr.c | 144 struct bpf_rb_node *rb; in __read_from_tree() local 150 rb = bpf_rbtree_first(root); in __read_from_tree() 151 if (!rb) { in __read_from_tree() 156 n = container_of(rb, struct node_data, r); in __read_from_tree() 164 rb = bpf_rbtree_remove(root, rb); in __read_from_tree() 166 if (!rb) in __read_from_tree() 168 n = container_of(rb, struct node_data, r); in __read_from_tree() 509 struct bpf_rb_node *rb; in BPF_PROG() local 519 rb = bpf_rbtree_first(&root); in BPF_PROG() 520 if (!rb) in BPF_PROG() [all …]
|
/linux/net/sunrpc/xprtrdma/ |
H A D | verbs.c | 82 static void rpcrdma_regbuf_dma_unmap(struct rpcrdma_regbuf *rb); 83 static void rpcrdma_regbuf_free(struct rpcrdma_regbuf *rb); 861 struct rpcrdma_regbuf *rb; in rpcrdma_req_setup() local 868 rb = rpcrdma_regbuf_alloc(__roundup_pow_of_two(maxhdrsize), in rpcrdma_req_setup() 870 if (!rb) in rpcrdma_req_setup() 873 if (!__rpcrdma_regbuf_dma_map(r_xprt, rb)) in rpcrdma_req_setup() 876 req->rl_rdmabuf = rb; in rpcrdma_req_setup() 877 xdr_buf_init(&req->rl_hdrbuf, rdmab_data(rb), rdmab_length(rb)); in rpcrdma_req_setup() 881 rpcrdma_regbuf_free(rb); in rpcrdma_req_setup() 1246 struct rpcrdma_regbuf *rb; in rpcrdma_regbuf_alloc_node() local [all …]
|