| /linux/kernel/events/ |
| H A D | ring_buffer.c | 22 atomic_set(&handle->rb->poll, EPOLLIN | EPOLLRDNORM); in perf_output_wakeup() 42 struct perf_buffer *rb = handle->rb; in perf_output_get_handle() local 50 (*(volatile unsigned int *)&rb->nest)++; in perf_output_get_handle() 51 handle->wakeup = local_read(&rb->wakeup); in perf_output_get_handle() 56 struct perf_buffer *rb = handle->rb; in perf_output_put_handle() local 62 * @rb->user_page->data_head. in perf_output_put_handle() 64 nest = READ_ONCE(rb->nest); in perf_output_put_handle() 66 WRITE_ONCE(rb->nest, nest - 1); in perf_output_put_handle() 73 * we must ensure the load of @rb->head happens after we've in perf_output_put_handle() 74 * incremented @rb->nest. in perf_output_put_handle() [all …]
|
| H A D | internal.h | 63 extern void rb_free(struct perf_buffer *rb); 67 struct perf_buffer *rb; in rb_free_rcu() local 69 rb = container_of(rcu_head, struct perf_buffer, rcu_head); in rb_free_rcu() 70 free_uid(rb->mmap_user); in rb_free_rcu() 71 rb_free(rb); in rb_free_rcu() 74 static inline void rb_toggle_paused(struct perf_buffer *rb, bool pause) in rb_toggle_paused() argument 76 if (!pause && rb->nr_pages) in rb_toggle_paused() 77 rb->paused = 0; in rb_toggle_paused() 79 rb->paused = 1; in rb_toggle_paused() 85 extern int rb_alloc_aux(struct perf_buffer *rb, struct perf_event *event, [all …]
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| H A D | bfa_ioc_ct.c | 49 static enum bfa_status bfa_ioc_ct_pll_init(void __iomem *rb, 51 static enum bfa_status bfa_ioc_ct2_pll_init(void __iomem *rb, 251 void __iomem *rb; in bfa_ioc_ct_reg_init() local 254 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init() 256 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init() 257 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init() 258 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init() 261 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init() 262 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init() 263 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init() [all …]
|
| /linux/drivers/scsi/bfa/ |
| H A D | bfa_ioc_ct.c | 185 void __iomem *rb; in bfa_ioc_ct_reg_init() local 188 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init() 190 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init() 191 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init() 192 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init() 195 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init() 196 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init() 197 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init() 198 ioc->ioc_regs.hfn_mbox_cmd = rb + ct_p0reg[pcifn].hfn; in bfa_ioc_ct_reg_init() 199 ioc->ioc_regs.lpu_mbox_cmd = rb + ct_p0reg[pcifn].lpu; in bfa_ioc_ct_reg_init() [all …]
|
| H A D | bfa_ioc_cb.c | 138 void __iomem *rb; in bfa_ioc_cb_reg_init() local 141 rb = bfa_ioc_bar0(ioc); in bfa_ioc_cb_reg_init() 143 ioc->ioc_regs.hfn_mbox = rb + iocreg_fnreg[pcifn].hfn_mbox; in bfa_ioc_cb_reg_init() 144 ioc->ioc_regs.lpu_mbox = rb + iocreg_fnreg[pcifn].lpu_mbox; in bfa_ioc_cb_reg_init() 145 ioc->ioc_regs.host_page_num_fn = rb + iocreg_fnreg[pcifn].hfn_pgn; in bfa_ioc_cb_reg_init() 148 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_cb_reg_init() 149 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_cb_reg_init() 150 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_cb_reg_init() 152 ioc->ioc_regs.heartbeat = (rb + BFA_IOC1_HBEAT_REG); in bfa_ioc_cb_reg_init() 153 ioc->ioc_regs.ioc_fwstate = (rb + BFA_IOC1_STATE_REG); in bfa_ioc_cb_reg_init() [all …]
|
| /linux/tools/testing/selftests/bpf/benchs/ |
| H A D | run_bench_ringbufs.sh | 10 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 15 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 16 summarize $b "$($RUN_RB_BENCH --rb-sampled $b)" 20 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do 21 summarize $b "$($RUN_RB_BENCH --rb-b2b $b)" 22 summarize $b-sampled "$($RUN_RB_BENCH --rb-sampled --rb [all...] |
| /linux/arch/x86/crypto/ |
| H A D | serpent-sse2-i586-asm_32.S | 28 #define RB %xmm1 macro 513 read_blocks(%eax, RA, RB, RC, RD, RT0, RT1, RE); 515 K(RA, RB, RC, RD, RE, 0); 516 S0(RA, RB, RC, RD, RE); LK(RC, RB, RD, RA, RE, 1); 517 S1(RC, RB, RD, RA, RE); LK(RE, RD, RA, RC, RB, 2); 518 S2(RE, RD, RA, RC, RB); LK(RB, RD, RE, RC, RA, 3); 519 S3(RB, RD, RE, RC, RA); LK(RC, RA, RD, RB, RE, 4); 520 S4(RC, RA, RD, RB, RE); LK(RA, RD, RB, RE, RC, 5); 521 S5(RA, RD, RB, RE, RC); LK(RC, RA, RD, RE, RB, 6); 522 S6(RC, RA, RD, RE, RB); LK(RD, RB, RA, RE, RC, 7); [all …]
|
| H A D | serpent-sse2-x86_64-asm_64.S | 636 K2(RA, RB, RC, RD, RE, 0); 637 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1); 638 S(S1, RC, RB, RD, RA, RE); LK2(RE, RD, RA, RC, RB, 2); 639 S(S2, RE, RD, RA, RC, RB); LK2(RB, RD, RE, RC, RA, 3); 640 S(S3, RB, RD, RE, RC, RA); LK2(RC, RA, RD, RB, RE, 4); 641 S(S4, RC, RA, RD, RB, RE); LK2(RA, RD, RB, RE, RC, 5); 642 S(S5, RA, RD, RB, RE, RC); LK2(RC, RA, RD, RE, RB, 6); 643 S(S6, RC, RA, RD, RE, RB); LK2(RD, RB, RA, RE, RC, 7); 644 S(S7, RD, RB, RA, RE, RC); LK2(RC, RA, RE, RD, RB, 8); 645 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9); [all …]
|
| H A D | serpent-avx2-asm_64.S | 566 K2(RA, RB, RC, RD, RE, 0); 567 S(S0, RA, RB, RC, RD, RE); LK2(RC, RB, RD, RA, RE, 1); 568 S(S1, RC, RB, RD, RA, RE); LK2(RE, RD, RA, RC, RB, 2); 569 S(S2, RE, RD, RA, RC, RB); LK2(RB, RD, RE, RC, RA, 3); 570 S(S3, RB, RD, RE, RC, RA); LK2(RC, RA, RD, RB, RE, 4); 571 S(S4, RC, RA, RD, RB, RE); LK2(RA, RD, RB, RE, RC, 5); 572 S(S5, RA, RD, RB, RE, RC); LK2(RC, RA, RD, RE, RB, 6); 573 S(S6, RC, RA, RD, RE, RB); LK2(RD, RB, RA, RE, RC, 7); 574 S(S7, RD, RB, RA, RE, RC); LK2(RC, RA, RE, RD, RB, 8); 575 S(S0, RC, RA, RE, RD, RB); LK2(RE, RA, RD, RC, RB, 9); [all …]
|
| /linux/fs/xfs/scrub/ |
| H A D | bmap_repair.c | 99 struct xrep_bmap *rb, in xrep_bmap_discover_shared() argument 103 struct xfs_scrub *sc = rb->sc; in xrep_bmap_discover_shared() 123 rb->reflink_scan = RLS_SET_IFLAG; in xrep_bmap_discover_shared() 131 struct xrep_bmap *rb, in xrep_bmap_from_rmap() argument 143 struct xfs_scrub *sc = rb->sc; in xrep_bmap_from_rmap() 151 if (rb->reflink_scan == RLS_UNKNOWN && !unwritten) { in xrep_bmap_from_rmap() 152 error = xrep_bmap_discover_shared(rb, startblock, blockcount); in xrep_bmap_from_rmap() 163 fa = xfs_bmap_validate_extent(sc->ip, rb->whichfork, &irec); in xrep_bmap_from_rmap() 169 trace_xrep_bmap_found(sc->ip, rb->whichfork, &irec); in xrep_bmap_from_rmap() 174 error = xfarray_append(rb->bmap_records, &rbe); in xrep_bmap_from_rmap() [all …]
|
| /linux/drivers/hid/intel-ish-hid/ishtp/ |
| H A D | client-buffers.c | 23 struct ishtp_cl_rb *rb; in ishtp_cl_alloc_rx_ring() local 28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring() 29 if (!rb) { in ishtp_cl_alloc_rx_ring() 33 ret = ishtp_io_rb_alloc_buf(rb, len); in ishtp_cl_alloc_rx_ring() 37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring() 99 struct ishtp_cl_rb *rb; in ishtp_cl_free_rx_ring() local 105 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, in ishtp_cl_free_rx_ring() 107 list_del(&rb->list); in ishtp_cl_free_rx_ring() 108 kfree(rb->buffer.data); in ishtp_cl_free_rx_ring() 109 kfree(rb); in ishtp_cl_free_rx_ring() [all …]
|
| H A D | client.c | 25 struct ishtp_cl_rb *rb; in ishtp_read_list_flush() local 30 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush() 31 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush() 32 list_del(&rb->list); in ishtp_read_list_flush() 34 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_read_list_flush() 593 struct ishtp_cl_rb *rb; in ishtp_cl_read_start() local 617 /* The current rb is the head of the free rb list */ in ishtp_cl_read_start() 623 rb = NULL; in ishtp_cl_read_start() 627 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, list); in ishtp_cl_read_start() 628 list_del_init(&rb->list); in ishtp_cl_read_start() [all …]
|
| /linux/lib/ |
| H A D | rbtree_test.c | 14 __param(int, nnodes, 100, "Number of nodes in the rb-tree"); 15 __param(int, perf_loops, 1000, "Number of iterations modifying the rb-tree"); 16 __param(int, check_loops, 100, "Number of iterations modifying and verifying the rb-tree"); 21 struct rb_node rb; member 40 if (key < rb_entry(parent, struct test_node, rb)->key) in insert() 46 rb_link_node(&node->rb, parent, new); in insert() 47 rb_insert_color(&node->rb, &root->rb_root); in insert() 58 if (key < rb_entry(parent, struct test_node, rb)->key) in insert_cached() 66 rb_link_node(&node->rb, parent, new); in insert_cached() 67 rb_insert_color_cached(&node->rb, root, leftmost); in insert_cached() [all …]
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | rbtree.rst | 271 node = rb_entry(root->rb_node, struct interval_tree_node, rb); 274 if (node->rb.rb_left) { 276 rb_entry(node->rb.rb_left, 277 struct interval_tree_node, rb); 294 if (node->rb.rb_right) { 295 node = rb_entry(node->rb.rb_right, 296 struct interval_tree_node, rb); 311 if (node->rb.rb_left) { 312 subtree_last = rb_entry(node->rb.rb_left, 313 struct interval_tree_node, rb)->__subtree_last; [all …]
|
| /linux/tools/testing/selftests/powerpc/stringloops/ |
| H A D | memcmp_64.S | 15 #define rB r10 macro 121 lbz rB,0(r4) 122 subf. rC,rB,rA 127 lbz rB,1(r4) 128 subf. rC,rB,rA 133 lbz rB,2(r4) 134 subf. rC,rB,rA 139 lbz rB,3(r4) 140 subf. rC,rB,rA 173 LD rB,0,r4 [all …]
|
| /linux/arch/powerpc/lib/ |
| H A D | memcmp_64.S | 15 #define rB r10 macro 121 lbz rB,0(r4) 122 subf. rC,rB,rA 127 lbz rB,1(r4) 128 subf. rC,rB,rA 133 lbz rB,2(r4) 134 subf. rC,rB,rA 139 lbz rB,3(r4) 140 subf. rC,rB,rA 173 LD rB,0,r4 [all …]
|
| /linux/drivers/misc/mchp_pci1xxxx/ |
| H A D | mchp_pci1xxxx_otpe2p.c | 99 void __iomem *rb = priv->reg_base; in is_eeprom_responsive() local 104 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive() 106 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive() 111 true, rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive() 124 void __iomem *rb = priv->reg_base; in pci1xxxx_eeprom_read() local 141 writel(EEPROM_CMD_EPC_BUSY_BIT | (off + byte), rb + in pci1xxxx_eeprom_read() 148 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in pci1xxxx_eeprom_read() 154 buf[byte] = readl(rb + MMAP_EEPROM_OFFSET(EEPROM_DATA_REG)); in pci1xxxx_eeprom_read() 165 void __iomem *rb = priv->reg_base; in pci1xxxx_eeprom_write() local 182 writel(*(value + byte), rb + MMAP_EEPROM_OFFSET(EEPROM_DATA_REG)); in pci1xxxx_eeprom_write() [all …]
|
| /linux/lib/crypto/powerpc/ |
| H A D | sha1-powerpc-asm.S | 28 #define RB(t) ((((t)+3)%6)+7) macro 40 andc r0,RD(t),RB(t); \ 41 and r6,RB(t),RC(t); \ 48 rotlwi RB(t),RB(t),30; \ 52 and r6,RB(t),RC(t); \ 53 andc r0,RD(t),RB(t); \ 55 rotlwi RB(t),RB(t),30; \ 67 xor r6,RB(t),RC(t); \ 69 rotlwi RB(t),RB(t),30; \ 77 xor r6,RB(t),RC(t); \ [all …]
|
| /linux/arch/powerpc/xmon/ |
| H A D | ppc-opc.c | 538 /* The RB field in an X, XO, M, or MDS form instruction. */ 539 #define RB RAOPT + 1 macro 543 /* The RB field in an X form instruction when it must be the same as 546 #define RBS RB + 1 549 /* The RB field in an lswx instruction, which has special value 554 /* The RB field of the dccci and iccci instructions, which are optional. */ 1782 /* The RB field in an X form instruction when it must be the same as 1807 /* The RB field in an lswx instruction, which has special value 2628 /* An XX1_MASK with the RB field fixed. */ 2680 /* An X_MASK with the RB field fixed. */ [all …]
|
| /linux/drivers/net/wireless/intel/iwlwifi/pcie/ |
| H A D | iwl-context-info.h | 24 * @IWL_CTXT_INFO_RB_SIZE: RB size mask 26 * @IWL_CTXT_INFO_RB_SIZE_1K: Value for 1K RB size 27 * @IWL_CTXT_INFO_RB_SIZE_2K: Value for 2K RB size 28 * @IWL_CTXT_INFO_RB_SIZE_4K: Value for 4K RB size 29 * @IWL_CTXT_INFO_RB_SIZE_8K: Value for 8K RB size 30 * @IWL_CTXT_INFO_RB_SIZE_12K: Value for 12K RB size 31 * @IWL_CTXT_INFO_RB_SIZE_16K: Value for 16K RB size 32 * @IWL_CTXT_INFO_RB_SIZE_20K: Value for 20K RB size 33 * @IWL_CTXT_INFO_RB_SIZE_24K: Value for 24K RB size 34 * @IWL_CTXT_INFO_RB_SIZE_28K: Value for 28K RB size [all …]
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | pin_system.c | 17 struct mmu_rb_node rb; member 74 return node->rb.handler->mn.mm; in mm_from_sdma_node() 111 return container_of(rb_node, struct sdma_mmu_node, rb); in find_system_node() 151 node->rb.addr = start_address; in pin_system_pages() 152 node->rb.len = length; in pin_system_pages() 181 kref_init(&node->rb.refcount); in add_system_pinning() 184 kref_get(&node->rb.refcount); in add_system_pinning() 189 ret = hfi1_mmu_rb_insert(pq->handler, &node->rb); in add_system_pinning() 239 if (node->rb.addr <= start) { in get_system_cache_entry() 248 SDMA_DBG(req, "prepend: node->rb.addr %lx, node->rb.refcount %d", in get_system_cache_entry() [all …]
|
| /linux/fs/jffs2/ |
| H A D | nodelist.h | 230 struct rb_node rb; member 271 struct rb_node rb; member 334 return rb_entry(node, struct jffs2_node_frag, rb); in frag_first() 344 return rb_entry(node, struct jffs2_node_frag, rb); in frag_last() 347 #define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb) 348 #define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb) 349 #define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb) 350 #define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb) 351 #define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb) 352 #define frag_erase(frag, list) rb_erase(&frag->rb, list) [all …]
|
| /linux/arch/arm/lib/ |
| H A D | getuser.S | 49 rb .req ip label 51 3: ldrbt rb, [r0], #0 53 rb .req r0 label 55 3: ldrb rb, [r0, #1] 58 orr r2, r2, rb, lsl #8 60 orr r2, rb, r2, lsl #8 117 rb .req ip label 119 10: ldrbt rb, [r0], #0 121 rb .req r0 label 123 10: ldrb rb, [r0, #1] [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | refcounted_kptr.c | 144 struct bpf_rb_node *rb; in __read_from_tree() local 150 rb = bpf_rbtree_first(root); in __read_from_tree() 151 if (!rb) { in __read_from_tree() 156 n = container_of(rb, struct node_data, r); in __read_from_tree() 164 rb = bpf_rbtree_remove(root, rb); in __read_from_tree() 166 if (!rb) in __read_from_tree() 168 n = container_of(rb, struct node_data, r); in __read_from_tree() 509 struct bpf_rb_node *rb; in BPF_PROG() local 519 rb in BPF_PROG() 543 struct bpf_rb_node *rb; BPF_PROG() local [all...] |
| /linux/tools/testing/selftests/perf_events/ |
| H A D | mmap.c | 25 /* Reserve space for rb, aux with space for shrink-beyond-vma testing. */ 47 FIXTURE_VARIANT_ADD(perf_mmap, rb) in FIXTURE_VARIANT_ADD() argument 85 struct perf_event_mmap_page *rb; in FIXTURE_SETUP() local 115 rb = mmap(region, RB_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED | MAP_FIXED, fd, 0); in FIXTURE_SETUP() 116 if (rb == MAP_FAILED) { in FIXTURE_SETUP() 130 rb->aux_offset = AUX_OFFS; in FIXTURE_SETUP() 131 rb->aux_size = AUX_SIZE; in FIXTURE_SETUP() 137 munmap(rb, RB_SIZE); in FIXTURE_SETUP() 145 munmap(rb, RB_SIZE); in FIXTURE_SETUP() 161 rb = mmap(region, RB_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED | MAP_FIXED, self->fd, 0); in FIXTURE_SETUP() [all …]
|