Home
last modified time | relevance | path

Searched refs:rb (Results 1 – 25 of 306) sorted by relevance

12345678910>>...13

/linux/kernel/events/
H A Dring_buffer.c22 atomic_set(&handle->rb->poll, EPOLLIN | EPOLLRDNORM); in perf_output_wakeup()
42 struct perf_buffer *rb = handle->rb; in perf_output_get_handle() local
50 (*(volatile unsigned int *)&rb->nest)++; in perf_output_get_handle()
51 handle->wakeup = local_read(&rb->wakeup); in perf_output_get_handle()
56 struct perf_buffer *rb = handle->rb; in perf_output_put_handle() local
64 nest = READ_ONCE(rb->nest); in perf_output_put_handle()
66 WRITE_ONCE(rb->nest, nest - 1); in perf_output_put_handle()
80 head = local_read(&rb->head); in perf_output_put_handle()
114 WRITE_ONCE(rb->user_page->data_head, head); in perf_output_put_handle()
122 WRITE_ONCE(rb->nest, 0); in perf_output_put_handle()
[all …]
H A Dinternal.h63 extern void rb_free(struct perf_buffer *rb);
67 struct perf_buffer *rb; in rb_free_rcu() local
69 rb = container_of(rcu_head, struct perf_buffer, rcu_head); in rb_free_rcu()
70 rb_free(rb); in rb_free_rcu()
73 static inline void rb_toggle_paused(struct perf_buffer *rb, bool pause) in rb_toggle_paused() argument
75 if (!pause && rb->nr_pages) in rb_toggle_paused()
76 rb->paused = 0; in rb_toggle_paused()
78 rb->paused = 1; in rb_toggle_paused()
84 extern int rb_alloc_aux(struct perf_buffer *rb, struct perf_event *event,
86 extern void rb_free_aux(struct perf_buffer *rb);
[all …]
/linux/drivers/scsi/bfa/
H A Dbfa_ioc_ct.c185 void __iomem *rb; in bfa_ioc_ct_reg_init() local
188 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init()
190 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init()
191 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init()
192 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init()
195 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init()
196 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init()
197 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init()
198 ioc->ioc_regs.hfn_mbox_cmd = rb + ct_p0reg[pcifn].hfn; in bfa_ioc_ct_reg_init()
199 ioc->ioc_regs.lpu_mbox_cmd = rb + ct_p0reg[pcifn].lpu; in bfa_ioc_ct_reg_init()
[all …]
H A Dbfa_ioc_cb.c138 void __iomem *rb; in bfa_ioc_cb_reg_init() local
141 rb = bfa_ioc_bar0(ioc); in bfa_ioc_cb_reg_init()
143 ioc->ioc_regs.hfn_mbox = rb + iocreg_fnreg[pcifn].hfn_mbox; in bfa_ioc_cb_reg_init()
144 ioc->ioc_regs.lpu_mbox = rb + iocreg_fnreg[pcifn].lpu_mbox; in bfa_ioc_cb_reg_init()
145 ioc->ioc_regs.host_page_num_fn = rb + iocreg_fnreg[pcifn].hfn_pgn; in bfa_ioc_cb_reg_init()
148 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_cb_reg_init()
149 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_cb_reg_init()
150 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_cb_reg_init()
152 ioc->ioc_regs.heartbeat = (rb + BFA_IOC1_HBEAT_REG); in bfa_ioc_cb_reg_init()
153 ioc->ioc_regs.ioc_fwstate = (rb + BFA_IOC1_STATE_REG); in bfa_ioc_cb_reg_init()
[all …]
/linux/drivers/net/ethernet/brocade/bna/
H A Dbfa_ioc_ct.c49 static enum bfa_status bfa_ioc_ct_pll_init(void __iomem *rb,
51 static enum bfa_status bfa_ioc_ct2_pll_init(void __iomem *rb,
251 void __iomem *rb; in bfa_ioc_ct_reg_init() local
254 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init()
256 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init()
257 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init()
258 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init()
261 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init()
262 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init()
263 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init()
[all …]
/linux/fs/xfs/scrub/
H A Dbmap_repair.c99 struct xrep_bmap *rb, in xrep_bmap_discover_shared() argument
103 struct xfs_scrub *sc = rb->sc; in xrep_bmap_discover_shared()
123 rb->reflink_scan = RLS_SET_IFLAG; in xrep_bmap_discover_shared()
131 struct xrep_bmap *rb, in xrep_bmap_from_rmap() argument
143 struct xfs_scrub *sc = rb->sc; in xrep_bmap_from_rmap()
151 if (rb->reflink_scan == RLS_UNKNOWN && !unwritten) { in xrep_bmap_from_rmap()
152 error = xrep_bmap_discover_shared(rb, startblock, blockcount); in xrep_bmap_from_rmap()
163 fa = xfs_bmap_validate_extent(sc->ip, rb->whichfork, &irec); in xrep_bmap_from_rmap()
169 trace_xrep_bmap_found(sc->ip, rb->whichfork, &irec); in xrep_bmap_from_rmap()
174 error = xfarray_append(rb->bmap_records, &rbe); in xrep_bmap_from_rmap()
[all …]
/linux/kernel/bpf/
H A Dringbuf.c84 struct bpf_ringbuf *rb;
101 struct bpf_ringbuf *rb; in bpf_ringbuf_area_alloc()
138 rb = vmap(pages, nr_meta_pages + 2 * nr_data_pages, in bpf_ringbuf_area_alloc()
140 if (rb) { in bpf_ringbuf_area_alloc()
142 rb->pages = pages; in bpf_ringbuf_area_alloc()
143 rb->nr_pages = nr_pages; in bpf_ringbuf_area_alloc()
144 return rb; in bpf_ringbuf_area_alloc()
156 struct bpf_ringbuf *rb = container_of(work, struct bpf_ringbuf, work); in bpf_ringbuf_notify()
158 wake_up_all(&rb->waitq); in bpf_ringbuf_notify()
174 struct bpf_ringbuf *rb; in bpf_ringbuf_alloc()
83 struct bpf_ringbuf *rb; global() member
100 struct bpf_ringbuf *rb; bpf_ringbuf_area_alloc() local
155 struct bpf_ringbuf *rb = container_of(work, struct bpf_ringbuf, work); bpf_ringbuf_notify() local
173 struct bpf_ringbuf *rb; bpf_ringbuf_alloc() local
227 bpf_ringbuf_free(struct bpf_ringbuf * rb) bpf_ringbuf_free() argument
314 ringbuf_avail_data_sz(struct bpf_ringbuf * rb) ringbuf_avail_data_sz() argument
330 ringbuf_total_data_sz(const struct bpf_ringbuf * rb) ringbuf_total_data_sz() argument
363 struct bpf_ringbuf *rb; ringbuf_map_mem_usage() local
412 bpf_ringbuf_rec_pg_off(struct bpf_ringbuf * rb,struct bpf_ringbuf_hdr * hdr) bpf_ringbuf_rec_pg_off() argument
430 bpf_ringbuf_has_space(const struct bpf_ringbuf * rb,unsigned long new_prod_pos,unsigned long cons_pos,unsigned long pend_pos) bpf_ringbuf_has_space() argument
462 __bpf_ringbuf_reserve(struct bpf_ringbuf * rb,u64 size) __bpf_ringbuf_reserve() argument
562 struct bpf_ringbuf *rb; bpf_ringbuf_commit() local
642 struct bpf_ringbuf *rb; BPF_CALL_2() local
747 __bpf_user_ringbuf_peek(struct bpf_ringbuf * rb,void ** sample,u32 * size) __bpf_user_ringbuf_peek() argument
802 __bpf_user_ringbuf_sample_release(struct bpf_ringbuf * rb,size_t size,u64 flags) __bpf_user_ringbuf_sample_release() argument
819 struct bpf_ringbuf *rb; BPF_CALL_4() local
[all...]
H A Drange_tree.c44 static struct range_node *rb_to_range_node(struct rb_node *rb)
46 return rb_entry(rb, struct range_node, rb_range_size); in rb_to_range_node()
57 struct rb_node *rb = rt->range_size_root.rb_root.rb_node; in __find_range()
60 while (rb) { in __find_range()
61 struct range_node *rn = rb_to_range_node(rb); in __find_range()
65 rb = rb->rb_right; in __find_range()
67 rb = rb->rb_left; in __find_range()
88 struct rb_node **link = &root->rb_root.rb_node, *rb in __range_size_insert()
45 rb_to_range_node(struct rb_node * rb) rb_to_range_node() argument
58 struct rb_node *rb = rt->range_size_root.rb_root.rb_node; __find_range() local
89 struct rb_node **link = &root->rb_root.rb_node, *rb = NULL; __range_size_insert() local
[all...]
/linux/drivers/hid/intel-ish-hid/ishtp/
H A Dclient-buffers.c23 struct ishtp_cl_rb *rb; in ishtp_cl_alloc_rx_ring() local
28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring()
29 if (!rb) { in ishtp_cl_alloc_rx_ring()
33 ret = ishtp_io_rb_alloc_buf(rb, len); in ishtp_cl_alloc_rx_ring()
37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring()
99 struct ishtp_cl_rb *rb; in ishtp_cl_free_rx_ring() local
105 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, in ishtp_cl_free_rx_ring()
107 list_del(&rb->list); in ishtp_cl_free_rx_ring()
108 kfree(rb->buffer.data); in ishtp_cl_free_rx_ring()
109 kfree(rb); in ishtp_cl_free_rx_ring()
[all …]
H A Dclient.c25 struct ishtp_cl_rb *rb; in ishtp_read_list_flush() local
30 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush()
31 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush()
32 list_del(&rb->list); in ishtp_read_list_flush()
34 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_read_list_flush()
593 struct ishtp_cl_rb *rb; in ishtp_cl_read_start() local
623 rb = NULL; in ishtp_cl_read_start()
627 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, list); in ishtp_cl_read_start()
628 list_del_init(&rb->list); in ishtp_cl_read_start()
631 rb->cl = cl; in ishtp_cl_read_start()
[all …]
/linux/lib/
H A Drbtree_test.c21 struct rb_node rb; member
40 if (key < rb_entry(parent, struct test_node, rb)->key) in insert()
46 rb_link_node(&node->rb, parent, new); in insert()
47 rb_insert_color(&node->rb, &root->rb_root); in insert()
58 if (key < rb_entry(parent, struct test_node, rb)->key) in insert_cached()
66 rb_link_node(&node->rb, parent, new); in insert_cached()
67 rb_insert_color_cached(&node->rb, root, leftmost); in insert_cached()
72 rb_erase(&node->rb, &root->rb_root); in erase()
77 rb_erase_cached(&node->rb, root); in erase_cached()
84 struct test_node, rb, u32, augmented, NODE_VAL) in RB_DECLARE_CALLBACKS_MAX() argument
[all …]
/linux/kernel/printk/
H A Dprintk_ringbuffer.c582 static bool data_make_reusable(struct printk_ringbuffer *rb, in data_make_reusable() argument
588 struct prb_data_ring *data_ring = &rb->text_data_ring; in data_make_reusable()
589 struct prb_desc_ring *desc_ring = &rb->desc_ring; in data_make_reusable()
650 static bool data_push_tail(struct printk_ringbuffer *rb, unsigned long lpos) in data_push_tail() argument
652 struct prb_data_ring *data_ring = &rb->text_data_ring; in data_push_tail()
693 if (!data_make_reusable(rb, tail_lpos, lpos, &next_lpos)) { in data_push_tail()
785 static bool desc_push_tail(struct printk_ringbuffer *rb, in desc_push_tail() argument
788 struct prb_desc_ring *desc_ring = &rb->desc_ring; in desc_push_tail()
830 if (!data_push_tail(rb, desc.text_blk_lpos.next)) in desc_push_tail()
893 static bool desc_reserve(struct printk_ringbuffer *rb, unsigned long *id_out) in desc_reserve() argument
[all …]
H A Dprintk_ringbuffer.h104 * @rb: Ringbuffer where the entry is reserved.
114 struct printk_ringbuffer *rb;
327 bool prb_reserve(struct prb_reserved_entry *e, struct printk_ringbuffer *rb,
329 bool prb_reserve_in_last(struct prb_reserved_entry *e, struct printk_ringbuffer *rb,
334 void prb_init(struct printk_ringbuffer *rb,
367 * @rb: The ringbuffer to iterate over.
377 #define prb_for_each_record(from, rb, s, r) \
378 for ((s) = from; prb_read_valid(rb, s, r); (s) = (r)->info->seq + 1)
384 * @rb: The ringbuffer to iterate over.
395 #define prb_for_each_info(from, rb,
109 struct printk_ringbuffer *rb; global() member
372 prb_for_each_record(from,rb,s,r) global() argument
390 prb_for_each_info(from,rb,s,i,lc) global() argument
406 __ulseq_to_u64seq(rb,ulseq) global() argument
407 ULSEQ_MAX(rb) global() argument
412 ULSEQ_MAX(rb) global() argument
414 __ulseq_to_u64seq(struct printk_ringbuffer * rb,u32 ulseq) __ulseq_to_u64seq() argument
[all...]
/linux/drivers/misc/mchp_pci1xxxx/
H A Dmchp_pci1xxxx_otpe2p.c99 void __iomem *rb = priv->reg_base; in is_eeprom_responsive() local
104 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive()
106 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive()
111 true, rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive()
124 void __iomem *rb = priv->reg_base; in pci1xxxx_eeprom_read() local
141 writel(EEPROM_CMD_EPC_BUSY_BIT | (off + byte), rb + in pci1xxxx_eeprom_read()
148 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in pci1xxxx_eeprom_read()
154 buf[byte] = readl(rb + MMAP_EEPROM_OFFSET(EEPROM_DATA_REG)); in pci1xxxx_eeprom_read()
165 void __iomem *rb = priv->reg_base; in pci1xxxx_eeprom_write() local
182 writel(*(value + byte), rb + MMAP_EEPROM_OFFSET(EEPROM_DATA_REG)); in pci1xxxx_eeprom_write()
[all …]
/linux/Documentation/translations/zh_CN/core-api/
H A Drbtree.rst271 node = rb_entry(root->rb_node, struct interval_tree_node, rb);
274 if (node->rb.rb_left) {
276 rb_entry(node->rb.rb_left,
277 struct interval_tree_node, rb);
294 if (node->rb.rb_right) {
295 node = rb_entry(node->rb.rb_right,
296 struct interval_tree_node, rb);
311 if (node->rb.rb_left) {
312 subtree_last = rb_entry(node->rb.rb_left,
313 struct interval_tree_node, rb)->__subtree_last;
[all …]
/linux/arch/arm64/crypto/
H A Dsm3-neon-core.S42 #define rb w4 macro
356 ldp ra, rb, [RSTATE, #0]
401 R1(ra, rb, rc, rd, re, rf, rg, rh, k_even, KL, 0, 0, IW, _, 0)
402 R1(rd, ra, rb, rc, rh, re, rf, rg, k_odd, _, 1, 1, IW, _, 0)
403 R1(rc, rd, ra, rb, rg, rh, re, rf, k_even, KL, 2, 2, IW, _, 0)
404 R1(rb, rc, rd, ra, rf, rg, rh, re, k_odd, _, 3, 3, IW, _, 0)
407 R1(ra, rb, rc, rd, re, rf, rg, rh, k_even, KL, 4, 0, IW, _, 0)
408 R1(rd, ra, rb, rc, rh, re, rf, rg, k_odd, _, 5, 1, IW, _, 0)
409 R1(rc, rd, ra, rb, rg, rh, re, rf, k_even, KL, 6, 2, IW, SCHED_W_W0W1W2W3W4W5_1, 12)
410 R1(rb, rc, rd, ra, rf, rg, rh, re, k_odd, _, 7, 3, IW, SCHED_W_W0W1W2W3W4W5_2, 12)
[all …]
/linux/mm/
H A Dinterval_tree.c23 INTERVAL_TREE_DEFINE(struct vm_area_struct, shared.rb,
38 if (!prev->shared.rb.rb_right) { in vma_interval_tree_insert_after()
40 link = &prev->shared.rb.rb_right; in vma_interval_tree_insert_after()
42 parent = rb_entry(prev->shared.rb.rb_right, in vma_interval_tree_insert_after()
43 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after()
46 while (parent->shared.rb.rb_left) { in vma_interval_tree_insert_after()
47 parent = rb_entry(parent->shared.rb.rb_left, in vma_interval_tree_insert_after()
48 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after()
52 link = &parent->shared.rb.rb_left; in vma_interval_tree_insert_after()
56 rb_link_node(&node->shared.rb, &parent->shared.rb, link); in vma_interval_tree_insert_after()
[all …]
/linux/drivers/gpu/drm/
H A Ddrm_prime.c101 struct rb_node **p, *rb; in drm_prime_add_buf_handle() local
111 rb = NULL; in drm_prime_add_buf_handle()
116 rb = *p; in drm_prime_add_buf_handle()
117 pos = rb_entry(rb, struct drm_prime_member, dmabuf_rb); in drm_prime_add_buf_handle()
119 p = &rb->rb_right; in drm_prime_add_buf_handle()
121 p = &rb->rb_left; in drm_prime_add_buf_handle()
123 rb_link_node(&member->dmabuf_rb, rb, p); in drm_prime_add_buf_handle()
126 rb = NULL; in drm_prime_add_buf_handle()
131 rb = *p; in drm_prime_add_buf_handle()
132 pos = rb_entry(rb, struct drm_prime_member, handle_rb); in drm_prime_add_buf_handle()
[all …]
/linux/fs/jffs2/
H A Dnodelist.h230 struct rb_node rb; member
271 struct rb_node rb; member
334 return rb_entry(node, struct jffs2_node_frag, rb); in frag_first()
344 return rb_entry(node, struct jffs2_node_frag, rb); in frag_last()
347 #define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb)
348 #define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb)
349 #define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb)
350 #define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb)
351 #define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb)
352 #define frag_erase(frag, list) rb_erase(&frag->rb, list)
[all …]
/linux/drivers/firmware/arm_scmi/
H A Draw_mode.c261 struct scmi_raw_buffer *rb = NULL; in scmi_raw_buffer_get() local
266 rb = list_first_entry(head, struct scmi_raw_buffer, node); in scmi_raw_buffer_get()
267 list_del_init(&rb->node); in scmi_raw_buffer_get()
271 return rb; in scmi_raw_buffer_get()
275 struct scmi_raw_buffer *rb) in scmi_raw_buffer_put() argument
280 rb->msg.len = rb->max_len; in scmi_raw_buffer_put()
283 list_add_tail(&rb->node, &q->free_bufs); in scmi_raw_buffer_put()
288 struct scmi_raw_buffer *rb) in scmi_raw_buffer_enqueue() argument
293 list_add_tail(&rb->node, &q->msg_q); in scmi_raw_buffer_enqueue()
302 struct scmi_raw_buffer *rb = NULL; in scmi_raw_buffer_dequeue_unlocked() local
[all …]
/linux/net/sunrpc/xprtrdma/
H A Dverbs.c82 static void rpcrdma_regbuf_dma_unmap(struct rpcrdma_regbuf *rb);
83 static void rpcrdma_regbuf_free(struct rpcrdma_regbuf *rb);
861 struct rpcrdma_regbuf *rb; in rpcrdma_req_setup() local
868 rb = rpcrdma_regbuf_alloc(__roundup_pow_of_two(maxhdrsize), in rpcrdma_req_setup()
870 if (!rb) in rpcrdma_req_setup()
873 if (!__rpcrdma_regbuf_dma_map(r_xprt, rb)) in rpcrdma_req_setup()
876 req->rl_rdmabuf = rb; in rpcrdma_req_setup()
877 xdr_buf_init(&req->rl_hdrbuf, rdmab_data(rb), rdmab_length(rb)); in rpcrdma_req_setup()
881 rpcrdma_regbuf_free(rb); in rpcrdma_req_setup()
1246 struct rpcrdma_regbuf *rb; in rpcrdma_regbuf_alloc_node() local
[all …]
H A Dxprt_rdma.h129 static inline u64 rdmab_addr(struct rpcrdma_regbuf *rb) in rdmab_addr() argument
131 return rb->rg_iov.addr; in rdmab_addr()
134 static inline u32 rdmab_length(struct rpcrdma_regbuf *rb) in rdmab_length() argument
136 return rb->rg_iov.length; in rdmab_length()
139 static inline u32 rdmab_lkey(struct rpcrdma_regbuf *rb) in rdmab_lkey() argument
141 return rb->rg_iov.lkey; in rdmab_lkey()
144 static inline struct ib_device *rdmab_device(struct rpcrdma_regbuf *rb) in rdmab_device() argument
146 return rb->rg_device; in rdmab_device()
149 static inline void *rdmab_data(const struct rpcrdma_regbuf *rb) in rdmab_data() argument
151 return rb->rg_data; in rdmab_data()
[all …]
/linux/tools/testing/selftests/bpf/benchs/
H A Drun_bench_ringbufs.sh10 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
15 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
16 summarize $b "$($RUN_RB_BENCH --rb-sampled $b)"
20 for b in rb-libbpf rb-custom pb-libbpf pb-custom; do
21 summarize $b "$($RUN_RB_BENCH --rb-b2b $b)"
22 summarize $b-sampled "$($RUN_RB_BENCH --rb-sampled --rb
[all...]
/linux/drivers/xen/xenbus/
H A Dxenbus_dev_frontend.c130 struct read_buffer *rb; in xenbus_file_read() local
148 rb = list_entry(u->read_buffers.next, struct read_buffer, list); in xenbus_file_read()
151 size_t sz = min_t(size_t, len - i, rb->len - rb->cons); in xenbus_file_read()
153 ret = copy_to_user(ubuf + i, &rb->msg[rb->cons], sz); in xenbus_file_read()
156 rb->cons += sz - ret; in xenbus_file_read()
165 if (rb->cons == rb->len) { in xenbus_file_read()
166 list_del(&rb->list); in xenbus_file_read()
167 kfree(rb); in xenbus_file_read()
170 rb = list_entry(u->read_buffers.next, in xenbus_file_read()
191 struct read_buffer *rb; in queue_reply() local
[all …]
/linux/drivers/infiniband/hw/hfi1/
H A Dpin_system.c17 struct mmu_rb_node rb; member
74 return node->rb.handler->mn.mm; in mm_from_sdma_node()
111 return container_of(rb_node, struct sdma_mmu_node, rb); in find_system_node()
151 node->rb.addr = start_address; in pin_system_pages()
152 node->rb.len = length; in pin_system_pages()
181 kref_init(&node->rb.refcount); in add_system_pinning()
184 kref_get(&node->rb.refcount); in add_system_pinning()
189 ret = hfi1_mmu_rb_insert(pq->handler, &node->rb); in add_system_pinning()
239 if (node->rb.addr <= start) { in get_system_cache_entry()
249 node->rb.addr, kref_read(&node->rb.refcount)); in get_system_cache_entry()
[all …]

12345678910>>...13