Lines Matching defs:rb_map

197 	struct bpf_ringbuf_map *rb_map;
213 rb_map = bpf_map_area_alloc(sizeof(*rb_map), NUMA_NO_NODE);
214 if (!rb_map)
217 bpf_map_init_from_attr(&rb_map->map, attr);
219 rb_map->rb = bpf_ringbuf_alloc(attr->max_entries, rb_map->map.numa_node, overwrite_mode);
220 if (!rb_map->rb) {
221 bpf_map_area_free(rb_map);
225 return &rb_map->map;
246 struct bpf_ringbuf_map *rb_map;
248 rb_map = container_of(map, struct bpf_ringbuf_map, map);
249 bpf_ringbuf_free(rb_map->rb);
250 bpf_map_area_free(rb_map);
277 struct bpf_ringbuf_map *rb_map;
279 rb_map = container_of(map, struct bpf_ringbuf_map, map);
287 return remap_vmalloc_range(vma, rb_map->rb,
293 struct bpf_ringbuf_map *rb_map;
295 rb_map = container_of(map, struct bpf_ringbuf_map, map);
306 return remap_vmalloc_range(vma, rb_map->rb, vma->vm_pgoff + RINGBUF_PGOFF);
339 struct bpf_ringbuf_map *rb_map;
341 rb_map = container_of(map, struct bpf_ringbuf_map, map);
342 poll_wait(filp, &rb_map->rb->waitq, pts);
344 if (ringbuf_avail_data_sz(rb_map->rb))
352 struct bpf_ringbuf_map *rb_map;
354 rb_map = container_of(map, struct bpf_ringbuf_map, map);
355 poll_wait(filp, &rb_map->rb->waitq, pts);
357 if (ringbuf_avail_data_sz(rb_map->rb) < ringbuf_total_data_sz(rb_map->rb))
542 struct bpf_ringbuf_map *rb_map;
547 rb_map = container_of(map, struct bpf_ringbuf_map, map);
548 return (unsigned long)__bpf_ringbuf_reserve(rb_map->rb, size);
616 struct bpf_ringbuf_map *rb_map;
622 rb_map = container_of(map, struct bpf_ringbuf_map, map);
623 rec = __bpf_ringbuf_reserve(rb_map->rb, size);
673 struct bpf_ringbuf_map *rb_map;
688 rb_map = container_of(map, struct bpf_ringbuf_map, map);
690 sample = __bpf_ringbuf_reserve(rb_map->rb, size);