| /linux/tools/bpf/bpftool/ |
| H A D | pids.c | 34 struct obj_refs *refs; in add_ref() local 40 refs = entry->pvalue; in add_ref() 42 for (i = 0; i < refs->ref_cnt; i++) { in add_ref() 43 if (refs->refs[i].pid == e->pid) in add_ref() 47 tmp = realloc(refs->refs, (refs->ref_cnt + 1) * sizeof(*ref)); in add_ref() 53 refs->refs = tmp; in add_ref() 54 ref = &refs->refs[refs->ref_cnt]; in add_ref() 58 refs->ref_cnt++; in add_ref() 64 refs = calloc(1, sizeof(*refs)); in add_ref() 65 if (!refs) { in add_ref() [all …]
|
| /linux/drivers/media/v4l2-core/ |
| H A D | v4l2-h264.c | 60 b->refs[i].longterm = true; in v4l2_h264_init_reflist_builder() 68 if (!b->refs[i].longterm && dpb[i].frame_num > cur_frame_num) in v4l2_h264_init_reflist_builder() 69 b->refs[i].frame_num = (int)dpb[i].frame_num - in v4l2_h264_init_reflist_builder() 72 b->refs[i].frame_num = dpb[i].frame_num; in v4l2_h264_init_reflist_builder() 74 b->refs[i].top_field_order_cnt = dpb[i].top_field_order_cnt; in v4l2_h264_init_reflist_builder() 75 b->refs[i].bottom_field_order_cnt = dpb[i].bottom_field_order_cnt; in v4l2_h264_init_reflist_builder() 113 return min(b->refs[ref->index].top_field_order_cnt, in v4l2_h264_get_poc() 114 b->refs[ref->index].bottom_field_order_cnt); in v4l2_h264_get_poc() 116 return b->refs[ref->index].top_field_order_cnt; in v4l2_h264_get_poc() 118 return b->refs[ref->index].bottom_field_order_cnt; in v4l2_h264_get_poc() [all …]
|
| /linux/drivers/xen/ |
| H A D | gntdev-dmabuf.c | 44 grant_ref_t *refs; member 429 int count, u32 domid, u32 *refs, u32 *fd) in dmabuf_exp_from_refs() argument 441 map->grants[i].ref = refs[i]; in dmabuf_exp_from_refs() 480 dmabuf_imp_grant_foreign_access(unsigned long *gfns, u32 *refs, in dmabuf_imp_grant_foreign_access() argument 504 refs[i] = cur_ref; in dmabuf_imp_grant_foreign_access() 514 static void dmabuf_imp_end_foreign_access(u32 *refs, int count) in dmabuf_imp_end_foreign_access() argument 519 if (refs[i] != INVALID_GRANT_REF) in dmabuf_imp_end_foreign_access() 520 gnttab_end_foreign_access(refs[i], NULL); in dmabuf_imp_end_foreign_access() 525 kfree(gntdev_dmabuf->u.imp.refs); in dmabuf_imp_free_storage() 538 gntdev_dmabuf->u.imp.refs = kcalloc(count, in dmabuf_imp_alloc_storage() [all …]
|
| /linux/fs/btrfs/ |
| H A D | delayed-ref.h | 66 * refs rbtree in the corresponding delayed ref head 90 refcount_t refs; member 93 * how many refs is this entry adding or deleting. For 94 * head refs, this may be a negative number because it is keeping 96 * For individual refs, this will always be a positive number 120 * the head refs are used to hold a lock on a given extent, which allows us 121 * to make sure that only one process is running the delayed refs 129 * the mutex is held while running the refs, and it is also 134 refcount_t refs; member 145 * This is used to track the final ref_mod from all the refs associated [all …]
|
| H A D | extent-tree.c | 94 * the delayed refs are not processed. 98 u64 offset, int metadata, u64 *refs, u64 *flags, in btrfs_lookup_extent_info() 186 refcount_inc(&head->refs); in btrfs_lookup_extent_info() 211 if (refs) in btrfs_lookup_extent_info() 212 *refs = num_refs; in btrfs_lookup_extent_info() 222 * Back reference rules. Back refs have three main goals: 235 * There are two kinds of back refs. The implicit back refs is optimized 237 * back refs of this kind provide information about the block's owner tree 239 * b-tree searching. The full back refs i 97 btrfs_lookup_extent_info(struct btrfs_trans_handle * trans,struct btrfs_fs_info * fs_info,u64 bytenr,u64 offset,int metadata,u64 * refs,u64 * flags,u64 * owning_root) btrfs_lookup_extent_info() argument 1002 u64 refs; setup_inline_extent_backref() local 1096 u64 refs; update_inline_extent_backref() local 1494 u64 refs; __btrfs_inc_extent_ref() local 3101 u64 refs; __btrfs_free_extent() local 5240 u64 refs[BTRFS_MAX_LEVEL]; global() member 5354 u64 refs; reada_walk_down() local [all...] |
| H A D | backref.c | 218 * Return 0 when both refs are for the same block (and can be merged). 286 * Add @newref to the @root rbtree, merging identical refs. 302 /* Identical refs, merge them and free @newref */ in prelim_ref_insert() 356 * delayed refs 369 * on disk refs (inline or keyed) 417 /* direct refs use root == 0, key == NULL */ 427 /* indirect refs use parent == 0 */ 605 * adding new delayed refs. To deal with this we need to look in cache in resolve_indirect_ref() 715 * We maintain three separate rbtrees: one for direct refs, one for 716 * indirect refs whic 2412 struct ulist *refs; iterate_extent_inodes() local [all...] |
| H A D | delayed-inode.c | 59 refcount_set(&delayed_node->refs, 0); in btrfs_init_delayed_node() 78 refcount_inc(&node->refs); in btrfs_get_delayed_node() 88 refcount_inc(&node->refs); /* can be accessed */ in btrfs_get_delayed_node() 111 if (refcount_inc_not_zero(&node->refs)) { in btrfs_get_delayed_node() 112 refcount_inc(&node->refs); in btrfs_get_delayed_node() 178 refcount_set(&node->refs, 2); in btrfs_get_or_create_delayed_node() 206 refcount_inc(&node->refs); /* inserted into list */ in btrfs_queue_delayed_node() 223 refcount_dec(&node->refs); /* not in the list */ in btrfs_dequeue_delayed_node() 242 refcount_inc(&node->refs); in btrfs_first_delayed_node() 271 refcount_inc(&next->refs); in btrfs_next_delayed_node() [all …]
|
| /linux/io_uring/ |
| H A D | refs.h | 12 ((unsigned int) atomic_read(&(req->refs)) + 127u <= 127u) 17 return atomic_inc_not_zero(&req->refs); in req_ref_inc_not_zero() 24 return atomic_dec_and_test(&req->refs); in req_ref_put_and_test_atomic() 33 return atomic_dec_and_test(&req->refs); in req_ref_put_and_test() 40 atomic_inc(&req->refs); in req_ref_get() 47 atomic_dec(&req->refs); in req_ref_put() 54 atomic_set(&req->refs, nr); in __io_req_set_refcount()
|
| H A D | waitid.c | 29 atomic_t refs; member 133 WARN_ON_ONCE(!(atomic_read(&iw->refs) & IO_WAITID_REF_MASK)); in io_waitid_complete() 156 atomic_or(IO_WAITID_CANCEL_FLAG, &iw->refs); in __io_waitid_cancel() 159 if (atomic_fetch_inc(&iw->refs) & IO_WAITID_REF_MASK) in __io_waitid_cancel() 183 if (!atomic_sub_return(1, &iw->refs)) in io_waitid_drop_issue_ref() 219 if (!(atomic_read(&iw->refs) & IO_WAITID_CANCEL_FLAG)) { in io_waitid_cb() 252 if (atomic_fetch_inc(&iw->refs) & IO_WAITID_REF_MASK) in io_waitid_wait() 298 atomic_set(&iw->refs, 1); in io_waitid()
|
| /linux/fs/smb/client/ |
| H A D | dfs_cache.c | 251 static inline void dump_refs(const struct dfs_info3_param *refs, int numrefs) in dump_refs() argument 257 const struct dfs_info3_param *ref = &refs[i]; in dump_refs() 383 static int copy_ref_data(const struct dfs_info3_param *refs, int numrefs, in copy_ref_data() argument 389 ce->ttl = max_t(int, refs[0].ttl, CACHE_MIN_TTL); in copy_ref_data() 391 ce->srvtype = refs[0].server_type; in copy_ref_data() 392 ce->hdr_flags = refs[0].flags; in copy_ref_data() 393 ce->ref_flags = refs[0].ref_flag; in copy_ref_data() 394 ce->path_consumed = refs[0].path_consumed; in copy_ref_data() 399 t = alloc_target(refs[i].node_name, refs[i].path_consumed); in copy_ref_data() 421 static struct cache_entry *alloc_cache_entry(struct dfs_info3_param *refs, int numrefs) in alloc_cache_entry() argument [all …]
|
| /linux/include/media/ |
| H A D | v4l2-h264.h | 18 * @refs.top_field_order_cnt: top field order count 19 * @refs.bottom_field_order_cnt: bottom field order count 20 * @refs.frame_num: reference frame number 21 * @refs.longterm: set to true for a long term reference 22 * @refs: array of references 27 * @num_valid: number of valid references in the refs array 39 } refs[V4L2_H264_NUM_DPB_ENTRIES]; member
|
| /linux/drivers/gpu/drm/nouveau/nvkm/core/ |
| H A D | event.c | 34 if (--event->refs[index * event->types_nr + type] == 0) { in nvkm_event_put() 51 if (++event->refs[index * event->types_nr + type] == 1) { in nvkm_event_get() 175 if (!event->refs || WARN_ON(id >= event->index_nr)) in nvkm_event_ntfy() 194 if (event->refs) { in nvkm_event_fini() 195 kfree(event->refs); in nvkm_event_fini() 196 event->refs = NULL; in nvkm_event_fini() 204 event->refs = kzalloc(array3_size(index_nr, types_nr, sizeof(*event->refs)), GFP_KERNEL); in __nvkm_event_init() 205 if (!event->refs) in __nvkm_event_init()
|
| /linux/tools/lib/python/abi/ |
| H A D | system_symbols.py | 175 def check_file(self, refs, found): argument 181 for names in refs: 235 """Interactor to group refs into chunks""" 238 refs = [] 241 refs.append(ref) 245 yield refs 247 refs = [] 249 yield refs 321 for refs in self.get_fileref(all_refs, chunk_size): 322 if refs: [all …]
|
| /linux/Documentation/userspace-api/media/dvb/ |
| H A D | headers.rst | 15 :generate-cross-refs: 24 :generate-cross-refs: 33 :generate-cross-refs: 42 :generate-cross-refs:
|
| /linux/net/ieee802154/6lowpan/ |
| H A D | reassembly.c | 35 int *refs); 49 int refs = 1; in lowpan_frag_expire() local 58 inet_frag_kill(&fq->q, &refs); in lowpan_frag_expire() 61 inet_frag_putn(&fq->q, refs); in lowpan_frag_expire() 88 int *refs) in lowpan_frag_queue() argument 149 res = lowpan_frag_reasm(fq, skb, prev_tail, ldev, refs); in lowpan_frag_queue() 169 int *refs) in lowpan_frag_reasm() argument 173 inet_frag_kill(&fq->q, refs); in lowpan_frag_reasm() 310 int ret, refs = 0; in lowpan_frag_rcv() local 313 ret = lowpan_frag_queue(fq, skb, frag_type, &refs); in lowpan_frag_rcv() [all …]
|
| /linux/net/ipv6/netfilter/ |
| H A D | nf_conntrack_reasm.c | 127 int *refs); 172 int *refs) in nf_ct_frag6_queue() argument 226 inet_frag_kill(&fq->q, refs); in nf_ct_frag6_queue() 292 err = nf_ct_frag6_reasm(fq, skb, prev, dev, refs); in nf_ct_frag6_queue() 306 inet_frag_kill(&fq->q, refs); in nf_ct_frag6_queue() 321 int *refs) in nf_ct_frag6_reasm() argument 327 inet_frag_kill(&fq->q, refs); in nf_ct_frag6_reasm() 378 inet_frag_kill(&fq->q, refs); in nf_ct_frag6_reasm() 453 int refs = 0; in nf_ct_frag6_gather() local 491 ret = nf_ct_frag6_queue(fq, skb, fhdr, nhoff, &refs); in nf_ct_frag6_gather() [all …]
|
| /linux/drivers/acpi/acpica/ |
| H A D | acpredef.h | 225 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 229 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 233 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 237 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 241 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 245 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 249 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 253 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 257 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ 261 METHOD_RETURNS(ACPI_RTYPE_PACKAGE)}}, /* Variable-length (Refs) */ [all …]
|
| /linux/drivers/android/binder/ |
| H A D | process.rs | 344 /// Used to store this `NodeRefInfo` in the node's `refs` list. 354 /// The id used for the `Node::refs` list. 552 let mut refs = self.node_refs.lock(); in debug_print_stats() localVariable 554 for r in refs.by_handle.values_mut() { in debug_print_stats() 561 seq_print!(m, " refs: {count} s {strong} w {weak}\n"); in debug_print_stats() 614 let mut refs = self.node_refs.lock(); in debug_print() localVariable 615 for r in refs.by_handle.values_mut() { in debug_print() 808 let mut refs = self.node_refs.lock(); in insert_or_update_handle() localVariable 811 if let Some(handle_ref) = refs.by_node.get(&node_ref.node.global_id()) { in insert_or_update_handle() 813 let info = refs.by_handle.get_mut(&handle).unwrap(); in insert_or_update_handle() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | map_kptr.c | 318 if (p_st->cnt.refs.counter != ref) { in test_map_kptr_ref_pre() 328 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre() 336 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre() 348 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_pre() 365 if (!p_st || p_st->cnt.refs.counter != ref) in test_map_kptr_ref_post() 371 if (p_st->cnt.refs.counter != ref) { in test_map_kptr_ref_post() 381 if (p_st->cnt.refs.counter != ref) in test_map_kptr_ref_post() 481 if (p->cnt.refs.counter != ref) { in test_map_kptr_ref3()
|
| /linux/net/ipv6/ |
| H A D | reassembly.c | 72 int *refs); 110 u32 *prob_offset, int *refs) in ip6_frag_queue() argument 224 err = ip6_frag_reasm(fq, skb, prev_tail, dev, refs); in ip6_frag_queue() 242 inet_frag_kill(&fq->q, refs); in ip6_frag_queue() 259 int *refs) in ip6_frag_reasm() argument 267 inet_frag_kill(&fq->q, refs); in ip6_frag_reasm() 321 inet_frag_kill(&fq->q, refs); in ip6_frag_reasm() 382 int ret, refs = 0; in ipv6_frag_rcv() local 388 &prob_offset, &refs); in ipv6_frag_rcv() 392 inet_frag_putn(&fq->q, refs); in ipv6_frag_rcv()
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
| H A D | cgrp.c | 39 if (refcount_dec_and_test(&ectx->refs)) { in nvkm_cgrp_ectx_put() 65 refcount_inc(&ectx->refs); in nvkm_cgrp_ectx_get() 76 refcount_set(&ectx->refs, 1); in nvkm_cgrp_ectx_get() 100 if (refcount_dec_and_test(&vctx->refs)) { in nvkm_cgrp_vctx_put() 130 refcount_inc(&vctx->refs); in nvkm_cgrp_vctx_get() 151 refcount_set(&vctx->refs, 1); in nvkm_cgrp_vctx_get()
|
| /linux/tools/testing/selftests/net/ |
| H A D | nl_netdev.py | 198 refs = sum([pp["inflight"] for pp in pp_list]) 199 ksft_eq(refs, 0) 210 refs = sum([pp["inflight"] for pp in pp_list if pp.get("ifindex") == nsim.ifindex]) 211 ksft_ge(refs, 1) 217 refs = sum([pp["inflight"] for pp in pp_list]) 218 ksft_eq(refs, 1)
|
| /linux/include/uapi/xen/ |
| H A D | gntdev.h | 70 struct ioctl_gntdev_grant_ref refs[1]; member 226 * Create a dma-buf [1] from grant references @refs of count @count provided 247 /* Number of grant references in @refs array. */ 256 __u32 refs[1]; member 281 * to the pages of that dma-buf into array @refs of size @count. 290 /* Number of grant references in @refs array. */ 298 __u32 refs[1]; member
|
| /linux/fs/btrfs/tests/ |
| H A D | qgroup-tests.c | 73 u64 refs; in add_tree_ref() 96 refs = btrfs_extent_refs(path->nodes[0], item); in add_tree_ref() 97 btrfs_set_extent_refs(path->nodes[0], item, refs + 1); in add_tree_ref() 151 u64 refs; in remove_extent_ref() 174 refs = btrfs_extent_refs(path->nodes[0], item); in remove_extent_ref() 175 btrfs_set_extent_refs(path->nodes[0], item, refs - 1); in remove_extent_ref() 221 * Since the test trans doesn't have the complicated delayed refs, in test_no_shared_qgroup() 321 test_msg("running qgroup multiple refs test"); in test_multiple_refs() 75 u64 refs; add_tree_ref() local 157 u64 refs; remove_extent_ref() local
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | port.c | 69 table->refs[i] = 0; in mlx4_init_mac_table() 83 table->refs[i] = 0; in mlx4_init_vlan_table() 118 if (table->refs[i] && in find_index() 212 dup_table->refs[index_at_port]) { in __mlx4_register_mac() 222 if (!table->refs[index_at_dup_port] || in __mlx4_register_mac() 231 if (!table->refs[i]) { in __mlx4_register_mac() 235 if (!dup_table->refs[i]) in __mlx4_register_mac() 245 ++table->refs[i]; in __mlx4_register_mac() 288 table->refs[free] = 1; in __mlx4_register_mac() 292 dup_table->refs[free] = 0; in __mlx4_register_mac() [all …]
|