| /linux/include/linux/ |
| H A D | leafops.h | 38 * softleaf_mk_none() - Create an empty ('none') leaf entry. 39 * Returns: empty leaf entry. 47 * softleaf_from_pte() - Obtain a leaf entry from a PTE entry. 48 * @pte: PTE entry. 50 * If @pte is present (therefore not a leaf entry) the function returns an empty 51 * leaf entry. Otherwise, it returns a leaf entry. 53 * Returns: Leaf entry. 70 * softleaf_to_pte() - Obtain a PTE entry fro 78 softleaf_to_pte(softleaf_t entry) softleaf_to_pte() argument 129 softleaf_is_none(softleaf_t entry) softleaf_is_none() argument 140 softleaf_type(softleaf_t entry) softleaf_type() argument 188 softleaf_is_swap(softleaf_t entry) softleaf_is_swap() argument 200 softleaf_is_migration_write(softleaf_t entry) softleaf_is_migration_write() argument 212 softleaf_is_migration_read(softleaf_t entry) softleaf_is_migration_read() argument 225 softleaf_is_migration_read_exclusive(softleaf_t entry) softleaf_is_migration_read_exclusive() argument 236 softleaf_is_migration(softleaf_t entry) softleaf_is_migration() argument 256 softleaf_is_device_private_write(softleaf_t entry) softleaf_is_device_private_write() argument 267 softleaf_is_device_private(softleaf_t entry) softleaf_is_device_private() argument 284 softleaf_is_device_exclusive(softleaf_t entry) softleaf_is_device_exclusive() argument 295 softleaf_is_hwpoison(softleaf_t entry) softleaf_is_hwpoison() argument 306 softleaf_is_marker(softleaf_t entry) softleaf_is_marker() argument 317 softleaf_to_marker(softleaf_t entry) softleaf_to_marker() argument 335 softleaf_has_pfn(softleaf_t entry) softleaf_has_pfn() argument 358 softleaf_to_pfn(softleaf_t entry) softleaf_to_pfn() argument 366 softleaf_migration_sync(softleaf_t entry,struct folio * folio) softleaf_migration_sync() argument 389 softleaf_to_page(softleaf_t entry) softleaf_to_page() argument 406 softleaf_to_folio(softleaf_t entry) softleaf_to_folio() argument 425 softleaf_is_poison_marker(softleaf_t entry) softleaf_is_poison_marker() argument 439 softleaf_is_guard_marker(softleaf_t entry) softleaf_is_guard_marker() argument 456 softleaf_is_uffd_wp_marker(softleaf_t entry) softleaf_is_uffd_wp_marker() argument 477 softleaf_is_migration_young(softleaf_t entry) softleaf_is_migration_young() argument 496 softleaf_is_migration_dirty(softleaf_t entry) softleaf_is_migration_dirty() argument 508 softleaf_is_migration_young(softleaf_t entry) softleaf_is_migration_young() argument 513 softleaf_is_migration_dirty(softleaf_t entry) softleaf_is_migration_dirty() argument 539 const softleaf_t entry = softleaf_from_pte(pte); pte_is_uffd_wp_marker() local 556 const softleaf_t entry = softleaf_from_pte(pte); pte_is_uffd_marker() local 621 const softleaf_t entry = softleaf_from_pmd(pmd); pmd_is_valid_softleaf() local [all...] |
| H A D | xarray.h | 28 * The bottom two bits of the entry determine how the XArray interprets 31 * 00: Pointer entry 32 * 10: Internal entry 33 * x1: Value entry or tagged pointer 41 * 256: Retry entry 42 * 257: Zero entry 52 * xa_mk_value() - Create an XArray entry from an integer. 56 * Return: An entry suitable for storing in the XArray. 65 * xa_to_value() - Get value stored in an XArray entry. 66 * @entry: XArray entry. [all …]
|
| /linux/net/atm/ |
| H A D | mpoa_caches.c | 38 in_cache_entry *entry; in in_cache_get() local 41 entry = client->in_cache; in in_cache_get() 42 while (entry != NULL) { in in_cache_get() 43 if (entry->ctrl_info.in_dst_ip == dst_ip) { in in_cache_get() 44 refcount_inc(&entry->use); in in_cache_get() 46 return entry; in in_cache_get() 48 entry = entry->next; in in_cache_get() 59 in_cache_entry *entry; in in_cache_get_with_mask() local 62 entry = client->in_cache; in in_cache_get_with_mask() 63 while (entry != NULL) { in in_cache_get_with_mask() [all …]
|
| H A D | lec.c | 106 static inline void lec_arp_hold(struct lec_arp_table *entry) in lec_arp_hold() argument 108 refcount_inc(&entry->usage); in lec_arp_hold() 111 static inline void lec_arp_put(struct lec_arp_table *entry) in lec_arp_put() argument 113 if (refcount_dec_and_test(&entry->usage)) in lec_arp_put() 114 kfree(entry); in lec_arp_put() 222 struct lec_arp_table *entry; in lec_start_xmit() local 293 entry = NULL; in lec_start_xmit() 294 vcc = lec_arp_resolve(priv, dst, is_rdesc, &entry); in lec_start_xmit() 295 pr_debug("%s:vcc:%p vcc_flags:%lx, entry:%p\n", in lec_start_xmit() 296 dev->name, vcc, vcc ? vcc->flags : 0, entry); in lec_start_xmit() [all …]
|
| /linux/sound/core/ |
| H A D | info.c | 54 struct snd_info_entry *entry; member 59 static void snd_info_clear_entries(struct snd_info_entry *entry); 73 static int alloc_info_private(struct snd_info_entry *entry, in alloc_info_private() argument 78 if (!entry || !entry->p) in alloc_info_private() 80 if (!try_module_get(entry->module)) in alloc_info_private() 84 module_put(entry->module); in alloc_info_private() 87 data->entry = entry; in alloc_info_private() 107 struct snd_info_entry *entry; in snd_info_entry_llseek() local 111 entry = data->entry; in snd_info_entry_llseek() 112 guard(mutex)(&entry->access); in snd_info_entry_llseek() [all …]
|
| /linux/fs/ |
| H A D | mbcache.c | 21 * identifies a cache entry. 24 * and a special "delete entry with given key-value pair" operation. Fixed 63 * mb_cache_entry_create - create entry in cache 64 * @cache - cache where the entry should be created 65 * @mask - gfp mask with which the entry should be allocated 66 * @key - key of the entry 67 * @value - value of the entry 68 * @reusable - is the entry reusable by others? 70 * Creates entry in @cache with key @key and value @value. The function returns 71 * -EBUSY if entry wit 77 struct mb_cache_entry *entry, *dup; mb_cache_entry_create() local 127 __mb_cache_entry_free(struct mb_cache * cache,struct mb_cache_entry * entry) __mb_cache_entry_free() argument 146 mb_cache_entry_wait_unused(struct mb_cache_entry * entry) mb_cache_entry_wait_unused() argument 153 __entry_find(struct mb_cache * cache,struct mb_cache_entry * entry,u32 key) __entry_find() argument 210 mb_cache_entry_find_next(struct mb_cache * cache,struct mb_cache_entry * entry) mb_cache_entry_find_next() argument 227 struct mb_cache_entry *entry; mb_cache_entry_get() local 257 struct mb_cache_entry *entry; mb_cache_entry_delete_or_get() local 287 mb_cache_entry_touch(struct mb_cache * cache,struct mb_cache_entry * entry) mb_cache_entry_touch() argument 305 struct mb_cache_entry *entry; mb_cache_shrink() local 407 struct mb_cache_entry *entry, *next; mb_cache_destroy() local [all...] |
| H A D | dax.c | 53 * for pages. We use one bit for locking, one bit for the entry size (PMD) 54 * and two more to tell us if the entry is a zero page or an empty entry that 57 * If the PMD bit isn't set the entry has size PAGE_SIZE, and if the ZERO_PAGE 58 * and EMPTY bits aren't set the entry is a normal DAX entry with a filesystem 67 static unsigned long dax_to_pfn(void *entry) 69 return xa_to_value(entry) >> DAX_SHIFT; in dax_to_pfn() 72 static struct folio *dax_to_folio(void *entry) 74 return page_folio(pfn_to_page(dax_to_pfn(entry))); in dax_to_folio() 68 dax_to_pfn(void * entry) dax_to_pfn() argument 73 dax_to_folio(void * entry) dax_to_folio() argument 83 dax_is_locked(void * entry) dax_is_locked() argument 88 dax_entry_order(void * entry) dax_entry_order() argument 95 dax_is_pmd_entry(void * entry) dax_is_pmd_entry() argument 100 dax_is_pte_entry(void * entry) dax_is_pte_entry() argument 105 dax_is_zero_entry(void * entry) dax_is_zero_entry() argument 110 dax_is_empty_entry(void * entry) dax_is_empty_entry() argument 119 dax_is_conflict(void * entry) dax_is_conflict() argument 148 dax_entry_waitqueue(struct xa_state * xas,void * entry,struct exceptional_entry_key * key) dax_entry_waitqueue() argument 185 dax_wake_entry(struct xa_state * xas,void * entry,enum dax_wake_mode mode) dax_wake_entry() argument 215 void *entry; get_next_unlocked_entry() local 247 wait_entry_unlocked_exclusive(struct xa_state * xas,void * entry) wait_entry_unlocked_exclusive() argument 278 wait_entry_unlocked(struct xa_state * xas,void * entry) wait_entry_unlocked() argument 299 put_unlocked_entry(struct xa_state * xas,void * entry,enum dax_wake_mode mode) put_unlocked_entry() argument 311 dax_unlock_entry(struct xa_state * xas,void * entry) dax_unlock_entry() argument 327 dax_lock_entry(struct xa_state * xas,void * entry) dax_lock_entry() argument 333 dax_entry_size(void * entry) dax_entry_size() argument 421 dax_folio_init(void * entry) dax_folio_init() argument 441 dax_associate_entry(void * entry,struct address_space * mapping,struct vm_area_struct * vma,unsigned long address,bool shared) dax_associate_entry() argument 468 dax_disassociate_entry(void * entry,struct address_space * mapping,bool trunc) dax_disassociate_entry() argument 479 dax_busy_page(void * entry) dax_busy_page() argument 503 void *entry; dax_lock_folio() local 571 void *entry; dax_lock_mapping_entry() local 655 void *entry; grab_mapping_entry() local 757 void *entry; dax_layout_busy_page_range() local 820 void *entry; __dax_invalidate_entry() local 845 void *entry; __dax_clear_dirty_range() local 891 void *entry; dax_delete_mapping_range() local 1043 dax_insert_entry(struct xa_state * xas,struct vm_fault * vmf,const struct iomap_iter * iter,void * entry,unsigned long pfn,unsigned long flags) dax_insert_entry() argument 1101 dax_writeback_one(struct xa_state * xas,struct dax_device * dax_dev,struct address_space * mapping,void * entry) dax_writeback_one() argument 1205 void *entry; dax_writeback_mapping_range() local 1359 dax_load_hole(struct xa_state * xas,struct vm_fault * vmf,const struct iomap_iter * iter,void ** entry) dax_load_hole() argument 1375 dax_pmd_load_hole(struct xa_state * xas,struct vm_fault * vmf,const struct iomap_iter * iter,void ** entry) dax_pmd_load_hole() argument 1399 dax_pmd_load_hole(struct xa_state * xas,struct vm_fault * vmf,const struct iomap_iter * iter,void ** entry) dax_pmd_load_hole() argument 1807 dax_fault_iter(struct vm_fault * vmf,const struct iomap_iter * iter,unsigned long * pfnp,struct xa_state * xas,void ** entry,bool pmd) dax_fault_iter() argument 1874 void *entry; dax_iomap_pte_fault() local 1984 void *entry; dax_iomap_pmd_fault() local 2099 void *entry; dax_insert_pfn_mkwrite() local [all...] |
| /linux/drivers/firmware/ |
| H A D | memmap.c | 22 * Firmware map entry. Because firmware memory maps are flat and not 34 struct list_head list; /* entry for the linked list */ 35 struct kobject kobj; /* kobject for each entry */ 43 static ssize_t start_show(struct firmware_map_entry *entry, char *buf); 44 static ssize_t end_show(struct firmware_map_entry *entry, char *buf); 45 static ssize_t type_show(struct firmware_map_entry *entry, char *buf); 56 ssize_t (*show)(struct firmware_map_entry *entry, char *buf); 64 * These are default attributes that are added for every memmap entry. 85 * map entry is allocated by bootmem, we need to remember the storage and 100 struct firmware_map_entry *entry = to_memmap_entry(kobj); in release_firmware_map_entry() local [all …]
|
| /linux/drivers/net/dsa/sja1105/ |
| H A D | sja1105_static_config.c | 94 struct sja1105_avb_params_entry *entry = entry_ptr; in sja1105et_avb_params_entry_packing() local 96 sja1105_packing(buf, &entry->destmeta, 95, 48, size, op); in sja1105et_avb_params_entry_packing() 97 sja1105_packing(buf, &entry->srcmeta, 47, 0, size, op); in sja1105et_avb_params_entry_packing() 105 struct sja1105_avb_params_entry *entry = entry_ptr; in sja1105pqrs_avb_params_entry_packing() local 107 sja1105_packing(buf, &entry->cas_master, 126, 126, size, op); in sja1105pqrs_avb_params_entry_packing() 108 sja1105_packing(buf, &entry->destmeta, 125, 78, size, op); in sja1105pqrs_avb_params_entry_packing() 109 sja1105_packing(buf, &entry->srcmeta, 77, 30, size, op); in sja1105pqrs_avb_params_entry_packing() 117 struct sja1105_general_params_entry *entry = entry_ptr; in sja1105et_general_params_entry_packing() local 119 sja1105_packing(buf, &entry->vllupformat, 319, 319, size, op); in sja1105et_general_params_entry_packing() 120 sja1105_packing(buf, &entry->mirr_ptacu, 318, 318, size, op); in sja1105et_general_params_entry_packing() [all …]
|
| /linux/drivers/media/platform/nvidia/tegra-vde/ |
| H A D | dmabuf-cache.c | 32 static void tegra_vde_release_entry(struct tegra_vde_cache_entry *entry) in tegra_vde_release_entry() argument 34 struct dma_buf *dmabuf = entry->a->dmabuf; in tegra_vde_release_entry() 36 WARN_ON_ONCE(entry->refcnt); in tegra_vde_release_entry() 38 if (entry->vde->domain) in tegra_vde_release_entry() 39 tegra_vde_iommu_unmap(entry->vde, entry->iova); in tegra_vde_release_entry() 41 dma_buf_unmap_attachment_unlocked(entry->a, entry->sgt, entry->dma_dir); in tegra_vde_release_entry() 42 dma_buf_detach(dmabuf, entry->a); in tegra_vde_release_entry() 45 list_del(&entry->list); in tegra_vde_release_entry() 46 kfree(entry); in tegra_vde_release_entry() 51 struct tegra_vde_cache_entry *entry; in tegra_vde_delayed_unmap() local [all …]
|
| /linux/tools/perf/ui/browsers/ |
| H A D | annotate-data.c | 52 struct browser_entry *entry, in get_member_overhead() argument 55 struct annotated_member *member = entry->data; in get_member_overhead() 70 update_hist_entry(&entry->hists[k++], &h->addr[offset]); in get_member_overhead() 83 struct browser_entry *entry; in add_child_entries() local 86 entry = zalloc(sizeof(*entry)); in add_child_entries() 87 if (entry == NULL) in add_child_entries() 90 entry->hists = calloc(browser->nr_events, sizeof(*entry->hists)); in add_child_entries() 91 if (entry->hists == NULL) { in add_child_entries() 92 free(entry); in add_child_entries() 96 entry->data = member; in add_child_entries() [all …]
|
| /linux/net/netlabel/ |
| H A D | netlabel_domainhash.c | 55 * netlbl_domhsh_free_entry - Frees a domain hash table entry 56 * @entry: the entry's RCU field 60 * function so that the memory allocated to a hash table entry can be released 64 static void netlbl_domhsh_free_entry(struct rcu_head *entry) in netlbl_domhsh_free_entry() argument 74 ptr = container_of(entry, struct netlbl_dom_map, rcu); in netlbl_domhsh_free_entry() 125 * netlbl_domhsh_search - Search for a domain entry 131 * entry if found, otherwise NULL is returned. @family may be %AF_UNSPEC 159 * netlbl_domhsh_search_def - Search for a domain entry 165 * entry if an exact match is found, if an exact match is not present in the 166 * hash table then the default entry is returned if valid otherwise NULL is [all …]
|
| H A D | netlabel_addrlist.c | 36 * netlbl_af4list_search - Search for a matching IPv4 address entry 41 * Searches the IPv4 address list given by @head. If a matching address entry 59 * netlbl_af4list_search_exact - Search for an exact IPv4 address entry 86 * netlbl_af6list_search - Search for a matching IPv6 address entry 91 * Searches the IPv6 address list given by @head. If a matching address entry 110 * netlbl_af6list_search_exact - Search for an exact IPv6 address entry 138 * netlbl_af4list_add - Add a new IPv4 address entry to a list 139 * @entry: address entry 143 * Add a new address entry to the list pointed to by @head. On success zero is 148 int netlbl_af4list_add(struct netlbl_af4list *entry, struct list_head *head) in netlbl_af4list_add() argument [all …]
|
| /linux/drivers/gpu/drm/amd/pm/powerplay/smumgr/ |
| H A D | vega10_smumgr.c | 46 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, in vega10_copy_table_from_smc() 48 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, in vega10_copy_table_from_smc() 52 upper_32_bits(priv->smu_tables.entry[table_id].mc_addr), in vega10_copy_table_from_smc() 56 lower_32_bits(priv->smu_tables.entry[table_id].mc_addr), in vega10_copy_table_from_smc() 60 priv->smu_tables.entry[table_id].table_id, in vega10_copy_table_from_smc() 65 memcpy(table, priv->smu_tables.entry[table_id].table, in vega10_copy_table_from_smc() 66 priv->smu_tables.entry[table_id].size); in vega10_copy_table_from_smc() 85 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, in vega10_copy_table_to_smc() 87 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, in vega10_copy_table_to_smc() 90 memcpy(priv->smu_tables.entry[table_id].table, table, in vega10_copy_table_to_smc() [all …]
|
| H A D | vega12_smumgr.c | 49 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, in vega12_copy_table_from_smc() 51 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, in vega12_copy_table_from_smc() 55 upper_32_bits(priv->smu_tables.entry[table_id].mc_addr), in vega12_copy_table_from_smc() 60 lower_32_bits(priv->smu_tables.entry[table_id].mc_addr), in vega12_copy_table_from_smc() 73 memcpy(table, priv->smu_tables.entry[table_id].table, in vega12_copy_table_from_smc() 74 priv->smu_tables.entry[table_id].size); in vega12_copy_table_from_smc() 93 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, in vega12_copy_table_to_smc() 95 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, in vega12_copy_table_to_smc() 98 memcpy(priv->smu_tables.entry[table_id].table, table, in vega12_copy_table_to_smc() 99 priv->smu_tables.entry[table_id].size); in vega12_copy_table_to_smc() [all …]
|
| /linux/drivers/net/ethernet/rocker/ |
| H A D | rocker_ofdpa.c | 93 struct hlist_node entry; member 103 struct hlist_node entry; member 130 struct hlist_node entry; 142 struct hlist_node entry; 149 struct hlist_node entry; 307 const struct ofdpa_flow_tbl_entry *entry) in ofdpa_cmd_flow_tbl_add_ig_port() 310 entry->key.ig_port.in_pport)) in ofdpa_cmd_flow_tbl_add_ig_port() 313 entry->key.ig_port.in_pport_mask)) in ofdpa_cmd_flow_tbl_add_ig_port() 316 entry->key.ig_port.goto_tbl)) in ofdpa_cmd_flow_tbl_add_ig_port() 324 const struct ofdpa_flow_tbl_entry *entry) in ofdpa_cmd_flow_tbl_add_vlan() 129 struct hlist_node entry; global() member 141 struct hlist_node entry; global() member 148 struct hlist_node entry; global() member 306 ofdpa_cmd_flow_tbl_add_ig_port(struct rocker_desc_info * desc_info,const struct ofdpa_flow_tbl_entry * entry) ofdpa_cmd_flow_tbl_add_ig_port() argument 323 ofdpa_cmd_flow_tbl_add_vlan(struct rocker_desc_info * desc_info,const struct ofdpa_flow_tbl_entry * entry) ofdpa_cmd_flow_tbl_add_vlan() argument 347 ofdpa_cmd_flow_tbl_add_term_mac(struct rocker_desc_info * desc_info,const struct ofdpa_flow_tbl_entry * entry) ofdpa_cmd_flow_tbl_add_term_mac() argument 383 ofdpa_cmd_flow_tbl_add_ucast_routing(struct rocker_desc_info * desc_info,const struct ofdpa_flow_tbl_entry * entry) ofdpa_cmd_flow_tbl_add_ucast_routing() argument 406 ofdpa_cmd_flow_tbl_add_bridge(struct rocker_desc_info * desc_info,const struct ofdpa_flow_tbl_entry * entry) ofdpa_cmd_flow_tbl_add_bridge() argument 440 ofdpa_cmd_flow_tbl_add_acl(struct rocker_desc_info * desc_info,const struct ofdpa_flow_tbl_entry * entry) ofdpa_cmd_flow_tbl_add_acl() argument 509 const struct ofdpa_flow_tbl_entry *entry = priv; ofdpa_cmd_flow_tbl_add() local 566 const struct ofdpa_flow_tbl_entry *entry = priv; ofdpa_cmd_flow_tbl_del() local 584 ofdpa_cmd_group_tbl_add_l2_interface(struct rocker_desc_info * desc_info,struct ofdpa_group_tbl_entry * entry) ofdpa_cmd_group_tbl_add_l2_interface() argument 598 ofdpa_cmd_group_tbl_add_l2_rewrite(struct rocker_desc_info * desc_info,const struct ofdpa_group_tbl_entry * entry) ofdpa_cmd_group_tbl_add_l2_rewrite() argument 621 ofdpa_cmd_group_tbl_add_group_ids(struct rocker_desc_info * desc_info,const struct ofdpa_group_tbl_entry * entry) ofdpa_cmd_group_tbl_add_group_ids() argument 647 ofdpa_cmd_group_tbl_add_l3_unicast(struct rocker_desc_info * desc_info,const struct ofdpa_group_tbl_entry * entry) ofdpa_cmd_group_tbl_add_l3_unicast() argument 675 struct ofdpa_group_tbl_entry *entry = priv; ofdpa_cmd_group_tbl_add() local 720 const struct ofdpa_group_tbl_entry *entry = priv; ofdpa_cmd_group_tbl_del() local 827 ofdpa_flow_tbl_do(struct ofdpa_port * ofdpa_port,int flags,struct ofdpa_flow_tbl_entry * entry) ofdpa_flow_tbl_do() argument 839 struct ofdpa_flow_tbl_entry *entry; ofdpa_flow_tbl_ig_port() local 861 struct ofdpa_flow_tbl_entry *entry; ofdpa_flow_tbl_vlan() local 887 struct ofdpa_flow_tbl_entry *entry; ofdpa_flow_tbl_term_mac() local 923 struct ofdpa_flow_tbl_entry *entry; ofdpa_flow_tbl_bridge() local 977 struct ofdpa_flow_tbl_entry *entry; ofdpa_flow_tbl_ucast4_routing() local 1007 struct ofdpa_flow_tbl_entry *entry; ofdpa_flow_tbl_acl() local 1062 ofdpa_group_tbl_entry_free(struct ofdpa_group_tbl_entry * entry) ofdpa_group_tbl_entry_free() argument 1139 ofdpa_group_tbl_do(struct ofdpa_port * ofdpa_port,int flags,struct ofdpa_group_tbl_entry * entry) ofdpa_group_tbl_do() argument 1151 struct ofdpa_group_tbl_entry *entry; ofdpa_group_l2_interface() local 1167 struct ofdpa_group_tbl_entry *entry; ofdpa_group_l2_fan_out() local 1200 struct ofdpa_group_tbl_entry *entry; ofdpa_group_l3_unicast() local 1232 ofdpa_neigh_add(struct ofdpa * ofdpa,struct ofdpa_neigh_tbl_entry * entry) ofdpa_neigh_add() argument 1240 ofdpa_neigh_del(struct ofdpa_neigh_tbl_entry * entry) ofdpa_neigh_del() argument 1248 ofdpa_neigh_update(struct ofdpa_neigh_tbl_entry * entry,const u8 * eth_dst,bool ttl_check) ofdpa_neigh_update() argument 1263 struct ofdpa_neigh_tbl_entry *entry; ofdpa_port_ipv4_neigh() local 1380 struct ofdpa_neigh_tbl_entry *entry; ofdpa_port_ipv4_nh() local 1987 struct ofdpa_fdb_tbl_entry *entry; ofdpa_fdb_cleanup() local 2230 struct ofdpa_internal_vlan_tbl_entry *entry; ofdpa_port_internal_vlan_id_get() local [all...] |
| /linux/security/integrity/ima/ |
| H A D | ima_policy.c | 383 static void ima_lsm_free_rule(struct ima_rule_entry *entry) in ima_lsm_free_rule() argument 388 ima_filter_rule_free(entry->lsm[i].rule); in ima_lsm_free_rule() 389 kfree(entry->lsm[i].args_p); in ima_lsm_free_rule() 393 static void ima_free_rule(struct ima_rule_entry *entry) in ima_free_rule() argument 395 if (!entry) in ima_free_rule() 399 * entry->template->fields may be allocated in ima_parse_rule() but that in ima_free_rule() 403 kfree(entry->fsname); in ima_free_rule() 404 kfree(entry->fs_subtype); in ima_free_rule() 405 ima_free_rule_opt_list(entry->keyrings); in ima_free_rule() 406 ima_lsm_free_rule(entry); in ima_free_rule() 410 ima_lsm_copy_rule(struct ima_rule_entry * entry,gfp_t gfp) ima_lsm_copy_rule() argument 444 ima_lsm_update_rule(struct ima_rule_entry * entry) ima_lsm_update_rule() argument 468 ima_rule_contains_lsm_cond(struct ima_rule_entry * entry) ima_rule_contains_lsm_cond() argument 486 struct ima_rule_entry *entry, *e; ima_lsm_update_rules() local 758 struct ima_rule_entry *entry; ima_match_policy() local 826 struct ima_rule_entry *entry; ima_update_policy_flags() local 883 struct ima_rule_entry *entry; add_rules() local 1144 ima_lsm_rule_init(struct ima_rule_entry * entry,substring_t * args,int lsm_rule,int audit_type) ima_lsm_rule_init() argument 1256 ima_validate_rule(struct ima_rule_entry * entry) ima_validate_rule() argument 1424 ima_parse_rule(char * rule,struct ima_rule_entry * entry) ima_parse_rule() argument 1967 struct ima_rule_entry *entry; ima_parse_add_rule() local 2010 struct ima_rule_entry *entry, *tmp; ima_delete_rules() local 2040 struct ima_rule_entry *entry; ima_policy_start() local 2057 struct ima_rule_entry *entry = v; ima_policy_next() local 2114 struct ima_rule_entry *entry = v; ima_policy_show() local 2338 struct ima_rule_entry *entry; ima_appraise_signature() local [all...] |
| /linux/tools/perf/util/ |
| H A D | block-range.c | 19 struct block_range *entry = rb_entry(rb, struct block_range, node); in block_range__debug() local 21 assert(old < entry->start); in block_range__debug() 22 assert(entry->start <= entry->end); /* single instruction block; jump to a jump */ in block_range__debug() 24 old = entry->end; in block_range__debug() 33 struct block_range *entry; in block_range__find() local 37 entry = rb_entry(parent, struct block_range, node); in block_range__find() 39 if (addr < entry->start) in block_range__find() 41 else if (addr > entry->end) in block_range__find() 44 return entry; in block_range__find() 81 struct block_range *next, *entry = NULL; in block_range__create() local [all …]
|
| /linux/drivers/acpi/ |
| H A D | nvs.c | 97 struct nvs_page *entry, *next; in suspend_nvs_register() local 105 entry = kzalloc_obj(struct nvs_page); in suspend_nvs_register() 106 if (!entry) in suspend_nvs_register() 109 list_add_tail(&entry->node, &nvs_list); in suspend_nvs_register() 110 entry->phys_start = start; in suspend_nvs_register() 112 entry->size = (size < nr_bytes) ? size : nr_bytes; in suspend_nvs_register() 114 start += entry->size; in suspend_nvs_register() 115 size -= entry->size; in suspend_nvs_register() 120 list_for_each_entry_safe(entry, next, &nvs_list, node) { in suspend_nvs_register() 121 list_del(&entry->node); in suspend_nvs_register() [all …]
|
| /linux/lib/ |
| H A D | xarray.c | 24 * @index is the index of the entry being operated on 30 * @entry refers to something stored in a slot in the xarray 120 * xas_squash_marks() - Merge all marks to the first entry 123 * Set a mark on the first entry if any entry has it set. Clear marks on 184 void *entry; in xas_start() local 191 entry = xa_head(xas->xa); in xas_start() 192 if (!xa_is_node(entry)) { in xas_start() 196 if ((xas->xa_index >> xa_to_node(entry)->shift) > XA_CHUNK_MASK) in xas_start() 201 return entry; in xas_start() 208 void *entry = xa_entry(xas->xa, node, offset); in xas_descend() local [all …]
|
| /linux/fs/squashfs/ |
| H A D | cache.c | 56 struct squashfs_cache_entry *entry; in squashfs_cache_get() local 62 if (cache->entry[i].block == block) { in squashfs_cache_get() 84 * At least one unused cache entry. A simple in squashfs_cache_get() 85 * round-robin strategy is used to choose the entry to in squashfs_cache_get() 90 if (cache->entry[i].refcount == 0) in squashfs_cache_get() 96 entry = &cache->entry[i]; in squashfs_cache_get() 99 * Initialise chosen cache entry, and fill it in from in squashfs_cache_get() 103 entry->block = block; in squashfs_cache_get() 104 entry->refcount = 1; in squashfs_cache_get() 105 entry->pending = 1; in squashfs_cache_get() [all …]
|
| /linux/drivers/infiniband/core/ |
| H A D | ib_core_uverbs.c | 20 * @entry: entry into the mmap_xa that needs to be linked with 36 struct rdma_user_mmap_entry *entry) in rdma_umap_priv_init() argument 41 if (entry) { in rdma_umap_priv_init() 42 kref_get(&entry->ref); in rdma_umap_priv_init() 43 priv->entry = entry; in rdma_umap_priv_init() 62 * @entry: mmap_entry retrieved from rdma_user_mmap_entry_get(), or NULL 73 struct rdma_user_mmap_entry *entry) in rdma_user_mmap_io() argument 100 rdma_umap_priv_init(priv, vma, entry); in rdma_user_mmap_io() 124 struct rdma_user_mmap_entry *entry; rdma_user_mmap_entry_get_pgoff() local 168 struct rdma_user_mmap_entry *entry; rdma_user_mmap_entry_get() local 185 struct rdma_user_mmap_entry *entry = rdma_user_mmap_entry_free() local 218 rdma_user_mmap_entry_put(struct rdma_user_mmap_entry * entry) rdma_user_mmap_entry_put() argument 234 rdma_user_mmap_entry_remove(struct rdma_user_mmap_entry * entry) rdma_user_mmap_entry_remove() argument 284 rdma_user_mmap_entry_insert_range(struct ib_ucontext * ucontext,struct rdma_user_mmap_entry * entry,size_t length,u32 min_pgoff,u32 max_pgoff) rdma_user_mmap_entry_insert_range() argument 385 rdma_user_mmap_entry_insert(struct ib_ucontext * ucontext,struct rdma_user_mmap_entry * entry,size_t length) rdma_user_mmap_entry_insert() argument [all...] |
| /linux/kernel/dma/ |
| H A D | debug.c | 171 static inline void dump_entry_trace(struct dma_debug_entry *entry) in dump_entry_trace() argument 174 if (entry) { in dump_entry_trace() 176 stack_trace_print(entry->stack_entries, entry->stack_len, 0); in dump_entry_trace() 222 #define err_printk(dev, entry, format, arg...) do { \ argument 229 dump_entry_trace(entry); \ 241 static int hash_fn(struct dma_debug_entry *entry) in hash_fn() argument 247 return (entry->dev_addr >> HASH_FN_SHIFT) & HASH_FN_MASK; in hash_fn() 253 static struct hash_bucket *get_hash_bucket(struct dma_debug_entry *entry, in get_hash_bucket() argument 257 int idx = hash_fn(entry); in get_hash_bucket() 295 * Search a given entry in the hash bucket list [all …]
|
| /linux/arch/x86/kvm/ |
| H A D | cpuid.c | 112 * If the index isn't significant, use the first entry with a in kvm_find_cpuid_entry2() 121 * Similarly, use the first matching entry if KVM is doing a in kvm_find_cpuid_entry2() 211 struct kvm_cpuid_entry2 *entry; in kvm_get_hypervisor_cpuid() local 215 entry = kvm_find_cpuid_entry(vcpu, base); in kvm_get_hypervisor_cpuid() 217 if (entry) { in kvm_get_hypervisor_cpuid() 220 signature[0] = entry->ebx; in kvm_get_hypervisor_cpuid() 221 signature[1] = entry->ecx; in kvm_get_hypervisor_cpuid() 222 signature[2] = entry->edx; in kvm_get_hypervisor_cpuid() 226 cpuid.limit = entry->eax; in kvm_get_hypervisor_cpuid() 281 struct kvm_cpuid_entry2 *entry, in kvm_update_feature_runtime() argument 329 struct kvm_cpuid_entry2 *entry; kvm_cpuid_has_hyperv() local 340 struct kvm_cpuid_entry2 *entry; guest_cpuid_is_amd_or_hygon() local 355 cpuid_get_reg_unsafe(struct kvm_cpuid_entry2 * entry,u32 reg) cpuid_get_reg_unsafe() argument 379 struct kvm_cpuid_entry2 *entry; kvm_vcpu_after_set_cpuid() local 677 struct kvm_cpuid_entry2 entry; raw_cpuid_get() local 1330 struct kvm_cpuid_entry2 *entry = get_next_cpuid(array); do_host_cpuid() local 1370 cpuid_func_emulated(struct kvm_cpuid_entry2 * entry,u32 func,bool include_partially_emulated) cpuid_func_emulated() argument 1418 struct kvm_cpuid_entry2 *entry; __do_cpuid_func() local 2091 struct kvm_cpuid_entry2 *entry; kvm_cpuid() local [all...] |
| /linux/drivers/net/ethernet/marvell/mvpp2/ |
| H A D | mvpp2_debugfs.c | 60 struct mvpp2_dbgfs_flow_tbl_entry *entry = s->private; in mvpp2_dbgfs_flow_flt_hits_show() local 62 u32 hits = mvpp2_cls_flow_hits(entry->priv, entry->id); in mvpp2_dbgfs_flow_flt_hits_show() 73 struct mvpp2_dbgfs_flow_entry *entry = s->private; in mvpp2_dbgfs_flow_dec_hits_show() local 75 u32 hits = mvpp2_cls_lookup_hits(entry->priv, entry->flow); in mvpp2_dbgfs_flow_dec_hits_show() 86 struct mvpp2_dbgfs_flow_entry *entry = s->private; in mvpp2_dbgfs_flow_type_show() local 90 f = mvpp2_cls_flow_get(entry->flow); in mvpp2_dbgfs_flow_type_show() 126 const struct mvpp2_dbgfs_flow_entry *entry = s->private; in mvpp2_dbgfs_flow_id_show() local 129 f = mvpp2_cls_flow_get(entry->flow); in mvpp2_dbgfs_flow_id_show() 142 struct mvpp2_dbgfs_port_flow_entry *entry = s->private; in mvpp2_dbgfs_port_flow_hash_opt_show() local 143 struct mvpp2_port *port = entry->port; in mvpp2_dbgfs_port_flow_hash_opt_show() [all …]
|