/linux/net/atm/ |
H A D | mpoa_caches.c | 38 in_cache_entry *entry; in in_cache_get() local 41 entry = client->in_cache; in in_cache_get() 42 while (entry != NULL) { in in_cache_get() 43 if (entry->ctrl_info.in_dst_ip == dst_ip) { in in_cache_get() 44 refcount_inc(&entry->use); in in_cache_get() 46 return entry; in in_cache_get() 48 entry = entry->next; in in_cache_get() 59 in_cache_entry *entry; in in_cache_get_with_mask() local 62 entry = client->in_cache; in in_cache_get_with_mask() 63 while (entry != NULL) { in in_cache_get_with_mask() [all …]
|
H A D | lec.c | 106 static inline void lec_arp_hold(struct lec_arp_table *entry) in lec_arp_hold() argument 108 refcount_inc(&entry->usage); in lec_arp_hold() 111 static inline void lec_arp_put(struct lec_arp_table *entry) in lec_arp_put() argument 113 if (refcount_dec_and_test(&entry->usage)) in lec_arp_put() 114 kfree(entry); in lec_arp_put() 211 struct lec_arp_table *entry; in lec_start_xmit() local 282 entry = NULL; in lec_start_xmit() 283 vcc = lec_arp_resolve(priv, dst, is_rdesc, &entry); in lec_start_xmit() 284 pr_debug("%s:vcc:%p vcc_flags:%lx, entry:%p\n", in lec_start_xmit() 285 dev->name, vcc, vcc ? vcc->flags : 0, entry); in lec_start_xmit() [all …]
|
/linux/sound/core/ |
H A D | info.c | 54 struct snd_info_entry *entry; member 59 static void snd_info_clear_entries(struct snd_info_entry *entry); 73 static int alloc_info_private(struct snd_info_entry *entry, in alloc_info_private() argument 78 if (!entry || !entry->p) in alloc_info_private() 80 if (!try_module_get(entry->module)) in alloc_info_private() 84 module_put(entry->module); in alloc_info_private() 87 data->entry = entry; in alloc_info_private() 107 struct snd_info_entry *entry; in snd_info_entry_llseek() local 111 entry = data->entry; in snd_info_entry_llseek() 112 guard(mutex)(&entry->access); in snd_info_entry_llseek() [all …]
|
/linux/fs/ |
H A D | mbcache.c | 21 * identifies a cache entry. 24 * and a special "delete entry with given key-value pair" operation. Fixed 63 * mb_cache_entry_create - create entry in cache 64 * @cache - cache where the entry should be created 65 * @mask - gfp mask with which the entry should be allocated 66 * @key - key of the entry 67 * @value - value of the entry 68 * @reusable - is the entry reusable by others? 70 * Creates entry in @cache with key @key and value @value. The function returns 71 * -EBUSY if entry with the same key and value already exists in cache. [all …]
|
H A D | dax.c | 55 * for pages. We use one bit for locking, one bit for the entry size (PMD) 56 * and two more to tell us if the entry is a zero page or an empty entry that 59 * If the PMD bit isn't set the entry has size PAGE_SIZE, and if the ZERO_PAGE 60 * and EMPTY bits aren't set the entry is a normal DAX entry with a filesystem 69 static unsigned long dax_to_pfn(void *entry) in dax_to_pfn() argument 71 return xa_to_value(entry) >> DAX_SHIFT; in dax_to_pfn() 79 static bool dax_is_locked(void *entry) in dax_is_locked() argument 81 return xa_to_value(entry) & DAX_LOCKED; in dax_is_locked() 84 static unsigned int dax_entry_order(void *entry) in dax_entry_order() argument 86 if (xa_to_value(entry) & DAX_PMD) in dax_entry_order() [all …]
|
/linux/drivers/scsi/libfc/ |
H A D | fc_encode.h | 168 static inline void fc_ct_ms_fill_attr(struct fc_fdmi_attr_entry *entry, in fc_ct_ms_fill_attr() argument 173 copied = strscpy(entry->value, in, len); in fc_ct_ms_fill_attr() 175 memset(entry->value + copied + 1, 0, len - copied - 1); in fc_ct_ms_fill_attr() 194 struct fc_fdmi_attr_entry *entry; in fc_ct_ms_fill() local 243 entry = (struct fc_fdmi_attr_entry *)hba_attrs->attr; in fc_ct_ms_fill() 248 &entry->type); in fc_ct_ms_fill() 249 put_unaligned_be16(len, &entry->len); in fc_ct_ms_fill() 251 (__be64 *)&entry->value); in fc_ct_ms_fill() 254 entry = (struct fc_fdmi_attr_entry *)((char *)entry->value + in fc_ct_ms_fill() 259 &entry->type); in fc_ct_ms_fill() [all …]
|
/linux/tools/perf/ui/browsers/ |
H A D | annotate-data.c | 52 struct browser_entry *entry, in get_member_overhead() argument 55 struct annotated_member *member = entry->data; in get_member_overhead() 70 update_hist_entry(&entry->hists[k++], &h->addr[offset]); in get_member_overhead() 83 struct browser_entry *entry; in add_child_entries() local 86 entry = zalloc(sizeof(*entry)); in add_child_entries() 87 if (entry == NULL) in add_child_entries() 90 entry->hists = calloc(browser->nr_events, sizeof(*entry->hists)); in add_child_entries() 91 if (entry->hists == NULL) { in add_child_entries() 92 free(entry); in add_child_entries() 96 entry->data = member; in add_child_entries() [all …]
|
/linux/drivers/media/platform/nvidia/tegra-vde/ |
H A D | dmabuf-cache.c | 32 static void tegra_vde_release_entry(struct tegra_vde_cache_entry *entry) in tegra_vde_release_entry() argument 34 struct dma_buf *dmabuf = entry->a->dmabuf; in tegra_vde_release_entry() 36 WARN_ON_ONCE(entry->refcnt); in tegra_vde_release_entry() 38 if (entry->vde->domain) in tegra_vde_release_entry() 39 tegra_vde_iommu_unmap(entry->vde, entry->iova); in tegra_vde_release_entry() 41 dma_buf_unmap_attachment_unlocked(entry->a, entry->sgt, entry->dma_dir); in tegra_vde_release_entry() 42 dma_buf_detach(dmabuf, entry->a); in tegra_vde_release_entry() 45 list_del(&entry->list); in tegra_vde_release_entry() 46 kfree(entry); in tegra_vde_release_entry() 51 struct tegra_vde_cache_entry *entry; in tegra_vde_delayed_unmap() local [all …]
|
/linux/include/linux/ |
H A D | xarray.h | 28 * The bottom two bits of the entry determine how the XArray interprets 31 * 00: Pointer entry 32 * 10: Internal entry 33 * x1: Value entry or tagged pointer 41 * 256: Retry entry 42 * 257: Zero entry 52 * xa_mk_value() - Create an XArray entry from an integer. 56 * Return: An entry suitable for storing in the XArray. in xa_mk_value() 65 * xa_to_value() - Get value stored in an XArray entry. 66 * @entry 67 xa_to_value(const void * entry) xa_to_value() argument 79 xa_is_value(const void * entry) xa_is_value() argument 112 xa_untag_pointer(void * entry) xa_untag_pointer() argument 127 xa_pointer_tag(void * entry) xa_pointer_tag() argument 157 xa_to_internal(const void * entry) xa_to_internal() argument 169 xa_is_internal(const void * entry) xa_is_internal() argument 185 xa_is_zero(const void * entry) xa_is_zero() argument 201 xa_is_err(const void * entry) xa_is_err() argument 219 xa_err(void * entry) xa_err() argument 447 xa_for_each_range(xa,index,entry,start,last) global() argument 476 xa_for_each_start(xa,index,entry,start) global() argument 500 xa_for_each(xa,index,entry) global() argument 527 xa_for_each_marked(xa,index,entry,filter) global() argument 586 xa_store_bh(struct xarray * xa,unsigned long index,void * entry,gfp_t gfp) xa_store_bh() argument 613 xa_store_irq(struct xarray * xa,unsigned long index,void * entry,gfp_t gfp) xa_store_irq() argument 640 void *entry; xa_erase_bh() local 664 void *entry; xa_erase_irq() local 689 xa_cmpxchg(struct xarray * xa,unsigned long index,void * old,void * entry,gfp_t gfp) xa_cmpxchg() argument 717 xa_cmpxchg_bh(struct xarray * xa,unsigned long index,void * old,void * entry,gfp_t gfp) xa_cmpxchg_bh() argument 745 xa_cmpxchg_irq(struct xarray * xa,unsigned long index,void * old,void * entry,gfp_t gfp) xa_cmpxchg_irq() argument 775 xa_insert(struct xarray * xa,unsigned long index,void * entry,gfp_t gfp) xa_insert() argument 805 xa_insert_bh(struct xarray * xa,unsigned long index,void * entry,gfp_t gfp) xa_insert_bh() argument 835 xa_insert_irq(struct xarray * xa,unsigned long index,void * entry,gfp_t gfp) xa_insert_irq() argument 868 xa_alloc(struct xarray * xa,u32 * id,void * entry,struct xa_limit limit,gfp_t gfp) xa_alloc() argument 901 xa_alloc_bh(struct xarray * xa,u32 * id,void * entry,struct xa_limit limit,gfp_t gfp) xa_alloc_bh() argument 934 xa_alloc_irq(struct xarray * xa,u32 * id,void * entry,struct xa_limit limit,gfp_t gfp) xa_alloc_irq() argument 970 xa_alloc_cyclic(struct xarray * xa,u32 * id,void * entry,struct xa_limit limit,u32 * next,gfp_t gfp) xa_alloc_cyclic() argument 1007 xa_alloc_cyclic_bh(struct xarray * xa,u32 * id,void * entry,struct xa_limit limit,u32 * next,gfp_t gfp) xa_alloc_cyclic_bh() argument 1044 xa_alloc_cyclic_irq(struct xarray * xa,u32 * id,void * entry,struct xa_limit limit,u32 * next,gfp_t gfp) xa_alloc_cyclic_irq() argument 1252 xa_to_node(const void * entry) xa_to_node() argument 1258 xa_is_node(const void * entry) xa_is_node() argument 1270 xa_to_sibling(const void * entry) xa_to_sibling() argument 1281 xa_is_sibling(const void * entry) xa_is_sibling() argument 1295 xa_is_retry(const void * entry) xa_is_retry() argument 1306 xa_is_advanced(const void * entry) xa_is_advanced() argument 1522 xas_retry(struct xa_state * xas,const void * entry) xas_retry() argument 1559 xas_split(struct xa_state * xas,void * entry,unsigned int order) xas_split() argument 1565 xas_split_alloc(struct xa_state * xas,void * entry,unsigned int order,gfp_t gfp) xas_split_alloc() argument 1588 void *entry; xas_reload() local 1691 void *entry; xas_next_entry() local 1749 void *entry; xas_next_marked() local 1788 xas_for_each(xas,entry,max) global() argument 1806 xas_for_each_marked(xas,entry,max,mark) global() argument 1822 xas_for_each_conflict(xas,entry) global() argument [all...] |
H A D | bpf_mprog.h | 23 * struct bpf_mprog_entry *entry, *entry_new; 27 * // fetch active @entry from attach location 29 * ret = bpf_mprog_attach(entry, &entry_new, [...]); 31 * if (entry != entry_new) { 32 * // swap @entry to @entry_new at attach location 33 * // ensure there are no inflight users of @entry: 36 * bpf_mprog_commit(entry); 44 * struct bpf_mprog_entry *entry, *entry_new; 48 * // fetch active @entry from attach location 50 * ret = bpf_mprog_detach(entry, &entry_new, [...]); [all …]
|
/linux/drivers/firmware/broadcom/ |
H A D | bcm47xx_sprom.c | 175 #define ENTRY(_revmask, _type, _prefix, _name, _val, _allset, _fallback) \ macro 195 ENTRY(0xfffffffe, u16, pre, "devid", dev_id, 0, fallback); in bcm47xx_sprom_fill_auto() 197 ENTRY(0xfffffffe, u16, pre, "boardrev", board_rev, 0, true); in bcm47xx_sprom_fill_auto() 198 ENTRY(0xfffffffe, u32, pre, "boardflags", boardflags, 0, fb); in bcm47xx_sprom_fill_auto() 199 ENTRY(0xfffffff0, u32, pre, "boardflags2", boardflags2, 0, fb); in bcm47xx_sprom_fill_auto() 200 ENTRY(0xfffff800, u32, pre, "boardflags3", boardflags3, 0, fb); in bcm47xx_sprom_fill_auto() 201 ENTRY(0x00000002, u16, pre, "boardflags", boardflags_lo, 0, fb); in bcm47xx_sprom_fill_auto() 202 ENTRY(0xfffffffc, u16, pre, "boardtype", board_type, 0, true); in bcm47xx_sprom_fill_auto() 203 ENTRY(0xfffffffe, u16, pre, "boardnum", board_num, 0, fb); in bcm47xx_sprom_fill_auto() 204 ENTRY(0x00000002, u8, pre, "cc", country_code, 0, fb); in bcm47xx_sprom_fill_auto() [all …]
|
/linux/arch/x86/kernel/ |
H A D | asm-offsets_64.c | 28 #define ENTRY(entry) OFFSET(pt_regs_ ## entry, pt_regs, entry) in main() argument 29 ENTRY(bx); in main() 30 ENTRY(cx); in main() 31 ENTRY(dx); in main() 32 ENTRY(sp); in main() 33 ENTRY(bp); in main() 34 ENTRY(si); in main() 35 ENTRY(di); in main() 36 ENTRY(r8); in main() 37 ENTRY(r9); in main() [all …]
|
/linux/fs/squashfs/ |
H A D | cache.c | 56 struct squashfs_cache_entry *entry; in squashfs_cache_get() local 62 if (cache->entry[i].block == block) { in squashfs_cache_get() 84 * At least one unused cache entry. A simple in squashfs_cache_get() 85 * round-robin strategy is used to choose the entry to in squashfs_cache_get() 90 if (cache->entry[i].refcount == 0) in squashfs_cache_get() 96 entry = &cache->entry[i]; in squashfs_cache_get() 99 * Initialise chosen cache entry, and fill it in from in squashfs_cache_get() 103 entry->block = block; in squashfs_cache_get() 104 entry in squashfs_cache_get() 173 squashfs_cache_put(struct squashfs_cache_entry * entry) squashfs_cache_put() argument 253 struct squashfs_cache_entry *entry = &cache->entry[i]; squashfs_cache_init() local 293 squashfs_copy_data(void * buffer,struct squashfs_cache_entry * entry,int offset,int length) squashfs_copy_data() argument 336 struct squashfs_cache_entry *entry; squashfs_read_metadata() local [all...] |
/linux/arch/sparc/lib/ |
H A D | Memcpy_utils.S | 8 ENTRY(__restore_asi_fp) 14 ENTRY(__restore_asi) 19 ENTRY(memcpy_retl_o2) 23 ENTRY(memcpy_retl_o2_plus_1) 27 ENTRY(memcpy_retl_o2_plus_3) 31 ENTRY(memcpy_retl_o2_plus_4) 35 ENTRY(memcpy_retl_o2_plus_5) 39 ENTRY(memcpy_retl_o2_plus_6) 43 ENTRY(memcpy_retl_o2_plus_7) 47 ENTRY(memcpy_retl_o2_plus_8) [all …]
|
/linux/drivers/gpu/drm/amd/pm/powerplay/smumgr/ |
H A D | vega10_smumgr.c | 46 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, in vega10_copy_table_from_smc() 48 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, in vega10_copy_table_from_smc() 52 upper_32_bits(priv->smu_tables.entry[table_id].mc_addr), in vega10_copy_table_from_smc() 56 lower_32_bits(priv->smu_tables.entry[table_id].mc_addr), in vega10_copy_table_from_smc() 60 priv->smu_tables.entry[table_id].table_id, in vega10_copy_table_from_smc() 65 memcpy(table, priv->smu_tables.entry[table_id].table, in vega10_copy_table_from_smc() 66 priv->smu_tables.entry[table_id].size); in vega10_copy_table_from_smc() 85 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, in vega10_copy_table_to_smc() 87 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, in vega10_copy_table_to_smc() 90 memcpy(priv->smu_tables.entry[table_id].table, table, in vega10_copy_table_to_smc() [all …]
|
/linux/net/netlabel/ |
H A D | netlabel_domainhash.c | 55 * netlbl_domhsh_free_entry - Frees a domain hash table entry 56 * @entry: the entry's RCU field 60 * function so that the memory allocated to a hash table entry can be released 64 static void netlbl_domhsh_free_entry(struct rcu_head *entry) in netlbl_domhsh_free_entry() argument 74 ptr = container_of(entry, struct netlbl_dom_map, rcu); in netlbl_domhsh_free_entry() 125 * netlbl_domhsh_search - Search for a domain entry 131 * entry if found, otherwise NULL is returned. @family may be %AF_UNSPEC 159 * netlbl_domhsh_search_def - Search for a domain entry 165 * entry if an exact match is found, if an exact match is not present in the 166 * hash table then the default entry is returned if valid otherwise NULL is [all …]
|
/linux/drivers/acpi/ |
H A D | nvs.c | 97 struct nvs_page *entry, *next; in suspend_nvs_register() local 105 entry = kzalloc(sizeof(struct nvs_page), GFP_KERNEL); in suspend_nvs_register() 106 if (!entry) in suspend_nvs_register() 109 list_add_tail(&entry->node, &nvs_list); in suspend_nvs_register() 110 entry->phys_start = start; in suspend_nvs_register() 112 entry->size = (size < nr_bytes) ? size : nr_bytes; in suspend_nvs_register() 114 start += entry->size; in suspend_nvs_register() 115 size -= entry->size; in suspend_nvs_register() 120 list_for_each_entry_safe(entry, next, &nvs_list, node) { in suspend_nvs_register() 121 list_del(&entry->node); in suspend_nvs_register() [all …]
|
/linux/tools/perf/util/ |
H A D | block-range.c | 19 struct block_range *entry = rb_entry(rb, struct block_range, node); in block_range__debug() local 21 assert(old < entry->start); in block_range__debug() 22 assert(entry->start <= entry->end); /* single instruction block; jump to a jump */ in block_range__debug() 24 old = entry->end; in block_range__debug() 33 struct block_range *entry; in block_range__find() local 37 entry = rb_entry(parent, struct block_range, node); in block_range__find() 39 if (addr < entry->start) in block_range__find() 41 else if (addr > entry->end) in block_range__find() 44 return entry; in block_range__find() 81 struct block_range *next, *entry = NULL; in block_range__create() local [all …]
|
/linux/drivers/infiniband/core/ |
H A D | ib_core_uverbs.c | 16 * @entry: entry into the mmap_xa that needs to be linked with 32 struct rdma_user_mmap_entry *entry) in rdma_umap_priv_init() argument 37 if (entry) { in rdma_umap_priv_init() 38 kref_get(&entry->ref); in rdma_umap_priv_init() 39 priv->entry = entry; in rdma_umap_priv_init() 58 * @entry: mmap_entry retrieved from rdma_user_mmap_entry_get(), or NULL 69 struct rdma_user_mmap_entry *entry) in rdma_user_mmap_io() argument 96 rdma_umap_priv_init(priv, vma, entry); in rdma_user_mmap_io() 102 * rdma_user_mmap_entry_get_pgoff() - Get an entry from the mmap_xa 111 * entry so that it won't be deleted from the xarray in the meantime. [all …]
|
/linux/lib/ |
H A D | xarray.c | 24 * @index is the index of the entry being operated on 30 * @entry refers to something stored in a slot in the xarray 120 * xas_squash_marks() - Merge all marks to the first entry 123 * Set a mark on the first entry if any entry has it set. Clear marks on 184 void *entry; in xas_start() 191 entry = xa_head(xas->xa); in xas_start() 192 if (!xa_is_node(entry)) { in xas_start() 196 if ((xas->xa_index >> xa_to_node(entry)->shift) > XA_CHUNK_MASK) in xas_start() 201 return entry; in xas_start() 183 void *entry; xas_start() local 206 void *entry = xa_entry(xas->xa, node, offset); xas_descend() local 237 void *entry = xas_start(xas); xas_load() local 430 max_index(void * entry) max_index() argument 443 void *entry; xas_shrink() local 529 void *entry = xa_entry_locked(xas->xa, node, offset); xas_free_nodes() local 642 void *entry; xas_create() local 775 xas_store(struct xa_state * xas,void * entry) xas_store() argument 1017 xas_split_alloc(struct xa_state * xas,void * entry,unsigned int order,gfp_t gfp) xas_split_alloc() argument 1068 xas_split(struct xa_state * xas,void * entry,unsigned int order) xas_split() argument 1167 void *entry; __xas_prev() local 1206 void *entry; __xas_next() local 1254 void *entry; xas_find() local 1325 void *entry; xas_find_marked() local 1472 void *entry; xa_load() local 1528 void *entry; xa_erase() local 1553 __xa_store(struct xarray * xa,unsigned long index,void * entry,gfp_t gfp) __xa_store() argument 1590 xa_store(struct xarray * xa,unsigned long index,void * entry,gfp_t gfp) xa_store() argument 1619 __xa_cmpxchg(struct xarray * xa,unsigned long index,void * old,void * entry,gfp_t gfp) __xa_cmpxchg() argument 1656 __xa_insert(struct xarray * xa,unsigned long index,void * entry,gfp_t gfp) __xa_insert() argument 1732 xa_store_range(struct xarray * xa,unsigned long first,unsigned long last,void * entry,gfp_t gfp) xa_store_range() argument 1777 void *entry; xa_get_order() local 1828 __xa_alloc(struct xarray * xa,u32 * id,void * entry,struct xa_limit limit,gfp_t gfp) __xa_alloc() argument 1880 __xa_alloc_cyclic(struct xarray * xa,u32 * id,void * entry,struct xa_limit limit,u32 * next,gfp_t gfp) __xa_alloc_cyclic() argument 1922 void *entry = xas_load(&xas); __xa_set_mark() local 1940 void *entry = xas_load(&xas); __xa_clear_mark() local 1962 void *entry; xa_get_mark() local 2036 void *entry; xa_find() local 2086 void *entry; xa_find_after() local 2116 void *entry; xas_extract_present() local 2135 void *entry; xas_extract_marked() local 2230 void *entry; xa_destroy() local 2279 xa_dump_entry(const void * entry,unsigned long index,unsigned long shift) xa_dump_entry() argument 2314 void *entry = xa->xa_head; xa_dump() local [all...] |
/linux/drivers/platform/x86/intel/pmt/ |
H A D | class.c | 87 struct intel_pmt_entry *entry = container_of(attr, in intel_pmt_read() local 94 if (off >= entry->size) in intel_pmt_read() 97 if (count > entry->size - off) in intel_pmt_read() 98 count = entry->size - off; in intel_pmt_read() 100 count = pmt_telem_read_mmio(entry->ep->pcidev, entry->cb, entry->header.guid, buf, in intel_pmt_read() 101 entry->base, off, count); in intel_pmt_read() 110 struct intel_pmt_entry *entry = container_of(attr, in intel_pmt_mmap() local 115 unsigned long phys = entry->base_addr; in intel_pmt_mmap() 122 psize = (PFN_UP(entry->base_addr + entry->size) - pfn) * PAGE_SIZE; in intel_pmt_mmap() 139 struct intel_pmt_entry *entry = dev_get_drvdata(dev); in guid_show() local [all …]
|
H A D | crashlog.c | 52 /* entry must be first member of struct */ 53 struct intel_pmt_entry entry; member 59 struct crashlog_entry entry[]; member 65 static bool pmt_crashlog_complete(struct intel_pmt_entry *entry) in pmt_crashlog_complete() argument 67 u32 control = readl(entry->disc_table + CONTROL_OFFSET); in pmt_crashlog_complete() 73 static bool pmt_crashlog_disabled(struct intel_pmt_entry *entry) in pmt_crashlog_disabled() argument 75 u32 control = readl(entry->disc_table + CONTROL_OFFSET); in pmt_crashlog_disabled() 81 static bool pmt_crashlog_supported(struct intel_pmt_entry *entry) in pmt_crashlog_supported() argument 83 u32 discovery_header = readl(entry->disc_table + CONTROL_OFFSET); in pmt_crashlog_supported() 96 static void pmt_crashlog_set_disable(struct intel_pmt_entry *entry, in pmt_crashlog_set_disable() argument [all …]
|
/linux/drivers/net/ethernet/rocker/ |
H A D | rocker_ofdpa.c | 93 struct hlist_node entry; member 103 struct hlist_node entry; member 129 struct hlist_node entry; member 141 struct hlist_node entry; member 148 struct hlist_node entry; member 306 const struct ofdpa_flow_tbl_entry *entry) in ofdpa_cmd_flow_tbl_add_ig_port() argument 309 entry->key.ig_port.in_pport)) in ofdpa_cmd_flow_tbl_add_ig_port() 312 entry->key.ig_port.in_pport_mask)) in ofdpa_cmd_flow_tbl_add_ig_port() 315 entry->key.ig_port.goto_tbl)) in ofdpa_cmd_flow_tbl_add_ig_port() 323 const struct ofdpa_flow_tbl_entry *entry) in ofdpa_cmd_flow_tbl_add_vlan() argument [all …]
|
/linux/kernel/dma/ |
H A D | debug.c | 166 static inline void dump_entry_trace(struct dma_debug_entry *entry) in dump_entry_trace() argument 169 if (entry) { in dump_entry_trace() 171 stack_trace_print(entry->stack_entries, entry->stack_len, 0); in dump_entry_trace() 217 #define err_printk(dev, entry, format, arg...) do { \ argument 224 dump_entry_trace(entry); \ 236 static int hash_fn(struct dma_debug_entry *entry) in hash_fn() argument 242 return (entry->dev_addr >> HASH_FN_SHIFT) & HASH_FN_MASK; in hash_fn() 248 static struct hash_bucket *get_hash_bucket(struct dma_debug_entry *entry, in get_hash_bucket() argument 252 int idx = hash_fn(entry); in get_hash_bucket() 290 * Search a given entry in the hash bucket list [all …]
|
/linux/arch/x86/kvm/ |
H A D | cpuid.c | 119 * If the index isn't significant, use the first entry with a in cpuid_entry2_find() 128 * Similarly, use the first matching entry if KVM is doing a in cpuid_entry2_find() 236 struct kvm_cpuid_entry2 *entry; in kvm_get_hypervisor_cpuid() local 240 entry = kvm_find_cpuid_entry(vcpu, base); in kvm_get_hypervisor_cpuid() 242 if (entry) { in kvm_get_hypervisor_cpuid() 245 signature[0] = entry->ebx; in kvm_get_hypervisor_cpuid() 246 signature[1] = entry->ecx; in kvm_get_hypervisor_cpuid() 247 signature[2] = entry->edx; in kvm_get_hypervisor_cpuid() 251 cpuid.limit = entry->eax; in kvm_get_hypervisor_cpuid() 295 struct kvm_cpuid_entry2 *entry, in kvm_update_feature_runtime() argument [all …]
|