Home
last modified time | relevance | path

Searched refs:entries (Results 1 – 25 of 1272) sorted by relevance

12345678910>>...51

/linux/drivers/net/ethernet/engleder/
H A Dtsnep_selftests.c357 qopt = kzalloc(struct_size(qopt, entries, 255), GFP_KERNEL); in tsnep_test_taprio()
361 qopt->entries[i].command = TC_TAPRIO_CMD_SET_GATES; in tsnep_test_taprio()
367 qopt->entries[0].gate_mask = 0x02; in tsnep_test_taprio()
368 qopt->entries[0].interval = 200000; in tsnep_test_taprio()
369 qopt->entries[1].gate_mask = 0x03; in tsnep_test_taprio()
370 qopt->entries[1].interval = 800000; in tsnep_test_taprio()
371 qopt->entries[2].gate_mask = 0x07; in tsnep_test_taprio()
372 qopt->entries[2].interval = 240000; in tsnep_test_taprio()
373 qopt->entries[3].gate_mask = 0x01; in tsnep_test_taprio()
374 qopt->entries[3].interval = 80000; in tsnep_test_taprio()
[all …]
/linux/lib/
H A Dstackdepot.c141 static void init_stack_table(unsigned long entries) in init_stack_table() argument
145 for (i = 0; i < entries; i++) in init_stack_table()
152 unsigned long entries = 0; in stack_depot_early_init() local
188 entries = 1UL << stack_bucket_number_order; in stack_depot_early_init()
192 entries, in stack_depot_early_init()
204 if (!entries) { in stack_depot_early_init()
209 entries = stack_hash_mask + 1; in stack_depot_early_init()
211 init_stack_table(entries); in stack_depot_early_init()
219 memblock_free(stack_table, entries * sizeof(struct list_head)); in stack_depot_early_init()
231 unsigned long entries; in stack_depot_init() local
[all …]
H A Dtest_rhashtable.c32 MODULE_PARM_DESC(parm_entries, "Number of entries to add (default: 50000)");
74 unsigned int entries; member
139 unsigned int entries) in test_rht_lookup() argument
143 for (i = 0; i < entries; i++) { in test_rht_lookup()
176 static void test_bucket_stats(struct rhashtable *ht, unsigned int entries) in test_bucket_stats() argument
202 pr_info(" Traversal complete: counted=%u, nelems=%u, entries=%d, table-jumps=%u\n", in test_bucket_stats()
203 total, atomic_read(&ht->nelems), entries, chain_len); in test_bucket_stats()
205 if (total != atomic_read(&ht->nelems) || total != entries) in test_bucket_stats()
210 unsigned int entries) in test_rhashtable() argument
219 * Insert entries int in test_rhashtable()
270 test_rhltable(unsigned int entries) test_rhltable() argument
436 test_rhashtable_max(struct test_obj * array,unsigned int entries) test_rhashtable_max() argument
589 unsigned int entries = tdata->entries; thread_lookup_test() local
684 unsigned int entries; test_rht_init() local
[all...]
/linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/
H A Dsmu_helper.c224 vvalue = vol_table->entries[i].value; in phm_trim_voltage_table()
228 if (vvalue == table->entries[j].value) { in phm_trim_voltage_table()
235 table->entries[table->count].value = vvalue; in phm_trim_voltage_table()
236 table->entries[table->count].smio_low = in phm_trim_voltage_table()
237 vol_table->entries[i].smio_low; in phm_trim_voltage_table()
265 vol_table->entries[i].value = dep_table->entries[i].mvdd; in phm_get_svi2_mvdd_voltage_table()
266 vol_table->entries[i].smio_low = 0; in phm_get_svi2_mvdd_voltage_table()
293 vol_table->entries[i].value = dep_table->entries[i].vddci; in phm_get_svi2_vddci_voltage_table()
294 vol_table->entries[i].smio_low = 0; in phm_get_svi2_vddci_voltage_table()
321 vol_table->entries[i].value = lookup_table->entries[i].us_vdd; in phm_get_svi2_vdd_voltage_table()
[all …]
H A Dvega10_processpptables.c319 (ATOM_Vega10_GFXCLK_Dependency_Record_V2 *)gfxclk_dep_table->entries; in init_over_drive_limits()
353 mm_table = kzalloc(struct_size(mm_table, entries, mm_dependency_table->ucNumEntries), in get_mm_clock_voltage_table()
361 mm_dependency_record = &mm_dependency_table->entries[i]; in get_mm_clock_voltage_table()
362 mm_table->entries[i].vddcInd = mm_dependency_record->ucVddcInd; in get_mm_clock_voltage_table()
363 mm_table->entries[i].samclock = in get_mm_clock_voltage_table()
365 mm_table->entries[i].eclk = le32_to_cpu(mm_dependency_record->ulEClk); in get_mm_clock_voltage_table()
366 mm_table->entries[i].vclk = le32_to_cpu(mm_dependency_record->ulVClk); in get_mm_clock_voltage_table()
367 mm_table->entries[i].dclk = le32_to_cpu(mm_dependency_record->ulDClk); in get_mm_clock_voltage_table()
576 clk_table = kzalloc(struct_size(clk_table, entries, clk_dep_table->ucNumEntries), in get_socclk_voltage_dependency_table()
584 clk_table->entries[i].vddInd = in get_socclk_voltage_dependency_table()
[all …]
/linux/arch/powerpc/mm/book3s64/
H A Diommu_api.c34 u64 entries; /* number of entries in hpas/hpages[] */ member
57 unsigned long entries, unsigned long dev_hpa, in mm_iommu_do_alloc() argument
66 ret = account_locked_vm(mm, entries, true); in mm_iommu_do_alloc()
70 locked_entries = entries; in mm_iommu_do_alloc()
80 mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
91 mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
92 mem->hpas = vzalloc(array_size(entries, sizeof(mem->hpas[0]))); in mm_iommu_do_alloc()
102 chunk = min(chunk, entries); in mm_iommu_do_alloc()
103 for (entry = 0; entry < entries; entry += chunk) { in mm_iommu_do_alloc()
104 unsigned long n = min(entries - entry, chunk); in mm_iommu_do_alloc()
[all …]
/linux/tools/lib/api/fd/
H A Darray.c15 fda->entries = NULL; in fdarray__init()
27 struct pollfd *entries = realloc(fda->entries, size); in fdarray__grow() local
29 if (entries == NULL) in fdarray__grow()
34 free(entries); in fdarray__grow()
38 memset(&entries[fda->nr_alloc], 0, sizeof(struct pollfd) * nr); in fdarray__grow()
42 fda->entries = entries; in fdarray__grow()
65 free(fda->entries); in fdarray__exit()
84 fda->entries[fda->nr].fd = fd; in fdarray__add()
85 fda->entries[fda->nr].events = revents; in fdarray__add()
99 entry = &from->entries[pos]; in fdarray__dup_entry_from()
[all …]
/linux/tools/perf/util/
H A Dmem2node.c50 struct phys_entry *entries, *tmp_entries; in mem2node__init() local
62 entries = zalloc(sizeof(*entries) * max); in mem2node__init()
63 if (!entries) in mem2node__init()
84 struct phys_entry *prev = &entries[j - 1]; in mem2node__init()
93 phys_entry__init(&entries[j++], start, bsize, n->node); in mem2node__init()
98 tmp_entries = realloc(entries, sizeof(*entries) * j); in mem2node__init()
101 entries = tmp_entries; in mem2node__init()
105 entries[i].node, entries[i].start, entries[i].end); in mem2node__init()
107 phys_entry__insert(&entries[i], &map->root); in mem2node__init()
110 map->entries = entries; in mem2node__init()
[all …]
H A Darm64-frame-pointer-unwind-support.c13 struct entries {
31 struct entries *entries = arg; in get_leaf_frame_caller_aarch64()
33 entries->stack[entries->length++] = entry->ip; in get_leaf_frame_caller_aarch64()
40 struct entries entries = {}; in get_leaf_frame_caller_aarch64()
63 ret = unwind__get_entries(add_entry, &entries, thread, sample, 2, true); in get_leaf_frame_caller_aarch64()
66 if (ret || entries.length != 2)
69 return callchain_param.order == ORDER_CALLER ? entries
12 struct entries { global() struct
25 struct entries *entries = arg; add_entry() local
34 struct entries entries = {}; get_leaf_frame_caller_aarch64() local
[all...]
H A Dpstack.c18 void *entries[]; member
45 if (pstack->entries[i] == key) { in pstack__remove()
47 memmove(pstack->entries + i, in pstack__remove()
48 pstack->entries + i + 1, in pstack__remove()
63 pstack->entries[pstack->top++] = key; in pstack__push()
70 return pstack->entries[pstack->top - 1]; in pstack__pop()
H A Drb_resort.h
/linux/drivers/net/ethernet/netronome/nfp/nfpcore/
H A Dnfp_nsp_eth.c294 union eth_table_entry *entries; in __nfp_eth_read_ports() local
298 entries = kzalloc(NSP_ETH_TABLE_SIZE, GFP_KERNEL); in __nfp_eth_read_ports()
299 if (!entries) in __nfp_eth_read_ports()
302 ret = nfp_nsp_read_eth_table(nsp, entries, NSP_ETH_TABLE_SIZE); in __nfp_eth_read_ports()
309 if (entries[i].port & NSP_ETH_PORT_LANES_MASK) in __nfp_eth_read_ports()
328 if (entries[i].port & NSP_ETH_PORT_LANES_MASK) in __nfp_eth_read_ports()
329 nfp_eth_port_translate(nsp, &entries[i], i, in __nfp_eth_read_ports()
338 kfree(entries); in __nfp_eth_read_ports()
343 kfree(entries); in __nfp_eth_read_ports()
349 union eth_table_entry *entries; in nfp_eth_config_start() local
[all …]
/linux/drivers/misc/vmw_vmci/
H A Dvmci_handle_array.c22 array = kmalloc(struct_size(array, entries, capacity), GFP_ATOMIC); in vmci_handle_arr_create()
48 size_t new_size = struct_size(array, entries, in vmci_handle_arr_append_entry()
62 array->entries[array->size] = handle; in vmci_handle_arr_append_entry()
78 if (vmci_handle_is_equal(array->entries[i], entry_handle)) { in vmci_handle_arr_remove_entry()
79 handle = array->entries[i]; in vmci_handle_arr_remove_entry()
81 array->entries[i] = array->entries[array->size]; in vmci_handle_arr_remove_entry()
82 array->entries[array->size] = VMCI_INVALID_HANDLE; in vmci_handle_arr_remove_entry()
99 handle = array->entries[array->size]; in vmci_handle_arr_remove_tail()
100 array->entries[array->size] = VMCI_INVALID_HANDLE; in vmci_handle_arr_remove_tail()
115 return array->entries[index]; in vmci_handle_arr_get_entry()
[all …]
/linux/drivers/net/dsa/sja1105/
H A Dsja1105_vl.c27 if (list_empty(&gating_cfg->entries)) { in sja1105_insert_gate_entry()
28 list_add(&e->list, &gating_cfg->entries); in sja1105_insert_gate_entry()
32 list_for_each_entry(p, &gating_cfg->entries, list) { in sja1105_insert_gate_entry()
65 list_for_each_entry(e, &gating_cfg->entries, list) { in sja1105_gating_cfg_time_to_interval()
70 if (prev == &gating_cfg->entries) in sja1105_gating_cfg_time_to_interval()
76 last_e = list_last_entry(&gating_cfg->entries, in sja1105_gating_cfg_time_to_interval()
85 list_for_each_entry_safe(e, n, &gating_cfg->entries, list) { in sja1105_free_gating_config()
144 u8 gate_state = rule->vl.entries[i].gate_state; in sja1105_compose_gating_subschedule()
157 time += rule->vl.entries[i].interval; in sja1105_compose_gating_subschedule()
353 kfree(table->entries); in sja1105_init_virtual_links()
[all …]
H A Dsja1105_tas.c49 if (!list_empty(&gating_cfg->entries)) { in sja1105_tas_set_runtime_params()
185 kfree(table->entries); in sja1105_init_scheduling()
192 kfree(table->entries); in sja1105_init_scheduling()
199 kfree(table->entries); in sja1105_init_scheduling()
206 kfree(table->entries); in sja1105_init_scheduling()
218 if (!list_empty(&gating_cfg->entries)) { in sja1105_init_scheduling()
231 table->entries = kcalloc(num_entries, table->ops->unpacked_entry_size, in sja1105_init_scheduling()
233 if (!table->entries) in sja1105_init_scheduling()
236 schedule = table->entries; in sja1105_init_scheduling()
240 table->entries = kcalloc(SJA1105_MAX_SCHEDULE_ENTRY_POINTS_PARAMS_COUNT, in sja1105_init_scheduling()
[all …]
/linux/arch/x86/xen/
H A Dmulticalls.c40 struct multicall_entry entries[MC_BATCH]; member
49 struct multicall_entry entries[MC_BATCH]; member
121 unsigned int opidx = mcdb->entries[idx].op & 0xff; in print_debug_data()
125 mcdb->entries[idx].op, b->entries[idx].result, in print_debug_data()
132 pr_cont("%lx ", mcdb->entries[idx].args[arg]); in print_debug_data()
161 memcpy(mcdb->entries, b->entries, in xen_mc_flush()
174 mc = &b->entries[0]; in xen_mc_flush()
183 if (HYPERVISOR_multicall(b->entries, b->mcidx) != 0) in xen_mc_flush()
186 if (b->entries[i].result < 0) in xen_mc_flush()
196 } else if (b->entries[i].result < 0) { in xen_mc_flush()
[all …]
/linux/tools/perf/trace/beauty/
H A Dioctl.c41 if (nr < strarray__ioctl_tty_cmd.nr_entries && strarray__ioctl_tty_cmd.entries[nr] != NULL) in ioctl__scnprintf_tty_cmd()
42 return scnprintf(bf, size, "%s", strarray__ioctl_tty_cmd.entries[nr]); in ioctl__scnprintf_tty_cmd()
52 if (nr < strarray__drm_ioctl_cmds.nr_entries && strarray__drm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_drm_cmd()
53 return scnprintf(bf, size, "DRM_%s", strarray__drm_ioctl_cmds.entries[nr]); in ioctl__scnprintf_drm_cmd()
63 …if (nr < strarray__sndrv_pcm_ioctl_cmds.nr_entries && strarray__sndrv_pcm_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_pcm_cmd()
64 return scnprintf(bf, size, "SNDRV_PCM_%s", strarray__sndrv_pcm_ioctl_cmds.entries[nr]); in ioctl__scnprintf_sndrv_pcm_cmd()
74 …if (nr < strarray__sndrv_ctl_ioctl_cmds.nr_entries && strarray__sndrv_ctl_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_ctl_cmd()
75 return scnprintf(bf, size, "SNDRV_CTL_%s", strarray__sndrv_ctl_ioctl_cmds.entries[nr]); in ioctl__scnprintf_sndrv_ctl_cmd()
85 if (nr < strarray__kvm_ioctl_cmds.nr_entries && strarray__kvm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_kvm_cmd()
86 return scnprintf(bf, size, "KVM_%s", strarray__kvm_ioctl_cmds.entries[nr]); in ioctl__scnprintf_kvm_cmd()
[all …]
/linux/arch/x86/kernel/cpu/
H A Dintel.c653 short entries = desc->entries; in intel_tlb_lookup() local
657 tlb_lli_4k = max(tlb_lli_4k, entries); in intel_tlb_lookup()
658 tlb_lld_4k = max(tlb_lld_4k, entries); in intel_tlb_lookup()
661 tlb_lli_4k = max(tlb_lli_4k, entries); in intel_tlb_lookup()
662 tlb_lld_4k = max(tlb_lld_4k, entries); in intel_tlb_lookup()
663 tlb_lli_2m = max(tlb_lli_2m, entries); in intel_tlb_lookup()
664 tlb_lld_2m = max(tlb_lld_2m, entries); in intel_tlb_lookup()
665 tlb_lli_4m = max(tlb_lli_4m, entries); in intel_tlb_lookup()
666 tlb_lld_4m = max(tlb_lld_4m, entries); in intel_tlb_lookup()
669 tlb_lli_4k = max(tlb_lli_4k, entries); in intel_tlb_lookup()
[all …]
/linux/drivers/tty/vt/
H A Dgen_ucs_fallback_table.py235 for page, entries in page_groups.items():
238 while i < len(entries):
239 start_offset, fallback = entries[i]
243 while (j < len(entries) and
244 entries[j][0] == entries[j-1][0] + 1 and # consecutive offsets
245 entries[j][1] == fallback): # same fallback
249 end_offset = entries[j-1][0]
259 compressed_entries.append(entries[k])
286 total_compressed_entries = sum(len(entries) for entries in compressed_pages.values())
316 for page, entries in sorted_pages:
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Dport.c68 table->entries[i] = 0; in mlx4_init_mac_table()
82 table->entries[i] = 0; in mlx4_init_vlan_table()
105 if (index < 0 || index >= table->max || !table->entries[index]) { in validate_index()
120 (MLX4_MAC_MASK & be64_to_cpu(table->entries[i]))) in find_index()
128 __be64 *entries) in mlx4_set_port_mac_table() argument
138 memcpy(mailbox->buf, entries, MLX4_MAC_TABLE_SIZE); in mlx4_set_port_mac_table()
196 if (((MLX4_MAC_MASK & mac) == (MLX4_MAC_MASK & be64_to_cpu(table->entries[i])))) in __mlx4_register_mac()
198 if (((MLX4_MAC_MASK & mac) == (MLX4_MAC_MASK & be64_to_cpu(dup_table->entries[i])))) in __mlx4_register_mac()
223 ((MLX4_MAC_MASK & mac) == (MLX4_MAC_MASK & be64_to_cpu(table->entries[index_at_dup_port])))) in __mlx4_register_mac()
242 (MLX4_MAC_MASK & be64_to_cpu(table->entries[i]))) { in __mlx4_register_mac()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn31/
H A Ddcn31_fpu.c458 if (dc->clk_mgr->bw_params->wm_table.entries[WM_A].valid) { in dcn31_update_soc_for_wm_a()
459 …context->bw_ctx.dml.soc.dram_clock_change_latency_us = dc->clk_mgr->bw_params->wm_table.entries[WM… in dcn31_update_soc_for_wm_a()
460 …context->bw_ctx.dml.soc.sr_enter_plus_exit_time_us = dc->clk_mgr->bw_params->wm_table.entries[WM_A… in dcn31_update_soc_for_wm_a()
461 …context->bw_ctx.dml.soc.sr_exit_time_us = dc->clk_mgr->bw_params->wm_table.entries[WM_A].sr_exit_t… in dcn31_update_soc_for_wm_a()
469 if (dc->clk_mgr->bw_params->wm_table.entries[WM_A].valid) { in dcn315_update_soc_for_wm_a()
474 …context->bw_ctx.dml.soc.dram_clock_change_latency_us = dc->clk_mgr->bw_params->wm_table.entries[WM… in dcn315_update_soc_for_wm_a()
476 dc->clk_mgr->bw_params->wm_table.entries[WM_A].sr_enter_plus_exit_time_us; in dcn315_update_soc_for_wm_a()
478 dc->clk_mgr->bw_params->wm_table.entries[WM_A].sr_exit_time_us; in dcn315_update_soc_for_wm_a()
611 if (clk_table->entries[i].dispclk_mhz > max_dispclk_mhz) in dcn31_update_bw_bounding_box()
612 max_dispclk_mhz = clk_table->entries[i].dispclk_mhz; in dcn31_update_bw_bounding_box()
[all …]
/linux/drivers/gpu/drm/radeon/
H A Dr600_dpm.c824 radeon_table->entries = kcalloc(atom_table->ucNumEntries, in r600_parse_clk_voltage_dep_table()
827 if (!radeon_table->entries) in r600_parse_clk_voltage_dep_table()
830 entry = &atom_table->entries[0]; in r600_parse_clk_voltage_dep_table()
832 radeon_table->entries[i].clk = le16_to_cpu(entry->usClockLow) | in r600_parse_clk_voltage_dep_table()
834 radeon_table->entries[i].v = le16_to_cpu(entry->usVoltage); in r600_parse_clk_voltage_dep_table()
936 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); in r600_parse_extended_power_table()
947 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); in r600_parse_extended_power_table()
948 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); in r600_parse_extended_power_table()
959 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); in r600_parse_extended_power_table()
960 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); in r600_parse_extended_power_table()
[all …]
/linux/kernel/
H A Dstacktrace.c24 void stack_trace_print(const unsigned long *entries, unsigned int nr_entries, in stack_trace_print() argument
29 if (WARN_ON(!entries)) in stack_trace_print()
33 printk("%*c%pS\n", 1 + spaces, ' ', (void *)entries[i]); in stack_trace_print()
47 int stack_trace_snprint(char *buf, size_t size, const unsigned long *entries, in stack_trace_snprint() argument
52 if (WARN_ON(!entries)) in stack_trace_snprint()
57 (void *)entries[i]); in stack_trace_snprint()
272 .entries = store, in stack_trace_save()
296 .entries = store, in stack_trace_save_tsk()
320 .entries = store, in stack_trace_save_regs()
346 .entries = store, in stack_trace_save_tsk_reliable()
[all …]
/linux/drivers/media/dvb-frontends/
H A Ddvb-pll.c66 } entries[]; member
79 .entries = {
102 .entries = {
125 .entries = {
143 .entries = {
158 .entries = {
180 .entries = {
194 .entries = {
220 .entries = {
253 .entries = {
[all …]
/linux/tools/perf/tests/
H A Dfdarray.c14 fda->entries[fd].fd = fda->nr - fd; in fdarray__init_revents()
15 fda->entries[fd].events = revents; in fdarray__init_revents()
16 fda->entries[fd].revents = revents; in fdarray__init_revents()
58 fda->entries[2].revents = POLLIN; in test__fdarray__filter()
70 fda->entries[0].revents = POLLIN; in test__fdarray__filter()
71 fda->entries[3].revents = POLLIN; in test__fdarray__filter()
103 if (fda->entries[_idx].fd != _fd) { \ in test__fdarray__add()
105 __LINE__, _idx, fda->entries[1].fd, _fd); \ in test__fdarray__add()
108 if (fda->entries[_idx].events != (_revents)) { \ in test__fdarray__add()
110 __LINE__, _idx, fda->entries[_idx].fd, _revents); \ in test__fdarray__add()
[all …]

12345678910>>...51