/linux/drivers/net/ethernet/engleder/ |
H A D | tsnep_selftests.c | 357 qopt = kzalloc(struct_size(qopt, entries, 255), GFP_KERNEL); in tsnep_test_taprio() 361 qopt->entries[i].command = TC_TAPRIO_CMD_SET_GATES; in tsnep_test_taprio() 367 qopt->entries[0].gate_mask = 0x02; in tsnep_test_taprio() 368 qopt->entries[0].interval = 200000; in tsnep_test_taprio() 369 qopt->entries[1].gate_mask = 0x03; in tsnep_test_taprio() 370 qopt->entries[1].interval = 800000; in tsnep_test_taprio() 371 qopt->entries[2].gate_mask = 0x07; in tsnep_test_taprio() 372 qopt->entries[2].interval = 240000; in tsnep_test_taprio() 373 qopt->entries[3].gate_mask = 0x01; in tsnep_test_taprio() 374 qopt->entries[3].interval = 80000; in tsnep_test_taprio() [all …]
|
/linux/tools/testing/selftests/net/netfilter/ |
H A D | nft_audit.sh | 51 lsplit='s/^\(.*\) entries=\([^ ]*\) \(.*\)$/pfx="\1"\nval="\2"\nsfx="\3"/' 61 echo "$tpfx entries=$sum $tsfx" 67 echo "$tpfx entries=$sum $tsfx" 88 "table=$table family=2 entries=1 op=nft_register_table" 91 "table=$table family=2 entries=1 op=nft_register_chain" 94 "table=$table family=2 entries=2 op=nft_register_chain" 99 "table=$table family=2 entries=1 op=nft_register_rule" 102 "table=$table family=2 entries=2 op=nft_register_rule" 113 "table=$table family=2 entries=6 op=nft_register_rule" 120 'table=t2 family=2 entries=500 op=nft_register_rule' [all …]
|
/linux/kernel/ |
H A D | stacktrace.c | 19 * stack_trace_print - Print the entries in the stack trace 20 * @entries: Pointer to storage array 21 * @nr_entries: Number of entries in the storage array 24 void stack_trace_print(const unsigned long *entries, unsigned int nr_entries, in stack_trace_print() argument 29 if (WARN_ON(!entries)) in stack_trace_print() 33 printk("%*c%pS\n", 1 + spaces, ' ', (void *)entries[i]); in stack_trace_print() 38 * stack_trace_snprint - Print the entries in the stack trace into a buffer 41 * @entries: Pointer to storage array 42 * @nr_entries: Number of entries in the storage array 47 int stack_trace_snprint(char *buf, size_t size, const unsigned long *entries, in stack_trace_snprint() argument [all …]
|
/linux/lib/ |
H A D | stackdepot.c | 141 static void init_stack_table(unsigned long entries) in init_stack_table() argument 145 for (i = 0; i < entries; i++) in init_stack_table() 152 unsigned long entries = 0; in stack_depot_early_init() local 184 * If stack_bucket_number_order is not set, leave entries as 0 to rely in stack_depot_early_init() 188 entries = 1UL << stack_bucket_number_order; in stack_depot_early_init() 192 entries, in stack_depot_early_init() 204 if (!entries) { in stack_depot_early_init() 206 * Obtain the number of entries that was calculated by in stack_depot_early_init() 209 entries = stack_hash_mask + 1; in stack_depot_early_init() 211 init_stack_table(entries); in stack_depot_early_init() [all …]
|
H A D | test_rhashtable.c | 32 MODULE_PARM_DESC(parm_entries, "Number of entries to add (default: 50000)"); 74 unsigned int entries; member 139 unsigned int entries) in test_rht_lookup() argument 143 for (i = 0; i < entries; i++) { in test_rht_lookup() 176 static void test_bucket_stats(struct rhashtable *ht, unsigned int entries) in test_bucket_stats() argument 202 pr_info(" Traversal complete: counted=%u, nelems=%u, entries=%d, table-jumps=%u\n", in test_bucket_stats() 203 total, atomic_read(&ht->nelems), entries, chain_len); in test_bucket_stats() 205 if (total != atomic_read(&ht->nelems) || total != entries) in test_bucket_stats() 210 unsigned int entries) in test_rhashtable() argument 219 * Insert entries int in test_rhashtable() 270 test_rhltable(unsigned int entries) test_rhltable() argument 436 test_rhashtable_max(struct test_obj * array,unsigned int entries) test_rhashtable_max() argument 589 unsigned int entries = tdata->entries; thread_lookup_test() local 684 unsigned int entries; test_rht_init() local [all...] |
/linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
H A D | vega10_pptable.h | 131 UCHAR ucNumEntries; /* Number of entries. */ 132 ATOM_Vega10_State states[]; /* Dynamically allocate entries. */ 165 UCHAR ucNumEntries; /* Number of entries. */ 166 ATOM_Vega10_GFXCLK_Dependency_Record entries[]; /* Dynamically allocate entries. */ member 171 UCHAR ucNumEntries; /* Number of entries. */ 172 ATOM_Vega10_MCLK_Dependency_Record entries[]; /* Dynamically allocate entries. */ member 177 UCHAR ucNumEntries; /* Number of entries. */ 178 ATOM_Vega10_CLK_Dependency_Record entries[]; /* Dynamically allocate entries. */ member 183 UCHAR ucNumEntries; /* Number of entries. */ 184 ATOM_Vega10_CLK_Dependency_Record entries[]; /* Dynamically allocate entries. */ member [all …]
|
H A D | smu_helper.c | 224 vvalue = vol_table->entries[i].value; in phm_trim_voltage_table() 228 if (vvalue == table->entries[j].value) { in phm_trim_voltage_table() 235 table->entries[table->count].value = vvalue; in phm_trim_voltage_table() 236 table->entries[table->count].smio_low = in phm_trim_voltage_table() 237 vol_table->entries[i].smio_low; in phm_trim_voltage_table() 265 vol_table->entries[i].value = dep_table->entries[i].mvdd; in phm_get_svi2_mvdd_voltage_table() 266 vol_table->entries[i].smio_low = 0; in phm_get_svi2_mvdd_voltage_table() 293 vol_table->entries[i].value = dep_table->entries[i].vddci; in phm_get_svi2_vddci_voltage_table() 294 vol_table->entries[i].smio_low = 0; in phm_get_svi2_vddci_voltage_table() 321 vol_table->entries[i].value = lookup_table->entries[i].us_vdd; in phm_get_svi2_vdd_voltage_table() [all …]
|
H A D | vega10_processpptables.c | 319 (ATOM_Vega10_GFXCLK_Dependency_Record_V2 *)gfxclk_dep_table->entries; in init_over_drive_limits() 353 mm_table = kzalloc(struct_size(mm_table, entries, mm_dependency_table->ucNumEntries), in get_mm_clock_voltage_table() 361 mm_dependency_record = &mm_dependency_table->entries[i]; in get_mm_clock_voltage_table() 362 mm_table->entries[i].vddcInd = mm_dependency_record->ucVddcInd; in get_mm_clock_voltage_table() 363 mm_table->entries[i].samclock = in get_mm_clock_voltage_table() 365 mm_table->entries[i].eclk = le32_to_cpu(mm_dependency_record->ulEClk); in get_mm_clock_voltage_table() 366 mm_table->entries[i].vclk = le32_to_cpu(mm_dependency_record->ulVClk); in get_mm_clock_voltage_table() 367 mm_table->entries[i].dclk = le32_to_cpu(mm_dependency_record->ulDClk); in get_mm_clock_voltage_table() 576 clk_table = kzalloc(struct_size(clk_table, entries, clk_dep_table->ucNumEntries), in get_socclk_voltage_dependency_table() 584 clk_table->entries[i].vddInd = in get_socclk_voltage_dependency_table() [all …]
|
H A D | hwmgr_ppt.h | 50 uint32_t count; /* Number of entries. */ 51 phm_ppt_v1_clock_voltage_dependency_record entries[]; /* Dynamically allocate count entries. */ member 73 uint32_t count; /* Number of entries. */ 74 phm_ppt_v1_mm_clock_voltage_dependency_record entries[]; /* Dynamically allocate count entries. */ member 89 phm_ppt_v1_voltage_lookup_record entries[]; /* Dynamically allocate count entries. */ member 105 uint32_t count; /* Number of entries. */ 106 phm_ppt_v1_pcie_record entries[]; /* Dynamically allocate count entries. */ member
|
/linux/drivers/net/dsa/microchip/ |
H A D | ksz9477_acl.c | 124 * index within the ksz9477 device's ACL table. It omits printing entries that 145 /* no need to print empty entries */ in ksz9477_dump_acl_index() 156 * ksz9477_dump_acl - Print ACL entries 171 dev_err(dev->dev, " Empty ACL entries were skipped\n"); in ksz9477_dump_acl() 206 * ksz9477_acl_get_cont_entr - Get count of contiguous ACL entries and validate 213 * in an ACL entry indicates which entries contain Matching rules linked to it. 222 * This function checks that, for complex rules, the entries containing the 224 * and returns the number of these contiguous entries. 229 * - The number of contiguous entries if it is the root entry of a complex 246 entry = &acles->entries[index].entry[0]; in ksz9477_acl_get_cont_entr() [all …]
|
/linux/arch/powerpc/mm/book3s64/ |
H A D | iommu_api.c | 34 u64 entries; /* number of entries in hpas/hpages[] */ member 57 unsigned long entries, unsigned long dev_hpa, in mm_iommu_do_alloc() argument 66 ret = account_locked_vm(mm, entries, true); in mm_iommu_do_alloc() 70 locked_entries = entries; in mm_iommu_do_alloc() 80 mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc() 88 * we use @ua and @entries natural alignment to allow IOMMU pages in mm_iommu_do_alloc() 91 mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc() 92 mem->hpas = vzalloc(array_size(entries, sizeof(mem->hpas[0]))); in mm_iommu_do_alloc() 102 chunk = min(chunk, entries); in mm_iommu_do_alloc() 103 for (entry = 0; entry < entries; entry += chunk) { in mm_iommu_do_alloc() [all …]
|
/linux/tools/testing/selftests/net/ |
H A D | fdb_flush.sh | 4 # This test is for checking functionality of flushing FDB entries. 152 local entries=$($BRIDGE fdb show dev $dev | grep "$filter" | wc -l) 154 [[ $entries -eq $exp_entries ]] 157 log_test $rc 0 "$dev: Expected $exp_entries FDB entries, got $entries" 176 log_test $? 0 "Flush FDB by dev vx10 - test vx10 entries" 179 log_test $? 0 "Flush FDB by dev vx10 - test vx20 entries" 198 log_test $? 0 "Test entries with vni $vni_1" 201 log_test $? 0 "Test entries with vni $vni_2" 206 # Set some entries with {vni=x,src_vni=y} and some with the opposite - 207 # {vni=y,src_vni=x}, to verify that when we flush by src_vni=x, entries [all …]
|
/linux/tools/lib/api/fd/ |
H A D | array.c | 15 fda->entries = NULL; in fdarray__init() 27 struct pollfd *entries = realloc(fda->entries, size); in fdarray__grow() local 29 if (entries == NULL) in fdarray__grow() 34 free(entries); in fdarray__grow() 38 memset(&entries[fda->nr_alloc], 0, sizeof(struct pollfd) * nr); in fdarray__grow() 42 fda->entries = entries; in fdarray__grow() 65 free(fda->entries); in fdarray__exit() 84 fda->entries[fda->nr].fd = fd; in fdarray__add() 85 fda->entries[fda->nr].events = revents; in fdarray__add() 99 entry = &from->entries[pos]; in fdarray__dup_entry_from() [all …]
|
/linux/tools/perf/util/ |
H A D | mem2node.c | 50 struct phys_entry *entries, *tmp_entries; in mem2node__init() local 62 entries = zalloc(sizeof(*entries) * max); in mem2node__init() 63 if (!entries) in mem2node__init() 84 struct phys_entry *prev = &entries[j - 1]; in mem2node__init() 93 phys_entry__init(&entries[j++], start, bsize, n->node); in mem2node__init() 97 /* Cut unused entries, due to merging. */ in mem2node__init() 98 tmp_entries = realloc(entries, sizeof(*entries) * j); in mem2node__init() 101 entries = tmp_entries; in mem2node__init() 105 entries[i].node, entries[i].start, entries[i].end); in mem2node__init() 107 phys_entry__insert(&entries[i], &map->root); in mem2node__init() [all …]
|
H A D | arm64-frame-pointer-unwind-support.c | 13 struct entries { 31 struct entries *entries = arg; in get_leaf_frame_caller_aarch64() 33 entries->stack[entries->length++] = entry->ip; in get_leaf_frame_caller_aarch64() 40 struct entries entries = {}; in get_leaf_frame_caller_aarch64() 63 ret = unwind__get_entries(add_entry, &entries, thread, sample, 2, true); in get_leaf_frame_caller_aarch64() 66 if (ret || entries.length != 2) 69 return callchain_param.order == ORDER_CALLER ? entries 12 struct entries { global() struct 25 struct entries *entries = arg; add_entry() local 34 struct entries entries = {}; get_leaf_frame_caller_aarch64() local [all...] |
H A D | rb_resort.h |
|
/linux/drivers/net/ethernet/netronome/nfp/nfpcore/ |
H A D | nfp_nsp_eth.c | 294 union eth_table_entry *entries; in __nfp_eth_read_ports() local 298 entries = kzalloc(NSP_ETH_TABLE_SIZE, GFP_KERNEL); in __nfp_eth_read_ports() 299 if (!entries) in __nfp_eth_read_ports() 302 ret = nfp_nsp_read_eth_table(nsp, entries, NSP_ETH_TABLE_SIZE); in __nfp_eth_read_ports() 309 if (entries[i].port & NSP_ETH_PORT_LANES_MASK) in __nfp_eth_read_ports() 317 nfp_err(cpp, "table entry count reported (%d) does not match entries present (%d)\n", in __nfp_eth_read_ports() 328 if (entries[i].port & NSP_ETH_PORT_LANES_MASK) in __nfp_eth_read_ports() 329 nfp_eth_port_translate(nsp, &entries[i], i, in __nfp_eth_read_ports() 338 kfree(entries); in __nfp_eth_read_ports() 343 kfree(entries); in __nfp_eth_read_ports() [all …]
|
/linux/arch/x86/kernel/ |
H A D | e820.c | 28 * the first 128 E820 memory entries in boot_params.e820_table and the remaining 29 * (if any) entries of the SETUP_E820_EXT nodes. We use this to: 46 * the entries to create an E820 table for the kexec kernel. 83 struct e820_entry *entry = &table->entries[i]; in _e820__mapped_any() 118 struct e820_entry *entry = &e820_table->entries[i]; in __e820__mapped_all() 170 if (x >= ARRAY_SIZE(table->entries)) { in __e820__range_add() 171 pr_err("too many entries; ignoring [mem %#010llx-%#010llx]\n", in __e820__range_add() 176 table->entries[x].addr = start; in __e820__range_add() 177 table->entries[x].size = size; in __e820__range_add() 178 table->entries[x].type = type; in __e820__range_add() [all …]
|
/linux/kernel/events/ |
H A D | callchain.c | 52 struct callchain_cpus_entries *entries; in release_callchain_buffers_rcu() local 55 entries = container_of(head, struct callchain_cpus_entries, rcu_head); in release_callchain_buffers_rcu() 58 kfree(entries->cpu_entries[cpu]); in release_callchain_buffers_rcu() 60 kfree(entries); in release_callchain_buffers_rcu() 65 struct callchain_cpus_entries *entries; in release_callchain_buffers() local 67 entries = callchain_cpus_entries; in release_callchain_buffers() 69 call_rcu(&entries->rcu_head, release_callchain_buffers_rcu); in release_callchain_buffers() 76 struct callchain_cpus_entries *entries; in alloc_callchain_buffers() local 85 entries = kzalloc(size, GFP_KERNEL); in alloc_callchain_buffers() 86 if (!entries) in alloc_callchain_buffers() [all …]
|
/linux/tools/perf/tests/ |
H A D | fdarray.c | 14 fda->entries[fd].fd = fda->nr - fd; in fdarray__init_revents() 15 fda->entries[fd].events = revents; in fdarray__init_revents() 16 fda->entries[fd].revents = revents; in fdarray__init_revents() 58 fda->entries[2].revents = POLLIN; in test__fdarray__filter() 60 pr_debug("\nfiltering all but fda->entries[2]:"); in test__fdarray__filter() 70 fda->entries[0].revents = POLLIN; in test__fdarray__filter() 71 fda->entries[3].revents = POLLIN; in test__fdarray__filter() 73 pr_debug("\nfiltering all but (fda->entries[0], fda->entries[3]):"); in test__fdarray__filter() 103 if (fda->entries[_idx].fd != _fd) { \ in test__fdarray__add() 104 pr_debug("\n%d: fda->entries[%d](%d) != %d!", \ in test__fdarray__add() [all …]
|
/linux/drivers/misc/vmw_vmci/ |
H A D | vmci_handle_array.c | 22 array = kmalloc(struct_size(array, entries, capacity), GFP_ATOMIC); in vmci_handle_arr_create() 48 size_t new_size = struct_size(array, entries, in vmci_handle_arr_append_entry() 62 array->entries[array->size] = handle; in vmci_handle_arr_append_entry() 78 if (vmci_handle_is_equal(array->entries[i], entry_handle)) { in vmci_handle_arr_remove_entry() 79 handle = array->entries[i]; in vmci_handle_arr_remove_entry() 81 array->entries[i] = array->entries[array->size]; in vmci_handle_arr_remove_entry() 82 array->entries[array->size] = VMCI_INVALID_HANDLE; in vmci_handle_arr_remove_entry() 99 handle = array->entries[array->size]; in vmci_handle_arr_remove_tail() 100 array->entries[array->size] = VMCI_INVALID_HANDLE; in vmci_handle_arr_remove_tail() 115 return array->entries[index]; in vmci_handle_arr_get_entry() [all …]
|
/linux/drivers/gpu/drm/radeon/ |
H A D | r600_dpm.c | 824 radeon_table->entries = kcalloc(atom_table->ucNumEntries, in r600_parse_clk_voltage_dep_table() 827 if (!radeon_table->entries) in r600_parse_clk_voltage_dep_table() 830 entry = &atom_table->entries[0]; in r600_parse_clk_voltage_dep_table() 832 radeon_table->entries[i].clk = le16_to_cpu(entry->usClockLow) | in r600_parse_clk_voltage_dep_table() 834 radeon_table->entries[i].v = le16_to_cpu(entry->usVoltage); in r600_parse_clk_voltage_dep_table() 936 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); in r600_parse_extended_power_table() 947 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); in r600_parse_extended_power_table() 948 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); in r600_parse_extended_power_table() 959 kfree(rdev->pm.dpm.dyn_state.vddc_dependency_on_sclk.entries); in r600_parse_extended_power_table() 960 kfree(rdev->pm.dpm.dyn_state.vddci_dependency_on_mclk.entries); in r600_parse_extended_power_table() [all …]
|
/linux/drivers/tty/vt/ |
H A D | gen_ucs_fallback_table.py | 223 # Sort each page's entries by offset 230 """Compress consecutive entries with the same fallback character into ranges. 231 A range is only compressed if it contains 3 or more consecutive entries.""" 235 for page, entries in page_groups.items(): 238 while i < len(entries): 239 start_offset, fallback = entries[i] 241 # Look ahead to find consecutive entries with the same fallback 243 while (j < len(entries) and 244 entries[j][0] == entries[j-1][0] + 1 and # consecutive offsets 245 entries[j][1] == fallback): # same fallback [all …]
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-firmware-efi-esrt | 13 Description: The number of entries in the ESRT 18 Description: The maximum number of entries that /could/ be registered 27 What: /sys/firmware/efi/esrt/entries/entry<N>/ 31 subdirectory under entries/ . 32 example: /sys/firmware/efi/esrt/entries/entry0/ 34 What: /sys/firmware/efi/esrt/entries/entry<N>/fw_type 46 What: /sys/firmware/efi/esrt/entries/entry<N>/fw_class 51 What: /sys/firmware/efi/esrt/entries/entry<N>/fw_version 57 What: /sys/firmware/efi/esrt/entries/entry<N>/lowest_supported_fw_version 62 What: /sys/firmware/efi/esrt/entries/entry<N>/capsule_flags [all …]
|
H A D | sysfs-firmware-dmi-entries | 1 What: /sys/firmware/dmi/entries/ 15 DMI is structured as a large table of entries, where 19 entries. 21 Some entries are required by the specification, but many 27 Multiple entries of the same type are allowed. In order 31 to say, if there are 'N' multiple entries with the same type 34 entries "T-0" through "T-(N-1)": 38 /sys/firmware/dmi/entries/17-0 39 /sys/firmware/dmi/entries/17-1 40 /sys/firmware/dmi/entries/17-2 [all …]
|