/linux/drivers/net/ethernet/engleder/ |
H A D | tsnep_selftests.c | 357 qopt = kzalloc(struct_size(qopt, entries, 255), GFP_KERNEL); in tsnep_test_taprio() 361 qopt->entries[i].command = TC_TAPRIO_CMD_SET_GATES; in tsnep_test_taprio() 367 qopt->entries[0].gate_mask = 0x02; in tsnep_test_taprio() 368 qopt->entries[0].interval = 200000; in tsnep_test_taprio() 369 qopt->entries[1].gate_mask = 0x03; in tsnep_test_taprio() 370 qopt->entries[1].interval = 800000; in tsnep_test_taprio() 371 qopt->entries[2].gate_mask = 0x07; in tsnep_test_taprio() 372 qopt->entries[2].interval = 240000; in tsnep_test_taprio() 373 qopt->entries[3].gate_mask = 0x01; in tsnep_test_taprio() 374 qopt->entries[3].interval = 80000; in tsnep_test_taprio() [all …]
|
/linux/tools/testing/selftests/net/netfilter/ |
H A D | nft_audit.sh | 51 lsplit='s/^\(.*\) entries=\([^ ]*\) \(.*\)$/pfx="\1"\nval="\2"\nsfx="\3"/' 61 echo "$tpfx entries=$sum $tsfx" 67 echo "$tpfx entries=$sum $tsfx" 88 "table=$table family=2 entries=1 op=nft_register_table" 91 "table=$table family=2 entries=1 op=nft_register_chain" 94 "table=$table family=2 entries=2 op=nft_register_chain" 99 "table=$table family=2 entries=1 op=nft_register_rule" 102 "table=$table family=2 entries=2 op=nft_register_rule" 113 "table=$table family=2 entries=6 op=nft_register_rule" 120 'table=t2 family=2 entries=500 op=nft_register_rule' [all …]
|
/linux/lib/ |
H A D | hashtable_test.c | 90 /* Both entries should have been visited exactly once. */ in hashtable_test_hash_add() 125 struct hashtable_test_entry entries[3]; in hashtable_test_hash_for_each() local 130 /* Add three entries to the hashtable. */ in hashtable_test_hash_for_each() 132 entries[i].key = i; in hashtable_test_hash_for_each() 133 entries[i].data = i + 10; in hashtable_test_hash_for_each() 134 entries[i].visited = 0; in hashtable_test_hash_for_each() 135 hash_add(hash, &entries[i].node, entries[i].key); in hashtable_test_hash_for_each() 149 KUNIT_EXPECT_EQ(test, entries[j].visited, 1); in hashtable_test_hash_for_each() 154 struct hashtable_test_entry entries[3]; in hashtable_test_hash_for_each_safe() local 160 /* Add three entries to the hashtable. */ in hashtable_test_hash_for_each_safe() [all …]
|
H A D | list-test.c | 391 struct list_head entries[3], *cur; in list_test_list_cut_position() local 396 list_add_tail(&entries[0], &list1); in list_test_list_cut_position() 397 list_add_tail(&entries[1], &list1); in list_test_list_cut_position() 398 list_add_tail(&entries[2], &list1); in list_test_list_cut_position() 400 /* before: [list1] -> entries[0] -> entries[1] -> entries[2] */ in list_test_list_cut_position() 401 list_cut_position(&list2, &list1, &entries[1]); in list_test_list_cut_position() 402 /* after: [list2] -> entries[0] -> entries[1], [list1] -> entries[2] */ in list_test_list_cut_position() 405 KUNIT_EXPECT_PTR_EQ(test, cur, &entries[i]); in list_test_list_cut_position() 412 KUNIT_EXPECT_PTR_EQ(test, cur, &entries[i]); in list_test_list_cut_position() 419 struct list_head entries[3], *cur; in list_test_list_cut_before() local [all …]
|
H A D | stackdepot.c | 113 static void init_stack_table(unsigned long entries) in init_stack_table() argument 117 for (i = 0; i < entries; i++) in init_stack_table() 124 unsigned long entries = 0; in stack_depot_early_init() local 156 * If stack_bucket_number_order is not set, leave entries as 0 to rely in stack_depot_early_init() 160 entries = 1UL << stack_bucket_number_order; in stack_depot_early_init() 164 entries, in stack_depot_early_init() 176 if (!entries) { in stack_depot_early_init() 178 * Obtain the number of entries that was calculated by in stack_depot_early_init() 181 entries = stack_hash_mask + 1; in stack_depot_early_init() 183 init_stack_table(entries); in stack_depot_early_init() 192 unsigned long entries; stack_depot_init() local 381 depot_alloc_stack(unsigned long * entries,unsigned int nr_entries,u32 hash,depot_flags_t flags,void ** prealloc) depot_alloc_stack() argument 512 hash_stack(unsigned long * entries,unsigned int size) hash_stack() argument 536 find_stack(struct list_head * bucket,unsigned long * entries,int size,u32 hash,depot_flags_t flags) find_stack() argument 584 stack_depot_save_flags(unsigned long * entries,unsigned int nr_entries,gfp_t alloc_flags,depot_flags_t depot_flags) stack_depot_save_flags() argument 681 stack_depot_save(unsigned long * entries,unsigned int nr_entries,gfp_t alloc_flags) stack_depot_save() argument 699 stack_depot_fetch(depot_stack_handle_t handle,unsigned long ** entries) stack_depot_fetch() argument 748 unsigned long *entries; stack_depot_print() local 760 unsigned long *entries; stack_depot_snprint() local [all...] |
H A D | test_rhashtable.c | 32 MODULE_PARM_DESC(parm_entries, "Number of entries to add (default: 50000)"); 74 unsigned int entries; member 139 unsigned int entries) in test_rht_lookup() argument 143 for (i = 0; i < entries; i++) { in test_rht_lookup() 176 static void test_bucket_stats(struct rhashtable *ht, unsigned int entries) in test_bucket_stats() argument 202 pr_info(" Traversal complete: counted=%u, nelems=%u, entries=%d, table-jumps=%u\n", in test_bucket_stats() 203 total, atomic_read(&ht->nelems), entries, chain_len); in test_bucket_stats() 205 if (total != atomic_read(&ht->nelems) || total != entries) in test_bucket_stats() 210 unsigned int entries) in test_rhashtable() argument 219 * Insert entries int in test_rhashtable() 270 test_rhltable(unsigned int entries) test_rhltable() argument 436 test_rhashtable_max(struct test_obj * array,unsigned int entries) test_rhashtable_max() argument 589 unsigned int entries = tdata->entries; thread_lookup_test() local 684 unsigned int entries; test_rht_init() local [all...] |
/linux/drivers/gpu/drm/amd/display/dc/dml2/dml21/src/dml2_mcg/ |
H A D | dml2_mcg_dcn4.c | 63 …min_table->dram_bw_table.entries[i].pre_derate_dram_bw_kbps = uclk_to_dram_bw_kbps(soc_bb->clk_tab… in build_min_clk_table_fine_grained() 65 …min_table->dram_bw_table.entries[i].min_fclk_khz = (unsigned long)((((double)min_table->dram_bw_ta… in build_min_clk_table_fine_grained() 69 …// To create the minium table, effectively shift "up" all the dcfclk/fclk entries by 1, and then r… in build_min_clk_table_fine_grained() 71 prev_100 = min_table->dram_bw_table.entries[i - 1].min_fclk_khz; in build_min_clk_table_fine_grained() 72 cur_50 = min_table->dram_bw_table.entries[i].min_fclk_khz / 2; in build_min_clk_table_fine_grained() 73 min_table->dram_bw_table.entries[i].min_fclk_khz = prev_100 > cur_50 ? prev_100 : cur_50; in build_min_clk_table_fine_grained() 76 …min_table->dram_bw_table.entries[i].min_fclk_khz = round_up_to_quantized_values(min_table->dram_bw… in build_min_clk_table_fine_grained() 79 min_table->dram_bw_table.entries[0].min_fclk_khz /= 2; in build_min_clk_table_fine_grained() 83 if (min_table->dram_bw_table.entries[i].min_dcfclk_khz < min_dcfclk_khz) in build_min_clk_table_fine_grained() 84 min_table->dram_bw_table.entries[i].min_dcfclk_khz = min_dcfclk_khz; in build_min_clk_table_fine_grained() [all …]
|
/linux/kernel/ |
H A D | stacktrace.c | 19 * stack_trace_print - Print the entries in the stack trace 20 * @entries: Pointer to storage array 21 * @nr_entries: Number of entries in the storage array 24 void stack_trace_print(const unsigned long *entries, unsigned int nr_entries, in stack_trace_print() argument 29 if (WARN_ON(!entries)) in stack_trace_print() 33 printk("%*c%pS\n", 1 + spaces, ' ', (void *)entries[i]); in stack_trace_print() 38 * stack_trace_snprint - Print the entries in the stack trace into a buffer 41 * @entries: Pointer to storage array 42 * @nr_entries: Number of entries in the storage array 47 int stack_trace_snprint(char *buf, size_t size, const unsigned long *entries, in stack_trace_snprint() argument [all …]
|
/linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
H A D | vega10_pptable.h | 131 UCHAR ucNumEntries; /* Number of entries. */ 132 ATOM_Vega10_State states[]; /* Dynamically allocate entries. */ 165 UCHAR ucNumEntries; /* Number of entries. */ 166 ATOM_Vega10_GFXCLK_Dependency_Record entries[]; /* Dynamically allocate entries. */ member 171 UCHAR ucNumEntries; /* Number of entries. */ 172 ATOM_Vega10_MCLK_Dependency_Record entries[]; /* Dynamically allocate entries. */ member 177 UCHAR ucNumEntries; /* Number of entries. */ 178 ATOM_Vega10_CLK_Dependency_Record entries[]; /* Dynamically allocate entries. */ member 183 UCHAR ucNumEntries; /* Number of entries. */ 184 ATOM_Vega10_CLK_Dependency_Record entries[]; /* Dynamically allocate entries. */ member [all …]
|
H A D | smu_helper.c | 224 vvalue = vol_table->entries[i].value; in phm_trim_voltage_table() 228 if (vvalue == table->entries[j].value) { in phm_trim_voltage_table() 235 table->entries[table->count].value = vvalue; in phm_trim_voltage_table() 236 table->entries[table->count].smio_low = in phm_trim_voltage_table() 237 vol_table->entries[i].smio_low; in phm_trim_voltage_table() 265 vol_table->entries[i].value = dep_table->entries[i].mvdd; in phm_get_svi2_mvdd_voltage_table() 266 vol_table->entries[i].smio_low = 0; in phm_get_svi2_mvdd_voltage_table() 293 vol_table->entries[i].value = dep_table->entries[i].vddci; in phm_get_svi2_vddci_voltage_table() 294 vol_table->entries[i].smio_low = 0; in phm_get_svi2_vddci_voltage_table() 321 vol_table->entries[i].value = lookup_table->entries[i].us_vdd; in phm_get_svi2_vdd_voltage_table() [all …]
|
H A D | vega10_processpptables.c | 319 (ATOM_Vega10_GFXCLK_Dependency_Record_V2 *)gfxclk_dep_table->entries; in init_over_drive_limits() 353 mm_table = kzalloc(struct_size(mm_table, entries, mm_dependency_table->ucNumEntries), in get_mm_clock_voltage_table() 361 mm_dependency_record = &mm_dependency_table->entries[i]; in get_mm_clock_voltage_table() 362 mm_table->entries[i].vddcInd = mm_dependency_record->ucVddcInd; in get_mm_clock_voltage_table() 363 mm_table->entries[i].samclock = in get_mm_clock_voltage_table() 365 mm_table->entries[i].eclk = le32_to_cpu(mm_dependency_record->ulEClk); in get_mm_clock_voltage_table() 366 mm_table->entries[i].vclk = le32_to_cpu(mm_dependency_record->ulVClk); in get_mm_clock_voltage_table() 367 mm_table->entries[i].dclk = le32_to_cpu(mm_dependency_record->ulDClk); in get_mm_clock_voltage_table() 576 clk_table = kzalloc(struct_size(clk_table, entries, clk_dep_table->ucNumEntries), in get_socclk_voltage_dependency_table() 584 clk_table->entries[i].vddInd = in get_socclk_voltage_dependency_table() [all …]
|
/linux/drivers/net/dsa/microchip/ |
H A D | ksz9477_acl.c | 124 * index within the ksz9477 device's ACL table. It omits printing entries that 145 /* no need to print empty entries */ in ksz9477_dump_acl_index() 156 * ksz9477_dump_acl - Print ACL entries 171 dev_err(dev->dev, " Empty ACL entries were skipped\n"); in ksz9477_dump_acl() 206 * ksz9477_acl_get_cont_entr - Get count of contiguous ACL entries and validate 213 * in an ACL entry indicates which entries contain Matching rules linked to it. 222 * This function checks that, for complex rules, the entries containing the 224 * and returns the number of these contiguous entries. 229 * - The number of contiguous entries if it is the root entry of a complex 246 entry = &acles->entries[index].entry[0]; in ksz9477_acl_get_cont_entr() [all …]
|
/linux/arch/powerpc/mm/book3s64/ |
H A D | iommu_api.c | 34 u64 entries; /* number of entries in hpas/hpages[] */ member 57 unsigned long entries, unsigned long dev_hpa, in mm_iommu_do_alloc() argument 66 ret = account_locked_vm(mm, entries, true); in mm_iommu_do_alloc() 70 locked_entries = entries; in mm_iommu_do_alloc() 80 mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc() 88 * we use @ua and @entries natural alignment to allow IOMMU pages in mm_iommu_do_alloc() 91 mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc() 92 mem->hpas = vzalloc(array_size(entries, sizeof(mem->hpas[0]))); in mm_iommu_do_alloc() 102 chunk = min(chunk, entries); in mm_iommu_do_alloc() 103 for (entry = 0; entry < entries; entry += chunk) { in mm_iommu_do_alloc() [all …]
|
/linux/arch/x86/kernel/cpu/ |
H A D | intel.c | 344 * the TLB when any changes are made to any of the page table entries. in early_init_intel() 720 { 0xb1, TLB_INST_2M_4M, 4, " TLB_INST 2M pages, 4-way, 8 entries or 4M pages, 4-way entries" }, 750 if (tlb_lli_4k[ENTRIES] < intel_tlb_table[k].entries) in intel_tlb_lookup() 751 tlb_lli_4k[ENTRIES] = intel_tlb_table[k].entries; in intel_tlb_lookup() 752 if (tlb_lld_4k[ENTRIES] < intel_tlb_table[k].entries) in intel_tlb_lookup() 753 tlb_lld_4k[ENTRIES] = intel_tlb_table[k].entries; in intel_tlb_lookup() 756 if (tlb_lli_4k[ENTRIES] < intel_tlb_table[k].entries) in intel_tlb_lookup() 757 tlb_lli_4k[ENTRIES] = intel_tlb_table[k].entries; in intel_tlb_lookup() 758 if (tlb_lld_4k[ENTRIES] < intel_tlb_table[k].entries) in intel_tlb_lookup() 759 tlb_lld_4k[ENTRIES] = intel_tlb_table[k].entries; in intel_tlb_lookup() [all …]
|
/linux/drivers/gpu/drm/amd/display/dc/clk_mgr/dcn315/ |
H A D | dcn315_clk_mgr.c | 256 .entries = { 304 .entries = { 341 .entries = { 388 /* skip empty entries, the smu array has no holes*/ in dcn315_build_watermark_ranges() 389 if (!bw_params->wm_table.entries[i].valid) in dcn315_build_watermark_ranges() 392 table->WatermarkRow[WM_DCFCLK][num_valid_sets].WmSetting = bw_params->wm_table.entries[i].wm_inst; in dcn315_build_watermark_ranges() 393 table->WatermarkRow[WM_DCFCLK][num_valid_sets].WmType = bw_params->wm_table.entries[i].wm_type; in dcn315_build_watermark_ranges() 404 bw_params->clk_table.entries[i - 1].dcfclk_mhz + 1; in dcn315_build_watermark_ranges() 407 bw_params->clk_table.entries[i].dcfclk_mhz; in dcn315_build_watermark_ranges() 487 …struct clk_limit_table_entry def_max = bw_params->clk_table.entries[bw_params->clk_table.num_entri… in dcn315_clk_mgr_helper_populate_bw_params() [all …]
|
/linux/tools/lib/api/fd/ |
H A D | array.c | 15 fda->entries = NULL; in fdarray__init() 27 struct pollfd *entries = realloc(fda->entries, size); in fdarray__grow() local 29 if (entries == NULL) in fdarray__grow() 34 free(entries); in fdarray__grow() 38 memset(&entries[fda->nr_alloc], 0, sizeof(struct pollfd) * nr); in fdarray__grow() 42 fda->entries = entries; in fdarray__grow() 65 free(fda->entries); in fdarray__exit() 84 fda->entries[fda->nr].fd = fd; in fdarray__add() 85 fda->entries[fda->nr].events = revents; in fdarray__add() 99 entry = &from->entries[pos]; in fdarray__dup_entry_from() [all …]
|
/linux/tools/testing/selftests/net/ |
H A D | fdb_flush.sh | 4 # This test is for checking functionality of flushing FDB entries. 152 local entries=$($BRIDGE fdb show dev $dev | grep "$filter" | wc -l) 154 [[ $entries -eq $exp_entries ]] 157 log_test $rc 0 "$dev: Expected $exp_entries FDB entries, got $entries" 176 log_test $? 0 "Flush FDB by dev vx10 - test vx10 entries" 179 log_test $? 0 "Flush FDB by dev vx10 - test vx20 entries" 198 log_test $? 0 "Test entries with vni $vni_1" 201 log_test $? 0 "Test entries with vni $vni_2" 206 # Set some entries with {vni=x,src_vni=y} and some with the opposite - 207 # {vni=y,src_vni=x}, to verify that when we flush by src_vni=x, entries [all …]
|
/linux/tools/perf/util/ |
H A D | mem2node.c | 50 struct phys_entry *entries, *tmp_entries; in mem2node__init() local 62 entries = zalloc(sizeof(*entries) * max); in mem2node__init() 63 if (!entries) in mem2node__init() 84 struct phys_entry *prev = &entries[j - 1]; in mem2node__init() 93 phys_entry__init(&entries[j++], start, bsize, n->node); in mem2node__init() 97 /* Cut unused entries, due to merging. */ in mem2node__init() 98 tmp_entries = realloc(entries, sizeof(*entries) * j); in mem2node__init() 101 entries = tmp_entries; in mem2node__init() 105 entries[i].node, entries[i].start, entries[i].end); in mem2node__init() 107 phys_entry__insert(&entries[i], &map->root); in mem2node__init() [all …]
|
H A D | arm64-frame-pointer-unwind-support.c | 12 struct entries { struct 25 struct entries *entries = arg; in add_entry() local 27 entries->stack[entries->length++] = entry->ip; in add_entry() 34 struct entries entries = {}; in get_leaf_frame_caller_aarch64() local 56 ret = unwind__get_entries(add_entry, &entries, thread, sample, 2, true); in get_leaf_frame_caller_aarch64() 59 if (ret || entries.length != 2) in get_leaf_frame_caller_aarch64() 62 return callchain_param.order == ORDER_CALLER ? entries.stack[0] : entries.stack[1]; in get_leaf_frame_caller_aarch64()
|
/linux/kernel/events/ |
H A D | callchain.c | 51 struct callchain_cpus_entries *entries; in release_callchain_buffers_rcu() local 54 entries = container_of(head, struct callchain_cpus_entries, rcu_head); in release_callchain_buffers_rcu() 57 kfree(entries->cpu_entries[cpu]); in release_callchain_buffers_rcu() 59 kfree(entries); in release_callchain_buffers_rcu() 64 struct callchain_cpus_entries *entries; in release_callchain_buffers() local 66 entries = callchain_cpus_entries; in release_callchain_buffers() 68 call_rcu(&entries->rcu_head, release_callchain_buffers_rcu); in release_callchain_buffers() 75 struct callchain_cpus_entries *entries; in alloc_callchain_buffers() local 84 entries = kzalloc(size, GFP_KERNEL); in alloc_callchain_buffers() 85 if (!entries) in alloc_callchain_buffers() [all …]
|
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn321/ |
H A D | dcn321_fpu.c | 260 * sort_entries_with_same_bw - Sort entries sharing the same bandwidth by DCFCLK 294 …* remove_inconsistent_entries - Ensure entries with the same bandwidth have MEMCLK and FCLK monoto… 295 * and remove entries that do not follow this order 365 if (bw_params->clk_table.entries[i].dcfclk_mhz > max_clk_data.dcfclk_mhz) in build_synthetic_soc_states() 366 max_clk_data.dcfclk_mhz = bw_params->clk_table.entries[i].dcfclk_mhz; in build_synthetic_soc_states() 367 if (bw_params->clk_table.entries[i].fclk_mhz > max_clk_data.fclk_mhz) in build_synthetic_soc_states() 368 max_clk_data.fclk_mhz = bw_params->clk_table.entries[i].fclk_mhz; in build_synthetic_soc_states() 369 if (bw_params->clk_table.entries[i].memclk_mhz > max_clk_data.memclk_mhz) in build_synthetic_soc_states() 370 max_clk_data.memclk_mhz = bw_params->clk_table.entries[i].memclk_mhz; in build_synthetic_soc_states() 371 if (bw_params->clk_table.entries[i].dispclk_mhz > max_clk_data.dispclk_mhz) in build_synthetic_soc_states() [all …]
|
/linux/drivers/gpu/drm/amd/display/dc/clk_mgr/dcn314/ |
H A D | dcn314_clk_mgr.c | 404 .entries = { 441 .entries = { 493 /* skip empty entries, the smu array has no holes*/ in dcn314_build_watermark_ranges() 494 if (!bw_params->wm_table.entries[i].valid) in dcn314_build_watermark_ranges() 497 table->WatermarkRow[WM_DCFCLK][num_valid_sets].WmSetting = bw_params->wm_table.entries[i].wm_inst; in dcn314_build_watermark_ranges() 498 table->WatermarkRow[WM_DCFCLK][num_valid_sets].WmType = bw_params->wm_table.entries[i].wm_type; in dcn314_build_watermark_ranges() 509 bw_params->clk_table.entries[i - 1].dcfclk_mhz + 1; in dcn314_build_watermark_ranges() 512 bw_params->clk_table.entries[i].dcfclk_mhz; in dcn314_build_watermark_ranges() 622 …struct clk_limit_table_entry def_max = bw_params->clk_table.entries[bw_params->clk_table.num_entri… in dcn314_clk_mgr_helper_populate_bw_params() 644 /* Invalid number of entries in the table from PMFW. */ in dcn314_clk_mgr_helper_populate_bw_params() [all …]
|
/linux/drivers/misc/vmw_vmci/ |
H A D | vmci_handle_array.c | 22 array = kmalloc(struct_size(array, entries, capacity), GFP_ATOMIC); in vmci_handle_arr_create() 48 size_t new_size = struct_size(array, entries, in vmci_handle_arr_append_entry() 62 array->entries[array->size] = handle; in vmci_handle_arr_append_entry() 78 if (vmci_handle_is_equal(array->entries[i], entry_handle)) { in vmci_handle_arr_remove_entry() 79 handle = array->entries[i]; in vmci_handle_arr_remove_entry() 81 array->entries[i] = array->entries[array->size]; in vmci_handle_arr_remove_entry() 82 array->entries[array->size] = VMCI_INVALID_HANDLE; in vmci_handle_arr_remove_entry() 99 handle = array->entries[array->size]; in vmci_handle_arr_remove_tail() 100 array->entries[array->size] = VMCI_INVALID_HANDLE; in vmci_handle_arr_remove_tail() 115 return array->entries[index]; in vmci_handle_arr_get_entry() [all …]
|
/linux/tools/perf/tests/ |
H A D | fdarray.c | 14 fda->entries[fd].fd = fda->nr - fd; in fdarray__init_revents() 15 fda->entries[fd].events = revents; in fdarray__init_revents() 16 fda->entries[fd].revents = revents; in fdarray__init_revents() 58 fda->entries[2].revents = POLLIN; in test__fdarray__filter() 60 pr_debug("\nfiltering all but fda->entries[2]:"); in test__fdarray__filter() 70 fda->entries[0].revents = POLLIN; in test__fdarray__filter() 71 fda->entries[3].revents = POLLIN; in test__fdarray__filter() 73 pr_debug("\nfiltering all but (fda->entries[0], fda->entries[3]):"); in test__fdarray__filter() 103 if (fda->entries[_idx].fd != _fd) { \ in test__fdarray__add() 104 pr_debug("\n%d: fda->entries[%d](%d) != %d!", \ in test__fdarray__add() [all …]
|
/linux/fs/nfs_common/ |
H A D | nfsacl.c | 13 * four instead of three entries. 16 * the ACL_MASK and ACL_GROUP_OBJ entries may differ.) 18 * entries contain the identifiers of the owner and owning group. 20 * - ACL entries in the kernel are kept sorted in ascending order 96 int entries = (acl && acl->a_count) ? max_t(int, acl->a_count, 4) : 0; in nfsacl_encode() local 100 .array_len = encode_entries ? entries : 0, in nfsacl_encode() 111 if (entries > NFS_ACL_MAX_ENTRIES || in nfsacl_encode() 112 xdr_encode_word(buf, base, entries)) in nfsacl_encode() 123 /* Insert entries in canonical order: other orders seem in nfsacl_encode() 158 u32 entries = (acl && acl->a_count) ? max_t(int, acl->a_count, 4) : 0; in nfs_stream_encode_acl() local [all …]
|