/linux/drivers/gpu/drm/amd/display/dc/spl/ |
H A D | dc_spl_scl_easf_filters.c | 1523 unsigned int num_entries) in spl_easf_get_scale_ratio_to_reg_value() argument 1529 lookup_table_index_ptr = (lookup_table_base_ptr + num_entries - 1); in spl_easf_get_scale_ratio_to_reg_value() 1532 while (count < num_entries) { in spl_easf_get_scale_ratio_to_reg_value() 1552 unsigned int num_entries = sizeof(easf_v_bf3_mode_lookup) / in spl_get_v_bf3_mode() local 1555 easf_v_bf3_mode_lookup, num_entries); in spl_get_v_bf3_mode() 1561 unsigned int num_entries = sizeof(easf_h_bf3_mode_lookup) / in spl_get_h_bf3_mode() local 1564 easf_h_bf3_mode_lookup, num_entries); in spl_get_h_bf3_mode() 1570 unsigned int num_entries; in spl_get_reducer_gain6() local 1573 num_entries = sizeof(easf_reducer_gain6_4tap_lookup) / in spl_get_reducer_gain6() 1576 easf_reducer_gain6_4tap_lookup, num_entries); in spl_get_reducer_gain6() [all …]
|
/linux/drivers/gpu/drm/i915/display/ |
H A D | intel_ddi_buf_trans.c | 32 .num_entries = ARRAY_SIZE(_hsw_trans_dp), 49 .num_entries = ARRAY_SIZE(_hsw_trans_fdi), 70 .num_entries = ARRAY_SIZE(_hsw_trans_hdmi), 88 .num_entries = ARRAY_SIZE(_bdw_trans_edp), 105 .num_entries = ARRAY_SIZE(_bdw_trans_dp), 122 .num_entries = ARRAY_SIZE(_bdw_trans_fdi), 141 .num_entries = ARRAY_SIZE(_bdw_trans_hdmi), 160 .num_entries = ARRAY_SIZE(_skl_trans_dp), 178 .num_entries = ARRAY_SIZE(_skl_u_trans_dp), 196 .num_entries = ARRAY_SIZE(_skl_y_trans_dp), [all …]
|
/linux/drivers/char/agp/ |
H A D | generic.c | 321 int num_entries; in agp_num_entries() local 328 num_entries = A_SIZE_8(temp)->num_entries; in agp_num_entries() 331 num_entries = A_SIZE_16(temp)->num_entries; in agp_num_entries() 334 num_entries = A_SIZE_32(temp)->num_entries; in agp_num_entries() 337 num_entries = A_SIZE_LVL2(temp)->num_entries; in agp_num_entries() 340 num_entries = A_SIZE_FIX(temp)->num_entries; in agp_num_entries() 343 num_entries = 0; in agp_num_entries() 347 num_entries -= agp_memory_reserved>>PAGE_SHIFT; in agp_num_entries() 348 if (num_entries<0) in agp_num_entries() 349 num_entries = 0; in agp_num_entries() [all …]
|
H A D | efficeon-agp.c | 198 int num_entries, l1_pages; in efficeon_create_gatt_table() local 200 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_create_gatt_table() 202 printk(KERN_DEBUG PFX "efficeon_create_gatt_table(%d)\n", num_entries); in efficeon_create_gatt_table() 205 BUG_ON(num_entries & 0x3ff); in efficeon_create_gatt_table() 206 l1_pages = num_entries >> 10; in efficeon_create_gatt_table() 238 int i, count = mem->page_count, num_entries; in efficeon_insert_memory() local 245 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_insert_memory() 246 if ((pg_start + mem->page_count) > num_entries) in efficeon_insert_memory() 287 int i, count = mem->page_count, num_entries; in efficeon_remove_memory() local 291 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_remove_memory() [all …]
|
H A D | alpha-agp.c | 88 int num_entries, status; in alpha_core_agp_insert_memory() local 95 num_entries = A_SIZE_FIX(temp)->num_entries; in alpha_core_agp_insert_memory() 96 if ((pg_start + mem->page_count) > num_entries) in alpha_core_agp_insert_memory() 169 aper_size->num_entries = agp->aperture.size / PAGE_SIZE; in alpha_core_agp_setup() 170 aper_size->page_order = __ffs(aper_size->num_entries / 1024); in alpha_core_agp_setup()
|
/linux/include/net/tc_act/ |
H A D | tc_gate.h | 33 size_t num_entries; member 101 u32 num_entries; in tcf_gate_num_entries() local 103 num_entries = to_gate(a)->param.num_entries; in tcf_gate_num_entries() 105 return num_entries; in tcf_gate_num_entries() 114 u32 num_entries; in tcf_gate_get_list() local 118 num_entries = p->num_entries; in tcf_gate_get_list() 123 if (i != num_entries) in tcf_gate_get_list() 126 oe = kcalloc(num_entries, sizeof(*oe), GFP_ATOMIC); in tcf_gate_get_list()
|
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn321/ |
H A D | dcn321_fpu.c | 211 unsigned int *num_entries, in dcn321_insert_entry_into_table_sorted() argument 219 if (*num_entries == 0) { in dcn321_insert_entry_into_table_sorted() 221 (*num_entries)++; in dcn321_insert_entry_into_table_sorted() 225 if (index >= *num_entries) in dcn321_insert_entry_into_table_sorted() 229 for (i = *num_entries; i > index; i--) in dcn321_insert_entry_into_table_sorted() 233 (*num_entries)++; in dcn321_insert_entry_into_table_sorted() 237 …ove_entry_from_table_at_index(struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries, in remove_entry_from_table_at_index() argument 242 if (*num_entries == 0) in remove_entry_from_table_at_index() 245 for (i = index; i < *num_entries - 1; i++) { in remove_entry_from_table_at_index() 248 memset(&table[--(*num_entries)], 0, sizeof(struct _vcs_dpi_voltage_scaling_st)); in remove_entry_from_table_at_index() [all …]
|
/linux/arch/riscv/kernel/ |
H A D | module-sections.c | 16 int i = got_sec->num_entries; in module_emit_got_entry() 26 got_sec->num_entries++; in module_emit_got_entry() 27 BUG_ON(got_sec->num_entries > got_sec->max_entries); in module_emit_got_entry() 38 int i = plt_sec->num_entries; in module_emit_plt_entry() 51 plt_sec->num_entries++; in module_emit_plt_entry() 52 got_plt_sec->num_entries++; in module_emit_plt_entry() 53 BUG_ON(plt_sec->num_entries > plt_sec->max_entries); in module_emit_plt_entry() 142 mod->arch.plt.num_entries = 0; in module_frob_arch_sections() 149 mod->arch.got.num_entries = 0; in module_frob_arch_sections() 156 mod->arch.got_plt.num_entries = 0; in module_frob_arch_sections()
|
/linux/arch/loongarch/kernel/ |
H A D | module-sections.c | 15 int i = got_sec->num_entries; in module_emit_got_entry() 25 got_sec->num_entries++; in module_emit_got_entry() 26 if (got_sec->num_entries > got_sec->max_entries) { in module_emit_got_entry() 49 nr = plt_sec->num_entries; in module_emit_plt_entry() 57 plt_sec->num_entries++; in module_emit_plt_entry() 58 plt_idx_sec->num_entries++; in module_emit_plt_entry() 59 BUG_ON(plt_sec->num_entries > plt_sec->max_entries); in module_emit_plt_entry() 158 mod->arch.got.num_entries = 0; in module_frob_arch_sections() 166 mod->arch.plt.num_entries = 0; in module_frob_arch_sections() 174 mod->arch.plt_idx.num_entries = 0; in module_frob_arch_sections()
|
H A D | unwind_orc.c | 59 unsigned int num_entries, unsigned long ip) in __orc_find() argument 63 int *last = ip_table + num_entries - 1; in __orc_find() 65 if (!num_entries) in __orc_find() 247 unsigned int num_entries = orc_ip_size / sizeof(int); in unwind_module_init() local 251 num_entries != orc_size / sizeof(*orc)); in unwind_module_init() 261 sort(orc_ip, num_entries, sizeof(int), orc_sort_cmp, orc_sort_swap); in unwind_module_init() 266 mod->arch.num_orcs = num_entries; in unwind_module_init() 275 size_t num_entries = orc_ip_size / sizeof(int); in unwind_init() local 278 if (!num_entries || orc_ip_size % sizeof(int) != 0 || in unwind_init() 280 num_entries != orc_size / sizeof(struct orc_entry)) { in unwind_init() [all …]
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_bo_list.h | 53 unsigned num_entries; member 59 struct amdgpu_bo_list_entry entries[] __counted_by(num_entries); 71 size_t num_entries, 76 e != &list->entries[list->num_entries]; \ 81 e != &list->entries[list->num_entries]; \
|
H A D | amdgpu_bo_list.c | 72 size_t num_entries, struct amdgpu_bo_list **result) in amdgpu_bo_list_create() argument 74 unsigned last_entry = 0, first_userptr = num_entries; in amdgpu_bo_list_create() 81 list = kvzalloc(struct_size(list, entries, num_entries), GFP_KERNEL); in amdgpu_bo_list_create() 87 list->num_entries = num_entries; in amdgpu_bo_list_create() 90 for (i = 0; i < num_entries; ++i) { in amdgpu_bo_list_create() 136 trace_amdgpu_cs_bo_status(list->num_entries, total_size); in amdgpu_bo_list_create() 145 for (i = first_userptr; i < num_entries; ++i) in amdgpu_bo_list_create()
|
/linux/drivers/net/wwan/iosm/ |
H A D | iosm_ipc_coredump.c | 66 u32 byte_read, num_entries, file_size; in ipc_coredump_get_list() local 95 num_entries = le32_to_cpu(cd_table->list.num_entries); in ipc_coredump_get_list() 96 if (num_entries == 0 || num_entries > IOSM_NOF_CD_REGION) { in ipc_coredump_get_list() 101 for (i = 0; i < num_entries; i++) { in ipc_coredump_get_list()
|
/linux/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_shared_buf.c | 81 unsigned int i, num_entries, entry_sz; in nfp_shared_buf_register() local 92 num_entries = n; in nfp_shared_buf_register() 95 num_entries * sizeof(pf->shared_bufs[0]), in nfp_shared_buf_register() 100 entry_sz = nfp_cpp_area_size(sb_desc_area) / num_entries; in nfp_shared_buf_register() 102 pf->shared_bufs = kmalloc_array(num_entries, sizeof(pf->shared_bufs[0]), in nfp_shared_buf_register() 109 for (i = 0; i < num_entries; i++) { in nfp_shared_buf_register() 125 pf->num_shared_bufs = num_entries; in nfp_shared_buf_register()
|
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn35/ |
H A D | dcn35_fpu.c | 244 ASSERT(clk_table->num_entries); in dcn35_update_bw_bounding_box_fpu() 247 for (i = 0; i < clk_table->num_entries; ++i) { in dcn35_update_bw_bounding_box_fpu() 254 for (i = 0; i < clk_table->num_entries; i++) { in dcn35_update_bw_bounding_box_fpu() 264 if (clk_table->num_entries == 1) { in dcn35_update_bw_bounding_box_fpu() 273 if (clk_table->num_entries == 1 && in dcn35_update_bw_bounding_box_fpu() 316 if (clk_table->num_entries) in dcn35_update_bw_bounding_box_fpu() 317 dcn3_5_soc.num_states = clk_table->num_entries; in dcn35_update_bw_bounding_box_fpu() 353 if (clk_table->num_entries > 2) { in dcn35_update_bw_bounding_box_fpu() 355 for (i = 0; i < clk_table->num_entries; i++) { in dcn35_update_bw_bounding_box_fpu() 357 clk_table->num_entries; in dcn35_update_bw_bounding_box_fpu() [all …]
|
/linux/drivers/net/dsa/sja1105/ |
H A D | sja1105_tas.c | 172 int num_entries = 0; in sja1105_init_scheduling() local 213 num_entries += tas_data->offload[port]->num_entries; in sja1105_init_scheduling() 219 num_entries += gating_cfg->num_entries; in sja1105_init_scheduling() 231 table->entries = kcalloc(num_entries, table->ops->unpacked_entry_size, in sja1105_init_scheduling() 235 table->entry_count = num_entries; in sja1105_init_scheduling() 283 schedule_end_idx = k + offload->num_entries - 1; in sja1105_init_scheduling() 310 for (i = 0; i < offload->num_entries; i++, k++) { in sja1105_init_scheduling() 329 schedule_end_idx = k + gating_cfg->num_entries - 1; in sja1105_init_scheduling() 425 i < offload->num_entries; in sja1105_tas_check_conflicts() 431 j < admin->num_entries; in sja1105_tas_check_conflicts() [all …]
|
/linux/drivers/net/ethernet/engleder/ |
H A D | tsnep_selftests.c | 381 qopt->num_entries = 7; in tsnep_test_taprio() 405 qopt->num_entries = 8; in tsnep_test_taprio() 434 qopt->num_entries = 10; in tsnep_test_taprio() 468 qopt->num_entries = 2; in tsnep_test_taprio_change() 501 qopt->num_entries = 3; in tsnep_test_taprio_change() 513 qopt->num_entries = 2; in tsnep_test_taprio_change() 527 qopt->num_entries = 4; in tsnep_test_taprio_change() 539 qopt->num_entries = 2; in tsnep_test_taprio_change() 551 qopt->num_entries = 3; in tsnep_test_taprio_change() 567 qopt->num_entries = 4; in tsnep_test_taprio_change() [all …]
|
/linux/fs/exfat/ |
H A D | dir.c | 49 for (i = ES_IDX_FIRST_FILENAME; i < es.num_entries; i++) { in exfat_get_uniname_from_ext_entry() 484 void exfat_init_ext_entry(struct exfat_entry_set_cache *es, int num_entries, in exfat_init_ext_entry() argument 492 ep->dentry.file.num_ext = (unsigned char)(num_entries - 1); in exfat_init_ext_entry() 498 for (i = ES_IDX_FIRST_FILENAME; i < num_entries; i++) { in exfat_init_ext_entry() 513 for (i = order; i < es->num_entries; i++) { in exfat_remove_entries() 522 if (order < es->num_entries) in exfat_remove_entries() 532 for (i = ES_IDX_FILE; i < es->num_entries; i++) { in exfat_update_dir_chksum() 754 unsigned int num_entries) in __exfat_get_dentry_set() argument 782 if (num_entries == ES_ALL_ENTRIES) { in __exfat_get_dentry_set() 791 num_entries = ep->dentry.file.num_ext + 1; in __exfat_get_dentry_set() [all …]
|
/linux/drivers/gpu/drm/amd/display/dc/dml2/dml21/src/dml2_mcg/ |
H A D | dml2_mcg_dcn4.c | 67 min_table->dram_bw_table.num_entries = soc_bb->clk_table.uclk.num_clk_values; in build_min_clk_table_fine_grained() 70 for (i = min_table->dram_bw_table.num_entries - 1; i > 0; i--) { in build_min_clk_table_fine_grained() 82 for (i = 0; i < (int)min_table->dram_bw_table.num_entries; i++) { in build_min_clk_table_fine_grained() 106 for (i = 0; i < (int)min_table->dram_bw_table.num_entries; i++) { in build_min_clk_table_fine_grained() 109 min_table->dram_bw_table.num_entries = i; in build_min_clk_table_fine_grained() 115 for (i = 0; i < (int)min_table->dram_bw_table.num_entries - 1; i++) { in build_min_clk_table_fine_grained() 121 for (j = i + 1; j < min_table->dram_bw_table.num_entries; j++) { in build_min_clk_table_fine_grained() 126 min_table->dram_bw_table.num_entries--; in build_min_clk_table_fine_grained() 142 min_table->dram_bw_table.num_entries = soc_bb->clk_table.uclk.num_clk_values; in build_min_clk_table_coarse_grained()
|
/linux/drivers/iommu/ |
H A D | io-pgtable-dart.c | 120 dart_iopte prot, int num_entries, in dart_init_pte() argument 127 for (i = 0; i < num_entries; i++) in dart_init_pte() 141 for (i = 0; i < num_entries; i++) in dart_init_pte() 238 int ret = 0, tbl, num_entries, max_entries, map_idx_start; in dart_map_pages() local 277 num_entries = min_t(int, pgcount, max_entries); in dart_map_pages() 279 ret = dart_init_pte(data, iova, paddr, prot, num_entries, ptep); in dart_map_pages() 281 *mapped += num_entries * pgsize; in dart_map_pages() 298 int i = 0, num_entries, max_entries, unmap_idx_start; in dart_unmap_pages() local 314 num_entries = min_t(int, pgcount, max_entries); in dart_unmap_pages() 316 while (i < num_entries) { in dart_unmap_pages()
|
/linux/kernel/kcsan/ |
H A D | report.c | 277 static int get_stack_skipnr(const unsigned long stack_entries[], int num_entries) in get_stack_skipnr() argument 283 for (skip = 0; skip < num_entries; ++skip) { in get_stack_skipnr() 315 replace_stack_entry(unsigned long stack_entries[], int num_entries, unsigned long ip, in replace_stack_entry() argument 327 for (skip = 0; skip < num_entries; ++skip) { in replace_stack_entry() 344 return get_stack_skipnr(stack_entries, num_entries); in replace_stack_entry() 348 sanitize_stack_entries(unsigned long stack_entries[], int num_entries, unsigned long ip, in sanitize_stack_entries() argument 351 return ip ? replace_stack_entry(stack_entries, num_entries, ip, replaced) : in sanitize_stack_entries() 352 get_stack_skipnr(stack_entries, num_entries); in sanitize_stack_entries() 368 print_stack_trace(unsigned long stack_entries[], int num_entries, unsigned long reordered_to) in print_stack_trace() argument 370 stack_trace_print(stack_entries, num_entries, 0); in print_stack_trace()
|
/linux/drivers/parisc/ |
H A D | iosapic.c | 222 static struct irt_entry *iosapic_alloc_irt(int num_entries) in iosapic_alloc_irt() argument 224 return kcalloc(num_entries, sizeof(struct irt_entry), GFP_KERNEL); in iosapic_alloc_irt() 259 unsigned long num_entries = 0UL; in iosapic_load_irt() local 266 status = pdc_pat_get_irt_size(&num_entries, cell_num); in iosapic_load_irt() 270 BUG_ON(num_entries == 0); in iosapic_load_irt() 278 table = iosapic_alloc_irt(num_entries); in iosapic_load_irt() 299 status = pdc_pci_irt_size(&num_entries, 0); in iosapic_load_irt() 307 BUG_ON(num_entries == 0); in iosapic_load_irt() 309 table = iosapic_alloc_irt(num_entries); in iosapic_load_irt() 317 status = pdc_pci_irt(num_entries, 0, table); in iosapic_load_irt() [all …]
|
/linux/drivers/pci/ |
H A D | tph.c | 469 int num_entries, i, offset; in pci_restore_tph_state() local 488 num_entries = get_st_table_size(pdev); in pci_restore_tph_state() 489 for (i = 0; i < num_entries; i++) { in pci_restore_tph_state() 499 int num_entries, i, offset; in pci_save_tph_state() local 520 num_entries = get_st_table_size(pdev); in pci_save_tph_state() 521 for (i = 0; i < num_entries; i++) { in pci_save_tph_state() 537 int num_entries; in pci_tph_init() local 544 num_entries = get_st_table_size(pdev); in pci_tph_init() 545 save_size = sizeof(u32) + num_entries * sizeof(u16); in pci_tph_init()
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_pt.c | 611 struct xe_vm_pgtable_update *entries, u32 *num_entries) in xe_pt_stage_bind() argument 698 *num_entries = xe_walk.wupd.num_used_entries; in xe_pt_stage_bind() 861 u32 num_entries) in xe_pt_cancel_bind() argument 865 for (i = 0; i < num_entries; i++) { in xe_pt_cancel_bind() 907 u32 num_entries, struct llist_head *deferred) in xe_pt_commit() argument 913 for (i = 0; i < num_entries; i++) { in xe_pt_commit() 933 u32 num_entries, bool rebind) in xe_pt_abort_bind() argument 939 for (i = num_entries - 1; i >= 0; --i) { in xe_pt_abort_bind() 963 u32 num_entries, bool rebind) in xe_pt_commit_prepare_bind() argument 969 for (i = 0; i < num_entries; i++) { in xe_pt_commit_prepare_bind() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | spectrum_nve.c | 65 unsigned int num_entries; member 311 WARN_ON(mc_record->num_entries); in mlxsw_sp_nve_mc_record_destroy() 323 unsigned int num_entries = mc_record->num_entries; in mlxsw_sp_nve_mc_record_get() local 327 num_entries < nve->num_max_mc_entries[proto]) in mlxsw_sp_nve_mc_record_get() 337 if (mc_record->num_entries != 0) in mlxsw_sp_nve_mc_record_put() 368 unsigned int num_entries = 0; in mlxsw_sp_nve_mc_record_refresh() local 383 next_kvdl_index, mc_record->num_entries); in mlxsw_sp_nve_mc_record_refresh() 393 num_entries++); in mlxsw_sp_nve_mc_record_refresh() 396 WARN_ON(num_entries != mc_record->num_entries); in mlxsw_sp_nve_mc_record_refresh() 449 mc_record->num_entries++; in mlxsw_sp_nve_mc_record_ip_add() [all …]
|