| /linux/drivers/net/ethernet/marvell/octeontx2/af/ |
| H A D | rvu_npc_hash.c | 448 hash &= table->mem_table.hash_mask; in rvu_exact_calculate_hash() 449 hash += table->mem_table.hash_offset; in rvu_exact_calculate_hash() 472 depth = table->mem_table.depth; in rvu_npc_exact_alloc_mem_table_entry() 476 for (i = 0; i < table->mem_table.ways; i++) { in rvu_npc_exact_alloc_mem_table_entry() 477 if (test_bit(hash + i * depth, table->mem_table.bmap)) in rvu_npc_exact_alloc_mem_table_entry() 480 set_bit(hash + i * depth, table->mem_table.bmap); in rvu_npc_exact_alloc_mem_table_entry() 493 bitmap_weight(table->mem_table.bmap, table->mem_table.depth)); in rvu_npc_exact_alloc_mem_table_entry() 677 rvu->hw->table->mem_table.hash_mask = (depth - 1); in rvu_exact_config_result_ctrl() 681 rvu->hw->table->mem_table.hash_offset = 0; in rvu_exact_config_result_ctrl() 710 rvu->hw->table->mem_table.mask = mask; in rvu_exact_config_table_mask() [all …]
|
| H A D | rvu_npc_hash.h | 182 } mem_table; member
|
| /linux/arch/parisc/kernel/ |
| H A D | inventory.c | 319 struct pdc_pat_pd_addr_map_entry mem_table[PAT_MAX_RANGES+1]; in pat_memconfig() local 329 status = pdc_pat_pd_get_addr_map(&actual_len, mem_table, length, 0L); in pat_memconfig() 358 mtbl_ptr = mem_table; in pat_memconfig() 413 struct pdc_memory_table mem_table[MAX_PHYSMEM_RANGES]; in sprockets_memconfig() local 420 status = pdc_mem_mem_table(&r_addr,mem_table, in sprockets_memconfig() 443 mtbl_ptr = mem_table; in sprockets_memconfig()
|
| /linux/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
| H A D | vega20_hwmgr.c | 621 dpm_table = &(data->dpm_table.mem_table); in vega20_setup_memclk_dpm_table() 673 dpm_table = &(data->dpm_table.mem_table); in vega20_setup_default_dpm_tables() 1078 data->dpm_table.mem_table.dpm_levels[data->dpm_table.mem_table.count - 2].value; in vega20_od8_set_feature_capabilities() 1527 &(data->dpm_table.mem_table); in vega20_get_mclk_od() 1529 &(data->golden_dpm_table.mem_table); in vega20_get_mclk_od() 1546 &(data->golden_dpm_table.mem_table); in vega20_set_mclk_od() 1572 struct vega20_single_dpm_table *mem_table = &(data->dpm_table.mem_table); in vega20_populate_umdpstate_clocks() local 1575 mem_table->count > VEGA20_UMD_PSTATE_MCLK_LEVEL) { in vega20_populate_umdpstate_clocks() 1577 hwmgr->pstate_mclk = mem_table->dpm_levels[VEGA20_UMD_PSTATE_MCLK_LEVEL].value; in vega20_populate_umdpstate_clocks() 1580 hwmgr->pstate_mclk = mem_table->dpm_levels[0].value; in vega20_populate_umdpstate_clocks() [all …]
|
| H A D | vega12_hwmgr.c | 680 dpm_table = &(data->dpm_table.mem_table); in vega12_setup_default_dpm_tables() 1042 struct vega12_single_dpm_table *mem_dpm_table = &(data->dpm_table.mem_table); in vega12_populate_umdpstate_clocks() 1176 min_freq = data->dpm_table.mem_table.dpm_state.soft_min_level; in vega12_upload_dpm_min_level() 1184 min_freq = data->dpm_table.mem_table.dpm_state.hard_min_level; in vega12_upload_dpm_min_level() 1268 max_freq = data->dpm_table.mem_table.dpm_state.soft_max_level; in vega12_upload_dpm_max_level() 1667 soft_level = vega12_find_highest_dpm_level(&(data->dpm_table.mem_table)); in vega12_force_dpm_highest() 1669 data->dpm_table.mem_table.dpm_state.soft_min_level = in vega12_force_dpm_highest() 1670 data->dpm_table.mem_table.dpm_state.soft_max_level = in vega12_force_dpm_highest() 1671 data->dpm_table.mem_table.dpm_levels[soft_level].value; in vega12_force_dpm_highest() 1696 soft_level = vega12_find_lowest_dpm_level(&(data->dpm_table.mem_table)); in vega12_force_dpm_lowest() [all …]
|
| H A D | vega10_hwmgr.c | 1370 data->dpm_table.mem_table.count = 0; in vega10_setup_default_dpm_tables() 1371 dpm_table = &(data->dpm_table.mem_table); in vega10_setup_default_dpm_tables() 1882 &(data->dpm_table.mem_table); in vega10_populate_all_memory_levels() 3439 struct vega10_single_dpm_table *mclk_table = &(data->dpm_table.mem_table); in vega10_find_dpm_states_clocks_in_dpm_table() 3500 for (count = 0; count < dpm_table->mem_table.count; count++) in vega10_populate_and_upload_sclk_mclk_dpm_levels() 3501 dpm_table->mem_table.dpm_levels[count].value = odn_clk_table->entries[count].clk; in vega10_populate_and_upload_sclk_mclk_dpm_levels() 3584 &(data->dpm_table.mem_table), in vega10_trim_dpm_states() 3661 data->dpm_table.mem_table.dpm_state.soft_min_level) { in vega10_upload_dpm_bootup_level() 3675 data->dpm_table.mem_table.dpm_state.soft_min_level = in vega10_upload_dpm_bootup_level() 3718 data->dpm_table.mem_table.dpm_state.soft_max_level) { in vega10_upload_dpm_max_level() [all …]
|
| H A D | vega12_hwmgr.h | 127 struct vega12_single_dpm_table mem_table; member
|
| H A D | vega10_hwmgr.h | 149 struct vega10_single_dpm_table mem_table; member
|
| H A D | vega20_hwmgr.h | 180 struct vega20_single_dpm_table mem_table; member
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu13/ |
| H A D | aldebaran_ppt.h | 62 struct aldebaran_single_dpm_table mem_table; member
|
| H A D | aldebaran_ppt.c | 587 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in aldebaran_populate_umd_state_clk() local 597 pstate_table->uclk_pstate.min = SMU_DPM_TABLE_MIN(mem_table); in aldebaran_populate_umd_state_clk() 598 pstate_table->uclk_pstate.peak = SMU_DPM_TABLE_MAX(mem_table); in aldebaran_populate_umd_state_clk() 599 pstate_table->uclk_pstate.curr.min = SMU_DPM_TABLE_MIN(mem_table); in aldebaran_populate_umd_state_clk() 600 pstate_table->uclk_pstate.curr.max = SMU_DPM_TABLE_MAX(mem_table); in aldebaran_populate_umd_state_clk() 608 mem_table->count > ALDEBARAN_UMD_PSTATE_MCLK_LEVEL && in aldebaran_populate_umd_state_clk() 613 mem_table->dpm_levels[ALDEBARAN_UMD_PSTATE_MCLK_LEVEL].value; in aldebaran_populate_umd_state_clk()
|
| H A D | smu_v13_0.c | 1588 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in smu_v13_0_set_performance_level() local 1608 mclk_min = mclk_max = SMU_DPM_TABLE_MAX(mem_table); in smu_v13_0_set_performance_level() 1616 mclk_min = mclk_max = SMU_DPM_TABLE_MIN(mem_table); in smu_v13_0_set_performance_level() 1625 mclk_min = SMU_DPM_TABLE_MIN(mem_table); in smu_v13_0_set_performance_level() 1626 mclk_max = SMU_DPM_TABLE_MAX(mem_table); in smu_v13_0_set_performance_level()
|
| H A D | smu_v13_0_6_ppt.c | 1197 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in smu_v13_0_6_populate_umd_state_clk() local 1206 pstate_table->uclk_pstate.min = SMU_DPM_TABLE_MIN(mem_table); in smu_v13_0_6_populate_umd_state_clk() 1207 pstate_table->uclk_pstate.peak = SMU_DPM_TABLE_MAX(mem_table); in smu_v13_0_6_populate_umd_state_clk() 1208 pstate_table->uclk_pstate.curr.min = SMU_DPM_TABLE_MIN(mem_table); in smu_v13_0_6_populate_umd_state_clk() 1209 pstate_table->uclk_pstate.curr.max = SMU_DPM_TABLE_MAX(mem_table); in smu_v13_0_6_populate_umd_state_clk() 1217 mem_table->count > SMU_13_0_6_UMD_PSTATE_MCLK_LEVEL && in smu_v13_0_6_populate_umd_state_clk() 1222 mem_table->dpm_levels[SMU_13_0_6_UMD_PSTATE_MCLK_LEVEL].value; in smu_v13_0_6_populate_umd_state_clk()
|
| H A D | smu_v13_0_7_ppt.c | 2241 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in smu_v13_0_7_populate_umd_state_clk() local 2260 pstate_table->uclk_pstate.min = SMU_DPM_TABLE_MIN(mem_table); in smu_v13_0_7_populate_umd_state_clk() 2261 pstate_table->uclk_pstate.peak = SMU_DPM_TABLE_MAX(mem_table); in smu_v13_0_7_populate_umd_state_clk() 2281 pstate_table->uclk_pstate.standard = SMU_DPM_TABLE_MAX(mem_table); in smu_v13_0_7_populate_umd_state_clk()
|
| H A D | smu_v13_0_0_ppt.c | 2239 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in smu_v13_0_0_populate_umd_state_clk() local 2258 pstate_table->uclk_pstate.min = SMU_DPM_TABLE_MIN(mem_table); in smu_v13_0_0_populate_umd_state_clk() 2259 pstate_table->uclk_pstate.peak = SMU_DPM_TABLE_MAX(mem_table); in smu_v13_0_0_populate_umd_state_clk() 2279 pstate_table->uclk_pstate.standard = SMU_DPM_TABLE_MAX(mem_table); in smu_v13_0_0_populate_umd_state_clk()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu11/ |
| H A D | arcturus_ppt.h | 62 struct arcturus_single_dpm_table mem_table; member
|
| H A D | arcturus_ppt.c | 566 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in arcturus_populate_umd_state_clk() local 574 pstate_table->uclk_pstate.min = SMU_DPM_TABLE_MIN(mem_table); in arcturus_populate_umd_state_clk() 575 pstate_table->uclk_pstate.peak = SMU_DPM_TABLE_MAX(mem_table); in arcturus_populate_umd_state_clk() 581 mem_table->count > ARCTURUS_UMD_PSTATE_MCLK_LEVEL && in arcturus_populate_umd_state_clk() 586 mem_table->dpm_levels[ARCTURUS_UMD_PSTATE_MCLK_LEVEL].value; in arcturus_populate_umd_state_clk()
|
| H A D | smu_v11_0.c | 1863 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in smu_v11_0_set_performance_level() local 1877 mclk_min = mclk_max = SMU_DPM_TABLE_MAX(mem_table); in smu_v11_0_set_performance_level() 1882 mclk_min = mclk_max = SMU_DPM_TABLE_MIN(mem_table); in smu_v11_0_set_performance_level() 1888 mclk_min = SMU_DPM_TABLE_MIN(mem_table); in smu_v11_0_set_performance_level() 1889 mclk_max = SMU_DPM_TABLE_MAX(mem_table); in smu_v11_0_set_performance_level()
|
| H A D | navi10_ppt.c | 1462 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in navi10_populate_umd_state_clk() local 1518 pstate_table->uclk_pstate.min = SMU_DPM_TABLE_MIN(mem_table); in navi10_populate_umd_state_clk() 1519 pstate_table->uclk_pstate.peak = SMU_DPM_TABLE_MAX(mem_table); in navi10_populate_umd_state_clk() 1525 SMU_DPM_TABLE_MAX(mem_table) > NAVI10_UMD_PSTATE_PROFILING_MEMCLK && in navi10_populate_umd_state_clk()
|
| H A D | sienna_cichlid_ppt.c | 1452 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in sienna_cichlid_populate_umd_state_clk() local 1461 pstate_table->uclk_pstate.min = SMU_DPM_TABLE_MIN(mem_table); in sienna_cichlid_populate_umd_state_clk() 1462 pstate_table->uclk_pstate.peak = SMU_DPM_TABLE_MAX(mem_table); in sienna_cichlid_populate_umd_state_clk()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu15/ |
| H A D | smu_v15_0.c | 1134 struct smu_15_0_dpm_table *mem_table = in smu_v15_0_set_performance_level() local 1159 mclk_min = mclk_max = mem_table->max; in smu_v15_0_set_performance_level() 1167 mclk_min = mclk_max = mem_table->min; in smu_v15_0_set_performance_level() 1176 mclk_min = mem_table->min; in smu_v15_0_set_performance_level() 1177 mclk_max = mem_table->max; in smu_v15_0_set_performance_level()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu14/ |
| H A D | smu_v14_0.c | 1250 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in smu_v14_0_set_performance_level() local 1270 mclk_min = mclk_max = SMU_DPM_TABLE_MAX(mem_table); in smu_v14_0_set_performance_level() 1278 mclk_min = mclk_max = SMU_DPM_TABLE_MIN(mem_table); in smu_v14_0_set_performance_level() 1287 mclk_min = SMU_DPM_TABLE_MIN(mem_table); in smu_v14_0_set_performance_level() 1288 mclk_max = SMU_DPM_TABLE_MAX(mem_table); in smu_v14_0_set_performance_level()
|
| H A D | smu_v14_0_2_ppt.c | 1478 struct smu_dpm_table *mem_table = &dpm_context->dpm_tables.uclk_table; in smu_v14_0_2_populate_umd_state_clk() local 1497 pstate_table->uclk_pstate.min = SMU_DPM_TABLE_MIN(mem_table); in smu_v14_0_2_populate_umd_state_clk() 1498 pstate_table->uclk_pstate.peak = SMU_DPM_TABLE_MAX(mem_table); in smu_v14_0_2_populate_umd_state_clk() 1518 pstate_table->uclk_pstate.standard = SMU_DPM_TABLE_MAX(mem_table); in smu_v14_0_2_populate_umd_state_clk()
|