| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | vcn_v2_5.c | 120 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) { in vcn_v2_5_idle_work_handler() 121 struct amdgpu_vcn_inst *v = &adev->vcn.inst[i]; in vcn_v2_5_idle_work_handler() 123 if (adev->vcn.harvest_config & (1 << i)) in vcn_v2_5_idle_work_handler() 148 if (!fences && !atomic_read(&adev->vcn.inst[0].total_submission_cnt)) { in vcn_v2_5_idle_work_handler() 153 schedule_delayed_work(&adev->vcn.inst[0].idle_work, VCN_IDLE_TIMEOUT); in vcn_v2_5_idle_work_handler() 160 struct amdgpu_vcn_inst *v = &adev->vcn.inst[ring->me]; in vcn_v2_5_ring_begin_use() 162 atomic_inc(&adev->vcn.inst[0].total_submission_cnt); in vcn_v2_5_ring_begin_use() 164 cancel_delayed_work_sync(&adev->vcn.inst[0].idle_work); in vcn_v2_5_ring_begin_use() 170 mutex_lock(&adev->vcn.inst[0].vcn_pg_lock); in vcn_v2_5_ring_begin_use() 196 mutex_unlock(&adev->vcn.inst[0].vcn_pg_lock); in vcn_v2_5_ring_begin_use() [all …]
|
| H A D | vcn_v5_0_0.c | 100 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) in vcn_v5_0_0_early_init() 102 adev->vcn.inst[i].num_enc_rings = 1; in vcn_v5_0_0_early_init() 107 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) { in vcn_v5_0_0_early_init() 108 adev->vcn.inst[i].set_pg_state = vcn_v5_0_0_set_pg_state; in vcn_v5_0_0_early_init() 131 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in vcn_v5_0_0_sw_init() 134 if (adev->vcn.harvest_config & (1 << i)) in vcn_v5_0_0_sw_init() 147 atomic_set(&adev->vcn.inst[i].sched_score, 0); in vcn_v5_0_0_sw_init() 151 VCN_5_0__SRCID__UVD_ENC_GENERAL_PURPOSE, &adev->vcn.inst[i].irq); in vcn_v5_0_0_sw_init() 157 VCN_5_0__SRCID_UVD_POISON, &adev->vcn.inst[i].irq); in vcn_v5_0_0_sw_init() 161 ring = &adev->vcn.inst[i].ring_enc[0]; in vcn_v5_0_0_sw_init() [all …]
|
| H A D | vcn_v4_0_5.c | 119 adev->vcn.per_inst_fw = true; in vcn_v4_0_5_early_init() 121 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) in vcn_v4_0_5_early_init() 123 adev->vcn.inst[i].num_enc_rings = 1; in vcn_v4_0_5_early_init() 127 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) { in vcn_v4_0_5_early_init() 128 adev->vcn.inst[i].set_pg_state = vcn_v4_0_5_set_pg_state; in vcn_v4_0_5_early_init() 151 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in vcn_v4_0_5_sw_init() 154 if (adev->vcn.harvest_config & (1 << i)) in vcn_v4_0_5_sw_init() 167 atomic_set(&adev->vcn.inst[i].sched_score, 0); in vcn_v4_0_5_sw_init() 171 VCN_4_0__SRCID__UVD_ENC_GENERAL_PURPOSE, &adev->vcn.inst[i].irq); in vcn_v4_0_5_sw_init() 177 VCN_4_0__SRCID_UVD_POISON, &adev->vcn.inst[i].irq); in vcn_v4_0_5_sw_init() [all …]
|
| H A D | vega10_reg_init.c | 83 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_DOORBELL64_VCN0_1; in vega10_doorbell_index_init() 84 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_DOORBELL64_VCN2_3; in vega10_doorbell_index_init() 85 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_DOORBELL64_VCN4_5; in vega10_doorbell_index_init() 86 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_DOORBELL64_VCN6_7; in vega10_doorbell_index_init()
|
| H A D | vega20_reg_init.c | 89 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_VEGA20_DOORBELL64_VCN0_1; in vega20_doorbell_index_init() 90 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_VEGA20_DOORBELL64_VCN2_3; in vega20_doorbell_index_init() 91 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_VEGA20_DOORBELL64_VCN4_5; in vega20_doorbell_index_init() 92 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_VEGA20_DOORBELL64_VCN6_7; in vega20_doorbell_index_init()
|
| H A D | soc21.c | 172 if (adev->vcn.num_vcn_inst == hweight8(adev->vcn.harvest_config)) in soc21_query_video_codecs() 181 if ((adev->vcn.harvest_config & AMDGPU_VCN_HARVEST_VCN0) || in soc21_query_video_codecs() 194 if ((adev->vcn.harvest_config & AMDGPU_VCN_HARVEST_VCN0)) { in soc21_query_video_codecs() 542 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_NAVI10_DOORBELL64_VCN0_1; in soc21_init_doorbell_index() 543 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_NAVI10_DOORBELL64_VCN2_3; in soc21_init_doorbell_index() 544 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_NAVI10_DOORBELL64_VCN4_5; in soc21_init_doorbell_index() 545 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_NAVI10_DOORBELL64_VCN6_7; in soc21_init_doorbell_index() 887 if ((adev->vcn.harvest_config & AMDGPU_VCN_HARVEST_VCN0) || in soc21_common_late_init()
|
| H A D | amdgpu_discovery.c | 424 adev->vcn.harvest_config |= AMDGPU_VCN_HARVEST_VCN1; in amdgpu_discovery_harvest_config_quirk() 425 adev->vcn.inst_mask &= ~AMDGPU_VCN_HARVEST_VCN1; in amdgpu_discovery_harvest_config_quirk() 741 adev->vcn.harvest_config |= AMDGPU_VCN_HARVEST_VCN0; in amdgpu_discovery_read_harvest_bit_per_ip() 742 adev->vcn.inst_mask &= in amdgpu_discovery_read_harvest_bit_per_ip() 747 adev->vcn.harvest_config |= AMDGPU_VCN_HARVEST_VCN1; in amdgpu_discovery_read_harvest_bit_per_ip() 748 adev->vcn.inst_mask &= in amdgpu_discovery_read_harvest_bit_per_ip() 796 adev->vcn.harvest_config |= in amdgpu_discovery_read_from_harvest_table() 801 adev->vcn.inst_mask &= in amdgpu_discovery_read_from_harvest_table() 1094 harvest = ((1 << inst) & adev->vcn.inst_mask) == 0; in amdgpu_discovery_get_harvest_info() 1410 adev->vcn.inst_mask = 0; in amdgpu_discovery_reg_base_init() [all …]
|
| H A D | amdgpu_kms.c | 229 fw_info->ver = adev->vcn.fw_version; in amdgpu_firmware_info() 531 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in amdgpu_hw_ip_info() 532 if (adev->vcn.harvest_config & (1 << i)) in amdgpu_hw_ip_info() 535 if (adev->vcn.inst[i].ring_dec.sched.ready && in amdgpu_hw_ip_info() 536 !adev->vcn.inst[i].ring_dec.no_user_submission) in amdgpu_hw_ip_info() 544 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in amdgpu_hw_ip_info() 545 if (adev->vcn.harvest_config & (1 << i)) in amdgpu_hw_ip_info() 548 for (j = 0; j < adev->vcn.inst[i].num_enc_rings; j++) in amdgpu_hw_ip_info() 549 if (adev->vcn.inst[i].ring_enc[j].sched.ready && in amdgpu_hw_ip_info() 550 !adev->vcn.inst[i].ring_enc[j].no_user_submission) in amdgpu_hw_ip_info() [all …]
|
| H A D | amdgpu_dev_coredump.c | 167 adev->vcn.fw_version); in amdgpu_devcoredump_fw_info()
|
| H A D | umsch_mm_v4_0.c | 226 (adev->doorbell_index.vcn.vcn_ring0_1 << 1), 0); in umsch_mm_v4_0_ring_start()
|
| H A D | amdgpu_debugfs.c | 2062 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in amdgpu_debugfs_init() 2066 if (adev->vcn.harvest_config & (1 << i)) in amdgpu_debugfs_init() 2069 amdgpu_debugfs_vcn_fwlog_init(adev, i, &adev->vcn.inst[i]); in amdgpu_debugfs_init()
|
| /linux/fs/ntfs3/ |
| H A D | ntfs_fs.h | 449 struct runs_tree *run_da, CLST vcn, CLST lcn, 469 int attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn, 472 int attr_data_get_block_locked(struct ntfs_inode *ni, CLST vcn, CLST clen, 478 CLST vcn); 507 u8 name_len, const CLST *vcn); 562 u8 name_len, const CLST *vcn, 858 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, 860 void run_truncate(struct runs_tree *run, CLST vcn); 861 void run_truncate_head(struct runs_tree *run, CLST vcn); 862 void run_truncate_around(struct runs_tree *run, CLST vcn); [all …]
|
| H A D | file.c | 318 CLST vcn = from >> sbi->cluster_bits; in ntfs_file_mmap_prepare() local 322 for (; vcn < end; vcn += len) { in ntfs_file_mmap_prepare() 323 err = attr_data_get_block(ni, vcn, 1, &lcn, in ntfs_file_mmap_prepare() 644 CLST vcn = vbo >> cluster_bits; in ntfs_fallocate() local 657 for (; vcn < cend_v; vcn += clen) { in ntfs_fallocate() 658 err = attr_data_get_block(ni, vcn, cend_v - vcn, in ntfs_fallocate() 676 for (; vcn < cend; vcn += clen) { in ntfs_fallocate() 677 err = attr_data_get_block(ni, vcn, cend - vcn, in ntfs_fallocate()
|
| H A D | frecord.c | 194 const __le16 *name, u8 name_len, const CLST *vcn, in ni_find_attr() argument 213 le = al_find_ex(ni, le_o ? *le_o : NULL, type, name, name_len, vcn); in ni_find_attr() 231 if (vcn && *vcn) in ni_find_attr() 233 } else if (!vcn) { in ni_find_attr() 236 } else if (le64_to_cpu(attr->nres.svcn) > *vcn || in ni_find_attr() 237 *vcn > le64_to_cpu(attr->nres.evcn)) { in ni_find_attr() 496 if (svcn != le64_to_cpu(le->vcn)) { in ni_repack() 569 attr->nres.svcn = le->vcn = cpu_to_le64(next_svcn); in ni_repack() 643 if (le->vcn) in ni_try_remove_attr_list() 792 le->vcn = 0; in ni_create_attr_list() [all …]
|
| H A D | fslog.c | 154 __le64 vcn; // 0x10: Vcn of dirty page member 654 u32 target_attr, u64 vcn) in find_dp() argument 660 u64 dp_vcn = le64_to_cpu(dp->vcn); in find_dp() 662 if (dp->target_attr == ta && vcn >= dp_vcn && in find_dp() 663 vcn < dp_vcn + le32_to_cpu(dp->lcns_follow)) { in find_dp() 3756 u64 size, vcn, undo_next_lsn; in log_replay() local 4216 memmove(&dp->vcn, &dp0->vcn_low, in log_replay() 4237 next->vcn == dp->vcn) { in log_replay() 4530 dp->vcn = cpu_to_le64(t64 & ~((u64)t32 - 1)); in log_replay() 4542 le64_to_cpu(dp->vcn)); in log_replay() [all …]
|
| H A D | index.c | 210 CLST vcn = off >> sbi->cluster_bits; in bmp_buf_get() local 220 &vcn, &bbuf->mi); in bmp_buf_get() 377 CLST lcn, clen, vcn, vcn_next; in scan_nres_bitmap() local 388 vcn = vbo >> sbi->cluster_bits; in scan_nres_bitmap() 391 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap() 401 name->name_len, run, vcn); in scan_nres_bitmap() 406 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap() 448 vcn_next = vcn + clen; in scan_nres_bitmap() 450 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && vcn == vcn_next; in scan_nres_bitmap() 452 vcn = vcn_next; in scan_nres_bitmap()
|
| H A D | super.c | 1258 CLST vcn, lcn, len; in ntfs_fill_super() local 1493 for (i = 0; run_get_entry(&ni->file.run, i, &vcn, &lcn, &len); i++) { in ntfs_fill_super()
|
| /linux/drivers/md/dm-vdo/indexer/ |
| H A D | volume.c | 1275 u64 vcn = BAD_CHAPTER; in probe_chapter() local 1296 if (vcn == BAD_CHAPTER) { in probe_chapter() 1297 vcn = page->virtual_chapter_number; in probe_chapter() 1298 } else if (page->virtual_chapter_number != vcn) { in probe_chapter() 1300 chapter_number, i, (unsigned long long) vcn, in probe_chapter() 1318 if (chapter_number != uds_map_to_physical_chapter(geometry, vcn)) { in probe_chapter() 1320 (unsigned long long) vcn, geometry->chapters_per_volume); in probe_chapter() 1324 *virtual_chapter_number = vcn; in probe_chapter() 1335 u64 vcn = 0; in find_real_end_of_volume() local 1337 probe_chapter(volume, chapter, &vcn); in find_real_end_of_volume() [all …]
|
| H A D | index.c | 825 static int rebuild_index_page_map(struct uds_index *index, u64 vcn) in rebuild_index_page_map() argument 830 u32 chapter = uds_map_to_physical_chapter(geometry, vcn); in rebuild_index_page_map() 856 uds_update_index_page_map(index->volume->index_page_map, vcn, chapter, in rebuild_index_page_map()
|
| H A D | delta-index.c | 498 u64 vcn; in uds_initialize_delta_index_page() local 510 vcn = get_unaligned_le64(vcn_addr); in uds_initialize_delta_index_page() 517 vcn = get_unaligned_be64(vcn_addr); in uds_initialize_delta_index_page() 536 delta_index_page->virtual_chapter_number = vcn; in uds_initialize_delta_index_page()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu15/ |
| H A D | smu_v15_0.c | 1258 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v15_0_set_performance_level() 1259 if (adev->vcn.harvest_config & (1 << i)) in smu_v15_0_set_performance_level() 1274 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v15_0_set_performance_level() 1275 if (adev->vcn.harvest_config & (1 << i)) in smu_v15_0_set_performance_level() 1458 if (adev->vcn.harvest_config & (1 << inst)) in smu_v15_0_set_vcn_enable()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu14/ |
| H A D | smu_v14_0.c | 1369 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v14_0_set_performance_level() 1370 if (adev->vcn.harvest_config & (1 << i)) in smu_v14_0_set_performance_level() 1385 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v14_0_set_performance_level() 1386 if (adev->vcn.harvest_config & (1 << i)) in smu_v14_0_set_performance_level() 1566 if (adev->vcn.harvest_config & (1 << inst)) in smu_v14_0_set_vcn_enable()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/ |
| H A D | amdgpu_smu.c | 840 for (i = 0; i < adev->vcn.num_vcn_inst; i++) in smu_set_default_dpm_table() 847 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_set_default_dpm_table() 869 for (i = 0; i < adev->vcn.num_vcn_inst; i++) in smu_set_default_dpm_table() 1365 for (i = 0; i < adev->vcn.num_vcn_inst; i++) in smu_sw_init() 1930 for (i = 0; i < adev->vcn.num_vcn_inst; i++) in smu_hw_init() 2139 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_hw_fini() 2141 adev->vcn.inst[i].cur_state = AMD_PG_STATE_GATE; in smu_hw_fini() 3177 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_read_sensor()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu13/ |
| H A D | smu_v13_0.c | 1720 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v13_0_set_performance_level() 1721 if (adev->vcn.harvest_config & (1 << i)) in smu_v13_0_set_performance_level() 1736 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v13_0_set_performance_level() 1737 if (adev->vcn.harvest_config & (1 << i)) in smu_v13_0_set_performance_level() 1995 if (adev->vcn.harvest_config & (1 << inst)) in smu_v13_0_set_vcn_enable()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/smu11/ |
| H A D | sienna_cichlid_ppt.c | 1033 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in sienna_cichlid_set_default_dpm_table() 1034 if (adev->vcn.harvest_config & (1 << i)) in sienna_cichlid_set_default_dpm_table() 1056 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in sienna_cichlid_set_default_dpm_table() 1057 if (adev->vcn.harvest_config & (1 << i)) in sienna_cichlid_set_default_dpm_table() 1160 if (adev->vcn.harvest_config & (1 << inst)) in sienna_cichlid_dpm_set_vcn_enable()
|