Home
last modified time | relevance | path

Searched refs:vcn (Results 1 – 25 of 35) sorted by relevance

12

/linux/drivers/gpu/drm/amd/amdgpu/
H A Dvcn_v2_5.c120 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) { in vcn_v2_5_idle_work_handler()
121 struct amdgpu_vcn_inst *v = &adev->vcn.inst[i]; in vcn_v2_5_idle_work_handler()
123 if (adev->vcn.harvest_config & (1 << i)) in vcn_v2_5_idle_work_handler()
148 if (!fences && !atomic_read(&adev->vcn.inst[0].total_submission_cnt)) { in vcn_v2_5_idle_work_handler()
153 schedule_delayed_work(&adev->vcn.inst[0].idle_work, VCN_IDLE_TIMEOUT); in vcn_v2_5_idle_work_handler()
160 struct amdgpu_vcn_inst *v = &adev->vcn.inst[ring->me]; in vcn_v2_5_ring_begin_use()
162 atomic_inc(&adev->vcn.inst[0].total_submission_cnt); in vcn_v2_5_ring_begin_use()
164 cancel_delayed_work_sync(&adev->vcn.inst[0].idle_work); in vcn_v2_5_ring_begin_use()
170 mutex_lock(&adev->vcn.inst[0].vcn_pg_lock); in vcn_v2_5_ring_begin_use()
196 mutex_unlock(&adev->vcn.inst[0].vcn_pg_lock); in vcn_v2_5_ring_begin_use()
[all …]
H A Dvcn_v5_0_0.c100 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) in vcn_v5_0_0_early_init()
102 adev->vcn.inst[i].num_enc_rings = 1; in vcn_v5_0_0_early_init()
107 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) { in vcn_v5_0_0_early_init()
108 adev->vcn.inst[i].set_pg_state = vcn_v5_0_0_set_pg_state; in vcn_v5_0_0_early_init()
131 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in vcn_v5_0_0_sw_init()
134 if (adev->vcn.harvest_config & (1 << i)) in vcn_v5_0_0_sw_init()
147 atomic_set(&adev->vcn.inst[i].sched_score, 0); in vcn_v5_0_0_sw_init()
151 VCN_5_0__SRCID__UVD_ENC_GENERAL_PURPOSE, &adev->vcn.inst[i].irq); in vcn_v5_0_0_sw_init()
157 VCN_5_0__SRCID_UVD_POISON, &adev->vcn.inst[i].irq); in vcn_v5_0_0_sw_init()
161 ring = &adev->vcn.inst[i].ring_enc[0]; in vcn_v5_0_0_sw_init()
[all …]
H A Dvcn_v4_0_5.c119 adev->vcn.per_inst_fw = true; in vcn_v4_0_5_early_init()
121 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) in vcn_v4_0_5_early_init()
123 adev->vcn.inst[i].num_enc_rings = 1; in vcn_v4_0_5_early_init()
127 for (i = 0; i < adev->vcn.num_vcn_inst; ++i) { in vcn_v4_0_5_early_init()
128 adev->vcn.inst[i].set_pg_state = vcn_v4_0_5_set_pg_state; in vcn_v4_0_5_early_init()
151 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in vcn_v4_0_5_sw_init()
154 if (adev->vcn.harvest_config & (1 << i)) in vcn_v4_0_5_sw_init()
167 atomic_set(&adev->vcn.inst[i].sched_score, 0); in vcn_v4_0_5_sw_init()
171 VCN_4_0__SRCID__UVD_ENC_GENERAL_PURPOSE, &adev->vcn.inst[i].irq); in vcn_v4_0_5_sw_init()
177 VCN_4_0__SRCID_UVD_POISON, &adev->vcn.inst[i].irq); in vcn_v4_0_5_sw_init()
[all …]
H A Dvega10_reg_init.c83 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_DOORBELL64_VCN0_1; in vega10_doorbell_index_init()
84 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_DOORBELL64_VCN2_3; in vega10_doorbell_index_init()
85 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_DOORBELL64_VCN4_5; in vega10_doorbell_index_init()
86 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_DOORBELL64_VCN6_7; in vega10_doorbell_index_init()
H A Dvega20_reg_init.c89 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_VEGA20_DOORBELL64_VCN0_1; in vega20_doorbell_index_init()
90 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_VEGA20_DOORBELL64_VCN2_3; in vega20_doorbell_index_init()
91 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_VEGA20_DOORBELL64_VCN4_5; in vega20_doorbell_index_init()
92 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_VEGA20_DOORBELL64_VCN6_7; in vega20_doorbell_index_init()
H A Dsoc21.c172 if (adev->vcn.num_vcn_inst == hweight8(adev->vcn.harvest_config)) in soc21_query_video_codecs()
181 if ((adev->vcn.harvest_config & AMDGPU_VCN_HARVEST_VCN0) || in soc21_query_video_codecs()
194 if ((adev->vcn.harvest_config & AMDGPU_VCN_HARVEST_VCN0)) { in soc21_query_video_codecs()
542 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_NAVI10_DOORBELL64_VCN0_1; in soc21_init_doorbell_index()
543 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_NAVI10_DOORBELL64_VCN2_3; in soc21_init_doorbell_index()
544 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_NAVI10_DOORBELL64_VCN4_5; in soc21_init_doorbell_index()
545 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_NAVI10_DOORBELL64_VCN6_7; in soc21_init_doorbell_index()
887 if ((adev->vcn.harvest_config & AMDGPU_VCN_HARVEST_VCN0) || in soc21_common_late_init()
H A Damdgpu_discovery.c424 adev->vcn.harvest_config |= AMDGPU_VCN_HARVEST_VCN1; in amdgpu_discovery_harvest_config_quirk()
425 adev->vcn.inst_mask &= ~AMDGPU_VCN_HARVEST_VCN1; in amdgpu_discovery_harvest_config_quirk()
741 adev->vcn.harvest_config |= AMDGPU_VCN_HARVEST_VCN0; in amdgpu_discovery_read_harvest_bit_per_ip()
742 adev->vcn.inst_mask &= in amdgpu_discovery_read_harvest_bit_per_ip()
747 adev->vcn.harvest_config |= AMDGPU_VCN_HARVEST_VCN1; in amdgpu_discovery_read_harvest_bit_per_ip()
748 adev->vcn.inst_mask &= in amdgpu_discovery_read_harvest_bit_per_ip()
796 adev->vcn.harvest_config |= in amdgpu_discovery_read_from_harvest_table()
801 adev->vcn.inst_mask &= in amdgpu_discovery_read_from_harvest_table()
1094 harvest = ((1 << inst) & adev->vcn.inst_mask) == 0; in amdgpu_discovery_get_harvest_info()
1410 adev->vcn.inst_mask = 0; in amdgpu_discovery_reg_base_init()
[all …]
H A Damdgpu_kms.c229 fw_info->ver = adev->vcn.fw_version; in amdgpu_firmware_info()
531 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in amdgpu_hw_ip_info()
532 if (adev->vcn.harvest_config & (1 << i)) in amdgpu_hw_ip_info()
535 if (adev->vcn.inst[i].ring_dec.sched.ready && in amdgpu_hw_ip_info()
536 !adev->vcn.inst[i].ring_dec.no_user_submission) in amdgpu_hw_ip_info()
544 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in amdgpu_hw_ip_info()
545 if (adev->vcn.harvest_config & (1 << i)) in amdgpu_hw_ip_info()
548 for (j = 0; j < adev->vcn.inst[i].num_enc_rings; j++) in amdgpu_hw_ip_info()
549 if (adev->vcn.inst[i].ring_enc[j].sched.ready && in amdgpu_hw_ip_info()
550 !adev->vcn.inst[i].ring_enc[j].no_user_submission) in amdgpu_hw_ip_info()
[all …]
H A Damdgpu_dev_coredump.c167 adev->vcn.fw_version); in amdgpu_devcoredump_fw_info()
H A Dumsch_mm_v4_0.c226 (adev->doorbell_index.vcn.vcn_ring0_1 << 1), 0); in umsch_mm_v4_0_ring_start()
H A Damdgpu_debugfs.c2062 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in amdgpu_debugfs_init()
2066 if (adev->vcn.harvest_config & (1 << i)) in amdgpu_debugfs_init()
2069 amdgpu_debugfs_vcn_fwlog_init(adev, i, &adev->vcn.inst[i]); in amdgpu_debugfs_init()
/linux/fs/ntfs3/
H A Dntfs_fs.h449 struct runs_tree *run_da, CLST vcn, CLST lcn,
469 int attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn,
472 int attr_data_get_block_locked(struct ntfs_inode *ni, CLST vcn, CLST clen,
478 CLST vcn);
507 u8 name_len, const CLST *vcn);
562 u8 name_len, const CLST *vcn,
858 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn,
860 void run_truncate(struct runs_tree *run, CLST vcn);
861 void run_truncate_head(struct runs_tree *run, CLST vcn);
862 void run_truncate_around(struct runs_tree *run, CLST vcn);
[all …]
H A Dfile.c318 CLST vcn = from >> sbi->cluster_bits; in ntfs_file_mmap_prepare() local
322 for (; vcn < end; vcn += len) { in ntfs_file_mmap_prepare()
323 err = attr_data_get_block(ni, vcn, 1, &lcn, in ntfs_file_mmap_prepare()
644 CLST vcn = vbo >> cluster_bits; in ntfs_fallocate() local
657 for (; vcn < cend_v; vcn += clen) { in ntfs_fallocate()
658 err = attr_data_get_block(ni, vcn, cend_v - vcn, in ntfs_fallocate()
676 for (; vcn < cend; vcn += clen) { in ntfs_fallocate()
677 err = attr_data_get_block(ni, vcn, cend - vcn, in ntfs_fallocate()
H A Dfrecord.c194 const __le16 *name, u8 name_len, const CLST *vcn, in ni_find_attr() argument
213 le = al_find_ex(ni, le_o ? *le_o : NULL, type, name, name_len, vcn); in ni_find_attr()
231 if (vcn && *vcn) in ni_find_attr()
233 } else if (!vcn) { in ni_find_attr()
236 } else if (le64_to_cpu(attr->nres.svcn) > *vcn || in ni_find_attr()
237 *vcn > le64_to_cpu(attr->nres.evcn)) { in ni_find_attr()
496 if (svcn != le64_to_cpu(le->vcn)) { in ni_repack()
569 attr->nres.svcn = le->vcn = cpu_to_le64(next_svcn); in ni_repack()
643 if (le->vcn) in ni_try_remove_attr_list()
792 le->vcn = 0; in ni_create_attr_list()
[all …]
H A Dfslog.c154 __le64 vcn; // 0x10: Vcn of dirty page member
654 u32 target_attr, u64 vcn) in find_dp() argument
660 u64 dp_vcn = le64_to_cpu(dp->vcn); in find_dp()
662 if (dp->target_attr == ta && vcn >= dp_vcn && in find_dp()
663 vcn < dp_vcn + le32_to_cpu(dp->lcns_follow)) { in find_dp()
3756 u64 size, vcn, undo_next_lsn; in log_replay() local
4216 memmove(&dp->vcn, &dp0->vcn_low, in log_replay()
4237 next->vcn == dp->vcn) { in log_replay()
4530 dp->vcn = cpu_to_le64(t64 & ~((u64)t32 - 1)); in log_replay()
4542 le64_to_cpu(dp->vcn)); in log_replay()
[all …]
H A Dindex.c210 CLST vcn = off >> sbi->cluster_bits; in bmp_buf_get() local
220 &vcn, &bbuf->mi); in bmp_buf_get()
377 CLST lcn, clen, vcn, vcn_next; in scan_nres_bitmap() local
388 vcn = vbo >> sbi->cluster_bits; in scan_nres_bitmap()
391 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap()
401 name->name_len, run, vcn); in scan_nres_bitmap()
406 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap()
448 vcn_next = vcn + clen; in scan_nres_bitmap()
450 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && vcn == vcn_next; in scan_nres_bitmap()
452 vcn = vcn_next; in scan_nres_bitmap()
H A Dsuper.c1258 CLST vcn, lcn, len; in ntfs_fill_super() local
1493 for (i = 0; run_get_entry(&ni->file.run, i, &vcn, &lcn, &len); i++) { in ntfs_fill_super()
/linux/drivers/md/dm-vdo/indexer/
H A Dvolume.c1275 u64 vcn = BAD_CHAPTER; in probe_chapter() local
1296 if (vcn == BAD_CHAPTER) { in probe_chapter()
1297 vcn = page->virtual_chapter_number; in probe_chapter()
1298 } else if (page->virtual_chapter_number != vcn) { in probe_chapter()
1300 chapter_number, i, (unsigned long long) vcn, in probe_chapter()
1318 if (chapter_number != uds_map_to_physical_chapter(geometry, vcn)) { in probe_chapter()
1320 (unsigned long long) vcn, geometry->chapters_per_volume); in probe_chapter()
1324 *virtual_chapter_number = vcn; in probe_chapter()
1335 u64 vcn = 0; in find_real_end_of_volume() local
1337 probe_chapter(volume, chapter, &vcn); in find_real_end_of_volume()
[all …]
H A Dindex.c825 static int rebuild_index_page_map(struct uds_index *index, u64 vcn) in rebuild_index_page_map() argument
830 u32 chapter = uds_map_to_physical_chapter(geometry, vcn); in rebuild_index_page_map()
856 uds_update_index_page_map(index->volume->index_page_map, vcn, chapter, in rebuild_index_page_map()
H A Ddelta-index.c498 u64 vcn; in uds_initialize_delta_index_page() local
510 vcn = get_unaligned_le64(vcn_addr); in uds_initialize_delta_index_page()
517 vcn = get_unaligned_be64(vcn_addr); in uds_initialize_delta_index_page()
536 delta_index_page->virtual_chapter_number = vcn; in uds_initialize_delta_index_page()
/linux/drivers/gpu/drm/amd/pm/swsmu/smu15/
H A Dsmu_v15_0.c1258 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v15_0_set_performance_level()
1259 if (adev->vcn.harvest_config & (1 << i)) in smu_v15_0_set_performance_level()
1274 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v15_0_set_performance_level()
1275 if (adev->vcn.harvest_config & (1 << i)) in smu_v15_0_set_performance_level()
1458 if (adev->vcn.harvest_config & (1 << inst)) in smu_v15_0_set_vcn_enable()
/linux/drivers/gpu/drm/amd/pm/swsmu/smu14/
H A Dsmu_v14_0.c1369 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v14_0_set_performance_level()
1370 if (adev->vcn.harvest_config & (1 << i)) in smu_v14_0_set_performance_level()
1385 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v14_0_set_performance_level()
1386 if (adev->vcn.harvest_config & (1 << i)) in smu_v14_0_set_performance_level()
1566 if (adev->vcn.harvest_config & (1 << inst)) in smu_v14_0_set_vcn_enable()
/linux/drivers/gpu/drm/amd/pm/swsmu/
H A Damdgpu_smu.c840 for (i = 0; i < adev->vcn.num_vcn_inst; i++) in smu_set_default_dpm_table()
847 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_set_default_dpm_table()
869 for (i = 0; i < adev->vcn.num_vcn_inst; i++) in smu_set_default_dpm_table()
1365 for (i = 0; i < adev->vcn.num_vcn_inst; i++) in smu_sw_init()
1930 for (i = 0; i < adev->vcn.num_vcn_inst; i++) in smu_hw_init()
2139 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_hw_fini()
2141 adev->vcn.inst[i].cur_state = AMD_PG_STATE_GATE; in smu_hw_fini()
3177 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_read_sensor()
/linux/drivers/gpu/drm/amd/pm/swsmu/smu13/
H A Dsmu_v13_0.c1720 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v13_0_set_performance_level()
1721 if (adev->vcn.harvest_config & (1 << i)) in smu_v13_0_set_performance_level()
1736 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in smu_v13_0_set_performance_level()
1737 if (adev->vcn.harvest_config & (1 << i)) in smu_v13_0_set_performance_level()
1995 if (adev->vcn.harvest_config & (1 << inst)) in smu_v13_0_set_vcn_enable()
/linux/drivers/gpu/drm/amd/pm/swsmu/smu11/
H A Dsienna_cichlid_ppt.c1033 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in sienna_cichlid_set_default_dpm_table()
1034 if (adev->vcn.harvest_config & (1 << i)) in sienna_cichlid_set_default_dpm_table()
1056 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in sienna_cichlid_set_default_dpm_table()
1057 if (adev->vcn.harvest_config & (1 << i)) in sienna_cichlid_set_default_dpm_table()
1160 if (adev->vcn.harvest_config & (1 << inst)) in sienna_cichlid_dpm_set_vcn_enable()

12