Home
last modified time | relevance | path

Searched full:gt (Results 1 – 25 of 692) sorted by relevance

12345678910>>...28

/linux/drivers/gpu/drm/xe/
H A Dxe_gt.c65 struct xe_gt *gt = arg; in gt_fini() local
67 destroy_workqueue(gt->ordered_wq); in gt_fini()
72 struct xe_gt *gt; in xe_gt_alloc() local
75 gt = drmm_kzalloc(&tile_to_xe(tile)->drm, sizeof(*gt), GFP_KERNEL); in xe_gt_alloc()
76 if (!gt) in xe_gt_alloc()
79 gt->tile = tile; in xe_gt_alloc()
80 gt->ordered_wq = alloc_ordered_workqueue("gt-ordered-wq", 0); in xe_gt_alloc()
82 err = drmm_add_action_or_reset(&gt_to_xe(gt)->drm, gt_fini, gt); in xe_gt_alloc()
86 return gt; in xe_gt_alloc()
89 void xe_gt_sanitize(struct xe_gt *gt) in xe_gt_sanitize() argument
[all …]
H A Dxe_gt_sriov_pf_policy.c37 static int pf_send_policy_klvs(struct xe_gt *gt, const u32 *klvs, u32 num_dwords) in pf_send_policy_klvs() argument
40 struct xe_tile *tile = gt_to_tile(gt); in pf_send_policy_klvs()
42 struct xe_guc *guc = &gt->uc.guc; in pf_send_policy_klvs()
67 static int pf_push_policy_klvs(struct xe_gt *gt, u32 num_klvs, in pf_push_policy_klvs() argument
72 xe_gt_assert(gt, num_klvs == xe_guc_klv_count(klvs, num_dwords)); in pf_push_policy_klvs()
74 ret = pf_send_policy_klvs(gt, klvs, num_dwords); in pf_push_policy_klvs()
78 struct drm_printer p = xe_gt_info_printer(gt); in pf_push_policy_klvs()
80 xe_gt_sriov_notice(gt, "Failed to push %u policy KLV%s (%pe)\n", in pf_push_policy_klvs()
89 static int pf_push_policy_u32(struct xe_gt *gt, u16 key, u32 value) in pf_push_policy_u32() argument
96 return pf_push_policy_klvs(gt, 1, klv, ARRAY_SIZE(klv)); in pf_push_policy_u32()
[all …]
H A Dxe_gt_sriov_pf_config.c58 static int pf_send_vf_cfg_reset(struct xe_gt *gt, u32 vfid) in pf_send_vf_cfg_reset() argument
60 struct xe_guc *guc = &gt->uc.guc; in pf_send_vf_cfg_reset()
72 static int pf_send_vf_cfg_klvs(struct xe_gt *gt, u32 vfid, const u32 *klvs, u32 num_dwords) in pf_send_vf_cfg_klvs() argument
75 struct xe_tile *tile = gt_to_tile(gt); in pf_send_vf_cfg_klvs()
77 struct xe_guc *guc = &gt->uc.guc; in pf_send_vf_cfg_klvs()
103 static int pf_push_vf_cfg_klvs(struct xe_gt *gt, unsigned int vfid, u32 num_klvs, in pf_push_vf_cfg_klvs() argument
108 xe_gt_assert(gt, num_klvs == xe_guc_klv_count(klvs, num_dwords)); in pf_push_vf_cfg_klvs()
110 ret = pf_send_vf_cfg_klvs(gt, vfid, klvs, num_dwords); in pf_push_vf_cfg_klvs()
114 struct drm_printer p = xe_gt_info_printer(gt); in pf_push_vf_cfg_klvs()
117 xe_gt_sriov_notice(gt, "Failed to push %s %u config KLV%s (%pe)\n", in pf_push_vf_cfg_klvs()
[all …]
H A Dxe_gt_sriov_vf.c49 static int vf_reset_guc_state(struct xe_gt *gt) in vf_reset_guc_state() argument
51 struct xe_guc *guc = &gt->uc.guc; in vf_reset_guc_state()
56 xe_gt_sriov_err(gt, "Failed to reset GuC state (%pe)\n", ERR_PTR(err)); in vf_reset_guc_state()
93 static void vf_minimum_guc_version(struct xe_gt *gt, u32 *branch, u32 *major, u32 *minor) in vf_minimum_guc_version() argument
95 struct xe_device *xe = gt_to_xe(gt); in vf_minimum_guc_version()
113 static void vf_wanted_guc_version(struct xe_gt *gt, u32 *branch, u32 *major, u32 *minor) in vf_wanted_guc_version() argument
116 return vf_minimum_guc_version(gt, branch, major, minor); in vf_wanted_guc_version()
119 static int vf_handshake_with_guc(struct xe_gt *gt) in vf_handshake_with_guc() argument
121 struct xe_gt_sriov_vf_guc_version *guc_version = &gt->sriov.vf.guc_version; in vf_handshake_with_guc()
122 struct xe_guc *guc = &gt->uc.guc; in vf_handshake_with_guc()
[all …]
H A Dxe_gt_tlb_invalidation.c29 static long tlb_timeout_jiffies(struct xe_gt *gt) in tlb_timeout_jiffies() argument
35 long delay = xe_guc_ct_queue_proc_time_jiffies(&gt->uc.guc.ct); in tlb_timeout_jiffies()
61 struct xe_gt *gt = container_of(work, struct xe_gt, in xe_gt_tlb_fence_timeout() local
63 struct xe_device *xe = gt_to_xe(gt); in xe_gt_tlb_fence_timeout()
66 spin_lock_irq(&gt->tlb_invalidation.pending_lock); in xe_gt_tlb_fence_timeout()
68 &gt->tlb_invalidation.pending_fences, link) { in xe_gt_tlb_fence_timeout()
72 if (msecs_to_jiffies(since_inval_ms) < tlb_timeout_jiffies(gt)) in xe_gt_tlb_fence_timeout()
76 xe_gt_err(gt, "TLB invalidation fence timeout, seqno=%d recv=%d", in xe_gt_tlb_fence_timeout()
77 fence->seqno, gt->tlb_invalidation.seqno_recv); in xe_gt_tlb_fence_timeout()
82 if (!list_empty(&gt->tlb_invalidation.pending_fences)) in xe_gt_tlb_fence_timeout()
[all …]
H A Dxe_gt_throttle.c17 * DOC: Xe GT Throttle
19 * Provides sysfs entries and other helpers for frequency throttle reasons in GT
21 * device/gt#/freq0/throttle/status - Overall status
22 * device/gt#/freq0/throttle/reason_pl1 - Frequency throttle due to PL1
23 * device/gt#/freq0/throttle/reason_pl2 - Frequency throttle due to PL2
24 * device/gt#/freq0/throttle/reason_pl4 - Frequency throttle due to PL4, Iccmax etc.
25 * device/gt#/freq0/throttle/reason_thermal - Frequency throttle due to thermal
26 * device/gt#/freq0/throttle/reason_prochot - Frequency throttle due to prochot
27 * device/gt#/freq0/throttle/reason_ratl - Frequency throttle due to RATL
28 * device/gt#/freq0/throttle/reason_vr_thermalert - Frequency throttle due to VR THERMALERT
[all …]
H A Dxe_gt_sriov_pf_service.c23 static void pf_init_versions(struct xe_gt *gt) in pf_init_versions() argument
29 gt->sriov.pf.service.version.base.major = GUC_RELAY_VERSION_BASE_MAJOR; in pf_init_versions()
30 gt->sriov.pf.service.version.base.minor = GUC_RELAY_VERSION_BASE_MINOR; in pf_init_versions()
33 gt->sriov.pf.service.version.latest.major = GUC_RELAY_VERSION_LATEST_MAJOR; in pf_init_versions()
34 gt->sriov.pf.service.version.latest.minor = GUC_RELAY_VERSION_LATEST_MINOR; in pf_init_versions()
38 static int pf_negotiate_version(struct xe_gt *gt, in pf_negotiate_version() argument
42 struct xe_gt_sriov_pf_service_version base = gt->sriov.pf.service.version.base; in pf_negotiate_version()
43 struct xe_gt_sriov_pf_service_version latest = gt->sriov.pf.service.version.latest; in pf_negotiate_version()
45 xe_gt_assert(gt, base.major); in pf_negotiate_version()
46 xe_gt_assert(gt, base.major <= latest.major); in pf_negotiate_version()
[all …]
H A Dxe_gt_idle.c21 * DOC: Xe GT Idle
23 * Contains functions that init GT idle features like C6
25 * device/gt#/gtidle/name - name of the state
26 * device/gt#/gtidle/idle_residency_ms - Provides residency of the idle state in ms
27 * device/gt#/gtidle/idle_status - Provides current idle state
51 struct xe_gt *gt = container_of(guc, struct xe_gt, uc.guc); in pc_to_xe() local
53 return gt_to_xe(gt); in pc_to_xe()
60 return "gt-c0"; in gt_idle_state_to_string()
62 return "gt-c6"; in gt_idle_state_to_string()
98 void xe_gt_idle_enable_pg(struct xe_gt *gt) in xe_gt_idle_enable_pg() argument
[all …]
H A Dxe_gt_ccs_mode.c17 static void __xe_gt_apply_ccs_mode(struct xe_gt *gt, u32 num_engines) in __xe_gt_apply_ccs_mode() argument
20 int num_slices = hweight32(CCS_MASK(gt)); in __xe_gt_apply_ccs_mode()
21 struct xe_device *xe = gt_to_xe(gt); in __xe_gt_apply_ccs_mode()
25 xe_assert(xe, xe_gt_ccs_mode_enabled(gt)); in __xe_gt_apply_ccs_mode()
52 for_each_hw_engine(hwe, gt, id) { in __xe_gt_apply_ccs_mode()
62 while ((CCS_MASK(gt) & BIT(cslice)) == 0) in __xe_gt_apply_ccs_mode()
71 xe_mmio_write32(gt, CCS_MODE, mode); in __xe_gt_apply_ccs_mode()
73 xe_gt_dbg(gt, "CCS_MODE=%x config:%08x, num_engines:%d, num_slices:%d\n", in __xe_gt_apply_ccs_mode()
77 void xe_gt_apply_ccs_mode(struct xe_gt *gt) in xe_gt_apply_ccs_mode() argument
79 if (!gt->ccs_mode || IS_SRIOV_VF(gt_to_xe(gt))) in xe_gt_apply_ccs_mode()
[all …]
H A Dxe_force_wake.c35 void xe_force_wake_init_gt(struct xe_gt *gt, struct xe_force_wake *fw) in xe_force_wake_init_gt() argument
37 struct xe_device *xe = gt_to_xe(gt); in xe_force_wake_init_gt()
39 fw->gt = gt; in xe_force_wake_init_gt()
43 xe_gt_assert(gt, GRAPHICS_VER(gt_to_xe(gt)) >= 11); in xe_force_wake_init_gt()
58 void xe_force_wake_init_engines(struct xe_gt *gt, struct xe_force_wake *fw) in xe_force_wake_init_engines() argument
63 xe_gt_assert(gt, GRAPHICS_VER(gt_to_xe(gt)) >= 11); in xe_force_wake_init_engines()
65 if (!xe_gt_is_media_type(gt)) in xe_force_wake_init_engines()
72 if (!(gt->info.engine_mask & BIT(i))) in xe_force_wake_init_engines()
82 if (!(gt->info.engine_mask & BIT(i))) in xe_force_wake_init_engines()
91 if (gt->info.engine_mask & BIT(XE_HW_ENGINE_GSCCS0)) in xe_force_wake_init_engines()
[all …]
H A Dxe_gt_sriov_pf_monitor.c18 * @gt: the &xe_gt
24 void xe_gt_sriov_pf_monitor_flr(struct xe_gt *gt, u32 vfid) in xe_gt_sriov_pf_monitor_flr() argument
28 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in xe_gt_sriov_pf_monitor_flr()
29 xe_gt_sriov_pf_assert_vfid(gt, vfid); in xe_gt_sriov_pf_monitor_flr()
32 gt->sriov.pf.vfs[vfid].monitor.guc.events[e] = 0; in xe_gt_sriov_pf_monitor_flr()
35 static void pf_update_event_counter(struct xe_gt *gt, u32 vfid, in pf_update_event_counter() argument
38 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in pf_update_event_counter()
39 xe_gt_assert(gt, e < XE_GUC_KLV_NUM_THRESHOLDS); in pf_update_event_counter()
41 gt->sriov.pf.vfs[vfid].monitor.guc.events[e]++; in pf_update_event_counter()
44 static int pf_handle_vf_threshold_event(struct xe_gt *gt, u32 vfid, u32 threshold) in pf_handle_vf_threshold_event() argument
[all …]
H A Dxe_gt_mcr.h15 void xe_gt_mcr_init_early(struct xe_gt *gt);
16 void xe_gt_mcr_init(struct xe_gt *gt);
18 void xe_gt_mcr_set_implicit_defaults(struct xe_gt *gt);
20 u32 xe_gt_mcr_unicast_read(struct xe_gt *gt, struct xe_reg_mcr mcr_reg,
22 u32 xe_gt_mcr_unicast_read_any(struct xe_gt *gt, struct xe_reg_mcr mcr_reg);
24 void xe_gt_mcr_unicast_write(struct xe_gt *gt, struct xe_reg_mcr mcr_reg,
26 void xe_gt_mcr_multicast_write(struct xe_gt *gt, struct xe_reg_mcr mcr_reg,
29 void xe_gt_mcr_steering_dump(struct xe_gt *gt, struct drm_printer *p);
30 void xe_gt_mcr_get_dss_steering(struct xe_gt *gt, unsigned int dss, u16 *group, u16 *instance);
36 * @gt: GT structure
[all …]
/linux/drivers/gpu/drm/i915/gt/
H A Dintel_gt_pm.c28 static void user_forcewake(struct intel_gt *gt, bool suspend) in user_forcewake() argument
30 int count = atomic_read(&gt->user_wakeref); in user_forcewake()
37 wakeref = intel_gt_pm_get(gt); in user_forcewake()
39 GEM_BUG_ON(count > atomic_read(&gt->wakeref.count)); in user_forcewake()
40 atomic_sub(count, &gt->wakeref.count); in user_forcewake()
42 atomic_add(count, &gt->wakeref.count); in user_forcewake()
44 intel_gt_pm_put(gt, wakeref); in user_forcewake()
47 static void runtime_begin(struct intel_gt *gt) in runtime_begin() argument
50 write_seqcount_begin(&gt->stats.lock); in runtime_begin()
51 gt->stats.start = ktime_get(); in runtime_begin()
[all …]
H A Dintel_gt.c40 void intel_gt_common_init_early(struct intel_gt *gt) in intel_gt_common_init_early() argument
42 spin_lock_init(gt->irq_lock); in intel_gt_common_init_early()
44 INIT_LIST_HEAD(&gt->closed_vma); in intel_gt_common_init_early()
45 spin_lock_init(&gt->closed_lock); in intel_gt_common_init_early()
47 init_llist_head(&gt->watchdog.list); in intel_gt_common_init_early()
48 INIT_WORK(&gt->watchdog.work, intel_gt_watchdog_work); in intel_gt_common_init_early()
50 intel_gt_init_buffer_pool(gt); in intel_gt_common_init_early()
51 intel_gt_init_reset(gt); in intel_gt_common_init_early()
52 intel_gt_init_requests(gt); in intel_gt_common_init_early()
53 intel_gt_init_timelines(gt); in intel_gt_common_init_early()
[all …]
H A Dintel_gt_irq.c31 gen11_gt_engine_identity(struct intel_gt *gt, in gen11_gt_engine_identity() argument
34 void __iomem * const regs = intel_uncore_regs(gt->uncore); in gen11_gt_engine_identity()
38 lockdep_assert_held(gt->irq_lock); in gen11_gt_engine_identity()
53 gt_err(gt, "INTR_IDENTITY_REG%u:%u 0x%08x not valid!\n", in gen11_gt_engine_identity()
65 gen11_other_irq_handler(struct intel_gt *gt, const u8 instance, in gen11_other_irq_handler() argument
68 struct intel_gt *media_gt = gt->i915->media_gt; in gen11_other_irq_handler()
71 return guc_irq_handler(gt_to_guc(gt), iir); in gen11_other_irq_handler()
76 return gen11_rps_irq_handler(&gt->rps, iir); in gen11_other_irq_handler()
81 return intel_pxp_irq_handler(gt->i915->pxp, iir); in gen11_other_irq_handler()
84 return intel_gsc_irq_handler(gt, iir); in gen11_other_irq_handler()
[all …]
H A Dintel_gt_pm.h14 static inline bool intel_gt_pm_is_awake(const struct intel_gt *gt) in intel_gt_pm_is_awake() argument
16 return intel_wakeref_is_active(&gt->wakeref); in intel_gt_pm_is_awake()
19 static inline void intel_gt_pm_get_untracked(struct intel_gt *gt) in intel_gt_pm_get_untracked() argument
21 intel_wakeref_get(&gt->wakeref); in intel_gt_pm_get_untracked()
24 static inline intel_wakeref_t intel_gt_pm_get(struct intel_gt *gt) in intel_gt_pm_get() argument
26 intel_gt_pm_get_untracked(gt); in intel_gt_pm_get()
27 return intel_wakeref_track(&gt->wakeref); in intel_gt_pm_get()
30 static inline void __intel_gt_pm_get(struct intel_gt *gt) in __intel_gt_pm_get() argument
32 __intel_wakeref_get(&gt->wakeref); in __intel_gt_pm_get()
35 static inline intel_wakeref_t intel_gt_pm_get_if_awake(struct intel_gt *gt) in intel_gt_pm_get_if_awake() argument
[all …]
H A Dintel_gt_mcr.c13 * DOC: GT Multicast/Replicated (MCR) Register Support
15 * Some GT registers are designed as "multicast" or "replicated" registers:
108 void intel_gt_mcr_init(struct intel_gt *gt) in intel_gt_mcr_init() argument
110 struct drm_i915_private *i915 = gt->i915; in intel_gt_mcr_init()
114 spin_lock_init(&gt->mcr_lock); in intel_gt_mcr_init()
121 gt->info.mslice_mask = in intel_gt_mcr_init()
122 intel_slicemask_from_xehp_dssmask(gt->info.sseu.subslice_mask, in intel_gt_mcr_init()
124 gt->info.mslice_mask |= in intel_gt_mcr_init()
125 (intel_uncore_read(gt->uncore, GEN10_MIRROR_FUSE3) & in intel_gt_mcr_init()
128 if (!gt->info.mslice_mask) /* should be impossible! */ in intel_gt_mcr_init()
[all …]
H A Dselftest_reset.c18 __igt_reset_stolen(struct intel_gt *gt, in __igt_reset_stolen() argument
22 struct i915_ggtt *ggtt = gt->ggtt; in __igt_reset_stolen()
23 const struct resource *dsm = &gt->i915->dsm.stolen; in __igt_reset_stolen()
51 igt_global_reset_lock(gt); in __igt_reset_stolen()
52 wakeref = intel_runtime_pm_get(gt->uncore->rpm); in __igt_reset_stolen()
54 err = igt_spinner_init(&spin, gt); in __igt_reset_stolen()
58 for_each_engine(engine, gt, id) { in __igt_reset_stolen()
89 i915_gem_get_pat_index(gt->i915, in __igt_reset_stolen()
98 if (!__drm_mm_interval_first(&gt->i915->mm.stolen, in __igt_reset_stolen()
114 intel_gt_reset(gt, mask, NULL); in __igt_reset_stolen()
[all …]
H A Dintel_gt_pm_irq.c13 static void write_pm_imr(struct intel_gt *gt) in write_pm_imr() argument
15 struct drm_i915_private *i915 = gt->i915; in write_pm_imr()
16 struct intel_uncore *uncore = gt->uncore; in write_pm_imr()
17 u32 mask = gt->pm_imr; in write_pm_imr()
32 static void gen6_gt_pm_update_irq(struct intel_gt *gt, in gen6_gt_pm_update_irq() argument
40 lockdep_assert_held(gt->irq_lock); in gen6_gt_pm_update_irq()
42 new_val = gt->pm_imr; in gen6_gt_pm_update_irq()
46 if (new_val != gt->pm_imr) { in gen6_gt_pm_update_irq()
47 gt->pm_imr = new_val; in gen6_gt_pm_update_irq()
48 write_pm_imr(gt); in gen6_gt_pm_update_irq()
[all …]
H A Dintel_gt_sysfs_pm.c29 int (func)(struct intel_gt *gt, u32 val), u32 val) in sysfs_gt_attribute_w_func() argument
31 struct intel_gt *gt; in sysfs_gt_attribute_w_func() local
39 for_each_gt(gt, i915, i) { in sysfs_gt_attribute_w_func()
40 ret = func(gt, val); in sysfs_gt_attribute_w_func()
45 gt = intel_gt_sysfs_get_drvdata(kobj, attr->name); in sysfs_gt_attribute_w_func()
46 ret = func(gt, val); in sysfs_gt_attribute_w_func()
54 u32 (func)(struct intel_gt *gt), in sysfs_gt_attribute_r_func() argument
57 struct intel_gt *gt; in sysfs_gt_attribute_r_func() local
67 for_each_gt(gt, i915, i) { in sysfs_gt_attribute_r_func()
68 u32 val = func(gt); in sysfs_gt_attribute_r_func()
[all …]
H A Dintel_tlb.c33 return intel_gt_mcr_wait_for_reg(engine->gt, in wait_for_invalidate()
40 return __intel_wait_for_register_fw(engine->gt->uncore, in wait_for_invalidate()
49 static void mmio_invalidate_full(struct intel_gt *gt) in mmio_invalidate_full() argument
51 struct drm_i915_private *i915 = gt->i915; in mmio_invalidate_full()
52 struct intel_uncore *uncore = gt->uncore; in mmio_invalidate_full()
63 intel_gt_mcr_lock(gt, &flags); in mmio_invalidate_full()
64 spin_lock(&uncore->lock); /* serialise invalidate with GT reset */ in mmio_invalidate_full()
67 for_each_engine(engine, gt, id) { in mmio_invalidate_full()
72 intel_gt_mcr_multicast_write_fw(gt, in mmio_invalidate_full()
83 GT_TRACE(gt, "invalidated engines %08x\n", awake); in mmio_invalidate_full()
[all …]
H A Dintel_gt_sysfs.c23 return !strncmp(kobj->name, "gt", 2); in is_object_gt()
31 * has been called, whether it's called from gt/ or from in intel_gt_sysfs_get_drvdata()
35 * If the interface is called from gt/ then private data is in intel_gt_sysfs_get_drvdata()
49 static struct kobject *gt_get_parent_obj(struct intel_gt *gt) in gt_get_parent_obj() argument
51 return &gt->i915->drm.primary->kdev->kobj; in gt_get_parent_obj()
58 struct intel_gt *gt = intel_gt_sysfs_get_drvdata(kobj, attr->attr.name); in id_show() local
60 return sysfs_emit(buf, "%u\n", gt->info.id); in id_show()
81 void intel_gt_sysfs_register(struct intel_gt *gt) in intel_gt_sysfs_register() argument
88 * We generate the files only for gt 0 in intel_gt_sysfs_register()
91 if (gt_is_root(gt)) in intel_gt_sysfs_register()
[all …]
/linux/drivers/gpu/drm/i915/
H A DMakefile74 gt-y += \
75 gt/gen2_engine_cs.o \
76 gt/gen6_engine_cs.o \
77 gt/gen6_ppgtt.o \
78 gt/gen7_renderclear.o \
79 gt/gen8_engine_cs.o \
80 gt/gen8_ppgtt.o \
81 gt/intel_breadcrumbs.o \
82 gt/intel_context.o \
83 gt/intel_context_sseu.o \
[all …]
/linux/drivers/gpu/drm/xe/tests/
H A Dxe_gt_sriov_pf_service_test.c20 struct xe_gt *gt; in pf_service_test_init() local
28 gt = xe_device_get_gt(xe, 0); in pf_service_test_init()
29 pf_init_versions(gt); in pf_service_test_init()
36 KUNIT_ASSERT_NE(test, 0, gt->sriov.pf.service.version.base.major); in pf_service_test_init()
37 KUNIT_ASSERT_NE(test, 0, gt->sriov.pf.service.version.latest.major); in pf_service_test_init()
38 KUNIT_ASSERT_LE(test, gt->sriov.pf.service.version.base.major, in pf_service_test_init()
39 gt->sriov.pf.service.version.latest.major); in pf_service_test_init()
40 if (gt->sriov.pf.service.version.base.major == gt->sriov.pf.service.version.latest.major) in pf_service_test_init()
41 KUNIT_ASSERT_LE(test, gt->sriov.pf.service.version.base.minor, in pf_service_test_init()
42 gt->sriov.pf.service.version.latest.minor); in pf_service_test_init()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/device/
H A Dpci.c80 { 0x106b, 0x0605, "GeForce GT 130" },
86 { 0x1043, 0x202d, "GeForce GT 220M" },
92 { 0x152d, 0x0850, "GeForce GT 240M LE" },
98 { 0x1043, 0x14a2, "GeForce GT 320M" },
99 { 0x1043, 0x14d2, "GeForce GT 320M" },
105 { 0x106b, 0x0633, "GeForce GT 120" },
111 { 0x106b, 0x0693, "GeForce GT 120" },
232 { 0x1025, 0x0692, "GeForce GT 620M" },
233 { 0x1025, 0x0725, "GeForce GT 620M" },
234 { 0x1025, 0x0728, "GeForce GT 62
[all...]

12345678910>>...28