Home
last modified time | relevance | path

Searched refs:device (Results 1 – 25 of 6986) sorted by relevance

12345678910>>...280

/linux/drivers/gpu/drm/mediatek/
H A Dmtk_disp_drv.h15 int mtk_aal_clk_enable(struct device *dev);
16 void mtk_aal_clk_disable(struct device *dev);
17 void mtk_aal_config(struct device *dev, unsigned int w,
20 unsigned int mtk_aal_gamma_get_lut_size(struct device *dev);
21 void mtk_aal_gamma_set(struct device *dev, struct drm_crtc_state *state);
22 void mtk_aal_start(struct device *dev);
23 void mtk_aal_stop(struct device *dev);
25 void mtk_ccorr_ctm_set(struct device *dev, struct drm_crtc_state *state);
26 int mtk_ccorr_clk_enable(struct device *dev);
27 void mtk_ccorr_clk_disable(struct device *dev);
[all …]
/linux/drivers/block/drbd/
H A Ddrbd_actlog.c82 void *drbd_md_get_buffer(struct drbd_device *device, const char *intent) in drbd_md_get_buffer() argument
86 wait_event(device->misc_wait, in drbd_md_get_buffer()
87 (r = atomic_cmpxchg(&device->md_io.in_use, 0, 1)) == 0 || in drbd_md_get_buffer()
88 device->state.disk <= D_FAILED); in drbd_md_get_buffer()
93 device->md_io.current_use = intent; in drbd_md_get_buffer()
94 device->md_io.start_jif = jiffies; in drbd_md_get_buffer()
95 device->md_io.submit_jif = device->md_io.start_jif - 1; in drbd_md_get_buffer()
96 return page_address(device->md_io.page); in drbd_md_get_buffer()
99 void drbd_md_put_buffer(struct drbd_device *device) in drbd_md_put_buffer() argument
101 if (atomic_dec_and_test(&device->md_io.in_use)) in drbd_md_put_buffer()
[all …]
H A Ddrbd_worker.c52 struct drbd_device *device; in drbd_md_endio() local
54 device = bio->bi_private; in drbd_md_endio()
55 device->md_io.error = blk_status_to_errno(bio->bi_status); in drbd_md_endio()
58 if (device->ldev) in drbd_md_endio()
59 put_ldev(device); in drbd_md_endio()
73 drbd_md_put_buffer(device); in drbd_md_endio()
74 device->md_io.done = 1; in drbd_md_endio()
75 wake_up(&device->misc_wait); in drbd_md_endio()
85 struct drbd_device *device = peer_device->device; in drbd_endio_read_sec_final() local
87 spin_lock_irqsave(&device->resource->req_lock, flags); in drbd_endio_read_sec_final()
[all …]
/linux/include/linux/
H A Dpm_runtime.h68 extern int pm_generic_runtime_suspend(struct device *dev);
69 extern int pm_generic_runtime_resume(struct device *dev);
70 extern int pm_runtime_force_suspend(struct device *dev);
72 extern int __pm_runtime_idle(struct device *dev, int rpmflags);
73 extern int __pm_runtime_suspend(struct device *dev, int rpmflags);
74 extern int __pm_runtime_resume(struct device *dev, int rpmflags);
75 extern int pm_runtime_get_if_active(struct device *dev);
76 extern int pm_runtime_get_if_in_use(struct device *dev);
77 extern int pm_schedule_suspend(struct device *dev, unsigned int delay);
78 extern int __pm_runtime_set_status(struct device *dev, unsigned int status);
[all …]
H A Ddevice.h36 struct device;
67 int (*add_dev)(struct device *dev, struct subsys_interface *sif);
68 void (*remove_dev)(struct device *dev, struct subsys_interface *sif);
91 int (*uevent)(const struct device *dev, struct kobj_uevent_env *env);
92 char *(*devnode)(const struct device *dev, umode_t *mode,
94 void (*release)(struct device *dev);
107 ssize_t (*show)(struct device *dev, struct device_attribute *attr,
109 ssize_t (*store)(struct device *dev, struct device_attribute *attr,
123 ssize_t device_show_ulong(struct device *dev, struct device_attribute *attr,
125 ssize_t device_store_ulong(struct device *dev, struct device_attribute *attr,
[all …]
H A Ddma-mapping.h112 void debug_dma_mapping_error(struct device *dev, dma_addr_t dma_addr);
113 void debug_dma_map_single(struct device *dev, const void *addr,
116 static inline void debug_dma_mapping_error(struct device *dev, in debug_dma_mapping_error()
120 static inline void debug_dma_map_single(struct device *dev, const void *addr, in debug_dma_map_single()
127 static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr) in dma_mapping_error()
136 dma_addr_t dma_map_page_attrs(struct device *dev, struct page *page,
139 void dma_unmap_page_attrs(struct device *dev, dma_addr_t addr, size_t size,
141 dma_addr_t dma_map_phys(struct device *dev, phys_addr_t phys, size_t size,
143 void dma_unmap_phys(struct device *dev, dma_addr_t addr, size_t size,
145 unsigned int dma_map_sg_attrs(struct device *dev, struct scatterlist *sg,
[all …]
/linux/drivers/base/power/
H A Dpower.h4 static inline void device_pm_init_common(struct device *dev) in device_pm_init_common()
15 static inline void pm_runtime_early_init(struct device *dev) in pm_runtime_early_init()
21 extern void pm_runtime_init(struct device *dev);
22 extern void pm_runtime_reinit(struct device *dev);
23 extern void pm_runtime_remove(struct device *dev);
24 extern u64 pm_runtime_active_time(struct device *dev);
35 struct device *dev;
43 extern void dev_pm_enable_wake_irq_check(struct device *dev,
45 extern void dev_pm_disable_wake_irq_check(struct device *dev, bool cond_disable);
46 extern void dev_pm_enable_wake_irq_complete(struct device *dev);
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/
H A Dnv50.c35 return nvkm_rd32(gr->engine.subdev.device, 0x1540); in nv50_gr_units()
46 int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, in nv50_gr_object_bind()
73 int ret = nvkm_gpuobj_new(gr->base.engine.subdev.device, gr->size, in nv50_gr_chan_bind()
77 nv50_grctx_fill(gr->base.engine.subdev.device, *pgpuobj); in nv50_gr_chan_bind()
243 struct nvkm_device *device = subdev->device; in nv50_gr_prop_trap() local
244 u32 e0c = nvkm_rd32(device, ustatus_addr + 0x04); in nv50_gr_prop_trap()
245 u32 e10 = nvkm_rd32(device, ustatus_addr + 0x08); in nv50_gr_prop_trap()
246 u32 e14 = nvkm_rd32(device, ustatus_addr + 0x0c); in nv50_gr_prop_trap()
247 u32 e18 = nvkm_rd32(device, ustatus_addr + 0x10); in nv50_gr_prop_trap()
248 u32 e1c = nvkm_rd32(device, ustatus_addr + 0x14); in nv50_gr_prop_trap()
[all …]
H A Dnv04.c445 nv04_gr_set_ctx1(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx1() argument
447 int subc = (nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR) >> 13) & 0x7; in nv04_gr_set_ctx1()
450 tmp = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx1()
453 nvkm_wr32(device, 0x700000 + inst, tmp); in nv04_gr_set_ctx1()
455 nvkm_wr32(device, NV04_PGRAPH_CTX_SWITCH1, tmp); in nv04_gr_set_ctx1()
456 nvkm_wr32(device, NV04_PGRAPH_CTX_CACHE1 + (subc << 2), tmp); in nv04_gr_set_ctx1()
460 nv04_gr_set_ctx_val(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx_val() argument
465 ctx1 = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx_val()
469 tmp = nvkm_rd32(device, 0x70000c + inst); in nv04_gr_set_ctx_val()
472 nvkm_wr32(device, 0x70000c + inst, tmp); in nv04_gr_set_ctx_val()
[all …]
H A Dnv10.c417 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr); \
419 state[__i] = nvkm_rd32(device, NV10_PGRAPH_PIPE_DATA); \
425 nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr); \
427 nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, state[__i]); \
433 struct nvkm_device *device = chan->object.engine->subdev.device; in nv17_gr_mthd_lma_window() local
447 PIPE_SAVE(device, pipe_0x0040, 0x0040); in nv17_gr_mthd_lma_window()
448 PIPE_SAVE(device, pipe->pipe_0x0200, 0x0200); in nv17_gr_mthd_lma_window()
450 PIPE_RESTORE(device, chan->lma_window, 0x6790); in nv17_gr_mthd_lma_window()
454 xfmode0 = nvkm_rd32(device, NV10_PGRAPH_XFMODE0); in nv17_gr_mthd_lma_window()
455 xfmode1 = nvkm_rd32(device, NV10_PGRAPH_XFMODE1); in nv17_gr_mthd_lma_window()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/
H A Dvga.c27 nvkm_rdport(struct nvkm_device *device, int head, u16 port) in nvkm_rdport() argument
29 if (device->card_type >= NV_50) in nvkm_rdport()
30 return nvkm_rd08(device, 0x601000 + port); in nvkm_rdport()
35 return nvkm_rd08(device, 0x601000 + (head * 0x2000) + port); in nvkm_rdport()
40 if (device->card_type < NV_40) in nvkm_rdport()
42 return nvkm_rd08(device, 0x0c0000 + (head * 0x2000) + port); in nvkm_rdport()
49 nvkm_wrport(struct nvkm_device *device, int head, u16 port, u8 data) in nvkm_wrport() argument
51 if (device->card_type >= NV_50) in nvkm_wrport()
52 nvkm_wr08(device, 0x601000 + port, data); in nvkm_wrport()
57 nvkm_wr08(device, 0x601000 + (head * 0x2000) + port, data); in nvkm_wrport()
[all …]
H A Dgf119.c39 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_device_entry() local
42 nvkm_mask(device, 0x616548 + hoff, 0x00000070, head << 4); in gf119_sor_hda_device_entry()
48 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_eld() local
53 nvkm_wr32(device, 0x10ec00 + soff, (i << 8) | data[i]); in gf119_sor_hda_eld()
55 nvkm_wr32(device, 0x10ec00 + soff, (i << 8)); in gf119_sor_hda_eld()
56 nvkm_mask(device, 0x10ec10 + soff, 0x80000002, 0x80000002); in gf119_sor_hda_eld()
62 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_hpd() local
74 nvkm_mask(device, 0x10ec10 + soff, mask, data); in gf119_sor_hda_hpd()
87 struct nvkm_device *device = sor->disp->engine.subdev.device; in gf119_sor_dp_watermark() local
90 nvkm_mask(device, 0x616610 + hoff, 0x0800003f, 0x08000000 | watermark); in gf119_sor_dp_watermark()
[all …]
H A Dg94.c36 struct nvkm_device *device = sor->disp->engine.subdev.device; in g94_sor_dp_watermark() local
39 nvkm_mask(device, 0x61c128 + loff, 0x0000003f, watermark); in g94_sor_dp_watermark()
46 struct nvkm_device *device = sor->disp->engine.subdev.device; in g94_sor_dp_activesym() local
49 nvkm_mask(device, 0x61c10c + loff, 0x000001fc, TU << 2); in g94_sor_dp_activesym()
50 nvkm_mask(device, 0x61c128 + loff, 0x010f7f00, VTUa << 24 | VTUf << 16 | VTUi << 8); in g94_sor_dp_activesym()
56 struct nvkm_device *device = sor->disp->engine.subdev.device; in g94_sor_dp_audio_sym() local
59 nvkm_mask(device, 0x61c1e8 + soff, 0x0000ffff, h); in g94_sor_dp_audio_sym()
60 nvkm_mask(device, 0x61c1ec + soff, 0x00ffffff, v); in g94_sor_dp_audio_sym()
66 struct nvkm_device *device = sor->disp->engine.subdev.device; in g94_sor_dp_drive() local
71 data[0] = nvkm_rd32(device, 0x61c118 + loff) & ~(0x000000ff << shift); in g94_sor_dp_drive()
[all …]
H A Dtu102.c36 struct nvkm_device *device = sor->disp->engine.subdev.device; in tu102_sor_dp_vcpi() local
39 nvkm_mask(device, 0x61657c + hoff, 0xffffffff, (aligned << 16) | pbn); in tu102_sor_dp_vcpi()
40 nvkm_mask(device, 0x616578 + hoff, 0x00003f3f, (slot_nr << 8) | slot); in tu102_sor_dp_vcpi()
46 struct nvkm_device *device = sor->disp->engine.subdev.device; in tu102_sor_dp_links() local
59 nvkm_mask(device, 0x612300 + soff, 0x007c0000, clksor); in tu102_sor_dp_links()
62 nvkm_msec(device, 40, NVKM_DELAY); in tu102_sor_dp_links()
63 nvkm_mask(device, 0x612300 + soff, 0x00030000, 0x00010000); in tu102_sor_dp_links()
64 nvkm_mask(device, 0x61c10c + loff, 0x00000003, 0x00000001); in tu102_sor_dp_links()
66 nvkm_mask(device, 0x61c10c + loff, 0x401f4000, dpctrl); in tu102_sor_dp_links()
101 struct nvkm_device *device = disp->engine.subdev.device; in tu102_sor_new() local
[all …]
H A Dgt215.c37 struct nvkm_device *device = ior->disp->engine.subdev.device; in gt215_sor_hda_eld() local
42 nvkm_wr32(device, 0x61c440 + soff, (i << 8) | data[i]); in gt215_sor_hda_eld()
44 nvkm_wr32(device, 0x61c440 + soff, (i << 8)); in gt215_sor_hda_eld()
45 nvkm_mask(device, 0x61c448 + soff, 0x80000002, 0x80000002); in gt215_sor_hda_eld()
51 struct nvkm_device *device = ior->disp->engine.subdev.device; in gt215_sor_hda_hpd() local
58 nvkm_mask(device, 0x61c448 + ior->id * 0x800, mask, data); in gt215_sor_hda_hpd()
70 struct nvkm_device *device = sor->disp->engine.subdev.device; in gt215_sor_dp_audio() local
75 nvkm_mask(device, 0x61c1e0 + soff, mask, data); in gt215_sor_dp_audio()
76 nvkm_msec(device, 2000, in gt215_sor_dp_audio()
77 if (!(nvkm_rd32(device, 0x61c1e0 + soff) & 0x80000000)) in gt215_sor_dp_audio()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/
H A Dgt215.c34 struct nvkm_device *device = subdev->device; in gt215_pmu_send() local
39 addr = nvkm_rd32(device, 0x10a4a0); in gt215_pmu_send()
40 if (nvkm_msec(device, 2000, in gt215_pmu_send()
41 u32 tmp = nvkm_rd32(device, 0x10a4b0); in gt215_pmu_send()
60 nvkm_wr32(device, 0x10a580, 0x00000001); in gt215_pmu_send()
61 } while (nvkm_rd32(device, 0x10a580) != 0x00000001); in gt215_pmu_send()
64 nvkm_wr32(device, 0x10a1c0, 0x01000000 | (((addr & 0x07) << 4) + in gt215_pmu_send()
66 nvkm_wr32(device, 0x10a1c4, process); in gt215_pmu_send()
67 nvkm_wr32(device, 0x10a1c4, message); in gt215_pmu_send()
68 nvkm_wr32(device, 0x10a1c4, data0); in gt215_pmu_send()
[all …]
H A Dgk104.c34 magic_(struct nvkm_device *device, u32 ctrl, int size) in magic_() argument
36 nvkm_wr32(device, 0x00c800, 0x00000000); in magic_()
37 nvkm_wr32(device, 0x00c808, 0x00000000); in magic_()
38 nvkm_wr32(device, 0x00c800, ctrl); in magic_()
39 nvkm_msec(device, 2000, in magic_()
40 if (nvkm_rd32(device, 0x00c800) & 0x40000000) { in magic_()
42 nvkm_wr32(device, 0x00c804, 0x00000000); in magic_()
46 nvkm_wr32(device, 0x00c800, 0x00000000); in magic_()
50 magic(struct nvkm_device *device, u32 ctrl) in magic() argument
52 magic_(device, 0x8000a41f | ctrl, 6); in magic()
[all …]
/linux/drivers/hid/
H A Dhid-roccat.c45 struct device *dev;
62 struct roccat_device *device; member
77 struct roccat_device *device = reader->device; in roccat_read() local
82 mutex_lock(&device->cbuf_lock); in roccat_read()
85 if (reader->cbuf_start == device->cbuf_end) { in roccat_read()
86 add_wait_queue(&device->wait, &wait); in roccat_read()
90 while (reader->cbuf_start == device->cbuf_end) { in roccat_read()
99 if (!device->exist) { in roccat_read()
104 mutex_unlock(&device->cbuf_lock); in roccat_read()
106 mutex_lock(&device->cbuf_lock); in roccat_read()
[all …]
/linux/tools/testing/selftests/vfio/lib/
H A Dvfio_pci_device.c27 static void vfio_pci_irq_set(struct vfio_pci_device *device, in vfio_pci_irq_set() argument
47 ioctl_assert(device->fd, VFIO_DEVICE_SET_IRQS, irq); in vfio_pci_irq_set()
50 void vfio_pci_irq_trigger(struct vfio_pci_device *device, u32 index, u32 vector) in vfio_pci_irq_trigger() argument
60 ioctl_assert(device->fd, VFIO_DEVICE_SET_IRQS, &irq); in vfio_pci_irq_trigger()
71 void vfio_pci_irq_enable(struct vfio_pci_device *device, u32 index, u32 vector, in vfio_pci_irq_enable() argument
79 VFIO_ASSERT_LT(device->msi_eventfds[i], 0); in vfio_pci_irq_enable()
80 device->msi_eventfds[i] = eventfd(0, 0); in vfio_pci_irq_enable()
81 VFIO_ASSERT_GE(device->msi_eventfds[i], 0); in vfio_pci_irq_enable()
84 vfio_pci_irq_set(device, index, vector, count, device->msi_eventfds + vector); in vfio_pci_irq_enable()
87 void vfio_pci_irq_disable(struct vfio_pci_device *device, u32 index) in vfio_pci_irq_disable() argument
[all …]
/linux/drivers/vfio/
H A Dvfio_main.c67 int vfio_assign_device_set(struct vfio_device *device, void *set_id) in vfio_assign_device_set() argument
110 device->dev_set = dev_set; in vfio_assign_device_set()
111 list_add_tail(&device->dev_set_list, &dev_set->device_list); in vfio_assign_device_set()
117 static void vfio_release_device_set(struct vfio_device *device) in vfio_release_device_set() argument
119 struct vfio_device_set *dev_set = device->dev_set; in vfio_release_device_set()
125 list_del(&device->dev_set_list); in vfio_release_device_set()
153 struct device *dev) in vfio_find_device_in_devset()
170 void vfio_device_put_registration(struct vfio_device *device) in vfio_device_put_registration() argument
172 if (refcount_dec_and_test(&device->refcount)) in vfio_device_put_registration()
173 complete(&device->comp); in vfio_device_put_registration()
[all …]
/linux/drivers/s390/block/
H A Ddasd_3990_erp.c70 struct dasd_device *device = erp->startdev; in dasd_3990_erp_block_queue() local
73 DBF_DEV_EVENT(DBF_INFO, device, in dasd_3990_erp_block_queue()
76 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_block_queue()
77 dasd_device_set_stop_bits(device, DASD_STOPPED_PENDING); in dasd_3990_erp_block_queue()
78 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_block_queue()
83 dasd_device_set_timer(device, expires); in dasd_3990_erp_block_queue()
102 struct dasd_device *device = erp->startdev; in dasd_3990_erp_int_req() local
115 dev_err(&device->cdev->dev, in dasd_3990_erp_int_req()
143 struct dasd_device *device = erp->startdev; in dasd_3990_erp_alternate_path() local
148 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_alternate_path()
[all …]
/linux/tools/testing/selftests/vfio/lib/drivers/ioat/
H A Dioat.c26 static inline struct ioat_state *to_ioat_state(struct vfio_pci_device *device) in to_ioat_state() argument
28 return device->driver.region.vaddr; in to_ioat_state()
31 static inline void *ioat_channel_registers(struct vfio_pci_device *device) in ioat_channel_registers() argument
33 return device->bars[0].vaddr + IOAT_CHANNEL_MMIO_SIZE; in ioat_channel_registers()
36 static int ioat_probe(struct vfio_pci_device *device) in ioat_probe() argument
41 if (!vfio_pci_device_match(device, PCI_VENDOR_ID_INTEL, in ioat_probe()
45 VFIO_ASSERT_NOT_NULL(device->bars[0].vaddr); in ioat_probe()
47 version = readb(device->bars[0].vaddr + IOAT_VER_OFFSET); in ioat_probe()
54 dev_err(device, "ioat: Unsupported version: 0x%x\n", version); in ioat_probe()
65 static void ioat_clear_errors(struct vfio_pci_device *device) in ioat_clear_errors() argument
[all …]
/linux/tools/testing/selftests/vfio/lib/drivers/dsa/
H A Ddsa.c47 static inline struct dsa_state *to_dsa_state(struct vfio_pci_device *device) in to_dsa_state() argument
49 return device->driver.region.vaddr; in to_dsa_state()
52 static bool dsa_int_handle_request_required(struct vfio_pci_device *device) in dsa_int_handle_request_required() argument
54 void *bar0 = device->bars[0].vaddr; in dsa_int_handle_request_required()
66 static int dsa_probe(struct vfio_pci_device *device) in dsa_probe() argument
68 if (!vfio_pci_device_match(device, PCI_VENDOR_ID_INTEL, in dsa_probe()
72 if (dsa_int_handle_request_required(device)) { in dsa_probe()
73 dev_err(device, "Device requires requesting interrupt handles\n"); in dsa_probe()
80 static void dsa_check_sw_err(struct vfio_pci_device *device) in dsa_check_sw_err() argument
82 void *reg = device->bars[0].vaddr + IDXD_SWERR_OFFSET; in dsa_check_sw_err()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/ltc/
H A Dgf100.c33 struct nvkm_device *device = ltc->subdev.device; in gf100_ltc_cbc_clear() local
34 nvkm_wr32(device, 0x17e8cc, start); in gf100_ltc_cbc_clear()
35 nvkm_wr32(device, 0x17e8d0, limit); in gf100_ltc_cbc_clear()
36 nvkm_wr32(device, 0x17e8c8, 0x00000004); in gf100_ltc_cbc_clear()
42 struct nvkm_device *device = ltc->subdev.device; in gf100_ltc_cbc_wait() local
47 nvkm_msec(device, 2000, in gf100_ltc_cbc_wait()
48 if (!nvkm_rd32(device, addr)) in gf100_ltc_cbc_wait()
58 struct nvkm_device *device = ltc->subdev.device; in gf100_ltc_zbc_clear_color() local
59 nvkm_mask(device, 0x17ea44, 0x0000000f, i); in gf100_ltc_zbc_clear_color()
60 nvkm_wr32(device, 0x17ea48, color[0]); in gf100_ltc_zbc_clear_color()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/devinit/
H A Dnv04.c39 struct nvkm_device *device = subdev->device; in nv04_devinit_meminit() local
45 fb = fbmem_init(device); in nv04_devinit_meminit()
52 nvkm_wrvgas(device, 0, 1, nvkm_rdvgas(device, 0, 1) | 0x20); in nv04_devinit_meminit()
53 nvkm_mask(device, NV04_PFB_DEBUG_0, 0, NV04_PFB_DEBUG_0_REFRESH_OFF); in nv04_devinit_meminit()
55 nvkm_mask(device, NV04_PFB_BOOT_0, ~0, in nv04_devinit_meminit()
66 nvkm_mask(device, NV04_PFB_BOOT_0, in nv04_devinit_meminit()
69 nvkm_mask(device, NV04_PFB_DEBUG_0, in nv04_devinit_meminit()
76 nvkm_mask(device, NV04_PFB_BOOT_0, in nv04_devinit_meminit()
82 nvkm_mask(device, NV04_PFB_BOOT_0, in nv04_devinit_meminit()
89 nvkm_mask(device, NV04_PFB_BOOT_0, in nv04_devinit_meminit()
[all …]

12345678910>>...280