Home
last modified time | relevance | path

Searched refs:device (Results 1 – 25 of 8054) sorted by relevance

12345678910>>...323

/linux/drivers/gpu/drm/mediatek/
H A Dmtk_disp_drv.h15 int mtk_aal_clk_enable(struct device *dev);
16 void mtk_aal_clk_disable(struct device *dev);
17 void mtk_aal_config(struct device *dev, unsigned int w,
20 unsigned int mtk_aal_gamma_get_lut_size(struct device *dev);
21 void mtk_aal_gamma_set(struct device *dev, struct drm_crtc_state *state);
22 void mtk_aal_start(struct device *dev);
23 void mtk_aal_stop(struct device *dev);
25 void mtk_ccorr_ctm_set(struct device *dev, struct drm_crtc_state *state);
26 int mtk_ccorr_clk_enable(struct device *dev);
27 void mtk_ccorr_clk_disable(struct device *dev);
[all …]
/linux/drivers/block/drbd/
H A Ddrbd_actlog.c82 void *drbd_md_get_buffer(struct drbd_device *device, const char *intent) in drbd_md_get_buffer() argument
86 wait_event(device->misc_wait, in drbd_md_get_buffer()
87 (r = atomic_cmpxchg(&device->md_io.in_use, 0, 1)) == 0 || in drbd_md_get_buffer()
88 device->state.disk <= D_FAILED); in drbd_md_get_buffer()
93 device->md_io.current_use = intent; in drbd_md_get_buffer()
94 device->md_io.start_jif = jiffies; in drbd_md_get_buffer()
95 device->md_io.submit_jif = device->md_io.start_jif - 1; in drbd_md_get_buffer()
96 return page_address(device->md_io.page); in drbd_md_get_buffer()
99 void drbd_md_put_buffer(struct drbd_device *device) in drbd_md_put_buffer() argument
101 if (atomic_dec_and_test(&device->md_io.in_use)) in drbd_md_put_buffer()
[all …]
H A Ddrbd_worker.c52 struct drbd_device *device; in drbd_md_endio() local
54 device = bio->bi_private; in drbd_md_endio()
55 device->md_io.error = blk_status_to_errno(bio->bi_status); in drbd_md_endio()
58 if (device->ldev) in drbd_md_endio()
59 put_ldev(device); in drbd_md_endio()
73 drbd_md_put_buffer(device); in drbd_md_endio()
74 device->md_io.done = 1; in drbd_md_endio()
75 wake_up(&device->misc_wait); in drbd_md_endio()
85 struct drbd_device *device = peer_device->device; in drbd_endio_read_sec_final() local
87 spin_lock_irqsave(&device->resource->req_lock, flags); in drbd_endio_read_sec_final()
[all …]
/linux/drivers/s390/char/
H A Dtape_core.c93 tape_medium_state_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_medium_state_show()
105 tape_first_minor_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_first_minor_show()
117 tape_state_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_state_show()
130 tape_operation_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_operation_show()
157 tape_blocksize_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_blocksize_show()
186 tape_state_set(struct tape_device *device, enum tape_state newstate) in tape_state_set() argument
190 if (device->tape_state == TS_NOT_OPER) { in tape_state_set()
194 DBF_EVENT(4, "ts. dev: %x\n", device->first_minor); in tape_state_set()
196 if (device->tape_state < TS_SIZE && device->tape_state >=0 ) in tape_state_set()
197 str = tape_state_verbose[device->tape_state]; in tape_state_set()
[all …]
/linux/include/linux/
H A Dpm_runtime.h68 extern int pm_generic_runtime_suspend(struct device *dev);
69 extern int pm_generic_runtime_resume(struct device *dev);
70 extern int pm_runtime_force_suspend(struct device *dev);
72 extern int __pm_runtime_idle(struct device *dev, int rpmflags);
73 extern int __pm_runtime_suspend(struct device *dev, int rpmflags);
74 extern int __pm_runtime_resume(struct device *dev, int rpmflags);
75 extern int pm_runtime_get_if_active(struct device *dev);
76 extern int pm_runtime_get_if_in_use(struct device *dev);
77 extern int pm_schedule_suspend(struct device *dev, unsigned int delay);
78 extern int __pm_runtime_set_status(struct device *dev, unsigned int status);
[all …]
H A Ddevice.h3 * device.h - generic, centralized driver model
29 #include <linux/device/bus.h>
30 #include <linux/device/class.h>
31 #include <linux/device/devres.h>
32 #include <linux/device/driver.h>
34 #include <asm/device.h>
36 struct device;
51 * struct subsys_interface - interfaces to device functions
52 * @name: name of the device function
55 * @add_dev: device hooku
565 struct device { global() struct
567 parentdevice global() argument
569 pdevice global() argument
571 init_namedevice global() argument
572 typedevice global() argument
574 busdevice global() argument
575 driverdevice global() argument
577 platform_datadevice global() argument
579 driver_datadevice global() argument
604 dma_maskdevice global() argument
613 dma_parmsdevice global() argument
636 of_nodedevice global() argument
637 fwnodedevice global() argument
640 numa_nodedevice global() argument
643 iddevice global() argument
651 releasedevice global() argument
655 physical_locationdevice global() argument
657 removabledevice global() argument
682 dma_iommudevice global() argument
967 DEFINE_GUARD(device,struct device *,device_lock (_T),device_unlock (_T)) DEFINE_GUARD() argument
[all...]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/
H A Dnv40.c36 return nvkm_rd32(gr->engine.subdev.device, 0x1540); in nv40_gr_units()
47 int ret = nvkm_gpuobj_new(object->engine->subdev.device, 20, align, in nv40_gr_object_bind()
79 int ret = nvkm_gpuobj_new(gr->base.engine.subdev.device, gr->size, in nv40_gr_chan_bind()
84 nv40_grctx_fill(gr->base.engine.subdev.device, *pgpuobj); in nv40_gr_chan_bind()
97 struct nvkm_device *device = subdev->device; in nv40_gr_chan_fini() local
101 nvkm_mask(device, 0x400720, 0x00000001, 0x00000000); in nv40_gr_chan_fini()
103 if (nvkm_rd32(device, 0x40032c) == inst) { in nv40_gr_chan_fini()
105 nvkm_wr32(device, 0x400720, 0x00000000); in nv40_gr_chan_fini()
106 nvkm_wr32(device, 0x400784, inst); in nv40_gr_chan_fini()
107 nvkm_mask(device, 0x400310, 0x00000020, 0x00000020); in nv40_gr_chan_fini()
[all …]
H A Dnv50.c35 return nvkm_rd32(gr->engine.subdev.device, 0x1540); in nv50_gr_units()
46 int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, in nv50_gr_object_bind()
73 int ret = nvkm_gpuobj_new(gr->base.engine.subdev.device, gr->size, in nv50_gr_chan_bind()
77 nv50_grctx_fill(gr->base.engine.subdev.device, *pgpuobj); in nv50_gr_chan_bind()
243 struct nvkm_device *device = subdev->device; in nv50_gr_prop_trap() local
244 u32 e0c = nvkm_rd32(device, ustatus_addr + 0x04); in nv50_gr_prop_trap()
245 u32 e10 = nvkm_rd32(device, ustatus_addr + 0x08); in nv50_gr_prop_trap()
246 u32 e14 = nvkm_rd32(device, ustatus_addr + 0x0c); in nv50_gr_prop_trap()
247 u32 e18 = nvkm_rd32(device, ustatus_addr + 0x10); in nv50_gr_prop_trap()
248 u32 e1c = nvkm_rd32(device, ustatus_addr + 0x14); in nv50_gr_prop_trap()
[all …]
H A Dnv20.c34 struct nvkm_device *device = gr->base.engine.subdev.device; in nv20_gr_chan_fini() local
38 nvkm_mask(device, 0x400720, 0x00000001, 0x00000000); in nv20_gr_chan_fini()
39 if (nvkm_rd32(device, 0x400144) & 0x00010000) in nv20_gr_chan_fini()
40 chid = (nvkm_rd32(device, 0x400148) & 0x1f000000) >> 24; in nv20_gr_chan_fini()
42 nvkm_wr32(device, 0x400784, inst >> 4); in nv20_gr_chan_fini()
43 nvkm_wr32(device, 0x400788, 0x00000002); in nv20_gr_chan_fini()
44 nvkm_msec(device, 2000, in nv20_gr_chan_fini()
45 if (!nvkm_rd32(device, 0x400700)) in nv20_gr_chan_fini()
48 nvkm_wr32(device, 0x400144, 0x10000000); in nv20_gr_chan_fini()
49 nvkm_mask(device, 0x400148, 0xff000000, 0x1f000000); in nv20_gr_chan_fini()
[all …]
H A Dnv04.c445 nv04_gr_set_ctx1(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx1() argument
447 int subc = (nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR) >> 13) & 0x7; in nv04_gr_set_ctx1()
450 tmp = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx1()
453 nvkm_wr32(device, 0x700000 + inst, tmp); in nv04_gr_set_ctx1()
455 nvkm_wr32(device, NV04_PGRAPH_CTX_SWITCH1, tmp); in nv04_gr_set_ctx1()
456 nvkm_wr32(device, NV04_PGRAPH_CTX_CACHE1 + (subc << 2), tmp); in nv04_gr_set_ctx1()
460 nv04_gr_set_ctx_val(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx_val() argument
465 ctx1 = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx_val()
469 tmp = nvkm_rd32(device, 0x70000c + inst); in nv04_gr_set_ctx_val()
472 nvkm_wr32(device, 0x70000c + inst, tmp); in nv04_gr_set_ctx_val()
[all …]
/linux/drivers/s390/block/
H A Ddasd.c87 struct dasd_device *device; in dasd_alloc_device() local
89 device = kzalloc_obj(struct dasd_device, GFP_ATOMIC); in dasd_alloc_device()
90 if (!device) in dasd_alloc_device()
94 device->ccw_mem = (void *) __get_free_pages(GFP_ATOMIC | GFP_DMA, 1); in dasd_alloc_device()
95 if (!device->ccw_mem) { in dasd_alloc_device()
96 kfree(device); in dasd_alloc_device()
100 device->erp_mem = (void *) get_zeroed_page(GFP_ATOMIC | GFP_DMA); in dasd_alloc_device()
101 if (!device->erp_mem) { in dasd_alloc_device()
102 free_pages((unsigned long) device->ccw_mem, 1); in dasd_alloc_device()
103 kfree(device); in dasd_alloc_device()
[all …]
H A Ddasd_devmap.c49 struct dasd_device *device; member
431 new->device = NULL; in dasd_add_busid()
494 BUG_ON(devmap->device != NULL); in dasd_forget_ranges()
509 struct dasd_device *device; in dasd_device_from_devindex() local
521 if (devmap && devmap->device) { in dasd_device_from_devindex()
522 device = devmap->device; in dasd_device_from_devindex()
523 dasd_get_device(device); in dasd_device_from_devindex()
525 device = ERR_PTR(-ENODEV); in dasd_device_from_devindex()
527 return device; in dasd_device_from_devindex()
553 struct dasd_device *device; in dasd_create_device() local
[all …]
H A Ddasd_3990_erp.c70 struct dasd_device *device = erp->startdev; in dasd_3990_erp_block_queue() local
73 DBF_DEV_EVENT(DBF_INFO, device, in dasd_3990_erp_block_queue()
76 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_block_queue()
77 dasd_device_set_stop_bits(device, DASD_STOPPED_PENDING); in dasd_3990_erp_block_queue()
78 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_block_queue()
83 dasd_device_set_timer(device, expires); in dasd_3990_erp_block_queue()
102 struct dasd_device *device = erp->startdev; in dasd_3990_erp_int_req() local
115 dev_err(&device->cdev->dev, in dasd_3990_erp_int_req()
143 struct dasd_device *device = erp->startdev; in dasd_3990_erp_alternate_path() local
148 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_alternate_path()
[all …]
/linux/drivers/base/power/
H A Dpower.h4 static inline void device_pm_init_common(struct device *dev) in device_pm_init_common()
15 static inline void pm_runtime_early_init(struct device *dev) in pm_runtime_early_init()
21 extern void pm_runtime_init(struct device *dev);
22 extern void pm_runtime_reinit(struct device *dev);
23 extern void pm_runtime_remove(struct device *dev);
24 extern u64 pm_runtime_active_time(struct device *dev);
35 struct device *dev;
43 extern void dev_pm_enable_wake_irq_check(struct device *dev,
45 extern void dev_pm_disable_wake_irq_check(struct device *dev, bool cond_disable);
46 extern void dev_pm_enable_wake_irq_complete(struct device *dev);
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/
H A Dvga.c27 nvkm_rdport(struct nvkm_device *device, int head, u16 port) in nvkm_rdport() argument
29 if (device->card_type >= NV_50) in nvkm_rdport()
30 return nvkm_rd08(device, 0x601000 + port); in nvkm_rdport()
35 return nvkm_rd08(device, 0x601000 + (head * 0x2000) + port); in nvkm_rdport()
40 if (device->card_type < NV_40) in nvkm_rdport()
42 return nvkm_rd08(device, 0x0c0000 + (head * 0x2000) + port); in nvkm_rdport()
49 nvkm_wrport(struct nvkm_device *device, int head, u16 port, u8 data) in nvkm_wrport() argument
51 if (device->card_type >= NV_50) in nvkm_wrport()
52 nvkm_wr08(device, 0x601000 + port, data); in nvkm_wrport()
57 nvkm_wr08(device, 0x601000 + (head * 0x2000) + port, data); in nvkm_wrport()
[all …]
H A Dgf119.c39 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_device_entry() local
42 nvkm_mask(device, 0x616548 + hoff, 0x00000070, head << 4); in gf119_sor_hda_device_entry()
48 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_eld() local
53 nvkm_wr32(device, 0x10ec00 + soff, (i << 8) | data[i]); in gf119_sor_hda_eld()
55 nvkm_wr32(device, 0x10ec00 + soff, (i << 8)); in gf119_sor_hda_eld()
56 nvkm_mask(device, 0x10ec10 + soff, 0x80000002, 0x80000002); in gf119_sor_hda_eld()
62 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_hpd() local
74 nvkm_mask(device, 0x10ec10 + soff, mask, data); in gf119_sor_hda_hpd()
87 struct nvkm_device *device = sor->disp->engine.subdev.device; in gf119_sor_dp_watermark() local
90 nvkm_mask(device, 0x616610 + hoff, 0x0800003f, 0x08000000 | watermark); in gf119_sor_dp_watermark()
[all …]
H A Dgv100.c40 struct nvkm_device *device = ior->disp->engine.subdev.device; in gv100_sor_hda_device_entry() local
43 nvkm_mask(device, 0x616528 + hoff, 0x00000070, head << 4); in gv100_sor_hda_device_entry()
56 struct nvkm_device *device = sor->disp->engine.subdev.device; in gv100_sor_dp_watermark() local
59 nvkm_mask(device, 0x616550 + hoff, 0x0c00003f, 0x08000000 | watermark); in gv100_sor_dp_watermark()
65 struct nvkm_device *device = sor->disp->engine.subdev.device; in gv100_sor_dp_audio_sym() local
68 nvkm_mask(device, 0x616568 + hoff, 0x0000ffff, h); in gv100_sor_dp_audio_sym()
69 nvkm_mask(device, 0x61656c + hoff, 0x00ffffff, v); in gv100_sor_dp_audio_sym()
75 struct nvkm_device *device = sor->disp->engine.subdev.device; in gv100_sor_dp_audio() local
80 nvkm_mask(device, 0x616560 + hoff, mask, data); in gv100_sor_dp_audio()
81 nvkm_msec(device, 2000, in gv100_sor_dp_audio()
[all …]
/linux/drivers/firewire/
H A Dcore-device.c191 static const struct ieee1394_device_id *unit_match(struct device *dev, in unit_match()
207 static bool is_fw_unit(const struct device *dev);
209 static int fw_unit_match(struct device *dev, const struct device_driver *drv) in fw_unit_match()
215 static int fw_unit_probe(struct device *dev) in fw_unit_probe()
223 static void fw_unit_remove(struct device *dev) in fw_unit_remove()
242 static int fw_unit_uevent(const struct device *dev, struct kobj_uevent_env *env) in fw_unit_uevent()
263 int fw_device_enable_phys_dma(struct fw_device *device) in fw_device_enable_phys_dma() argument
265 int generation = device->generation; in fw_device_enable_phys_dma()
270 return device->card->driver->enable_phys_dma(device->card, in fw_device_enable_phys_dma()
271 device->node_id, in fw_device_enable_phys_dma()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/
H A Dgt215.c34 struct nvkm_device *device = subdev->device; in gt215_pmu_send() local
39 addr = nvkm_rd32(device, 0x10a4a0); in gt215_pmu_send()
40 if (nvkm_msec(device, 2000, in gt215_pmu_send()
41 u32 tmp = nvkm_rd32(device, 0x10a4b0); in gt215_pmu_send()
60 nvkm_wr32(device, 0x10a580, 0x00000001); in gt215_pmu_send()
61 } while (nvkm_rd32(device, 0x10a580) != 0x00000001); in gt215_pmu_send()
64 nvkm_wr32(device, 0x10a1c0, 0x01000000 | (((addr & 0x07) << 4) + in gt215_pmu_send()
66 nvkm_wr32(device, 0x10a1c4, process); in gt215_pmu_send()
67 nvkm_wr32(device, 0x10a1c4, message); in gt215_pmu_send()
68 nvkm_wr32(device, 0x10a1c4, data0); in gt215_pmu_send()
[all …]
H A Dgk104.c34 magic_(struct nvkm_device *device, u32 ctrl, int size) in magic_() argument
36 nvkm_wr32(device, 0x00c800, 0x00000000); in magic_()
37 nvkm_wr32(device, 0x00c808, 0x00000000); in magic_()
38 nvkm_wr32(device, 0x00c800, ctrl); in magic_()
39 nvkm_msec(device, 2000, in magic_()
40 if (nvkm_rd32(device, 0x00c800) & 0x40000000) { in magic_()
42 nvkm_wr32(device, 0x00c804, 0x00000000); in magic_()
46 nvkm_wr32(device, 0x00c800, 0x00000000); in magic_()
50 magic(struct nvkm_device *device, u32 ctrl) in magic() argument
52 magic_(device, 0x8000a41f | ctrl, 6); in magic()
[all …]
/linux/drivers/hid/
H A Dhid-roccat.c45 struct device *dev;
62 struct roccat_device *device; member
77 struct roccat_device *device = reader->device; in roccat_read() local
82 mutex_lock(&device->cbuf_lock); in roccat_read()
85 if (reader->cbuf_start == device->cbuf_end) { in roccat_read()
86 add_wait_queue(&device->wait, &wait); in roccat_read()
90 while (reader->cbuf_start == device->cbuf_end) { in roccat_read()
99 if (!device->exist) { in roccat_read()
104 mutex_unlock(&device->cbuf_lock); in roccat_read()
106 mutex_lock(&device->cbuf_lock); in roccat_read()
[all …]
/linux/tools/testing/selftests/vfio/lib/
H A Dvfio_pci_device.c30 static void vfio_pci_irq_set(struct vfio_pci_device *device, in vfio_pci_irq_set() argument
52 ioctl_assert(device->fd, VFIO_DEVICE_SET_IRQS, irq);
55 void vfio_pci_irq_trigger(struct vfio_pci_device *device, u32 index, u32 vector) in vfio_pci_irq_trigger()
65 ioctl_assert(device->fd, VFIO_DEVICE_SET_IRQS, &irq);
76 void vfio_pci_irq_enable(struct vfio_pci_device *device, u32 index, u32 vector, in vfio_pci_irq_enable()
84 VFIO_ASSERT_LT(device->msi_eventfds[i], 0); in vfio_pci_irq_enable()
85 device->msi_eventfds[i] = eventfd(0, 0); in vfio_pci_irq_enable()
86 VFIO_ASSERT_GE(device->msi_eventfds[i], 0); in vfio_pci_irq_enable()
89 vfio_pci_irq_set(device, index, vector, count, device in vfio_pci_irq_enable()
53 vfio_pci_irq_trigger(struct vfio_pci_device * device,u32 index,u32 vector) vfio_pci_irq_trigger() argument
74 vfio_pci_irq_enable(struct vfio_pci_device * device,u32 index,u32 vector,int count) vfio_pci_irq_enable() argument
90 vfio_pci_irq_disable(struct vfio_pci_device * device,u32 index) vfio_pci_irq_disable() argument
107 vfio_pci_irq_get(struct vfio_pci_device * device,u32 index,struct vfio_irq_info * irq_info) vfio_pci_irq_get() argument
116 vfio_pci_region_get(struct vfio_pci_device * device,int index,struct vfio_region_info * info) vfio_pci_region_get() argument
127 vfio_pci_bar_map(struct vfio_pci_device * device,int index) vfio_pci_bar_map() argument
164 vfio_pci_bar_unmap(struct vfio_pci_device * device,int index) vfio_pci_bar_unmap() argument
175 vfio_pci_bar_unmap_all(struct vfio_pci_device * device) vfio_pci_bar_unmap_all() argument
185 vfio_pci_config_access(struct vfio_pci_device * device,bool write,size_t config,size_t size,void * data) vfio_pci_config_access() argument
200 vfio_pci_device_reset(struct vfio_pci_device * device) vfio_pci_device_reset() argument
223 vfio_pci_group_setup(struct vfio_pci_device * device,const char * bdf) vfio_pci_group_setup() argument
243 vfio_pci_container_setup(struct vfio_pci_device * device,const char * bdf) vfio_pci_container_setup() argument
265 vfio_pci_device_setup(struct vfio_pci_device * device) vfio_pci_device_setup() argument
343 vfio_pci_iommufd_setup(struct vfio_pci_device * device,const char * bdf) vfio_pci_iommufd_setup() argument
357 struct vfio_pci_device *device; vfio_pci_device_init() local
377 vfio_pci_device_cleanup(struct vfio_pci_device * device) vfio_pci_device_cleanup() argument
[all...]
/linux/drivers/vfio/
H A Dvfio_main.c15 #include <linux/device.h>
62 MODULE_PARM_DESC(enable_unsafe_noiommu_mode, "Enable UNSAFE, no-IOMMU mode. This mode provides no device isolation, no DMA translation, no host kernel protection, cannot be used for device assignment to virtual machines, requires RAWIO permissions, and will taint the kernel. If you do not know what this is for, step away. (default: false)");
67 int vfio_assign_device_set(struct vfio_device *device, void *set_id) in vfio_assign_device_set() argument
110 device->dev_set = dev_set; in vfio_assign_device_set()
111 list_add_tail(&device->dev_set_list, &dev_set->device_list); in vfio_assign_device_set()
117 static void vfio_release_device_set(struct vfio_device *device) in vfio_release_device_set() argument
119 struct vfio_device_set *dev_set = device->dev_set; in vfio_release_device_set()
125 list_del(&device->dev_set_list); in vfio_release_device_set()
153 struct device *de in vfio_find_device_in_devset()
170 vfio_device_put_registration(struct vfio_device * device) vfio_device_put_registration() argument
177 vfio_device_try_get_registration(struct vfio_device * device) vfio_device_try_get_registration() argument
189 struct vfio_device *device = vfio_device_release() local
221 struct vfio_device *device; _vfio_alloc_device() local
273 vfio_init_device(struct vfio_device * device,struct device * dev,const struct vfio_device_ops * ops) vfio_init_device() argument
315 __vfio_register_dev(struct vfio_device * device,enum vfio_group_type type) __vfio_register_dev() argument
369 vfio_register_group_dev(struct vfio_device * device) vfio_register_group_dev() argument
379 vfio_register_emulated_iommu_dev(struct vfio_device * device) vfio_register_emulated_iommu_dev() argument
388 vfio_unregister_group_dev(struct vfio_device * device) vfio_unregister_group_dev() argument
436 vfio_device_get_kvm_safe(struct vfio_device * device,struct kvm * kvm) vfio_device_get_kvm_safe() argument
468 vfio_device_put_kvm(struct vfio_device * device) vfio_device_put_kvm() argument
488 vfio_assert_device_open(struct vfio_device * device) vfio_assert_device_open() argument
494 vfio_allocate_device_file(struct vfio_device * device) vfio_allocate_device_file() argument
510 struct vfio_device *device = df->device; vfio_df_device_first_open() local
545 struct vfio_device *device = df->device; vfio_df_device_last_close() local
561 struct vfio_device *device = df->device; vfio_df_open() local
585 struct vfio_device *device = df->device; vfio_df_close() local
600 vfio_device_pm_runtime_get(struct vfio_device * device) vfio_device_pm_runtime_get() argument
621 vfio_device_pm_runtime_put(struct vfio_device * device) vfio_device_pm_runtime_put() argument
635 struct vfio_device *device = df->device; vfio_device_fops_release() local
664 vfio_mig_get_next_state(struct vfio_device * device,enum vfio_device_mig_state cur_fsm,enum vfio_device_mig_state new_fsm,enum vfio_device_mig_state * next_fsm) vfio_mig_get_next_state() argument
890 vfio_ioctl_device_feature_mig_device_state(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_mig_device_state() argument
940 vfio_ioctl_device_feature_migration_data_size(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_migration_data_size() argument
967 vfio_ioctl_device_feature_migration(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_migration() argument
1050 vfio_ioctl_device_feature_logging_start(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_logging_start() argument
1134 vfio_ioctl_device_feature_logging_stop(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_logging_stop() argument
1155 struct vfio_device *device = opaque; vfio_device_log_read_and_clear() local
1161 vfio_ioctl_device_feature_logging_report(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_logging_report() argument
1205 vfio_ioctl_device_feature(struct vfio_device * device,struct vfio_device_feature __user * arg) vfio_ioctl_device_feature() argument
1263 vfio_get_region_info(struct vfio_device * device,struct vfio_region_info __user * arg) vfio_get_region_info() argument
1312 struct vfio_device *device = df->device; vfio_device_fops_unl_ioctl() local
1365 struct vfio_device *device = df->device; vfio_device_fops_read() local
1382 struct vfio_device *device = df->device; vfio_device_fops_write() local
1397 struct vfio_device *device = df->device; vfio_device_fops_mmap() local
1414 struct vfio_device *device = df->device; vfio_device_show_fdinfo() local
1470 struct vfio_device *device; vfio_file_enforced_coherent() local
1660 vfio_pin_pages(struct vfio_device * device,dma_addr_t iova,int npage,int prot,struct page ** pages) vfio_pin_pages() argument
1701 vfio_unpin_pages(struct vfio_device * device,dma_addr_t iova,int npage) vfio_unpin_pages() argument
1740 vfio_dma_rw(struct vfio_device * device,dma_addr_t iova,void * data,size_t len,bool write) vfio_dma_rw() argument
[all...]
/linux/tools/testing/selftests/vfio/lib/drivers/ioat/
H A Dioat.c26 static inline struct ioat_state *to_ioat_state(struct vfio_pci_device *device) in to_ioat_state() argument
28 return device->driver.region.vaddr; in to_ioat_state()
31 static inline void *ioat_channel_registers(struct vfio_pci_device *device) in ioat_channel_registers() argument
33 return device->bars[0].vaddr + IOAT_CHANNEL_MMIO_SIZE; in ioat_channel_registers()
36 static int ioat_probe(struct vfio_pci_device *device) in ioat_probe() argument
41 if (!vfio_pci_device_match(device, PCI_VENDOR_ID_INTEL, in ioat_probe()
45 VFIO_ASSERT_NOT_NULL(device->bars[0].vaddr); in ioat_probe()
47 version = readb(device->bars[0].vaddr + IOAT_VER_OFFSET); in ioat_probe()
54 dev_err(device, "ioat: Unsupported version: 0x%x\n", version); in ioat_probe()
65 static void ioat_clear_errors(struct vfio_pci_device *device) in ioat_clear_errors() argument
[all …]
/linux/tools/testing/selftests/vfio/lib/drivers/dsa/
H A Ddsa.c28 /* Cached device registers (and derived data) for easy access */
47 static inline struct dsa_state *to_dsa_state(struct vfio_pci_device *device) in to_dsa_state() argument
49 return device->driver.region.vaddr; in to_dsa_state()
52 static bool dsa_int_handle_request_required(struct vfio_pci_device *device) in dsa_int_handle_request_required() argument
54 void *bar0 = device->bars[0].vaddr; in dsa_int_handle_request_required()
66 static int dsa_probe(struct vfio_pci_device *device) in dsa_probe() argument
68 const u16 vendor_id = vfio_pci_config_readw(device, PCI_VENDOR_ID); in dsa_probe()
69 const u16 device_id = vfio_pci_config_readw(device, PCI_DEVICE_ID); in dsa_probe()
83 if (dsa_int_handle_request_required(device)) { in dsa_check_sw_err()
84 dev_err(device, "Devic in dsa_check_sw_err()
80 dsa_check_sw_err(struct vfio_pci_device * device) dsa_check_sw_err() argument
115 dsa_command(struct vfio_pci_device * device,u32 cmd) dsa_command() argument
140 dsa_wq_init(struct vfio_pci_device * device) dsa_wq_init() argument
170 dsa_group_init(struct vfio_pci_device * device) dsa_group_init() argument
184 dsa_register_cache_init(struct vfio_pci_device * device) dsa_register_cache_init() argument
207 dsa_init(struct vfio_pci_device * device) dsa_init() argument
235 dsa_remove(struct vfio_pci_device * device) dsa_remove() argument
241 dsa_completion_wait(struct vfio_pci_device * device,struct dsa_completion_record * completion) dsa_completion_wait() argument
263 dsa_copy_desc_init(struct vfio_pci_device * device,struct dsa_hw_desc * desc,iova_t src,iova_t dst,u64 size,bool interrupt) dsa_copy_desc_init() argument
288 dsa_batch_desc_init(struct vfio_pci_device * device,struct dsa_hw_desc * desc,u64 count) dsa_batch_desc_init() argument
304 dsa_desc_write(struct vfio_pci_device * device,struct dsa_hw_desc * desc) dsa_desc_write() argument
310 dsa_memcpy_one(struct vfio_pci_device * device,iova_t src,iova_t dst,u64 size,bool interrupt) dsa_memcpy_one() argument
321 dsa_memcpy_batch(struct vfio_pci_device * device,iova_t src,iova_t dst,u64 size,u64 count) dsa_memcpy_batch() argument
367 dsa_memcpy_start(struct vfio_pci_device * device,iova_t src,iova_t dst,u64 size,u64 count) dsa_memcpy_start() argument
381 dsa_memcpy_wait(struct vfio_pci_device * device) dsa_memcpy_wait() argument
396 dsa_send_msi(struct vfio_pci_device * device) dsa_send_msi() argument
[all...]

12345678910>>...323