| /linux/drivers/gpu/drm/mediatek/ |
| H A D | mtk_disp_drv.h | 15 int mtk_aal_clk_enable(struct device *dev); 16 void mtk_aal_clk_disable(struct device *dev); 17 void mtk_aal_config(struct device *dev, unsigned int w, 20 unsigned int mtk_aal_gamma_get_lut_size(struct device *dev); 21 void mtk_aal_gamma_set(struct device *dev, struct drm_crtc_state *state); 22 void mtk_aal_start(struct device *dev); 23 void mtk_aal_stop(struct device *dev); 25 void mtk_ccorr_ctm_set(struct device *dev, struct drm_crtc_state *state); 26 int mtk_ccorr_clk_enable(struct device *dev); 27 void mtk_ccorr_clk_disable(struct device *dev); [all …]
|
| /linux/drivers/block/drbd/ |
| H A D | drbd_actlog.c | 82 void *drbd_md_get_buffer(struct drbd_device *device, const char *intent) in drbd_md_get_buffer() argument 86 wait_event(device->misc_wait, in drbd_md_get_buffer() 87 (r = atomic_cmpxchg(&device->md_io.in_use, 0, 1)) == 0 || in drbd_md_get_buffer() 88 device->state.disk <= D_FAILED); in drbd_md_get_buffer() 93 device->md_io.current_use = intent; in drbd_md_get_buffer() 94 device->md_io.start_jif = jiffies; in drbd_md_get_buffer() 95 device->md_io.submit_jif = device->md_io.start_jif - 1; in drbd_md_get_buffer() 96 return page_address(device->md_io.page); in drbd_md_get_buffer() 99 void drbd_md_put_buffer(struct drbd_device *device) in drbd_md_put_buffer() argument 101 if (atomic_dec_and_test(&device->md_io.in_use)) in drbd_md_put_buffer() [all …]
|
| H A D | drbd_worker.c | 52 struct drbd_device *device; in drbd_md_endio() local 54 device = bio->bi_private; in drbd_md_endio() 55 device->md_io.error = blk_status_to_errno(bio->bi_status); in drbd_md_endio() 58 if (device->ldev) in drbd_md_endio() 59 put_ldev(device); in drbd_md_endio() 73 drbd_md_put_buffer(device); in drbd_md_endio() 74 device->md_io.done = 1; in drbd_md_endio() 75 wake_up(&device->misc_wait); in drbd_md_endio() 85 struct drbd_device *device = peer_device->device; in drbd_endio_read_sec_final() local 87 spin_lock_irqsave(&device->resource->req_lock, flags); in drbd_endio_read_sec_final() [all …]
|
| /linux/drivers/s390/char/ |
| H A D | tape_core.c | 93 tape_medium_state_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_medium_state_show() 105 tape_first_minor_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_first_minor_show() 117 tape_state_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_state_show() 130 tape_operation_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_operation_show() 157 tape_blocksize_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_blocksize_show() 186 tape_state_set(struct tape_device *device, enum tape_state newstate) in tape_state_set() argument 190 if (device->tape_state == TS_NOT_OPER) { in tape_state_set() 194 DBF_EVENT(4, "ts. dev: %x\n", device->first_minor); in tape_state_set() 196 if (device->tape_state < TS_SIZE && device->tape_state >=0 ) in tape_state_set() 197 str = tape_state_verbose[device->tape_state]; in tape_state_set() [all …]
|
| /linux/include/linux/ |
| H A D | pm_runtime.h | 68 extern int pm_generic_runtime_suspend(struct device *dev); 69 extern int pm_generic_runtime_resume(struct device *dev); 70 extern int pm_runtime_force_suspend(struct device *dev); 72 extern int __pm_runtime_idle(struct device *dev, int rpmflags); 73 extern int __pm_runtime_suspend(struct device *dev, int rpmflags); 74 extern int __pm_runtime_resume(struct device *dev, int rpmflags); 75 extern int pm_runtime_get_if_active(struct device *dev); 76 extern int pm_runtime_get_if_in_use(struct device *dev); 77 extern int pm_schedule_suspend(struct device *dev, unsigned int delay); 78 extern int __pm_runtime_set_status(struct device *dev, unsigned int status); [all …]
|
| H A D | device.h | 3 * device.h - generic, centralized driver model 29 #include <linux/device/bus.h> 30 #include <linux/device/class.h> 31 #include <linux/device/devres.h> 32 #include <linux/device/driver.h> 34 #include <asm/device.h> 36 struct device; 51 * struct subsys_interface - interfaces to device functions 52 * @name: name of the device function 55 * @add_dev: device hooku 565 struct device { global() struct 567 parentdevice global() argument 569 pdevice global() argument 571 init_namedevice global() argument 572 typedevice global() argument 574 busdevice global() argument 575 driverdevice global() argument 577 platform_datadevice global() argument 579 driver_datadevice global() argument 604 dma_maskdevice global() argument 613 dma_parmsdevice global() argument 636 of_nodedevice global() argument 637 fwnodedevice global() argument 640 numa_nodedevice global() argument 643 iddevice global() argument 651 releasedevice global() argument 655 physical_locationdevice global() argument 657 removabledevice global() argument 682 dma_iommudevice global() argument 967 DEFINE_GUARD(device,struct device *,device_lock (_T),device_unlock (_T)) DEFINE_GUARD() argument [all...] |
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
| H A D | nv40.c | 36 return nvkm_rd32(gr->engine.subdev.device, 0x1540); in nv40_gr_units() 47 int ret = nvkm_gpuobj_new(object->engine->subdev.device, 20, align, in nv40_gr_object_bind() 79 int ret = nvkm_gpuobj_new(gr->base.engine.subdev.device, gr->size, in nv40_gr_chan_bind() 84 nv40_grctx_fill(gr->base.engine.subdev.device, *pgpuobj); in nv40_gr_chan_bind() 97 struct nvkm_device *device = subdev->device; in nv40_gr_chan_fini() local 101 nvkm_mask(device, 0x400720, 0x00000001, 0x00000000); in nv40_gr_chan_fini() 103 if (nvkm_rd32(device, 0x40032c) == inst) { in nv40_gr_chan_fini() 105 nvkm_wr32(device, 0x400720, 0x00000000); in nv40_gr_chan_fini() 106 nvkm_wr32(device, 0x400784, inst); in nv40_gr_chan_fini() 107 nvkm_mask(device, 0x400310, 0x00000020, 0x00000020); in nv40_gr_chan_fini() [all …]
|
| H A D | nv50.c | 35 return nvkm_rd32(gr->engine.subdev.device, 0x1540); in nv50_gr_units() 46 int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, in nv50_gr_object_bind() 73 int ret = nvkm_gpuobj_new(gr->base.engine.subdev.device, gr->size, in nv50_gr_chan_bind() 77 nv50_grctx_fill(gr->base.engine.subdev.device, *pgpuobj); in nv50_gr_chan_bind() 243 struct nvkm_device *device = subdev->device; in nv50_gr_prop_trap() local 244 u32 e0c = nvkm_rd32(device, ustatus_addr + 0x04); in nv50_gr_prop_trap() 245 u32 e10 = nvkm_rd32(device, ustatus_addr + 0x08); in nv50_gr_prop_trap() 246 u32 e14 = nvkm_rd32(device, ustatus_addr + 0x0c); in nv50_gr_prop_trap() 247 u32 e18 = nvkm_rd32(device, ustatus_addr + 0x10); in nv50_gr_prop_trap() 248 u32 e1c = nvkm_rd32(device, ustatus_addr + 0x14); in nv50_gr_prop_trap() [all …]
|
| H A D | nv20.c | 34 struct nvkm_device *device = gr->base.engine.subdev.device; in nv20_gr_chan_fini() local 38 nvkm_mask(device, 0x400720, 0x00000001, 0x00000000); in nv20_gr_chan_fini() 39 if (nvkm_rd32(device, 0x400144) & 0x00010000) in nv20_gr_chan_fini() 40 chid = (nvkm_rd32(device, 0x400148) & 0x1f000000) >> 24; in nv20_gr_chan_fini() 42 nvkm_wr32(device, 0x400784, inst >> 4); in nv20_gr_chan_fini() 43 nvkm_wr32(device, 0x400788, 0x00000002); in nv20_gr_chan_fini() 44 nvkm_msec(device, 2000, in nv20_gr_chan_fini() 45 if (!nvkm_rd32(device, 0x400700)) in nv20_gr_chan_fini() 48 nvkm_wr32(device, 0x400144, 0x10000000); in nv20_gr_chan_fini() 49 nvkm_mask(device, 0x400148, 0xff000000, 0x1f000000); in nv20_gr_chan_fini() [all …]
|
| H A D | nv04.c | 445 nv04_gr_set_ctx1(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx1() argument 447 int subc = (nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR) >> 13) & 0x7; in nv04_gr_set_ctx1() 450 tmp = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx1() 453 nvkm_wr32(device, 0x700000 + inst, tmp); in nv04_gr_set_ctx1() 455 nvkm_wr32(device, NV04_PGRAPH_CTX_SWITCH1, tmp); in nv04_gr_set_ctx1() 456 nvkm_wr32(device, NV04_PGRAPH_CTX_CACHE1 + (subc << 2), tmp); in nv04_gr_set_ctx1() 460 nv04_gr_set_ctx_val(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx_val() argument 465 ctx1 = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx_val() 469 tmp = nvkm_rd32(device, 0x70000c + inst); in nv04_gr_set_ctx_val() 472 nvkm_wr32(device, 0x70000c + inst, tmp); in nv04_gr_set_ctx_val() [all …]
|
| /linux/drivers/s390/block/ |
| H A D | dasd.c | 87 struct dasd_device *device; in dasd_alloc_device() local 89 device = kzalloc_obj(struct dasd_device, GFP_ATOMIC); in dasd_alloc_device() 90 if (!device) in dasd_alloc_device() 94 device->ccw_mem = (void *) __get_free_pages(GFP_ATOMIC | GFP_DMA, 1); in dasd_alloc_device() 95 if (!device->ccw_mem) { in dasd_alloc_device() 96 kfree(device); in dasd_alloc_device() 100 device->erp_mem = (void *) get_zeroed_page(GFP_ATOMIC | GFP_DMA); in dasd_alloc_device() 101 if (!device->erp_mem) { in dasd_alloc_device() 102 free_pages((unsigned long) device->ccw_mem, 1); in dasd_alloc_device() 103 kfree(device); in dasd_alloc_device() [all …]
|
| H A D | dasd_devmap.c | 49 struct dasd_device *device; member 431 new->device = NULL; in dasd_add_busid() 494 BUG_ON(devmap->device != NULL); in dasd_forget_ranges() 509 struct dasd_device *device; in dasd_device_from_devindex() local 521 if (devmap && devmap->device) { in dasd_device_from_devindex() 522 device = devmap->device; in dasd_device_from_devindex() 523 dasd_get_device(device); in dasd_device_from_devindex() 525 device = ERR_PTR(-ENODEV); in dasd_device_from_devindex() 527 return device; in dasd_device_from_devindex() 553 struct dasd_device *device; in dasd_create_device() local [all …]
|
| H A D | dasd_3990_erp.c | 70 struct dasd_device *device = erp->startdev; in dasd_3990_erp_block_queue() local 73 DBF_DEV_EVENT(DBF_INFO, device, in dasd_3990_erp_block_queue() 76 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_block_queue() 77 dasd_device_set_stop_bits(device, DASD_STOPPED_PENDING); in dasd_3990_erp_block_queue() 78 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_block_queue() 83 dasd_device_set_timer(device, expires); in dasd_3990_erp_block_queue() 102 struct dasd_device *device = erp->startdev; in dasd_3990_erp_int_req() local 115 dev_err(&device->cdev->dev, in dasd_3990_erp_int_req() 143 struct dasd_device *device = erp->startdev; in dasd_3990_erp_alternate_path() local 148 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_alternate_path() [all …]
|
| /linux/drivers/base/power/ |
| H A D | power.h | 4 static inline void device_pm_init_common(struct device *dev) in device_pm_init_common() 15 static inline void pm_runtime_early_init(struct device *dev) in pm_runtime_early_init() 21 extern void pm_runtime_init(struct device *dev); 22 extern void pm_runtime_reinit(struct device *dev); 23 extern void pm_runtime_remove(struct device *dev); 24 extern u64 pm_runtime_active_time(struct device *dev); 35 struct device *dev; 43 extern void dev_pm_enable_wake_irq_check(struct device *dev, 45 extern void dev_pm_disable_wake_irq_check(struct device *dev, bool cond_disable); 46 extern void dev_pm_enable_wake_irq_complete(struct device *dev); [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| H A D | vga.c | 27 nvkm_rdport(struct nvkm_device *device, int head, u16 port) in nvkm_rdport() argument 29 if (device->card_type >= NV_50) in nvkm_rdport() 30 return nvkm_rd08(device, 0x601000 + port); in nvkm_rdport() 35 return nvkm_rd08(device, 0x601000 + (head * 0x2000) + port); in nvkm_rdport() 40 if (device->card_type < NV_40) in nvkm_rdport() 42 return nvkm_rd08(device, 0x0c0000 + (head * 0x2000) + port); in nvkm_rdport() 49 nvkm_wrport(struct nvkm_device *device, int head, u16 port, u8 data) in nvkm_wrport() argument 51 if (device->card_type >= NV_50) in nvkm_wrport() 52 nvkm_wr08(device, 0x601000 + port, data); in nvkm_wrport() 57 nvkm_wr08(device, 0x601000 + (head * 0x2000) + port, data); in nvkm_wrport() [all …]
|
| H A D | gf119.c | 39 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_device_entry() local 42 nvkm_mask(device, 0x616548 + hoff, 0x00000070, head << 4); in gf119_sor_hda_device_entry() 48 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_eld() local 53 nvkm_wr32(device, 0x10ec00 + soff, (i << 8) | data[i]); in gf119_sor_hda_eld() 55 nvkm_wr32(device, 0x10ec00 + soff, (i << 8)); in gf119_sor_hda_eld() 56 nvkm_mask(device, 0x10ec10 + soff, 0x80000002, 0x80000002); in gf119_sor_hda_eld() 62 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_hpd() local 74 nvkm_mask(device, 0x10ec10 + soff, mask, data); in gf119_sor_hda_hpd() 87 struct nvkm_device *device = sor->disp->engine.subdev.device; in gf119_sor_dp_watermark() local 90 nvkm_mask(device, 0x616610 + hoff, 0x0800003f, 0x08000000 | watermark); in gf119_sor_dp_watermark() [all …]
|
| H A D | gv100.c | 40 struct nvkm_device *device = ior->disp->engine.subdev.device; in gv100_sor_hda_device_entry() local 43 nvkm_mask(device, 0x616528 + hoff, 0x00000070, head << 4); in gv100_sor_hda_device_entry() 56 struct nvkm_device *device = sor->disp->engine.subdev.device; in gv100_sor_dp_watermark() local 59 nvkm_mask(device, 0x616550 + hoff, 0x0c00003f, 0x08000000 | watermark); in gv100_sor_dp_watermark() 65 struct nvkm_device *device = sor->disp->engine.subdev.device; in gv100_sor_dp_audio_sym() local 68 nvkm_mask(device, 0x616568 + hoff, 0x0000ffff, h); in gv100_sor_dp_audio_sym() 69 nvkm_mask(device, 0x61656c + hoff, 0x00ffffff, v); in gv100_sor_dp_audio_sym() 75 struct nvkm_device *device = sor->disp->engine.subdev.device; in gv100_sor_dp_audio() local 80 nvkm_mask(device, 0x616560 + hoff, mask, data); in gv100_sor_dp_audio() 81 nvkm_msec(device, 2000, in gv100_sor_dp_audio() [all …]
|
| /linux/drivers/firewire/ |
| H A D | core-device.c | 191 static const struct ieee1394_device_id *unit_match(struct device *dev, in unit_match() 207 static bool is_fw_unit(const struct device *dev); 209 static int fw_unit_match(struct device *dev, const struct device_driver *drv) in fw_unit_match() 215 static int fw_unit_probe(struct device *dev) in fw_unit_probe() 223 static void fw_unit_remove(struct device *dev) in fw_unit_remove() 242 static int fw_unit_uevent(const struct device *dev, struct kobj_uevent_env *env) in fw_unit_uevent() 263 int fw_device_enable_phys_dma(struct fw_device *device) in fw_device_enable_phys_dma() argument 265 int generation = device->generation; in fw_device_enable_phys_dma() 270 return device->card->driver->enable_phys_dma(device->card, in fw_device_enable_phys_dma() 271 device->node_id, in fw_device_enable_phys_dma() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
| H A D | gt215.c | 34 struct nvkm_device *device = subdev->device; in gt215_pmu_send() local 39 addr = nvkm_rd32(device, 0x10a4a0); in gt215_pmu_send() 40 if (nvkm_msec(device, 2000, in gt215_pmu_send() 41 u32 tmp = nvkm_rd32(device, 0x10a4b0); in gt215_pmu_send() 60 nvkm_wr32(device, 0x10a580, 0x00000001); in gt215_pmu_send() 61 } while (nvkm_rd32(device, 0x10a580) != 0x00000001); in gt215_pmu_send() 64 nvkm_wr32(device, 0x10a1c0, 0x01000000 | (((addr & 0x07) << 4) + in gt215_pmu_send() 66 nvkm_wr32(device, 0x10a1c4, process); in gt215_pmu_send() 67 nvkm_wr32(device, 0x10a1c4, message); in gt215_pmu_send() 68 nvkm_wr32(device, 0x10a1c4, data0); in gt215_pmu_send() [all …]
|
| H A D | gk104.c | 34 magic_(struct nvkm_device *device, u32 ctrl, int size) in magic_() argument 36 nvkm_wr32(device, 0x00c800, 0x00000000); in magic_() 37 nvkm_wr32(device, 0x00c808, 0x00000000); in magic_() 38 nvkm_wr32(device, 0x00c800, ctrl); in magic_() 39 nvkm_msec(device, 2000, in magic_() 40 if (nvkm_rd32(device, 0x00c800) & 0x40000000) { in magic_() 42 nvkm_wr32(device, 0x00c804, 0x00000000); in magic_() 46 nvkm_wr32(device, 0x00c800, 0x00000000); in magic_() 50 magic(struct nvkm_device *device, u32 ctrl) in magic() argument 52 magic_(device, 0x8000a41f | ctrl, 6); in magic() [all …]
|
| /linux/drivers/hid/ |
| H A D | hid-roccat.c | 45 struct device *dev; 62 struct roccat_device *device; member 77 struct roccat_device *device = reader->device; in roccat_read() local 82 mutex_lock(&device->cbuf_lock); in roccat_read() 85 if (reader->cbuf_start == device->cbuf_end) { in roccat_read() 86 add_wait_queue(&device->wait, &wait); in roccat_read() 90 while (reader->cbuf_start == device->cbuf_end) { in roccat_read() 99 if (!device->exist) { in roccat_read() 104 mutex_unlock(&device->cbuf_lock); in roccat_read() 106 mutex_lock(&device->cbuf_lock); in roccat_read() [all …]
|
| /linux/tools/testing/selftests/vfio/lib/ |
| H A D | vfio_pci_device.c | 30 static void vfio_pci_irq_set(struct vfio_pci_device *device, in vfio_pci_irq_set() argument 52 ioctl_assert(device->fd, VFIO_DEVICE_SET_IRQS, irq); 55 void vfio_pci_irq_trigger(struct vfio_pci_device *device, u32 index, u32 vector) in vfio_pci_irq_trigger() 65 ioctl_assert(device->fd, VFIO_DEVICE_SET_IRQS, &irq); 76 void vfio_pci_irq_enable(struct vfio_pci_device *device, u32 index, u32 vector, in vfio_pci_irq_enable() 84 VFIO_ASSERT_LT(device->msi_eventfds[i], 0); in vfio_pci_irq_enable() 85 device->msi_eventfds[i] = eventfd(0, 0); in vfio_pci_irq_enable() 86 VFIO_ASSERT_GE(device->msi_eventfds[i], 0); in vfio_pci_irq_enable() 89 vfio_pci_irq_set(device, index, vector, count, device in vfio_pci_irq_enable() 53 vfio_pci_irq_trigger(struct vfio_pci_device * device,u32 index,u32 vector) vfio_pci_irq_trigger() argument 74 vfio_pci_irq_enable(struct vfio_pci_device * device,u32 index,u32 vector,int count) vfio_pci_irq_enable() argument 90 vfio_pci_irq_disable(struct vfio_pci_device * device,u32 index) vfio_pci_irq_disable() argument 107 vfio_pci_irq_get(struct vfio_pci_device * device,u32 index,struct vfio_irq_info * irq_info) vfio_pci_irq_get() argument 116 vfio_pci_region_get(struct vfio_pci_device * device,int index,struct vfio_region_info * info) vfio_pci_region_get() argument 127 vfio_pci_bar_map(struct vfio_pci_device * device,int index) vfio_pci_bar_map() argument 164 vfio_pci_bar_unmap(struct vfio_pci_device * device,int index) vfio_pci_bar_unmap() argument 175 vfio_pci_bar_unmap_all(struct vfio_pci_device * device) vfio_pci_bar_unmap_all() argument 185 vfio_pci_config_access(struct vfio_pci_device * device,bool write,size_t config,size_t size,void * data) vfio_pci_config_access() argument 200 vfio_pci_device_reset(struct vfio_pci_device * device) vfio_pci_device_reset() argument 223 vfio_pci_group_setup(struct vfio_pci_device * device,const char * bdf) vfio_pci_group_setup() argument 243 vfio_pci_container_setup(struct vfio_pci_device * device,const char * bdf) vfio_pci_container_setup() argument 265 vfio_pci_device_setup(struct vfio_pci_device * device) vfio_pci_device_setup() argument 343 vfio_pci_iommufd_setup(struct vfio_pci_device * device,const char * bdf) vfio_pci_iommufd_setup() argument 357 struct vfio_pci_device *device; vfio_pci_device_init() local 377 vfio_pci_device_cleanup(struct vfio_pci_device * device) vfio_pci_device_cleanup() argument [all...] |
| /linux/drivers/vfio/ |
| H A D | vfio_main.c | 15 #include <linux/device.h> 62 MODULE_PARM_DESC(enable_unsafe_noiommu_mode, "Enable UNSAFE, no-IOMMU mode. This mode provides no device isolation, no DMA translation, no host kernel protection, cannot be used for device assignment to virtual machines, requires RAWIO permissions, and will taint the kernel. If you do not know what this is for, step away. (default: false)"); 67 int vfio_assign_device_set(struct vfio_device *device, void *set_id) in vfio_assign_device_set() argument 110 device->dev_set = dev_set; in vfio_assign_device_set() 111 list_add_tail(&device->dev_set_list, &dev_set->device_list); in vfio_assign_device_set() 117 static void vfio_release_device_set(struct vfio_device *device) in vfio_release_device_set() argument 119 struct vfio_device_set *dev_set = device->dev_set; in vfio_release_device_set() 125 list_del(&device->dev_set_list); in vfio_release_device_set() 153 struct device *de in vfio_find_device_in_devset() 170 vfio_device_put_registration(struct vfio_device * device) vfio_device_put_registration() argument 177 vfio_device_try_get_registration(struct vfio_device * device) vfio_device_try_get_registration() argument 189 struct vfio_device *device = vfio_device_release() local 221 struct vfio_device *device; _vfio_alloc_device() local 273 vfio_init_device(struct vfio_device * device,struct device * dev,const struct vfio_device_ops * ops) vfio_init_device() argument 315 __vfio_register_dev(struct vfio_device * device,enum vfio_group_type type) __vfio_register_dev() argument 369 vfio_register_group_dev(struct vfio_device * device) vfio_register_group_dev() argument 379 vfio_register_emulated_iommu_dev(struct vfio_device * device) vfio_register_emulated_iommu_dev() argument 388 vfio_unregister_group_dev(struct vfio_device * device) vfio_unregister_group_dev() argument 436 vfio_device_get_kvm_safe(struct vfio_device * device,struct kvm * kvm) vfio_device_get_kvm_safe() argument 468 vfio_device_put_kvm(struct vfio_device * device) vfio_device_put_kvm() argument 488 vfio_assert_device_open(struct vfio_device * device) vfio_assert_device_open() argument 494 vfio_allocate_device_file(struct vfio_device * device) vfio_allocate_device_file() argument 510 struct vfio_device *device = df->device; vfio_df_device_first_open() local 545 struct vfio_device *device = df->device; vfio_df_device_last_close() local 561 struct vfio_device *device = df->device; vfio_df_open() local 585 struct vfio_device *device = df->device; vfio_df_close() local 600 vfio_device_pm_runtime_get(struct vfio_device * device) vfio_device_pm_runtime_get() argument 621 vfio_device_pm_runtime_put(struct vfio_device * device) vfio_device_pm_runtime_put() argument 635 struct vfio_device *device = df->device; vfio_device_fops_release() local 664 vfio_mig_get_next_state(struct vfio_device * device,enum vfio_device_mig_state cur_fsm,enum vfio_device_mig_state new_fsm,enum vfio_device_mig_state * next_fsm) vfio_mig_get_next_state() argument 890 vfio_ioctl_device_feature_mig_device_state(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_mig_device_state() argument 940 vfio_ioctl_device_feature_migration_data_size(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_migration_data_size() argument 967 vfio_ioctl_device_feature_migration(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_migration() argument 1050 vfio_ioctl_device_feature_logging_start(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_logging_start() argument 1134 vfio_ioctl_device_feature_logging_stop(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_logging_stop() argument 1155 struct vfio_device *device = opaque; vfio_device_log_read_and_clear() local 1161 vfio_ioctl_device_feature_logging_report(struct vfio_device * device,u32 flags,void __user * arg,size_t argsz) vfio_ioctl_device_feature_logging_report() argument 1205 vfio_ioctl_device_feature(struct vfio_device * device,struct vfio_device_feature __user * arg) vfio_ioctl_device_feature() argument 1263 vfio_get_region_info(struct vfio_device * device,struct vfio_region_info __user * arg) vfio_get_region_info() argument 1312 struct vfio_device *device = df->device; vfio_device_fops_unl_ioctl() local 1365 struct vfio_device *device = df->device; vfio_device_fops_read() local 1382 struct vfio_device *device = df->device; vfio_device_fops_write() local 1397 struct vfio_device *device = df->device; vfio_device_fops_mmap() local 1414 struct vfio_device *device = df->device; vfio_device_show_fdinfo() local 1470 struct vfio_device *device; vfio_file_enforced_coherent() local 1660 vfio_pin_pages(struct vfio_device * device,dma_addr_t iova,int npage,int prot,struct page ** pages) vfio_pin_pages() argument 1701 vfio_unpin_pages(struct vfio_device * device,dma_addr_t iova,int npage) vfio_unpin_pages() argument 1740 vfio_dma_rw(struct vfio_device * device,dma_addr_t iova,void * data,size_t len,bool write) vfio_dma_rw() argument [all...] |
| /linux/tools/testing/selftests/vfio/lib/drivers/ioat/ |
| H A D | ioat.c | 26 static inline struct ioat_state *to_ioat_state(struct vfio_pci_device *device) in to_ioat_state() argument 28 return device->driver.region.vaddr; in to_ioat_state() 31 static inline void *ioat_channel_registers(struct vfio_pci_device *device) in ioat_channel_registers() argument 33 return device->bars[0].vaddr + IOAT_CHANNEL_MMIO_SIZE; in ioat_channel_registers() 36 static int ioat_probe(struct vfio_pci_device *device) in ioat_probe() argument 41 if (!vfio_pci_device_match(device, PCI_VENDOR_ID_INTEL, in ioat_probe() 45 VFIO_ASSERT_NOT_NULL(device->bars[0].vaddr); in ioat_probe() 47 version = readb(device->bars[0].vaddr + IOAT_VER_OFFSET); in ioat_probe() 54 dev_err(device, "ioat: Unsupported version: 0x%x\n", version); in ioat_probe() 65 static void ioat_clear_errors(struct vfio_pci_device *device) in ioat_clear_errors() argument [all …]
|
| /linux/tools/testing/selftests/vfio/lib/drivers/dsa/ |
| H A D | dsa.c | 28 /* Cached device registers (and derived data) for easy access */ 47 static inline struct dsa_state *to_dsa_state(struct vfio_pci_device *device) in to_dsa_state() argument 49 return device->driver.region.vaddr; in to_dsa_state() 52 static bool dsa_int_handle_request_required(struct vfio_pci_device *device) in dsa_int_handle_request_required() argument 54 void *bar0 = device->bars[0].vaddr; in dsa_int_handle_request_required() 66 static int dsa_probe(struct vfio_pci_device *device) in dsa_probe() argument 68 const u16 vendor_id = vfio_pci_config_readw(device, PCI_VENDOR_ID); in dsa_probe() 69 const u16 device_id = vfio_pci_config_readw(device, PCI_DEVICE_ID); in dsa_probe() 83 if (dsa_int_handle_request_required(device)) { in dsa_check_sw_err() 84 dev_err(device, "Devic in dsa_check_sw_err() 80 dsa_check_sw_err(struct vfio_pci_device * device) dsa_check_sw_err() argument 115 dsa_command(struct vfio_pci_device * device,u32 cmd) dsa_command() argument 140 dsa_wq_init(struct vfio_pci_device * device) dsa_wq_init() argument 170 dsa_group_init(struct vfio_pci_device * device) dsa_group_init() argument 184 dsa_register_cache_init(struct vfio_pci_device * device) dsa_register_cache_init() argument 207 dsa_init(struct vfio_pci_device * device) dsa_init() argument 235 dsa_remove(struct vfio_pci_device * device) dsa_remove() argument 241 dsa_completion_wait(struct vfio_pci_device * device,struct dsa_completion_record * completion) dsa_completion_wait() argument 263 dsa_copy_desc_init(struct vfio_pci_device * device,struct dsa_hw_desc * desc,iova_t src,iova_t dst,u64 size,bool interrupt) dsa_copy_desc_init() argument 288 dsa_batch_desc_init(struct vfio_pci_device * device,struct dsa_hw_desc * desc,u64 count) dsa_batch_desc_init() argument 304 dsa_desc_write(struct vfio_pci_device * device,struct dsa_hw_desc * desc) dsa_desc_write() argument 310 dsa_memcpy_one(struct vfio_pci_device * device,iova_t src,iova_t dst,u64 size,bool interrupt) dsa_memcpy_one() argument 321 dsa_memcpy_batch(struct vfio_pci_device * device,iova_t src,iova_t dst,u64 size,u64 count) dsa_memcpy_batch() argument 367 dsa_memcpy_start(struct vfio_pci_device * device,iova_t src,iova_t dst,u64 size,u64 count) dsa_memcpy_start() argument 381 dsa_memcpy_wait(struct vfio_pci_device * device) dsa_memcpy_wait() argument 396 dsa_send_msi(struct vfio_pci_device * device) dsa_send_msi() argument [all...] |