| /linux/drivers/gpu/drm/mediatek/ |
| H A D | mtk_disp_drv.h | 15 int mtk_aal_clk_enable(struct device *dev); 16 void mtk_aal_clk_disable(struct device *dev); 17 void mtk_aal_config(struct device *dev, unsigned int w, 20 unsigned int mtk_aal_gamma_get_lut_size(struct device *dev); 21 void mtk_aal_gamma_set(struct device *dev, struct drm_crtc_state *state); 22 void mtk_aal_start(struct device *dev); 23 void mtk_aal_stop(struct device *dev); 25 void mtk_ccorr_ctm_set(struct device *dev, struct drm_crtc_state *state); 26 int mtk_ccorr_clk_enable(struct device *dev); 27 void mtk_ccorr_clk_disable(struct device *dev); [all …]
|
| /linux/drivers/block/drbd/ |
| H A D | drbd_actlog.c | 82 void *drbd_md_get_buffer(struct drbd_device *device, const char *intent) in drbd_md_get_buffer() argument 86 wait_event(device->misc_wait, in drbd_md_get_buffer() 87 (r = atomic_cmpxchg(&device->md_io.in_use, 0, 1)) == 0 || in drbd_md_get_buffer() 88 device->state.disk <= D_FAILED); in drbd_md_get_buffer() 93 device->md_io.current_use = intent; in drbd_md_get_buffer() 94 device->md_io.start_jif = jiffies; in drbd_md_get_buffer() 95 device->md_io.submit_jif = device->md_io.start_jif - 1; in drbd_md_get_buffer() 96 return page_address(device->md_io.page); in drbd_md_get_buffer() 99 void drbd_md_put_buffer(struct drbd_device *device) in drbd_md_put_buffer() argument 101 if (atomic_dec_and_test(&device->md_io.in_use)) in drbd_md_put_buffer() [all …]
|
| H A D | drbd_worker.c | 52 struct drbd_device *device; in drbd_md_endio() local 54 device = bio->bi_private; in drbd_md_endio() 55 device->md_io.error = blk_status_to_errno(bio->bi_status); in drbd_md_endio() 58 if (device->ldev) in drbd_md_endio() 59 put_ldev(device); in drbd_md_endio() 73 drbd_md_put_buffer(device); in drbd_md_endio() 74 device->md_io.done = 1; in drbd_md_endio() 75 wake_up(&device->misc_wait); in drbd_md_endio() 85 struct drbd_device *device = peer_device->device; in drbd_endio_read_sec_final() local 87 spin_lock_irqsave(&device->resource->req_lock, flags); in drbd_endio_read_sec_final() [all …]
|
| /linux/drivers/s390/char/ |
| H A D | tape_core.c | 93 tape_medium_state_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_medium_state_show() 105 tape_first_minor_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_first_minor_show() 117 tape_state_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_state_show() 130 tape_operation_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_operation_show() 157 tape_blocksize_show(struct device *dev, struct device_attribute *attr, char *buf) in tape_blocksize_show() 186 tape_state_set(struct tape_device *device, enum tape_state newstate) in tape_state_set() argument 190 if (device->tape_state == TS_NOT_OPER) { in tape_state_set() 194 DBF_EVENT(4, "ts. dev: %x\n", device->first_minor); in tape_state_set() 196 if (device->tape_state < TS_SIZE && device->tape_state >=0 ) in tape_state_set() 197 str = tape_state_verbose[device->tape_state]; in tape_state_set() [all …]
|
| /linux/include/linux/ |
| H A D | pm_runtime.h | 68 extern int pm_generic_runtime_suspend(struct device *dev); 69 extern int pm_generic_runtime_resume(struct device *dev); 70 extern int pm_runtime_force_suspend(struct device *dev); 72 extern int __pm_runtime_idle(struct device *dev, int rpmflags); 73 extern int __pm_runtime_suspend(struct device *dev, int rpmflags); 74 extern int __pm_runtime_resume(struct device *dev, int rpmflags); 75 extern int pm_runtime_get_if_active(struct device *dev); 76 extern int pm_runtime_get_if_in_use(struct device *dev); 77 extern int pm_schedule_suspend(struct device *dev, unsigned int delay); 78 extern int __pm_runtime_set_status(struct device *dev, unsigned int status); [all …]
|
| H A D | device.h | 36 struct device; 67 int (*add_dev)(struct device *dev, struct subsys_interface *sif); 68 void (*remove_dev)(struct device *dev, struct subsys_interface *sif); 91 int (*uevent)(const struct device *dev, struct kobj_uevent_env *env); 92 char *(*devnode)(const struct device *dev, umode_t *mode, 94 void (*release)(struct device *dev); 107 ssize_t (*show)(struct device *dev, struct device_attribute *attr, 109 ssize_t (*store)(struct device *dev, struct device_attribute *attr, 123 ssize_t device_show_ulong(struct device *dev, struct device_attribute *attr, 125 ssize_t device_store_ulong(struct device *dev, struct device_attribute *attr, [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
| H A D | nv40.c | 36 return nvkm_rd32(gr->engine.subdev.device, 0x1540); in nv40_gr_units() 47 int ret = nvkm_gpuobj_new(object->engine->subdev.device, 20, align, in nv40_gr_object_bind() 79 int ret = nvkm_gpuobj_new(gr->base.engine.subdev.device, gr->size, in nv40_gr_chan_bind() 84 nv40_grctx_fill(gr->base.engine.subdev.device, *pgpuobj); in nv40_gr_chan_bind() 97 struct nvkm_device *device = subdev->device; in nv40_gr_chan_fini() local 101 nvkm_mask(device, 0x400720, 0x00000001, 0x00000000); in nv40_gr_chan_fini() 103 if (nvkm_rd32(device, 0x40032c) == inst) { in nv40_gr_chan_fini() 105 nvkm_wr32(device, 0x400720, 0x00000000); in nv40_gr_chan_fini() 106 nvkm_wr32(device, 0x400784, inst); in nv40_gr_chan_fini() 107 nvkm_mask(device, 0x400310, 0x00000020, 0x00000020); in nv40_gr_chan_fini() [all …]
|
| H A D | nv50.c | 35 return nvkm_rd32(gr->engine.subdev.device, 0x1540); in nv50_gr_units() 46 int ret = nvkm_gpuobj_new(object->engine->subdev.device, 16, in nv50_gr_object_bind() 73 int ret = nvkm_gpuobj_new(gr->base.engine.subdev.device, gr->size, in nv50_gr_chan_bind() 77 nv50_grctx_fill(gr->base.engine.subdev.device, *pgpuobj); in nv50_gr_chan_bind() 243 struct nvkm_device *device = subdev->device; in nv50_gr_prop_trap() local 244 u32 e0c = nvkm_rd32(device, ustatus_addr + 0x04); in nv50_gr_prop_trap() 245 u32 e10 = nvkm_rd32(device, ustatus_addr + 0x08); in nv50_gr_prop_trap() 246 u32 e14 = nvkm_rd32(device, ustatus_addr + 0x0c); in nv50_gr_prop_trap() 247 u32 e18 = nvkm_rd32(device, ustatus_addr + 0x10); in nv50_gr_prop_trap() 248 u32 e1c = nvkm_rd32(device, ustatus_addr + 0x14); in nv50_gr_prop_trap() [all …]
|
| H A D | nv20.c | 34 struct nvkm_device *device = gr->base.engine.subdev.device; in nv20_gr_chan_fini() local 38 nvkm_mask(device, 0x400720, 0x00000001, 0x00000000); in nv20_gr_chan_fini() 39 if (nvkm_rd32(device, 0x400144) & 0x00010000) in nv20_gr_chan_fini() 40 chid = (nvkm_rd32(device, 0x400148) & 0x1f000000) >> 24; in nv20_gr_chan_fini() 42 nvkm_wr32(device, 0x400784, inst >> 4); in nv20_gr_chan_fini() 43 nvkm_wr32(device, 0x400788, 0x00000002); in nv20_gr_chan_fini() 44 nvkm_msec(device, 2000, in nv20_gr_chan_fini() 45 if (!nvkm_rd32(device, 0x400700)) in nv20_gr_chan_fini() 48 nvkm_wr32(device, 0x400144, 0x10000000); in nv20_gr_chan_fini() 49 nvkm_mask(device, 0x400148, 0xff000000, 0x1f000000); in nv20_gr_chan_fini() [all …]
|
| H A D | nv04.c | 445 nv04_gr_set_ctx1(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx1() argument 447 int subc = (nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR) >> 13) & 0x7; in nv04_gr_set_ctx1() 450 tmp = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx1() 453 nvkm_wr32(device, 0x700000 + inst, tmp); in nv04_gr_set_ctx1() 455 nvkm_wr32(device, NV04_PGRAPH_CTX_SWITCH1, tmp); in nv04_gr_set_ctx1() 456 nvkm_wr32(device, NV04_PGRAPH_CTX_CACHE1 + (subc << 2), tmp); in nv04_gr_set_ctx1() 460 nv04_gr_set_ctx_val(struct nvkm_device *device, u32 inst, u32 mask, u32 value) in nv04_gr_set_ctx_val() argument 465 ctx1 = nvkm_rd32(device, 0x700000 + inst); in nv04_gr_set_ctx_val() 469 tmp = nvkm_rd32(device, 0x70000c + inst); in nv04_gr_set_ctx_val() 472 nvkm_wr32(device, 0x70000c + inst, tmp); in nv04_gr_set_ctx_val() [all …]
|
| /linux/drivers/base/power/ |
| H A D | power.h | 4 static inline void device_pm_init_common(struct device *dev) in device_pm_init_common() 15 static inline void pm_runtime_early_init(struct device *dev) in pm_runtime_early_init() 21 extern void pm_runtime_init(struct device *dev); 22 extern void pm_runtime_reinit(struct device *dev); 23 extern void pm_runtime_remove(struct device *dev); 24 extern u64 pm_runtime_active_time(struct device *dev); 35 struct device *dev; 43 extern void dev_pm_enable_wake_irq_check(struct device *dev, 45 extern void dev_pm_disable_wake_irq_check(struct device *dev, bool cond_disable); 46 extern void dev_pm_enable_wake_irq_complete(struct device *dev); [all …]
|
| /linux/drivers/s390/block/ |
| H A D | dasd.c | 87 struct dasd_device *device; in dasd_alloc_device() local 89 device = kzalloc_obj(struct dasd_device, GFP_ATOMIC); in dasd_alloc_device() 90 if (!device) in dasd_alloc_device() 94 device->ccw_mem = (void *) __get_free_pages(GFP_ATOMIC | GFP_DMA, 1); in dasd_alloc_device() 95 if (!device->ccw_mem) { in dasd_alloc_device() 96 kfree(device); in dasd_alloc_device() 100 device->erp_mem = (void *) get_zeroed_page(GFP_ATOMIC | GFP_DMA); in dasd_alloc_device() 101 if (!device->erp_mem) { in dasd_alloc_device() 102 free_pages((unsigned long) device->ccw_mem, 1); in dasd_alloc_device() 103 kfree(device); in dasd_alloc_device() [all …]
|
| H A D | dasd_devmap.c | 49 struct dasd_device *device; member 431 new->device = NULL; in dasd_add_busid() 494 BUG_ON(devmap->device != NULL); in dasd_forget_ranges() 509 struct dasd_device *device; in dasd_device_from_devindex() local 521 if (devmap && devmap->device) { in dasd_device_from_devindex() 522 device = devmap->device; in dasd_device_from_devindex() 523 dasd_get_device(device); in dasd_device_from_devindex() 525 device = ERR_PTR(-ENODEV); in dasd_device_from_devindex() 527 return device; in dasd_device_from_devindex() 553 struct dasd_device *device; in dasd_create_device() local [all …]
|
| H A D | dasd_3990_erp.c | 70 struct dasd_device *device = erp->startdev; in dasd_3990_erp_block_queue() local 73 DBF_DEV_EVENT(DBF_INFO, device, in dasd_3990_erp_block_queue() 76 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_block_queue() 77 dasd_device_set_stop_bits(device, DASD_STOPPED_PENDING); in dasd_3990_erp_block_queue() 78 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_block_queue() 83 dasd_device_set_timer(device, expires); in dasd_3990_erp_block_queue() 102 struct dasd_device *device = erp->startdev; in dasd_3990_erp_int_req() local 115 dev_err(&device->cdev->dev, in dasd_3990_erp_int_req() 143 struct dasd_device *device = erp->startdev; in dasd_3990_erp_alternate_path() local 148 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags); in dasd_3990_erp_alternate_path() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| H A D | vga.c | 27 nvkm_rdport(struct nvkm_device *device, int head, u16 port) in nvkm_rdport() argument 29 if (device->card_type >= NV_50) in nvkm_rdport() 30 return nvkm_rd08(device, 0x601000 + port); in nvkm_rdport() 35 return nvkm_rd08(device, 0x601000 + (head * 0x2000) + port); in nvkm_rdport() 40 if (device->card_type < NV_40) in nvkm_rdport() 42 return nvkm_rd08(device, 0x0c0000 + (head * 0x2000) + port); in nvkm_rdport() 49 nvkm_wrport(struct nvkm_device *device, int head, u16 port, u8 data) in nvkm_wrport() argument 51 if (device->card_type >= NV_50) in nvkm_wrport() 52 nvkm_wr08(device, 0x601000 + port, data); in nvkm_wrport() 57 nvkm_wr08(device, 0x601000 + (head * 0x2000) + port, data); in nvkm_wrport() [all …]
|
| H A D | gf119.c | 39 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_device_entry() local 42 nvkm_mask(device, 0x616548 + hoff, 0x00000070, head << 4); in gf119_sor_hda_device_entry() 48 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_eld() local 53 nvkm_wr32(device, 0x10ec00 + soff, (i << 8) | data[i]); in gf119_sor_hda_eld() 55 nvkm_wr32(device, 0x10ec00 + soff, (i << 8)); in gf119_sor_hda_eld() 56 nvkm_mask(device, 0x10ec10 + soff, 0x80000002, 0x80000002); in gf119_sor_hda_eld() 62 struct nvkm_device *device = ior->disp->engine.subdev.device; in gf119_sor_hda_hpd() local 74 nvkm_mask(device, 0x10ec10 + soff, mask, data); in gf119_sor_hda_hpd() 87 struct nvkm_device *device = sor->disp->engine.subdev.device; in gf119_sor_dp_watermark() local 90 nvkm_mask(device, 0x616610 + hoff, 0x0800003f, 0x08000000 | watermark); in gf119_sor_dp_watermark() [all …]
|
| H A D | gv100.c | 40 struct nvkm_device *device = ior->disp->engine.subdev.device; in gv100_sor_hda_device_entry() local 43 nvkm_mask(device, 0x616528 + hoff, 0x00000070, head << 4); in gv100_sor_hda_device_entry() 56 struct nvkm_device *device = sor->disp->engine.subdev.device; in gv100_sor_dp_watermark() local 59 nvkm_mask(device, 0x616550 + hoff, 0x0c00003f, 0x08000000 | watermark); in gv100_sor_dp_watermark() 65 struct nvkm_device *device = sor->disp->engine.subdev.device; in gv100_sor_dp_audio_sym() local 68 nvkm_mask(device, 0x616568 + hoff, 0x0000ffff, h); in gv100_sor_dp_audio_sym() 69 nvkm_mask(device, 0x61656c + hoff, 0x00ffffff, v); in gv100_sor_dp_audio_sym() 75 struct nvkm_device *device = sor->disp->engine.subdev.device; in gv100_sor_dp_audio() local 80 nvkm_mask(device, 0x616560 + hoff, mask, data); in gv100_sor_dp_audio() 81 nvkm_msec(device, 2000, in gv100_sor_dp_audio() [all …]
|
| /linux/drivers/firewire/ |
| H A D | core-device.c | 191 static const struct ieee1394_device_id *unit_match(struct device *dev, in unit_match() 207 static bool is_fw_unit(const struct device *dev); 209 static int fw_unit_match(struct device *dev, const struct device_driver *drv) in fw_unit_match() 215 static int fw_unit_probe(struct device *dev) in fw_unit_probe() 223 static void fw_unit_remove(struct device *dev) in fw_unit_remove() 242 static int fw_unit_uevent(const struct device *dev, struct kobj_uevent_env *env) in fw_unit_uevent() 263 int fw_device_enable_phys_dma(struct fw_device *device) in fw_device_enable_phys_dma() argument 265 int generation = device->generation; in fw_device_enable_phys_dma() 270 return device->card->driver->enable_phys_dma(device->card, in fw_device_enable_phys_dma() 271 device->node_id, in fw_device_enable_phys_dma() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
| H A D | gt215.c | 34 struct nvkm_device *device = subdev->device; in gt215_pmu_send() local 39 addr = nvkm_rd32(device, 0x10a4a0); in gt215_pmu_send() 40 if (nvkm_msec(device, 2000, in gt215_pmu_send() 41 u32 tmp = nvkm_rd32(device, 0x10a4b0); in gt215_pmu_send() 60 nvkm_wr32(device, 0x10a580, 0x00000001); in gt215_pmu_send() 61 } while (nvkm_rd32(device, 0x10a580) != 0x00000001); in gt215_pmu_send() 64 nvkm_wr32(device, 0x10a1c0, 0x01000000 | (((addr & 0x07) << 4) + in gt215_pmu_send() 66 nvkm_wr32(device, 0x10a1c4, process); in gt215_pmu_send() 67 nvkm_wr32(device, 0x10a1c4, message); in gt215_pmu_send() 68 nvkm_wr32(device, 0x10a1c4, data0); in gt215_pmu_send() [all …]
|
| H A D | gk104.c | 34 magic_(struct nvkm_device *device, u32 ctrl, int size) in magic_() argument 36 nvkm_wr32(device, 0x00c800, 0x00000000); in magic_() 37 nvkm_wr32(device, 0x00c808, 0x00000000); in magic_() 38 nvkm_wr32(device, 0x00c800, ctrl); in magic_() 39 nvkm_msec(device, 2000, in magic_() 40 if (nvkm_rd32(device, 0x00c800) & 0x40000000) { in magic_() 42 nvkm_wr32(device, 0x00c804, 0x00000000); in magic_() 46 nvkm_wr32(device, 0x00c800, 0x00000000); in magic_() 50 magic(struct nvkm_device *device, u32 ctrl) in magic() argument 52 magic_(device, 0x8000a41f | ctrl, 6); in magic() [all …]
|
| /linux/drivers/hid/ |
| H A D | hid-roccat.c | 45 struct device *dev; 62 struct roccat_device *device; member 77 struct roccat_device *device = reader->device; in roccat_read() local 82 mutex_lock(&device->cbuf_lock); in roccat_read() 85 if (reader->cbuf_start == device->cbuf_end) { in roccat_read() 86 add_wait_queue(&device->wait, &wait); in roccat_read() 90 while (reader->cbuf_start == device->cbuf_end) { in roccat_read() 99 if (!device->exist) { in roccat_read() 104 mutex_unlock(&device->cbuf_lock); in roccat_read() 106 mutex_lock(&device->cbuf_lock); in roccat_read() [all …]
|
| /linux/tools/testing/selftests/vfio/lib/ |
| H A D | vfio_pci_device.c | 30 static void vfio_pci_irq_set(struct vfio_pci_device *device, in vfio_pci_irq_set() argument 50 ioctl_assert(device->fd, VFIO_DEVICE_SET_IRQS, irq); in vfio_pci_irq_set() 53 void vfio_pci_irq_trigger(struct vfio_pci_device *device, u32 index, u32 vector) in vfio_pci_irq_trigger() argument 63 ioctl_assert(device->fd, VFIO_DEVICE_SET_IRQS, &irq); in vfio_pci_irq_trigger() 74 void vfio_pci_irq_enable(struct vfio_pci_device *device, u32 index, u32 vector, in vfio_pci_irq_enable() argument 82 VFIO_ASSERT_LT(device->msi_eventfds[i], 0); in vfio_pci_irq_enable() 83 device->msi_eventfds[i] = eventfd(0, 0); in vfio_pci_irq_enable() 84 VFIO_ASSERT_GE(device->msi_eventfds[i], 0); in vfio_pci_irq_enable() 87 vfio_pci_irq_set(device, index, vector, count, device->msi_eventfds + vector); in vfio_pci_irq_enable() 90 void vfio_pci_irq_disable(struct vfio_pci_device *device, u32 index) in vfio_pci_irq_disable() argument [all …]
|
| /linux/drivers/vfio/ |
| H A D | vfio_main.c | 67 int vfio_assign_device_set(struct vfio_device *device, void *set_id) in vfio_assign_device_set() argument 110 device->dev_set = dev_set; in vfio_assign_device_set() 111 list_add_tail(&device->dev_set_list, &dev_set->device_list); in vfio_assign_device_set() 117 static void vfio_release_device_set(struct vfio_device *device) in vfio_release_device_set() argument 119 struct vfio_device_set *dev_set = device->dev_set; in vfio_release_device_set() 125 list_del(&device->dev_set_list); in vfio_release_device_set() 153 struct device *dev) in vfio_find_device_in_devset() 170 void vfio_device_put_registration(struct vfio_device *device) in vfio_device_put_registration() argument 172 if (refcount_dec_and_test(&device->refcount)) in vfio_device_put_registration() 173 complete(&device->comp); in vfio_device_put_registration() [all …]
|
| /linux/tools/testing/selftests/vfio/lib/drivers/ioat/ |
| H A D | ioat.c | 26 static inline struct ioat_state *to_ioat_state(struct vfio_pci_device *device) in to_ioat_state() argument 28 return device->driver.region.vaddr; in to_ioat_state() 31 static inline void *ioat_channel_registers(struct vfio_pci_device *device) in ioat_channel_registers() argument 33 return device->bars[0].vaddr + IOAT_CHANNEL_MMIO_SIZE; in ioat_channel_registers() 36 static int ioat_probe(struct vfio_pci_device *device) in ioat_probe() argument 41 if (!vfio_pci_device_match(device, PCI_VENDOR_ID_INTEL, in ioat_probe() 45 VFIO_ASSERT_NOT_NULL(device->bars[0].vaddr); in ioat_probe() 47 version = readb(device->bars[0].vaddr + IOAT_VER_OFFSET); in ioat_probe() 54 dev_err(device, "ioat: Unsupported version: 0x%x\n", version); in ioat_probe() 65 static void ioat_clear_errors(struct vfio_pci_device *device) in ioat_clear_errors() argument [all …]
|
| /linux/tools/testing/selftests/vfio/lib/drivers/dsa/ |
| H A D | dsa.c | 47 static inline struct dsa_state *to_dsa_state(struct vfio_pci_device *device) in to_dsa_state() argument 49 return device->driver.region.vaddr; in to_dsa_state() 52 static bool dsa_int_handle_request_required(struct vfio_pci_device *device) in dsa_int_handle_request_required() argument 54 void *bar0 = device->bars[0].vaddr; in dsa_int_handle_request_required() 66 static int dsa_probe(struct vfio_pci_device *device) in dsa_probe() argument 68 if (!vfio_pci_device_match(device, PCI_VENDOR_ID_INTEL, in dsa_probe() 72 if (dsa_int_handle_request_required(device)) { in dsa_probe() 73 dev_err(device, "Device requires requesting interrupt handles\n"); in dsa_probe() 80 static void dsa_check_sw_err(struct vfio_pci_device *device) in dsa_check_sw_err() argument 82 void *reg = device->bars[0].vaddr + IDXD_SWERR_OFFSET; in dsa_check_sw_err() [all …]
|