Home
last modified time | relevance | path

Searched refs:bdev (Results 1 – 25 of 319) sorted by relevance

12345678910>>...13

/linux/drivers/bluetooth/
H A Dbtmtkuart.c85 #define btmtkuart_is_standalone(bdev) \ argument
86 ((bdev)->data->flags & BTMTKUART_FLAG_STANDALONE_HW)
87 #define btmtkuart_is_builtin_soc(bdev) \ argument
88 !((bdev)->data->flags & BTMTKUART_FLAG_STANDALONE_HW)
93 struct btmtkuart_dev *bdev = hci_get_drvdata(hdev); in mtk_hci_wmt_sync() local
121 set_bit(BTMTKUART_TX_WAIT_VND_EVT, &bdev->tx_state); in mtk_hci_wmt_sync()
125 clear_bit(BTMTKUART_TX_WAIT_VND_EVT, &bdev->tx_state); in mtk_hci_wmt_sync()
138 err = wait_on_bit_timeout(&bdev->tx_state, BTMTKUART_TX_WAIT_VND_EVT, in mtk_hci_wmt_sync()
142 clear_bit(BTMTKUART_TX_WAIT_VND_EVT, &bdev->tx_state); in mtk_hci_wmt_sync()
148 clear_bit(BTMTKUART_TX_WAIT_VND_EVT, &bdev->tx_state); in mtk_hci_wmt_sync()
[all …]
H A Dhci_bcm.c324 struct bcm_device *bdev = data; in bcm_host_wake() local
326 bt_dev_dbg(bdev, "Host wake IRQ"); in bcm_host_wake()
328 pm_runtime_get(bdev->dev); in bcm_host_wake()
329 pm_runtime_put_autosuspend(bdev->dev); in bcm_host_wake()
336 struct bcm_device *bdev = bcm->dev; in bcm_request_irq() local
340 if (!bcm_device_exists(bdev)) { in bcm_request_irq()
345 if (bdev->irq <= 0) { in bcm_request_irq()
350 err = devm_request_irq(bdev->dev, bdev->irq, bcm_host_wake, in bcm_request_irq()
351 bdev->irq_active_low ? IRQF_TRIGGER_FALLING : in bcm_request_irq()
353 "host_wake", bdev); in bcm_request_irq()
[all …]
/linux/drivers/dma/qcom/
H A Dbam_dma.c351 struct bam_device *bdev; member
410 static inline void __iomem *bam_addr(struct bam_device *bdev, u32 pipe, in bam_addr() argument
413 const struct reg_offset_data r = bdev->layout[reg]; in bam_addr()
415 return bdev->regs + r.base_offset + in bam_addr()
418 r.ee_mult * bdev->ee; in bam_addr()
425 static void bam_reset(struct bam_device *bdev) in bam_reset() argument
431 val = readl_relaxed(bam_addr(bdev, 0, BAM_CTRL)); in bam_reset()
433 writel_relaxed(val, bam_addr(bdev, 0, BAM_CTRL)); in bam_reset()
435 writel_relaxed(val, bam_addr(bdev, 0, BAM_CTRL)); in bam_reset()
442 writel_relaxed(val, bam_addr(bdev, 0, BAM_CTRL)); in bam_reset()
[all …]
/linux/block/
H A Dioctl.c21 static int blkpg_do_ioctl(struct block_device *bdev, in blkpg_do_ioctl() argument
24 struct gendisk *disk = bdev->bd_disk; in blkpg_do_ioctl()
32 if (bdev_is_partition(bdev)) in blkpg_do_ioctl()
44 if (!IS_ALIGNED(p.start | p.length, bdev_logical_block_size(bdev))) in blkpg_do_ioctl()
67 static int blkpg_ioctl(struct block_device *bdev, in blkpg_ioctl() argument
76 return blkpg_do_ioctl(bdev, udata, op); in blkpg_ioctl()
87 static int compat_blkpg_ioctl(struct block_device *bdev, in compat_blkpg_ioctl() argument
96 return blkpg_do_ioctl(bdev, compat_ptr(udata), op); in compat_blkpg_ioctl()
105 static int blk_validate_byte_range(struct block_device *bdev, in blk_validate_byte_range() argument
108 unsigned int bs_mask = bdev_logical_block_size(bdev) - 1; in blk_validate_byte_range()
[all …]
H A Dholder.c13 static struct bd_holder_disk *bd_find_holder_disk(struct block_device *bdev, in bd_find_holder_disk() argument
19 if (holder->holder_dir == bdev->bd_holder_dir) in bd_find_holder_disk()
62 int bd_link_disk_holder(struct block_device *bdev, struct gendisk *disk) in bd_link_disk_holder() argument
70 if (bdev->bd_disk == disk) in bd_link_disk_holder()
77 mutex_lock(&bdev->bd_disk->open_mutex); in bd_link_disk_holder()
78 if (!disk_live(bdev->bd_disk)) { in bd_link_disk_holder()
79 mutex_unlock(&bdev->bd_disk->open_mutex); in bd_link_disk_holder()
82 kobject_get(bdev->bd_holder_dir); in bd_link_disk_holder()
83 mutex_unlock(&bdev->bd_disk->open_mutex); in bd_link_disk_holder()
86 WARN_ON_ONCE(!bdev->bd_holder); in bd_link_disk_holder()
[all …]
H A Dfops.c39 static bool blkdev_dio_invalid(struct block_device *bdev, struct kiocb *iocb, in blkdev_dio_invalid() argument
43 (bdev_logical_block_size(bdev) - 1); in blkdev_dio_invalid()
47 struct iov_iter *iter, struct block_device *bdev) in blkdev_iov_iter_get_pages() argument
50 bdev_logical_block_size(bdev) - 1); in blkdev_iov_iter_get_pages()
56 struct iov_iter *iter, struct block_device *bdev, in __blkdev_direct_IO_simple() argument
74 bio_init(&bio, bdev, vecs, nr_pages, REQ_OP_READ); in __blkdev_direct_IO_simple()
78 bio_init(&bio, bdev, vecs, nr_pages, dio_bio_write_op(iocb)); in __blkdev_direct_IO_simple()
87 ret = blkdev_iov_iter_get_pages(&bio, iter, bdev); in __blkdev_direct_IO_simple()
176 struct block_device *bdev, unsigned int nr_pages) in __blkdev_direct_IO() argument
186 bio = bio_alloc_bioset(bdev, nr_pages, opf, GFP_KERNEL, in __blkdev_direct_IO()
[all …]
H A Dblk-zoned.c191 bool bdev_zone_is_seq(struct block_device *bdev, sector_t sector) in bdev_zone_is_seq() argument
193 struct gendisk *disk = bdev->bd_disk; in bdev_zone_is_seq()
198 if (!bdev_is_zoned(bdev)) in bdev_zone_is_seq()
222 static int blkdev_do_report_zones(struct block_device *bdev, sector_t sector, in blkdev_do_report_zones() argument
226 struct gendisk *disk = bdev->bd_disk; in blkdev_do_report_zones()
228 if (!bdev_is_zoned(bdev) || WARN_ON_ONCE(!disk->fops->report_zones)) in blkdev_do_report_zones()
256 int blkdev_report_zones(struct block_device *bdev, sector_t sector, in blkdev_report_zones() argument
264 return blkdev_do_report_zones(bdev, sector, nr_zones, &args); in blkdev_report_zones()
268 static int blkdev_zone_reset_all(struct block_device *bdev) in blkdev_zone_reset_all() argument
272 bio_init(&bio, bdev, NULL, 0, REQ_OP_ZONE_RESET_ALL | REQ_SYNC); in blkdev_zone_reset_all()
[all …]
/linux/drivers/staging/media/atomisp/pci/hmm/
H A Dhmm_bo.c36 static int __bo_init(struct hmm_bo_device *bdev, struct hmm_buffer_object *bo, in __bo_init() argument
39 check_bodev_null_return(bdev, -EINVAL); in __bo_init()
52 bo->bdev = bdev; in __bo_init()
55 bo->start = bdev->start; in __bo_init()
98 rb_erase(&this->node, &this->bdev->free_rbtree); in __bo_search_and_remove_from_free_rbtree()
223 static struct hmm_buffer_object *__bo_break_up(struct hmm_bo_device *bdev, in __bo_break_up() argument
231 new_bo = kmem_cache_alloc(bdev->bo_cache, GFP_KERNEL); in __bo_break_up()
236 ret = __bo_init(bdev, new_bo, pgnr); in __bo_break_up()
239 kmem_cache_free(bdev->bo_cache, new_bo); in __bo_break_up()
248 spin_lock_irqsave(&bdev->list_lock, flags); in __bo_break_up()
[all …]
/linux/drivers/s390/block/
H A Dscm_blk.c144 static bool scm_permit_request(struct scm_blk_dev *bdev, struct request *req) in scm_permit_request() argument
146 return rq_data_dir(req) != WRITE || bdev->state != SCM_WR_PROHIBIT; in scm_permit_request()
179 struct scm_blk_dev *bdev = scmrq->bdev; in scm_request_prepare() local
180 struct scm_device *scmdev = bdev->gendisk->private_data; in scm_request_prepare()
216 static inline void scm_request_init(struct scm_blk_dev *bdev, in scm_request_init() argument
225 aobrq->scmdev = bdev->scmdev; in scm_request_init()
228 scmrq->bdev = bdev; in scm_request_init()
237 struct scm_blk_dev *bdev = scmrq->bdev; in scm_request_requeue() local
243 atomic_dec(&bdev->queued_reqs); in scm_request_requeue()
245 blk_mq_kick_requeue_list(bdev->rq); in scm_request_requeue()
[all …]
H A Dscm_drv.c18 struct scm_blk_dev *bdev = dev_get_drvdata(&scmdev->dev); in scm_notify() local
30 scm_blk_set_available(bdev); in scm_notify()
37 struct scm_blk_dev *bdev; in scm_probe() local
46 bdev = kzalloc_obj(*bdev); in scm_probe()
47 if (!bdev) in scm_probe()
50 dev_set_drvdata(&scmdev->dev, bdev); in scm_probe()
51 ret = scm_blk_dev_setup(bdev, scmdev); in scm_probe()
54 kfree(bdev); in scm_probe()
64 struct scm_blk_dev *bdev = dev_get_drvdata(&scmdev->dev); in scm_remove() local
66 scm_blk_dev_cleanup(bdev); in scm_remove()
[all …]
/linux/drivers/comedi/drivers/
H A Dcomedi_bond.c81 struct bonded_device *bdev = *devs++; in bonding_dio_insn_bits() local
83 if (base_chan < bdev->nchans) { in bonding_dio_insn_bits()
92 b_chans = bdev->nchans - base_chan; in bonding_dio_insn_bits()
100 ret = comedi_dio_bitfield2(bdev->dev, bdev->subdev, in bonding_dio_insn_bits()
117 base_chan -= bdev->nchans; in bonding_dio_insn_bits()
131 struct bonded_device *bdev; in bonding_dio_insn_config() local
138 for (bdev = *devs++; chan >= bdev->nchans; bdev = *devs++) in bonding_dio_insn_config()
139 chan -= bdev->nchans; in bonding_dio_insn_config()
154 ret = comedi_dio_config(bdev->dev, bdev->subdev, chan, data[0]); in bonding_dio_insn_config()
157 ret = comedi_dio_get_config(bdev->dev, bdev->subdev, chan, in bonding_dio_insn_config()
[all …]
/linux/drivers/input/misc/
H A Dcobalt_btns.c37 struct buttons_dev *bdev = input_get_drvdata(input); in handle_buttons() local
41 status = ~readl(bdev->reg) >> 24; in handle_buttons()
43 for (i = 0; i < ARRAY_SIZE(bdev->keymap); i++) { in handle_buttons()
45 if (++bdev->count[i] == BUTTONS_COUNT_THRESHOLD) { in handle_buttons()
47 input_report_key(input, bdev->keymap[i], 1); in handle_buttons()
51 if (bdev->count[i] >= BUTTONS_COUNT_THRESHOLD) { in handle_buttons()
53 input_report_key(input, bdev->keymap[i], 0); in handle_buttons()
56 bdev->count[i] = 0; in handle_buttons()
63 struct buttons_dev *bdev; in cobalt_buttons_probe() local
68 bdev = devm_kzalloc(&pdev->dev, sizeof(*bdev), GFP_KERNEL); in cobalt_buttons_probe()
[all …]
H A Dsgi_btns.c54 struct buttons_dev *bdev = input_get_drvdata(input); in handle_buttons() local
60 for (i = 0; i < ARRAY_SIZE(bdev->keymap); i++) { in handle_buttons()
62 if (++bdev->count[i] == BUTTONS_COUNT_THRESHOLD) { in handle_buttons()
64 input_report_key(input, bdev->keymap[i], 1); in handle_buttons()
68 if (bdev->count[i] >= BUTTONS_COUNT_THRESHOLD) { in handle_buttons()
70 input_report_key(input, bdev->keymap[i], 0); in handle_buttons()
73 bdev->count[i] = 0; in handle_buttons()
80 struct buttons_dev *bdev; in sgi_buttons_probe() local
84 bdev = devm_kzalloc(&pdev->dev, sizeof(*bdev), GFP_KERNEL); in sgi_buttons_probe()
85 if (!bdev) in sgi_buttons_probe()
[all …]
/linux/drivers/gpu/drm/ttm/
H A Dttm_bo.c65 man = ttm_manager_type(bo->bdev, mem_type); in ttm_bo_mem_space_debug()
110 spin_lock(&bo->bdev->lru_lock); in ttm_bo_set_bulk_move()
116 spin_unlock(&bo->bdev->lru_lock); in ttm_bo_set_bulk_move()
125 struct ttm_device *bdev = bo->bdev; in ttm_bo_handle_move_mem() local
129 old_use_tt = !bo->resource || ttm_manager_type(bdev, bo->resource->mem_type)->use_tt; in ttm_bo_handle_move_mem()
130 new_use_tt = ttm_manager_type(bdev, mem->mem_type)->use_tt; in ttm_bo_handle_move_mem()
157 ret = bdev->funcs->move(bo, evict, ctx, mem, hop); in ttm_bo_handle_move_mem()
184 if (bo->bdev->funcs->delete_mem_notify) in ttm_bo_cleanup_memtype_use()
185 bo->bdev->funcs->delete_mem_notify(bo); in ttm_bo_cleanup_memtype_use()
210 spin_lock(&bo->bdev->lru_lock); in ttm_bo_individualize_resv()
[all …]
H A Dttm_tt.c70 struct ttm_device *bdev = bo->bdev; in ttm_tt_create() local
98 if (ttm_pool_uses_dma_alloc(&bdev->pool) && in ttm_tt_create()
104 bo->ttm = bdev->funcs->ttm_tt_create(bo, page_flags); in ttm_tt_create()
147 void ttm_tt_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) in ttm_tt_destroy() argument
149 bdev->funcs->ttm_tt_destroy(bdev, ttm); in ttm_tt_destroy()
281 long ttm_tt_backup(struct ttm_device *bdev, struct ttm_tt *tt, in ttm_tt_backup() argument
289 ret = ttm_pool_backup(&bdev->pool, tt, &flags); in ttm_tt_backup()
298 int ttm_tt_restore(struct ttm_device *bdev, struct ttm_tt *tt, in ttm_tt_restore() argument
301 int ret = ttm_pool_restore_and_alloc(&bdev->pool, tt, ctx); in ttm_tt_restore()
322 int ttm_tt_swapout(struct ttm_device *bdev, struct ttm_tt *ttm, in ttm_tt_swapout() argument
[all …]
H A Dttm_bo_util.c49 int ttm_mem_io_reserve(struct ttm_device *bdev, in ttm_mem_io_reserve() argument
56 if (!bdev->funcs->io_mem_reserve) in ttm_mem_io_reserve()
59 return bdev->funcs->io_mem_reserve(bdev, mem); in ttm_mem_io_reserve()
62 void ttm_mem_io_free(struct ttm_device *bdev, in ttm_mem_io_free() argument
71 if (bdev->funcs->io_mem_free) in ttm_mem_io_free()
72 bdev->funcs->io_mem_free(bdev, mem); in ttm_mem_io_free()
150 struct ttm_device *bdev = bo->bdev; in ttm_bo_move_memcpy() local
152 ttm_manager_type(bo->bdev, dst_mem->mem_type); in ttm_bo_move_memcpy()
167 src_man = ttm_manager_type(bdev, src_mem->mem_type); in ttm_bo_move_memcpy()
175 dst_iter = ttm_kmap_iter_linear_io_init(&_dst_iter.io, bdev, dst_mem); in ttm_bo_move_memcpy()
[all …]
/linux/drivers/hid/
H A Dhid-picolcd_backlight.c16 static int picolcd_get_brightness(struct backlight_device *bdev) in picolcd_get_brightness() argument
18 struct picolcd_data *data = bl_get_data(bdev); in picolcd_get_brightness()
22 static int picolcd_set_brightness(struct backlight_device *bdev) in picolcd_set_brightness() argument
24 struct picolcd_data *data = bl_get_data(bdev); in picolcd_set_brightness()
31 data->lcd_brightness = bdev->props.brightness & 0x0ff; in picolcd_set_brightness()
32 data->lcd_power = bdev->props.power; in picolcd_set_brightness()
50 struct backlight_device *bdev; in picolcd_init_backlight() local
63 bdev = backlight_device_register(dev_name(dev), dev, data, in picolcd_init_backlight()
65 if (IS_ERR(bdev)) { in picolcd_init_backlight()
67 return PTR_ERR(bdev); in picolcd_init_backlight()
[all …]
/linux/drivers/input/keyboard/
H A Dgpio_keys_polled.c48 struct gpio_keys_polled_dev *bdev = input_get_drvdata(input); in gpio_keys_button_event() local
54 __set_bit(button->code, bdev->rel_axis_seen); in gpio_keys_button_event()
59 __set_bit(button->code, bdev->abs_axis_seen); in gpio_keys_button_event()
89 struct gpio_keys_polled_dev *bdev = input_get_drvdata(input); in gpio_keys_polled_poll() local
90 const struct gpio_keys_platform_data *pdata = bdev->pdata; in gpio_keys_polled_poll()
93 memset(bdev->rel_axis_seen, 0, sizeof(bdev->rel_axis_seen)); in gpio_keys_polled_poll()
94 memset(bdev->abs_axis_seen, 0, sizeof(bdev->abs_axis_seen)); in gpio_keys_polled_poll()
97 struct gpio_keys_button_data *bdata = &bdev->data[i]; in gpio_keys_polled_poll()
110 if (!test_bit(i, bdev->rel_axis_seen)) in gpio_keys_polled_poll()
115 if (!test_bit(i, bdev->abs_axis_seen)) in gpio_keys_polled_poll()
[all …]
/linux/drivers/nvme/host/
H A Dpr.c52 static int nvme_send_ns_head_pr_command(struct block_device *bdev, in nvme_send_ns_head_pr_command() argument
55 struct nvme_ns_head *head = bdev->bd_disk->private_data; in nvme_send_ns_head_pr_command()
95 static int __nvme_send_pr_command(struct block_device *bdev, u32 cdw10, in __nvme_send_pr_command() argument
104 if (nvme_disk_is_ns_head(bdev->bd_disk)) in __nvme_send_pr_command()
105 return nvme_send_ns_head_pr_command(bdev, &c, data, data_len); in __nvme_send_pr_command()
106 return nvme_send_ns_pr_command(bdev->bd_disk->private_data, &c, in __nvme_send_pr_command()
110 static int nvme_send_pr_command(struct block_device *bdev, u32 cdw10, u32 cdw11, in nvme_send_pr_command() argument
115 ret = __nvme_send_pr_command(bdev, cdw10, cdw11, op, data, data_len); in nvme_send_pr_command()
119 static int nvme_pr_register(struct block_device *bdev, u64 old_key, u64 new_key, in nvme_pr_register() argument
136 return nvme_send_pr_command(bdev, cdw10, 0, nvme_cmd_resv_register, in nvme_pr_register()
[all …]
/linux/drivers/gpu/drm/radeon/
H A Dradeon_ttm.c56 static int radeon_ttm_tt_bind(struct ttm_device *bdev, struct ttm_tt *ttm,
58 static void radeon_ttm_tt_unbind(struct ttm_device *bdev, struct ttm_tt *ttm);
60 struct radeon_device *radeon_get_rdev(struct ttm_device *bdev) in radeon_get_rdev() argument
65 mman = container_of(bdev, struct radeon_mman, bdev); in radeon_get_rdev()
72 return ttm_range_man_init(&rdev->mman.bdev, TTM_PL_VRAM, in radeon_ttm_init_vram()
78 return ttm_range_man_init(&rdev->mman.bdev, TTM_PL_TT, in radeon_ttm_init_gtt()
144 rdev = radeon_get_rdev(bo->bdev); in radeon_move_blit()
198 r = radeon_ttm_tt_bind(bo->bdev, bo->ttm, new_mem); in radeon_bo_move()
207 rdev = radeon_get_rdev(bo->bdev); in radeon_bo_move()
221 radeon_ttm_tt_unbind(bo->bdev, bo->ttm); in radeon_bo_move()
[all …]
/linux/drivers/nvme/target/
H A Dzns.c37 u8 zasl = nvmet_zasl(bdev_max_zone_append_sectors(ns->bdev)); in nvmet_bdev_zns_enable()
38 struct gendisk *bd_disk = ns->bdev->bd_disk; in nvmet_bdev_zns_enable()
52 if (get_capacity(bd_disk) & (bdev_zone_sectors(ns->bdev) - 1)) in nvmet_bdev_zns_enable()
59 ret = blkdev_report_zones(ns->bdev, 0, bdev_nr_zones(ns->bdev), in nvmet_bdev_zns_enable()
64 ns->blksize_shift = blksize_bits(bdev_logical_block_size(ns->bdev)); in nvmet_bdev_zns_enable()
123 if (!bdev_is_zoned(req->ns->bdev)) { in nvmet_execute_identify_ns_zns()
129 zsze = (bdev_zone_sectors(req->ns->bdev) << 9) >> in nvmet_execute_identify_ns_zns()
133 mor = bdev_max_open_zones(req->ns->bdev); in nvmet_execute_identify_ns_zns()
140 mar = bdev_max_active_zones(req->ns->bdev); in nvmet_execute_identify_ns_zns()
159 if (sect >= get_capacity(req->ns->bdev->bd_disk)) { in nvmet_bdev_validate_zone_mgmt_recv()
[all …]
/linux/drivers/pci/controller/
H A Dpcie-iproc-bcma.c29 struct bcma_device *bdev = container_of(pcie->dev, struct bcma_device, dev); in iproc_bcma_pcie_map_irq() local
31 return bcma_core_irq(bdev, 5); in iproc_bcma_pcie_map_irq()
34 static int iproc_bcma_pcie_probe(struct bcma_device *bdev) in iproc_bcma_pcie_probe() argument
36 struct device *dev = &bdev->dev; in iproc_bcma_pcie_probe()
50 pcie->base = bdev->io_addr; in iproc_bcma_pcie_probe()
56 pcie->base_addr = bdev->addr; in iproc_bcma_pcie_probe()
58 pcie->mem.start = bdev->addr_s[0]; in iproc_bcma_pcie_probe()
59 pcie->mem.end = bdev->addr_s[0] + SZ_128M - 1; in iproc_bcma_pcie_probe()
69 bcma_set_drvdata(bdev, pcie); in iproc_bcma_pcie_probe()
74 static void iproc_bcma_pcie_remove(struct bcma_device *bdev) in iproc_bcma_pcie_remove() argument
[all …]
/linux/drivers/gpu/drm/ttm/tests/
H A Dttm_mock_manager.c97 int ttm_mock_manager_init(struct ttm_device *bdev, u32 mem_type, u32 size) in ttm_mock_manager_init() argument
121 ttm_resource_manager_init(base, bdev, size); in ttm_mock_manager_init()
122 ttm_set_driver_manager(bdev, mem_type, base); in ttm_mock_manager_init()
129 void ttm_mock_manager_fini(struct ttm_device *bdev, u32 mem_type) in ttm_mock_manager_fini() argument
135 man = ttm_manager_type(bdev, mem_type); in ttm_mock_manager_fini()
138 err = ttm_resource_manager_evict_all(bdev, man); in ttm_mock_manager_fini()
148 ttm_set_driver_manager(bdev, mem_type, NULL); in ttm_mock_manager_fini()
193 int ttm_bad_manager_init(struct ttm_device *bdev, u32 mem_type, u32 size) in ttm_bad_manager_init() argument
203 ttm_resource_manager_init(man, bdev, size); in ttm_bad_manager_init()
204 ttm_set_driver_manager(bdev, mem_type, man); in ttm_bad_manager_init()
[all …]
/linux/drivers/gpu/drm/loongson/
H A Dlsdc_ttm.c87 static void lsdc_ttm_tt_destroy(struct ttm_device *bdev, struct ttm_tt *tt) in lsdc_ttm_tt_destroy() argument
112 static int lsdc_ttm_tt_populate(struct ttm_device *bdev, in lsdc_ttm_tt_populate() argument
126 return ttm_pool_alloc(&bdev->pool, ttm, ctx); in lsdc_ttm_tt_populate()
129 static void lsdc_ttm_tt_unpopulate(struct ttm_device *bdev, in lsdc_ttm_tt_unpopulate() argument
137 return ttm_pool_free(&bdev->pool, ttm); in lsdc_ttm_tt_unpopulate()
220 static int lsdc_bo_reserve_io_mem(struct ttm_device *bdev, in lsdc_bo_reserve_io_mem() argument
223 struct lsdc_device *ldev = tdev_to_ldev(bdev); in lsdc_bo_reserve_io_mem()
291 struct lsdc_device *ldev = tdev_to_ldev(tbo->bdev); in lsdc_bo_pin()
326 struct lsdc_device *ldev = tdev_to_ldev(tbo->bdev); in lsdc_bo_unpin()
406 struct ttm_device *bdev = &ldev->bdev; in lsdc_bo_evict_vram() local
[all …]
/linux/drivers/gpu/drm/qxl/
H A Dqxl_ttm.c41 static struct qxl_device *qxl_get_qdev(struct ttm_device *bdev) in qxl_get_qdev() argument
46 mman = container_of(bdev, struct qxl_mman, bdev); in qxl_get_qdev()
72 int qxl_ttm_io_mem_reserve(struct ttm_device *bdev, in qxl_ttm_io_mem_reserve() argument
75 struct qxl_device *qdev = qxl_get_qdev(bdev); in qxl_ttm_io_mem_reserve()
101 static void qxl_ttm_backend_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) in qxl_ttm_backend_destroy() argument
188 return ttm_range_man_init(&qdev->mman.bdev, type, false, size); in qxl_ttm_init_mem_type()
197 r = ttm_device_init(&qdev->mman.bdev, &qxl_bo_driver, NULL, in qxl_ttm_init()
229 ttm_range_man_fini(&qdev->mman.bdev, TTM_PL_VRAM); in qxl_ttm_fini()
230 ttm_range_man_fini(&qdev->mman.bdev, TTM_PL_PRIV); in qxl_ttm_fini()
231 ttm_device_fini(&qdev->mman.bdev); in qxl_ttm_fini()
[all …]

12345678910>>...13