Home
last modified time | relevance | path

Searched refs:tile (Results 1 – 25 of 166) sorted by relevance

1234567

/linux/drivers/gpu/drm/xe/
H A Dxe_tile.c87 static int xe_tile_alloc(struct xe_tile *tile) in xe_tile_alloc() argument
89 struct drm_device *drm = &tile_to_xe(tile)->drm; in xe_tile_alloc()
91 tile->mem.ggtt = drmm_kzalloc(drm, sizeof(*tile->mem.ggtt), in xe_tile_alloc()
93 if (!tile->mem.ggtt) in xe_tile_alloc()
95 tile->mem.ggtt->tile = tile; in xe_tile_alloc()
97 tile->mem.vram_mgr = drmm_kzalloc(drm, sizeof(*tile->mem.vram_mgr), GFP_KERNEL); in xe_tile_alloc()
98 if (!tile->mem.vram_mgr) in xe_tile_alloc()
115 int xe_tile_init_early(struct xe_tile *tile, struct xe_device *xe, u8 id) in xe_tile_init_early() argument
119 tile->xe = xe; in xe_tile_init_early()
120 tile->id = id; in xe_tile_init_early()
[all …]
H A Dxe_pcode.c32 static int pcode_mailbox_status(struct xe_tile *tile) in pcode_mailbox_status() argument
47 err = xe_mmio_read32(&tile->mmio, PCODE_MAILBOX) & PCODE_ERROR_MASK; in pcode_mailbox_status()
49 drm_err(&tile_to_xe(tile)->drm, "PCODE Mailbox failed: %d %s", err, in pcode_mailbox_status()
57 static int __pcode_mailbox_rw(struct xe_tile *tile, u32 mbox, u32 *data0, u32 *data1, in __pcode_mailbox_rw() argument
61 struct xe_mmio *mmio = &tile->mmio; in __pcode_mailbox_rw()
64 if (tile_to_xe(tile)->info.skip_pcode) in __pcode_mailbox_rw()
85 return pcode_mailbox_status(tile); in __pcode_mailbox_rw()
88 static int pcode_mailbox_rw(struct xe_tile *tile, u32 mbox, u32 *data0, u32 *data1, in pcode_mailbox_rw() argument
92 if (tile_to_xe(tile)->info.skip_pcode) in pcode_mailbox_rw()
95 lockdep_assert_held(&tile->pcode.lock); in pcode_mailbox_rw()
[all …]
H A Dxe_irq.c45 drm_WARN(&mmio->tile->xe->drm, 1, in assert_iir_is_zero()
58 static void unmask_and_enable(struct xe_tile *tile, u32 irqregs, u32 bits) in unmask_and_enable() argument
60 struct xe_mmio *mmio = &tile->mmio; in unmask_and_enable()
76 static void mask_and_disable(struct xe_tile *tile, u32 irqregs) in mask_and_disable() argument
78 struct xe_mmio *mmio = &tile->mmio; in mask_and_disable()
267 static struct xe_gt *pick_engine_gt(struct xe_tile *tile, in pick_engine_gt() argument
271 struct xe_device *xe = tile_to_xe(tile); in pick_engine_gt()
274 return tile->primary_gt; in pick_engine_gt()
279 return tile->media_gt; in pick_engine_gt()
285 return tile->media_gt; in pick_engine_gt()
[all …]
H A Dxe_vram.c218 static int tile_vram_size(struct xe_tile *tile, u64 *vram_size, in tile_vram_size() argument
221 struct xe_device *xe = tile_to_xe(tile); in tile_vram_size()
222 struct xe_gt *gt = tile->primary_gt; in tile_vram_size()
233 for_each_if(t->id < tile->id) in tile_vram_size()
261 offset = xe_mmio_read64_2x32(&tile->mmio, GSMBASE); in tile_vram_size()
275 struct xe_tile *tile; in vram_fini() local
283 for_each_tile(tile, xe, id) in vram_fini()
284 tile->mem.vram.mapping = NULL; in vram_fini()
297 struct xe_tile *tile; in xe_vram_probe() local
311 tile = xe_device_get_root_tile(xe); in xe_vram_probe()
[all …]
H A Dxe_tile_sysfs.c27 struct xe_tile *tile = arg; in tile_sysfs_fini() local
29 kobject_put(tile->sysfs); in tile_sysfs_fini()
32 int xe_tile_sysfs_init(struct xe_tile *tile) in xe_tile_sysfs_init() argument
34 struct xe_device *xe = tile_to_xe(tile); in xe_tile_sysfs_init()
44 kt->tile = tile; in xe_tile_sysfs_init()
46 err = kobject_add(&kt->base, &dev->kobj, "tile%d", tile->id); in xe_tile_sysfs_init()
52 tile->sysfs = &kt->base; in xe_tile_sysfs_init()
54 err = xe_vram_freq_sysfs_init(tile); in xe_tile_sysfs_init()
58 return devm_add_action_or_reset(xe->drm.dev, tile_sysfs_fini, tile); in xe_tile_sysfs_init()
H A Dxe_ggtt.c111 struct xe_tile *tile = ggtt->tile; in ggtt_update_access_counter() local
112 struct xe_gt *affected_gt = XE_WA(tile->primary_gt, 22019338487) ? in ggtt_update_access_counter()
113 tile->primary_gt : tile->media_gt; in ggtt_update_access_counter()
115 u32 max_gtt_writes = XE_WA(ggtt->tile->primary_gt, 22019338487) ? 1100 : 63; in ggtt_update_access_counter()
131 xe_tile_assert(ggtt->tile, !(addr & XE_PTE_MASK)); in xe_ggtt_set_pte()
132 xe_tile_assert(ggtt->tile, addr < ggtt->size); in xe_ggtt_set_pte()
145 u16 pat_index = tile_to_xe(ggtt->tile)->pat.idx[XE_CACHE_WB]; in xe_ggtt_clear()
149 xe_tile_assert(ggtt->tile, start < end); in xe_ggtt_clear()
217 struct xe_device *xe = tile_to_xe(ggtt->tile); in xe_ggtt_init_early()
232 ggtt->gsm = ggtt->tile->mmio.regs + SZ_8M; in xe_ggtt_init_early()
[all …]
H A Dxe_mmio.c29 struct xe_tile *tile; in tiles_fini() local
32 for_each_remote_tile(tile, xe, id) in tiles_fini()
33 tile->mmio.regs = NULL; in tiles_fini()
57 struct xe_tile *tile; in mmio_multi_tile_setup() local
98 for_each_tile(tile, xe, id) { in mmio_multi_tile_setup()
99 tile->mmio.regs_size = SZ_4M; in mmio_multi_tile_setup()
100 tile->mmio.regs = regs; in mmio_multi_tile_setup()
101 tile->mmio.tile = tile; in mmio_multi_tile_setup()
127 struct xe_tile *tile; in mmio_extension_setup() local
135 for_each_tile(tile, xe, id) { in mmio_extension_setup()
[all …]
H A Dxe_bo_evict.c31 struct xe_tile *tile; in xe_bo_evict_all() local
93 for_each_tile(tile, xe, id) in xe_bo_evict_all()
94 xe_tile_migrate_wait(tile); in xe_bo_evict_all()
155 struct xe_tile *tile; in xe_bo_restore_kernel() local
158 for_each_tile(tile, xe, id) { in xe_bo_restore_kernel()
159 if (tile != bo->tile && !(bo->flags & XE_BO_FLAG_GGTTx(tile))) in xe_bo_restore_kernel()
162 mutex_lock(&tile->mem.ggtt->lock); in xe_bo_restore_kernel()
163 xe_ggtt_map_bo(tile->mem.ggtt, bo); in xe_bo_restore_kernel()
164 mutex_unlock(&tile->mem.ggtt->lock); in xe_bo_restore_kernel()
196 struct xe_tile *tile; in xe_bo_restore_user() local
[all …]
H A Dxe_pt.c59 static u64 __xe_pt_empty_pte(struct xe_tile *tile, struct xe_vm *vm, in __xe_pt_empty_pte() argument
62 struct xe_device *xe = tile_to_xe(tile); in __xe_pt_empty_pte()
64 u8 id = tile->id; in __xe_pt_empty_pte()
100 struct xe_pt *xe_pt_create(struct xe_vm *vm, struct xe_tile *tile, in xe_pt_create() argument
118 bo = xe_bo_create_pin_map(vm->xe, tile, vm, SZ_4K, in xe_pt_create()
120 XE_BO_FLAG_VRAM_IF_DGFX(tile) | in xe_pt_create()
135 xe_tile_assert(tile, level <= XE_VM_MAX_LEVEL); in xe_pt_create()
155 void xe_pt_populate_empty(struct xe_tile *tile, struct xe_vm *vm, in xe_pt_populate_empty() argument
169 empty = __xe_pt_empty_pte(tile, vm, pt->level); in xe_pt_populate_empty()
265 struct xe_tile *tile; member
[all …]
H A Dxe_bo.h26 #define XE_BO_FLAG_VRAM_IF_DGFX(tile) (IS_DGFX(tile_to_xe(tile)) ? \ argument
27 XE_BO_FLAG_VRAM0 << (tile)->id : \
55 #define XE_BO_FLAG_GGTTx(tile) \ argument
56 (XE_BO_FLAG_GGTT0 << (tile)->id)
84 struct xe_tile *tile, struct dma_resv *resv,
90 struct xe_tile *tile, struct xe_vm *vm,
93 struct xe_bo *xe_bo_create_locked(struct xe_device *xe, struct xe_tile *tile,
96 struct xe_bo *xe_bo_create(struct xe_device *xe, struct xe_tile *tile,
99 struct xe_bo *xe_bo_create_user(struct xe_device *xe, struct xe_tile *tile,
103 struct xe_bo *xe_bo_create_pin_map(struct xe_device *xe, struct xe_tile *tile,
[all …]
H A Dxe_pcode.h13 void xe_pcode_init(struct xe_tile *tile);
16 int xe_pcode_init_min_freq_table(struct xe_tile *tile, u32 min_gt_freq,
18 int xe_pcode_read(struct xe_tile *tile, u32 mbox, u32 *val, u32 *val1);
19 int xe_pcode_write_timeout(struct xe_tile *tile, u32 mbox, u32 val,
21 #define xe_pcode_write(tile, mbox, val) \ argument
22 xe_pcode_write_timeout(tile, mbox, val, 1)
24 int xe_pcode_request(struct xe_tile *tile, u32 mbox, u32 request,
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
H A Dnv20.c31 u32 flags, struct nvkm_fb_tile *tile) in nv20_fb_tile_init() argument
33 tile->addr = 0x00000001 | addr; in nv20_fb_tile_init()
34 tile->limit = max(1u, addr + size) - 1; in nv20_fb_tile_init()
35 tile->pitch = pitch; in nv20_fb_tile_init()
37 fb->func->tile.comp(fb, i, size, flags, tile); in nv20_fb_tile_init()
38 tile->addr |= 2; in nv20_fb_tile_init()
44 struct nvkm_fb_tile *tile) in nv20_fb_tile_comp() argument
48 if (!nvkm_mm_head(&fb->tags.mm, 0, 1, tags, tags, 1, &tile->tag)) { in nv20_fb_tile_comp()
49 if (!(flags & 2)) tile->zcomp = 0x00000000; /* Z16 */ in nv20_fb_tile_comp()
50 else tile->zcomp = 0x04000000; /* Z24S8 */ in nv20_fb_tile_comp()
[all …]
H A Dnv30.c31 u32 flags, struct nvkm_fb_tile *tile) in nv30_fb_tile_init() argument
35 tile->addr = (0 << 4); in nv30_fb_tile_init()
37 if (fb->func->tile.comp) /* z compression */ in nv30_fb_tile_init()
38 fb->func->tile.comp(fb, i, size, flags, tile); in nv30_fb_tile_init()
39 tile->addr = (1 << 4); in nv30_fb_tile_init()
42 tile->addr |= 0x00000001; /* enable */ in nv30_fb_tile_init()
43 tile->addr |= addr; in nv30_fb_tile_init()
44 tile->limit = max(1u, addr + size) - 1; in nv30_fb_tile_init()
45 tile->pitch = pitch; in nv30_fb_tile_init()
50 struct nvkm_fb_tile *tile) in nv30_fb_tile_comp() argument
[all …]
H A Dnv10.c31 u32 flags, struct nvkm_fb_tile *tile) in nv10_fb_tile_init() argument
33 tile->addr = 0x80000000 | addr; in nv10_fb_tile_init()
34 tile->limit = max(1u, addr + size) - 1; in nv10_fb_tile_init()
35 tile->pitch = pitch; in nv10_fb_tile_init()
39 nv10_fb_tile_fini(struct nvkm_fb *fb, int i, struct nvkm_fb_tile *tile) in nv10_fb_tile_fini() argument
41 tile->addr = 0; in nv10_fb_tile_fini()
42 tile->limit = 0; in nv10_fb_tile_fini()
43 tile->pitch = 0; in nv10_fb_tile_fini()
44 tile->zcomp = 0; in nv10_fb_tile_fini()
48 nv10_fb_tile_prog(struct nvkm_fb *fb, int i, struct nvkm_fb_tile *tile) in nv10_fb_tile_prog() argument
[all …]
H A Dnv35.c31 struct nvkm_fb_tile *tile) in nv35_fb_tile_comp() argument
35 if (!nvkm_mm_head(&fb->tags.mm, 0, 1, tags, tags, 1, &tile->tag)) { in nv35_fb_tile_comp()
36 if (flags & 2) tile->zcomp |= 0x04000000; /* Z16 */ in nv35_fb_tile_comp()
37 else tile->zcomp |= 0x08000000; /* Z24S8 */ in nv35_fb_tile_comp()
38 tile->zcomp |= ((tile->tag->offset ) >> 6); in nv35_fb_tile_comp()
39 tile->zcomp |= ((tile->tag->offset + tags - 1) >> 6) << 13; in nv35_fb_tile_comp()
41 tile->zcomp |= 0x40000000; in nv35_fb_tile_comp()
50 .tile.regions = 8,
51 .tile.init = nv30_fb_tile_init,
52 .tile.comp = nv35_fb_tile_comp,
[all …]
H A Dnv36.c31 struct nvkm_fb_tile *tile) in nv36_fb_tile_comp() argument
35 if (!nvkm_mm_head(&fb->tags.mm, 0, 1, tags, tags, 1, &tile->tag)) { in nv36_fb_tile_comp()
36 if (flags & 2) tile->zcomp |= 0x10000000; /* Z16 */ in nv36_fb_tile_comp()
37 else tile->zcomp |= 0x20000000; /* Z24S8 */ in nv36_fb_tile_comp()
38 tile->zcomp |= ((tile->tag->offset ) >> 6); in nv36_fb_tile_comp()
39 tile->zcomp |= ((tile->tag->offset + tags - 1) >> 6) << 14; in nv36_fb_tile_comp()
41 tile->zcomp |= 0x80000000; in nv36_fb_tile_comp()
50 .tile.regions = 8,
51 .tile.init = nv30_fb_tile_init,
52 .tile.comp = nv36_fb_tile_comp,
[all …]
H A Dnv40.c31 struct nvkm_fb_tile *tile) in nv40_fb_tile_comp() argument
36 !nvkm_mm_head(&fb->tags.mm, 0, 1, tags, tags, 1, &tile->tag)) { in nv40_fb_tile_comp()
37 tile->zcomp = 0x28000000; /* Z24S8_SPLIT_GRAD */ in nv40_fb_tile_comp()
38 tile->zcomp |= ((tile->tag->offset ) >> 8); in nv40_fb_tile_comp()
39 tile->zcomp |= ((tile->tag->offset + tags - 1) >> 8) << 13; in nv40_fb_tile_comp()
41 tile->zcomp |= 0x40000000; in nv40_fb_tile_comp()
56 .tile.regions = 8,
57 .tile.init = nv30_fb_tile_init,
58 .tile.comp = nv40_fb_tile_comp,
59 .tile.fini = nv20_fb_tile_fini,
[all …]
H A Dnv44.c31 u32 flags, struct nvkm_fb_tile *tile) in nv44_fb_tile_init() argument
33 tile->addr = 0x00000001; /* mode = vram */ in nv44_fb_tile_init()
34 tile->addr |= addr; in nv44_fb_tile_init()
35 tile->limit = max(1u, addr + size) - 1; in nv44_fb_tile_init()
36 tile->pitch = pitch; in nv44_fb_tile_init()
40 nv44_fb_tile_prog(struct nvkm_fb *fb, int i, struct nvkm_fb_tile *tile) in nv44_fb_tile_prog() argument
43 nvkm_wr32(device, 0x100604 + (i * 0x10), tile->limit); in nv44_fb_tile_prog()
44 nvkm_wr32(device, 0x100608 + (i * 0x10), tile->pitch); in nv44_fb_tile_prog()
45 nvkm_wr32(device, 0x100600 + (i * 0x10), tile->addr); in nv44_fb_tile_prog()
60 .tile.regions = 12,
[all …]
H A Dnv25.c31 struct nvkm_fb_tile *tile) in nv25_fb_tile_comp() argument
35 if (!nvkm_mm_head(&fb->tags.mm, 0, 1, tags, tags, 1, &tile->tag)) { in nv25_fb_tile_comp()
36 if (!(flags & 2)) tile->zcomp = 0x00100000; /* Z16 */ in nv25_fb_tile_comp()
37 else tile->zcomp = 0x00200000; /* Z24S8 */ in nv25_fb_tile_comp()
38 tile->zcomp |= tile->tag->offset; in nv25_fb_tile_comp()
40 tile->zcomp |= 0x01000000; in nv25_fb_tile_comp()
48 .tile.regions = 8,
49 .tile.init = nv20_fb_tile_init,
50 .tile.comp = nv25_fb_tile_comp,
51 .tile.fini = nv20_fb_tile_fini,
[all …]
H A Dnv46.c31 u32 flags, struct nvkm_fb_tile *tile) in nv46_fb_tile_init() argument
34 if (!(flags & 4)) tile->addr = (0 << 3); in nv46_fb_tile_init()
35 else tile->addr = (1 << 3); in nv46_fb_tile_init()
37 tile->addr |= 0x00000001; /* mode = vram */ in nv46_fb_tile_init()
38 tile->addr |= addr; in nv46_fb_tile_init()
39 tile->limit = max(1u, addr + size) - 1; in nv46_fb_tile_init()
40 tile->pitch = pitch; in nv46_fb_tile_init()
46 .tile.regions = 15,
47 .tile.init = nv46_fb_tile_init,
48 .tile.fini = nv20_fb_tile_fini,
[all …]
H A Dnv41.c30 nv41_fb_tile_prog(struct nvkm_fb *fb, int i, struct nvkm_fb_tile *tile) in nv41_fb_tile_prog() argument
33 nvkm_wr32(device, 0x100604 + (i * 0x10), tile->limit); in nv41_fb_tile_prog()
34 nvkm_wr32(device, 0x100608 + (i * 0x10), tile->pitch); in nv41_fb_tile_prog()
35 nvkm_wr32(device, 0x100600 + (i * 0x10), tile->addr); in nv41_fb_tile_prog()
37 nvkm_wr32(device, 0x100700 + (i * 0x04), tile->zcomp); in nv41_fb_tile_prog()
50 .tile.regions = 12,
51 .tile.init = nv30_fb_tile_init,
52 .tile.comp = nv40_fb_tile_comp,
53 .tile.fini = nv20_fb_tile_fini,
54 .tile.prog = nv41_fb_tile_prog,
H A Dbase.c35 nvkm_fb_tile_fini(struct nvkm_fb *fb, int region, struct nvkm_fb_tile *tile) in nvkm_fb_tile_fini() argument
37 fb->func->tile.fini(fb, region, tile); in nvkm_fb_tile_fini()
42 u32 pitch, u32 flags, struct nvkm_fb_tile *tile) in nvkm_fb_tile_init() argument
44 fb->func->tile.init(fb, region, addr, size, pitch, flags, tile); in nvkm_fb_tile_init()
48 nvkm_fb_tile_prog(struct nvkm_fb *fb, int region, struct nvkm_fb_tile *tile) in nvkm_fb_tile_prog() argument
51 if (fb->func->tile.prog) { in nvkm_fb_tile_prog()
52 fb->func->tile.prog(fb, region, tile); in nvkm_fb_tile_prog()
201 for (i = 0; i < fb->tile.regions; i++) in nvkm_fb_init()
202 fb->func->tile.prog(fb, i, &fb->tile.region[i]); in nvkm_fb_init()
240 for (i = 0; i < fb->tile.regions; i++) in nvkm_fb_dtor()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/
H A Dnv44.c31 nv44_gr_tile(struct nvkm_gr *base, int i, struct nvkm_fb_tile *tile) in nv44_gr_tile() argument
44 nvkm_wr32(device, NV20_PGRAPH_TSIZE(i), tile->pitch); in nv44_gr_tile()
45 nvkm_wr32(device, NV20_PGRAPH_TLIMIT(i), tile->limit); in nv44_gr_tile()
46 nvkm_wr32(device, NV20_PGRAPH_TILE(i), tile->addr); in nv44_gr_tile()
53 nvkm_wr32(device, NV47_PGRAPH_TSIZE(i), tile->pitch); in nv44_gr_tile()
54 nvkm_wr32(device, NV47_PGRAPH_TLIMIT(i), tile->limit); in nv44_gr_tile()
55 nvkm_wr32(device, NV47_PGRAPH_TILE(i), tile->addr); in nv44_gr_tile()
56 nvkm_wr32(device, NV40_PGRAPH_TSIZE1(i), tile->pitch); in nv44_gr_tile()
57 nvkm_wr32(device, NV40_PGRAPH_TLIMIT1(i), tile->limit); in nv44_gr_tile()
58 nvkm_wr32(device, NV40_PGRAPH_TILE1(i), tile->addr); in nv44_gr_tile()
[all …]
/linux/drivers/gpu/drm/i915/gem/selftests/
H A Di915_gem_mman.c30 struct tile { struct
44 static u64 tiled_offset(const struct tile *tile, u64 v) in tiled_offset() argument
48 if (tile->tiling == I915_TILING_NONE) in tiled_offset()
51 y = div64_u64_rem(v, tile->stride, &x); in tiled_offset()
52 v = div64_u64_rem(y, tile->height, &y) * tile->stride * tile->height; in tiled_offset()
54 if (tile->tiling == I915_TILING_X) { in tiled_offset()
55 v += y * tile->width; in tiled_offset()
56 v += div64_u64_rem(x, tile->width, &x) << tile->size; in tiled_offset()
58 } else if (tile->width == 128) { in tiled_offset()
74 switch (tile->swizzle) { in tiled_offset()
[all …]
/linux/drivers/gpu/ipu-v3/
H A Dipu-image-convert.c126 struct ipu_image_tile tile[MAX_TILES]; member
677 in_tile = &in->tile[tile_idx]; in fill_tile_column()
678 out_tile = &out->tile[ctx->out_tile_map[tile_idx]]; in fill_tile_column()
709 in_tile = &in->tile[tile_idx]; in fill_tile_row()
710 out_tile = &out->tile[ctx->out_tile_map[tile_idx]]; in fill_tile_row()
870 struct ipu_image_tile *tile; in calc_tile_dimensions() local
875 tile = &image->tile[ctx->out_tile_map[i]]; in calc_tile_dimensions()
877 tile = &image->tile[i]; in calc_tile_dimensions()
879 tile->size = ((tile->height * image->fmt->bpp) >> 3) * in calc_tile_dimensions()
880 tile->width; in calc_tile_dimensions()
[all …]

1234567