/linux/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | gk20a.c | 54 struct gk20a_instmem *imem; member 145 struct gk20a_instmem *imem = obj->base.imem; in gk20a_instobj_iommu_recycle_vaddr() local 151 imem->vaddr_use -= nvkm_memory_size(&obj->base.base.memory); in gk20a_instobj_iommu_recycle_vaddr() 152 nvkm_debug(&imem->base.subdev, "vaddr used: %x/%x\n", imem->vaddr_use, in gk20a_instobj_iommu_recycle_vaddr() 153 imem->vaddr_max); in gk20a_instobj_iommu_recycle_vaddr() 160 gk20a_instmem_vaddr_gc(struct gk20a_instmem *imem, const u64 size) in gk20a_instmem_vaddr_gc() argument 162 while (imem->vaddr_use + size > imem->vaddr_max) { in gk20a_instmem_vaddr_gc() 164 if (list_empty(&imem->vaddr_lru)) in gk20a_instmem_vaddr_gc() 168 list_first_entry(&imem->vaddr_lru, in gk20a_instmem_vaddr_gc() 177 struct gk20a_instmem *imem = node->imem; in gk20a_instobj_acquire_dma() local [all …]
|
H A D | nv04.c | 42 struct nv04_instmem *imem; member 50 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_wr32() 58 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_rd32() 77 struct nvkm_device *device = iobj->imem->base.subdev.device; in nv04_instobj_acquire() 103 mutex_lock(&iobj->imem->base.mutex); in nv04_instobj_dtor() 104 nvkm_mm_free(&iobj->imem->heap, &iobj->node); in nv04_instobj_dtor() 105 mutex_unlock(&iobj->imem->base.mutex); in nv04_instobj_dtor() 106 nvkm_instobj_dtor(&iobj->imem->base, &iobj->base); in nv04_instobj_dtor() 124 struct nv04_instmem *imem = nv04_instmem(base); in nv04_instobj_new() local 132 nvkm_instobj_ctor(&nv04_instobj_func, &imem->base, &iobj->base); in nv04_instobj_new() [all …]
|
/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_parser_rt.c | 115 struct ice_imem_item *imem) in ice_bst_key_init() argument 123 if (imem->b_kb.tsr_ctrl) in ice_bst_key_init() 126 key[idd] = imem->b_kb.prio; in ice_bst_key_init() 198 struct ice_imem_item *imem) in ice_imem_pgk_init() argument 201 rt->pg_key.next_proto = ice_pk_build(rt, &imem->np_kb); in ice_imem_pgk_init() 205 if (imem->pg_kb.flag0_ena) in ice_imem_pgk_init() 206 rt->pg_key.flag0 = ice_flag_get(rt, imem->pg_kb.flag0_idx); in ice_imem_pgk_init() 207 if (imem->pg_kb.flag1_ena) in ice_imem_pgk_init() 208 rt->pg_key.flag1 = ice_flag_get(rt, imem->pg_kb.flag1_idx); in ice_imem_pgk_init() 209 if (imem->pg_kb.flag2_ena) in ice_imem_pgk_init() [all …]
|
/linux/drivers/net/wwan/iosm/ |
H A D | iosm_ipc_imem.h | 387 * Returns: Initialized imem pointer on success else NULL 395 * @ipc_imem: Pointer to imem data-struct 403 * @ipc_imem: Pointer to imem data-struct 410 * @ipc_imem: Pointer to imem data-struct 416 * @ipc_imem: Pointer to imem data-struct 422 * @ipc_imem: Pointer to imem data-struct 429 * @ipc_imem: Pointer to imem instance 437 * @ipc_imem: Pointer to imem data-struct 446 * @ipc_imem: Pointer to imem data-struct 453 * @ipc_imem: Pointer to imem data-struct [all …]
|
H A D | iosm_ipc_mux.c | 13 channel_id = ipc_imem_channel_alloc(ipc_mux->imem, ipc_mux->instance_id, in ipc_mux_channel_create() 25 ipc_mux->channel = ipc_imem_channel_open(ipc_mux->imem, channel_id, in ipc_mux_channel_create() 187 ipc_imem_channel_close(ipc_mux->imem, ipc_mux->channel_id); in ipc_mux_channel_close() 223 ipc_imem_td_update_timer_suspend(ipc_mux->imem, true); in ipc_mux_schedule() 227 ipc_imem_td_update_timer_suspend(ipc_mux->imem, false); in ipc_mux_schedule() 236 ipc_imem_td_update_timer_suspend(ipc_mux->imem, true); in ipc_mux_schedule() 240 ipc_imem_td_update_timer_suspend(ipc_mux->imem, false); in ipc_mux_schedule() 279 struct iosm_imem *imem) in ipc_mux_init() argument 295 ipc_mux->pcie = imem->pcie; in ipc_mux_init() 296 ipc_mux->imem = imem; in ipc_mux_init() [all …]
|
H A D | iosm_ipc_protocol.c | 17 int index = ipc_protocol_msg_prep(ipc_protocol->imem, msg_type, in ipc_protocol_tq_msg_send() 25 ipc_protocol_msg_hp_update(ipc_protocol->imem); in ipc_protocol_tq_msg_send() 83 index = ipc_task_queue_send_task(ipc_protocol->imem, in ipc_protocol_msg_send() 100 ipc_task_queue_send_task(ipc_protocol->imem, in ipc_protocol_msg_send() 182 ipc_task_queue_send_task(ipc_protocol->imem, in ipc_protocol_suspend() 235 ipc_protocol->imem = ipc_imem; in ipc_protocol_init()
|
H A D | iosm_ipc_imem_ops.h | 38 * @ipc_imem: Imem instance. 61 * @ipc_imem: Imem instance. 71 * @ipc_imem: Imem instance. 80 * @ipc_imem: Imem instance.
|
H A D | iosm_ipc_task_queue.c | 140 int ipc_task_queue_send_task(struct iosm_imem *imem, in ipc_task_queue_send_task() argument 157 ret = ipc_task_queue_add_task(imem, arg, copy, func, in ipc_task_queue_send_task() 160 dev_err(imem->ipc_task->dev, in ipc_task_queue_send_task()
|
H A D | iosm_ipc_pcie.h | 54 * @imem: Pointer to imem data struct 69 struct iosm_imem *imem; member
|
H A D | iosm_ipc_task_queue.h | 82 * @imem: Pointer to iosm_imem struct 92 int ipc_task_queue_send_task(struct iosm_imem *imem,
|
H A D | iosm_ipc_mux.h | 317 * @imem: Pointer to iosm_imem 335 * wwan/imem layer. 351 struct iosm_imem *imem; member 387 * @ipc_imem: Pointer to imem data-struct
|
H A D | iosm_ipc_mux_codec.c | 20 ipc_imem_ul_send(ipc_mux->imem); in ipc_mux_tq_cmd_send() 28 int ret = ipc_task_queue_send_task(ipc_mux->imem, ipc_mux_tq_cmd_send, in ipc_mux_acb_send() 49 ipc_uevent_send(ipc_mux->imem->dev, UEVENT_MDM_TIMEOUT); in ipc_mux_acb_send() 252 adb_timer = &ipc_mux->imem->adb_timer; in ipc_mux_dl_cmds_decode_process() 900 ul_data_pend = ipc_imem_ul_write_td(ipc_mux->imem); in ipc_mux_ul_adb_finish() 904 ipc_imem_td_update_timer_start(ipc_mux->imem); in ipc_mux_ul_adb_finish() 1003 (void)ipc_imem_ul_write_td(ipc_mux->imem); in ipc_mux_lite_send_qlt() 1161 (void)ipc_imem_ul_write_td(ipc_mux->imem); in ipc_mux_ul_adgh_encode() 1491 ipc_imem_adb_timer_start(ipc_mux->imem); in ipc_mux_tq_ul_trigger_encode() 1494 ipc_imem_td_update_timer_start(ipc_mux->imem); in ipc_mux_tq_ul_trigger_encode() [all …]
|
/linux/drivers/remoteproc/ |
H A D | qcom_pil_info.c | 33 struct resource imem; in qcom_pil_info_init() local 45 ret = of_address_to_resource(np, 0, &imem); in qcom_pil_info_init() 50 base = ioremap(imem.start, resource_size(&imem)); in qcom_pil_info_init() 56 memset_io(base, 0, resource_size(&imem)); in qcom_pil_info_init() 59 _reloc.num_entries = (u32)resource_size(&imem) / PIL_RELOC_ENTRY_SIZE; in qcom_pil_info_init() 65 * qcom_pil_info_store() - store PIL information of image in IMEM
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/mpeg/ |
H A D | nv40.c | 33 struct nvkm_instmem *imem = device->imem; in nv40_mpeg_mthd_dma() local 37 u32 dma0 = nvkm_instmem_rd32(imem, inst + 0); in nv40_mpeg_mthd_dma() 38 u32 dma1 = nvkm_instmem_rd32(imem, inst + 4); in nv40_mpeg_mthd_dma() 39 u32 dma2 = nvkm_instmem_rd32(imem, inst + 8); in nv40_mpeg_mthd_dma()
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | nv17.c | 40 struct nvkm_memory *ramfc = chan->cgrp->runl->fifo->engine.subdev.device->imem->ramfc; in nv17_chan_ramfc_write() 96 struct nvkm_instmem *imem = device->imem; in nv17_fifo_init() local 97 struct nvkm_ramht *ramht = imem->ramht; in nv17_fifo_init() 98 struct nvkm_memory *ramro = imem->ramro; in nv17_fifo_init() 99 struct nvkm_memory *ramfc = imem->ramfc; in nv17_fifo_init()
|
/linux/Documentation/devicetree/bindings/clock/ |
H A D | tesla,fsd-clock.yaml | 25 - tesla,fsd-clock-imem 65 const: tesla,fsd-clock-imem 71 - description: IMEM TCU clock (from CMU_CMU) 72 - description: IMEM bus clock (from CMU_CMU) 73 - description: IMEM DMA clock (from CMU_CMU)
|
H A D | samsung,exynos5433-clock.yaml | 27 # IMEM/FSYS/G3D/GSCL/HEVC/MSCL/G2D/MFC/PERIC/PERIS domains and bus 69 - samsung,exynos5433-cmu-imem 483 const: samsung,exynos5433-cmu-imem
|
/linux/Documentation/devicetree/bindings/remoteproc/ |
H A D | qcom,pil-info.yaml | 13 The Qualcomm peripheral image loader relocation memory region, in IMEM, is 32 imem@146bf000 { 33 compatible = "qcom,sdm630-imem", "syscon", "simple-mfd";
|
/linux/drivers/memory/ |
H A D | brcmstb_dpfe.c | 150 * 3 IMEM byte size 152 * IMEM 184 void __iomem *imem; member 571 u32 __iomem *imem = priv->imem; in __verify_fw_checksum() local 590 sum += readl_relaxed(imem + i); in __verify_fw_checksum() 626 const u32 *dmem, *imem; in brcmstb_dpfe_download_firmware() local 675 /* IMEM comes right after the header. */ in brcmstb_dpfe_download_firmware() 676 imem = fw_blob; in brcmstb_dpfe_download_firmware() 677 /* DMEM follows after IMEM. */ in brcmstb_dpfe_download_firmware() 683 ret = __write_firmware(priv->imem, imem, imem_size, is_big_endian); in brcmstb_dpfe_download_firmware() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | memory.c | 141 struct nvkm_instmem *imem = device->imem; in nvkm_memory_new() local 146 if (unlikely(!imem)) in nvkm_memory_new() 159 ret = nvkm_instobj_new(imem, size, align, zero, preserve, &memory); in nvkm_memory_new()
|
/linux/Documentation/devicetree/bindings/memory-controllers/ |
H A D | brcm,dpfe-cpu.yaml | 31 - const: dpfe-imem 47 reg-names = "dpfe-cpu", "dpfe-dmem", "dpfe-imem";
|
/linux/Documentation/devicetree/bindings/dma/ |
H A D | st_fdma.txt | 15 - reg-names : Must contain "slimcore", "dmem", "peripherals", "imem" entries 31 reg-names = "slimcore", "dmem", "peripherals", "imem";
|
/linux/drivers/gpu/drm/nouveau/include/nvkm/core/ |
H A D | falcon.h | 7 IMEM, enumerator 16 case IMEM: return "imem"; in nvkm_falcon_mem()
|
/linux/drivers/net/wireless/realtek/rtlwifi/rtl8192se/ |
H A D | fw.c | 38 /* Polling IMEM Ready after CPU has refilled. */ in _rtl92s_firmware_enable_cpu() 43 "IMEM Ready after CPU has refilled\n"); in _rtl92s_firmware_enable_cpu() 205 /* Polling IMEM code done. */ in _rtl92s_firmware_checkready() 346 /* 2. Retrieve IMEM image. */ in rtl92s_download_fw() 349 pr_err("memory for data image is less than IMEM required\n"); in rtl92s_download_fw()
|
/linux/drivers/gpu/drm/nouveau/nvkm/falcon/ |
H A D | ga102.c | 52 if (mem_type == IMEM) in ga102_flcn_dma_init() 136 IMEM, fw->imem_base, fw->imem_size, true); in ga102_flcn_fw_load()
|