| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_sriov_vf_ccs.c | 151 ctx->ctx_id ? "Restore" : "Save", bb_pool_size / SZ_1M); in alloc_bb_pool() 210 enum xe_sriov_vf_ccs_rw_ctxs ctx_id; in xe_sriov_vf_ccs_rebase() local 215 for_each_ccs_rw_ctx(ctx_id) { in xe_sriov_vf_ccs_rebase() 217 &xe->sriov.vf.ccs.contexts[ctx_id]; in xe_sriov_vf_ccs_rebase() 227 switch (ctx->ctx_id) { in register_save_restore_context() 253 enum xe_sriov_vf_ccs_rw_ctxs ctx_id; in xe_sriov_vf_ccs_register_context() local 259 for_each_ccs_rw_ctx(ctx_id) { in xe_sriov_vf_ccs_register_context() 260 ctx = &xe->sriov.vf.ccs.contexts[ctx_id]; in xe_sriov_vf_ccs_register_context() 333 enum xe_sriov_vf_ccs_rw_ctxs ctx_id; in xe_sriov_vf_ccs_init() local 346 for_each_ccs_rw_ctx(ctx_id) { in xe_sriov_vf_ccs_init() [all …]
|
| H A D | xe_execlist.c | 44 u32 ctx_id) in __start_lrc() argument 55 xe_gt_assert(hwe->gt, FIELD_FIT(XEHP_SW_CTX_ID, ctx_id)); in __start_lrc() 56 lrc_desc |= FIELD_PREP(XEHP_SW_CTX_ID, ctx_id); in __start_lrc() 58 xe_gt_assert(hwe->gt, FIELD_FIT(SW_CTX_ID, ctx_id)); in __start_lrc() 59 lrc_desc |= FIELD_PREP(SW_CTX_ID, ctx_id); in __start_lrc()
|
| /linux/include/linux/sunrpc/ |
| H A D | gss_api.h | 51 struct gss_ctx **ctx_id, 55 struct gss_ctx *ctx_id, 59 struct gss_ctx *ctx_id, 63 struct gss_ctx *ctx_id, 68 struct gss_ctx *ctx_id, 73 struct gss_ctx **ctx_id); 113 struct gss_ctx *ctx_id, 117 struct gss_ctx *ctx_id, 121 struct gss_ctx *ctx_id, 125 struct gss_ctx *ctx_id, [all …]
|
| /linux/net/sunrpc/auth_gss/ |
| H A D | gss_mech_switch.c | 354 struct gss_ctx **ctx_id, in gss_import_sec_context() argument 358 if (!(*ctx_id = kzalloc_obj(**ctx_id, gfp_mask))) in gss_import_sec_context() 360 (*ctx_id)->mech_type = gss_mech_get(mech); in gss_import_sec_context() 363 *ctx_id, endtime, gfp_mask); in gss_import_sec_context() 407 gss_wrap(struct gss_ctx *ctx_id, in gss_wrap() argument 412 return ctx_id->mech_type->gm_ops in gss_wrap() 413 ->gss_wrap(ctx_id, offset, buf, inpages); in gss_wrap() 417 gss_unwrap(struct gss_ctx *ctx_id, in gss_unwrap() argument 422 return ctx_id->mech_type->gm_ops in gss_unwrap() 423 ->gss_unwrap(ctx_id, offset, len, buf); in gss_unwrap()
|
| /linux/drivers/media/platform/mediatek/vcodec/common/ |
| H A D | mtk_vcodec_intr.c | 19 int ctx_id, ctx_type, status = 0; in mtk_vcodec_wait_for_done_ctx() local 28 ctx_id = ctx->id; in mtk_vcodec_wait_for_done_ctx() 38 ctx_id = ctx->id; in mtk_vcodec_wait_for_done_ctx() 54 ctx_id, command, ctx_type, timeout_ms, in mtk_vcodec_wait_for_done_ctx() 59 ctx_id, command, ctx_type, in mtk_vcodec_wait_for_done_ctx()
|
| H A D | mtk_vcodec_dbgfs.h | 54 void mtk_vcodec_dbgfs_remove(struct mtk_vcodec_dec_dev *vcodec_dev, int ctx_id); 62 static inline void mtk_vcodec_dbgfs_remove(struct mtk_vcodec_dec_dev *vcodec_dev, int ctx_id) in mtk_vcodec_dbgfs_remove() argument
|
| /linux/tools/testing/selftests/drivers/net/hw/ |
| H A D | rss_ctx.py | 109 def _ntuple_rule_check(cfg, rule_id, ctx_id): argument 112 pattern = f"RSS Context (ID: )?{ctx_id}" 211 ctx_id = 0 214 ctx_id = ethtool_create(cfg, "-X", "context new") 215 ctx_ref = f"context {ctx_id}" 219 data = get_rss(cfg, context=ctx_id) 240 flow = f"flow-type tcp{cfg.addr_ipver} dst-ip {cfg.addr} dst-port {port} context {ctx_id}" 249 data = get_rss(cfg, context=ctx_id) 261 raise Exception(f"Driver didn't prevent us from deactivating a used queue (context {ctx_id})") 265 flow = f"flow-type tcp{cfg.addr_ipver} dst-ip {cfg.addr} dst-port {port} context {ctx_id} actio [all...] |
| H A D | rss_api.py | 125 ctx_id = _ethtool_create(cfg, "-X", "context new") 129 "context": ctx_id, "indir": [1]}) 131 "context": ctx_id}) 138 "context": ctx_id, "indir": [0, 1]}) 140 "context": ctx_id}) 188 ctx_id = _ethtool_create(cfg, "-X", "context new") 200 ksft_eq(ntf["msg"].get("context"), ctx_id) 417 ctx_id = ctx.get("context") 420 "context": ctx_id, 424 ksft_eq(ctx.get("context"), ctx_id) [all …]
|
| H A D | iou-zcrx.py | 35 _set_flow_rule_rss(cfg, port, ctx_id) global() argument
|
| /linux/sound/soc/fsl/ |
| H A D | fsl_easrc.c | 240 unsigned int ctx_id, int mem_type) in fsl_easrc_coeff_mem_ptr_reset() argument 253 if (ctx_id >= EASRC_CTX_MAX_NUM) { in fsl_easrc_coeff_mem_ptr_reset() 254 dev_err(dev, "Invalid context id[%d]\n", ctx_id); in fsl_easrc_coeff_mem_ptr_reset() 258 reg = REG_EASRC_CCE1(ctx_id); in fsl_easrc_coeff_mem_ptr_reset() 428 static int fsl_easrc_write_pf_coeff_mem(struct fsl_asrc *easrc, int ctx_id, in fsl_easrc_write_pf_coeff_mem() argument 450 ret = fsl_easrc_coeff_mem_ptr_reset(easrc, ctx_id, EASRC_PF_COEFF_MEM); in fsl_easrc_write_pf_coeff_mem() 460 regmap_write(easrc->regmap, REG_EASRC_PCF(ctx_id), in fsl_easrc_write_pf_coeff_mem() 462 regmap_write(easrc->regmap, REG_EASRC_PCF(ctx_id), in fsl_easrc_write_pf_coeff_mem() 470 unsigned int ctx_id) in fsl_easrc_prefilter_config() argument 488 if (ctx_id >= EASRC_CTX_MAX_NUM) { in fsl_easrc_prefilter_config() [all …]
|
| /linux/drivers/media/platform/st/sti/hva/ |
| H A D | hva-hw.c | 120 u8 ctx_id = 0; in hva_hw_its_irq_thread() local 130 ctx_id = (hva->sts_reg & 0xFF00) >> 8; in hva_hw_its_irq_thread() 131 if (ctx_id >= HVA_MAX_INSTANCES) { in hva_hw_its_irq_thread() 133 HVA_PREFIX, __func__, ctx_id); in hva_hw_its_irq_thread() 137 ctx = hva->instances[ctx_id]; in hva_hw_its_irq_thread() 223 u8 ctx_id = 0; in hva_hw_err_irq_thread() local 233 ctx_id = (hva->sts_reg & 0xFF00) >> 8; in hva_hw_err_irq_thread() 234 if (ctx_id >= HVA_MAX_INSTANCES) { in hva_hw_err_irq_thread() 236 ctx_id); in hva_hw_err_irq_thread() 240 ctx = hva->instances[ctx_id]; in hva_hw_err_irq_thread()
|
| /linux/drivers/gpu/drm/imx/dcss/ |
| H A D | dcss-ctxld.c | 332 void dcss_ctxld_write_irqsafe(struct dcss_ctxld *ctxld, u32 ctx_id, u32 val, in dcss_ctxld_write_irqsafe() argument 341 int item_idx = ctxld->ctx_size[curr_ctx][ctx_id]; in dcss_ctxld_write_irqsafe() 343 if (item_idx + 1 > dcss_ctxld_ctx_size[ctx_id]) { in dcss_ctxld_write_irqsafe() 348 ctx[ctx_id][item_idx].val = val; in dcss_ctxld_write_irqsafe() 349 ctx[ctx_id][item_idx].ofs = reg_ofs; in dcss_ctxld_write_irqsafe() 350 ctxld->ctx_size[curr_ctx][ctx_id] += 1; in dcss_ctxld_write_irqsafe() 353 void dcss_ctxld_write(struct dcss_ctxld *ctxld, u32 ctx_id, in dcss_ctxld_write() argument 357 dcss_ctxld_write_irqsafe(ctxld, ctx_id, val, reg_ofs); in dcss_ctxld_write()
|
| H A D | dcss-ss.c | 68 u32 ctx_id; member 78 dcss_ctxld_write(ss->ctxld, ss->ctx_id, val, in dcss_ss_write() 101 ss->ctx_id = CTX_SB_HP; in dcss_ss_init()
|
| /linux/drivers/accel/habanalabs/common/ |
| H A D | command_buffer.c | 104 int ctx_id, bool internal_cb) in hl_cb_alloc() argument 118 if (ctx_id == HL_KERNEL_ASID_ID && !hdev->disabled) in hl_cb_alloc() 137 } else if (ctx_id == HL_KERNEL_ASID_ID) { in hl_cb_alloc() 186 int rc, ctx_id = cb_args->ctx->asid; in hl_cb_mmap_mem_alloc() local 194 if (ctx_id == HL_KERNEL_ASID_ID && in hl_cb_mmap_mem_alloc() 212 cb = hl_cb_alloc(cb_args->hdev, cb_args->cb_size, ctx_id, cb_args->internal_cb); in hl_cb_mmap_mem_alloc() 226 if (ctx_id == HL_KERNEL_ASID_ID) { in hl_cb_mmap_mem_alloc() 278 int ctx_id = ctx->asid; in hl_cb_create() local 280 if ((hdev->disabled) || (hdev->reset_info.in_reset && (ctx_id != HL_KERNEL_ASID_ID))) { in hl_cb_create() 294 ctx_id == HL_KERNEL_ASID_ID ? GFP_ATOMIC : GFP_KERNEL, &args); in hl_cb_create()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_sched.c | 63 unsigned ctx_id, in amdgpu_sched_context_priority_override() argument 78 ctx = amdgpu_ctx_get(fpriv, ctx_id); in amdgpu_sched_context_priority_override() 120 args->in.ctx_id, in amdgpu_sched_ioctl()
|
| /linux/drivers/accel/ivpu/ |
| H A D | ivpu_jsm_msg.c | 95 int ivpu_jsm_register_db(struct ivpu_device *vdev, u32 ctx_id, u32 db_id, in ivpu_jsm_register_db() argument 105 req.payload.register_db.host_ssid = ctx_id; in ivpu_jsm_register_db() 286 int ivpu_jsm_hws_create_cmdq(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_group, u32 cmdq_id, in ivpu_jsm_hws_create_cmdq() argument 293 req.payload.hws_create_cmdq.host_ssid = ctx_id; in ivpu_jsm_hws_create_cmdq() 309 int ivpu_jsm_hws_destroy_cmdq(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_id) in ivpu_jsm_hws_destroy_cmdq() argument 315 req.payload.hws_destroy_cmdq.host_ssid = ctx_id; in ivpu_jsm_hws_destroy_cmdq() 326 int ivpu_jsm_hws_register_db(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_id, u32 db_id, in ivpu_jsm_hws_register_db() argument 334 req.payload.hws_register_db.host_ssid = ctx_id; in ivpu_jsm_hws_register_db() 368 int ivpu_jsm_hws_set_context_sched_properties(struct ivpu_device *vdev, u32 ctx_id, u32 cmdq_id, in ivpu_jsm_hws_set_context_sched_properties() argument 375 req.payload.hws_set_context_sched_properties.host_ssid = ctx_id; in ivpu_jsm_hws_set_context_sched_properties()
|
| /linux/drivers/net/ethernet/netronome/nfp/flower/ |
| H A D | offload.c | 1501 u32 ctx_id = be32_to_cpu(link->sub_flow.flow->meta.host_ctx_id); in nfp_flower_remove_merge_flow() local 1503 parent_ctx = (parent_ctx << 32) | (u64)(ctx_id); in nfp_flower_remove_merge_flow() 1637 u32 ctx_id; in __nfp_flower_update_merge_stats() local 1639 ctx_id = be32_to_cpu(merge_flow->meta.host_ctx_id); in __nfp_flower_update_merge_stats() 1640 pkts = priv->stats[ctx_id].pkts; in __nfp_flower_update_merge_stats() 1644 bytes = priv->stats[ctx_id].bytes; in __nfp_flower_update_merge_stats() 1645 used = priv->stats[ctx_id].used; in __nfp_flower_update_merge_stats() 1648 priv->stats[ctx_id].pkts = 0; in __nfp_flower_update_merge_stats() 1649 priv->stats[ctx_id].bytes = 0; in __nfp_flower_update_merge_stats() 1657 ctx_id = be32_to_cpu(sub_flow->meta.host_ctx_id); in __nfp_flower_update_merge_stats() [all …]
|
| H A D | conntrack.c | 1994 u32 ctx_id; in nfp_fl_ct_sub_stats() local 2000 ctx_id = be32_to_cpu(nfp_flow->meta.host_ctx_id); in nfp_fl_ct_sub_stats() 2001 *m_pkts += priv->stats[ctx_id].pkts; in nfp_fl_ct_sub_stats() 2002 *m_bytes += priv->stats[ctx_id].bytes; in nfp_fl_ct_sub_stats() 2003 *m_used = max_t(u64, *m_used, priv->stats[ctx_id].used); in nfp_fl_ct_sub_stats() 2014 priv->stats[ctx_id].bytes, in nfp_fl_ct_sub_stats() 2015 priv->stats[ctx_id].pkts, in nfp_fl_ct_sub_stats() 2016 0, priv->stats[ctx_id].used, in nfp_fl_ct_sub_stats() 2021 priv->stats[ctx_id].bytes, in nfp_fl_ct_sub_stats() 2022 priv->stats[ctx_id].pkts, in nfp_fl_ct_sub_stats() [all …]
|
| H A D | metadata.c | 121 u32 ctx_id; in nfp_flower_rx_flow_stats() local 129 ctx_id = be32_to_cpu(stats->stats_con_id); in nfp_flower_rx_flow_stats() 130 priv->stats[ctx_id].pkts += be32_to_cpu(stats->pkt_count); in nfp_flower_rx_flow_stats() 131 priv->stats[ctx_id].bytes += be64_to_cpu(stats->byte_count); in nfp_flower_rx_flow_stats() 132 priv->stats[ctx_id].used = jiffies; in nfp_flower_rx_flow_stats() 435 nfp_flower_get_fl_payload_from_ctx(struct nfp_app *app, u32 ctx_id) in nfp_flower_get_fl_payload_from_ctx() argument 440 ctx_entry = rhashtable_lookup_fast(&priv->stats_ctx_table, &ctx_id, in nfp_flower_get_fl_payload_from_ctx()
|
| /linux/fs/ |
| H A D | aio.c | 1074 static struct kioctx *lookup_ioctx(unsigned long ctx_id) in lookup_ioctx() argument 1076 struct aio_ring __user *ring = (void __user *)ctx_id; in lookup_ioctx() 1093 if (ctx && ctx->user_id == ctx_id) { in lookup_ioctx() 2081 SYSCALL_DEFINE3(io_submit, aio_context_t, ctx_id, long, nr, in SYSCALL_DEFINE3() argument 2092 ctx = lookup_ioctx(ctx_id); in SYSCALL_DEFINE3() 2123 COMPAT_SYSCALL_DEFINE3(io_submit, compat_aio_context_t, ctx_id, in COMPAT_SYSCALL_DEFINE3() argument 2134 ctx = lookup_ioctx(ctx_id); in COMPAT_SYSCALL_DEFINE3() 2175 SYSCALL_DEFINE3(io_cancel, aio_context_t, ctx_id, struct iocb __user *, iocb, in SYSCALL_DEFINE3() argument 2189 ctx = lookup_ioctx(ctx_id); in SYSCALL_DEFINE3() 2217 static long do_io_getevents(aio_context_t ctx_id, in do_io_getevents() argument [all …]
|
| /linux/drivers/gpu/drm/virtio/ |
| H A D | virtgpu_vq.c | 1091 cmd_p->hdr.ctx_id = cpu_to_le32(id); in virtio_gpu_cmd_context_create() 1108 cmd_p->hdr.ctx_id = cpu_to_le32(id); in virtio_gpu_cmd_context_destroy() 1113 uint32_t ctx_id, in virtio_gpu_cmd_context_attach_resource() argument 1125 cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); in virtio_gpu_cmd_context_attach_resource() 1131 uint32_t ctx_id, in virtio_gpu_cmd_context_detach_resource() argument 1143 cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); in virtio_gpu_cmd_context_detach_resource() 1182 uint32_t ctx_id, in virtio_gpu_cmd_transfer_to_host_3d() argument 1205 cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); in virtio_gpu_cmd_transfer_to_host_3d() 1217 uint32_t ctx_id, in virtio_gpu_cmd_transfer_from_host_3d() argument 1235 cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); in virtio_gpu_cmd_transfer_from_host_3d() [all …]
|
| H A D | virtgpu_ioctl.c | 46 virtio_gpu_cmd_context_create(vgdev, vfpriv->ctx_id, in virtio_gpu_create_context_locked() 54 virtio_gpu_cmd_context_create(vgdev, vfpriv->ctx_id, in virtio_gpu_create_context_locked() 266 (vgdev, vfpriv->ctx_id, offset, args->level, args->stride, in virtio_gpu_transfer_from_host_ioctl() 326 vfpriv ? vfpriv->ctx_id : 0, offset, args->level, in virtio_gpu_transfer_to_host_ioctl() 478 params->ctx_id = vfpriv->ctx_id; in verify_blob() 527 vfpriv->ctx_id, NULL, NULL); in virtio_gpu_resource_create_blob_ioctl()
|
| /linux/drivers/net/ethernet/qlogic/qlcnic/ |
| H A D | qlcnic_83xx_hw.h | 172 u16 ctx_id; member 178 u16 ctx_id; 194 u16 ctx_id; member 196 u16 ctx_id; 238 u16 ctx_id; member 244 u16 ctx_id; member
|
| /linux/include/uapi/drm/ |
| H A D | amdgpu_drm.h | 299 __u32 ctx_id; member 306 __u32 ctx_id; member 614 __u32 ctx_id; member 764 __u32 ctx_id; member 778 __u32 ctx_id; member 967 __u32 ctx_id; member 1033 __u32 ctx_id; member
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_cmdbuf.c | 1000 int ctx_id, in vmw_cmdbuf_reserve_cur() argument 1012 ctx_id != cur->cb_header->dxContext))) in vmw_cmdbuf_reserve_cur() 1027 if (ctx_id != SVGA3D_INVALID_ID) { in vmw_cmdbuf_reserve_cur() 1029 cur->cb_header->dxContext = ctx_id; in vmw_cmdbuf_reserve_cur() 1074 int ctx_id, bool interruptible, in vmw_cmdbuf_reserve() argument 1078 return vmw_cmdbuf_reserve_cur(man, size, ctx_id, interruptible); in vmw_cmdbuf_reserve() 1083 if (ctx_id != SVGA3D_INVALID_ID) { in vmw_cmdbuf_reserve() 1085 header->cb_header->dxContext = ctx_id; in vmw_cmdbuf_reserve()
|