| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_fence.c | 57 struct nouveau_fence_chan *fctx = nouveau_fctx(fence); in nouveau_fence_signal() local 59 if (!--fctx->notify_ref) in nouveau_fence_signal() 78 nouveau_fence_context_kill(struct nouveau_fence_chan *fctx, int error) in nouveau_fence_context_kill() argument 83 spin_lock_irqsave(&fctx->lock, flags); in nouveau_fence_context_kill() 84 list_for_each_entry_safe(fence, tmp, &fctx->pending, head) { in nouveau_fence_context_kill() 89 nvif_event_block(&fctx->event); in nouveau_fence_context_kill() 91 fctx->killed = 1; in nouveau_fence_context_kill() 92 spin_unlock_irqrestore(&fctx->lock, flags); in nouveau_fence_context_kill() 96 nouveau_fence_context_del(struct nouveau_fence_chan *fctx) in nouveau_fence_context_del() argument 98 cancel_work_sync(&fctx->uevent_work); in nouveau_fence_context_del() [all …]
|
| H A D | nv04_fence.c | 70 struct nv04_fence_chan *fctx = chan->fence; in nv04_fence_context_del() local 71 nouveau_fence_context_del(&fctx->base); in nv04_fence_context_del() 73 nouveau_fence_context_free(&fctx->base); in nv04_fence_context_del() 79 struct nv04_fence_chan *fctx = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv04_fence_context_new() local 80 if (fctx) { in nv04_fence_context_new() 81 nouveau_fence_context_new(chan, &fctx->base); in nv04_fence_context_new() 82 fctx->base.emit = nv04_fence_emit; in nv04_fence_context_new() 83 fctx->base.sync = nv04_fence_sync; in nv04_fence_context_new() 84 fctx->base.read = nv04_fence_read; in nv04_fence_context_new() 85 chan->fence = fctx; in nv04_fence_context_new()
|
| H A D | gv100_fence.c | 72 struct nv84_fence_chan *fctx; in gv100_fence_context_new() local 79 fctx = chan->fence; in gv100_fence_context_new() 80 fctx->base.emit32 = gv100_fence_emit32; in gv100_fence_context_new() 81 fctx->base.sync32 = gv100_fence_sync32; in gv100_fence_context_new()
|
| H A D | nvc0_fence.c | 82 struct nv84_fence_chan *fctx = chan->fence; in nvc0_fence_context_new() local 83 fctx->base.emit32 = nvc0_fence_emit32; in nvc0_fence_context_new() 84 fctx->base.sync32 = nvc0_fence_sync32; in nvc0_fence_context_new()
|
| /linux/drivers/crypto/cavium/nitrox/ |
| H A D | nitrox_aead.c | 38 struct flexi_crypto_context *fctx; in nitrox_aes_gcm_setkey() local 46 fctx = nctx->u.fctx; in nitrox_aes_gcm_setkey() 47 flags.fu = be64_to_cpu(fctx->flags.f); in nitrox_aes_gcm_setkey() 49 fctx->flags.f = cpu_to_be64(flags.fu); in nitrox_aes_gcm_setkey() 52 memset(&fctx->crypto, 0, sizeof(fctx->crypto)); in nitrox_aes_gcm_setkey() 53 memcpy(fctx->crypto.u.key, key, keylen); in nitrox_aes_gcm_setkey() 62 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_aead_setauthsize() local 65 flags.fu = be64_to_cpu(fctx->flags.f); in nitrox_aead_setauthsize() 67 fctx->flags.f = cpu_to_be64(flags.fu); in nitrox_aead_setauthsize() 219 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_aes_gcm_enc() local [all …]
|
| H A D | nitrox_skcipher.c | 153 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_skcipher_exit() local 155 memzero_explicit(&fctx->crypto, sizeof(struct crypto_keys)); in nitrox_skcipher_exit() 156 memzero_explicit(&fctx->auth, sizeof(struct auth_keys)); in nitrox_skcipher_exit() 171 struct flexi_crypto_context *fctx; in nitrox_skcipher_setkey() local 184 fctx = nctx->u.fctx; in nitrox_skcipher_setkey() 185 flags = &fctx->flags; in nitrox_skcipher_setkey() 192 memcpy(fctx->crypto.u.key, key, keylen); in nitrox_skcipher_setkey() 341 struct flexi_crypto_context *fctx; in nitrox_aes_xts_setkey() local 354 fctx = nctx->u.fctx; in nitrox_aes_xts_setkey() 356 memcpy(fctx->auth.u.key2, (key + keylen), keylen); in nitrox_aes_xts_setkey() [all …]
|
| H A D | nitrox_req.h | 206 struct flexi_crypto_context *fctx; member
|
| /linux/drivers/crypto/cavium/cpt/ |
| H A D | cptvf_algs.c | 101 struct fc_context *fctx = &rctx->fctx; in create_ctx_hdr() local 121 fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type; in create_ctx_hdr() 122 fctx->enc.enc_ctrl.e.aes_key = ctx->key_type; in create_ctx_hdr() 123 fctx->enc.enc_ctrl.e.iv_source = FROM_DPTR; in create_ctx_hdr() 126 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len * 2); in create_ctx_hdr() 128 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len); in create_ctx_hdr() 129 ctrl_flags = (__be64 *)&fctx->enc.enc_ctrl.flags; in create_ctx_hdr() 130 *ctrl_flags = cpu_to_be64(fctx->enc.enc_ctrl.flags); in create_ctx_hdr() 142 req_info->in[*argcnt].vptr = (u8 *)fctx; in create_ctx_hdr() 198 struct fc_context *fctx = &rctx->fctx; in cvm_enc_dec() local [all …]
|
| H A D | cptvf_algs.h | 113 struct fc_context fctx; member
|
| /linux/drivers/gpu/drm/msm/ |
| H A D | msm_fence.h | 81 void msm_fence_context_free(struct msm_fence_context *fctx); 83 bool msm_fence_completed(struct msm_fence_context *fctx, uint32_t fence); 84 void msm_update_fence(struct msm_fence_context *fctx, uint32_t fence); 87 void msm_fence_init(struct dma_fence *fence, struct msm_fence_context *fctx);
|
| H A D | msm_gpu.c | 547 msm_update_fence(ring->fctx, fence); in recover_worker() 653 } else if (fence_before(fence, ring->fctx->last_fence) && in hangcheck_handler() 663 gpu->name, ring->fctx->last_fence); in hangcheck_handler() 669 if (fence_after(ring->fctx->last_fence, ring->hangcheck_fence)) in hangcheck_handler() 873 msm_update_fence(gpu->rb[i]->fctx, gpu->rb[i]->memptrs->fence); in msm_gpu_retire()
|
| H A D | msm_gpu.h | 316 if (fence_after(ring->fctx->last_fence, ring->memptrs->fence)) in msm_gpu_active()
|
| /linux/drivers/crypto/marvell/octeontx/ |
| H A D | otx_cptvf_algs.c | 108 if (memcmp(rctx->fctx.hmac.s.hmac_calc, in validate_hmac_cipher_null() 109 rctx->fctx.hmac.s.hmac_recv, in validate_hmac_cipher_null() 241 struct otx_cpt_fc_ctx *fctx = &rctx->fctx; in create_ctx_hdr() local 273 fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type; in create_ctx_hdr() 274 fctx->enc.enc_ctrl.e.aes_key = ctx->key_type; in create_ctx_hdr() 275 fctx->enc.enc_ctrl.e.iv_source = OTX_CPT_FROM_CPTR; in create_ctx_hdr() 278 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len * 2); in create_ctx_hdr() 280 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len); in create_ctx_hdr() 282 memcpy(fctx->enc.encr_iv, req->iv, crypto_skcipher_ivsize(stfm)); in create_ctx_hdr() 284 fctx->enc.enc_ctrl.flags = cpu_to_be64(fctx->enc.enc_ctrl.cflags); in create_ctx_hdr() [all …]
|
| H A D | otx_cptvf_algs.h | 162 struct otx_cpt_fc_ctx fctx; member
|
| /linux/security/apparmor/ |
| H A D | file.c | 457 static void update_file_ctx(struct aa_file_ctx *fctx, struct aa_label *label, in update_file_ctx() argument 463 spin_lock(&fctx->lock); in update_file_ctx() 464 old = rcu_dereference_protected(fctx->label, in update_file_ctx() 465 lockdep_is_held(&fctx->lock)); in update_file_ctx() 469 rcu_assign_pointer(fctx->label, l); in update_file_ctx() 473 fctx->allow |= request; in update_file_ctx() 475 spin_unlock(&fctx->lock); in update_file_ctx() 608 struct aa_file_ctx *fctx; in aa_file_perm() local 616 fctx = file_ctx(file); in aa_file_perm() 619 flabel = rcu_dereference(fctx->label); in aa_file_perm() [all …]
|
| H A D | lsm.c | 462 struct aa_file_ctx *fctx = file_ctx(file); in apparmor_file_open() local 478 fctx->allow = MAY_EXEC | MAY_READ | AA_EXEC_MMAP; in apparmor_file_open() 497 fctx->allow = aa_map_file_to_perms(file); in apparmor_file_open()
|
| /linux/fs/ |
| H A D | userfaultfd.c | 638 struct userfaultfd_fork_ctx *fctx; in dup_userfaultfd() local 649 list_for_each_entry(fctx, fcs, list) in dup_userfaultfd() 650 if (fctx->orig == octx) { in dup_userfaultfd() 651 ctx = fctx->new; in dup_userfaultfd() 656 fctx = kmalloc(sizeof(*fctx), GFP_KERNEL); in dup_userfaultfd() 657 if (!fctx) in dup_userfaultfd() 662 kfree(fctx); in dup_userfaultfd() 679 fctx->orig = octx; in dup_userfaultfd() 680 fctx->new = ctx; in dup_userfaultfd() 681 list_add_tail(&fctx->list, fcs); in dup_userfaultfd() [all …]
|
| /linux/drivers/gpu/drm/msm/adreno/ |
| H A D | adreno_gpu.c | 671 ring->memptrs->bv_fence = ring->fctx->completed_fence; in adreno_hw_init() 677 if (fence_before(ring->fctx->last_fence, ring->memptrs->fence)) { in adreno_hw_init() 678 ring->memptrs->fence = ring->fctx->last_fence; in adreno_hw_init() 769 state->ring[i].seqno = gpu->rb[i]->fctx->last_fence; in adreno_gpu_state_get() 1043 ring->fctx->last_fence); in adreno_dump_info()
|
| H A D | a8xx_gpu.c | 993 ring ? ring->id : -1, ring ? ring->fctx->last_fence : 0, in a8xx_fault_detect_irq()
|
| H A D | a5xx_gpu.c | 1248 ring ? ring->id : -1, ring ? ring->fctx->last_fence : 0, in a5xx_fault_detect_irq()
|
| H A D | a6xx_gpu.c | 1896 ring ? ring->id : -1, ring ? ring->fctx->last_fence : 0, in a6xx_fault_detect_irq()
|
| /linux/drivers/crypto/marvell/octeontx2/ |
| H A D | otx2_cptvf_algs.h | 151 struct otx2_cpt_fc_ctx fctx; member
|
| /linux/fs/fuse/ |
| H A D | dir.c | 661 struct fuse_secctx *fctx; in get_security_context() local 684 total_len += FUSE_REC_ALIGN(sizeof(*fctx) + namesize + in get_security_context() 697 fctx = ptr; in get_security_context() 698 fctx->size = lsmctx.len; in get_security_context() 699 ptr += sizeof(*fctx); in get_security_context()
|
| /linux/drivers/net/ethernet/broadcom/ |
| H A D | cnic.c | 2356 struct fcoe_context *fctx; in cnic_bnx2x_fcoe_ofld1() local 2390 fctx = cnic_get_bnx2x_ctx(dev, cid, 1, &ctx_addr); in cnic_bnx2x_fcoe_ofld1() 2391 if (fctx) { in cnic_bnx2x_fcoe_ofld1() 2397 fctx->xstorm_ag_context.cdu_reserved = val; in cnic_bnx2x_fcoe_ofld1() 2400 fctx->ustorm_ag_context.cdu_usage = val; in cnic_bnx2x_fcoe_ofld1()
|