| /linux/fs/btrfs/ |
| H A D | space-info.c | 547 struct reserve_ticket *ticket, int error) in remove_ticket() argument 551 if (!list_empty(&ticket->list)) { in remove_ticket() 552 list_del_init(&ticket->list); in remove_ticket() 553 ASSERT(space_info->reclaim_size >= ticket->bytes, in remove_ticket() 555 space_info->reclaim_size, ticket->bytes); in remove_ticket() 556 space_info->reclaim_size -= ticket->bytes; in remove_ticket() 559 spin_lock(&ticket->lock); in remove_ticket() 565 if (error && ticket->bytes > 0) in remove_ticket() 566 ticket->error = error; in remove_ticket() 568 ticket->bytes = 0; in remove_ticket() [all …]
|
| /linux/net/rxrpc/ |
| H A D | key.c | 96 memcpy(&token->kad->ticket, &xdr[8], tktlen); in rxrpc_preparse_xdr_rxkad() 110 token->kad->ticket[0], token->kad->ticket[1], in rxrpc_preparse_xdr_rxkad() 111 token->kad->ticket[2], token->kad->ticket[3], in rxrpc_preparse_xdr_rxkad() 112 token->kad->ticket[4], token->kad->ticket[5], in rxrpc_preparse_xdr_rxkad() 113 token->kad->ticket[6], token->kad->ticket[7]); in rxrpc_preparse_xdr_rxkad() 172 const __be32 *ticket, *key; in rxrpc_preparse_xdr_yfs_rxgk() local 190 ticket = xdr + (6 * 2 + 1 + (keylen / 4) + 1); in rxrpc_preparse_xdr_yfs_rxgk() 191 tktlen = ntohl(ticket[-1]); in rxrpc_preparse_xdr_yfs_rxgk() 226 token->rxgk->ticket.len = ntohl(ticket[-1]); in rxrpc_preparse_xdr_yfs_rxgk() 239 token->rxgk->ticket.data = kzalloc(round_up(token->rxgk->ticket.len, 4), in rxrpc_preparse_xdr_yfs_rxgk() [all …]
|
| H A D | rxgk_app.c | 51 void *payload, *ticket; in rxgk_yfs_decode_ticket() local 86 ticket = payload + pre_ticket_len; in rxgk_yfs_decode_ticket() 87 ret = skb_copy_bits(skb, ticket_offset, ticket, ticket_len); in rxgk_yfs_decode_ticket() 106 t = ticket + sizeof(__be32) * 2 + xdr_round_up(klen); in rxgk_yfs_decode_ticket() 125 memcpy(q, ticket + sizeof(__be32) * 2, klen); in rxgk_yfs_decode_ticket() 129 if (WARN_ON((unsigned long)q != (unsigned long)ticket)) { in rxgk_yfs_decode_ticket()
|
| H A D | rxkad.c | 901 ret = skb_store_bits(response, offset, token->kad->ticket, in rxkad_respond_to_challenge() 951 void *ticket, size_t ticket_len, in rxkad_decrypt_ticket() argument 969 ASSERTCMP((unsigned long) ticket & 7UL, ==, 0); in rxkad_decrypt_ticket() 977 sg_init_one(&sg[0], ticket, ticket_len); in rxkad_decrypt_ticket() 983 p = ticket; in rxkad_decrypt_ticket() 1118 void *ticket; in rxkad_verify_response() local 1179 ticket = kmalloc(ticket_len, GFP_NOFS); in rxkad_verify_response() 1180 if (!ticket) in rxkad_verify_response() 1184 ticket, ticket_len) < 0) { in rxkad_verify_response() 1190 ret = rxkad_decrypt_ticket(conn, server_key, skb, ticket, ticket_len, in rxkad_verify_response() [all …]
|
| H A D | rxgk.c | 811 h.ticket_len = htonl(gk->key->ticket.len); in rxgk_insert_response_header() 934 8 + (4 + xdr_round_up(gk->key->ticket.len)) + (4 + authx_len); in rxgk_construct_response() 948 ret = skb_store_bits(response, offset, gk->key->ticket.data, gk->key->ticket.len); in rxgk_construct_response() 951 offset += gk->key->ticket.len; in rxgk_construct_response() 952 ret = rxgk_pad_out(response, gk->key->ticket.len, offset); in rxgk_construct_response()
|
| /linux/fs/xfs/ |
| H A D | xfs_log.c | 26 struct xlog_ticket *ticket; member 60 struct xlog_ticket *ticket); 472 struct xlog_ticket *ticket) in xlog_state_release_iclog() argument 518 xlog_sync(log, iclog, ticket); in xlog_state_release_iclog() 806 struct xlog_ticket *ticket) in xlog_write_one_vec() argument 816 ticket->t_curr_res -= lv.lv_bytes; in xlog_write_one_vec() 819 return xlog_write(log, ctx, &lv_chain, ticket, lv.lv_bytes); in xlog_write_one_vec() 830 struct xlog_ticket *ticket) in xlog_write_unmount_record() argument 838 .oh_tid = cpu_to_be32(ticket->t_tid), in xlog_write_unmount_record() 851 return xlog_write_one_vec(log, NULL, ®, ticket); in xlog_write_unmount_record() [all …]
|
| H A D | xfs_log_cil.c | 129 ctx->ticket->t_curr_res += cilpcp->space_reserved; in xlog_cil_push_pcp_aggregate() 207 log->l_cilp->xc_ctx->ticket = xlog_cil_ticket_alloc(log); in xlog_cil_init_post_recovery() 685 ctx_res = ctx->ticket->t_unit_res; in xlog_cil_insert_items() 1175 return xlog_write(log, ctx, &ctx->lv_chain, ctx->ticket, chain_len); in xlog_cil_write_chain() 1191 .oh_tid = cpu_to_be32(ctx->ticket->t_tid), in xlog_cil_write_commit_record() 1207 error = xlog_write_one_vec(log, ctx, ®, ctx->ticket); in xlog_cil_write_commit_record() 1238 struct xlog_ticket *tic = ctx->ticket; in xlog_cil_build_trans_hdr() 1392 struct xlog_ticket *ticket; in xlog_cil_push_work() local 1395 new_ctx->ticket = xlog_cil_ticket_alloc(log); in xlog_cil_push_work() 1534 ticket = ctx->ticket; in xlog_cil_push_work() [all …]
|
| /linux/include/drm/ttm/ |
| H A D | ttm_execbuf_util.h | 63 void ttm_eu_backoff_reservation(struct ww_acquire_ctx *ticket, 99 int ttm_eu_reserve_buffers(struct ww_acquire_ctx *ticket, 115 void ttm_eu_fence_buffer_objects(struct ww_acquire_ctx *ticket,
|
| /linux/include/asm-generic/ |
| H A D | ticket_spinlock.h | 36 u16 ticket = val >> 16; in ticket_spin_lock() local 38 if (ticket == (u16)val) in ticket_spin_lock() 49 atomic_cond_read_acquire(&lock->val, ticket == (u16)VAL); in ticket_spin_lock()
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_validation.h | 62 struct ww_acquire_ctx ticket; member 122 return ttm_eu_reserve_buffers(&ctx->ticket, &ctx->bo_list, intr, in vmw_validation_bo_reserve() 138 ttm_eu_fence_buffer_objects(&ctx->ticket, &ctx->bo_list, in vmw_validation_bo_fence()
|
| /linux/drivers/gpu/drm/etnaviv/ |
| H A D | etnaviv_gem_submit.c | 121 struct ww_acquire_ctx *ticket) in submit_lock_objects() argument 135 ret = dma_resv_lock_interruptible(obj->resv, ticket); in submit_lock_objects() 145 ww_acquire_done(ticket); in submit_lock_objects() 162 ret = dma_resv_lock_slow_interruptible(obj->resv, ticket); in submit_lock_objects() 423 struct ww_acquire_ctx ticket; in etnaviv_ioctl_gem_submit() local 515 ww_acquire_init(&ticket, &reservation_ww_class); in etnaviv_ioctl_gem_submit() 580 ret = submit_lock_objects(submit, &ticket); in etnaviv_ioctl_gem_submit() 624 ww_acquire_fini(&ticket); in etnaviv_ioctl_gem_submit()
|
| /linux/drivers/gpu/drm/ttm/ |
| H A D | ttm_bo.c | 562 struct ww_acquire_ctx *ticket, in ttm_bo_evict_alloc() argument 571 .ticket = ticket, in ttm_bo_evict_alloc() 589 if (lret || !ticket) in ttm_bo_evict_alloc() 600 evict_walk.walk.arg.ticket = ticket; in ttm_bo_evict_alloc() 718 struct ww_acquire_ctx *ticket; in ttm_bo_alloc_resource() local 721 ticket = dma_resv_locking_ctx(bo->base.resv); in ttm_bo_alloc_resource() 753 ticket, res, limit_pool); in ttm_bo_alloc_resource()
|
| H A D | ttm_bo_util.c | 848 ret = dma_resv_lock_interruptible(bo->base.resv, arg->ticket); in ttm_lru_walk_ticketlock() 850 ret = dma_resv_lock(bo->base.resv, arg->ticket); in ttm_lru_walk_ticketlock() 860 arg->ticket = NULL; in ttm_lru_walk_ticketlock() 1002 else if (!arg->ticket || arg->ctx->no_wait_gpu || arg->trylock_only) in __ttm_bo_lru_cursor_next()
|
| /linux/arch/mips/include/asm/octeon/ |
| H A D | cvmx-cmd-queue.h | 154 uint64_t ticket[(CVMX_CMD_QUEUE_END >> 16) * 256]; member 278 …[ticket_ptr] "=" GCC_OFF_SMALL_ASM()(__cvmx_cmd_queue_state_ptr->ticket[__cvmx_cmd_queue_get_index… in __cvmx_cmd_queue_lock() 279 [now_serving] "=m"(qptr->now_serving), [ticket] "=r"(tmp), in __cvmx_cmd_queue_lock()
|
| /linux/tools/tracing/latency/ |
| H A D | latency-collector.c | 185 int ticket; member 666 req->ticket = r; in printstate_next_ticket() 675 if (req->ticket > printstate.ticket_completed) in printstate_mark_req_completed() 676 printstate.ticket_completed = req->ticket; in printstate_mark_req_completed() 682 return (printstate.ticket_counter != req->ticket); in printstate_has_new_req_arrived() 764 int diff = req->ticket - req->ticket_completed_ref; in table_get_probability() 1095 sec, us, req->ticket, str); in __print_skipmessage() 1098 sec, us, req->ticket); in __print_skipmessage() 1166 sec, us, req->ticket, slept_ms); in print_tracefile() 1170 us, req->ticket); in print_tracefile()
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_gem.c | 1628 bool (*shrink)(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket), in drm_gem_lru_scan() argument 1629 struct ww_acquire_ctx *ticket) in drm_gem_lru_scan() 1662 if (ticket) in drm_gem_lru_scan() 1663 ww_acquire_init(ticket, &reservation_ww_class); in drm_gem_lru_scan() 1670 if (!ww_mutex_trylock(&obj->resv->lock, ticket)) { in drm_gem_lru_scan() 1675 if (shrink(obj, ticket)) { in drm_gem_lru_scan() 1689 if (ticket) in drm_gem_lru_scan() 1690 ww_acquire_fini(ticket); in drm_gem_lru_scan()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_dma_buf.c | 470 struct ww_acquire_ctx *ticket = dma_resv_locking_ctx(obj->resv); in amdgpu_dma_buf_move_notify() local 495 if (ticket) { in amdgpu_dma_buf_move_notify() 500 r = dma_resv_lock(resv, ticket); in amdgpu_dma_buf_move_notify()
|
| H A D | amdgpu_vm.h | 513 struct ww_acquire_ctx *ticket, 524 struct ww_acquire_ctx *ticket); 570 void amdgpu_vm_bo_trace_cs(struct amdgpu_vm *vm, struct ww_acquire_ctx *ticket);
|
| H A D | amdgpu_vm.c | 593 struct ww_acquire_ctx *ticket, in amdgpu_vm_validate() argument 632 while (ticket && !list_empty(&vm->evicted_user)) { in amdgpu_vm_validate() 1609 struct ww_acquire_ctx *ticket) in amdgpu_vm_handle_moved() argument 1640 } else if (ticket && dma_resv_locking_ctx(resv) == ticket) { in amdgpu_vm_handle_moved() 2152 void amdgpu_vm_bo_trace_cs(struct amdgpu_vm *vm, struct ww_acquire_ctx *ticket) in amdgpu_vm_bo_trace_cs() argument 2166 ticket) in amdgpu_vm_bo_trace_cs()
|
| H A D | amdgpu_amdkfd_gpuvm.c | 482 struct ww_acquire_ctx *ticket) in vm_validate_pt_pd_bos() argument 488 ret = amdgpu_vm_validate(adev, vm, ticket, in vm_validate_pt_pd_bos() 1352 struct ww_acquire_ctx *ticket) in process_validate_vms() argument 1359 ret = vm_validate_pt_pd_bos(peer_vm, ticket); in process_validate_vms() 3001 ret = process_validate_vms(process_info, &exec.ticket); in amdgpu_amdkfd_gpuvm_restore_process_bos() 3042 ret = amdgpu_vm_handle_moved(adev, peer_vm, &exec.ticket); in amdgpu_amdkfd_gpuvm_restore_process_bos()
|
| /linux/drivers/gpu/drm/virtio/ |
| H A D | virtgpu_gem.c | 226 &objs->ticket); in virtio_gpu_array_lock_resv() 247 &objs->ticket); in virtio_gpu_array_unlock_resv()
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_gem.c | 407 struct ww_acquire_ctx ticket; member 452 ww_acquire_fini(&op->ticket); in validate_fini() 468 ww_acquire_init(&op->ticket, &reservation_ww_class); in validate_init() 501 ret = ttm_bo_reserve(&nvbo->bo, true, false, &op->ticket); in validate_init() 509 &op->ticket); in validate_init() 556 ww_acquire_done(&op->ticket); in validate_init()
|
| /linux/include/drm/ |
| H A D | drm_exec.h | 26 struct ww_acquire_ctx ticket; member
|
| /linux/Documentation/locking/ |
| H A D | ww-mutex-design.rst | 29 to be locked, the caller would be assigned a unique reservation id/ticket, 31 associated with a execbuf, the one with the lowest reservation ticket (i.e. 36 In the RDBMS literature, a reservation ticket is associated with a transaction. 65 acquired when starting the lock acquisition. This ticket is stored in the 105 better to avoid grabbing a deadlock avoidance ticket).
|
| /linux/arch/mips/cavium-octeon/executive/ |
| H A D | cvmx-cmd-queue.c | 188 ticket[__cvmx_cmd_queue_get_index(queue_id)] = 0; in cvmx_cmd_queue_initialize()
|