/linux/drivers/gpu/drm/ttm/ |
H A D | ttm_execbuf_util.c | 42 void ttm_eu_backoff_reservation(struct ww_acquire_ctx *ticket, in ttm_eu_backoff_reservation() argument 57 if (ticket) in ttm_eu_backoff_reservation() 58 ww_acquire_fini(ticket); in ttm_eu_backoff_reservation() 74 int ttm_eu_reserve_buffers(struct ww_acquire_ctx *ticket, in ttm_eu_reserve_buffers() argument 84 if (ticket) in ttm_eu_reserve_buffers() 85 ww_acquire_init(ticket, &reservation_ww_class); in ttm_eu_reserve_buffers() 91 ret = ttm_bo_reserve(bo, intr, (ticket == NULL), ticket); in ttm_eu_reserve_buffers() 115 ret = ttm_bo_reserve_slowpath(bo, intr, ticket); in ttm_eu_reserve_buffers() 123 if (ticket) { in ttm_eu_reserve_buffers() 124 ww_acquire_done(ticket); in ttm_eu_reserve_buffers() [all …]
|
H A D | ttm_bo.c | 547 struct ww_acquire_ctx *ticket, in ttm_bo_evict_alloc() argument 554 .ticket = ticket, in ttm_bo_evict_alloc() 564 if (lret || !ticket) in ttm_bo_evict_alloc() 571 evict_walk.walk.ticket = ticket; in ttm_bo_evict_alloc() 680 struct ww_acquire_ctx *ticket; in ttm_bo_alloc_resource() local 683 ticket = dma_resv_locking_ctx(bo->base.resv); in ttm_bo_alloc_resource() 710 ticket, res); in ttm_bo_alloc_resource()
|
H A D | ttm_bo_util.c | 801 ret = dma_resv_lock_interruptible(resv, walk->ticket); in ttm_lru_walk_ticketlock() 803 ret = dma_resv_lock(resv, walk->ticket); in ttm_lru_walk_ticketlock() 813 walk->ticket = NULL; in ttm_lru_walk_ticketlock() 881 else if (!walk->ticket || walk->ctx->no_wait_gpu || in ttm_lru_walk_for_evict()
|
/linux/fs/btrfs/ |
H A D | space-info.c | 459 struct reserve_ticket *ticket) in remove_ticket() argument 461 if (!list_empty(&ticket->list)) { in remove_ticket() 462 list_del_init(&ticket->list); in remove_ticket() 463 ASSERT(space_info->reclaim_size >= ticket->bytes); in remove_ticket() 464 space_info->reclaim_size -= ticket->bytes; in remove_ticket() 483 struct reserve_ticket *ticket; in btrfs_try_granting_tickets() local 486 ticket = list_first_entry(head, struct reserve_ticket, list); in btrfs_try_granting_tickets() 489 if ((used + ticket->bytes <= space_info->total_bytes) || in btrfs_try_granting_tickets() 490 btrfs_can_overcommit(fs_info, space_info, ticket->bytes, in btrfs_try_granting_tickets() 494 ticket->bytes); in btrfs_try_granting_tickets() [all …]
|
/linux/fs/xfs/ |
H A D | xfs_log.c | 47 struct xlog_ticket *ticket, 53 struct xlog_ticket *ticket); 521 struct xlog_ticket *ticket) in xlog_state_release_iclog() argument 567 xlog_sync(log, iclog, ticket); in xlog_state_release_iclog() 858 struct xlog_ticket *ticket) in xlog_write_unmount_record() argument 866 .oh_tid = cpu_to_be32(ticket->t_tid), in xlog_write_unmount_record() 890 ticket->t_curr_res -= sizeof(unmount_rec); in xlog_write_unmount_record() 892 return xlog_write(log, NULL, &lv_chain, ticket, reg.i_len); in xlog_write_unmount_record() 1796 struct xlog_ticket *ticket) in xlog_sync() argument 1813 if (ticket) { in xlog_sync() [all …]
|
H A D | xfs_log_cil.c | 129 ctx->ticket->t_curr_res += cilpcp->space_reserved; in xlog_cil_push_pcp_aggregate() 210 log->l_cilp->xc_ctx->ticket = xlog_cil_ticket_alloc(log); in xlog_cil_init_post_recovery() 599 ctx_res = ctx->ticket->t_unit_res; in xlog_cil_insert_items() 1088 return xlog_write(log, ctx, &ctx->lv_chain, ctx->ticket, chain_len); in xlog_cil_write_chain() 1104 .oh_tid = cpu_to_be32(ctx->ticket->t_tid), in xlog_cil_write_commit_record() 1128 ctx->ticket->t_curr_res -= reg.i_len; in xlog_cil_write_commit_record() 1129 error = xlog_write(log, ctx, &lv_chain, ctx->ticket, reg.i_len); in xlog_cil_write_commit_record() 1160 struct xlog_ticket *tic = ctx->ticket; in xlog_cil_build_trans_hdr() 1314 struct xlog_ticket *ticket; in xlog_cil_push_work() local 1317 new_ctx->ticket = xlog_cil_ticket_alloc(log); in xlog_cil_push_work() [all …]
|
H A D | xfs_log_priv.h | 226 struct xlog_ticket *ticket; /* chkpt ticket */ member 508 void xlog_print_tic_res(struct xfs_mount *mp, struct xlog_ticket *ticket); 513 void xfs_log_ticket_ungrant(struct xlog *log, struct xlog_ticket *ticket); 514 void xfs_log_ticket_regrant(struct xlog *log, struct xlog_ticket *ticket); 519 struct xlog_ticket *ticket);
|
H A D | xfs_log.h | 143 struct xlog_ticket **ticket, bool permanent); 148 struct xlog_ticket *xfs_log_ticket_get(struct xlog_ticket *ticket); 149 void xfs_log_ticket_put(struct xlog_ticket *ticket);
|
/linux/net/rxrpc/ |
H A D | key.c | 96 memcpy(&token->kad->ticket, &xdr[8], tktlen); in rxrpc_preparse_xdr_rxkad() 110 token->kad->ticket[0], token->kad->ticket[1], in rxrpc_preparse_xdr_rxkad() 111 token->kad->ticket[2], token->kad->ticket[3], in rxrpc_preparse_xdr_rxkad() 112 token->kad->ticket[4], token->kad->ticket[5], in rxrpc_preparse_xdr_rxkad() 113 token->kad->ticket[6], token->kad->ticket[7]); in rxrpc_preparse_xdr_rxkad() 333 v1->ticket[0], v1->ticket[1], in rxrpc_preparse() 334 v1->ticket[2], v1->ticket[3], in rxrpc_preparse() 335 v1->ticket[4], v1->ticket[5], in rxrpc_preparse() 336 v1->ticket[6], v1->ticket[7]); in rxrpc_preparse() 358 memcpy(&token->kad->ticket, v1->ticket, v1->ticket_length); in rxrpc_preparse() [all …]
|
H A D | rxkad.c | 721 iov[2].iov_base = (void *)s2->ticket; in rxkad_send_response() 863 void *ticket, size_t ticket_len, in rxkad_decrypt_ticket() argument 881 ASSERTCMP((unsigned long) ticket & 7UL, ==, 0); in rxkad_decrypt_ticket() 889 sg_init_one(&sg[0], ticket, ticket_len); in rxkad_decrypt_ticket() 895 p = ticket; in rxkad_decrypt_ticket() 1030 void *ticket; in rxkad_verify_response() local 1091 ticket = kmalloc(ticket_len, GFP_NOFS); in rxkad_verify_response() 1092 if (!ticket) in rxkad_verify_response() 1096 ticket, ticket_len) < 0) { in rxkad_verify_response() 1102 ret = rxkad_decrypt_ticket(conn, server_key, skb, ticket, ticket_len, in rxkad_verify_response() [all …]
|
/linux/drivers/gpu/drm/ |
H A D | drm_exec.c | 108 ww_acquire_fini(&exec->ticket); in drm_exec_fini() 124 ww_acquire_done(&exec->ticket); in drm_exec_cleanup() 130 ww_acquire_init(&exec->ticket, &reservation_ww_class); in drm_exec_cleanup() 174 &exec->ticket); in drm_exec_lock_contended() 178 dma_resv_lock_slow(obj->resv, &exec->ticket); in drm_exec_lock_contended() 222 ret = dma_resv_lock_interruptible(obj->resv, &exec->ticket); in drm_exec_lock_obj() 224 ret = dma_resv_lock(obj->resv, &exec->ticket); in drm_exec_lock_obj()
|
/linux/include/drm/ttm/ |
H A D | ttm_bo.h | 220 struct ww_acquire_ctx *ticket; member 280 struct ww_acquire_ctx *ticket) in ttm_bo_reserve() argument 287 if (WARN_ON(ticket)) in ttm_bo_reserve() 295 ret = dma_resv_lock_interruptible(bo->base.resv, ticket); in ttm_bo_reserve() 297 ret = dma_resv_lock(bo->base.resv, ticket); in ttm_bo_reserve() 315 struct ww_acquire_ctx *ticket) in ttm_bo_reserve_slowpath() argument 319 ticket); in ttm_bo_reserve_slowpath() 324 dma_resv_lock_slow(bo->base.resv, ticket); in ttm_bo_reserve_slowpath()
|
H A D | ttm_execbuf_util.h | 63 void ttm_eu_backoff_reservation(struct ww_acquire_ctx *ticket, 99 int ttm_eu_reserve_buffers(struct ww_acquire_ctx *ticket, 115 void ttm_eu_fence_buffer_objects(struct ww_acquire_ctx *ticket,
|
/linux/include/asm-generic/ |
H A D | spinlock.h | 36 u16 ticket = val >> 16; in arch_spin_lock() local 38 if (ticket == (u16)val) in arch_spin_lock() 49 atomic_cond_read_acquire(lock, ticket == (u16)VAL); in arch_spin_lock()
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_validation.h | 62 struct ww_acquire_ctx ticket; member 122 return ttm_eu_reserve_buffers(&ctx->ticket, &ctx->bo_list, intr, in vmw_validation_bo_reserve() 138 ttm_eu_fence_buffer_objects(&ctx->ticket, &ctx->bo_list, in vmw_validation_bo_fence()
|
H A D | vmwgfx_resource.c | 517 vmw_resource_check_buffer(struct ww_acquire_ctx *ticket, in vmw_resource_check_buffer() argument 538 ret = ttm_eu_reserve_buffers(ticket, &val_list, interruptible, NULL); in vmw_resource_check_buffer() 558 ttm_eu_backoff_reservation(ticket, &val_list); in vmw_resource_check_buffer() 611 vmw_resource_backoff_reservation(struct ww_acquire_ctx *ticket, in vmw_resource_backoff_reservation() argument 621 ttm_eu_backoff_reservation(ticket, &val_list); in vmw_resource_backoff_reservation() 634 static int vmw_resource_do_evict(struct ww_acquire_ctx *ticket, in vmw_resource_do_evict() argument 645 ret = vmw_resource_check_buffer(ticket, res, interruptible, &val_buf); in vmw_resource_do_evict() 660 vmw_resource_backoff_reservation(ticket, &val_buf); in vmw_resource_do_evict() 901 struct ww_acquire_ctx ticket; in vmw_resource_evict_type() local 916 ret = vmw_resource_do_evict(&ticket, evict_res, false); in vmw_resource_evict_type()
|
/linux/include/keys/ |
H A D | rxrpc-type.h | 31 u8 ticket[]; /* the encrypted ticket */ member 55 u8 ticket[]; member
|
/linux/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem_submit.c | 122 struct ww_acquire_ctx *ticket) in submit_lock_objects() argument 136 ret = dma_resv_lock_interruptible(obj->resv, ticket); in submit_lock_objects() 146 ww_acquire_done(ticket); in submit_lock_objects() 163 ret = dma_resv_lock_slow_interruptible(obj->resv, ticket); in submit_lock_objects() 424 struct ww_acquire_ctx ticket; in etnaviv_ioctl_gem_submit() local 516 ww_acquire_init(&ticket, &reservation_ww_class); in etnaviv_ioctl_gem_submit() 581 ret = submit_lock_objects(submit, &ticket); in etnaviv_ioctl_gem_submit() 625 ww_acquire_fini(&ticket); in etnaviv_ioctl_gem_submit()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_dma_buf.c | 337 struct ww_acquire_ctx *ticket = dma_resv_locking_ctx(obj->resv); in amdgpu_dma_buf_move_notify() local 362 if (ticket) { in amdgpu_dma_buf_move_notify() 367 r = dma_resv_lock(resv, ticket); in amdgpu_dma_buf_move_notify() 385 r = amdgpu_vm_handle_moved(adev, vm, ticket); in amdgpu_dma_buf_move_notify()
|
/linux/drivers/gpu/drm/qxl/ |
H A D | qxl_release.c | 224 ret = ttm_eu_reserve_buffers(&release->ticket, &release->bos, in qxl_release_reserve_list() 234 ttm_eu_backoff_reservation(&release->ticket, &release->bos); in qxl_release_reserve_list() 248 ttm_eu_backoff_reservation(&release->ticket, &release->bos); in qxl_release_backoff_reserve_list() 437 ww_acquire_fini(&release->ticket); in qxl_release_fence_buffer_objects()
|
/linux/arch/mips/include/asm/octeon/ |
H A D | cvmx-cmd-queue.h | 154 uint64_t ticket[(CVMX_CMD_QUEUE_END >> 16) * 256]; member 278 …[ticket_ptr] "=" GCC_OFF_SMALL_ASM()(__cvmx_cmd_queue_state_ptr->ticket[__cvmx_cmd_queue_get_index… in __cvmx_cmd_queue_lock() 279 [now_serving] "=m"(qptr->now_serving), [ticket] "=r"(tmp), in __cvmx_cmd_queue_lock()
|
/linux/tools/tracing/latency/ |
H A D | latency-collector.c | 185 int ticket; member 666 req->ticket = r; in printstate_next_ticket() 675 if (req->ticket > printstate.ticket_completed) in printstate_mark_req_completed() 676 printstate.ticket_completed = req->ticket; in printstate_mark_req_completed() 682 return (printstate.ticket_counter != req->ticket); in printstate_has_new_req_arrived() 764 int diff = req->ticket - req->ticket_completed_ref; in table_get_probability() 1095 sec, us, req->ticket, str); in __print_skipmessage() 1098 sec, us, req->ticket); in __print_skipmessage() 1166 sec, us, req->ticket, slept_ms); in print_tracefile() 1170 us, req->ticket); in print_tracefile() [all...] |
/linux/drivers/gpu/drm/virtio/ |
H A D | virtgpu_gem.c | 224 &objs->ticket); in virtio_gpu_array_lock_resv() 245 &objs->ticket); in virtio_gpu_array_unlock_resv()
|
/linux/include/drm/ |
H A D | drm_exec.h | 26 struct ww_acquire_ctx ticket; member
|
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_gem.c | 410 struct ww_acquire_ctx ticket; member 455 ww_acquire_fini(&op->ticket); in validate_fini() 471 ww_acquire_init(&op->ticket, &reservation_ww_class); in validate_init() 504 ret = ttm_bo_reserve(&nvbo->bo, true, false, &op->ticket); in validate_init() 512 &op->ticket); in validate_init() 559 ww_acquire_done(&op->ticket); in validate_init()
|