| /linux/tools/testing/selftests/sync/ |
| H A D | sync_fence.c | 38 ASSERT(valid, "Failure allocating timeline\n"); in test_fence_one_timeline_wait() 42 ASSERT(valid, "Failure allocating fence\n"); in test_fence_one_timeline_wait() 46 ASSERT(ret == 0, "Failure waiting on fence until timeout\n"); in test_fence_one_timeline_wait() 50 ASSERT(ret == 0, "Failure advancing timeline\n"); in test_fence_one_timeline_wait() 54 ASSERT(ret == 0, "Failure waiting on fence until timeout\n"); in test_fence_one_timeline_wait() 58 ASSERT(ret == 0, "Failure signaling the fence\n"); in test_fence_one_timeline_wait() 62 ASSERT(ret > 0, "Failure waiting on fence\n"); in test_fence_one_timeline_wait() 66 ASSERT(ret == 0, "Failure going further\n"); in test_fence_one_timeline_wait() 68 ASSERT(ret > 0, "Failure waiting ahead\n"); in test_fence_one_timeline_wait() 89 ASSERT(valid, "Failure allocating fences\n"); in test_fence_one_timeline_merge() [all …]
|
| H A D | sync_stress_consumer.c | 46 ASSERT(error == 0, "Error occurred on fence\n"); in busy_wait_on_fence() 74 ASSERT(valid, "Failure creating fence\n"); in mpsc_producer_thread() 82 ASSERT(sync_wait(fence, -1) > 0, in mpsc_producer_thread() 85 ASSERT(busy_wait_on_fence(fence) == 0, in mpsc_producer_thread() 97 ASSERT(sw_sync_timeline_inc(producer_timelines[id], 1) == 0, in mpsc_producer_thread() 126 ASSERT(valid, "Failure merging fences\n"); in mpcs_consumer_thread() 133 ASSERT(sync_wait(fence, -1) > 0, in mpcs_consumer_thread() 136 ASSERT(busy_wait_on_fence(fence) == 0, in mpcs_consumer_thread() 140 ASSERT(test_data_mpsc.counter == n * it, in mpcs_consumer_thread() 144 ASSERT(sw_sync_timeline_inc(consumer_timeline, 1) == 0, in mpcs_consumer_thread()
|
| H A D | sync_wait.c | 50 ASSERT(valid, "Failure merging fence from various timelines\n"); in test_fence_multi_timeline_wait() 54 ASSERT(active == 3, "Fence signaled too early!\n"); in test_fence_multi_timeline_wait() 57 ASSERT(ret == 0, in test_fence_multi_timeline_wait() 63 ASSERT(active == 2 && signaled == 1, in test_fence_multi_timeline_wait() 69 ASSERT(active == 1 && signaled == 2, in test_fence_multi_timeline_wait() 75 ASSERT(active == 0 && signaled == 3, in test_fence_multi_timeline_wait() 80 ASSERT(ret > 0, "Failure waiting on signaled fence\n"); in test_fence_multi_timeline_wait()
|
| H A D | sync_stress_parallelism.c | 51 ASSERT(valid, "Failure allocating fence\n"); in test_stress_two_threads_shared_timeline_thread() 55 ASSERT(ret > 0, "Problem occurred on prior thread\n"); in test_stress_two_threads_shared_timeline_thread() 61 ASSERT(test_data_two_threads.counter == i * 2 + thread_id, in test_stress_two_threads_shared_timeline_thread() 67 ASSERT(ret == 0, "Advancing timeline failed\n"); in test_stress_two_threads_shared_timeline_thread() 82 ASSERT(valid, "Failure allocating timeline\n"); in test_stress_two_threads_shared_timeline() 104 ASSERT(test_data_two_threads.counter == in test_stress_two_threads_shared_timeline()
|
| H A D | sync_merge.c | 38 ASSERT(valid, "Failure allocating timeline\n"); in test_fence_merge_same_fence() 42 ASSERT(valid, "Failure allocating fence\n"); in test_fence_merge_same_fence() 46 ASSERT(valid, "Failure merging fence\n"); in test_fence_merge_same_fence() 48 ASSERT(sync_fence_count_with_status(merged, FENCE_STATUS_SIGNALED) == 0, in test_fence_merge_same_fence() 52 ASSERT(sync_fence_count_with_status(merged, FENCE_STATUS_SIGNALED) == 1, in test_fence_merge_same_fence()
|
| H A D | sync_alloc.c | 38 ASSERT(valid, "Failure allocating timeline\n"); in test_alloc_timeline() 50 ASSERT(valid, "Failure allocating timeline\n"); in test_alloc_fence() 54 ASSERT(valid, "Failure allocating fence\n"); in test_alloc_fence() 66 ASSERT(timeline > 0, "Failure allocating timeline\n"); in test_alloc_fence_negative() 69 ASSERT(fence < 0, "Success allocating negative fence\n"); in test_alloc_fence_negative()
|
| H A D | sync_stress_merge.c | 53 ASSERT(valid, "Failure creating fence\n"); in test_merge_stress_random_merge() 82 ASSERT(valid, "Failure creating fence i\n"); in test_merge_stress_random_merge() 91 ASSERT(sync_fence_size(fence) == size, in test_merge_stress_random_merge() 98 ASSERT(ret == 0, in test_merge_stress_random_merge() 107 ASSERT(ret > 0, "Failure triggering fence\n"); in test_merge_stress_random_merge()
|
| /linux/fs/xfs/ |
| H A D | xfs_buf_item.c | 40 ASSERT(bip->bli_formats == NULL); in xfs_buf_item_get_format() 82 ASSERT(!test_bit(XFS_LI_IN_AIL, &bip->bli_item.li_flags)); in xfs_buf_item_relse() 83 ASSERT(atomic_read(&bip->bli_refcount) == 0); in xfs_buf_item_relse() 144 ASSERT(nbits > 0); in xfs_buf_item_size_segment() 210 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_buf_item_size() 218 ASSERT(bip->__bli_format.blf_flags & XFS_BLF_CANCEL); in xfs_buf_item_size() 226 ASSERT(bip->bli_flags & XFS_BLI_LOGGED); in xfs_buf_item_size() 318 ASSERT(blfp->blf_flags & XFS_BLF_CANCEL); in xfs_buf_item_format_segment() 327 ASSERT(first_bit >= 0); in xfs_buf_item_format_segment() 330 ASSERT(nbits > 0); in xfs_buf_item_format_segment() [all …]
|
| H A D | xfs_qm.c | 178 ASSERT(atomic_read(&dqp->q_pincount) == 0); in xfs_qm_dqpurge() 179 ASSERT(xlog_is_shutdown(dqp->q_logitem.qli_item.li_log) || in xfs_qm_dqpurge() 192 ASSERT(!list_empty(&dqp->q_lru)); in xfs_qm_dqpurge() 273 ASSERT(mp->m_rootip); in xfs_qm_unmount_quotas() 326 ASSERT(!xfs_is_metadir_inode(ip)); in xfs_qm_dqattach_locked() 333 ASSERT(ip->i_udquot); in xfs_qm_dqattach_locked() 341 ASSERT(ip->i_gdquot); in xfs_qm_dqattach_locked() 349 ASSERT(ip->i_pdquot); in xfs_qm_dqattach_locked() 393 ASSERT(!xfs_is_quota_inode(&ip->i_mount->m_sb, ip->i_ino)); in xfs_qm_dqdetach() 467 ASSERT(!XFS_DQ_IS_DIRTY(dqp)); in xfs_qm_dquot_isolate() [all …]
|
| H A D | xfs_buf_mem.c | 119 ASSERT(xfs_buftarg_is_mem(btp)); in xmbuf_free() 120 ASSERT(percpu_counter_sum(&btp->bt_readahead_count) == 0); in xmbuf_free() 140 ASSERT(xfs_buftarg_is_mem(bp->b_target)); in xmbuf_map_backing_mem() 147 ASSERT(offset_in_page(pos)); in xmbuf_map_backing_mem() 180 ASSERT(xfs_buftarg_is_mem(btp)); in xmbuf_verify_daddr() 193 ASSERT(xfs_buftarg_is_mem(bp->b_target)); in xmbuf_stale() 239 ASSERT(bli != NULL); in xmbuf_trans_bdetach()
|
| H A D | xfs_inode_item_recover.c | 86 ASSERT(in_f->ilf_fields & (XFS_ILOG_DOWNER|XFS_ILOG_AOWNER)); in xfs_recover_inode_owner_change() 93 ASSERT(dip->di_version >= 3); in xfs_recover_inode_owner_change() 100 ASSERT(in_f->ilf_fields & XFS_ILOG_DBROOT); in xfs_recover_inode_owner_change() 108 ASSERT(in_f->ilf_fields & XFS_ILOG_ABROOT); in xfs_recover_inode_owner_change() 295 ASSERT(0); in xlog_recover_inode_dbroot() 300 ASSERT(0); in xlog_recover_inode_dbroot() 355 ASSERT(in_f->ilf_fields & XFS_ILOG_CORE); in xlog_recover_inode_commit_pass2() 505 ASSERT(in_f->ilf_size <= 4); in xlog_recover_inode_commit_pass2() 506 ASSERT((in_f->ilf_size == 3) || (fields & XFS_ILOG_AFORK)); in xlog_recover_inode_commit_pass2() 507 ASSERT(!(fields & XFS_ILOG_DFORK) || in xlog_recover_inode_commit_pass2() [all …]
|
| H A D | xfs_dquot.c | 66 ASSERT(0); in xfs_dquot_mark_sick() 103 ASSERT(list_empty(&dqp->q_lru)); in xfs_qm_dqdestroy() 104 ASSERT(dqp->q_logitem.qli_item.li_buf == NULL); in xfs_qm_dqdestroy() 127 ASSERT(dq->q_id); in xfs_qm_adjust_dqlimits() 181 ASSERT(res->hardlimit == 0 || res->softlimit <= res->hardlimit); in xfs_qm_adjust_res_timer() 214 ASSERT(dq->q_id); in xfs_qm_adjust_dqtimers() 240 ASSERT(tp); in xfs_qm_init_dquot_blk() 241 ASSERT(xfs_buf_islocked(bp)); in xfs_qm_init_dquot_blk() 257 ASSERT(0); in xfs_qm_init_dquot_blk() 387 ASSERT(map.br_blockcount == XFS_DQUOT_CLUSTER_SIZE_FSB); in xfs_dquot_disk_alloc() [all …]
|
| H A D | xfs_buf.c | 70 ASSERT(xfs_buf_islocked(bp)); in xfs_buf_stale() 87 ASSERT(bp->b_hold >= 1); in xfs_buf_stale() 111 ASSERT(list_empty(&bp->b_lru)); in xfs_buf_free() 132 ASSERT(is_power_of_2(size)); in xfs_buf_alloc_kmem() 133 ASSERT(size < PAGE_SIZE); in xfs_buf_alloc_kmem() 355 ASSERT(bp->b_flags & XBF_STALE); in _xfs_buf_obj_cmp() 391 ASSERT(!(BBTOB(map->bm_len) < btp->bt_meta_sectorsize)); in xfs_buf_map_verify() 392 ASSERT(!(BBTOB(map->bm_bn) & (xfs_off_t)btp->bt_meta_sectormask)); in xfs_buf_map_verify() 433 ASSERT((bp->b_flags & _XBF_DELWRI_Q) == 0); in xfs_buf_find_lock() 638 ASSERT(bp->b_maps[0].bm_bn != XFS_BUF_DADDR_NULL); in _xfs_buf_read() [all …]
|
| /linux/arch/arm64/kernel/ |
| H A D | vmlinux.lds.S | 211 ASSERT(SIZEOF(.got.plt) == 0 || SIZEOF(.got.plt) == 0x18, 302 ASSERT(SIZEOF(.data.rel.ro) == 0, "Unexpected RELRO detected!") 364 ASSERT(SIZEOF(.plt) == 0, "Unexpected run-time procedure linkages detected!") 374 ASSERT(__hyp_idmap_text_end - __hyp_idmap_text_start <= PAGE_SIZE, 376 ASSERT(__idmap_text_end - (__idmap_text_start & ~(SZ_4K - 1)) <= SZ_4K, 379 ASSERT(__hibernate_exit_text_end - __hibernate_exit_text_start <= SZ_4K, 381 ASSERT(__hibernate_exit_text_start == swsusp_arch_suspend_exit, 385 ASSERT((__entry_tramp_text_end - __entry_tramp_text_start) <= 3*PAGE_SIZE, 389 ASSERT(__hyp_bss_start == __bss_start, "HYP and Host BSS are misaligned") 394 ASSERT(_text == KIMAGE_VADDR, "HEAD is misaligned") [all …]
|
| /linux/arch/x86/kernel/ |
| H A D | vmlinux.lds.S | 113 ASSERT(__relocate_kernel_end - __relocate_kernel_start <= KEXEC_CONTROL_CODE_MAX_SIZE, 337 ASSERT(__per_cpu_hot_end - __per_cpu_hot_start <= 64, "percpu cache hot data too large") 440 ASSERT(SIZEOF(.got.plt) == 0 || 455 ASSERT(SIZEOF(.got) == 0, "Unexpected GOT entries detected!") 460 ASSERT(SIZEOF(.plt) == 0, "Unexpected run-time procedure linkages detected!") 465 ASSERT(SIZEOF(.rel.dyn) == 0, "Unexpected run-time relocations (.rel) detected!") 470 ASSERT(SIZEOF(.rela.dyn) == 0, "Unexpected run-time relocations (.rela) detected!") 483 . = ASSERT((_end - LOAD_OFFSET <= KERNEL_IMAGE_SIZE), 493 . = ASSERT((retbleed_return_thunk & 0x3f) == 0, "retbleed_return_thunk not cacheline-aligned"); 497 . = ASSERT((srso_safe_ret & 0x3f) == 0, "srso_safe_ret not cacheline-aligned"); [all …]
|
| /linux/fs/btrfs/ |
| H A D | lzo.c | 140 ASSERT(out_folio && *out_folio); in write_and_queue_folio() 142 ASSERT(foffset + write_len <= fsize); in write_and_queue_folio() 156 ASSERT(IS_ALIGNED(*total_out, fsize)); in write_and_queue_folio() 208 ASSERT(out_folio); in copy_compressed_data_to_bio() 211 ASSERT(old_size); in copy_compressed_data_to_bio() 212 ASSERT(old_size == *total_out); in copy_compressed_data_to_bio() 218 ASSERT((old_size >> sectorsize_bits) == (old_size + LZO_LEN - 1) >> sectorsize_bits); in copy_compressed_data_to_bio() 269 ASSERT(*out_folio); in copy_compressed_data_to_bio() 296 ASSERT(bio->bi_iter.bi_size == 0); in lzo_compress_bio() 297 ASSERT(len); in lzo_compress_bio() [all …]
|
| H A D | raid56.c | 83 ASSERT((expr)); \ 95 ASSERT((expr)); \ 107 ASSERT((expr)); \ 119 ASSERT((expr)); \ 235 ASSERT(sector_nr < rbio->nr_sectors); in memcpy_from_bio_to_stripe() 241 ASSERT(dst != INVALID_PADDR); in memcpy_from_bio_to_stripe() 242 ASSERT(src != INVALID_PADDR); in memcpy_from_bio_to_stripe() 276 ASSERT(test_bit(i, rbio->stripe_uptodate_bitmap)); in cache_rbio_pages() 309 ASSERT(page_nr < rbio->nr_pages); in page_nr_to_sector_nr() 312 ASSERT(sector_nr < rbio->nr_sectors); in page_nr_to_sector_nr() [all …]
|
| H A D | free-space-tree.c | 241 ASSERT(found_key.objectid == block_group->start); in btrfs_convert_free_space_to_bitmaps() 242 ASSERT(found_key.offset == block_group->length); in btrfs_convert_free_space_to_bitmaps() 248 ASSERT(found_key.objectid >= start); in btrfs_convert_free_space_to_bitmaps() 249 ASSERT(found_key.objectid < end); in btrfs_convert_free_space_to_bitmaps() 250 ASSERT(found_key.objectid + found_key.offset <= end); in btrfs_convert_free_space_to_bitmaps() 262 ASSERT(0); in btrfs_convert_free_space_to_bitmaps() 381 ASSERT(found_key.objectid == block_group->start); in btrfs_convert_free_space_to_extents() 382 ASSERT(found_key.offset == block_group->length); in btrfs_convert_free_space_to_extents() 390 ASSERT(found_key.objectid >= start); in btrfs_convert_free_space_to_extents() 391 ASSERT(found_key.objectid < end); in btrfs_convert_free_space_to_extents() [all …]
|
| H A D | bio.c | 48 ASSERT(inode); in btrfs_bio_init() 114 ASSERT(in_task()); in btrfs_bio_end_io() 196 ASSERT(repair_bbio->saved_iter.bi_size == fs_info->sectorsize); in btrfs_end_repair_bio() 199 ASSERT(slot < nr_steps); in btrfs_end_repair_bio() 285 ASSERT(offset_in_page(paddrs[i]) + step <= PAGE_SIZE); in repair_one_sector() 289 ASSERT(ret == step); in repair_one_sector() 317 ASSERT(inode); in btrfs_check_read_bio() 411 ASSERT(in_task()); in btrfs_raid56_end_io() 508 ASSERT(btrfs_dev_is_sequential(dev, physical)); in btrfs_submit_dev_bio() 538 ASSERT(bio_op(orig_bio) != REQ_OP_READ); in btrfs_submit_mirrored_bio() [all …]
|
| H A D | lru_cache.c | 75 ASSERT(cache->size > 0); in btrfs_lru_cache_remove() 76 ASSERT(!mtree_empty(&cache->entries)); in btrfs_lru_cache_remove() 90 ASSERT(head == prev); in btrfs_lru_cache_remove() 125 ASSERT(head != NULL); in btrfs_lru_cache_store() 164 ASSERT(cache->size == 0); in btrfs_lru_cache_clear() 165 ASSERT(mtree_empty(&cache->entries)); in btrfs_lru_cache_clear()
|
| /linux/drivers/gpu/drm/amd/display/dc/basics/ |
| H A D | bw_fixed.c | 53 ASSERT(value < BW_FIXED_MAX_I32 && value > BW_FIXED_MIN_I32); in bw_int_to_fixed_nonconst() 70 ASSERT(denominator != 0); in bw_frc_to_fixed() 76 ASSERT(res_value <= BW_FIXED_MAX_I32); in bw_frc_to_fixed() 98 ASSERT(res_value <= MAX_I64 - summand); in bw_frc_to_fixed() 118 ASSERT(abs_i64(result.value) <= abs_i64(arg.value)); in bw_floor2() 159 ASSERT(res.value <= BW_FIXED_MAX_I32); in bw_mul() 165 ASSERT(tmp <= (uint64_t)(MAX_I64 - res.value)); in bw_mul() 171 ASSERT(tmp <= (uint64_t)(MAX_I64 - res.value)); in bw_mul() 180 ASSERT(tmp <= (uint64_t)(MAX_I64 - res.value)); in bw_mul()
|
| /linux/drivers/gpu/drm/amd/display/dc/dml/ |
| H A D | dml1_display_rq_dlg_calc.c | 183 ASSERT(prefill > 0.0 && prefill <= 8.0); in get_swath_need() 507 ASSERT((log2_dpte_req_width_ptes == 3 && log2_dpte_req_height_ptes == 0) || /* 8x1 */ in dml1_rq_dlg_get_row_heights() 526 ASSERT(log2_dpte_row_height_linear >= 3); in dml1_rq_dlg_get_row_heights() 831 ASSERT((log2_dpte_req_width_ptes == 3 && log2_dpte_req_height_ptes == 0) || /* 8x1 */ in get_surf_rq_param() 852 ASSERT(log2_dpte_row_height_linear >= 3); in get_surf_rq_param() 947 ASSERT(0); in get_surf_rq_param() 955 ASSERT(0); in get_surf_rq_param() 1151 ASSERT(ref_freq_to_pix_freq < 4.0); in dml1_rq_dlg_get_dlg_params() 1158 ASSERT(disp_dlg_regs->refcyc_h_blank_end < (unsigned int) dml_pow(2, 13)); in dml1_rq_dlg_get_dlg_params() 1176 ASSERT(disp_dlg_regs->min_dst_y_next_start < (unsigned int) dml_pow(2, 18)); in dml1_rq_dlg_get_dlg_params() [all …]
|
| /linux/fs/xfs/scrub/ |
| H A D | dirtree_repair.c | 92 ASSERT(oc->suspect == 0); in xrep_dirtree_delete_all_paths() 93 ASSERT(oc->good == 0); in xrep_dirtree_delete_all_paths() 131 ASSERT(foundit == false); in xrep_dirtree_find_surviving_path() 138 ASSERT(oc->suspect + oc->good == 1); in xrep_dirtree_find_surviving_path() 173 ASSERT(oc->suspect == 0); in xrep_dirtree_keep_one_good_path() 174 ASSERT(oc->good < 2); in xrep_dirtree_keep_one_good_path() 200 ASSERT(0); in xrep_dirtree_keep_one_suspect_path() 207 ASSERT(oc->suspect == 1); in xrep_dirtree_keep_one_suspect_path() 208 ASSERT(oc->good == 0); in xrep_dirtree_keep_one_suspect_path() 303 ASSERT(d_is_dir(parent_dentry)); in xrep_dirtree_purge_dentry() [all …]
|
| /linux/arch/x86/boot/compressed/ |
| H A D | vmlinux.lds.S | 97 ASSERT(SIZEOF(.got.plt) == 0 || 112 ASSERT(SIZEOF(.got) == 0, "Unexpected GOT entries detected!") 117 ASSERT(SIZEOF(.plt) == 0, "Unexpected run-time procedure linkages detected!") 122 ASSERT(SIZEOF(.rel.dyn) == 0, "Unexpected run-time relocations (.rel) detected!") 127 ASSERT(SIZEOF(.rela.dyn) == 0, "Unexpected run-time relocations (.rela) detected!")
|
| /linux/fs/xfs/libxfs/ |
| H A D | xfs_da_btree.c | 155 ASSERT(to->magic == XFS_DA3_NODE_MAGIC); in xfs_da3_node_hdr_from_disk() 163 ASSERT(to->magic == XFS_DA_NODE_MAGIC); in xfs_da3_node_hdr_from_disk() 176 ASSERT(from->magic == XFS_DA3_NODE_MAGIC); in xfs_da3_node_hdr_to_disk() 183 ASSERT(from->magic == XFS_DA_NODE_MAGIC); in xfs_da3_node_hdr_to_disk() 296 ASSERT(0); in xfs_da3_header_check() 519 ASSERT(level <= XFS_DA_NODE_MAXDEPTH); in xfs_da3_node_create() 578 ASSERT((max >= 0) && (max < XFS_DA_NODE_MAXDEPTH)); in xfs_da3_split() 579 ASSERT(state->path.blk[max].magic == XFS_ATTR_LEAF_MAGIC || in xfs_da3_split() 660 ASSERT(state->extravalid == 0 || in xfs_da3_split() 666 ASSERT(stat in xfs_da3_split() [all...] |