| /linux/block/ |
| H A D | badblocks.c | 457 static int prev_by_hint(struct badblocks *bb, sector_t s, int hint) in prev_by_hint() argument 460 u64 *p = bb->page; in prev_by_hint() 463 while ((hint < hint_end) && ((hint + 1) <= bb->count) && in prev_by_hint() 465 if ((hint + 1) == bb->count || BB_OFFSET(p[hint + 1]) > s) { in prev_by_hint() 481 static int prev_badblocks(struct badblocks *bb, struct badblocks_context *bad, in prev_badblocks() argument 489 if (!bb->count) in prev_badblocks() 493 ret = prev_by_hint(bb, s, hint); in prev_badblocks() 499 hi = bb->count; in prev_badblocks() 500 p = bb->page; in prev_badblocks() 534 static bool can_merge_front(struct badblocks *bb, int prev, in can_merge_front() argument [all …]
|
| /linux/tools/bpf/bpftool/ |
| H A D | cfg.c | 50 #define bb_prev(bb) list_prev_entry(bb, l) argument 51 #define bb_next(bb) list_next_entry(bb, l) argument 91 struct bb_node *new_bb, *bb; in func_append_bb() local 93 list_for_each_entry(bb, &func->bbs, l) { in func_append_bb() 94 if (bb->head == insn) in func_append_bb() 95 return bb; in func_append_bb() 96 else if (bb->head > insn) in func_append_bb() 100 bb = bb_prev(bb); in func_append_bb() 109 list_add(&new_bb->l, &bb->l); in func_append_bb() 116 struct bb_node *bb; in func_insert_dummy_bb() local [all …]
|
| /linux/arch/x86/events/intel/ |
| H A D | bts.c | 84 struct bts_buffer *bb; in bts_buffer_setup_aux() local 105 bb = kzalloc_node(struct_size(bb, buf, nr_buf), GFP_KERNEL, node); in bts_buffer_setup_aux() 106 if (!bb) in bts_buffer_setup_aux() 109 bb->nr_pages = nr_pages; in bts_buffer_setup_aux() 110 bb->nr_bufs = nr_buf; in bts_buffer_setup_aux() 111 bb->snapshot = overwrite; in bts_buffer_setup_aux() 112 bb->data_pages = pages; in bts_buffer_setup_aux() 113 bb->real_size = size - size % BTS_RECORD_SIZE; in bts_buffer_setup_aux() 115 for (pg = 0, nr_buf = 0, offset = 0, pad = 0; nr_buf < bb->nr_bufs; nr_buf++) { in bts_buffer_setup_aux() 120 bb->buf[nr_buf].page = page; in bts_buffer_setup_aux() [all …]
|
| /linux/include/linux/ |
| H A D | badblocks.h | 51 int badblocks_check(struct badblocks *bb, sector_t s, sector_t sectors, 53 bool badblocks_set(struct badblocks *bb, sector_t s, sector_t sectors, 55 bool badblocks_clear(struct badblocks *bb, sector_t s, sector_t sectors); 56 void ack_all_badblocks(struct badblocks *bb); 57 ssize_t badblocks_show(struct badblocks *bb, char *page, int unack); 58 ssize_t badblocks_store(struct badblocks *bb, const char *page, size_t len, 60 int badblocks_init(struct badblocks *bb, int enable); 61 void badblocks_exit(struct badblocks *bb); 63 int devm_init_badblocks(struct device *dev, struct badblocks *bb); 64 static inline void devm_exit_badblocks(struct device *dev, struct badblocks *bb) in devm_exit_badblocks() argument [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_migrate.c | 121 /* First slot is reserved for mapping of PT bo and bb, start from 1 */ in xe_migrate_vm_addr() 601 struct xe_bb *bb, u32 at_pt, in emit_pte() 625 bb->cs[bb->len++] = MI_STORE_DATA_IMM | MI_SDI_NUM_QW(chunk); in emit_pte() 626 bb->cs[bb->len++] = ofs; in emit_pte() 627 bb->cs[bb->len++] = 0; in emit_pte() 655 bb->cs[bb in emit_pte() 598 emit_pte(struct xe_migrate * m,struct xe_bb * bb,u32 at_pt,bool is_vram,bool is_comp_pte,struct xe_res_cursor * cur,u32 size,struct ttm_resource * res) emit_pte() argument 662 emit_copy_ccs(struct xe_gt * gt,struct xe_bb * bb,u64 dst_ofs,bool dst_is_indirect,u64 src_ofs,bool src_is_indirect,u32 size) emit_copy_ccs() argument 703 emit_xy_fast_copy(struct xe_gt * gt,struct xe_bb * bb,u64 src_ofs,u64 dst_ofs,unsigned int size,unsigned int pitch) emit_xy_fast_copy() argument 735 emit_mem_copy(struct xe_gt * gt,struct xe_bb * bb,u64 src_ofs,u64 dst_ofs,unsigned int size,unsigned int pitch) emit_mem_copy() argument 777 emit_copy(struct xe_gt * gt,struct xe_bb * bb,u64 src_ofs,u64 dst_ofs,unsigned int size,unsigned int pitch) emit_copy() argument 795 xe_migrate_ccs_copy(struct xe_migrate * m,struct xe_bb * bb,u64 src_ofs,bool src_is_indirect,u64 dst_ofs,bool dst_is_indirect,u32 dst_size,u64 ccs_ofs,bool copy_ccs) xe_migrate_ccs_copy() argument 908 struct xe_bb *bb; xe_migrate_copy() local 1107 struct xe_bb *bb = NULL; xe_migrate_ccs_rw_copy() local 1265 struct xe_bb *bb; xe_migrate_vram_copy_chunk() local 1336 emit_clear_link_copy(struct xe_gt * gt,struct xe_bb * bb,u64 src_ofs,u32 size,u32 pitch) emit_clear_link_copy() argument 1359 emit_clear_main_copy(struct xe_gt * gt,struct xe_bb * bb,u64 src_ofs,u32 size,u32 pitch,bool is_vram) emit_clear_main_copy() argument 1422 emit_clear(struct xe_gt * gt,struct xe_bb * bb,u64 src_ofs,u32 size,u32 pitch,bool is_vram) emit_clear() argument 1482 struct xe_bb *bb; xe_migrate_clear() local 1591 write_pgtable(struct xe_tile * tile,struct xe_bb * bb,u64 ppgtt_ofs,const struct xe_vm_pgtable_update_op * pt_op,const struct xe_vm_pgtable_update * update,struct xe_migrate_pt_update * pt_update) write_pgtable() argument 1717 struct xe_bb *bb; __xe_migrate_update_pgtables() local 1970 build_pt_update_batch_sram(struct xe_migrate * m,struct xe_bb * bb,u32 pt_offset,struct drm_pagemap_addr * sram_addr,u32 size,int level) build_pt_update_batch_sram() argument 2075 struct xe_bb *bb; xe_migrate_vram() local [all...] |
| H A D | xe_gsc_submit.c | 175 struct xe_bb *bb; in xe_gsc_pkt_submit_kernel() 186 bb = xe_bb_new(gt, 8, false); in xe_gsc_pkt_submit_kernel() 187 if (IS_ERR(bb)) in xe_gsc_pkt_submit_kernel() 188 return PTR_ERR(bb); in xe_gsc_pkt_submit_kernel() 190 bb->cs[bb->len++] = GSC_HECI_CMD_PKT; in xe_gsc_pkt_submit_kernel() 191 bb->cs[bb->len++] = lower_32_bits(addr_in); in xe_gsc_pkt_submit_kernel() 192 bb->cs[bb in xe_gsc_pkt_submit_kernel() 177 struct xe_bb *bb; xe_gsc_pkt_submit_kernel() local [all...] |
| H A D | xe_sriov_vf_ccs.c | 416 struct xe_bb *bb; in xe_sriov_vf_ccs_attach_bo() 424 bb = bo->bb_ccs[ctx_id]; 425 /* bb should be NULL here. Assert if not NULL */ 426 xe_assert(xe, !bb); 448 struct xe_bb *bb; in xe_sriov_vf_ccs_detach_bo() 456 bb = bo->bb_ccs[ctx_id]; 457 if (!bb) 486 drm_printf(p, "ccs %s bb suballoc info\n", ctx_id ? "write" : "read"); 400 struct xe_bb *bb; xe_sriov_vf_ccs_attach_bo() local 432 struct xe_bb *bb; xe_sriov_vf_ccs_detach_bo() local
|
| H A D | xe_gt.c | 173 static int emit_job_sync(struct xe_exec_queue *q, struct xe_bb *bb, in emit_job_sync() 180 job = xe_bb_create_job(q, bb); in emit_nop_job() 200 struct xe_bb *bb; in emit_wa_job() 203 bb = xe_bb_new(gt, 4, false); in emit_wa_job() 204 if (IS_ERR(bb)) in emit_wa_job() 205 return PTR_ERR(bb); in emit_wa_job() 207 ret = emit_job_sync(q, bb, HZ); in emit_wa_job() 208 xe_bb_free(bb, NULL); in emit_wa_job() 219 struct xe_bb *bb; in emit_wa_job() 246 bb in emit_wa_job() 150 emit_job_sync(struct xe_exec_queue * q,struct xe_bb * bb,long timeout_jiffies) emit_job_sync() argument 177 struct xe_bb *bb; emit_nop_job() local 196 struct xe_bb *bb; emit_wa_job() local [all...] |
| /linux/net/tipc/ |
| H A D | bcast.c | 98 struct tipc_bc_base *bb = tipc_bc_base(net); in tipc_bcbase_calc_bc_threshold() local 101 bb->bc_threshold = 1 + (cluster_size * bb->rc_ratio / 100); in tipc_bcbase_calc_bc_threshold() 109 struct tipc_bc_base *bb = tipc_bc_base(net); in tipc_bcbase_select_primary() local 110 int all_dests = tipc_link_bc_peers(bb->link); in tipc_bcbase_select_primary() 111 int max_win = tipc_link_max_win(bb->link); in tipc_bcbase_select_primary() 112 int min_win = tipc_link_min_win(bb->link); in tipc_bcbase_select_primary() 115 bb->primary_bearer = INVALID_BEARER_ID; in tipc_bcbase_select_primary() 116 bb->bcast_support = true; in tipc_bcbase_select_primary() 122 if (!bb->dests[i]) in tipc_bcbase_select_primary() 126 if (mtu < tipc_link_mtu(bb->link)) { in tipc_bcbase_select_primary() [all …]
|
| /linux/arch/alpha/boot/tools/ |
| H A D | objstrip.c | 109 unsigned long bb[64], sum = 0; in main() local 124 memset(bb, 0, sizeof(bb)); in main() 125 strcpy((char *) bb, "Linux SRM bootblock"); in main() 126 bb[60] = size / BLOCK_SIZE; /* count */ in main() 127 bb[61] = 1; /* starting sector # */ in main() 128 bb[62] = 0; /* flags---must be 0 */ in main() 130 sum += bb[i]; in main() 132 bb[63] = sum; in main() 133 if (write(ofd, bb, sizeof(bb)) != sizeof(bb)) { in main()
|
| /linux/arch/mips/lib/ |
| H A D | multi3.c | 35 TWunion res, aa, bb; in __multi3() local 38 bb.ti = b; in __multi3() 45 res.s.low = dmulu(aa.s.low, bb.s.low); in __multi3() 46 res.s.high = dmuhu(aa.s.low, bb.s.low); in __multi3() 47 res.s.high += dmulu(aa.s.high, bb.s.low); in __multi3() 48 res.s.high += dmulu(aa.s.low, bb.s.high); in __multi3()
|
| /linux/drivers/nvdimm/ |
| H A D | badrange.c | 165 static void set_badblock(struct badblocks *bb, sector_t s, int num) in set_badblock() argument 167 dev_dbg(bb->dev, "Found a bad range (0x%llx, 0x%llx)\n", in set_badblock() 170 if (!badblocks_set(bb, s, num, 1)) in set_badblock() 171 dev_info_once(bb->dev, "%s: failed for sector %llx\n", in set_badblock() 185 static void __add_badblock_range(struct badblocks *bb, u64 ns_offset, u64 len) in __add_badblock_range() argument 205 set_badblock(bb, s, done); in __add_badblock_range() 210 set_badblock(bb, start_sector, num_sectors); in __add_badblock_range() 214 struct badblocks *bb, const struct range *range) in badblocks_populate() argument 239 __add_badblock_range(bb, start - range->start, len); in badblocks_populate() 253 __add_badblock_range(bb, 0, len); in badblocks_populate() [all …]
|
| H A D | claim.c | 251 if (unlikely(is_bad_pmem(&nsio->bb, sector, sz_align))) in nsio_rw_bytes() 258 if (unlikely(is_bad_pmem(&nsio->bb, sector, sz_align))) { in nsio_rw_bytes() 270 badblocks_clear(&nsio->bb, sector, cleared); in nsio_rw_bytes() 302 if (devm_init_badblocks(dev, &nsio->bb)) in devm_nsio_enable() 304 nvdimm_badblocks_populate(to_nd_region(ndns->dev.parent), &nsio->bb, in devm_nsio_enable() 317 devm_exit_badblocks(dev, &nsio->bb); in devm_nsio_disable()
|
| /linux/fs/xfs/libxfs/ |
| H A D | xfs_rtrefcount_btree.h | 118 struct xfs_btree_block *bb, in xfs_rtrefcount_broot_ptr_addr() argument 122 return xfs_rtrefcount_ptr_addr(bb, index, in xfs_rtrefcount_broot_ptr_addr() 149 xfs_rtrefcount_broot_space(struct xfs_mount *mp, struct xfs_rtrefcount_root *bb) in xfs_rtrefcount_broot_space() argument 151 return xfs_rtrefcount_broot_space_calc(mp, be16_to_cpu(bb->bb_level), in xfs_rtrefcount_broot_space() 152 be16_to_cpu(bb->bb_numrecs)); in xfs_rtrefcount_broot_space() 174 xfs_rtrefcount_droot_space(struct xfs_btree_block *bb) in xfs_rtrefcount_droot_space() argument 176 return xfs_rtrefcount_droot_space_calc(be16_to_cpu(bb->bb_level), in xfs_rtrefcount_droot_space() 177 be16_to_cpu(bb->bb_numrecs)); in xfs_rtrefcount_droot_space()
|
| H A D | xfs_rtrmap_btree.h | 129 struct xfs_btree_block *bb, in xfs_rtrmap_broot_ptr_addr() argument 133 return xfs_rtrmap_ptr_addr(bb, index, in xfs_rtrmap_broot_ptr_addr() 160 xfs_rtrmap_broot_space(struct xfs_mount *mp, struct xfs_rtrmap_root *bb) in xfs_rtrmap_broot_space() argument 162 return xfs_rtrmap_broot_space_calc(mp, be16_to_cpu(bb->bb_level), in xfs_rtrmap_broot_space() 163 be16_to_cpu(bb->bb_numrecs)); in xfs_rtrmap_broot_space() 185 xfs_rtrmap_droot_space(struct xfs_btree_block *bb) in xfs_rtrmap_droot_space() argument 187 return xfs_rtrmap_droot_space_calc(be16_to_cpu(bb->bb_level), in xfs_rtrmap_droot_space() 188 be16_to_cpu(bb->bb_numrecs)); in xfs_rtrmap_droot_space()
|
| H A D | xfs_bmap_btree.h | 151 struct xfs_btree_block *bb, in xfs_bmap_broot_ptr_addr() argument 155 return xfs_bmbt_ptr_addr(mp, bb, i, xfs_bmbt_maxrecs(mp, sz, false)); in xfs_bmap_broot_ptr_addr() 178 struct xfs_bmdr_block *bb) in xfs_bmap_broot_space() argument 180 return xfs_bmap_broot_space_calc(mp, be16_to_cpu(bb->bb_numrecs)); in xfs_bmap_broot_space() 196 xfs_bmap_bmdr_space(struct xfs_btree_block *bb) in xfs_bmap_bmdr_space() argument 198 return xfs_bmdr_space_calc(be16_to_cpu(bb->bb_numrecs)); in xfs_bmap_bmdr_space()
|
| /linux/drivers/spi/ |
| H A D | spi-gpio.c | 347 struct spi_bitbang *bb; in spi_gpio_probe() local 385 bb = &spi_gpio->bitbang; in spi_gpio_probe() 386 bb->ctlr = host; in spi_gpio_probe() 393 bb->chipselect = spi_gpio_chipselect; in spi_gpio_probe() 394 bb->set_line_direction = spi_gpio_set_direction; in spi_gpio_probe() 395 bb->set_mosi_idle = spi_gpio_set_mosi_idle; in spi_gpio_probe() 398 bb->txrx_word[SPI_MODE_0] = spi_gpio_spec_txrx_word_mode0; in spi_gpio_probe() 399 bb->txrx_word[SPI_MODE_1] = spi_gpio_spec_txrx_word_mode1; in spi_gpio_probe() 400 bb->txrx_word[SPI_MODE_2] = spi_gpio_spec_txrx_word_mode2; in spi_gpio_probe() 401 bb in spi_gpio_probe() [all...] |
| /linux/scripts/gcc-plugins/ |
| H A D | latent_entropy_plugin.c | 340 static void perturb_local_entropy(basic_block bb, tree local_entropy) in perturb_local_entropy() argument 349 gsi = gsi_after_labels(bb); in perturb_local_entropy() 383 static bool handle_tail_calls(basic_block bb, tree local_entropy) in handle_tail_calls() argument 387 for (gsi = gsi_start_bb(bb); !gsi_end_p(gsi); gsi_next(&gsi)) { in handle_tail_calls() 431 static void init_local_entropy(basic_block bb, tree local_entropy) in init_local_entropy() argument 437 gimple_stmt_iterator gsi = gsi_after_labels(bb); in init_local_entropy() 503 basic_block bb; in latent_entropy_execute() local 511 bb = single_succ(ENTRY_BLOCK_PTR_FOR_FN(cfun)); in latent_entropy_execute() 512 if (!single_pred_p(bb)) { in latent_entropy_execute() 515 bb = single_succ(ENTRY_BLOCK_PTR_FOR_FN(cfun)); in latent_entropy_execute() [all …]
|
| H A D | stackleak_plugin.c | 59 basic_block bb; in add_stack_tracking_gcall() local 70 bb = gimple_bb(gimple_call); in add_stack_tracking_gcall() 74 gimple_call, bb->count, in add_stack_tracking_gcall() 75 compute_call_stmt_bb_frequency(current_function_decl, bb)); in add_stack_tracking_gcall() 173 basic_block bb, entry_bb; in stackleak_instrument_execute() local 190 FOR_EACH_BB_FN(bb, cfun) { in stackleak_instrument_execute() 191 for (gsi = gsi_start_bb(bb); !gsi_end_p(gsi); gsi_next(&gsi)) { in stackleak_instrument_execute() 210 if (bb == entry_bb) in stackleak_instrument_execute() 245 bb = entry_bb; in stackleak_instrument_execute() 246 if (!single_pred_p(bb)) { in stackleak_instrument_execute() [all …]
|
| /linux/net/netfilter/ |
| H A D | nft_set_pipapo.c | 449 if (likely(f->bb == 8)) in pipapo_get_slow() 679 * @bb: Number of bits grouped together in lookup table buckets 684 static ssize_t lt_calculate_size(unsigned int groups, unsigned int bb, in lt_calculate_size() argument 687 ssize_t ret = groups * NFT_PIPAPO_BUCKETS(bb) * sizeof(long); in lt_calculate_size() 736 lt_size = lt_calculate_size(f->groups, f->bb, new_bucket_size); in pipapo_resize() 748 for (bucket = 0; bucket < NFT_PIPAPO_BUCKETS(f->bb); bucket++) { in pipapo_resize() 789 pos += f->bsize * NFT_PIPAPO_BUCKETS(f->bb) * group; in pipapo_bucket_set() 923 unsigned int groups, bb; in pipapo_lt_bits_adjust() local 927 lt_size = f->groups * NFT_PIPAPO_BUCKETS(f->bb) * f->bsize * in pipapo_lt_bits_adjust() 930 if (f->bb in pipapo_lt_bits_adjust() [all...] |
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | selftest_ring_submission.c | 208 struct i915_vma *bb; in __live_ctx_switch_wa() 212 bb = create_wally(engine); in __live_ctx_switch_wa() 213 if (IS_ERR(bb)) in __live_ctx_switch_wa() 214 return PTR_ERR(bb); in __live_ctx_switch_wa() 216 result = i915_gem_object_pin_map_unlocked(bb->obj, I915_MAP_WC); in __live_ctx_switch_wa() 218 intel_context_put(bb->private); in __live_ctx_switch_wa() 219 i915_vma_unpin_and_release(&bb, 0); in __live_ctx_switch_wa() 224 engine->wa_ctx.vma = bb; in __live_ctx_switch_wa() 207 struct i915_vma *bb; __live_ctx_switch_wa() local
|
| /linux/tools/testing/selftests/net/forwarding/ |
| H A D | tc_tunnel_key.sh | 117 $MZ h1-et -c 1 -p 930 -a 00:aa:bb:cc:dd:ee -b 00:ee:dd:cc:bb:aa -t ip -q 121 $MZ h1-et -c 1 -p 931 -a 00:aa:bb:cc:dd:ee -b 00:ee:dd:cc:bb:aa -t ip -q 131 $MZ h1-et -c 1 -p 931 -a 00:aa:bb:cc:dd:ee -b 00:ee:dd:cc:bb:aa -t ip -q
|
| /linux/drivers/pci/ |
| H A D | p2pdma.c | 685 struct pci_dev *a = provider, *b = client, *bb; in calc_map_type_and_dist() 709 bb = b; in calc_map_type_and_dist() 711 while (bb) { in calc_map_type_and_dist() 712 if (a == bb) in calc_map_type_and_dist() 715 bb = pci_upstream_bridge(bb); in calc_map_type_and_dist() 727 bb = b; in calc_map_type_and_dist() 729 while (bb) { in calc_map_type_and_dist() 730 if (a == bb) in calc_map_type_and_dist() 733 if (pci_bridge_has_acs_redir(bb)) { in calc_map_type_and_dist() 677 struct pci_dev *a = provider, *b = client, *bb; calc_map_type_and_dist() local [all...] |
| /linux/lib/crypto/powerpc/ |
| H A D | curve25519.h | 139 fe51 da, cb, aa, bb; in curve25519_fe51() local 152 fsqr(bb, b); // BB = B^2 in curve25519_fe51() 157 fsub(e, aa, bb); // E = AA - BB in curve25519_fe51() 158 fmul(x2, aa, bb); // x2 = AA * BB in curve25519_fe51() 165 fadd(b, bb, z3); // BB + 121666 * E in curve25519_fe51()
|
| /linux/arch/parisc/kernel/ |
| H A D | relocate_kernel.S | 83 bb,<,n %r3, 31 - IND_DONE_BIT, boot 84 bb,>=,n %r3, 31 - IND_INDIRECTION_BIT, .Lnotind 90 bb,>=,n %r3, 31 - IND_DESTINATION_BIT, .Lnotdest 95 bb,>= %r3, 31 - IND_SOURCE_BIT, .Lloop
|