| /linux/drivers/md/dm-vdo/ |
| H A D | repair.c | 107 physical_block_number_t pbn; member 147 if (mapping1->block_map_slot.pbn != mapping2->block_map_slot.pbn) in mapping_is_less_than() 148 return mapping1->block_map_slot.pbn < mapping2->block_map_slot.pbn; in mapping_is_less_than() 437 if (mapping.pbn == VDO_ZERO_BLOCK) in process_slot() 440 if (!vdo_is_physical_data_block(depot, mapping.pbn)) { in process_slot() 449 result = vdo_adjust_reference_count_for_rebuild(depot, mapping.pbn, in process_slot() 457 slot, (unsigned long long) mapping.pbn); in process_slot() 484 if (vdo_get_block_map_page_pbn(page) == repair->last_slot.pbn) { in rebuild_reference_counts_from_page() 519 physical_block_number_t pbn = VDO_ZERO_BLOCK; in get_pbn_to_fetch() local 524 while ((pbn == VDO_ZERO_BLOCK) && (repair->page_to_fetch < repair->leaf_pages)) in get_pbn_to_fetch() [all …]
|
| H A D | block-map.c | 191 info->pbn = NO_PAGE; in initialize_info() 381 static int __must_check set_info_pbn(struct page_info *info, physical_block_number_t pbn) in set_info_pbn() argument 386 int result = VDO_ASSERT((pbn == NO_PAGE) || (info->pbn == NO_PAGE), in set_info_pbn() 391 if (info->pbn != NO_PAGE) in set_info_pbn() 392 vdo_int_map_remove(cache->page_map, info->pbn); in set_info_pbn() 394 info->pbn = pbn; in set_info_pbn() 396 if (pbn != NO_PAGE) { in set_info_pbn() 397 result = vdo_int_map_put(cache->page_map, pbn, info, true, NULL); in set_info_pbn() 450 physical_block_number_t pbn) in find_page() argument 452 if ((cache->last_found != NULL) && (cache->last_found->pbn == pbn)) in find_page() [all …]
|
| H A D | vio.c | 45 physical_block_number_t pbn = bio->bi_iter.bi_sector / VDO_SECTORS_PER_BLOCK; in pbn_from_vio_bio() local 47 return ((pbn == VDO_GEOMETRY_BLOCK_LOCATION) ? pbn : pbn + vdo->geometry.bio_offset); in pbn_from_vio_bio() 174 blk_opf_t bi_opf, physical_block_number_t pbn) in vdo_set_bio_properties() argument 179 pbn -= vdo->geometry.bio_offset; in vdo_set_bio_properties() 180 vio->bio_zone = ((pbn / config->thread_counts.bio_rotation_interval) % in vdo_set_bio_properties() 186 bio->bi_iter.bi_sector = pbn * VDO_SECTORS_PER_BLOCK; in vdo_set_bio_properties() 195 blk_opf_t bi_opf, physical_block_number_t pbn) in vio_reset_bio() argument 198 callback, bi_opf, pbn); in vio_reset_bio() 202 blk_opf_t bi_opf, physical_block_number_t pbn) in vio_reset_bio_with_size() argument 210 vdo_set_bio_properties(bio, vio, callback, bi_opf, pbn); in vio_reset_bio_with_size() [all …]
|
| H A D | dump.c | 163 wait_on, data_vio, data_vio->allocation.pbn, data_vio->logical.lbn, in dump_vio_waiters() 164 data_vio->duplicate.pbn, get_data_vio_operation_name(data_vio)); in dump_vio_waiters() 169 data_vio, data_vio->allocation.pbn, data_vio->logical.lbn, in dump_vio_waiters() 170 data_vio->duplicate.pbn, in dump_vio_waiters() 239 data_vio->allocation.pbn, data_vio->logical.lbn, in dump_data_vio() 240 data_vio->duplicate.pbn); in dump_data_vio() 244 data_vio->allocation.pbn, data_vio->logical.lbn); in dump_data_vio()
|
| H A D | block-map.h | 139 physical_block_number_t pbn; member 173 physical_block_number_t pbn; member 290 typedef int (*vdo_entry_callback_fn)(physical_block_number_t pbn, 302 struct block_map_zone *zone, physical_block_number_t pbn, 320 physical_block_number_t pbn, 364 physical_block_number_t pbn,
|
| H A D | slab-depot.h | 518 physical_block_number_t pbn, 525 physical_block_number_t pbn, 535 physical_block_number_t pbn, 548 physical_block_number_t pbn); 566 physical_block_number_t pbn); 569 physical_block_number_t pbn); 572 physical_block_number_t pbn);
|
| H A D | encodings.h | 155 __le64 pbn; member 840 vdo_pack_block_map_entry(physical_block_number_t pbn, enum block_mapping_state mapping_state) in vdo_pack_block_map_entry() argument 844 .pbn_high_nibble = ((pbn >> 32) & 0x0F), in vdo_pack_block_map_entry() 845 .pbn_low_word = __cpu_to_le32(pbn & UINT_MAX), in vdo_pack_block_map_entry() 855 .pbn = ((high4 << 32) | low32), in vdo_unpack_block_map_entry() 867 if (location->pbn == VDO_ZERO_BLOCK) in vdo_is_valid_location() 876 return __le64_to_cpu(page->header.pbn); in vdo_get_block_map_page_pbn() 880 physical_block_number_t pbn, 885 physical_block_number_t pbn); 911 .pbn_high_nibble = (entry->slot.pbn >> 32) & 0x0F, in vdo_pack_recovery_journal_entry() [all …]
|
| H A D | physical-zone.c | 417 physical_block_number_t pbn) in vdo_get_physical_zone_pbn_lock() argument 419 return ((zone == NULL) ? NULL : vdo_int_map_get(zone->pbn_operations, pbn)); in vdo_get_physical_zone_pbn_lock() 438 physical_block_number_t pbn, in vdo_attempt_physical_zone_pbn_lock() argument 455 result = vdo_int_map_put(zone->pbn_operations, pbn, new_lock, false, in vdo_attempt_physical_zone_pbn_lock() 466 (unsigned long long) pbn); in vdo_attempt_physical_zone_pbn_lock() 492 result = vdo_allocate_block(allocation->zone->allocator, &allocation->pbn); in allocate_and_lock_block() 496 result = vdo_attempt_physical_zone_pbn_lock(allocation->zone, allocation->pbn, in allocate_and_lock_block() 505 (unsigned long long) allocation->pbn, in allocate_and_lock_block()
|
| H A D | data-vio.c | 1338 (unsigned long long) data_vio->new_mapped.pbn, in enter_read_only_mode() 1339 (unsigned long long) data_vio->mapped.pbn, in enter_read_only_mode() 1340 (unsigned long long) data_vio->allocation.pbn, in enter_read_only_mode() 1387 VDO_ASSERT_LOG_ONLY((allocation->pbn == VDO_ZERO_BLOCK), in data_vio_allocate_data_block() 1407 physical_block_number_t locked_pbn = allocation->pbn; in release_data_vio_allocation_lock() 1412 allocation->pbn = VDO_ZERO_BLOCK; in release_data_vio_allocation_lock() 1551 if (data_vio->mapped.pbn == VDO_ZERO_BLOCK) { in read_block() 1560 read_endio, REQ_OP_READ, data_vio->mapped.pbn); in read_block() 1566 data_vio->mapped.pbn); in read_block() 1575 data_vio->mapped.pbn); in read_block() [all …]
|
| H A D | slab-depot.c | 320 physical_block_number_t pbn; in launch_write() local 345 pbn = (depot->summary_origin + in launch_write() 348 vdo_submit_metadata_vio(&block->vio, pbn, write_slab_summary_endio, in launch_write() 903 static void add_entry(struct slab_journal *journal, physical_block_number_t pbn, in add_entry() argument 934 pbn - journal->slab->start, operation, increment); in add_entry() 957 bool vdo_attempt_replay_into_slab(struct vdo_slab *slab, physical_block_number_t pbn, in vdo_attempt_replay_into_slab() argument 999 add_entry(journal, pbn, operation, increment, expanded); in vdo_attempt_replay_into_slab() 1188 physical_block_number_t pbn; in write_reference_block() local 1196 pbn = (block->slab->ref_counts_origin + block_offset); in write_reference_block() 1216 vdo_submit_metadata_vio(&pooled->vio, pbn, write_reference_block_endio, in write_reference_block() [all …]
|
| H A D | physical-zone.h | 100 physical_block_number_t pbn); 103 physical_block_number_t pbn,
|
| H A D | vio.h | 124 blk_opf_t bi_opf, physical_block_number_t pbn); 127 blk_opf_t bi_opf, physical_block_number_t pbn); 129 blk_opf_t bi_opf, physical_block_number_t pbn);
|
| H A D | dedupe.c | 512 data_vio->is_duplicate = (source.pbn != VDO_ZERO_BLOCK); in set_duplicate_location() 694 vdo_release_physical_zone_pbn_lock(agent->duplicate.zone, agent->duplicate.pbn, in unlock_duplicate_pbn() 1206 agent->duplicate.pbn); in start_verifying() 1282 struct vdo_slab *slab = vdo_get_slab(depot, agent->duplicate.pbn); in acquire_provisional_reference() 1283 int result = vdo_acquire_provisional_reference(slab, agent->duplicate.pbn, lock); in acquire_provisional_reference() 1292 agent->duplicate.pbn, lock); in acquire_provisional_reference() 1325 increment_limit = vdo_get_increment_limit(depot, agent->duplicate.pbn); in lock_duplicate_pbn() 1336 result = vdo_attempt_physical_zone_pbn_lock(zone, agent->duplicate.pbn, in lock_duplicate_pbn() 1620 advice->pbn = get_unaligned_le64(&encoding->data[offset]); in decode_uds_advice() 1625 if ((advice->state == VDO_MAPPING_STATE_UNMAPPED) || (advice->pbn == VDO_ZERO_BLOCK)) { in decode_uds_advice() [all …]
|
| H A D | types.h | 141 physical_block_number_t pbn; member 166 physical_block_number_t pbn; member
|
| H A D | packer.c | 260 .pbn = allocation->pbn, in release_compressed_write_waiter() 306 (unsigned long long) allocation->pbn); in handle_compressed_write_error() 469 REQ_OP_WRITE, agent->allocation.pbn); in write_bin()
|
| H A D | recovery-journal.c | 1203 has_decrement = (data_vio->decrement_updater.zpbn.pbn != VDO_ZERO_BLOCK); in continue_committed_waiter() 1204 if ((data_vio->increment_updater.zpbn.pbn != VDO_ZERO_BLOCK) || !has_decrement) in continue_committed_waiter() 1351 .pbn = data_vio->increment_updater.zpbn.pbn, in add_queued_recovery_entries() 1355 .pbn = data_vio->decrement_updater.zpbn.pbn, in add_queued_recovery_entries()
|
| H A D | encodings.c | 336 physical_block_number_t pbn, in vdo_format_block_map_page() argument 344 page->header.pbn = __cpu_to_le64(pbn); in vdo_format_block_map_page() 351 physical_block_number_t pbn) in vdo_validate_block_map_page() argument 360 if (pbn != vdo_get_block_map_page_pbn(page)) in vdo_validate_block_map_page()
|
| /linux/fs/nilfs2/ |
| H A D | gcinode.c | 56 sector_t pbn, __u64 vbn, in nilfs_gccache_submit_read_data() argument 69 if (pbn == 0) { in nilfs_gccache_submit_read_data() 72 err = nilfs_dat_translate(nilfs->ns_dat, vbn, &pbn); in nilfs_gccache_submit_read_data() 85 bh->b_blocknr = pbn; in nilfs_gccache_submit_read_data() 120 int nilfs_gccache_submit_read_node(struct inode *inode, sector_t pbn, in nilfs_gccache_submit_read_node() argument 126 ret = nilfs_btnode_submit_block(btnc_inode->i_mapping, vbn ? : pbn, pbn, in nilfs_gccache_submit_read_node() 127 REQ_OP_READ, out_bh, &pbn); in nilfs_gccache_submit_read_node()
|
| /linux/drivers/bluetooth/ |
| H A D | hci_ag6xx.c | 254 struct pbn_entry *pbn = (void *)fw_ptr; in ag6xx_setup() local 257 if (pbn->addr == 0xffffffff) { in ag6xx_setup() 263 addr = le32_to_cpu(pbn->addr); in ag6xx_setup() 264 plen = le32_to_cpu(pbn->plen); in ag6xx_setup() 266 if (fw->data + fw->size <= pbn->data + plen) { in ag6xx_setup() 274 err = intel_mem_write(hdev, addr, plen, pbn->data); in ag6xx_setup() 280 fw_ptr = pbn->data + plen; in ag6xx_setup()
|
| /linux/drivers/gpu/drm/amd/display/amdgpu_dm/ |
| H A D | amdgpu_dm_mst_types.c | 948 params[i].timing->dsc_cfg.mst_pbn = vars[i + k].pbn; in set_dsc_configs_from_fairness_vars() 962 vars[i + k].pbn); in set_dsc_configs_from_fairness_vars() 966 static int bpp_x16_from_pbn(struct dsc_mst_fairness_params param, int pbn) in bpp_x16_from_pbn() argument 977 kbps = div_u64((u64)pbn * 994 * 8 * 54, 64); in bpp_x16_from_pbn() 1011 kbps_to_peak_pbn(params[i].bw_range.max_kbps, fec_overhead_multiplier_x1000) - vars[i + k].pbn; in increase_dsc_bpp() 1038 link_timeslots_used += DIV_ROUND_UP(vars[i + k].pbn, dfixed_trunc(mst_state->pbn_div)); in increase_dsc_bpp() 1044 vars[next_index].pbn += fair_pbn_alloc; in increase_dsc_bpp() 1048 vars[next_index].pbn); in increase_dsc_bpp() 1054 vars[next_index].bpp_x16 = bpp_x16_from_pbn(params[next_index], vars[next_index].pbn); in increase_dsc_bpp() 1056 vars[next_index].pbn -= fair_pbn_alloc; in increase_dsc_bpp() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| H A D | tu102.c | 34 tu102_sor_dp_vcpi(struct nvkm_ior *sor, int head, u8 slot, u8 slot_nr, u16 pbn, u16 aligned) in tu102_sor_dp_vcpi() argument 39 nvkm_mask(device, 0x61657c + hoff, 0xffffffff, (aligned << 16) | pbn); in tu102_sor_dp_vcpi()
|
| /linux/drivers/gpu/drm/nouveau/include/nvif/ |
| H A D | outp.h | 115 u8 start_slot, u8 num_slots, u16 pbn, u16 aligned_pbn);
|
| H A D | if0012.h | 290 __u16 pbn; member
|
| /linux/drivers/gpu/drm/amd/display/dc/link/ |
| H A D | link_dpms.c | 1437 struct fixed31_32 pbn; in allocate_mst_payload() local 1515 pbn = get_pbn_from_timing(pipe_ctx); in allocate_mst_payload() 1516 avg_time_slots_per_mtp = dc_fixpt_div(pbn, pbn_per_slot); in allocate_mst_payload() 1746 struct fixed31_32 pbn; in link_reduce_mst_payload() local 1755 pbn = get_pbn_from_bw_in_kbps(bw_in_kbps); in link_reduce_mst_payload() 1756 avg_time_slots_per_mtp = dc_fixpt_div(pbn, pbn_per_slot); in link_reduce_mst_payload() 1834 struct fixed31_32 pbn; in link_increase_mst_payload() local 1901 pbn = get_pbn_from_bw_in_kbps(bw_in_kbps); in link_increase_mst_payload() 1903 avg_time_slots_per_mtp = dc_fixpt_div(pbn, pbn_per_slot); in link_increase_mst_payload()
|
| /linux/drivers/gpu/drm/nouveau/dispnv50/ |
| H A D | atom.h | 129 int pbn; member
|