/linux/fs/ntfs3/ |
H A D | run.c | 23 CLST lcn; /* Logical cluster number. */ member 124 if (n->lcn != SPARSE_LCN) in run_consolidate() 125 n->lcn += dl; in run_consolidate() 133 if ((n->lcn == SPARSE_LCN) != (r->lcn == SPARSE_LCN)) { in run_consolidate() 144 if (n->lcn != SPARSE_LCN && n->lcn != r->lcn + r->len) in run_consolidate() 192 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, in run_lookup_entry() argument 215 *lcn = r->lcn == SPARSE_LCN ? SPARSE_LCN : (r->lcn + gap); in run_lookup_entry() 241 if (r->lcn != SPARSE_LCN) in run_truncate_head() 242 r->lcn += dlen; in run_truncate_head() 316 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, in run_add_entry() argument [all …]
|
H A D | fsntfs.c | 352 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len, in ntfs_look_for_free_space() argument 380 lcn = wnd_zone_bit(wnd); in ntfs_look_for_free_space() 383 wnd_zone_set(wnd, lcn + alen, zlen - alen); in ntfs_look_for_free_space() 385 err = wnd_set_used(wnd, lcn, alen); in ntfs_look_for_free_space() 389 alcn = lcn; in ntfs_look_for_free_space() 396 if (!lcn) in ntfs_look_for_free_space() 397 lcn = sbi->used.next_free_lcn; in ntfs_look_for_free_space() 399 if (lcn >= wnd->nbits) in ntfs_look_for_free_space() 400 lcn = 0; in ntfs_look_for_free_space() 402 alen = wnd_find(wnd, len, lcn, BITMAP_FIND_MARK_AS_USED, &alcn); in ntfs_look_for_free_space() [all …]
|
H A D | attrib.c | 97 CLST vcn_next, vcn0 = vcn, lcn, clen, dn = 0; in run_deallocate_ex() local 103 if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) { in run_deallocate_ex() 119 if (lcn != SPARSE_LCN) { in run_deallocate_ex() 122 mark_as_free_ex(sbi, lcn, clen, trim); in run_deallocate_ex() 132 if (!run_get_entry(run, ++idx, &vcn, &lcn, &clen) || in run_deallocate_ex() 150 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, in attr_allocate_clusters() argument 159 err = ntfs_look_for_free_space(sbi, lcn, len + pre, &lcn, &flen, in attr_allocate_clusters() 175 *new_lcn = lcn; in attr_allocate_clusters() 181 if (!run_add_entry(run, vcn, lcn, flen, opt & ALLOCATE_MFT)) { in attr_allocate_clusters() 183 mark_as_free_ex(sbi, lcn, len, false); in attr_allocate_clusters() [all …]
|
H A D | file.c | 199 CLST vcn, lcn, clen; in ntfs_extend_initialized_size() local 205 err = attr_data_get_block(ni, vcn, 1, &lcn, &clen, NULL, in ntfs_extend_initialized_size() 210 if (lcn == SPARSE_LCN) { in ntfs_extend_initialized_size() 365 CLST lcn, len; in ntfs_file_mmap() local 371 err = attr_data_get_block(ni, vcn, 1, &lcn, in ntfs_file_mmap() 431 CLST lcn, clen; in ntfs_extend() local 442 err = attr_data_get_block(ni, vcn, cend_v - vcn, &lcn, in ntfs_extend() 451 err = attr_data_get_block(ni, vcn, cend - vcn, &lcn, in ntfs_extend() 731 CLST lcn, clen; in ntfs_fallocate() local 743 &lcn, &clen, &new, in ntfs_fallocate() [all …]
|
H A D | bitmap.c | 510 CLST lcn, clen; in wnd_rescan() local 546 &lcn, &clen, NULL)) { in wnd_rescan() 551 lbo = ((u64)lcn << cluster_bits) + off; in wnd_rescan() 683 CLST lcn, clen; in wnd_map() local 692 if (!run_lookup_entry(&wnd->run, vbo >> sbi->cluster_bits, &lcn, &clen, in wnd_map() 697 lbo = ((u64)lcn << sbi->cluster_bits) + (vbo & sbi->cluster_mask); in wnd_map() 1388 void wnd_zone_set(struct wnd_bitmap *wnd, size_t lcn, size_t len) in wnd_zone_set() argument 1396 wnd_remove_free_ext(wnd, lcn, len); in wnd_zone_set() 1398 wnd->zone_bit = lcn; in wnd_zone_set() 1399 wnd->zone_end = lcn + len; in wnd_zone_set() [all …]
|
H A D | super.c | 1164 CLST vcn, lcn, len; in ntfs_fill_super() local 1384 for (i = 0; run_get_entry(&ni->file.run, i, &vcn, &lcn, &len); i++) { in ntfs_fill_super() 1385 if (lcn == SPARSE_LCN) in ntfs_fill_super() 1393 if (wnd_set_used_safe(&sbi->used.bitmap, lcn, len, &tt) || tt) { in ntfs_fill_super() 1621 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len) in ntfs_unmap_meta() argument 1625 sector_t devblock = (u64)lcn * sbi->blocks_per_cluster; in ntfs_unmap_meta() 1650 int ntfs_discard(struct ntfs_sb_info *sbi, CLST lcn, CLST len) in ntfs_discard() argument 1656 if (sbi->used.next_free_lcn == lcn + len) in ntfs_discard() 1657 sbi->used.next_free_lcn = lcn; in ntfs_discard() 1665 lbo = (u64)lcn << sbi->cluster_bits; in ntfs_discard()
|
H A D | ntfs_fs.h | 424 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, 435 int attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn, 604 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len, 662 void mark_as_free_ex(struct ntfs_sb_info *sbi, CLST lcn, CLST len, bool trim); 798 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, 803 bool run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, 808 CLST *lcn, CLST *len); 830 void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len);
|
H A D | frecord.c | 1918 CLST lcn, clen; in ni_fiemap() local 1963 ok = run_lookup_entry(&run, vcn, &lcn, &clen, &idx); in ni_fiemap() 1967 ok = run_get_entry(&run, ++idx, &vcn, &lcn, &clen) && in ni_fiemap() 1981 ok = run_lookup_entry(&run, vcn, &lcn, &clen, &idx); in ni_fiemap() 1994 if (lcn == SPARSE_LCN) { in ni_fiemap() 2019 lbo = (u64)lcn << cluster_bits; in ni_fiemap() 2173 CLST vcn, cend, lcn, clen, end; in ni_decompress_file() local 2221 err = attr_data_get_block(ni, vcn, cend - vcn, &lcn, in ni_decompress_file()
|
H A D | fslog.c | 228 __le64 lcn; member 3755 CLST rno, lcn, lcn0, len0, clen; in log_replay() local 4554 u64 lcn0 = le64_to_cpu(r->lcn); in log_replay() 4852 lcn = le64_to_cpu(dp->page_lcns[i]); in log_replay() 4855 lcn0 != lcn) && in log_replay() 4856 !run_add_entry(oa->run1, vcn, lcn, 1, false)) { in log_replay() 4945 if (!run_lookup_entry(oa->run1, vcn, &lcn, NULL, NULL) || in log_replay() 4946 lcn == SPARSE_LCN) { in log_replay() 5069 &lcn, &clen, NULL); in log_replay() 5076 if (is_mapped && lcn != SPARSE_LCN && clen >= t16) in log_replay()
|
H A D | index.c | 379 CLST lcn, clen, vcn, vcn_next; in scan_nres_bitmap() local 393 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap() 408 ok = run_lookup_entry(run, vcn, &lcn, &clen, &idx); in scan_nres_bitmap() 415 block = (sector_t)lcn * sbi->blocks_per_cluster; in scan_nres_bitmap() 452 ok = run_get_entry(run, ++idx, &vcn, &lcn, &clen) && vcn == vcn_next; in scan_nres_bitmap()
|
H A D | inode.c | 574 CLST vcn, lcn, len; in ntfs_get_block_vbo() local 602 err = attr_data_get_block(ni, vcn, 1, &lcn, &len, create ? &new : NULL, in ntfs_get_block_vbo() 612 if (lcn >= sbi->used.bitmap.nbits) { in ntfs_get_block_vbo() 625 lbo = ((u64)lcn << cluster_bits) + off; in ntfs_get_block_vbo()
|
/linux/fs/erofs/ |
H A D | zmap.c | 13 unsigned long lcn; member 24 unsigned long lcn) in z_erofs_load_full_lcluster() argument 30 lcn * sizeof(struct z_erofs_lcluster_index); in z_erofs_load_full_lcluster() 37 m->lcn = lcn; in z_erofs_load_full_lcluster() 211 unsigned long lcn, bool lookahead) in z_erofs_load_compact_lcluster() argument 222 if (lcn >= totalidx || vi->z_logical_clusterbits > 14) in z_erofs_load_compact_lcluster() 225 m->lcn = lcn; in z_erofs_load_compact_lcluster() 238 if (lcn < compacted_4b_initial) { in z_erofs_load_compact_lcluster() 243 lcn -= compacted_4b_initial; in z_erofs_load_compact_lcluster() 245 if (lcn < compacted_2b) { in z_erofs_load_compact_lcluster() [all …]
|
/linux/drivers/net/wireless/broadcom/b43/ |
H A D | phy_lcn.c | 341 tx_pwr_idx = dev->phy.lcn->tx_pwr_curr_idx; in b43_phy_lcn_sense_setup() 537 if (!dev->phy.lcn->hw_pwr_ctl_capable) { in b43_phy_lcn_tx_pwr_ctl_init() 675 dev->phy.lcn = phy_lcn; in b43_phy_lcn_op_allocate() 683 struct b43_phy_lcn *phy_lcn = phy->lcn; in b43_phy_lcn_op_free() 686 phy->lcn = NULL; in b43_phy_lcn_op_free() 692 struct b43_phy_lcn *phy_lcn = phy->lcn; in b43_phy_lcn_op_prepare_structs()
|
H A D | phy_common.h | 215 struct b43_phy_lcn *lcn; member
|