1 /* SPDX-License-Identifier: GPL-2.0 */ 2 3 #ifndef BTRFS_SPACE_INFO_H 4 #define BTRFS_SPACE_INFO_H 5 6 #include "volumes.h" 7 8 struct btrfs_space_info { 9 spinlock_t lock; 10 11 u64 total_bytes; /* total bytes in the space, 12 this doesn't take mirrors into account */ 13 u64 bytes_used; /* total bytes used, 14 this doesn't take mirrors into account */ 15 u64 bytes_pinned; /* total bytes pinned, will be freed when the 16 transaction finishes */ 17 u64 bytes_reserved; /* total bytes the allocator has reserved for 18 current allocations */ 19 u64 bytes_may_use; /* number of bytes that may be used for 20 delalloc/allocations */ 21 u64 bytes_readonly; /* total bytes that are read only */ 22 /* Total bytes in the space, but only accounts active block groups. */ 23 u64 active_total_bytes; 24 u64 bytes_zone_unusable; /* total bytes that are unusable until 25 resetting the device zone */ 26 27 u64 max_extent_size; /* This will hold the maximum extent size of 28 the space info if we had an ENOSPC in the 29 allocator. */ 30 /* Chunk size in bytes */ 31 u64 chunk_size; 32 33 /* 34 * Once a block group drops below this threshold (percents) we'll 35 * schedule it for reclaim. 36 */ 37 int bg_reclaim_threshold; 38 39 int clamp; /* Used to scale our threshold for preemptive 40 flushing. The value is >> clamp, so turns 41 out to be a 2^clamp divisor. */ 42 43 unsigned int full:1; /* indicates that we cannot allocate any more 44 chunks for this space */ 45 unsigned int chunk_alloc:1; /* set if we are allocating a chunk */ 46 47 unsigned int flush:1; /* set if we are trying to make space */ 48 49 unsigned int force_alloc; /* set if we need to force a chunk 50 alloc for this space */ 51 52 u64 disk_used; /* total bytes used on disk */ 53 u64 disk_total; /* total bytes on disk, takes mirrors into 54 account */ 55 56 u64 flags; 57 58 struct list_head list; 59 /* Protected by the spinlock 'lock'. */ 60 struct list_head ro_bgs; 61 struct list_head priority_tickets; 62 struct list_head tickets; 63 64 /* 65 * Size of space that needs to be reclaimed in order to satisfy pending 66 * tickets 67 */ 68 u64 reclaim_size; 69 70 /* 71 * tickets_id just indicates the next ticket will be handled, so note 72 * it's not stored per ticket. 73 */ 74 u64 tickets_id; 75 76 struct rw_semaphore groups_sem; 77 /* for block groups in our same type */ 78 struct list_head block_groups[BTRFS_NR_RAID_TYPES]; 79 80 struct kobject kobj; 81 struct kobject *block_group_kobjs[BTRFS_NR_RAID_TYPES]; 82 }; 83 84 struct reserve_ticket { 85 u64 bytes; 86 int error; 87 bool steal; 88 struct list_head list; 89 wait_queue_head_t wait; 90 }; 91 92 static inline bool btrfs_mixed_space_info(struct btrfs_space_info *space_info) 93 { 94 return ((space_info->flags & BTRFS_BLOCK_GROUP_METADATA) && 95 (space_info->flags & BTRFS_BLOCK_GROUP_DATA)); 96 } 97 98 /* 99 * 100 * Declare a helper function to detect underflow of various space info members 101 */ 102 #define DECLARE_SPACE_INFO_UPDATE(name, trace_name) \ 103 static inline void \ 104 btrfs_space_info_update_##name(struct btrfs_fs_info *fs_info, \ 105 struct btrfs_space_info *sinfo, \ 106 s64 bytes) \ 107 { \ 108 const u64 abs_bytes = (bytes < 0) ? -bytes : bytes; \ 109 lockdep_assert_held(&sinfo->lock); \ 110 trace_update_##name(fs_info, sinfo, sinfo->name, bytes); \ 111 trace_btrfs_space_reservation(fs_info, trace_name, \ 112 sinfo->flags, abs_bytes, \ 113 bytes > 0); \ 114 if (bytes < 0 && sinfo->name < -bytes) { \ 115 WARN_ON(1); \ 116 sinfo->name = 0; \ 117 return; \ 118 } \ 119 sinfo->name += bytes; \ 120 } 121 122 DECLARE_SPACE_INFO_UPDATE(bytes_may_use, "space_info"); 123 DECLARE_SPACE_INFO_UPDATE(bytes_pinned, "pinned"); 124 125 int btrfs_init_space_info(struct btrfs_fs_info *fs_info); 126 void btrfs_update_space_info(struct btrfs_fs_info *info, u64 flags, 127 u64 total_bytes, u64 bytes_used, 128 u64 bytes_readonly, u64 bytes_zone_unusable, 129 bool active, struct btrfs_space_info **space_info); 130 void btrfs_update_space_info_chunk_size(struct btrfs_space_info *space_info, 131 u64 chunk_size); 132 struct btrfs_space_info *btrfs_find_space_info(struct btrfs_fs_info *info, 133 u64 flags); 134 u64 __pure btrfs_space_info_used(struct btrfs_space_info *s_info, 135 bool may_use_included); 136 void btrfs_clear_space_info_full(struct btrfs_fs_info *info); 137 void btrfs_dump_space_info(struct btrfs_fs_info *fs_info, 138 struct btrfs_space_info *info, u64 bytes, 139 int dump_block_groups); 140 int btrfs_reserve_metadata_bytes(struct btrfs_fs_info *fs_info, 141 struct btrfs_block_rsv *block_rsv, 142 u64 orig_bytes, 143 enum btrfs_reserve_flush_enum flush); 144 void btrfs_try_granting_tickets(struct btrfs_fs_info *fs_info, 145 struct btrfs_space_info *space_info); 146 int btrfs_can_overcommit(struct btrfs_fs_info *fs_info, 147 struct btrfs_space_info *space_info, u64 bytes, 148 enum btrfs_reserve_flush_enum flush); 149 150 static inline void btrfs_space_info_free_bytes_may_use( 151 struct btrfs_fs_info *fs_info, 152 struct btrfs_space_info *space_info, 153 u64 num_bytes) 154 { 155 spin_lock(&space_info->lock); 156 btrfs_space_info_update_bytes_may_use(fs_info, space_info, -num_bytes); 157 btrfs_try_granting_tickets(fs_info, space_info); 158 spin_unlock(&space_info->lock); 159 } 160 int btrfs_reserve_data_bytes(struct btrfs_fs_info *fs_info, u64 bytes, 161 enum btrfs_reserve_flush_enum flush); 162 #endif /* BTRFS_SPACE_INFO_H */ 163