Linux Audio

Check our new training course

Linux kernel drivers training

Mar 31-Apr 9, 2025, special US time zones
Register
Loading...
v6.2
  1/* SPDX-License-Identifier: GPL-2.0 */
  2
  3#ifndef BTRFS_EXTENT_IO_TREE_H
  4#define BTRFS_EXTENT_IO_TREE_H
  5
  6#include "misc.h"
  7
  8struct extent_changeset;
  9struct io_failure_record;
 10
 11/* Bits for the extent state */
 12enum {
 13	ENUM_BIT(EXTENT_DIRTY),
 14	ENUM_BIT(EXTENT_UPTODATE),
 15	ENUM_BIT(EXTENT_LOCKED),
 16	ENUM_BIT(EXTENT_NEW),
 17	ENUM_BIT(EXTENT_DELALLOC),
 18	ENUM_BIT(EXTENT_DEFRAG),
 19	ENUM_BIT(EXTENT_BOUNDARY),
 20	ENUM_BIT(EXTENT_NODATASUM),
 21	ENUM_BIT(EXTENT_CLEAR_META_RESV),
 22	ENUM_BIT(EXTENT_NEED_WAIT),
 23	ENUM_BIT(EXTENT_NORESERVE),
 24	ENUM_BIT(EXTENT_QGROUP_RESERVED),
 25	ENUM_BIT(EXTENT_CLEAR_DATA_RESV),
 26	/*
 27	 * Must be cleared only during ordered extent completion or on error
 28	 * paths if we did not manage to submit bios and create the ordered
 29	 * extents for the range.  Should not be cleared during page release
 30	 * and page invalidation (if there is an ordered extent in flight),
 31	 * that is left for the ordered extent completion.
 32	 */
 33	ENUM_BIT(EXTENT_DELALLOC_NEW),
 34	/*
 35	 * When an ordered extent successfully completes for a region marked as
 36	 * a new delalloc range, use this flag when clearing a new delalloc
 37	 * range to indicate that the VFS' inode number of bytes should be
 38	 * incremented and the inode's new delalloc bytes decremented, in an
 39	 * atomic way to prevent races with stat(2).
 40	 */
 41	ENUM_BIT(EXTENT_ADD_INODE_BYTES),
 42	/*
 43	 * Set during truncate when we're clearing an entire range and we just
 44	 * want the extent states to go away.
 45	 */
 46	ENUM_BIT(EXTENT_CLEAR_ALL_BITS),
 47};
 48
 49#define EXTENT_DO_ACCOUNTING    (EXTENT_CLEAR_META_RESV | \
 50				 EXTENT_CLEAR_DATA_RESV)
 51#define EXTENT_CTLBITS		(EXTENT_DO_ACCOUNTING | \
 52				 EXTENT_ADD_INODE_BYTES | \
 53				 EXTENT_CLEAR_ALL_BITS)
 54
 55/*
 56 * Redefined bits above which are used only in the device allocation tree,
 57 * shouldn't be using EXTENT_LOCKED / EXTENT_BOUNDARY / EXTENT_CLEAR_META_RESV
 58 * / EXTENT_CLEAR_DATA_RESV because they have special meaning to the bit
 59 * manipulation functions
 60 */
 61#define CHUNK_ALLOCATED				EXTENT_DIRTY
 62#define CHUNK_TRIMMED				EXTENT_DEFRAG
 63#define CHUNK_STATE_MASK			(CHUNK_ALLOCATED |		\
 64						 CHUNK_TRIMMED)
 65
 66enum {
 67	IO_TREE_FS_PINNED_EXTENTS,
 68	IO_TREE_FS_EXCLUDED_EXTENTS,
 69	IO_TREE_BTREE_INODE_IO,
 70	IO_TREE_INODE_IO,
 
 71	IO_TREE_RELOC_BLOCKS,
 72	IO_TREE_TRANS_DIRTY_PAGES,
 73	IO_TREE_ROOT_DIRTY_LOG_PAGES,
 74	IO_TREE_INODE_FILE_EXTENT,
 75	IO_TREE_LOG_CSUM_RANGE,
 76	IO_TREE_SELFTEST,
 77	IO_TREE_DEVICE_ALLOC_STATE,
 78};
 79
 80struct extent_io_tree {
 81	struct rb_root state;
 82	struct btrfs_fs_info *fs_info;
 83	/* Inode associated with this tree, or NULL. */
 84	struct btrfs_inode *inode;
 
 85
 86	/* Who owns this io tree, should be one of IO_TREE_* */
 87	u8 owner;
 88
 89	spinlock_t lock;
 
 90};
 91
 92struct extent_state {
 93	u64 start;
 94	u64 end; /* inclusive */
 95	struct rb_node rb_node;
 96
 97	/* ADD NEW ELEMENTS AFTER THIS */
 98	wait_queue_head_t wq;
 99	refcount_t refs;
100	u32 state;
 
 
101
102#ifdef CONFIG_BTRFS_DEBUG
103	struct list_head leak_list;
104#endif
105};
106
 
 
 
107void extent_io_tree_init(struct btrfs_fs_info *fs_info,
108			 struct extent_io_tree *tree, unsigned int owner);
 
109void extent_io_tree_release(struct extent_io_tree *tree);
110
111int lock_extent(struct extent_io_tree *tree, u64 start, u64 end,
112		struct extent_state **cached);
113
114int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end,
115		    struct extent_state **cached);
 
 
 
 
116
117int __init extent_state_init_cachep(void);
118void __cold extent_state_free_cachep(void);
119
120u64 count_range_bits(struct extent_io_tree *tree,
121		     u64 *start, u64 search_end,
122		     u64 max_bytes, u32 bits, int contig,
123		     struct extent_state **cached_state);
124
125void free_extent_state(struct extent_state *state);
126int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end,
127		   u32 bits, int filled, struct extent_state *cached_state);
 
128int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
129			     u32 bits, struct extent_changeset *changeset);
 
 
 
130int __clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
131		       u32 bits, struct extent_state **cached, gfp_t mask,
132		       struct extent_changeset *changeset);
 
133
134static inline int clear_extent_bit(struct extent_io_tree *tree, u64 start,
135				   u64 end, u32 bits,
136				   struct extent_state **cached)
137{
138	return __clear_extent_bit(tree, start, end, bits, cached,
139				  GFP_NOFS, NULL);
140}
141
142static inline int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end,
143				struct extent_state **cached)
144{
145	return __clear_extent_bit(tree, start, end, EXTENT_LOCKED, cached,
146				  GFP_NOFS, NULL);
 
 
 
 
 
 
 
147}
148
149static inline int clear_extent_bits(struct extent_io_tree *tree, u64 start,
150				    u64 end, u32 bits)
151{
152	return clear_extent_bit(tree, start, end, bits, NULL);
 
 
 
 
 
153}
154
155int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
156			   u32 bits, struct extent_changeset *changeset);
157int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
158		   u32 bits, struct extent_state **cached_state, gfp_t mask);
159
160static inline int set_extent_bits_nowait(struct extent_io_tree *tree, u64 start,
161					 u64 end, u32 bits)
162{
163	return set_extent_bit(tree, start, end, bits, NULL, GFP_NOWAIT);
164}
165
166static inline int set_extent_bits(struct extent_io_tree *tree, u64 start,
167		u64 end, u32 bits)
168{
169	return set_extent_bit(tree, start, end, bits, NULL, GFP_NOFS);
170}
171
172static inline int clear_extent_uptodate(struct extent_io_tree *tree, u64 start,
173		u64 end, struct extent_state **cached_state)
174{
175	return __clear_extent_bit(tree, start, end, EXTENT_UPTODATE,
176				  cached_state, GFP_NOFS, NULL);
177}
178
179static inline int set_extent_dirty(struct extent_io_tree *tree, u64 start,
180		u64 end, gfp_t mask)
181{
182	return set_extent_bit(tree, start, end, EXTENT_DIRTY, NULL, mask);
 
183}
184
185static inline int clear_extent_dirty(struct extent_io_tree *tree, u64 start,
186				     u64 end, struct extent_state **cached)
187{
188	return clear_extent_bit(tree, start, end,
189				EXTENT_DIRTY | EXTENT_DELALLOC |
190				EXTENT_DO_ACCOUNTING, cached);
191}
192
193int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
194		       u32 bits, u32 clear_bits,
195		       struct extent_state **cached_state);
196
197static inline int set_extent_delalloc(struct extent_io_tree *tree, u64 start,
198				      u64 end, u32 extra_bits,
199				      struct extent_state **cached_state)
200{
201	return set_extent_bit(tree, start, end,
202			      EXTENT_DELALLOC | extra_bits,
203			      cached_state, GFP_NOFS);
204}
205
206static inline int set_extent_defrag(struct extent_io_tree *tree, u64 start,
207		u64 end, struct extent_state **cached_state)
208{
209	return set_extent_bit(tree, start, end,
210			      EXTENT_DELALLOC | EXTENT_DEFRAG,
211			      cached_state, GFP_NOFS);
212}
213
214static inline int set_extent_new(struct extent_io_tree *tree, u64 start,
215		u64 end)
216{
217	return set_extent_bit(tree, start, end, EXTENT_NEW, NULL, GFP_NOFS);
 
 
 
 
 
 
 
 
218}
219
220int find_first_extent_bit(struct extent_io_tree *tree, u64 start,
221			  u64 *start_ret, u64 *end_ret, u32 bits,
222			  struct extent_state **cached_state);
223void find_first_clear_extent_bit(struct extent_io_tree *tree, u64 start,
224				 u64 *start_ret, u64 *end_ret, u32 bits);
225int find_contiguous_extent_bit(struct extent_io_tree *tree, u64 start,
226			       u64 *start_ret, u64 *end_ret, u32 bits);
 
 
227bool btrfs_find_delalloc_range(struct extent_io_tree *tree, u64 *start,
228			       u64 *end, u64 max_bytes,
229			       struct extent_state **cached_state);
230void wait_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, u32 bits,
231		     struct extent_state **cached_state);
 
 
 
 
 
 
 
 
 
 
 
 
232
233#endif /* BTRFS_EXTENT_IO_TREE_H */
v5.9
  1/* SPDX-License-Identifier: GPL-2.0 */
  2
  3#ifndef BTRFS_EXTENT_IO_TREE_H
  4#define BTRFS_EXTENT_IO_TREE_H
  5
 
 
  6struct extent_changeset;
  7struct io_failure_record;
  8
  9/* Bits for the extent state */
 10#define EXTENT_DIRTY		(1U << 0)
 11#define EXTENT_UPTODATE		(1U << 1)
 12#define EXTENT_LOCKED		(1U << 2)
 13#define EXTENT_NEW		(1U << 3)
 14#define EXTENT_DELALLOC		(1U << 4)
 15#define EXTENT_DEFRAG		(1U << 5)
 16#define EXTENT_BOUNDARY		(1U << 6)
 17#define EXTENT_NODATASUM	(1U << 7)
 18#define EXTENT_CLEAR_META_RESV	(1U << 8)
 19#define EXTENT_NEED_WAIT	(1U << 9)
 20#define EXTENT_DAMAGED		(1U << 10)
 21#define EXTENT_NORESERVE	(1U << 11)
 22#define EXTENT_QGROUP_RESERVED	(1U << 12)
 23#define EXTENT_CLEAR_DATA_RESV	(1U << 13)
 24#define EXTENT_DELALLOC_NEW	(1U << 14)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 25#define EXTENT_DO_ACCOUNTING    (EXTENT_CLEAR_META_RESV | \
 26				 EXTENT_CLEAR_DATA_RESV)
 27#define EXTENT_CTLBITS		(EXTENT_DO_ACCOUNTING)
 
 
 28
 29/*
 30 * Redefined bits above which are used only in the device allocation tree,
 31 * shouldn't be using EXTENT_LOCKED / EXTENT_BOUNDARY / EXTENT_CLEAR_META_RESV
 32 * / EXTENT_CLEAR_DATA_RESV because they have special meaning to the bit
 33 * manipulation functions
 34 */
 35#define CHUNK_ALLOCATED				EXTENT_DIRTY
 36#define CHUNK_TRIMMED				EXTENT_DEFRAG
 37#define CHUNK_STATE_MASK			(CHUNK_ALLOCATED |		\
 38						 CHUNK_TRIMMED)
 39
 40enum {
 41	IO_TREE_FS_PINNED_EXTENTS,
 42	IO_TREE_FS_EXCLUDED_EXTENTS,
 
 43	IO_TREE_INODE_IO,
 44	IO_TREE_INODE_IO_FAILURE,
 45	IO_TREE_RELOC_BLOCKS,
 46	IO_TREE_TRANS_DIRTY_PAGES,
 47	IO_TREE_ROOT_DIRTY_LOG_PAGES,
 48	IO_TREE_INODE_FILE_EXTENT,
 49	IO_TREE_LOG_CSUM_RANGE,
 50	IO_TREE_SELFTEST,
 
 51};
 52
 53struct extent_io_tree {
 54	struct rb_root state;
 55	struct btrfs_fs_info *fs_info;
 56	void *private_data;
 57	u64 dirty_bytes;
 58	bool track_uptodate;
 59
 60	/* Who owns this io tree, should be one of IO_TREE_* */
 61	u8 owner;
 62
 63	spinlock_t lock;
 64	const struct extent_io_ops *ops;
 65};
 66
 67struct extent_state {
 68	u64 start;
 69	u64 end; /* inclusive */
 70	struct rb_node rb_node;
 71
 72	/* ADD NEW ELEMENTS AFTER THIS */
 73	wait_queue_head_t wq;
 74	refcount_t refs;
 75	unsigned state;
 76
 77	struct io_failure_record *failrec;
 78
 79#ifdef CONFIG_BTRFS_DEBUG
 80	struct list_head leak_list;
 81#endif
 82};
 83
 84int __init extent_state_cache_init(void);
 85void __cold extent_state_cache_exit(void);
 86
 87void extent_io_tree_init(struct btrfs_fs_info *fs_info,
 88			 struct extent_io_tree *tree, unsigned int owner,
 89			 void *private_data);
 90void extent_io_tree_release(struct extent_io_tree *tree);
 91
 92int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
 93		     struct extent_state **cached);
 94
 95static inline int lock_extent(struct extent_io_tree *tree, u64 start, u64 end)
 96{
 97	return lock_extent_bits(tree, start, end, NULL);
 98}
 99
100int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end);
101
102int __init extent_io_init(void);
103void __cold extent_io_exit(void);
104
105u64 count_range_bits(struct extent_io_tree *tree,
106		     u64 *start, u64 search_end,
107		     u64 max_bytes, unsigned bits, int contig);
 
108
109void free_extent_state(struct extent_state *state);
110int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end,
111		   unsigned bits, int filled,
112		   struct extent_state *cached_state);
113int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
114		unsigned bits, struct extent_changeset *changeset);
115int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
116		     unsigned bits, int wake, int delete,
117		     struct extent_state **cached);
118int __clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
119		     unsigned bits, int wake, int delete,
120		     struct extent_state **cached, gfp_t mask,
121		     struct extent_changeset *changeset);
122
123static inline int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end)
 
 
124{
125	return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, NULL);
 
126}
127
128static inline int unlock_extent_cached(struct extent_io_tree *tree, u64 start,
129		u64 end, struct extent_state **cached)
130{
131	return __clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached,
132				GFP_NOFS, NULL);
133}
134
135static inline int unlock_extent_cached_atomic(struct extent_io_tree *tree,
136		u64 start, u64 end, struct extent_state **cached)
137{
138	return __clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached,
139				GFP_ATOMIC, NULL);
140}
141
142static inline int clear_extent_bits(struct extent_io_tree *tree, u64 start,
143		u64 end, unsigned bits)
144{
145	int wake = 0;
146
147	if (bits & EXTENT_LOCKED)
148		wake = 1;
149
150	return clear_extent_bit(tree, start, end, bits, wake, 0, NULL);
151}
152
153int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end,
154			   unsigned bits, struct extent_changeset *changeset);
155int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
156		   unsigned bits, u64 *failed_start,
157		   struct extent_state **cached_state, gfp_t mask);
158int set_extent_bits_nowait(struct extent_io_tree *tree, u64 start, u64 end,
159			   unsigned bits);
 
 
 
160
161static inline int set_extent_bits(struct extent_io_tree *tree, u64 start,
162		u64 end, unsigned bits)
163{
164	return set_extent_bit(tree, start, end, bits, NULL, NULL, GFP_NOFS);
165}
166
167static inline int clear_extent_uptodate(struct extent_io_tree *tree, u64 start,
168		u64 end, struct extent_state **cached_state)
169{
170	return __clear_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, 0,
171				cached_state, GFP_NOFS, NULL);
172}
173
174static inline int set_extent_dirty(struct extent_io_tree *tree, u64 start,
175		u64 end, gfp_t mask)
176{
177	return set_extent_bit(tree, start, end, EXTENT_DIRTY, NULL,
178			      NULL, mask);
179}
180
181static inline int clear_extent_dirty(struct extent_io_tree *tree, u64 start,
182				     u64 end, struct extent_state **cached)
183{
184	return clear_extent_bit(tree, start, end,
185				EXTENT_DIRTY | EXTENT_DELALLOC |
186				EXTENT_DO_ACCOUNTING, 0, 0, cached);
187}
188
189int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
190		       unsigned bits, unsigned clear_bits,
191		       struct extent_state **cached_state);
192
193static inline int set_extent_delalloc(struct extent_io_tree *tree, u64 start,
194				      u64 end, unsigned int extra_bits,
195				      struct extent_state **cached_state)
196{
197	return set_extent_bit(tree, start, end,
198			      EXTENT_DELALLOC | EXTENT_UPTODATE | extra_bits,
199			      NULL, cached_state, GFP_NOFS);
200}
201
202static inline int set_extent_defrag(struct extent_io_tree *tree, u64 start,
203		u64 end, struct extent_state **cached_state)
204{
205	return set_extent_bit(tree, start, end,
206			      EXTENT_DELALLOC | EXTENT_UPTODATE | EXTENT_DEFRAG,
207			      NULL, cached_state, GFP_NOFS);
208}
209
210static inline int set_extent_new(struct extent_io_tree *tree, u64 start,
211		u64 end)
212{
213	return set_extent_bit(tree, start, end, EXTENT_NEW, NULL, NULL,
214			GFP_NOFS);
215}
216
217static inline int set_extent_uptodate(struct extent_io_tree *tree, u64 start,
218		u64 end, struct extent_state **cached_state, gfp_t mask)
219{
220	return set_extent_bit(tree, start, end, EXTENT_UPTODATE, NULL,
221			      cached_state, mask);
222}
223
224int find_first_extent_bit(struct extent_io_tree *tree, u64 start,
225			  u64 *start_ret, u64 *end_ret, unsigned bits,
226			  struct extent_state **cached_state);
227void find_first_clear_extent_bit(struct extent_io_tree *tree, u64 start,
228				 u64 *start_ret, u64 *end_ret, unsigned bits);
229int find_contiguous_extent_bit(struct extent_io_tree *tree, u64 start,
230			       u64 *start_ret, u64 *end_ret, unsigned bits);
231int extent_invalidatepage(struct extent_io_tree *tree,
232			  struct page *page, unsigned long offset);
233bool btrfs_find_delalloc_range(struct extent_io_tree *tree, u64 *start,
234			       u64 *end, u64 max_bytes,
235			       struct extent_state **cached_state);
236
237/* This should be reworked in the future and put elsewhere. */
238struct io_failure_record *get_state_failrec(struct extent_io_tree *tree, u64 start);
239int set_state_failrec(struct extent_io_tree *tree, u64 start,
240		      struct io_failure_record *failrec);
241void btrfs_free_io_failure_record(struct btrfs_inode *inode, u64 start,
242		u64 end);
243int free_io_failure(struct extent_io_tree *failure_tree,
244		    struct extent_io_tree *io_tree,
245		    struct io_failure_record *rec);
246int clean_io_failure(struct btrfs_fs_info *fs_info,
247		     struct extent_io_tree *failure_tree,
248		     struct extent_io_tree *io_tree, u64 start,
249		     struct page *page, u64 ino, unsigned int pg_offset);
250
251#endif /* BTRFS_EXTENT_IO_TREE_H */