Loading...
1/* SPDX-License-Identifier: GPL-2.0 */
2/*
3 * Copyright (C) 2017 Western Digital Corporation or its affiliates.
4 *
5 * This file is released under the GPL.
6 */
7
8#ifndef DM_ZONED_H
9#define DM_ZONED_H
10
11#include <linux/types.h>
12#include <linux/blkdev.h>
13#include <linux/device-mapper.h>
14#include <linux/dm-kcopyd.h>
15#include <linux/list.h>
16#include <linux/spinlock.h>
17#include <linux/mutex.h>
18#include <linux/workqueue.h>
19#include <linux/rwsem.h>
20#include <linux/rbtree.h>
21#include <linux/radix-tree.h>
22#include <linux/shrinker.h>
23
24/*
25 * dm-zoned creates block devices with 4KB blocks, always.
26 */
27#define DMZ_BLOCK_SHIFT 12
28#define DMZ_BLOCK_SIZE (1 << DMZ_BLOCK_SHIFT)
29#define DMZ_BLOCK_MASK (DMZ_BLOCK_SIZE - 1)
30
31#define DMZ_BLOCK_SHIFT_BITS (DMZ_BLOCK_SHIFT + 3)
32#define DMZ_BLOCK_SIZE_BITS (1 << DMZ_BLOCK_SHIFT_BITS)
33#define DMZ_BLOCK_MASK_BITS (DMZ_BLOCK_SIZE_BITS - 1)
34
35#define DMZ_BLOCK_SECTORS_SHIFT (DMZ_BLOCK_SHIFT - SECTOR_SHIFT)
36#define DMZ_BLOCK_SECTORS (DMZ_BLOCK_SIZE >> SECTOR_SHIFT)
37#define DMZ_BLOCK_SECTORS_MASK (DMZ_BLOCK_SECTORS - 1)
38
39/*
40 * 4KB block <-> 512B sector conversion.
41 */
42#define dmz_blk2sect(b) ((sector_t)(b) << DMZ_BLOCK_SECTORS_SHIFT)
43#define dmz_sect2blk(s) ((sector_t)(s) >> DMZ_BLOCK_SECTORS_SHIFT)
44
45#define dmz_bio_block(bio) dmz_sect2blk((bio)->bi_iter.bi_sector)
46#define dmz_bio_blocks(bio) dmz_sect2blk(bio_sectors(bio))
47
48struct dmz_metadata;
49struct dmz_reclaim;
50
51/*
52 * Zoned block device information.
53 */
54struct dmz_dev {
55 struct block_device *bdev;
56 struct dmz_metadata *metadata;
57 struct dmz_reclaim *reclaim;
58
59 char name[BDEVNAME_SIZE];
60 uuid_t uuid;
61
62 sector_t capacity;
63
64 unsigned int dev_idx;
65
66 unsigned int nr_zones;
67 unsigned int zone_offset;
68
69 unsigned int flags;
70
71 sector_t zone_nr_sectors;
72
73 unsigned int nr_rnd;
74 atomic_t unmap_nr_rnd;
75 struct list_head unmap_rnd_list;
76 struct list_head map_rnd_list;
77
78 unsigned int nr_seq;
79 atomic_t unmap_nr_seq;
80 struct list_head unmap_seq_list;
81 struct list_head map_seq_list;
82};
83
84#define dmz_bio_chunk(zmd, bio) ((bio)->bi_iter.bi_sector >> \
85 dmz_zone_nr_sectors_shift(zmd))
86#define dmz_chunk_block(zmd, b) ((b) & (dmz_zone_nr_blocks(zmd) - 1))
87
88/* Device flags. */
89#define DMZ_BDEV_DYING (1 << 0)
90#define DMZ_CHECK_BDEV (2 << 0)
91#define DMZ_BDEV_REGULAR (4 << 0)
92
93/*
94 * Zone descriptor.
95 */
96struct dm_zone {
97 /* For listing the zone depending on its state */
98 struct list_head link;
99
100 /* Device containing this zone */
101 struct dmz_dev *dev;
102
103 /* Zone type and state */
104 unsigned long flags;
105
106 /* Zone activation reference count */
107 atomic_t refcount;
108
109 /* Zone id */
110 unsigned int id;
111
112 /* Zone write pointer block (relative to the zone start block) */
113 unsigned int wp_block;
114
115 /* Zone weight (number of valid blocks in the zone) */
116 unsigned int weight;
117
118 /* The chunk that the zone maps */
119 unsigned int chunk;
120
121 /*
122 * For a sequential data zone, pointer to the random zone
123 * used as a buffer for processing unaligned writes.
124 * For a buffer zone, this points back to the data zone.
125 */
126 struct dm_zone *bzone;
127};
128
129/*
130 * Zone flags.
131 */
132enum {
133 /* Zone write type */
134 DMZ_CACHE,
135 DMZ_RND,
136 DMZ_SEQ,
137
138 /* Zone critical condition */
139 DMZ_OFFLINE,
140 DMZ_READ_ONLY,
141
142 /* How the zone is being used */
143 DMZ_META,
144 DMZ_DATA,
145 DMZ_BUF,
146 DMZ_RESERVED,
147
148 /* Zone internal state */
149 DMZ_RECLAIM,
150 DMZ_SEQ_WRITE_ERR,
151 DMZ_RECLAIM_TERMINATE,
152};
153
154/*
155 * Zone data accessors.
156 */
157#define dmz_is_cache(z) test_bit(DMZ_CACHE, &(z)->flags)
158#define dmz_is_rnd(z) test_bit(DMZ_RND, &(z)->flags)
159#define dmz_is_seq(z) test_bit(DMZ_SEQ, &(z)->flags)
160#define dmz_is_empty(z) ((z)->wp_block == 0)
161#define dmz_is_offline(z) test_bit(DMZ_OFFLINE, &(z)->flags)
162#define dmz_is_readonly(z) test_bit(DMZ_READ_ONLY, &(z)->flags)
163#define dmz_in_reclaim(z) test_bit(DMZ_RECLAIM, &(z)->flags)
164#define dmz_is_reserved(z) test_bit(DMZ_RESERVED, &(z)->flags)
165#define dmz_seq_write_err(z) test_bit(DMZ_SEQ_WRITE_ERR, &(z)->flags)
166#define dmz_reclaim_should_terminate(z) \
167 test_bit(DMZ_RECLAIM_TERMINATE, &(z)->flags)
168
169#define dmz_is_meta(z) test_bit(DMZ_META, &(z)->flags)
170#define dmz_is_buf(z) test_bit(DMZ_BUF, &(z)->flags)
171#define dmz_is_data(z) test_bit(DMZ_DATA, &(z)->flags)
172
173#define dmz_weight(z) ((z)->weight)
174
175/*
176 * Message functions.
177 */
178#define dmz_dev_info(dev, format, args...) \
179 DMINFO("(%s): " format, (dev)->name, ## args)
180
181#define dmz_dev_err(dev, format, args...) \
182 DMERR("(%s): " format, (dev)->name, ## args)
183
184#define dmz_dev_warn(dev, format, args...) \
185 DMWARN("(%s): " format, (dev)->name, ## args)
186
187#define dmz_dev_debug(dev, format, args...) \
188 DMDEBUG("(%s): " format, (dev)->name, ## args)
189
190/*
191 * Functions defined in dm-zoned-metadata.c
192 */
193int dmz_ctr_metadata(struct dmz_dev *dev, int num_dev,
194 struct dmz_metadata **zmd, const char *devname);
195void dmz_dtr_metadata(struct dmz_metadata *zmd);
196int dmz_resume_metadata(struct dmz_metadata *zmd);
197
198void dmz_lock_map(struct dmz_metadata *zmd);
199void dmz_unlock_map(struct dmz_metadata *zmd);
200void dmz_lock_metadata(struct dmz_metadata *zmd);
201void dmz_unlock_metadata(struct dmz_metadata *zmd);
202void dmz_lock_flush(struct dmz_metadata *zmd);
203void dmz_unlock_flush(struct dmz_metadata *zmd);
204int dmz_flush_metadata(struct dmz_metadata *zmd);
205const char *dmz_metadata_label(struct dmz_metadata *zmd);
206
207sector_t dmz_start_sect(struct dmz_metadata *zmd, struct dm_zone *zone);
208sector_t dmz_start_block(struct dmz_metadata *zmd, struct dm_zone *zone);
209unsigned int dmz_nr_chunks(struct dmz_metadata *zmd);
210
211bool dmz_check_dev(struct dmz_metadata *zmd);
212bool dmz_dev_is_dying(struct dmz_metadata *zmd);
213
214#define DMZ_ALLOC_RND 0x01
215#define DMZ_ALLOC_CACHE 0x02
216#define DMZ_ALLOC_SEQ 0x04
217#define DMZ_ALLOC_RECLAIM 0x10
218
219struct dm_zone *dmz_alloc_zone(struct dmz_metadata *zmd,
220 unsigned int dev_idx, unsigned long flags);
221void dmz_free_zone(struct dmz_metadata *zmd, struct dm_zone *zone);
222
223void dmz_map_zone(struct dmz_metadata *zmd, struct dm_zone *zone,
224 unsigned int chunk);
225void dmz_unmap_zone(struct dmz_metadata *zmd, struct dm_zone *zone);
226unsigned int dmz_nr_zones(struct dmz_metadata *zmd);
227unsigned int dmz_nr_cache_zones(struct dmz_metadata *zmd);
228unsigned int dmz_nr_unmap_cache_zones(struct dmz_metadata *zmd);
229unsigned int dmz_nr_rnd_zones(struct dmz_metadata *zmd, int idx);
230unsigned int dmz_nr_unmap_rnd_zones(struct dmz_metadata *zmd, int idx);
231unsigned int dmz_nr_seq_zones(struct dmz_metadata *zmd, int idx);
232unsigned int dmz_nr_unmap_seq_zones(struct dmz_metadata *zmd, int idx);
233unsigned int dmz_zone_nr_blocks(struct dmz_metadata *zmd);
234unsigned int dmz_zone_nr_blocks_shift(struct dmz_metadata *zmd);
235unsigned int dmz_zone_nr_sectors(struct dmz_metadata *zmd);
236unsigned int dmz_zone_nr_sectors_shift(struct dmz_metadata *zmd);
237
238/*
239 * Activate a zone (increment its reference count).
240 */
241static inline void dmz_activate_zone(struct dm_zone *zone)
242{
243 atomic_inc(&zone->refcount);
244}
245
246int dmz_lock_zone_reclaim(struct dm_zone *zone);
247void dmz_unlock_zone_reclaim(struct dm_zone *zone);
248struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd,
249 unsigned int dev_idx, bool idle);
250
251struct dm_zone *dmz_get_chunk_mapping(struct dmz_metadata *zmd,
252 unsigned int chunk, int op);
253void dmz_put_chunk_mapping(struct dmz_metadata *zmd, struct dm_zone *zone);
254struct dm_zone *dmz_get_chunk_buffer(struct dmz_metadata *zmd,
255 struct dm_zone *dzone);
256
257int dmz_validate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone,
258 sector_t chunk_block, unsigned int nr_blocks);
259int dmz_invalidate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone,
260 sector_t chunk_block, unsigned int nr_blocks);
261int dmz_block_valid(struct dmz_metadata *zmd, struct dm_zone *zone,
262 sector_t chunk_block);
263int dmz_first_valid_block(struct dmz_metadata *zmd, struct dm_zone *zone,
264 sector_t *chunk_block);
265int dmz_copy_valid_blocks(struct dmz_metadata *zmd, struct dm_zone *from_zone,
266 struct dm_zone *to_zone);
267int dmz_merge_valid_blocks(struct dmz_metadata *zmd, struct dm_zone *from_zone,
268 struct dm_zone *to_zone, sector_t chunk_block);
269
270/*
271 * Functions defined in dm-zoned-reclaim.c
272 */
273int dmz_ctr_reclaim(struct dmz_metadata *zmd, struct dmz_reclaim **zrc, int idx);
274void dmz_dtr_reclaim(struct dmz_reclaim *zrc);
275void dmz_suspend_reclaim(struct dmz_reclaim *zrc);
276void dmz_resume_reclaim(struct dmz_reclaim *zrc);
277void dmz_reclaim_bio_acc(struct dmz_reclaim *zrc);
278void dmz_schedule_reclaim(struct dmz_reclaim *zrc);
279
280/*
281 * Functions defined in dm-zoned-target.c
282 */
283bool dmz_bdev_is_dying(struct dmz_dev *dmz_dev);
284bool dmz_check_bdev(struct dmz_dev *dmz_dev);
285
286/*
287 * Deactivate a zone. This decrement the zone reference counter
288 * indicating that all BIOs to the zone have completed when the count is 0.
289 */
290static inline void dmz_deactivate_zone(struct dm_zone *zone)
291{
292 dmz_reclaim_bio_acc(zone->dev->reclaim);
293 atomic_dec(&zone->refcount);
294}
295
296/*
297 * Test if a zone is active, that is, has a refcount > 0.
298 */
299static inline bool dmz_is_active(struct dm_zone *zone)
300{
301 return atomic_read(&zone->refcount);
302}
303
304#endif /* DM_ZONED_H */
1/* SPDX-License-Identifier: GPL-2.0 */
2/*
3 * Copyright (C) 2017 Western Digital Corporation or its affiliates.
4 *
5 * This file is released under the GPL.
6 */
7
8#ifndef DM_ZONED_H
9#define DM_ZONED_H
10
11#include <linux/types.h>
12#include <linux/blkdev.h>
13#include <linux/device-mapper.h>
14#include <linux/dm-kcopyd.h>
15#include <linux/list.h>
16#include <linux/spinlock.h>
17#include <linux/mutex.h>
18#include <linux/workqueue.h>
19#include <linux/rwsem.h>
20#include <linux/rbtree.h>
21#include <linux/radix-tree.h>
22#include <linux/shrinker.h>
23
24/*
25 * dm-zoned creates block devices with 4KB blocks, always.
26 */
27#define DMZ_BLOCK_SHIFT 12
28#define DMZ_BLOCK_SIZE (1 << DMZ_BLOCK_SHIFT)
29#define DMZ_BLOCK_MASK (DMZ_BLOCK_SIZE - 1)
30
31#define DMZ_BLOCK_SHIFT_BITS (DMZ_BLOCK_SHIFT + 3)
32#define DMZ_BLOCK_SIZE_BITS (1 << DMZ_BLOCK_SHIFT_BITS)
33#define DMZ_BLOCK_MASK_BITS (DMZ_BLOCK_SIZE_BITS - 1)
34
35#define DMZ_BLOCK_SECTORS_SHIFT (DMZ_BLOCK_SHIFT - SECTOR_SHIFT)
36#define DMZ_BLOCK_SECTORS (DMZ_BLOCK_SIZE >> SECTOR_SHIFT)
37#define DMZ_BLOCK_SECTORS_MASK (DMZ_BLOCK_SECTORS - 1)
38
39/*
40 * 4KB block <-> 512B sector conversion.
41 */
42#define dmz_blk2sect(b) ((sector_t)(b) << DMZ_BLOCK_SECTORS_SHIFT)
43#define dmz_sect2blk(s) ((sector_t)(s) >> DMZ_BLOCK_SECTORS_SHIFT)
44
45#define dmz_bio_block(bio) dmz_sect2blk((bio)->bi_iter.bi_sector)
46#define dmz_bio_blocks(bio) dmz_sect2blk(bio_sectors(bio))
47
48/*
49 * Zoned block device information.
50 */
51struct dmz_dev {
52 struct block_device *bdev;
53
54 char name[BDEVNAME_SIZE];
55
56 sector_t capacity;
57
58 unsigned int nr_zones;
59
60 unsigned int flags;
61
62 sector_t zone_nr_sectors;
63 unsigned int zone_nr_sectors_shift;
64
65 sector_t zone_nr_blocks;
66 sector_t zone_nr_blocks_shift;
67};
68
69#define dmz_bio_chunk(dev, bio) ((bio)->bi_iter.bi_sector >> \
70 (dev)->zone_nr_sectors_shift)
71#define dmz_chunk_block(dev, b) ((b) & ((dev)->zone_nr_blocks - 1))
72
73/* Device flags. */
74#define DMZ_BDEV_DYING (1 << 0)
75
76/*
77 * Zone descriptor.
78 */
79struct dm_zone {
80 /* For listing the zone depending on its state */
81 struct list_head link;
82
83 /* Zone type and state */
84 unsigned long flags;
85
86 /* Zone activation reference count */
87 atomic_t refcount;
88
89 /* Zone write pointer block (relative to the zone start block) */
90 unsigned int wp_block;
91
92 /* Zone weight (number of valid blocks in the zone) */
93 unsigned int weight;
94
95 /* The chunk that the zone maps */
96 unsigned int chunk;
97
98 /*
99 * For a sequential data zone, pointer to the random zone
100 * used as a buffer for processing unaligned writes.
101 * For a buffer zone, this points back to the data zone.
102 */
103 struct dm_zone *bzone;
104};
105
106/*
107 * Zone flags.
108 */
109enum {
110 /* Zone write type */
111 DMZ_RND,
112 DMZ_SEQ,
113
114 /* Zone critical condition */
115 DMZ_OFFLINE,
116 DMZ_READ_ONLY,
117
118 /* How the zone is being used */
119 DMZ_META,
120 DMZ_DATA,
121 DMZ_BUF,
122
123 /* Zone internal state */
124 DMZ_RECLAIM,
125 DMZ_SEQ_WRITE_ERR,
126};
127
128/*
129 * Zone data accessors.
130 */
131#define dmz_is_rnd(z) test_bit(DMZ_RND, &(z)->flags)
132#define dmz_is_seq(z) test_bit(DMZ_SEQ, &(z)->flags)
133#define dmz_is_empty(z) ((z)->wp_block == 0)
134#define dmz_is_offline(z) test_bit(DMZ_OFFLINE, &(z)->flags)
135#define dmz_is_readonly(z) test_bit(DMZ_READ_ONLY, &(z)->flags)
136#define dmz_in_reclaim(z) test_bit(DMZ_RECLAIM, &(z)->flags)
137#define dmz_seq_write_err(z) test_bit(DMZ_SEQ_WRITE_ERR, &(z)->flags)
138
139#define dmz_is_meta(z) test_bit(DMZ_META, &(z)->flags)
140#define dmz_is_buf(z) test_bit(DMZ_BUF, &(z)->flags)
141#define dmz_is_data(z) test_bit(DMZ_DATA, &(z)->flags)
142
143#define dmz_weight(z) ((z)->weight)
144
145/*
146 * Message functions.
147 */
148#define dmz_dev_info(dev, format, args...) \
149 DMINFO("(%s): " format, (dev)->name, ## args)
150
151#define dmz_dev_err(dev, format, args...) \
152 DMERR("(%s): " format, (dev)->name, ## args)
153
154#define dmz_dev_warn(dev, format, args...) \
155 DMWARN("(%s): " format, (dev)->name, ## args)
156
157#define dmz_dev_debug(dev, format, args...) \
158 DMDEBUG("(%s): " format, (dev)->name, ## args)
159
160struct dmz_metadata;
161struct dmz_reclaim;
162
163/*
164 * Functions defined in dm-zoned-metadata.c
165 */
166int dmz_ctr_metadata(struct dmz_dev *dev, struct dmz_metadata **zmd);
167void dmz_dtr_metadata(struct dmz_metadata *zmd);
168int dmz_resume_metadata(struct dmz_metadata *zmd);
169
170void dmz_lock_map(struct dmz_metadata *zmd);
171void dmz_unlock_map(struct dmz_metadata *zmd);
172void dmz_lock_metadata(struct dmz_metadata *zmd);
173void dmz_unlock_metadata(struct dmz_metadata *zmd);
174void dmz_lock_flush(struct dmz_metadata *zmd);
175void dmz_unlock_flush(struct dmz_metadata *zmd);
176int dmz_flush_metadata(struct dmz_metadata *zmd);
177
178unsigned int dmz_id(struct dmz_metadata *zmd, struct dm_zone *zone);
179sector_t dmz_start_sect(struct dmz_metadata *zmd, struct dm_zone *zone);
180sector_t dmz_start_block(struct dmz_metadata *zmd, struct dm_zone *zone);
181unsigned int dmz_nr_chunks(struct dmz_metadata *zmd);
182
183#define DMZ_ALLOC_RND 0x01
184#define DMZ_ALLOC_RECLAIM 0x02
185
186struct dm_zone *dmz_alloc_zone(struct dmz_metadata *zmd, unsigned long flags);
187void dmz_free_zone(struct dmz_metadata *zmd, struct dm_zone *zone);
188
189void dmz_map_zone(struct dmz_metadata *zmd, struct dm_zone *zone,
190 unsigned int chunk);
191void dmz_unmap_zone(struct dmz_metadata *zmd, struct dm_zone *zone);
192unsigned int dmz_nr_rnd_zones(struct dmz_metadata *zmd);
193unsigned int dmz_nr_unmap_rnd_zones(struct dmz_metadata *zmd);
194
195/*
196 * Activate a zone (increment its reference count).
197 */
198static inline void dmz_activate_zone(struct dm_zone *zone)
199{
200 atomic_inc(&zone->refcount);
201}
202
203/*
204 * Deactivate a zone. This decrement the zone reference counter
205 * indicating that all BIOs to the zone have completed when the count is 0.
206 */
207static inline void dmz_deactivate_zone(struct dm_zone *zone)
208{
209 atomic_dec(&zone->refcount);
210}
211
212/*
213 * Test if a zone is active, that is, has a refcount > 0.
214 */
215static inline bool dmz_is_active(struct dm_zone *zone)
216{
217 return atomic_read(&zone->refcount);
218}
219
220int dmz_lock_zone_reclaim(struct dm_zone *zone);
221void dmz_unlock_zone_reclaim(struct dm_zone *zone);
222struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd);
223
224struct dm_zone *dmz_get_chunk_mapping(struct dmz_metadata *zmd,
225 unsigned int chunk, int op);
226void dmz_put_chunk_mapping(struct dmz_metadata *zmd, struct dm_zone *zone);
227struct dm_zone *dmz_get_chunk_buffer(struct dmz_metadata *zmd,
228 struct dm_zone *dzone);
229
230int dmz_validate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone,
231 sector_t chunk_block, unsigned int nr_blocks);
232int dmz_invalidate_blocks(struct dmz_metadata *zmd, struct dm_zone *zone,
233 sector_t chunk_block, unsigned int nr_blocks);
234int dmz_block_valid(struct dmz_metadata *zmd, struct dm_zone *zone,
235 sector_t chunk_block);
236int dmz_first_valid_block(struct dmz_metadata *zmd, struct dm_zone *zone,
237 sector_t *chunk_block);
238int dmz_copy_valid_blocks(struct dmz_metadata *zmd, struct dm_zone *from_zone,
239 struct dm_zone *to_zone);
240int dmz_merge_valid_blocks(struct dmz_metadata *zmd, struct dm_zone *from_zone,
241 struct dm_zone *to_zone, sector_t chunk_block);
242
243/*
244 * Functions defined in dm-zoned-reclaim.c
245 */
246int dmz_ctr_reclaim(struct dmz_dev *dev, struct dmz_metadata *zmd,
247 struct dmz_reclaim **zrc);
248void dmz_dtr_reclaim(struct dmz_reclaim *zrc);
249void dmz_suspend_reclaim(struct dmz_reclaim *zrc);
250void dmz_resume_reclaim(struct dmz_reclaim *zrc);
251void dmz_reclaim_bio_acc(struct dmz_reclaim *zrc);
252void dmz_schedule_reclaim(struct dmz_reclaim *zrc);
253
254/*
255 * Functions defined in dm-zoned-target.c
256 */
257bool dmz_bdev_is_dying(struct dmz_dev *dmz_dev);
258
259#endif /* DM_ZONED_H */