Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 | // SPDX-License-Identifier: GPL-2.0 /* Maximum size of each resync request */ #define RESYNC_BLOCK_SIZE (64*1024) #define RESYNC_PAGES ((RESYNC_BLOCK_SIZE + PAGE_SIZE-1) / PAGE_SIZE) /* * Number of guaranteed raid bios in case of extreme VM load: */ #define NR_RAID_BIOS 256 /* when we get a read error on a read-only array, we redirect to another * device without failing the first device, or trying to over-write to * correct the read error. To keep track of bad blocks on a per-bio * level, we store IO_BLOCKED in the appropriate 'bios' pointer */ #define IO_BLOCKED ((struct bio *)1) /* When we successfully write to a known bad-block, we need to remove the * bad-block marking which must be done from process context. So we record * the success by setting devs[n].bio to IO_MADE_GOOD */ #define IO_MADE_GOOD ((struct bio *)2) #define BIO_SPECIAL(bio) ((unsigned long)bio <= 2) #define MAX_PLUG_BIO 32 /* for managing resync I/O pages */ struct resync_pages { void *raid_bio; struct page *pages[RESYNC_PAGES]; }; struct raid1_plug_cb { struct blk_plug_cb cb; struct bio_list pending; unsigned int count; }; static void rbio_pool_free(void *rbio, void *data) { kfree(rbio); } static inline int resync_alloc_pages(struct resync_pages *rp, gfp_t gfp_flags) { int i; for (i = 0; i < RESYNC_PAGES; i++) { rp->pages[i] = alloc_page(gfp_flags); if (!rp->pages[i]) goto out_free; } return 0; out_free: while (--i >= 0) put_page(rp->pages[i]); return -ENOMEM; } static inline void resync_free_pages(struct resync_pages *rp) { int i; for (i = 0; i < RESYNC_PAGES; i++) put_page(rp->pages[i]); } static inline void resync_get_all_pages(struct resync_pages *rp) { int i; for (i = 0; i < RESYNC_PAGES; i++) get_page(rp->pages[i]); } static inline struct page *resync_fetch_page(struct resync_pages *rp, unsigned idx) { if (WARN_ON_ONCE(idx >= RESYNC_PAGES)) return NULL; return rp->pages[idx]; } /* * 'strct resync_pages' stores actual pages used for doing the resync * IO, and it is per-bio, so make .bi_private points to it. */ static inline struct resync_pages *get_resync_pages(struct bio *bio) { return bio->bi_private; } /* generally called after bio_reset() for reseting bvec */ static void md_bio_reset_resync_pages(struct bio *bio, struct resync_pages *rp, int size) { int idx = 0; /* initialize bvec table again */ do { struct page *page = resync_fetch_page(rp, idx); int len = min_t(int, size, PAGE_SIZE); if (WARN_ON(!bio_add_page(bio, page, len, 0))) { bio->bi_status = BLK_STS_RESOURCE; bio_endio(bio); return; } size -= len; } while (idx++ < RESYNC_PAGES && size > 0); } static inline void raid1_submit_write(struct bio *bio) { struct md_rdev *rdev = (void *)bio->bi_bdev; bio->bi_next = NULL; bio_set_dev(bio, rdev->bdev); if (test_bit(Faulty, &rdev->flags)) bio_io_error(bio); else if (unlikely(bio_op(bio) == REQ_OP_DISCARD && !bdev_max_discard_sectors(bio->bi_bdev))) /* Just ignore it */ bio_endio(bio); else submit_bio_noacct(bio); } static inline bool raid1_add_bio_to_plug(struct mddev *mddev, struct bio *bio, blk_plug_cb_fn unplug, int copies) { struct raid1_plug_cb *plug = NULL; struct blk_plug_cb *cb; /* * If bitmap is not enabled, it's safe to submit the io directly, and * this can get optimal performance. */ if (!md_bitmap_enabled(mddev->bitmap)) { raid1_submit_write(bio); return true; } cb = blk_check_plugged(unplug, mddev, sizeof(*plug)); if (!cb) return false; plug = container_of(cb, struct raid1_plug_cb, cb); bio_list_add(&plug->pending, bio); if (++plug->count / MAX_PLUG_BIO >= copies) { list_del(&cb->list); cb->callback(cb, false); } return true; } /* * current->bio_list will be set under submit_bio() context, in this case bitmap * io will be added to the list and wait for current io submission to finish, * while current io submission must wait for bitmap io to be done. In order to * avoid such deadlock, submit bitmap io asynchronously. */ static inline void raid1_prepare_flush_writes(struct bitmap *bitmap) { if (current->bio_list) md_bitmap_unplug_async(bitmap); else md_bitmap_unplug(bitmap); } /* * Used by fix_read_error() to decay the per rdev read_errors. * We halve the read error count for every hour that has elapsed * since the last recorded read error. */ static inline void check_decay_read_errors(struct mddev *mddev, struct md_rdev *rdev) { long cur_time_mon; unsigned long hours_since_last; unsigned int read_errors = atomic_read(&rdev->read_errors); cur_time_mon = ktime_get_seconds(); if (rdev->last_read_error == 0) { /* first time we've seen a read error */ rdev->last_read_error = cur_time_mon; return; } hours_since_last = (long)(cur_time_mon - rdev->last_read_error) / 3600; rdev->last_read_error = cur_time_mon; /* * if hours_since_last is > the number of bits in read_errors * just set read errors to 0. We do this to avoid * overflowing the shift of read_errors by hours_since_last. */ if (hours_since_last >= 8 * sizeof(read_errors)) atomic_set(&rdev->read_errors, 0); else atomic_set(&rdev->read_errors, read_errors >> hours_since_last); } static inline bool exceed_read_errors(struct mddev *mddev, struct md_rdev *rdev) { int max_read_errors = atomic_read(&mddev->max_corr_read_errors); int read_errors; check_decay_read_errors(mddev, rdev); read_errors = atomic_inc_return(&rdev->read_errors); if (read_errors > max_read_errors) { pr_notice("md/"RAID_1_10_NAME":%s: %pg: Raid device exceeded read_error threshold [cur %d:max %d]\n", mdname(mddev), rdev->bdev, read_errors, max_read_errors); pr_notice("md/"RAID_1_10_NAME":%s: %pg: Failing raid device\n", mdname(mddev), rdev->bdev); md_error(mddev, rdev); return true; } return false; } /** * raid1_check_read_range() - check a given read range for bad blocks, * available read length is returned; * @rdev: the rdev to read; * @this_sector: read position; * @len: read length; * * helper function for read_balance() * * 1) If there are no bad blocks in the range, @len is returned; * 2) If the range are all bad blocks, 0 is returned; * 3) If there are partial bad blocks: * - If the bad block range starts after @this_sector, the length of first * good region is returned; * - If the bad block range starts before @this_sector, 0 is returned and * the @len is updated to the offset into the region before we get to the * good blocks; */ static inline int raid1_check_read_range(struct md_rdev *rdev, sector_t this_sector, int *len) { sector_t first_bad; int bad_sectors; /* no bad block overlap */ if (!is_badblock(rdev, this_sector, *len, &first_bad, &bad_sectors)) return *len; /* * bad block range starts offset into our range so we can return the * number of sectors before the bad blocks start. */ if (first_bad > this_sector) return first_bad - this_sector; /* read range is fully consumed by bad blocks. */ if (this_sector + *len <= first_bad + bad_sectors) return 0; /* * final case, bad block range starts before or at the start of our * range but does not cover our entire range so we still return 0 but * update the length with the number of sectors before we get to the * good ones. */ *len = first_bad + bad_sectors - this_sector; return 0; } /* * Check if read should choose the first rdev. * * Balance on the whole device if no resync is going on (recovery is ok) or * below the resync window. Otherwise, take the first readable disk. */ static inline bool raid1_should_read_first(struct mddev *mddev, sector_t this_sector, int len) { if ((mddev->recovery_cp < this_sector + len)) return true; if (mddev_is_clustered(mddev) && md_cluster_ops->area_resyncing(mddev, READ, this_sector, this_sector + len)) return true; return false; } |