root/fs/btrfs/dev-replace.c
// SPDX-License-Identifier: GPL-2.0
/*
 * Copyright (C) STRATO AG 2012.  All rights reserved.
 */

#include <linux/sched.h>
#include <linux/bio.h>
#include <linux/slab.h>
#include <linux/blkdev.h>
#include <linux/kthread.h>
#include <linux/math64.h>
#include "misc.h"
#include "ctree.h"
#include "disk-io.h"
#include "transaction.h"
#include "volumes.h"
#include "async-thread.h"
#include "dev-replace.h"
#include "sysfs.h"
#include "zoned.h"
#include "block-group.h"
#include "fs.h"
#include "accessors.h"
#include "scrub.h"

/*
 * Device replace overview
 *
 * [Objective]
 * To copy all extents (both new and on-disk) from source device to target
 * device, while still keeping the filesystem read-write.
 *
 * [Method]
 * There are two main methods involved:
 *
 * - Write duplication
 *
 *   All new writes will be written to both target and source devices, so even
 *   if replace gets canceled, sources device still contains up-to-date data.
 *
 *   Location:          handle_ops_on_dev_replace() from btrfs_map_block()
 *   Start:             btrfs_dev_replace_start()
 *   End:               btrfs_dev_replace_finishing()
 *   Content:           Latest data/metadata
 *
 * - Copy existing extents
 *
 *   This happens by reusing scrub facility, as scrub also iterates through
 *   existing extents from commit root.
 *
 *   Location:          scrub_write_block_to_dev_replace() from
 *                      scrub_block_complete()
 *   Content:           Data/meta from commit root.
 *
 * Due to the content difference, we need to avoid nocow write when dev-replace
 * is happening.  This is done by marking the block group read-only and waiting
 * for NOCOW writes.
 *
 * After replace is done, the finishing part is done by swapping the target and
 * source devices.
 *
 *   Location:          btrfs_dev_replace_update_device_in_mapping_tree() from
 *                      btrfs_dev_replace_finishing()
 */

static int btrfs_dev_replace_finishing(struct btrfs_fs_info *fs_info,
                                       int scrub_ret);
static int btrfs_dev_replace_kthread(void *data);

int btrfs_init_dev_replace(struct btrfs_fs_info *fs_info)
{
        struct btrfs_dev_lookup_args args = { .devid = BTRFS_DEV_REPLACE_DEVID };
        struct btrfs_key key;
        struct btrfs_root *dev_root = fs_info->dev_root;
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;
        struct extent_buffer *eb;
        int slot;
        int ret = 0;
        BTRFS_PATH_AUTO_FREE(path);
        int item_size;
        struct btrfs_dev_replace_item *ptr;
        u64 src_devid;

        if (!dev_root)
                return 0;

        path = btrfs_alloc_path();
        if (!path)
                return -ENOMEM;

        key.objectid = 0;
        key.type = BTRFS_DEV_REPLACE_KEY;
        key.offset = 0;
        ret = btrfs_search_slot(NULL, dev_root, &key, path, 0, 0);
        if (ret) {
no_valid_dev_replace_entry_found:
                /*
                 * We don't have a replace item or it's corrupted.  If there is
                 * a replace target, fail the mount.
                 */
                if (unlikely(btrfs_find_device(fs_info->fs_devices, &args))) {
                        btrfs_err(fs_info,
                        "found replace target device without a valid replace item");
                        return -EUCLEAN;
                }
                dev_replace->replace_state =
                        BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED;
                dev_replace->cont_reading_from_srcdev_mode =
                    BTRFS_DEV_REPLACE_ITEM_CONT_READING_FROM_SRCDEV_MODE_ALWAYS;
                dev_replace->time_started = 0;
                dev_replace->time_stopped = 0;
                atomic64_set(&dev_replace->num_write_errors, 0);
                atomic64_set(&dev_replace->num_uncorrectable_read_errors, 0);
                dev_replace->cursor_left = 0;
                dev_replace->committed_cursor_left = 0;
                dev_replace->cursor_left_last_write_of_item = 0;
                dev_replace->cursor_right = 0;
                dev_replace->srcdev = NULL;
                dev_replace->tgtdev = NULL;
                dev_replace->is_valid = 0;
                dev_replace->item_needs_writeback = 0;
                return 0;
        }
        slot = path->slots[0];
        eb = path->nodes[0];
        item_size = btrfs_item_size(eb, slot);
        ptr = btrfs_item_ptr(eb, slot, struct btrfs_dev_replace_item);

        if (item_size != sizeof(struct btrfs_dev_replace_item)) {
                btrfs_warn(fs_info,
                        "dev_replace entry found has unexpected size, ignore entry");
                goto no_valid_dev_replace_entry_found;
        }

        src_devid = btrfs_dev_replace_src_devid(eb, ptr);
        dev_replace->cont_reading_from_srcdev_mode =
                btrfs_dev_replace_cont_reading_from_srcdev_mode(eb, ptr);
        dev_replace->replace_state = btrfs_dev_replace_replace_state(eb, ptr);
        dev_replace->time_started = btrfs_dev_replace_time_started(eb, ptr);
        dev_replace->time_stopped =
                btrfs_dev_replace_time_stopped(eb, ptr);
        atomic64_set(&dev_replace->num_write_errors,
                     btrfs_dev_replace_num_write_errors(eb, ptr));
        atomic64_set(&dev_replace->num_uncorrectable_read_errors,
                     btrfs_dev_replace_num_uncorrectable_read_errors(eb, ptr));
        dev_replace->cursor_left = btrfs_dev_replace_cursor_left(eb, ptr);
        dev_replace->committed_cursor_left = dev_replace->cursor_left;
        dev_replace->cursor_left_last_write_of_item = dev_replace->cursor_left;
        dev_replace->cursor_right = btrfs_dev_replace_cursor_right(eb, ptr);
        dev_replace->is_valid = 1;

        dev_replace->item_needs_writeback = 0;
        switch (dev_replace->replace_state) {
        case BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_FINISHED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_CANCELED:
                /*
                 * We don't have an active replace item but if there is a
                 * replace target, fail the mount.
                 */
                if (unlikely(btrfs_find_device(fs_info->fs_devices, &args))) {
                        btrfs_err(fs_info,
"replace without active item, run 'device scan --forget' on the target device");
                        ret = -EUCLEAN;
                } else {
                        dev_replace->srcdev = NULL;
                        dev_replace->tgtdev = NULL;
                }
                break;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED:
                dev_replace->tgtdev = btrfs_find_device(fs_info->fs_devices, &args);
                args.devid = src_devid;
                dev_replace->srcdev = btrfs_find_device(fs_info->fs_devices, &args);

                /*
                 * allow 'btrfs dev replace_cancel' if src/tgt device is
                 * missing
                 */
                if (unlikely(!dev_replace->srcdev && !btrfs_test_opt(fs_info, DEGRADED))) {
                        ret = -EIO;
                        btrfs_warn(fs_info,
                           "cannot mount because device replace operation is ongoing and");
                        btrfs_warn(fs_info,
                           "srcdev (devid %llu) is missing, need to run 'btrfs dev scan'?",
                           src_devid);
                }
                if (unlikely(!dev_replace->tgtdev && !btrfs_test_opt(fs_info, DEGRADED))) {
                        ret = -EIO;
                        btrfs_warn(fs_info,
                           "cannot mount because device replace operation is ongoing and");
                        btrfs_warn(fs_info,
                           "tgtdev (devid %llu) is missing, need to run 'btrfs dev scan'?",
                                BTRFS_DEV_REPLACE_DEVID);
                }
                if (dev_replace->tgtdev) {
                        if (dev_replace->srcdev) {
                                dev_replace->tgtdev->total_bytes =
                                        dev_replace->srcdev->total_bytes;
                                dev_replace->tgtdev->disk_total_bytes =
                                        dev_replace->srcdev->disk_total_bytes;
                                dev_replace->tgtdev->commit_total_bytes =
                                        dev_replace->srcdev->commit_total_bytes;
                                dev_replace->tgtdev->bytes_used =
                                        dev_replace->srcdev->bytes_used;
                                dev_replace->tgtdev->commit_bytes_used =
                                        dev_replace->srcdev->commit_bytes_used;
                        }
                        set_bit(BTRFS_DEV_STATE_REPLACE_TGT,
                                &dev_replace->tgtdev->dev_state);

                        WARN_ON(fs_info->fs_devices->rw_devices == 0);
                        dev_replace->tgtdev->io_width = fs_info->sectorsize;
                        dev_replace->tgtdev->io_align = fs_info->sectorsize;
                        dev_replace->tgtdev->sector_size = fs_info->sectorsize;
                        dev_replace->tgtdev->fs_info = fs_info;
                        set_bit(BTRFS_DEV_STATE_IN_FS_METADATA,
                                &dev_replace->tgtdev->dev_state);
                }
                break;
        }

        return ret;
}

/*
 * Initialize a new device for device replace target from a given source dev
 * and path.
 *
 * Return 0 and new device in @device_out, otherwise return < 0
 */
static int btrfs_init_dev_replace_tgtdev(struct btrfs_fs_info *fs_info,
                                  const char *device_path,
                                  struct btrfs_device *srcdev,
                                  struct btrfs_device **device_out)
{
        struct btrfs_fs_devices *fs_devices = fs_info->fs_devices;
        struct btrfs_device *device;
        struct file *bdev_file;
        struct block_device *bdev;
        u64 devid = BTRFS_DEV_REPLACE_DEVID;
        int ret = 0;

        *device_out = NULL;
        if (srcdev->fs_devices->seeding) {
                btrfs_err(fs_info, "the filesystem is a seed filesystem!");
                return -EINVAL;
        }

        bdev_file = bdev_file_open_by_path(device_path, BLK_OPEN_WRITE,
                                           fs_info->sb, &fs_holder_ops);
        if (IS_ERR(bdev_file)) {
                btrfs_err(fs_info, "target device %s is invalid!", device_path);
                return PTR_ERR(bdev_file);
        }
        bdev = file_bdev(bdev_file);

        if (!btrfs_check_device_zone_type(fs_info, bdev)) {
                btrfs_err(fs_info,
                "dev-replace: zoned type of target device mismatch with filesystem");
                ret = -EINVAL;
                goto error;
        }

        sync_blockdev(bdev);

        list_for_each_entry(device, &fs_devices->devices, dev_list) {
                if (device->bdev == bdev) {
                        btrfs_err(fs_info,
                                  "target device is in the filesystem!");
                        ret = -EEXIST;
                        goto error;
                }
        }


        if (bdev_nr_bytes(bdev) < btrfs_device_get_total_bytes(srcdev)) {
                btrfs_err(fs_info,
                          "target device is smaller than source device!");
                ret = -EINVAL;
                goto error;
        }


        device = btrfs_alloc_device(NULL, &devid, NULL, device_path);
        if (IS_ERR(device)) {
                ret = PTR_ERR(device);
                goto error;
        }

        ret = lookup_bdev(device_path, &device->devt);
        if (ret)
                goto error;

        set_bit(BTRFS_DEV_STATE_WRITEABLE, &device->dev_state);
        device->generation = 0;
        device->io_width = fs_info->sectorsize;
        device->io_align = fs_info->sectorsize;
        device->sector_size = fs_info->sectorsize;
        device->total_bytes = btrfs_device_get_total_bytes(srcdev);
        device->disk_total_bytes = btrfs_device_get_disk_total_bytes(srcdev);
        device->bytes_used = btrfs_device_get_bytes_used(srcdev);
        device->commit_total_bytes = srcdev->commit_total_bytes;
        device->commit_bytes_used = device->bytes_used;
        device->fs_info = fs_info;
        device->bdev = bdev;
        device->bdev_file = bdev_file;
        set_bit(BTRFS_DEV_STATE_IN_FS_METADATA, &device->dev_state);
        set_bit(BTRFS_DEV_STATE_REPLACE_TGT, &device->dev_state);
        device->dev_stats_valid = 1;
        set_blocksize(bdev_file, BTRFS_BDEV_BLOCKSIZE);
        device->fs_devices = fs_devices;

        ret = btrfs_get_dev_zone_info(device, false);
        if (ret)
                goto error;

        mutex_lock(&fs_devices->device_list_mutex);
        list_add(&device->dev_list, &fs_devices->devices);
        fs_devices->num_devices++;
        fs_devices->open_devices++;
        mutex_unlock(&fs_devices->device_list_mutex);

        *device_out = device;
        return 0;

error:
        bdev_fput(bdev_file);
        return ret;
}

/*
 * called from commit_transaction. Writes changed device replace state to
 * disk.
 */
int btrfs_run_dev_replace(struct btrfs_trans_handle *trans)
{
        struct btrfs_fs_info *fs_info = trans->fs_info;
        int ret;
        struct btrfs_root *dev_root = fs_info->dev_root;
        BTRFS_PATH_AUTO_FREE(path);
        struct btrfs_key key;
        struct extent_buffer *eb;
        struct btrfs_dev_replace_item *ptr;
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;

        down_read(&dev_replace->rwsem);
        if (!dev_replace->is_valid ||
            !dev_replace->item_needs_writeback) {
                up_read(&dev_replace->rwsem);
                return 0;
        }
        up_read(&dev_replace->rwsem);

        key.objectid = 0;
        key.type = BTRFS_DEV_REPLACE_KEY;
        key.offset = 0;

        path = btrfs_alloc_path();
        if (!path)
                return -ENOMEM;

        ret = btrfs_search_slot(trans, dev_root, &key, path, -1, 1);
        if (ret < 0) {
                btrfs_warn(fs_info,
                           "error %d while searching for dev_replace item!",
                           ret);
                return ret;
        }

        if (ret == 0 &&
            btrfs_item_size(path->nodes[0], path->slots[0]) < sizeof(*ptr)) {
                /*
                 * need to delete old one and insert a new one.
                 * Since no attempt is made to recover any old state, if the
                 * dev_replace state is 'running', the data on the target
                 * drive is lost.
                 * It would be possible to recover the state: just make sure
                 * that the beginning of the item is never changed and always
                 * contains all the essential information. Then read this
                 * minimal set of information and use it as a base for the
                 * new state.
                 */
                ret = btrfs_del_item(trans, dev_root, path);
                if (ret != 0) {
                        btrfs_warn(fs_info,
                                   "delete too small dev_replace item failed %d!",
                                   ret);
                        return ret;
                }
                ret = 1;
        }

        if (ret == 1) {
                /* need to insert a new item */
                btrfs_release_path(path);
                ret = btrfs_insert_empty_item(trans, dev_root, path,
                                              &key, sizeof(*ptr));
                if (ret < 0) {
                        btrfs_warn(fs_info,
                                   "insert dev_replace item failed %d!", ret);
                        return ret;
                }
        }

        eb = path->nodes[0];
        ptr = btrfs_item_ptr(eb, path->slots[0],
                             struct btrfs_dev_replace_item);

        down_write(&dev_replace->rwsem);
        if (dev_replace->srcdev)
                btrfs_set_dev_replace_src_devid(eb, ptr,
                        dev_replace->srcdev->devid);
        else
                btrfs_set_dev_replace_src_devid(eb, ptr, (u64)-1);
        btrfs_set_dev_replace_cont_reading_from_srcdev_mode(eb, ptr,
                dev_replace->cont_reading_from_srcdev_mode);
        btrfs_set_dev_replace_replace_state(eb, ptr,
                dev_replace->replace_state);
        btrfs_set_dev_replace_time_started(eb, ptr, dev_replace->time_started);
        btrfs_set_dev_replace_time_stopped(eb, ptr, dev_replace->time_stopped);
        btrfs_set_dev_replace_num_write_errors(eb, ptr,
                atomic64_read(&dev_replace->num_write_errors));
        btrfs_set_dev_replace_num_uncorrectable_read_errors(eb, ptr,
                atomic64_read(&dev_replace->num_uncorrectable_read_errors));
        dev_replace->cursor_left_last_write_of_item =
                dev_replace->cursor_left;
        btrfs_set_dev_replace_cursor_left(eb, ptr,
                dev_replace->cursor_left_last_write_of_item);
        btrfs_set_dev_replace_cursor_right(eb, ptr,
                dev_replace->cursor_right);
        dev_replace->item_needs_writeback = 0;
        up_write(&dev_replace->rwsem);

        return ret;
}

static int mark_block_group_to_copy(struct btrfs_fs_info *fs_info,
                                    struct btrfs_device *src_dev)
{
        struct btrfs_path *path;
        struct btrfs_key key;
        struct btrfs_key found_key;
        struct btrfs_root *root = fs_info->dev_root;
        struct btrfs_dev_extent *dev_extent = NULL;
        struct btrfs_block_group *cache;
        struct btrfs_trans_handle *trans;
        int iter_ret = 0;
        int ret = 0;
        u64 chunk_offset;

        /* Do not use "to_copy" on non zoned filesystem for now */
        if (!btrfs_is_zoned(fs_info))
                return 0;

        mutex_lock(&fs_info->chunk_mutex);

        /* Ensure we don't have pending new block group */
        spin_lock(&fs_info->trans_lock);
        while (fs_info->running_transaction &&
               !list_empty(&fs_info->running_transaction->dev_update_list)) {
                spin_unlock(&fs_info->trans_lock);
                mutex_unlock(&fs_info->chunk_mutex);
                trans = btrfs_attach_transaction(root);
                if (IS_ERR(trans)) {
                        ret = PTR_ERR(trans);
                        mutex_lock(&fs_info->chunk_mutex);
                        if (ret == -ENOENT) {
                                spin_lock(&fs_info->trans_lock);
                                continue;
                        } else {
                                goto unlock;
                        }
                }

                ret = btrfs_commit_transaction(trans);
                mutex_lock(&fs_info->chunk_mutex);
                if (ret)
                        goto unlock;

                spin_lock(&fs_info->trans_lock);
        }
        spin_unlock(&fs_info->trans_lock);

        path = btrfs_alloc_path();
        if (!path) {
                ret = -ENOMEM;
                goto unlock;
        }

        path->reada = READA_FORWARD;
        path->search_commit_root = true;
        path->skip_locking = true;

        key.objectid = src_dev->devid;
        key.type = BTRFS_DEV_EXTENT_KEY;
        key.offset = 0;

        btrfs_for_each_slot(root, &key, &found_key, path, iter_ret) {
                struct extent_buffer *leaf = path->nodes[0];

                if (found_key.objectid != src_dev->devid)
                        break;

                if (found_key.type != BTRFS_DEV_EXTENT_KEY)
                        break;

                if (found_key.offset < key.offset)
                        break;

                dev_extent = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_dev_extent);

                chunk_offset = btrfs_dev_extent_chunk_offset(leaf, dev_extent);

                cache = btrfs_lookup_block_group(fs_info, chunk_offset);
                if (!cache)
                        continue;

                set_bit(BLOCK_GROUP_FLAG_TO_COPY, &cache->runtime_flags);
                btrfs_put_block_group(cache);
        }
        if (iter_ret < 0)
                ret = iter_ret;

        btrfs_free_path(path);
unlock:
        mutex_unlock(&fs_info->chunk_mutex);

        return ret;
}

bool btrfs_finish_block_group_to_copy(struct btrfs_device *srcdev,
                                      struct btrfs_block_group *cache,
                                      u64 physical)
{
        struct btrfs_fs_info *fs_info = cache->fs_info;
        struct btrfs_chunk_map *map;
        u64 chunk_offset = cache->start;
        int num_extents, cur_extent;
        int i;

        /* Do not use "to_copy" on non zoned filesystem for now */
        if (!btrfs_is_zoned(fs_info))
                return true;

        spin_lock(&cache->lock);
        if (test_bit(BLOCK_GROUP_FLAG_REMOVED, &cache->runtime_flags)) {
                spin_unlock(&cache->lock);
                return true;
        }
        spin_unlock(&cache->lock);

        map = btrfs_get_chunk_map(fs_info, chunk_offset, 1);
        ASSERT(!IS_ERR(map));

        num_extents = 0;
        cur_extent = 0;
        for (i = 0; i < map->num_stripes; i++) {
                /* We have more device extent to copy */
                if (srcdev != map->stripes[i].dev)
                        continue;

                num_extents++;
                if (physical == map->stripes[i].physical)
                        cur_extent = i;
        }

        btrfs_free_chunk_map(map);

        if (num_extents > 1 && cur_extent < num_extents - 1) {
                /*
                 * Has more stripes on this device. Keep this block group
                 * readonly until we finish all the stripes.
                 */
                return false;
        }

        /* Last stripe on this device */
        clear_bit(BLOCK_GROUP_FLAG_TO_COPY, &cache->runtime_flags);

        return true;
}

static int btrfs_dev_replace_start(struct btrfs_fs_info *fs_info,
                const char *tgtdev_name, u64 srcdevid, const char *srcdev_name,
                int read_src)
{
        struct btrfs_root *root = fs_info->dev_root;
        struct btrfs_trans_handle *trans;
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;
        int ret;
        struct btrfs_device *tgt_device = NULL;
        struct btrfs_device *src_device = NULL;

        src_device = btrfs_find_device_by_devspec(fs_info, srcdevid,
                                                  srcdev_name);
        if (IS_ERR(src_device))
                return PTR_ERR(src_device);

        if (btrfs_pinned_by_swapfile(fs_info, src_device)) {
                btrfs_warn(fs_info,
          "cannot replace device %s (devid %llu) due to active swapfile",
                        btrfs_dev_name(src_device), src_device->devid);
                return -ETXTBSY;
        }

        /*
         * Here we commit the transaction to make sure commit_total_bytes
         * of all the devices are updated.
         */
        trans = btrfs_attach_transaction(root);
        if (!IS_ERR(trans)) {
                ret = btrfs_commit_transaction(trans);
                if (ret)
                        return ret;
        } else if (PTR_ERR(trans) != -ENOENT) {
                return PTR_ERR(trans);
        }

        ret = btrfs_init_dev_replace_tgtdev(fs_info, tgtdev_name,
                                            src_device, &tgt_device);
        if (ret)
                return ret;

        ret = mark_block_group_to_copy(fs_info, src_device);
        if (ret)
                return ret;

        down_write(&dev_replace->rwsem);
        dev_replace->replace_task = current;
        switch (dev_replace->replace_state) {
        case BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_FINISHED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_CANCELED:
                break;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED:
                DEBUG_WARN("unexpected STARTED or SUSPENDED dev-replace state");
                ret = BTRFS_IOCTL_DEV_REPLACE_RESULT_ALREADY_STARTED;
                up_write(&dev_replace->rwsem);
                goto leave;
        }

        dev_replace->cont_reading_from_srcdev_mode = read_src;
        dev_replace->srcdev = src_device;
        dev_replace->tgtdev = tgt_device;

        btrfs_info(fs_info,
                      "dev_replace from %s (devid %llu) to %s started",
                      btrfs_dev_name(src_device),
                      src_device->devid,
                      btrfs_dev_name(tgt_device));

        /*
         * from now on, the writes to the srcdev are all duplicated to
         * go to the tgtdev as well (refer to btrfs_map_block()).
         */
        dev_replace->replace_state = BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED;
        dev_replace->time_started = ktime_get_real_seconds();
        dev_replace->cursor_left = 0;
        dev_replace->committed_cursor_left = 0;
        dev_replace->cursor_left_last_write_of_item = 0;
        dev_replace->cursor_right = 0;
        dev_replace->is_valid = 1;
        dev_replace->item_needs_writeback = 1;
        atomic64_set(&dev_replace->num_write_errors, 0);
        atomic64_set(&dev_replace->num_uncorrectable_read_errors, 0);
        up_write(&dev_replace->rwsem);

        ret = btrfs_sysfs_add_device(tgt_device);
        if (ret)
                btrfs_err(fs_info, "kobj add dev failed %d", ret);

        btrfs_wait_ordered_roots(fs_info, U64_MAX, NULL);

        /*
         * Commit dev_replace state and reserve 1 item for it.
         * This is crucial to ensure we won't miss copying extents for new block
         * groups that are allocated after we started the device replace, and
         * must be done after setting up the device replace state.
         */
        trans = btrfs_start_transaction(root, 1);
        if (IS_ERR(trans)) {
                ret = PTR_ERR(trans);
                down_write(&dev_replace->rwsem);
                dev_replace->replace_state =
                        BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED;
                dev_replace->srcdev = NULL;
                dev_replace->tgtdev = NULL;
                up_write(&dev_replace->rwsem);
                goto leave;
        }

        ret = btrfs_commit_transaction(trans);
        WARN_ON(ret);

        /* the disk copy procedure reuses the scrub code */
        ret = btrfs_scrub_dev(fs_info, src_device->devid, 0,
                              btrfs_device_get_total_bytes(src_device),
                              &dev_replace->scrub_progress, 0, 1);

        ret = btrfs_dev_replace_finishing(fs_info, ret);
        if (ret == -EINPROGRESS)
                ret = BTRFS_IOCTL_DEV_REPLACE_RESULT_SCRUB_INPROGRESS;

        return ret;

leave:
        btrfs_destroy_dev_replace_tgtdev(tgt_device);
        return ret;
}

static int btrfs_check_replace_dev_names(struct btrfs_ioctl_dev_replace_args *args)
{
        if (args->start.srcdevid == 0) {
                if (memchr(args->start.srcdev_name, 0,
                           sizeof(args->start.srcdev_name)) == NULL)
                        return -ENAMETOOLONG;
        } else {
                args->start.srcdev_name[0] = 0;
        }

        if (memchr(args->start.tgtdev_name, 0,
                   sizeof(args->start.tgtdev_name)) == NULL)
            return -ENAMETOOLONG;

        return 0;
}

int btrfs_dev_replace_by_ioctl(struct btrfs_fs_info *fs_info,
                            struct btrfs_ioctl_dev_replace_args *args)
{
        int ret;

        switch (args->start.cont_reading_from_srcdev_mode) {
        case BTRFS_IOCTL_DEV_REPLACE_CONT_READING_FROM_SRCDEV_MODE_ALWAYS:
        case BTRFS_IOCTL_DEV_REPLACE_CONT_READING_FROM_SRCDEV_MODE_AVOID:
                break;
        default:
                return -EINVAL;
        }
        ret = btrfs_check_replace_dev_names(args);
        if (ret < 0)
                return ret;

        ret = btrfs_dev_replace_start(fs_info, args->start.tgtdev_name,
                                        args->start.srcdevid,
                                        args->start.srcdev_name,
                                        args->start.cont_reading_from_srcdev_mode);
        args->result = ret;
        /* don't warn if EINPROGRESS, someone else might be running scrub */
        if (ret == BTRFS_IOCTL_DEV_REPLACE_RESULT_SCRUB_INPROGRESS ||
            ret == BTRFS_IOCTL_DEV_REPLACE_RESULT_NO_ERROR)
                return 0;

        return ret;
}

/*
 * blocked until all in-flight bios operations are finished.
 */
static void btrfs_rm_dev_replace_blocked(struct btrfs_fs_info *fs_info)
{
        set_bit(BTRFS_FS_STATE_DEV_REPLACING, &fs_info->fs_state);
        wait_event(fs_info->dev_replace.replace_wait, !percpu_counter_sum(
                   &fs_info->dev_replace.bio_counter));
}

/*
 * we have removed target device, it is safe to allow new bios request.
 */
static void btrfs_rm_dev_replace_unblocked(struct btrfs_fs_info *fs_info)
{
        clear_bit(BTRFS_FS_STATE_DEV_REPLACING, &fs_info->fs_state);
        wake_up(&fs_info->dev_replace.replace_wait);
}

/*
 * When finishing the device replace, before swapping the source device with the
 * target device we must update the chunk allocation state in the target device,
 * as it is empty because replace works by directly copying the chunks and not
 * through the normal chunk allocation path.
 */
static int btrfs_set_target_alloc_state(struct btrfs_device *srcdev,
                                        struct btrfs_device *tgtdev)
{
        struct extent_state *cached_state = NULL;
        u64 start = 0;
        u64 found_start;
        u64 found_end;
        int ret = 0;

        lockdep_assert_held(&srcdev->fs_info->chunk_mutex);

        while (btrfs_find_first_extent_bit(&srcdev->alloc_state, start,
                                           &found_start, &found_end,
                                           CHUNK_ALLOCATED, &cached_state)) {
                ret = btrfs_set_extent_bit(&tgtdev->alloc_state, found_start,
                                           found_end, CHUNK_ALLOCATED, NULL);
                if (ret)
                        break;
                start = found_end + 1;
        }

        btrfs_free_extent_state(cached_state);
        return ret;
}

static void btrfs_dev_replace_update_device_in_mapping_tree(
                                                struct btrfs_fs_info *fs_info,
                                                struct btrfs_device *srcdev,
                                                struct btrfs_device *tgtdev)
{
        struct rb_node *node;

        /*
         * The chunk mutex must be held so that no new chunks can be created
         * while we are updating existing chunks. This guarantees we don't miss
         * any new chunk that gets created for a range that falls before the
         * range of the last chunk we processed.
         */
        lockdep_assert_held(&fs_info->chunk_mutex);

        write_lock(&fs_info->mapping_tree_lock);
        node = rb_first_cached(&fs_info->mapping_tree);
        while (node) {
                struct rb_node *next = rb_next(node);
                struct btrfs_chunk_map *map;
                u64 next_start;

                map = rb_entry(node, struct btrfs_chunk_map, rb_node);
                next_start = map->start + map->chunk_len;

                for (int i = 0; i < map->num_stripes; i++)
                        if (srcdev == map->stripes[i].dev)
                                map->stripes[i].dev = tgtdev;

                if (cond_resched_rwlock_write(&fs_info->mapping_tree_lock)) {
                        map = btrfs_find_chunk_map_nolock(fs_info, next_start, U64_MAX);
                        if (!map)
                                break;
                        node = &map->rb_node;
                        /*
                         * Drop the lookup reference since we are holding the
                         * lock in write mode and no one can remove the chunk
                         * map from the tree and drop its tree reference.
                         */
                        btrfs_free_chunk_map(map);
                } else {
                        node = next;
                }
        }
        write_unlock(&fs_info->mapping_tree_lock);
}

static int btrfs_dev_replace_finishing(struct btrfs_fs_info *fs_info,
                                       int scrub_ret)
{
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;
        struct btrfs_fs_devices *fs_devices = fs_info->fs_devices;
        struct btrfs_device *tgt_device;
        struct btrfs_device *src_device;
        struct btrfs_root *root = fs_info->tree_root;
        u8 uuid_tmp[BTRFS_UUID_SIZE];
        struct btrfs_trans_handle *trans;
        int ret = 0;

        /* don't allow cancel or unmount to disturb the finishing procedure */
        mutex_lock(&dev_replace->lock_finishing_cancel_unmount);

        down_read(&dev_replace->rwsem);
        /* was the operation canceled, or is it finished? */
        if (dev_replace->replace_state !=
            BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED) {
                up_read(&dev_replace->rwsem);
                mutex_unlock(&dev_replace->lock_finishing_cancel_unmount);
                return 0;
        }

        tgt_device = dev_replace->tgtdev;
        src_device = dev_replace->srcdev;
        up_read(&dev_replace->rwsem);

        /*
         * flush all outstanding I/O and inode extent mappings before the
         * copy operation is declared as being finished
         */
        ret = btrfs_start_delalloc_roots(fs_info, LONG_MAX, false);
        if (ret) {
                mutex_unlock(&dev_replace->lock_finishing_cancel_unmount);
                return ret;
        }
        btrfs_wait_ordered_roots(fs_info, U64_MAX, NULL);

        /*
         * We have to use this loop approach because at this point src_device
         * has to be available for transaction commit to complete, yet new
         * chunks shouldn't be allocated on the device.
         */
        while (1) {
                trans = btrfs_start_transaction(root, 0);
                if (IS_ERR(trans)) {
                        mutex_unlock(&dev_replace->lock_finishing_cancel_unmount);
                        return PTR_ERR(trans);
                }
                ret = btrfs_commit_transaction(trans);
                WARN_ON(ret);

                /* Prevent write_all_supers() during the finishing procedure */
                mutex_lock(&fs_devices->device_list_mutex);
                /* Prevent new chunks being allocated on the source device */
                mutex_lock(&fs_info->chunk_mutex);

                if (!list_empty(&src_device->post_commit_list)) {
                        mutex_unlock(&fs_devices->device_list_mutex);
                        mutex_unlock(&fs_info->chunk_mutex);
                } else {
                        break;
                }
        }

        down_write(&dev_replace->rwsem);
        dev_replace->replace_state =
                scrub_ret ? BTRFS_IOCTL_DEV_REPLACE_STATE_CANCELED
                          : BTRFS_IOCTL_DEV_REPLACE_STATE_FINISHED;
        dev_replace->tgtdev = NULL;
        dev_replace->srcdev = NULL;
        dev_replace->time_stopped = ktime_get_real_seconds();
        dev_replace->item_needs_writeback = 1;

        /*
         * Update allocation state in the new device and replace the old device
         * with the new one in the mapping tree.
         */
        if (!scrub_ret) {
                scrub_ret = btrfs_set_target_alloc_state(src_device, tgt_device);
                if (scrub_ret)
                        goto error;
                btrfs_dev_replace_update_device_in_mapping_tree(fs_info,
                                                                src_device,
                                                                tgt_device);
        } else {
                if (scrub_ret != -ECANCELED)
                        btrfs_err(fs_info,
                                 "btrfs_scrub_dev(%s, %llu, %s) failed %d",
                                 btrfs_dev_name(src_device),
                                 src_device->devid,
                                 btrfs_dev_name(tgt_device), scrub_ret);
error:
                up_write(&dev_replace->rwsem);
                mutex_unlock(&fs_info->chunk_mutex);
                mutex_unlock(&fs_devices->device_list_mutex);
                btrfs_rm_dev_replace_blocked(fs_info);
                if (tgt_device)
                        btrfs_destroy_dev_replace_tgtdev(tgt_device);
                btrfs_rm_dev_replace_unblocked(fs_info);
                mutex_unlock(&dev_replace->lock_finishing_cancel_unmount);

                return scrub_ret;
        }

        btrfs_info(fs_info,
                          "dev_replace from %s (devid %llu) to %s finished",
                          btrfs_dev_name(src_device),
                          src_device->devid,
                          btrfs_dev_name(tgt_device));
        clear_bit(BTRFS_DEV_STATE_REPLACE_TGT, &tgt_device->dev_state);
        tgt_device->devid = src_device->devid;
        src_device->devid = BTRFS_DEV_REPLACE_DEVID;
        memcpy(uuid_tmp, tgt_device->uuid, sizeof(uuid_tmp));
        memcpy(tgt_device->uuid, src_device->uuid, sizeof(tgt_device->uuid));
        memcpy(src_device->uuid, uuid_tmp, sizeof(src_device->uuid));
        btrfs_device_set_total_bytes(tgt_device, src_device->total_bytes);
        btrfs_device_set_disk_total_bytes(tgt_device,
                                          src_device->disk_total_bytes);
        btrfs_device_set_bytes_used(tgt_device, src_device->bytes_used);
        tgt_device->commit_bytes_used = src_device->bytes_used;

        btrfs_assign_next_active_device(src_device, tgt_device);

        list_add(&tgt_device->dev_alloc_list, &fs_devices->alloc_list);
        fs_devices->rw_devices++;

        dev_replace->replace_task = NULL;
        up_write(&dev_replace->rwsem);
        btrfs_rm_dev_replace_blocked(fs_info);

        btrfs_rm_dev_replace_remove_srcdev(src_device);

        btrfs_rm_dev_replace_unblocked(fs_info);

        /*
         * Increment dev_stats_ccnt so that btrfs_run_dev_stats() will
         * update on-disk dev stats value during commit transaction
         */
        atomic_inc(&tgt_device->dev_stats_ccnt);

        /*
         * this is again a consistent state where no dev_replace procedure
         * is running, the target device is part of the filesystem, the
         * source device is not part of the filesystem anymore and its 1st
         * superblock is scratched out so that it is no longer marked to
         * belong to this filesystem.
         */
        mutex_unlock(&fs_info->chunk_mutex);
        mutex_unlock(&fs_devices->device_list_mutex);

        /* replace the sysfs entry */
        btrfs_sysfs_remove_device(src_device);
        btrfs_sysfs_update_devid(tgt_device);
        if (test_bit(BTRFS_DEV_STATE_WRITEABLE, &src_device->dev_state))
                btrfs_scratch_superblocks(fs_info, src_device);

        /* write back the superblocks */
        trans = btrfs_start_transaction(root, 0);
        if (!IS_ERR(trans))
                btrfs_commit_transaction(trans);

        mutex_unlock(&dev_replace->lock_finishing_cancel_unmount);

        btrfs_rm_dev_replace_free_srcdev(src_device);

        return 0;
}

/*
 * Read progress of device replace status according to the state and last
 * stored position. The value format is the same as for
 * btrfs_dev_replace::progress_1000
 */
static u64 btrfs_dev_replace_progress(struct btrfs_fs_info *fs_info)
{
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;
        u64 ret = 0;

        switch (dev_replace->replace_state) {
        case BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_CANCELED:
                ret = 0;
                break;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_FINISHED:
                ret = 1000;
                break;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED:
                ret = div64_u64(dev_replace->cursor_left,
                                div_u64(btrfs_device_get_total_bytes(
                                                dev_replace->srcdev), 1000));
                break;
        }

        return ret;
}

void btrfs_dev_replace_status(struct btrfs_fs_info *fs_info,
                              struct btrfs_ioctl_dev_replace_args *args)
{
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;

        down_read(&dev_replace->rwsem);
        /* even if !dev_replace_is_valid, the values are good enough for
         * the replace_status ioctl */
        args->result = BTRFS_IOCTL_DEV_REPLACE_RESULT_NO_ERROR;
        args->status.replace_state = dev_replace->replace_state;
        args->status.time_started = dev_replace->time_started;
        args->status.time_stopped = dev_replace->time_stopped;
        args->status.num_write_errors =
                atomic64_read(&dev_replace->num_write_errors);
        args->status.num_uncorrectable_read_errors =
                atomic64_read(&dev_replace->num_uncorrectable_read_errors);
        args->status.progress_1000 = btrfs_dev_replace_progress(fs_info);
        up_read(&dev_replace->rwsem);
}

int btrfs_dev_replace_cancel(struct btrfs_fs_info *fs_info)
{
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;
        struct btrfs_device *tgt_device = NULL;
        struct btrfs_device *src_device = NULL;
        struct btrfs_trans_handle *trans;
        struct btrfs_root *root = fs_info->tree_root;
        int result;
        int ret;

        if (sb_rdonly(fs_info->sb))
                return -EROFS;

        mutex_lock(&dev_replace->lock_finishing_cancel_unmount);
        down_write(&dev_replace->rwsem);
        switch (dev_replace->replace_state) {
        case BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_FINISHED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_CANCELED:
                result = BTRFS_IOCTL_DEV_REPLACE_RESULT_NOT_STARTED;
                up_write(&dev_replace->rwsem);
                break;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED:
                tgt_device = dev_replace->tgtdev;
                src_device = dev_replace->srcdev;
                up_write(&dev_replace->rwsem);
                ret = btrfs_scrub_cancel(fs_info);
                if (ret < 0) {
                        result = BTRFS_IOCTL_DEV_REPLACE_RESULT_NOT_STARTED;
                } else {
                        result = BTRFS_IOCTL_DEV_REPLACE_RESULT_NO_ERROR;
                        /*
                         * btrfs_dev_replace_finishing() will handle the
                         * cleanup part
                         */
                        btrfs_info(fs_info,
                                "dev_replace from %s (devid %llu) to %s canceled",
                                btrfs_dev_name(src_device), src_device->devid,
                                btrfs_dev_name(tgt_device));
                }
                break;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED:
                /*
                 * Scrub doing the replace isn't running so we need to do the
                 * cleanup step of btrfs_dev_replace_finishing() here
                 */
                result = BTRFS_IOCTL_DEV_REPLACE_RESULT_NO_ERROR;
                tgt_device = dev_replace->tgtdev;
                src_device = dev_replace->srcdev;
                dev_replace->tgtdev = NULL;
                dev_replace->srcdev = NULL;
                dev_replace->replace_state =
                                BTRFS_IOCTL_DEV_REPLACE_STATE_CANCELED;
                dev_replace->time_stopped = ktime_get_real_seconds();
                dev_replace->item_needs_writeback = 1;

                up_write(&dev_replace->rwsem);

                /* Scrub for replace must not be running in suspended state */
                btrfs_scrub_cancel(fs_info);

                trans = btrfs_start_transaction(root, 0);
                if (IS_ERR(trans)) {
                        mutex_unlock(&dev_replace->lock_finishing_cancel_unmount);
                        return PTR_ERR(trans);
                }
                ret = btrfs_commit_transaction(trans);
                WARN_ON(ret);

                btrfs_info(fs_info,
                "suspended dev_replace from %s (devid %llu) to %s canceled",
                        btrfs_dev_name(src_device), src_device->devid,
                        btrfs_dev_name(tgt_device));

                if (tgt_device)
                        btrfs_destroy_dev_replace_tgtdev(tgt_device);
                break;
        default:
                up_write(&dev_replace->rwsem);
                result = -EINVAL;
        }

        mutex_unlock(&dev_replace->lock_finishing_cancel_unmount);
        return result;
}

void btrfs_dev_replace_suspend_for_unmount(struct btrfs_fs_info *fs_info)
{
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;

        mutex_lock(&dev_replace->lock_finishing_cancel_unmount);
        down_write(&dev_replace->rwsem);

        switch (dev_replace->replace_state) {
        case BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_FINISHED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_CANCELED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED:
                break;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED:
                dev_replace->replace_state =
                        BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED;
                dev_replace->time_stopped = ktime_get_real_seconds();
                dev_replace->item_needs_writeback = 1;
                btrfs_info(fs_info, "suspending dev_replace for unmount");
                break;
        }

        up_write(&dev_replace->rwsem);
        mutex_unlock(&dev_replace->lock_finishing_cancel_unmount);
}

/* resume dev_replace procedure that was interrupted by unmount */
int btrfs_resume_dev_replace_async(struct btrfs_fs_info *fs_info)
{
        struct task_struct *task;
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;

        down_write(&dev_replace->rwsem);

        switch (dev_replace->replace_state) {
        case BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_FINISHED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_CANCELED:
                up_write(&dev_replace->rwsem);
                return 0;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED:
                break;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED:
                dev_replace->replace_state =
                        BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED;
                break;
        }
        if (!dev_replace->tgtdev || !dev_replace->tgtdev->bdev) {
                btrfs_info(fs_info,
                           "cannot continue dev_replace, tgtdev is missing");
                btrfs_info(fs_info,
                           "you may cancel the operation after 'mount -o degraded'");
                dev_replace->replace_state =
                                        BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED;
                up_write(&dev_replace->rwsem);
                return 0;
        }
        up_write(&dev_replace->rwsem);

        /*
         * This could collide with a paused balance, but the exclusive op logic
         * should never allow both to start and pause. We don't want to allow
         * dev-replace to start anyway.
         */
        if (!btrfs_exclop_start(fs_info, BTRFS_EXCLOP_DEV_REPLACE)) {
                down_write(&dev_replace->rwsem);
                dev_replace->replace_state =
                                        BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED;
                up_write(&dev_replace->rwsem);
                btrfs_info(fs_info,
                "cannot resume dev-replace, other exclusive operation running");
                return 0;
        }

        task = kthread_run(btrfs_dev_replace_kthread, fs_info, "btrfs-devrepl");
        return PTR_ERR_OR_ZERO(task);
}

static int btrfs_dev_replace_kthread(void *data)
{
        struct btrfs_fs_info *fs_info = data;
        struct btrfs_dev_replace *dev_replace = &fs_info->dev_replace;
        u64 progress;
        int ret;

        progress = btrfs_dev_replace_progress(fs_info);
        progress = div_u64(progress, 10);
        btrfs_info(fs_info,
                "continuing dev_replace from %s (devid %llu) to target %s @%u%%",
                btrfs_dev_name(dev_replace->srcdev),
                dev_replace->srcdev->devid,
                btrfs_dev_name(dev_replace->tgtdev),
                (unsigned int)progress);

        ret = btrfs_scrub_dev(fs_info, dev_replace->srcdev->devid,
                              dev_replace->committed_cursor_left,
                              btrfs_device_get_total_bytes(dev_replace->srcdev),
                              &dev_replace->scrub_progress, 0, 1);
        ret = btrfs_dev_replace_finishing(fs_info, ret);
        WARN_ON(ret && ret != -ECANCELED);

        btrfs_exclop_finish(fs_info);
        return 0;
}

bool __pure btrfs_dev_replace_is_ongoing(struct btrfs_dev_replace *dev_replace)
{
        if (!dev_replace->is_valid)
                return false;

        switch (dev_replace->replace_state) {
        case BTRFS_IOCTL_DEV_REPLACE_STATE_NEVER_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_FINISHED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_CANCELED:
                return false;
        case BTRFS_IOCTL_DEV_REPLACE_STATE_STARTED:
        case BTRFS_IOCTL_DEV_REPLACE_STATE_SUSPENDED:
                /*
                 * return true even if tgtdev is missing (this is
                 * something that can happen if the dev_replace
                 * procedure is suspended by an umount and then
                 * the tgtdev is missing (or "btrfs dev scan") was
                 * not called and the filesystem is remounted
                 * in degraded state. This does not stop the
                 * dev_replace procedure. It needs to be canceled
                 * manually if the cancellation is wanted.
                 */
                break;
        }
        return true;
}

void btrfs_bio_counter_sub(struct btrfs_fs_info *fs_info, s64 amount)
{
        percpu_counter_sub(&fs_info->dev_replace.bio_counter, amount);
        cond_wake_up_nomb(&fs_info->dev_replace.replace_wait);
}

void btrfs_bio_counter_inc_blocked(struct btrfs_fs_info *fs_info)
{
        while (1) {
                percpu_counter_inc(&fs_info->dev_replace.bio_counter);
                if (likely(!test_bit(BTRFS_FS_STATE_DEV_REPLACING,
                                     &fs_info->fs_state)))
                        break;

                btrfs_bio_counter_dec(fs_info);
                wait_event(fs_info->dev_replace.replace_wait,
                           !test_bit(BTRFS_FS_STATE_DEV_REPLACING,
                                     &fs_info->fs_state));
        }
}