diff options
-rw-r--r-- | block-migration.c | 149 |
1 files changed, 50 insertions, 99 deletions
diff --git a/block-migration.c b/block-migration.c index 2f89a4ee9a..99fe3331c5 100644 --- a/block-migration.c +++ b/block-migration.c @@ -85,31 +85,25 @@ static void blk_mig_read_cb(void *opaque, int ret) assert(block_mig_state.submitted >= 0); } -static int mig_read_device_bulk(QEMUFile *f, BlkMigDevState *bms) +static int mig_save_device_bulk(QEMUFile *f, BlkMigDevState *bmds, int is_async) { - int nr_sectors; - int64_t total_sectors, cur_sector = 0; - BlockDriverState *bs = bms->bs; + int64_t total_sectors = bmds->total_sectors; + int64_t cur_sector = bmds->cur_sector; + BlockDriverState *bs = bmds->bs; + int len, nr_sectors; BlkMigBlock *blk; + uint8_t *tmp_buf; - blk = qemu_malloc(sizeof(BlkMigBlock)); - blk->buf = qemu_malloc(BLOCK_SIZE); - - cur_sector = bms->cur_sector; - total_sectors = bms->total_sectors; - - if (bms->shared_base) { + if (bmds->shared_base) { while (cur_sector < total_sectors && - !bdrv_is_allocated(bms->bs, cur_sector, - MAX_IS_ALLOCATED_SEARCH, &nr_sectors)) { + !bdrv_is_allocated(bs, cur_sector, MAX_IS_ALLOCATED_SEARCH, + &nr_sectors)) { cur_sector += nr_sectors; } } if (cur_sector >= total_sectors) { - bms->cur_sector = total_sectors; - qemu_free(blk->buf); - qemu_free(blk); + bmds->cur_sector = total_sectors; return 1; } @@ -120,101 +114,65 @@ static int mig_read_device_bulk(QEMUFile *f, BlkMigDevState *bms) (BDRV_SECTORS_PER_DIRTY_CHUNK * 10000); } + cur_sector &= ~((int64_t)BDRV_SECTORS_PER_DIRTY_CHUNK - 1); + /* we are going to transfer a full block even if it is not allocated */ nr_sectors = BDRV_SECTORS_PER_DIRTY_CHUNK; - cur_sector &= ~((int64_t)BDRV_SECTORS_PER_DIRTY_CHUNK - 1); - if (total_sectors - cur_sector < BDRV_SECTORS_PER_DIRTY_CHUNK) { - nr_sectors = (total_sectors - cur_sector); + nr_sectors = total_sectors - cur_sector; } - bms->cur_sector = cur_sector + nr_sectors; - blk->sector = cur_sector; - blk->bmds = bms; + if (is_async) { + blk = qemu_malloc(sizeof(BlkMigBlock)); + blk->buf = qemu_malloc(BLOCK_SIZE); - blk->iov.iov_base = blk->buf; - blk->iov.iov_len = nr_sectors * BDRV_SECTOR_SIZE; - qemu_iovec_init_external(&blk->qiov, &blk->iov, 1); + bmds->cur_sector = cur_sector + nr_sectors; + blk->sector = cur_sector; + blk->bmds = bmds; - blk->aiocb = bdrv_aio_readv(bs, cur_sector, &blk->qiov, - nr_sectors, blk_mig_read_cb, blk); + blk->iov.iov_base = blk->buf; + blk->iov.iov_len = nr_sectors * BDRV_SECTOR_SIZE; + qemu_iovec_init_external(&blk->qiov, &blk->iov, 1); - if (!blk->aiocb) { - printf("Error reading sector %" PRId64 "\n", cur_sector); - qemu_free(blk->buf); - qemu_free(blk); - return 0; - } - - bdrv_reset_dirty(bms->bs, cur_sector, nr_sectors); - block_mig_state.submitted++; - - return (bms->cur_sector >= total_sectors); -} + blk->aiocb = bdrv_aio_readv(bs, cur_sector, &blk->qiov, + nr_sectors, blk_mig_read_cb, blk); -static int mig_save_device_bulk(QEMUFile *f, BlkMigDevState *bmds) -{ - int len, nr_sectors; - int64_t total_sectors = bmds->total_sectors, cur_sector = 0; - uint8_t *tmp_buf = NULL; - BlockDriverState *bs = bmds->bs; + if (!blk->aiocb) { + printf("Error reading sector %" PRId64 "\n", cur_sector); + qemu_free(blk->buf); + qemu_free(blk); + return 0; + } - tmp_buf = qemu_malloc(BLOCK_SIZE); + bdrv_reset_dirty(bs, cur_sector, nr_sectors); + block_mig_state.submitted++; - cur_sector = bmds->cur_sector; + } else { + tmp_buf = qemu_malloc(BLOCK_SIZE); - if (bmds->shared_base) { - while (cur_sector < total_sectors && - !bdrv_is_allocated(bmds->bs, cur_sector, - MAX_IS_ALLOCATED_SEARCH, &nr_sectors)) { - cur_sector += nr_sectors; + if (bdrv_read(bs, cur_sector, tmp_buf, nr_sectors) < 0) { + printf("Error reading sector %" PRId64 "\n", cur_sector); } - } - if (cur_sector >= total_sectors) { - bmds->cur_sector = total_sectors; - qemu_free(tmp_buf); - return 1; - } + bdrv_reset_dirty(bs, cur_sector, nr_sectors); - if (cur_sector >= block_mig_state.print_completion) { - printf("Completed %" PRId64 " %%\r", cur_sector * 100 / total_sectors); - fflush(stdout); - block_mig_state.print_completion += - (BDRV_SECTORS_PER_DIRTY_CHUNK * 10000); - } + /* sector number and flags */ + qemu_put_be64(f, (cur_sector << BDRV_SECTOR_BITS) + | BLK_MIG_FLAG_DEVICE_BLOCK); - cur_sector &= ~((int64_t)BDRV_SECTORS_PER_DIRTY_CHUNK - 1); + /* device name */ + len = strlen(bs->device_name); + qemu_put_byte(f, len); + qemu_put_buffer(f, (uint8_t *)bs->device_name, len); - /* we are going to transfer a full block even if it is not allocated */ - nr_sectors = BDRV_SECTORS_PER_DIRTY_CHUNK; + qemu_put_buffer(f, tmp_buf, BLOCK_SIZE); - if (total_sectors - cur_sector < BDRV_SECTORS_PER_DIRTY_CHUNK) { - nr_sectors = (total_sectors - cur_sector); - } + bmds->cur_sector = cur_sector + BDRV_SECTORS_PER_DIRTY_CHUNK; - if (bdrv_read(bs, cur_sector, tmp_buf, nr_sectors) < 0) { - printf("Error reading sector %" PRId64 "\n", cur_sector); + qemu_free(tmp_buf); } - bdrv_reset_dirty(bs, cur_sector, nr_sectors); - - /* sector number and flags */ - qemu_put_be64(f, (cur_sector << BDRV_SECTOR_BITS) - | BLK_MIG_FLAG_DEVICE_BLOCK); - - /* device name */ - len = strlen(bs->device_name); - qemu_put_byte(f, len); - qemu_put_buffer(f, (uint8_t *)bs->device_name, len); - - qemu_put_buffer(f, tmp_buf, BLOCK_SIZE); - - bmds->cur_sector = cur_sector + BDRV_SECTORS_PER_DIRTY_CHUNK; - - qemu_free(tmp_buf); - return (bmds->cur_sector >= total_sectors); } @@ -279,16 +237,9 @@ static int blk_mig_save_bulked_block(QEMUFile *f, int is_async) QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) { if (bmds->bulk_completed == 0) { - if (is_async) { - if (mig_read_device_bulk(f, bmds) == 1) { - /* completed bulk section for this device */ - bmds->bulk_completed = 1; - } - } else { - if (mig_save_device_bulk(f, bmds) == 1) { - /* completed bulk section for this device */ - bmds->bulk_completed = 1; - } + if (mig_save_device_bulk(f, bmds, is_async) == 1) { + /* completed bulk section for this device */ + bmds->bulk_completed = 1; } return 1; } |