aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--include/migration/migration.h1
-rw-r--r--migration/migration.c6
-rw-r--r--migration/ram.c1
3 files changed, 3 insertions, 5 deletions
diff --git a/include/migration/migration.h b/include/migration/migration.h
index 2f9c9736f9..e185d13c9c 100644
--- a/include/migration/migration.h
+++ b/include/migration/migration.h
@@ -155,7 +155,6 @@ struct MigrationState
int64_t downtime;
int64_t expected_downtime;
int64_t dirty_pages_rate;
- int64_t dirty_bytes_rate;
bool enabled_capabilities[MIGRATION_CAPABILITY__MAX];
int64_t xbzrle_cache_size;
int64_t setup_time;
diff --git a/migration/migration.c b/migration/migration.c
index 94647a81e3..dcf9b944cc 100644
--- a/migration/migration.c
+++ b/migration/migration.c
@@ -1118,7 +1118,6 @@ MigrationState *migrate_init(const MigrationParams *params)
s->downtime = 0;
s->expected_downtime = 0;
s->dirty_pages_rate = 0;
- s->dirty_bytes_rate = 0;
s->setup_time = 0;
s->start_postcopy = false;
s->postcopy_after_devices = false;
@@ -2008,8 +2007,9 @@ static void *migration_thread(void *opaque)
bandwidth, max_size);
/* if we haven't sent anything, we don't want to recalculate
10000 is a small enough number for our purposes */
- if (s->dirty_bytes_rate && transferred_bytes > 10000) {
- s->expected_downtime = s->dirty_bytes_rate / bandwidth;
+ if (s->dirty_pages_rate && transferred_bytes > 10000) {
+ s->expected_downtime = s->dirty_pages_rate *
+ (1ul << qemu_target_page_bits()) / bandwidth;
}
qemu_file_reset_rate_limit(s->to_dst_file);
diff --git a/migration/ram.c b/migration/ram.c
index 1580617b38..ef4f508b6c 100644
--- a/migration/ram.c
+++ b/migration/ram.c
@@ -728,7 +728,6 @@ static void migration_bitmap_sync(RAMState *rs)
}
s->dirty_pages_rate = rs->num_dirty_pages_period * 1000
/ (end_time - rs->time_last_bitmap_sync);
- s->dirty_bytes_rate = s->dirty_pages_rate * TARGET_PAGE_SIZE;
rs->time_last_bitmap_sync = end_time;
rs->num_dirty_pages_period = 0;
}