diff options
author | Juan Quintela <quintela@redhat.com> | 2017-03-28 15:02:43 +0200 |
---|---|---|
committer | Juan Quintela <quintela@redhat.com> | 2017-04-21 12:25:36 +0200 |
commit | a66cd90c74f2a521d2441ba591f822cc93291cab (patch) | |
tree | 04f1f113cd235c5fa10462919c015d859cbebc3b /migration | |
parent | c4bdf0cf4b62a90b07f27f55d4650dc1cb3f8d7e (diff) |
ram: Move num_dirty_pages_period into RAMState
Signed-off-by: Juan Quintela <quintela@redhat.com>
Reviewed-by: Dr. David Alan Gilbert <dgilbert@redhat.com>
Reviewed-by: Peter Xu <peterx@redhat.com>
Diffstat (limited to 'migration')
-rw-r--r-- | migration/ram.c | 20 |
1 files changed, 11 insertions, 9 deletions
diff --git a/migration/ram.c b/migration/ram.c index d13674f4b4..d6cf032342 100644 --- a/migration/ram.c +++ b/migration/ram.c @@ -159,6 +159,8 @@ struct RAMState { int64_t time_last_bitmap_sync; /* bytes transferred at start_time */ uint64_t bytes_xfer_prev; + /* number of dirty pages since start_time */ + int64_t num_dirty_pages_period; }; typedef struct RAMState RAMState; @@ -612,13 +614,13 @@ static inline bool migration_bitmap_clear_dirty(ram_addr_t addr) return ret; } -static int64_t num_dirty_pages_period; -static void migration_bitmap_sync_range(ram_addr_t start, ram_addr_t length) +static void migration_bitmap_sync_range(RAMState *rs, ram_addr_t start, + ram_addr_t length) { unsigned long *bitmap; bitmap = atomic_rcu_read(&migration_bitmap_rcu)->bmap; migration_dirty_pages += cpu_physical_memory_sync_dirty_bitmap(bitmap, - start, length, &num_dirty_pages_period); + start, length, &rs->num_dirty_pages_period); } /* Fix me: there are too many global variables used in migration process. */ @@ -629,7 +631,7 @@ static void migration_bitmap_sync_init(RAMState *rs) { rs->time_last_bitmap_sync = 0; rs->bytes_xfer_prev = 0; - num_dirty_pages_period = 0; + rs->num_dirty_pages_period = 0; xbzrle_cache_miss_prev = 0; iterations_prev = 0; } @@ -678,12 +680,12 @@ static void migration_bitmap_sync(RAMState *rs) qemu_mutex_lock(&migration_bitmap_mutex); rcu_read_lock(); QLIST_FOREACH_RCU(block, &ram_list.blocks, next) { - migration_bitmap_sync_range(block->offset, block->used_length); + migration_bitmap_sync_range(rs, block->offset, block->used_length); } rcu_read_unlock(); qemu_mutex_unlock(&migration_bitmap_mutex); - trace_migration_bitmap_sync_end(num_dirty_pages_period); + trace_migration_bitmap_sync_end(rs->num_dirty_pages_period); end_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME); @@ -698,7 +700,7 @@ static void migration_bitmap_sync(RAMState *rs) bytes_xfer_now = ram_bytes_transferred(); if (s->dirty_pages_rate && - (num_dirty_pages_period * TARGET_PAGE_SIZE > + (rs->num_dirty_pages_period * TARGET_PAGE_SIZE > (bytes_xfer_now - rs->bytes_xfer_prev) / 2) && (rs->dirty_rate_high_cnt++ >= 2)) { trace_migration_throttle(); @@ -718,11 +720,11 @@ static void migration_bitmap_sync(RAMState *rs) iterations_prev = acct_info.iterations; xbzrle_cache_miss_prev = acct_info.xbzrle_cache_miss; } - s->dirty_pages_rate = num_dirty_pages_period * 1000 + s->dirty_pages_rate = rs->num_dirty_pages_period * 1000 / (end_time - rs->time_last_bitmap_sync); s->dirty_bytes_rate = s->dirty_pages_rate * TARGET_PAGE_SIZE; rs->time_last_bitmap_sync = end_time; - num_dirty_pages_period = 0; + rs->num_dirty_pages_period = 0; } s->dirty_sync_count = rs->bitmap_sync_count; if (migrate_use_events()) { |