From patchwork Wed Aug 3 12:25:51 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Zhanghailiang X-Patchwork-Id: 9261151 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork.web.codeaurora.org (Postfix) with ESMTP id 915E66048B for ; Wed, 3 Aug 2016 12:32:10 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 82CC028521 for ; Wed, 3 Aug 2016 12:32:10 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 777042855E; Wed, 3 Aug 2016 12:32:10 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-6.9 required=2.0 tests=BAYES_00,RCVD_IN_DNSWL_HI autolearn=ham version=3.3.1 Received: from lists.gnu.org (lists.gnu.org [208.118.235.17]) (using TLSv1 with cipher AES256-SHA (256/256 bits)) (No client certificate requested) by mail.wl.linuxfoundation.org (Postfix) with ESMTPS id 7494B28521 for ; Wed, 3 Aug 2016 12:32:09 +0000 (UTC) Received: from localhost ([::1]:34198 helo=lists.gnu.org) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1bUvLD-0006tD-QC for patchwork-qemu-devel@patchwork.kernel.org; Wed, 03 Aug 2016 08:32:07 -0400 Received: from eggs.gnu.org ([2001:4830:134:3::10]:60718) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1bUvGI-0000eE-H8 for qemu-devel@nongnu.org; Wed, 03 Aug 2016 08:27:03 -0400 Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.71) (envelope-from ) id 1bUvGF-0006Yi-A2 for qemu-devel@nongnu.org; Wed, 03 Aug 2016 08:27:02 -0400 Received: from szxga03-in.huawei.com ([119.145.14.66]:42751) by eggs.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1bUvGE-0006X2-Io for qemu-devel@nongnu.org; Wed, 03 Aug 2016 08:26:59 -0400 Received: from 172.24.1.136 (EHLO szxeml427-hub.china.huawei.com) ([172.24.1.136]) by szxrg03-dlp.huawei.com (MOS 4.4.3-GA FastPath queued) with ESMTP id CFU48218; Wed, 03 Aug 2016 20:26:43 +0800 (CST) Received: from localhost (10.177.24.212) by szxeml427-hub.china.huawei.com (10.82.67.182) with Microsoft SMTP Server id 14.3.235.1; Wed, 3 Aug 2016 20:26:33 +0800 From: zhanghailiang To: Date: Wed, 3 Aug 2016 20:25:51 +0800 Message-ID: <1470227172-13704-14-git-send-email-zhang.zhanghailiang@huawei.com> X-Mailer: git-send-email 2.7.2.windows.1 In-Reply-To: <1470227172-13704-1-git-send-email-zhang.zhanghailiang@huawei.com> References: <1470227172-13704-1-git-send-email-zhang.zhanghailiang@huawei.com> MIME-Version: 1.0 X-Originating-IP: [10.177.24.212] X-CFilter-Loop: Reflected X-Mirapoint-Virus-RAPID-Raw: score=unknown(0), refid=str=0001.0A0B0207.57A1E304.00A3, ss=1, re=0.000, recu=0.000, reip=0.000, cl=1, cld=1, fgs=0, ip=0.0.0.0, so=2013-05-26 15:14:31, dmn=2013-03-21 17:37:32 X-Mirapoint-Loop-Id: 8a33147195847d3091d907f94f454d1b X-detected-operating-system: by eggs.gnu.org: GNU/Linux 2.4.x-2.6.x [generic] X-Received-From: 119.145.14.66 Subject: [Qemu-devel] [PATCH COLO-Frame v18 13/34] COLO: Flush PVM's cached RAM into SVM's memory X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.21 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: xiecl.fnst@cn.fujitsu.com, lizhijian@cn.fujitsu.com, quintela@redhat.com, eddie.dong@intel.com, peter.huangpeng@huawei.com, dgilbert@redhat.com, Gonglei , amit.shah@redhat.com, zhangchen.fnst@cn.fujitsu.com, zhanghailiang Errors-To: qemu-devel-bounces+patchwork-qemu-devel=patchwork.kernel.org@nongnu.org Sender: "Qemu-devel" X-Virus-Scanned: ClamAV using ClamSMTP During the time of VM's running, PVM may dirty some pages, we will transfer PVM's dirty pages to SVM and store them into SVM's RAM cache at next checkpoint time. So, the content of SVM's RAM cache will always be same with PVM's memory after checkpoint. Instead of flushing all content of PVM's RAM cache into SVM's MEMORY, we do this in a more efficient way: Only flush any page that dirtied by PVM since last checkpoint. In this way, we can ensure SVM's memory same with PVM's. Besides, we must ensure flush RAM cache before load device state. Signed-off-by: zhanghailiang Signed-off-by: Li Zhijian Signed-off-by: Gonglei Reviewed-by: Dr. David Alan Gilbert --- v12: - Add a trace point in the end of colo_flush_ram_cache() (Dave's suggestion) - Add Reviewed-by tag v11: - Move the place of 'need_flush' (Dave's suggestion) - Remove unused 'DPRINTF("Flush ram_cache\n")' v10: - trace the number of dirty pages that be received. --- include/migration/migration.h | 1 + migration/colo.c | 2 -- migration/ram.c | 38 ++++++++++++++++++++++++++++++++++++++ migration/trace-events | 2 ++ 4 files changed, 41 insertions(+), 2 deletions(-) diff --git a/include/migration/migration.h b/include/migration/migration.h index 55679b5..8ae5248 100644 --- a/include/migration/migration.h +++ b/include/migration/migration.h @@ -366,4 +366,5 @@ PostcopyState postcopy_state_set(PostcopyState new_state); /* ram cache */ int colo_init_ram_cache(void); void colo_release_ram_cache(void); +void colo_flush_ram_cache(void); #endif diff --git a/migration/colo.c b/migration/colo.c index 8fa2104..58bd5e1 100644 --- a/migration/colo.c +++ b/migration/colo.c @@ -407,8 +407,6 @@ void *colo_process_incoming_thread(void *opaque) } qemu_mutex_unlock_iothread(); - /* TODO: flush vm state */ - colo_send_message(mis->to_src_file, COLO_MESSAGE_VMSTATE_LOADED, &local_err); if (local_err) { diff --git a/migration/ram.c b/migration/ram.c index 99449c3..198dc53 100644 --- a/migration/ram.c +++ b/migration/ram.c @@ -2499,6 +2499,7 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id) * be atomic */ bool postcopy_running = postcopy_state_get() >= POSTCOPY_INCOMING_LISTENING; + bool need_flush = false; seq_iter++; @@ -2533,6 +2534,7 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id) /* After going into COLO, we should load the Page into colo_cache */ if (ram_cache_enable) { host = colo_cache_from_block_offset(block, addr); + need_flush = true; } else { host = host_from_ram_block_offset(block, addr); } @@ -2626,6 +2628,10 @@ static int ram_load(QEMUFile *f, void *opaque, int version_id) wait_for_decompress_done(); rcu_read_unlock(); + + if (!ret && ram_cache_enable && need_flush) { + colo_flush_ram_cache(); + } DPRINTF("Completed load of VM with exit code %d seq iteration " "%" PRIu64 "\n", ret, seq_iter); return ret; @@ -2698,6 +2704,38 @@ void colo_release_ram_cache(void) rcu_read_unlock(); } +/* + * Flush content of RAM cache into SVM's memory. + * Only flush the pages that be dirtied by PVM or SVM or both. + */ +void colo_flush_ram_cache(void) +{ + RAMBlock *block = NULL; + void *dst_host; + void *src_host; + ram_addr_t offset = 0; + + trace_colo_flush_ram_cache_begin(migration_dirty_pages); + rcu_read_lock(); + block = QLIST_FIRST_RCU(&ram_list.blocks); + while (block) { + ram_addr_t ram_addr_abs; + offset = migration_bitmap_find_dirty(block, offset, &ram_addr_abs); + migration_bitmap_clear_dirty(ram_addr_abs); + if (offset >= block->used_length) { + offset = 0; + block = QLIST_NEXT_RCU(block, next); + } else { + dst_host = block->host + offset; + src_host = block->colo_cache + offset; + memcpy(dst_host, src_host, TARGET_PAGE_SIZE); + } + } + rcu_read_unlock(); + trace_colo_flush_ram_cache_end(); + assert(migration_dirty_pages == 0); +} + static SaveVMHandlers savevm_ram_handlers = { .save_live_setup = ram_save_setup, .save_live_iterate = ram_save_iterate, diff --git a/migration/trace-events b/migration/trace-events index e98eaef..82dc7a0 100644 --- a/migration/trace-events +++ b/migration/trace-events @@ -65,6 +65,8 @@ migration_throttle(void) "" ram_load_postcopy_loop(uint64_t addr, int flags) "@%" PRIx64 " %x" ram_postcopy_send_discard_bitmap(void) "" ram_save_queue_pages(const char *rbname, size_t start, size_t len) "%s: start: %zx len: %zx" +colo_flush_ram_cache_begin(uint64_t dirty_pages) "dirty_pages %" PRIu64 +colo_flush_ram_cache_end(void) "" # migration/migration.c await_return_path_close_on_source_close(void) ""