From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-16.7 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 8AB30C433B4 for ; Fri, 9 Apr 2021 03:44:26 +0000 (UTC) Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 2934261042 for ; Fri, 9 Apr 2021 03:44:24 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 2934261042 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=intel.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Received: from localhost ([::1]:40584 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1lUi4F-0005JA-90 for qemu-devel@archiver.kernel.org; Thu, 08 Apr 2021 23:44:23 -0400 Received: from eggs.gnu.org ([2001:470:142:3::10]:52072) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1lUi2s-0003fO-BP for qemu-devel@nongnu.org; Thu, 08 Apr 2021 23:43:00 -0400 Received: from mga05.intel.com ([192.55.52.43]:12420) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1lUi2p-0007T4-KB for qemu-devel@nongnu.org; Thu, 08 Apr 2021 23:42:58 -0400 IronPort-SDR: fmsJs6bfNwy0wqZfGWrfq3EXQto5puMGdD4tyNSqYhd/1Fmjb4kM8K3jh9IREzsIDC2e2JFkKO PCrVZKebiJzg== X-IronPort-AV: E=McAfee;i="6000,8403,9948"; a="278957370" X-IronPort-AV: E=Sophos;i="5.82,208,1613462400"; d="scan'208";a="278957370" Received: from fmsmga008.fm.intel.com ([10.253.24.58]) by fmsmga105.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 08 Apr 2021 20:42:55 -0700 IronPort-SDR: vvuZ+A8Vq+LlWFR6A+2atACfpWSjlVJhde3R+f48B2/FvbFuGBskyCRj8jmCkySl4G2XV3EXXS L+OtHXNalrSQ== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.82,208,1613462400"; d="scan'208";a="416081168" Received: from unknown (HELO localhost.localdomain.bj.intel.com) ([10.240.192.103]) by fmsmga008.fm.intel.com with ESMTP; 08 Apr 2021 20:42:52 -0700 From: leirao To: chen.zhang@intel.com, lizhijian@cn.fujitsu.com, jasowang@redhat.com, quintela@redhat.com, dgilbert@redhat.com, pbonzini@redhat.com, lukasstraub2@web.de Subject: [PATCH v6 09/10] Add the function of colo_bitmap_clear_dirty Date: Thu, 8 Apr 2021 23:20:55 -0400 Message-Id: <1617938456-315058-10-git-send-email-lei.rao@intel.com> X-Mailer: git-send-email 1.8.3.1 In-Reply-To: <1617938456-315058-1-git-send-email-lei.rao@intel.com> References: <1617938456-315058-1-git-send-email-lei.rao@intel.com> Received-SPF: pass client-ip=192.55.52.43; envelope-from=lei.rao@intel.com; helo=mga05.intel.com X-Spam_score_int: -41 X-Spam_score: -4.2 X-Spam_bar: ---- X-Spam_report: (-4.2 / 5.0 requ) BAYES_00=-1.9, RCVD_IN_DNSWL_MED=-2.3, SPF_HELO_NONE=0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: "Rao, Lei" , qemu-devel@nongnu.org Errors-To: qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Sender: "Qemu-devel" From: "Rao, Lei" When we use continuous dirty memory copy for flushing ram cache on secondary VM, we can also clean up the bitmap of contiguous dirty page memory. This also can reduce the VM stop time during checkpoint. The performance test for COLO as follow: Server configuraton: CPU :Intel(R) Xeon(R) Gold 6140 CPU @ 2.30GHz MEM :251G(type:DDR4 Speed:2666 MT/s) SSD :Intel 730 and DC S35x0/3610/3700 Series SSDs dirty pages:3189376 migration_bitmap_clear_dirty time consuming(ns):105194000 dirty pages:3189784 migration_bitmap_clear_dirty time consuming(ns):105297000 dirty pages:3190501 migration_bitmap_clear_dirty time consuming(ns):105410000 dirty pages:3188734 migration_bitmap_clear_dirty time consuming(ns):105138000 dirty pages:3189464 migration_bitmap_clear_dirty time consuming(ns):111736000 dirty pages:3188558 migration_bitmap_clear_dirty time consuming(ns):105079000 dirty pages:3239489 migration_bitmap_clear_dirty time consuming(ns):106761000 dirty pages:3190240 colo_bitmap_clear_dirty time consuming(ns):8369000 dirty pages:3189293 colo_bitmap_clear_dirty time consuming(ns):8388000 dirty pages:3189171 colo_bitmap_clear_dirty time consuming(ns):8641000 dirty pages:3189099 colo_bitmap_clear_dirty time consuming(ns):8280000 dirty pages:3189974 colo_bitmap_clear_dirty time consuming(ns):8352000 dirty pages:3189471 colo_bitmap_clear_dirty time consuming(ns):8348000 dirty pages:3189681 colo_bitmap_clear_dirty time consuming(ns):8426000 it can be seen from the data that colo_bitmap_clear_dirty is more efficient. Signed-off-by: Lei Rao Reviewed-by: Lukas Straub Tested-by: Lukas Straub --- migration/ram.c | 36 +++++++++++++++++++++++++++++++----- 1 file changed, 31 insertions(+), 5 deletions(-) diff --git a/migration/ram.c b/migration/ram.c index 8661d82..11275cd 100644 --- a/migration/ram.c +++ b/migration/ram.c @@ -857,6 +857,36 @@ unsigned long colo_bitmap_find_dirty(RAMState *rs, RAMBlock *rb, return first; } +/** + * colo_bitmap_clear_dirty:when we flush ram cache to ram, we will use + * continuous memory copy, so we can also clean up the bitmap of contiguous + * dirty memory. + */ +static inline bool colo_bitmap_clear_dirty(RAMState *rs, + RAMBlock *rb, + unsigned long start, + unsigned long num) +{ + bool ret; + unsigned long i = 0; + + /* + * Since flush ram cache to ram can only happen on Secondary VM. + * and the clear bitmap always is NULL on destination side. + * Therefore, there is unnecessary to judge whether the + * clear_bitmap needs clear. + */ + QEMU_LOCK_GUARD(&rs->bitmap_mutex); + for (i = 0; i < num; i++) { + ret = test_and_clear_bit(start + i, rb->bmap); + if (ret) { + rs->migration_dirty_pages--; + } + } + + return ret; +} + static inline bool migration_bitmap_clear_dirty(RAMState *rs, RAMBlock *rb, unsigned long page) @@ -3774,11 +3804,7 @@ void colo_flush_ram_cache(void) num = 0; block = QLIST_NEXT_RCU(block, next); } else { - unsigned long i = 0; - - for (i = 0; i < num; i++) { - migration_bitmap_clear_dirty(ram_state, block, offset + i); - } + colo_bitmap_clear_dirty(ram_state, block, offset, num); dst_host = block->host + (((ram_addr_t)offset) << TARGET_PAGE_BITS); src_host = block->colo_cache -- 1.8.3.1