From patchwork Thu Apr 22 06:14:05 2010 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Xiao Guangrong X-Patchwork-Id: 94029 Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by demeter.kernel.org (8.14.3/8.14.3) with ESMTP id o3M6JZIY019186 for ; Thu, 22 Apr 2010 06:19:38 GMT Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752401Ab0DVGRM (ORCPT ); Thu, 22 Apr 2010 02:17:12 -0400 Received: from cn.fujitsu.com ([222.73.24.84]:50344 "EHLO song.cn.fujitsu.com" rhost-flags-OK-FAIL-OK-OK) by vger.kernel.org with ESMTP id S1752704Ab0DVGRJ (ORCPT ); Thu, 22 Apr 2010 02:17:09 -0400 Received: from tang.cn.fujitsu.com (tang.cn.fujitsu.com [10.167.250.3]) by song.cn.fujitsu.com (Postfix) with ESMTP id B31CA170126; Thu, 22 Apr 2010 14:17:08 +0800 (CST) Received: from fnst.cn.fujitsu.com (tang.cn.fujitsu.com [127.0.0.1]) by tang.cn.fujitsu.com (8.14.3/8.13.1) with ESMTP id o3M6FRc4022247; Thu, 22 Apr 2010 14:15:28 +0800 Received: from [10.167.141.99] (unknown [10.167.141.99]) by fnst.cn.fujitsu.com (Postfix) with ESMTPA id 33C1EDC2D2; Thu, 22 Apr 2010 14:20:09 +0800 (CST) Message-ID: <4BCFE92D.9000201@cn.fujitsu.com> Date: Thu, 22 Apr 2010 14:14:05 +0800 From: Xiao Guangrong User-Agent: Thunderbird 2.0.0.24 (Windows/20100228) MIME-Version: 1.0 To: Avi Kivity CC: Marcelo Tosatti , KVM list , LKML Subject: [PATCH 10/10] KVM MMU: optimize sync/update unsync-page References: <4BCFE581.8050305@cn.fujitsu.com> In-Reply-To: <4BCFE581.8050305@cn.fujitsu.com> Sender: kvm-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org X-Greylist: IP, sender and recipient auto-whitelisted, not delayed by milter-greylist-4.2.3 (demeter.kernel.org [140.211.167.41]); Thu, 22 Apr 2010 06:19:38 +0000 (UTC) diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c index f092e71..5bdcc17 100644 --- a/arch/x86/kvm/mmu.c +++ b/arch/x86/kvm/mmu.c @@ -2299,10 +2299,11 @@ static void paging_invlpg(struct kvm_vcpu *vcpu, gva_t gva) for_each_shadow_entry(vcpu, gva, iterator) { level = iterator.level; sptep = iterator.sptep; + sp = page_header(__pa(sptep)); if (is_last_spte(*sptep, level)) { - - sp = page_header(__pa(sptep)); + if (!sp->unsync) + break; if (!sp->role.cr4_pae) offset = sp->role.quadrant << PT64_LEVEL_BITS;; @@ -2320,7 +2321,7 @@ static void paging_invlpg(struct kvm_vcpu *vcpu, gva_t gva) break; } - if (!is_shadow_present_pte(*sptep)) + if (!is_shadow_present_pte(*sptep) || !sp->unsync_children) break; } @@ -2744,7 +2745,8 @@ void kvm_mmu_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa, restart: hlist_for_each_entry_safe(sp, node, n, bucket, hash_link) { - if (sp->gfn != gfn || sp->role.direct || sp->role.invalid) + if (sp->gfn != gfn || sp->role.direct || sp->role.invalid || + sp->unsync) continue; pte_size = sp->role.cr4_pae ? 8 : 4; misaligned = (offset ^ (offset + bytes - 1)) & ~(pte_size - 1);