From patchwork Mon Jan 9 06:24:34 2017 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Jintack Lim X-Patchwork-Id: 90355 Delivered-To: patch@linaro.org Received: by 10.140.20.101 with SMTP id 92csp708096qgi; Sun, 8 Jan 2017 22:31:43 -0800 (PST) X-Received: by 10.99.117.79 with SMTP id f15mr19082343pgn.69.1483943503721; Sun, 08 Jan 2017 22:31:43 -0800 (PST) Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id c82si8531491pfc.179.2017.01.08.22.31.43; Sun, 08 Jan 2017 22:31:43 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1163650AbdAIGbX (ORCPT + 25 others); Mon, 9 Jan 2017 01:31:23 -0500 Received: from outprodmail01.cc.columbia.edu ([128.59.72.39]:38506 "EHLO outprodmail01.cc.columbia.edu" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S939760AbdAIG0e (ORCPT ); Mon, 9 Jan 2017 01:26:34 -0500 Received: from hazelnut (hazelnut.cc.columbia.edu [128.59.213.250]) by outprodmail01.cc.columbia.edu (8.14.4/8.14.4) with ESMTP id v096P4f2017232 for ; Mon, 9 Jan 2017 01:26:21 -0500 Received: from hazelnut (localhost.localdomain [127.0.0.1]) by hazelnut (Postfix) with ESMTP id 29BA87E for ; Mon, 9 Jan 2017 01:26:21 -0500 (EST) Received: from sendprodmail03.cc.columbia.edu (sendprodmail03.cc.columbia.edu [128.59.72.15]) by hazelnut (Postfix) with ESMTP id 0FCF76D for ; Mon, 9 Jan 2017 01:26:21 -0500 (EST) Received: from mail-qk0-f197.google.com (mail-qk0-f197.google.com [209.85.220.197]) by sendprodmail03.cc.columbia.edu (8.14.4/8.14.4) with ESMTP id v096QKZ1057605 (version=TLSv1/SSLv3 cipher=AES128-GCM-SHA256 bits=128 verify=NOT) for ; Mon, 9 Jan 2017 01:26:20 -0500 Received: by mail-qk0-f197.google.com with SMTP id a195so63264270qkg.3 for ; Sun, 08 Jan 2017 22:26:20 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=Z53saGPK+A3MPULcfvwNl5jeGvvgUqLH/qz4VMpaqkk=; b=gR1vWoWgMA6TA0eLevhgdW6wss3uIi+zeH4e7n0PS/fnwwsLUMriqqioWzPhVnna4R /hZ1Utx0KTvRZtNrGNbjrkK7qbXfCRT/REPnQ7qrnPcJGBdNnn200hl/twp6U4v8DjpX ddwlwPcAfoCqa1skTrKtQhwuh9kEQB5JuGtLgfV5KeuAcSsNvrF05SOd9eCM8LkNq2qv lURPlUG22vvlEzRPq0vJeC/UcsadpxiHzL9tP1b3gfS85Uy4lTPkg8mBcDROvgNoR9b2 b5DmN7dB3O2xzr29A5VrX30LijVaE7rBmiaOneG7Vk3eL57NsLn2OS+pfeNBv3ymgGOa 1GBA== X-Gm-Message-State: AIkVDXJ9P+3lxrjPs/HMkhCuixANzgOpjTLYfLRW05hCO+fReP6bg3C/uu5ze/C/ewSUpHQI4UVLaMhc9DDzrkQNNPilT67LkC8Hq8QYEKoqB5fB0FAiG0NuHbq0wFoSpGqUU1xockJfXpa71elq654UHTM= X-Received: by 10.55.136.199 with SMTP id k190mr70784430qkd.108.1483943180151; Sun, 08 Jan 2017 22:26:20 -0800 (PST) X-Received: by 10.55.136.199 with SMTP id k190mr70784406qkd.108.1483943179921; Sun, 08 Jan 2017 22:26:19 -0800 (PST) Received: from jintack.cs.columbia.edu ([2001:18d8:ffff:16:21a:4aff:feaa:f900]) by smtp.gmail.com with ESMTPSA id h3sm8623257qtc.6.2017.01.08.22.26.18 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Sun, 08 Jan 2017 22:26:19 -0800 (PST) From: Jintack Lim To: christoffer.dall@linaro.org, marc.zyngier@arm.com, pbonzini@redhat.com, rkrcmar@redhat.com, linux@armlinux.org.uk, catalin.marinas@arm.com, will.deacon@arm.com, vladimir.murzin@arm.com, suzuki.poulose@arm.com, mark.rutland@arm.com, james.morse@arm.com, lorenzo.pieralisi@arm.com, kevin.brodsky@arm.com, wcohen@redhat.com, shankerd@codeaurora.org, geoff@infradead.org, andre.przywara@arm.com, eric.auger@redhat.com, anna-maria@linutronix.de, shihwei@cs.columbia.edu, linux-arm-kernel@lists.infradead.org, kvmarm@lists.cs.columbia.edu, kvm@vger.kernel.org, linux-kernel@vger.kernel.org Cc: jintack@cs.columbia.edu Subject: [RFC 38/55] KVM: arm/arm64: Make mmu functions non-static Date: Mon, 9 Jan 2017 01:24:34 -0500 Message-Id: <1483943091-1364-39-git-send-email-jintack@cs.columbia.edu> X-Mailer: git-send-email 1.9.1 In-Reply-To: <1483943091-1364-1-git-send-email-jintack@cs.columbia.edu> References: <1483943091-1364-1-git-send-email-jintack@cs.columbia.edu> X-No-Spam-Score: Local X-Scanned-By: MIMEDefang 2.78 on 128.59.72.15 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Christoffer Dall Make mmu functions non-static so that we can reuse those functions to support mmu for the nested VMs. Signed-off-by: Christoffer Dall Signed-off-by: Jintack Lim --- arch/arm/kvm/mmu.c | 90 +++++++++++++++++++++++----------------- arch/arm64/include/asm/kvm_mmu.h | 9 ++++ 2 files changed, 61 insertions(+), 38 deletions(-) -- 1.9.1 diff --git a/arch/arm/kvm/mmu.c b/arch/arm/kvm/mmu.c index 56358fa..98b42e8 100644 --- a/arch/arm/kvm/mmu.c +++ b/arch/arm/kvm/mmu.c @@ -301,7 +301,7 @@ static void unmap_stage2_puds(struct kvm_s2_mmu *mmu, pgd_t *pgd, } /** - * unmap_stage2_range -- Clear stage2 page table entries to unmap a range + * kvm_unmap_stage2_range -- Clear stage2 page table entries to unmap a range * @kvm: The VM pointer * @start: The intermediate physical base address of the range to unmap * @size: The size of the area to unmap @@ -311,8 +311,7 @@ static void unmap_stage2_puds(struct kvm_s2_mmu *mmu, pgd_t *pgd, * destroying the VM), otherwise another faulting VCPU may come in and mess * with things behind our backs. */ -static void unmap_stage2_range(struct kvm_s2_mmu *mmu, - phys_addr_t start, u64 size) +void kvm_unmap_stage2_range(struct kvm_s2_mmu *mmu, phys_addr_t start, u64 size) { pgd_t *pgd; phys_addr_t addr = start, end = start + size; @@ -374,11 +373,10 @@ static void stage2_flush_puds(pgd_t *pgd, } while (pud++, addr = next, addr != end); } -static void stage2_flush_memslot(struct kvm_s2_mmu *mmu, - struct kvm_memory_slot *memslot) +void kvm_stage2_flush_range(struct kvm_s2_mmu *mmu, + phys_addr_t start, phys_addr_t end) { - phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; - phys_addr_t end = addr + PAGE_SIZE * memslot->npages; + phys_addr_t addr = start; phys_addr_t next; pgd_t *pgd; @@ -389,6 +387,15 @@ static void stage2_flush_memslot(struct kvm_s2_mmu *mmu, } while (pgd++, addr = next, addr != end); } +static void stage2_flush_memslot(struct kvm_s2_mmu *mmu, + struct kvm_memory_slot *memslot) +{ + phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; + phys_addr_t end = start + PAGE_SIZE * memslot->npages; + + kvm_stage2_flush_range(mmu, start, end); +} + /** * stage2_flush_vm - Invalidate cache for pages mapped in stage 2 * @kvm: The struct kvm pointer @@ -745,21 +752,9 @@ int create_hyp_io_mappings(void *from, void *to, phys_addr_t phys_addr) __phys_to_pfn(phys_addr), PAGE_HYP_DEVICE); } -/** - * kvm_alloc_stage2_pgd - allocate level-1 table for stage-2 translation. - * @kvm: The KVM struct pointer for the VM. - * - * Allocates only the stage-2 HW PGD level table(s) (can support either full - * 40-bit input addresses or limited to 32-bit input addresses). Clears the - * allocated pages. - * - * Note we don't need locking here as this is only called when the VM is - * created, which can only be done once. - */ -int kvm_alloc_stage2_pgd(struct kvm *kvm) +int __kvm_alloc_stage2_pgd(struct kvm_s2_mmu *mmu) { pgd_t *pgd; - struct kvm_s2_mmu *mmu = &kvm->arch.mmu; if (mmu->pgd != NULL) { kvm_err("kvm_arch already initialized?\n"); @@ -776,6 +771,22 @@ int kvm_alloc_stage2_pgd(struct kvm *kvm) return 0; } +/** + * kvm_alloc_stage2_pgd - allocate level-1 table for stage-2 translation. + * @kvm: The KVM struct pointer for the VM. + * + * Allocates only the stage-2 HW PGD level table(s) (can support either full + * 40-bit input addresses or limited to 32-bit input addresses). Clears the + * allocated pages. + * + * Note we don't need locking here as this is only called when the VM is + * created, which can only be done once. + */ +int kvm_alloc_stage2_pgd(struct kvm *kvm) +{ + return __kvm_alloc_stage2_pgd(&kvm->arch.mmu); +} + static void stage2_unmap_memslot(struct kvm_s2_mmu *mmu, struct kvm_memory_slot *memslot) { @@ -811,7 +822,7 @@ static void stage2_unmap_memslot(struct kvm_s2_mmu *mmu, if (!(vma->vm_flags & VM_PFNMAP)) { gpa_t gpa = addr + (vm_start - memslot->userspace_addr); - unmap_stage2_range(mmu, gpa, vm_end - vm_start); + kvm_unmap_stage2_range(mmu, gpa, vm_end - vm_start); } hva = vm_end; } while (hva < reg_end); @@ -841,6 +852,17 @@ void stage2_unmap_vm(struct kvm *kvm) srcu_read_unlock(&kvm->srcu, idx); } +void __kvm_free_stage2_pgd(struct kvm_s2_mmu *mmu) +{ + if (mmu->pgd == NULL) + return; + + kvm_unmap_stage2_range(mmu, 0, KVM_PHYS_SIZE); + /* Free the HW pgd, one page at a time */ + free_pages_exact(mmu->pgd, S2_PGD_SIZE); + mmu->pgd = NULL; +} + /** * kvm_free_stage2_pgd - free all stage-2 tables * @kvm: The KVM struct pointer for the VM. @@ -854,15 +876,7 @@ void stage2_unmap_vm(struct kvm *kvm) */ void kvm_free_stage2_pgd(struct kvm *kvm) { - struct kvm_s2_mmu *mmu = &kvm->arch.mmu; - - if (mmu->pgd == NULL) - return; - - unmap_stage2_range(mmu, 0, KVM_PHYS_SIZE); - /* Free the HW pgd, one page at a time */ - free_pages_exact(mmu->pgd, S2_PGD_SIZE); - mmu->pgd = NULL; + __kvm_free_stage2_pgd(&kvm->arch.mmu); } static pud_t *stage2_get_pud(struct kvm_s2_mmu *mmu, @@ -1175,13 +1189,13 @@ static void stage2_wp_puds(pgd_t *pgd, phys_addr_t addr, phys_addr_t end) } /** - * stage2_wp_range() - write protect stage2 memory region range + * kvm_stage2_wp_range() - write protect stage2 memory region range * @kvm: The KVM pointer * @addr: Start address of range * @end: End address of range */ -static void stage2_wp_range(struct kvm *kvm, struct kvm_s2_mmu *mmu, - phys_addr_t addr, phys_addr_t end) +void kvm_stage2_wp_range(struct kvm *kvm, struct kvm_s2_mmu *mmu, + phys_addr_t addr, phys_addr_t end) { pgd_t *pgd; phys_addr_t next; @@ -1225,7 +1239,7 @@ void kvm_mmu_wp_memory_region(struct kvm *kvm, int slot) phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; spin_lock(&kvm->mmu_lock); - stage2_wp_range(kvm, &kvm->arch.mmu, start, end); + kvm_stage2_wp_range(kvm, &kvm->arch.mmu, start, end); spin_unlock(&kvm->mmu_lock); kvm_flush_remote_tlbs(kvm); } @@ -1249,7 +1263,7 @@ static void kvm_mmu_write_protect_pt_masked(struct kvm *kvm, phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; - stage2_wp_range(kvm, &kvm->arch.mmu, start, end); + kvm_stage2_wp_range(kvm, &kvm->arch.mmu, start, end); } /* @@ -1589,7 +1603,7 @@ static int handle_hva_to_gpa(struct kvm *kvm, static int kvm_unmap_hva_handler(struct kvm *kvm, gpa_t gpa, void *data) { - unmap_stage2_range(&kvm->arch.mmu, gpa, PAGE_SIZE); + kvm_unmap_stage2_range(&kvm->arch.mmu, gpa, PAGE_SIZE); return 0; } @@ -1900,7 +1914,7 @@ int kvm_arch_prepare_memory_region(struct kvm *kvm, spin_lock(&kvm->mmu_lock); if (ret) - unmap_stage2_range(&kvm->arch.mmu, mem->guest_phys_addr, + kvm_unmap_stage2_range(&kvm->arch.mmu, mem->guest_phys_addr, mem->memory_size); else stage2_flush_memslot(&kvm->arch.mmu, memslot); @@ -1944,7 +1958,7 @@ void kvm_arch_flush_shadow_memslot(struct kvm *kvm, phys_addr_t size = slot->npages << PAGE_SHIFT; spin_lock(&kvm->mmu_lock); - unmap_stage2_range(&kvm->arch.mmu, gpa, size); + kvm_unmap_stage2_range(&kvm->arch.mmu, gpa, size); spin_unlock(&kvm->mmu_lock); } diff --git a/arch/arm64/include/asm/kvm_mmu.h b/arch/arm64/include/asm/kvm_mmu.h index e3455c4..a504162 100644 --- a/arch/arm64/include/asm/kvm_mmu.h +++ b/arch/arm64/include/asm/kvm_mmu.h @@ -145,9 +145,18 @@ static inline unsigned long __kern_hyp_va(unsigned long v) void stage2_unmap_vm(struct kvm *kvm); int kvm_alloc_stage2_pgd(struct kvm *kvm); +int __kvm_alloc_stage2_pgd(struct kvm_s2_mmu *mmu); void kvm_free_stage2_pgd(struct kvm *kvm); +void __kvm_free_stage2_pgd(struct kvm_s2_mmu *mmu); int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa, phys_addr_t pa, unsigned long size, bool writable); +void kvm_unmap_stage2_range(struct kvm_s2_mmu *mmu, phys_addr_t start, + u64 size); +void kvm_stage2_wp_range(struct kvm *kvm, struct kvm_s2_mmu *mmu, + phys_addr_t addr, phys_addr_t end); +void kvm_stage2_flush_range(struct kvm_s2_mmu *mmu, + phys_addr_t start, phys_addr_t end); + int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run);