summaryrefslogtreecommitdiffstats
path: root/virt
diff options
context:
space:
mode:
authorWill Deacon <will@kernel.org>2020-08-11 11:27:24 +0100
committerPaolo Bonzini <pbonzini@redhat.com>2020-08-21 18:03:47 -0400
commitfdfe7cbd58806522e799e2a50a15aee7f2cbb7b6 (patch)
tree9b6f76e58e9945c50989d86f0037073fb8fa10df /virt
parentcb957adb4ea422bd758568df5b2478ea3bb34f35 (diff)
downloadlinux-fdfe7cbd58806522e799e2a50a15aee7f2cbb7b6.tar.gz
linux-fdfe7cbd58806522e799e2a50a15aee7f2cbb7b6.tar.bz2
linux-fdfe7cbd58806522e799e2a50a15aee7f2cbb7b6.zip
KVM: Pass MMU notifier range flags to kvm_unmap_hva_range()
The 'flags' field of 'struct mmu_notifier_range' is used to indicate whether invalidate_range_{start,end}() are permitted to block. In the case of kvm_mmu_notifier_invalidate_range_start(), this field is not forwarded on to the architecture-specific implementation of kvm_unmap_hva_range() and therefore the backend cannot sensibly decide whether or not to block. Add an extra 'flags' parameter to kvm_unmap_hva_range() so that architectures are aware as to whether or not they are permitted to block. Cc: <stable@vger.kernel.org> Cc: Marc Zyngier <maz@kernel.org> Cc: Suzuki K Poulose <suzuki.poulose@arm.com> Cc: James Morse <james.morse@arm.com> Signed-off-by: Will Deacon <will@kernel.org> Message-Id: <20200811102725.7121-2-will@kernel.org> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
Diffstat (limited to 'virt')
-rw-r--r--virt/kvm/kvm_main.c3
1 files changed, 2 insertions, 1 deletions
diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
index 2c2c0254c2d8..4eaa4e46c7d0 100644
--- a/virt/kvm/kvm_main.c
+++ b/virt/kvm/kvm_main.c
@@ -482,7 +482,8 @@ static int kvm_mmu_notifier_invalidate_range_start(struct mmu_notifier *mn,
* count is also read inside the mmu_lock critical section.
*/
kvm->mmu_notifier_count++;
- need_tlb_flush = kvm_unmap_hva_range(kvm, range->start, range->end);
+ need_tlb_flush = kvm_unmap_hva_range(kvm, range->start, range->end,
+ range->flags);
need_tlb_flush |= kvm->tlbs_dirty;
/* we've to flush the tlb before the pages can be freed */
if (need_tlb_flush)