c/s 885 added very coarse TLB flushing. Since these flushes always
follow single page updates, single page flushes (when available) are
sufficient.
Signed-off-by: Jan Beulich <jbeulich@xxxxxxxxxx>
--- head-2009-05-29.orig/arch/i386/mm/hypervisor.c 2009-06-03
09:30:17.000000000 +0200
+++ head-2009-05-29/arch/i386/mm/hypervisor.c 2009-06-03 09:31:39.000000000
+0200
@@ -140,6 +140,7 @@ void xen_invlpg_all(unsigned long ptr)
op.arg1.linear_addr = ptr & PAGE_MASK;
BUG_ON(HYPERVISOR_mmuext_op(&op, 1, NULL, DOMID_SELF) < 0);
}
+EXPORT_SYMBOL_GPL(xen_invlpg_all);
void xen_invlpg_mask(cpumask_t *mask, unsigned long ptr)
{
@@ -151,6 +152,7 @@ void xen_invlpg_mask(const cpumask_t *ma
set_xen_guest_handle(op.arg2.vcpumask, mask->bits);
BUG_ON(HYPERVISOR_mmuext_op(&op, 1, NULL, DOMID_SELF) < 0);
}
+EXPORT_SYMBOL_GPL(xen_invlpg_mask);
#endif /* CONFIG_SMP */
--- head-2009-05-29.orig/drivers/xen/blktap2/device.c 2009-06-03
09:01:12.000000000 +0200
+++ head-2009-05-29/drivers/xen/blktap2/device.c 2009-06-03
09:40:38.000000000 +0200
@@ -194,6 +194,16 @@ blktap_umap_uaddr(struct vm_area_struct
PAGE_SIZE, blktap_umap_uaddr_fn, mm);
}
+static inline void
+flush_tlb_kernel_page(unsigned long kvaddr)
+{
+#ifdef CONFIG_X86
+ xen_invlpg_all(kvaddr);
+#else
+ flush_tlb_kernel_range(kvaddr, kvaddr + PAGE_SIZE);
+#endif
+}
+
static void
blktap_device_end_dequeued_request(struct blktap_device *dev,
struct request *req, int uptodate)
@@ -319,7 +329,7 @@ blktap_unmap(struct blktap *tap, struct
if (request->handles[i].kernel == INVALID_GRANT_HANDLE) {
kvaddr = request_to_kaddr(request, i);
blktap_umap_uaddr(&init_mm, kvaddr);
- flush_tlb_kernel_range(kvaddr, kvaddr + PAGE_SIZE);
+ flush_tlb_kernel_page(kvaddr);
set_phys_to_machine(__pa(kvaddr) >> PAGE_SHIFT,
INVALID_P2M_ENTRY);
}
@@ -550,9 +560,9 @@ blktap_map(struct blktap *tap,
pte = mk_pte(page, ring->vma->vm_page_prot);
blktap_map_uaddr(ring->vma->vm_mm, uvaddr, pte_mkwrite(pte));
- flush_tlb_mm(ring->vma->vm_mm);
+ flush_tlb_page(ring->vma, uvaddr);
blktap_map_uaddr(&init_mm, kvaddr, mk_pte(page, PAGE_KERNEL));
- flush_tlb_kernel_range(kvaddr, kvaddr + PAGE_SIZE);
+ flush_tlb_kernel_page(kvaddr);
set_phys_to_machine(__pa(kvaddr) >> PAGE_SHIFT, pte_mfn(pte));
request->handles[seg].kernel = INVALID_GRANT_HANDLE;
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|