xen/arch/x86/mm/p2m.c | 38 ++++++++++++++++++++++++++++++++++++++
xen/include/asm-x86/p2m.h | 40 ++--------------------------------------
2 files changed, 40 insertions(+), 38 deletions(-)
Move the main query accessor to the p2m outside of an inline and into the
p2m code itself. This will allow for p2m internal locking to be added
to the accessor later.
Signed-off-by: Andres Lagar-Cavilla <andres@xxxxxxxxxxxxxxxx>
diff -r 0e8fb21ab4ff -r f07d4ebe5248 xen/arch/x86/mm/p2m.c
--- a/xen/arch/x86/mm/p2m.c
+++ b/xen/arch/x86/mm/p2m.c
@@ -144,6 +144,44 @@ void p2m_change_entry_type_global(struct
p2m_unlock(p2m);
}
+mfn_t gfn_to_mfn_type_p2m(struct p2m_domain *p2m, unsigned long gfn,
+ p2m_type_t *t, p2m_access_t *a, p2m_query_t q,
+ unsigned int *page_order)
+{
+ mfn_t mfn;
+
+ if ( !p2m || !paging_mode_translate(p2m->domain) )
+ {
+ /* Not necessarily true, but for non-translated guests, we claim
+ * it's the most generic kind of memory */
+ *t = p2m_ram_rw;
+ return _mfn(gfn);
+ }
+
+ mfn = p2m->get_entry(p2m, gfn, t, a, q, page_order);
+
+#ifdef __x86_64__
+ if ( q == p2m_unshare && p2m_is_shared(*t) )
+ {
+ ASSERT(!p2m_is_nestedp2m(p2m));
+ mem_sharing_unshare_page(p2m->domain, gfn, 0);
+ mfn = p2m->get_entry(p2m, gfn, t, a, q, page_order);
+ }
+#endif
+
+#ifdef __x86_64__
+ if (unlikely((p2m_is_broken(*t))))
+ {
+ /* Return invalid_mfn to avoid caller's access */
+ mfn = _mfn(INVALID_MFN);
+ if (q == p2m_guest)
+ domain_crash(p2m->domain);
+ }
+#endif
+
+ return mfn;
+}
+
int set_p2m_entry(struct p2m_domain *p2m, unsigned long gfn, mfn_t mfn,
unsigned int page_order, p2m_type_t p2mt, p2m_access_t p2ma)
{
diff -r 0e8fb21ab4ff -r f07d4ebe5248 xen/include/asm-x86/p2m.h
--- a/xen/include/asm-x86/p2m.h
+++ b/xen/include/asm-x86/p2m.h
@@ -313,45 +313,9 @@ struct p2m_domain *p2m_get_p2m(struct vc
* If the lookup succeeds, the return value is != INVALID_MFN and
* *page_order is filled in with the order of the superpage (if any) that
* the entry was found in. */
-static inline mfn_t
-gfn_to_mfn_type_p2m(struct p2m_domain *p2m, unsigned long gfn,
+mfn_t gfn_to_mfn_type_p2m(struct p2m_domain *p2m, unsigned long gfn,
p2m_type_t *t, p2m_access_t *a, p2m_query_t q,
- unsigned int *page_order)
-{
- mfn_t mfn;
-
- if ( !p2m || !paging_mode_translate(p2m->domain) )
- {
- /* Not necessarily true, but for non-translated guests, we claim
- * it's the most generic kind of memory */
- *t = p2m_ram_rw;
- return _mfn(gfn);
- }
-
- mfn = p2m->get_entry(p2m, gfn, t, a, q, page_order);
-
-#ifdef __x86_64__
- if ( q == p2m_unshare && p2m_is_shared(*t) )
- {
- ASSERT(!p2m_is_nestedp2m(p2m));
- mem_sharing_unshare_page(p2m->domain, gfn, 0);
- mfn = p2m->get_entry(p2m, gfn, t, a, q, page_order);
- }
-#endif
-
-#ifdef __x86_64__
- if (unlikely((p2m_is_broken(*t))))
- {
- /* Return invalid_mfn to avoid caller's access */
- mfn = _mfn(INVALID_MFN);
- if (q == p2m_guest)
- domain_crash(p2m->domain);
- }
-#endif
-
- return mfn;
-}
-
+ unsigned int *page_order);
/* General conversion function from gfn to mfn */
static inline mfn_t gfn_to_mfn_type(struct domain *d,
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|