WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-changelog

[Xen-changelog] [IA64] Changed from page to page_info

# HG changeset patch
# User awilliam@xxxxxxxxxxx
# Node ID d75f733f328c9b5df761c442f6450f904ab1b094
# Parent  9c0123c8a1b4de9414001876a262428ca61dc34c
[IA64] Changed from page to page_info

This patch changed from "struct page" to "struct page_info" in the below files.
 - xen/arch/ia64/xen/domain.c
 - xen/arch/ia64/xen/mm_init.c
 - xen/arch/ia64/xen/xenmem.c
 - xen/arch/ia64/xen/xenmisc.c
 - xen/include/asm-ia64/config.h
 - xen/include/asm-ia64/mm.h

This patch is "step1" which we showed by the following mail.
http://lists.xensource.com/archives/html/xen-ia64-devel/2006-03/msg00305.html

Signed-off-by: Akio Takebe <takebe_akio@xxxxxxxxxxxxxx>
Signed-off-by: Masaki Kanno <kanno.masaki@xxxxxxxxxxxxxx>

diff -r 9c0123c8a1b4 -r d75f733f328c xen/arch/ia64/xen/domain.c
--- a/xen/arch/ia64/xen/domain.c        Thu Mar 16 19:06:48 2006
+++ b/xen/arch/ia64/xen/domain.c        Thu Mar 16 19:10:22 2006
@@ -80,14 +80,14 @@
 /* this belongs in include/asm, but there doesn't seem to be a suitable place 
*/
 void arch_domain_destroy(struct domain *d)
 {
-       struct page *page;
+       struct page_info *page;
        struct list_head *ent, *prev;
 
        if (d->arch.mm->pgd != NULL)
        {
                list_for_each ( ent, &d->arch.mm->pt_list )
                {
-                       page = list_entry(ent, struct page, list);
+                       page = list_entry(ent, struct page_info, list);
                        prev = ent->prev;
                        list_del(ent);
                        free_xenheap_page(page_to_virt(page));
@@ -340,7 +340,7 @@
 static void relinquish_memory(struct domain *d, struct list_head *list)
 {
     struct list_head *ent;
-    struct page      *page;
+    struct page_info *page;
 #ifndef __ia64__
     unsigned long     x, y;
 #endif
@@ -350,7 +350,7 @@
     ent = list->next;
     while ( ent != list )
     {
-        page = list_entry(ent, struct page, list);
+        page = list_entry(ent, struct page_info, list);
         /* Grab a reference to the page so it won't disappear from under us. */
         if ( unlikely(!get_page(page, d)) )
         {
@@ -468,7 +468,7 @@
        }
 }
 
-static struct page * assign_new_domain0_page(unsigned long mpaddr)
+static struct page_info * assign_new_domain0_page(unsigned long mpaddr)
 {
        if (mpaddr < dom0_start || mpaddr >= dom0_start + dom0_size) {
                printk("assign_new_domain0_page: bad domain0 mpaddr 
0x%lx!\n",mpaddr);
@@ -480,10 +480,10 @@
 }
 
 /* allocate new page for domain and map it to the specified metaphysical addr 
*/
-struct page * assign_new_domain_page(struct domain *d, unsigned long mpaddr)
+struct page_info * assign_new_domain_page(struct domain *d, unsigned long 
mpaddr)
 {
        struct mm_struct *mm = d->arch.mm;
-       struct page *pt, *p = (struct page *)0;
+       struct page_info *pt, *p = (struct page_info *)0;
        pgd_t *pgd;
        pud_t *pud;
        pmd_t *pmd;
@@ -549,7 +549,7 @@
 void assign_domain_page(struct domain *d, unsigned long mpaddr, unsigned long 
physaddr)
 {
        struct mm_struct *mm = d->arch.mm;
-       struct page *pt;
+       struct page_info *pt;
        pgd_t *pgd;
        pud_t *pud;
        pmd_t *pmd;
@@ -746,7 +746,7 @@
        Elf_Phdr phdr;
        int h, filesz, memsz;
        unsigned long elfaddr, dom_mpaddr, dom_imva;
-       struct page *p;
+       struct page_info *p;
   
        copy_memory(&ehdr, (void *) image_start, sizeof(Elf_Ehdr));
        for ( h = 0; h < ehdr.e_phnum; h++ ) {
diff -r 9c0123c8a1b4 -r d75f733f328c xen/arch/ia64/xen/mm_init.c
--- a/xen/arch/ia64/xen/mm_init.c       Thu Mar 16 19:06:48 2006
+++ b/xen/arch/ia64/xen/mm_init.c       Thu Mar 16 19:10:22 2006
@@ -60,13 +60,13 @@
 #ifdef CONFIG_VIRTUAL_MEM_MAP
 unsigned long vmalloc_end = VMALLOC_END_INIT;
 EXPORT_SYMBOL(vmalloc_end);
-struct page *vmem_map;
+struct page_info *vmem_map;
 EXPORT_SYMBOL(vmem_map);
 #endif
 
 // static int pgt_cache_water[2] = { 25, 50 };
 
-struct page *zero_page_memmap_ptr;             /* map entry for zero page */
+struct page_info *zero_page_memmap_ptr;        /* map entry for zero page */
 EXPORT_SYMBOL(zero_page_memmap_ptr);
 
 #ifdef XEN
@@ -172,7 +172,7 @@
 pte_t fastcall * pte_alloc_map(struct mm_struct *mm, pmd_t *pmd, unsigned long 
address)
 {
        if (!pmd_present(*pmd)) {
-               struct page *new;
+               struct page_info *new;
 
                spin_unlock(&mm->page_table_lock);
                new = pte_alloc_one(mm, address);
@@ -202,7 +202,7 @@
 update_mmu_cache (struct vm_area_struct *vma, unsigned long vaddr, pte_t pte)
 {
        unsigned long addr;
-       struct page *page;
+       struct page_info *page;
 
        if (!pte_exec(pte))
                return;                         /* not an executable page... */
@@ -386,7 +386,7 @@
 create_mem_map_page_table (u64 start, u64 end, void *arg)
 {
        unsigned long address, start_page, end_page;
-       struct page *map_start, *map_end;
+       struct page_info *map_start, *map_end;
        int node;
        pgd_t *pgd;
        pmd_t *pmd;
@@ -417,8 +417,8 @@
 }
 
 struct memmap_init_callback_data {
-       struct page *start;
-       struct page *end;
+       struct page_info *start;
+       struct page_info *end;
        int nid;
        unsigned long zone;
 };
@@ -427,7 +427,7 @@
 virtual_memmap_init (u64 start, u64 end, void *arg)
 {
        struct memmap_init_callback_data *args;
-       struct page *map_start, *map_end;
+       struct page_info *map_start, *map_end;
 
        args = (struct memmap_init_callback_data *) arg;
 
@@ -440,13 +440,13 @@
                map_end = args->end;
 
        /*
-        * We have to initialize "out of bounds" struct page elements that fit 
completely
+        * We have to initialize "out of bounds" struct page_info elements that 
fit completely
         * on the same pages that were allocated for the "in bounds" elements 
because they
         * may be referenced later (and found to be "reserved").
         */
-       map_start -= ((unsigned long) map_start & (PAGE_SIZE - 1)) / 
sizeof(struct page);
+       map_start -= ((unsigned long) map_start & (PAGE_SIZE - 1)) / 
sizeof(struct page_info);
        map_end += ((PAGE_ALIGN((unsigned long) map_end) - (unsigned long) 
map_end)
-                   / sizeof(struct page));
+                   / sizeof(struct page_info));
 
        if (map_start < map_end)
                memmap_init_zone(map_start, (unsigned long) (map_end - 
map_start),
@@ -455,7 +455,7 @@
 }
 
 void
-memmap_init (struct page *start, unsigned long size, int nid,
+memmap_init (struct page_info *start, unsigned long size, int nid,
             unsigned long zone, unsigned long start_pfn)
 {
        if (!vmem_map)
@@ -476,7 +476,7 @@
 ia64_mfn_valid (unsigned long pfn)
 {
        char byte;
-       struct page *pg = mfn_to_page(pfn);
+       struct page_info *pg = mfn_to_page(pfn);
 
        return     (__get_user(byte, (char *) pg) == 0)
                && ((((u64)pg & PAGE_MASK) == (((u64)(pg + 1) - 1) & PAGE_MASK))
diff -r 9c0123c8a1b4 -r d75f733f328c xen/arch/ia64/xen/xenmem.c
--- a/xen/arch/ia64/xen/xenmem.c        Thu Mar 16 19:06:48 2006
+++ b/xen/arch/ia64/xen/xenmem.c        Thu Mar 16 19:10:22 2006
@@ -13,12 +13,12 @@
 #include <asm/pgtable.h>
 #include <xen/mm.h>
 
-extern struct page *zero_page_memmap_ptr;
+extern struct page_info *zero_page_memmap_ptr;
 struct page_info *frame_table;
 unsigned long frame_table_size;
 unsigned long max_page;
 
-struct page *mem_map;
+struct page_info *mem_map;
 #define MAX_DMA_ADDRESS ~0UL   // FIXME???
 
 #ifdef CONFIG_VIRTUAL_MEM_MAP
diff -r 9c0123c8a1b4 -r d75f733f328c xen/arch/ia64/xen/xenmisc.c
--- a/xen/arch/ia64/xen/xenmisc.c       Thu Mar 16 19:06:48 2006
+++ b/xen/arch/ia64/xen/xenmisc.c       Thu Mar 16 19:10:22 2006
@@ -171,7 +171,7 @@
        return (unsigned long)p;
 }
 
-void __free_pages(struct page *page, unsigned int order)
+void __free_pages(struct page_info *page, unsigned int order)
 {
        if (order) BUG();
        free_xenheap_page(page);
diff -r 9c0123c8a1b4 -r d75f733f328c xen/include/asm-ia64/config.h
--- a/xen/include/asm-ia64/config.h     Thu Mar 16 19:06:48 2006
+++ b/xen/include/asm-ia64/config.h     Thu Mar 16 19:10:22 2006
@@ -73,8 +73,11 @@
 extern unsigned long dom0_start;
 extern unsigned long dom0_size;
 
+// see include/asm-ia64/mm.h, handle remaining page_info uses until gone
+#define page_info page
+
 // from linux/include/linux/mm.h
-extern struct page *mem_map;
+extern struct page_info *mem_map;
 
 // xen/include/asm/config.h
 extern char _end[]; /* standard ELF symbol */
@@ -134,9 +137,6 @@
 #define smp_num_siblings 1
 #endif
 
-// from linux/include/linux/mm.h
-struct page;
-
 // function calls; see decl in xen/include/xen/sched.h
 #undef free_task_struct
 #undef alloc_task_struct
@@ -206,8 +206,6 @@
 #define _atomic_read(v) ((v).counter)
 #define atomic_compareandswap(old, new, v) ((atomic_t){ cmpxchg(v, 
_atomic_read(old), _atomic_read(new)) })
 
-// see include/asm-ia64/mm.h, handle remaining page_info uses until gone
-#define page_info page
 // Deprivated linux inf and put here for short time compatibility
 #define kmalloc(s, t) xmalloc_bytes((s))
 #define kfree(s) xfree((s))
diff -r 9c0123c8a1b4 -r d75f733f328c xen/include/asm-ia64/mm.h
--- a/xen/include/asm-ia64/mm.h Thu Mar 16 19:06:48 2006
+++ b/xen/include/asm-ia64/mm.h Thu Mar 16 19:10:22 2006
@@ -36,7 +36,7 @@
 
 #define PRtype_info "08x"
 
-struct page
+struct page_info
 {
     /* Each frame can be threaded onto a doubly-linked list. */
     struct list_head list;
@@ -228,7 +228,7 @@
 
 // prototype of misc memory stuff
 //unsigned long __get_free_pages(unsigned int mask, unsigned int order);
-//void __free_pages(struct page *page, unsigned int order);
+//void __free_pages(struct page_info *page, unsigned int order);
 void *pgtable_quicklist_alloc(void);
 void pgtable_quicklist_free(void *pgtable_entry);
 
@@ -348,11 +348,11 @@
 #define NODEZONE_SHIFT (sizeof(page_flags_t)*8 - MAX_NODES_SHIFT - 
MAX_ZONES_SHIFT)
 #define NODEZONE(node, zone)   ((node << ZONES_SHIFT) | zone)
 
-static inline unsigned long page_zonenum(struct page *page)
+static inline unsigned long page_zonenum(struct page_info *page)
 {
        return (page->flags >> NODEZONE_SHIFT) & (~(~0UL << ZONES_SHIFT));
 }
-static inline unsigned long page_to_nid(struct page *page)
+static inline unsigned long page_to_nid(struct page_info *page)
 {
        return (page->flags >> (NODEZONE_SHIFT + ZONES_SHIFT));
 }
@@ -360,12 +360,12 @@
 struct zone;
 extern struct zone *zone_table[];
 
-static inline struct zone *page_zone(struct page *page)
+static inline struct zone *page_zone(struct page_info *page)
 {
        return zone_table[page->flags >> NODEZONE_SHIFT];
 }
 
-static inline void set_page_zone(struct page *page, unsigned long nodezone_num)
+static inline void set_page_zone(struct page_info *page, unsigned long 
nodezone_num)
 {
        page->flags &= ~(~0UL << NODEZONE_SHIFT);
        page->flags |= nodezone_num << NODEZONE_SHIFT;
@@ -376,7 +376,7 @@
 extern unsigned long max_mapnr;
 #endif
 
-static inline void *lowmem_page_address(struct page *page)
+static inline void *lowmem_page_address(struct page_info *page)
 {
        return __va(page_to_mfn(page) << PAGE_SHIFT);
 }
@@ -395,8 +395,8 @@
 #endif
 
 #if defined(HASHED_PAGE_VIRTUAL)
-void *page_address(struct page *page);
-void set_page_address(struct page *page, void *virtual);
+void *page_address(struct page_info *page);
+void set_page_address(struct page_info *page, void *virtual);
 void page_address_init(void);
 #endif
 
@@ -409,7 +409,7 @@
 
 #ifndef CONFIG_DEBUG_PAGEALLOC
 static inline void
-kernel_map_pages(struct page *page, int numpages, int enable)
+kernel_map_pages(struct page_info *page, int numpages, int enable)
 {
 }
 #endif

_______________________________________________
Xen-changelog mailing list
Xen-changelog@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-changelog

<Prev in Thread] Current Thread [Next in Thread>
  • [Xen-changelog] [IA64] Changed from page to page_info, Xen patchbot -unstable <=