WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-devel

[Xen-devel] RE: [PATCH] mem_sharing: fix race condition of nominate and

To: <juihaochiang@xxxxxxxxx>, xen devel <xen-devel@xxxxxxxxxxxxxxxxxxx>
Subject: [Xen-devel] RE: [PATCH] mem_sharing: fix race condition of nominate and unshare
From: MaoXiaoyun <tinnycloud@xxxxxxxxxxx>
Date: Tue, 18 Jan 2011 17:42:32 +0800
Cc: tim.deegan@xxxxxxxxxx
Delivery-date: Tue, 18 Jan 2011 01:43:29 -0800
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
Importance: Normal
In-reply-to: <AANLkTikj6gJ2we+3FcfmqdeSkrvBFyj=p6R47UcmS3Rk@xxxxxxxxxxxxxx>
List-help: <mailto:xen-devel-request@lists.xensource.com?subject=help>
List-id: Xen developer discussion <xen-devel.lists.xensource.com>
List-post: <mailto:xen-devel@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=unsubscribe>
References: <BLU157-ds1E01DEBE18840E5FBB6D1DA0E0@xxxxxxx>, <AANLkTikBaB5awvRu3Sn3WfuoS3LRmeBH=pG8c7H1n4Cw@xxxxxxxxxxxxxx>, <AANLkTinoXve=zBzB9qN1qXRz+iJmhiQ+-gB7MwFoY5Dg@xxxxxxxxxxxxxx>, <20110112105405.GH5651@xxxxxxxxxxxxxxxxxxxxxxx>, <BLU157-w59C63325262D0BE99E6C43DAF10@xxxxxxx>, <20110112140223.GI5651@xxxxxxxxxxxxxxxxxxxxxxx>, <BLU157-w1861EFE53CB51FC710011FDAF10@xxxxxxx>, <AANLkTimOz_uauDEnu_XaPEgwD1EZJWEgOO1oiFccFNs1@xxxxxxxxxxxxxx>, <20110113092427.GJ5651@xxxxxxxxxxxxxxxxxxxxxxx>, <AANLkTinSga8xDkuH0BsqbhbBtvgwgbn=T0qmg9y9CeGr@xxxxxxxxxxxxxx>, <20110113155344.GN5651@xxxxxxxxxxxxxxxxxxxxxxx>, <BLU157-w507CBBB94539BFDB92B339DAF30@xxxxxxx>, <AANLkTimC9OYTiHYaeeytH5OCN-EF2v6L=KDVMwBjtB0z@xxxxxxxxxxxxxx>, <BLU157-w995512A939ADE678A6401DAF40@xxxxxxx>, <AANLkTikj6gJ2we+3FcfmqdeSkrvBFyj=p6R47UcmS3Rk@xxxxxxxxxxxxxx>
Sender: xen-devel-bounces@xxxxxxxxxxxxxxxxxxx
Hi Tim & Jui-Hao:
 
      When I use Linux HVM instead of Windows HVM, more bug shows up.
 
      I only start on VM, and when I destroy it , xen crashed on mem_sharing_unshare_page()
which in line709, hash_entry is NULL. Later I found the handle has been removed in
mem_sharing_share_pages(), please refer logs below.
 
----mem_sharing_unshare_page
708     /* Remove the gfn_info from the list */
709     hash_entry = mem_sharing_hash_lookup(handle);
710     list_for_each(le, &hash_entry->gfns)
711     {
712         gfn_info = list_entry(le, struct gfn_info, list);
713         if((gfn_info->gfn == gfn) && (gfn_info->domain == d->domain_id))
714             goto gfn_found;
715     }
716     printk("Could not find gfn_info for shared gfn: %lx\n", gfn);
717     BUG();
 
 
----mem_sharing_share_page
 
649     printk("del %lu\n", ch);
650     list_for_each_safe(le, te, &ce->gfns)
651     {
652         gfn = list_entry(le, struct gfn_info, list);
653         /* Get the source page and type, this should never fail
654          * because we are under shr lock, and got non-null se */
655         BUG_ON(!get_page_and_type(spage, dom_cow, PGT_shared_page));
656         /* Move the gfn_info from ce list to se list */
657         list_del(&gfn->list);
658         d = get_domain_by_id(gfn->domain);
659         BUG_ON(!d);
660          BUG_ON(set_shared_p2m_entry(d, gfn->gfn, se->mfn) == 0);
661         put_domain(d);
662         list_add(&gfn->list, &se->gfns);
663         put_page_and_type(cpage);
664         mem_sharing_debug_gfn(d, gfn->gfn);
665     }
666     ASSERT(list_empty(&ce->gfns));
667     mem_sharing_hash_delete(ch);
668     atomic_inc(&nr_saved_mfns);
669     /* Free the client page */
670     if(test_and_clear_bit(_PGC_allocated, &cpage->count_info))
671         put_page(cpage);
672     mem_sharing_debug_gfn(d, gfn->gfn);      &nb sp;                                                                                                          
673     ret = 0;
 
 
-------log------------
     
(XEN) del 31261
(XEN) Debug for domain=1, gfn=75fd5, Debug page: MFN=179fd5 is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fd5, Debug page: MFN=179fd5 is ci=4, ti=8400000000000001, owner_id=32755
(XEN) del 31262
(XEN) Debug for domain=1, gfn=75fd6, Debug page: MFN=179fd6 is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fd6, Debug page: MFN=179fd6 is ci=4, ti=8400000000000001, owner_id=32755
(XEN) del 31263
(XEN) Debug for domain=1, gfn=75fd7, Debug page: MFN=179fd7 is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fd7, Debug page: MFN=179fd7 is ci=4, ti=8400000000000001, owner_id=32755
(XEN) del 31264
(XEN) Debug for domain=1, gfn=75fd8, Debug page: MFN=179fd8 is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fd8, Debug page: MFN=179fd8 is ci=4, ti=8400000000000001, owner_id=32755
(XEN) del 31265
(XEN) Debug for domain=1, gfn=75fd9, Debug page: MFN=179fd9 is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fd9, Debug page: MFN=179fd9 is ci=4, ti=8400000000000001, owner_id=32755
(XEN) del 31266
(XEN) Debug for domain=1, gfn=75fda, Debug page: MFN=179fda is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fda, Debug page: MFN=179fda is ci=4, ti=8400000000000001, owner_id=32755
(XEN) del 31267
(XEN) Debug for domain=1, gfn=75fdb, Debug page: MFN=179fdb is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fdb, Debug page: MFN=179fdb is ci=4, ti=8400000000000001, owner_id=32755
(XEN) del 31268
(XEN) Debug for domain=1, gfn=75fdc, Debug page: MFN=179fdc is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fdc, Debug page: MFN= 179fdc is ci=4, ti=8400000000000001, owner_id=32755
(XEN) del 31269
(XEN) Debug for domain=1, gfn=75fdd, Debug page: MFN=179fdd is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fdd, Debug page: MFN=179fdd is ci=4, ti=8400000000000001, owner_id=32755
(XEN) del 31270
(XEN) Debug for domain=1, gfn=75fde, Debug page: MFN=179fde is ci=8000000000000005, ti=8400000000000001, owner_id=32755
(XEN) Debug for domain=1, gfn=75fde, Debug page: MFN=179fde is ci=4, ti=8400000000000001, owner_id=32755
blktap_sysfs_destroy
(XEN) handle 31261
(XEN) Debug for domain=1, gfn=75fd5, Debug page: MFN=179fd5 is ci=1, ti=8400000000000001, owner_id=32755
(XEN) ----[ Xen-4.0.0  x86_64  debug=n  Not tainted ]----
(XEN) CPU:    1
(XEN) RIP:    e008:[<ffff82c4801bfeeb>] mem_sharing_unshare_page+0x1ab/0x740
(XEN) RFLAGS: 0000000000010246   CONTEXT: hypervisor
(XEN ) rax: 0000000000000000   rbx: 0000000000179fd5   rcx: 0000000000000082
(XEN) rdx: 000000000000000a   rsi: 000000000000000a   rdi: ffff82c48021e9c4
(XEN) rbp: ffff8302dd6a0000   rsp: ffff83023ff3fcd0   r8:  0000000000000001
(XEN) r9:  0000000000000000   r10: 00000000fffffff8   r11: 0000000000000005
(XEN) r12: 0000000000075fd5   r13: 0000000000000002   r14: 0000000000000000
(XEN) r15: ffff82f602f3faa0   cr0: 000000008005003b   cr4: 00000000000026f0
(XEN) cr3: 000000031b83b000   cr2: 0000000000000018
(XEN) ds: 002b   es: 002b   fs: 0000   gs: 0000   ss: e010   cs: e008
(XEN) Xen stack trace from rsp=ffff83023ff3fcd0:
(XEN)    ffff83023fe60000 ffff830173da4410 0000000100000000 0000000000000000
(XEN)    0000000000007a1d 00000000000000 01 0000000000000009 ffff8302dd6a0000
(XEN)    ffff83023ff3fd40 ffff82c4801df7e9 ffff83023ff3ff28 ffff83023ff3fd94
(XEN)    0000000000075fd5 0000000d000001d5 ffff83033e950000 0000000000075fd5
(XEN)    ffff83063fde8ef0 ffff8302dd6a0000 ffff83023ff3fd94 ffff82c48037a980
(XEN)    ffff82c480253080 ffff82c4801b8949 ffff8302dd6a0180 ffff82c48037a980
(XEN)    0000000d80253080 ffff8302dd6a0000 ffff8302dd6a0000 00000000ffffffff
(XEN)    ffff8302dd6a0000 ffff82c4801b681c 0000000000000000 ffff82c480149b74
(XEN)    ffff8300bf560000 ffff8300bf560000 fffffffffffffff8 00000000ffffffff
(XEN)    ffff8302dd6a0000 ffff82c4801061fc ffff8300bf552060 0000000000000000
(XEN)    ffff82c4802531a0 0000000000000001 ffff82c480376980 ffff82c48012218c
(XEN)    0000000000000001 fffffffffffffffd ffff83023ff3ff28 ffff82c48011c588
(XEN)    ffff83023ff3ff28 ffff83063fdeb170 ffff83063fdeb230 ffff8300bf552000
(XEN)    000001831ea27db3 ffff82c480189c6a 7fffffffffffffff ffff82c4801441b5
(XEN)    ffff82c48037b7b0 ffff82c48011e474 0000000000000001 ffffffffffffffff
(XEN)    0000000000000000 0000000000000000 0000000080376980 00001833000116f2
(XEN)    ffffffffffffffff ffff83023ff3ff28 ffff82c480251b00 ffff83023ff3fe28
(XEN)    ffff8300bf552000 000001831ea27db3 ffff82c480253080 ffff82c480149ad6
(XEN)    0000000000000000 0000000000002000 ffff8300bf2fc000 0000000000000000
(XEN)    0000000000000000 0000000000000000 0000000000000000 ffff88015f8f9f10
(XEN) Xen call trace:
(XEN)    [<ffff82c4801bfeeb>] mem_sharing_unshare_page+0x1ab/0x740
(XEN)    [<ffff82c4801df7e9>] ept_get_entry+0xa9/0x1c0
(XEN)    [<ffff82 c4801b8949>] p2m_teardown+0x129/0x170
(XEN)    [<ffff82c4801b681c>] paging_final_teardown+0x2c/0x40
(XEN)    [<ffff82c480149b74>] arch_domain_destroy+0x44/0x170
(XEN)    [<ffff82c4801061fc>] complete_domain_destroy+0x6c/0x130
(XEN)    [<ffff82c48012218c>] rcu_process_callbacks+0xac/0x220
(XEN)    [<ffff82c48011c588>] __do_softirq+0x58/0x80
(XEN)    [<ffff82c480189c6a>] acpi_processor_idle+0x14a/0x740
(XEN)    [<ffff82c4801441b5>] reprogram_timer+0x55/0x90
(XEN)    [<ffff82c48011e474>] timer_softirq_action+0x1a4/0x360
(XEN)    [<ffff82c480149ad6>] idle_loop+0x26/0x80
(XEN)   
(XEN) Pagetable walk from 0000000000000018:
(XEN)  L4[0x000] = 00000001676bd067 0000000000156103
(XEN)  L3[0x000] = 000000031b947067 0000000000121c8d(XEN)  L2[0x000] = 0000000000000000 ffffffffffffffff
(XEN)
(XEN) ****************************************
(XEN) Panic on CPU 1:
(XEN) FATAL PAGE FAULT
(XEN) [error_code=0000]
(XEN) Faulting linear address: 0000000000000018
(XEN) ****************************************
(XEN)
(XEN) Manual reset required ('noreboot' specified
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
<Prev in Thread] Current Thread [Next in Thread>