|   | 
      | 
  
  
      | 
      | 
  
 
     | 
    | 
  
  
     | 
    | 
  
  
    |   | 
      | 
  
  
    | 
         
xen-devel
[Xen-devel] Re: Xen-devel Digest, Vol 68, Issue 57
 
Hi all,
  We are using xen4.0.1 with kernel 2.6.32.21, x86_64, running Windows 2003 server 64 bits at domU with gplpv drivers. 
  When starting several instances in sequence, we notice that processor load becomes very heavy; most of processor load goes to the qemu-dm processes. Even the processes for domU instances that were already running get loaded. 
Looking at xen load, only the first domU uses processing time; the other ones started in sequence do not use any process time: [root@localhost ~]# xl list Name                                        ID   Mem VCPUs    State    Time(s) 
Domain-0                                     0  3602     8        r--  19232.7 wk3-64-ent.1                                 1  4099     4        ---   1548.7 wk3-64-ent.2                                 2  4099     4        ---   2181.8 
wk3-64-ent.3                                 3  4099     4        ---   2362.0 wk3-64-ent.4                                 4  4099     4        r--   2407.3 wk3-64-ent.5                                 5  4099     4        ---   1662.9 
wk3-64-ent.6                                 6  4099     4        ---    960.3 wk3-64-ent.7                                 7  4099     1        r--      6.4 wk3-64-ent.8                                 8  4099     1        r--      5.6 
wk3-64-ent.9                                 9  4099     4        ---     10.9 wk3-64-ent.10                               10  4099     1        ---      4.1 wk3-64-ent.11                               11  4099     1        ---      3.4 
wk3-64-ent.12                               12  4099     1        ---      3.0 wk3-64-ent.13                               13  4099     1        ---      2.6 wk3-64-ent.14                               14  4099     1        ---      2.5 
wk3-64-ent.15                               15  4099     1        ---      2.3 wk3-64-ent.16                               16  4099     1        ---      2.5 wk3-64-ent.17                               17  4099     1        ---      2.2 
wk3-64-ent.18                               18  4099     1        ---      2.0 wk3-64-ent.19                               19  4099     1        ---      1.5 wk3-64-ent.20                               20  4099     1        ---      1.1 
 Looking at the console of any of the instances, I see the initial screen, waiting for qemu disks.
  Eventually, we get the exception below. After it, the server gets lost and needs to be restarted:
  OctOct  6 13:50:28 localhost kernel: [ 9667.958935] ------------[ cut here ]------------
 Oct  6 13:50:28 localhost kernel: [ 9667.958965] WARNING: at /root/xen4/xen-4.0-testing.hg/linux-2.6-pvops.git/net/sched/sch_generic.c:261 dev_watchdog+0x100/0x165
 ()
 Oct  6 13:50:28 localhost kernel: [ 9667.958996] Hardware name: PowerEdge M610
 Oct  6 13:50:28 localhost kernel: [ 9667.959011] NETDEV WATCHDOG: eth1 (bnx2): transmit queue 2 timed out
 Oct  6 13:50:28 localhost kernel: [ 9667.959013] Modules linked in: ebtable_nat tun bridge stp ebt_ip ebtable_filter ebtables bonding xt_state xt_multiport megarai
 d_sas
   6 13:50:28 localhost kernel: [ 9667.959102] Pid: 27902, comm: qemu-dm Not tainted 2.6.32.21 #2
 Oct  6 13:50:28 localhost kernel: [ 9667.959120] Call Trace:
 Oct  6 13:50:28 localhost kernel: [ 9667.959131]  <IRQ>  [<ffffffff81483960>] ? dev_watchdog+0x100/0x165
 Oct  6 13:50:28 localhost kernel: [ 9667.959162]  [<ffffffff81074dbf>] warn_slowpath_common+0x77/0x8f
 Oct  6 13:50:28 localhost kernel: [ 9667.959183]  [<ffffffff81074e87>] warn_slowpath_fmt+0x9f/0xa1
 Oct  6 13:50:28 localhost kernel: [ 9667.959204]  [<ffffffff810684c3>] ? __enqueue_entity+0x74/0x76
 Oct  6 13:50:28 localhost kernel: [ 9667.959225]  [<ffffffff81038731>] ? xen_force_evtchn_callback+0xd/0xf
 Oct  6 13:50:28 localhost kernel: [ 9667.959246]  [<ffffffff81038ea2>] ? check_events+0x12/0x20
 Oct  6 13:50:28 localhost kernel: [ 9667.959481]  [<ffffffff81064bce>] ? __raw_local_irq_save+0x12/0x18
 Oct  6 13:50:28 localhost kernel: [ 9667.959715]  [<ffffffff81482da6>] ? __netif_tx_lock+0x16/0x1f
 Oct  6 13:50:28 localhost kernel: [ 9667.959946]  [<ffffffff81482e27>] ? netif_tx_lock+0x41/0x69
 Oct  6 13:50:28 localhost kernel: [ 9667.960179]  [<ffffffff8146db0c>] ? netdev_drivername+0x43/0x4a
 Oct  6 13:50:28 localhost kernel: [ 9667.960413]  [<ffffffff81483960>] dev_watchdog+0x100/0x165
 Oct  6 13:50:28 localhost kernel: [ 9667.960649]  [<ffffffff8106c24b>] ? try_to_wake_up+0x2af/0x2c1
 Oct  6 13:50:28 localhost kernel: [ 9667.960885]  [<ffffffff8158f4d3>] ? _spin_unlock_irqrestore+0xe/0x10
 Oct  6 13:50:28 localhost kernel: [ 9667.961120]  [<ffffffff81080803>] ? process_timeout+0x0/0xb
 Oct  6 13:50:28 localhost kernel: [ 9667.961353]  [<ffffffff81483860>] ? dev_watchdog+0x0/0x165
 Oct  6 13:50:28 localhost kernel: [ 9667.961588]  [<ffffffff810804d0>] run_timer_softirq+0x172/0x20e
 Oct  6 13:50:28 localhost kernel: [ 9667.961824]  [<ffffffff8107aad1>] __do_softirq+0xcb/0x18b
 Oct  6 13:50:28 localhost kernel: [ 9667.962059]  [<ffffffff8103cdec>] call_softirq+0x1c/0x30
 Oct  6 13:50:28 localhost kernel: [ 9667.962290]  [<ffffffff8103e5e7>] do_softirq+0x4d/0x8e
 Oct  6 13:50:28 localhost kernel: [ 9667.962522]  [<ffffffff8107a964>] irq_exit+0x36/0x75
 Oct  6 13:50:28 localhost kernel: [ 9667.962755]  [<ffffffff812bbdd4>] xen_evtchn_do_upcall+0x33/0x43
 Oct  6 13:50:28 localhost kernel: [ 9667.962990]  [<ffffffff8103ce3e>] xen_do_hypervisor_callback+0x1e/0x30
 Oct  6 13:50:28 localhost kernel: [ 9667.963224]  <EOI>  [<ffffffff8158f521>] ? _spin_lock_irqsave+0xd/0x24
 Oct  6 13:50:28 localhost kernel: [ 9667.963467]  [<ffffffff8100922a>] ? hypercall_page+0x22a/0x1000
 Oct  6 13:50:28 localhost kernel: [ 9667.963701]  [<ffffffff8100922a>] ? hypercall_page+0x22a/0x1000
 Oct  6 13:50:28 localhost kernel: [ 9667.963935]  [<ffffffff81038731>] ? xen_force_evtchn_callback+0xd/0xf
 Oct  6 13:50:28 localhost kernel: [ 9667.964171]  [<ffffffff81038ea2>] ? check_events+0x12/0x20
 Oct  6 13:50:28 localhost kernel: [ 9667.964402]  [<ffffffff8158f521>] ? _spin_lock_irqsave+0xd/0x24
 Oct  6 13:50:28 localhost kernel: [ 9667.964636]  [<ffffffff81038e49>] ? xen_irq_enable_direct_end+0x0/0x7
 Oct  6 13:50:28 localhost kernel: [ 9667.964872]  [<ffffffff812c0e36>] ? __spin_unlock_irq+0x1b/0x1d
 Oct  6 13:50:28 localhost kernel: [ 9667.965107]  [<ffffffff812c1431>] ? evtchn_write+0xe9/0x103
 Oct  6 13:50:28 localhost kernel: [ 9667.965343]  [<ffffffff811107de>] ? vfs_write+0xab/0x105
 Oct  6 13:50:28 localhost kernel: [ 9667.965576]  [<ffffffff811108f2>] ? sys_write+0x47/0x6c
 Oct  6 13:50:28 localhost kernel: [ 9667.965808]  [<ffffffff8103bc82>] ? system_call_fastpath+0x16/0x1b
 Oct  6 13:50:28 localhost kernel: [ 9667.966044] ---[ end trace 5ecb364f57668210 ]---
  We would thank any help on this issue.
  Regards
  -- Roberto Scudeller 
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
 
 |   
 
| <Prev in Thread] | 
Current Thread | 
[Next in Thread> |  
- [Xen-devel] Re: Xen-devel Digest, Vol 68, Issue 57,
Roberto Scudeller <=
  
 |  
  
 | 
    | 
  
  
    |   | 
    |