WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-devel

Re: [Xen-devel] xen-unstable pvops 2.6.32.21 kernel/lockdep.c:2323 trace

To: Jeremy Fitzhardinge <jeremy@xxxxxxxx>
Subject: Re: [Xen-devel] xen-unstable pvops 2.6.32.21 kernel/lockdep.c:2323 trace_hardirqs_on_caller+0x12f/0x190()
From: Bruce Edge <bruce.edge@xxxxxxxxx>
Date: Thu, 16 Sep 2010 16:06:59 -0700
Cc: xen-devel@xxxxxxxxxxxxxxxxxxx
Delivery-date: Thu, 16 Sep 2010 16:07:49 -0700
Dkim-signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:mime-version:received:received:in-reply-to :references:date:message-id:subject:from:to:cc:content-type :content-transfer-encoding; bh=LpcdyD27ULjuTTdAatCC5GBrlFJu3rAVWXJ+h4lUSQU=; b=EIt5R2mD0HUxjoYA3MLdRvzXLMR58LrQ2jCQwG0NgFR7LwqLiO31MZz8nfX4Fgc9bY M4nP9ve9e/tas48iCCVy3N8Z2V+j3ngJbi+fyYtRsnSkujrT2QlqRj0Ba4wonbnfTZen tW3s790jQ8NbDr1tOKtXd3silnALr+fpSt9Xo=
Domainkey-signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc:content-type:content-transfer-encoding; b=Hmq3FZQJB4iOQcHxB5FL+xMnESmVPbpKqJjGtn3zWvhAlxMLFMfxNPHRx9QbCTLgbs cL8TKk0NA6dm2skUVK8RwJWxcEeG/dukBYrY6QB0mxuJp9V+vdrsO8tve1Ezxp3cLbwW fof1+Geyuogbe+hGqhYB9XOqz4Gk6KXcyEtxY=
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
In-reply-to: <4C9299D6.80700@xxxxxxxx>
List-help: <mailto:xen-devel-request@lists.xensource.com?subject=help>
List-id: Xen developer discussion <xen-devel.lists.xensource.com>
List-post: <mailto:xen-devel@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=unsubscribe>
References: <AANLkTikRHHKppqy-DfM5TJWTSuX4yAAU=nO1FeMMKyNc@xxxxxxxxxxxxxx> <4C9299D6.80700@xxxxxxxx>
Sender: xen-devel-bounces@xxxxxxxxxxxxxxxxxxx
On Thu, Sep 16, 2010 at 3:27 PM, Jeremy Fitzhardinge <jeremy@xxxxxxxx> wrote:
>  On 09/15/2010 04:37 PM, Bruce Edge wrote:
>> With top of xen-unstable and pvops 2.6.32.x I get this on the dom0
>> serial console when I start a linux pvops domU (using the same kernel
>> as the dom0):
>
> Do you know what the actual git revision is?

b297cdac0373625d3cd0e6f2b393570dcf2edba6. It's a current 2.6.32.x tree.

> Change 04cc1e6a6a85c4 fixes this problem.

I checked the source against that checkin and it's definitely applied
to the kernel, both dom0 and domU (I have the same kernel for both).

It only happens the first time on each dom0 boot. After the first domU
is created, it's not reported again.

Here's another occurrence on a completely different HW platform than
the first one I reported. The original was on a HP Proliant with 2
CPU, this is on a SuperMicro with 1 CPU.

[   88.876635] WARNING: at kernel/lockdep.c:2323
trace_hardirqs_on_caller+0x12f/0x190()
[   88.876669] Hardware name: X8ST3
[   88.876716] Modules linked in: xt_physdev ipmi_msghandler ipv6
osa_mfgdom0 xenfs xen_gntdev xen_evtchn xen_pciback tun bridge stp llc
serio_raw lp joydev ppdev ioatdma parport_pc dca parport usb_storage
e1000e usbhid hid
[   88.876930] Pid: 11, comm: xenwatch Not tainted 2.6.32.21-1 #1
[   88.876972] Call Trace:
[   88.876985]  <IRQ>  [<ffffffff810aa22f>] ?
trace_hardirqs_on_caller+0x12f/0x190
[   88.877027]  [<ffffffff8106bf70>] warn_slowpath_common+0x80/0xd0
[   88.877059]  [<ffffffff815f3f50>] ? _spin_unlock_irq+0x30/0x40
[   88.877090]  [<ffffffff8106bfd4>] warn_slowpath_null+0x14/0x20
[   88.877120]  [<ffffffff810aa22f>] trace_hardirqs_on_caller+0x12f/0x190
[   88.877151]  [<ffffffff810aa29d>] trace_hardirqs_on+0xd/0x10
[   88.877181]  [<ffffffff815f3f50>] _spin_unlock_irq+0x30/0x40
[   88.877212]  [<ffffffff813c5055>] add_to_net_schedule_list_tail+0x85/0xd0
[   88.877243]  [<ffffffff813c62a6>] netif_be_int+0x36/0x160
[   88.877269]  [<ffffffff810e1150>] handle_IRQ_event+0x70/0x180
[   88.877300]  [<ffffffff810e3769>] handle_edge_irq+0xc9/0x170
[   88.877330]  [<ffffffff813b8dff>] __xen_evtchn_do_upcall+0x1bf/0x1f0
[   88.877361]  [<ffffffff813b937d>] xen_evtchn_do_upcall+0x3d/0x60
[   88.877393]  [<ffffffff8101647e>] xen_do_hypervisor_callback+0x1e/0x30
[   88.877422]  <EOI>  [<ffffffff8100940a>] ? hypercall_page+0x40a/0x1010
[   88.877463]  [<ffffffff8100940a>] ? hypercall_page+0x40a/0x1010
[   88.877500]  [<ffffffff813bcee4>] ? xb_write+0x1e4/0x290
[   88.877538]  [<ffffffff813bd95a>] ? xs_talkv+0x6a/0x1f0
[   88.877575]  [<ffffffff813bd968>] ? xs_talkv+0x78/0x1f0
[   88.877604]  [<ffffffff813bdc5d>] ? xs_single+0x4d/0x60
[   88.877629]  [<ffffffff813be592>] ? xenbus_read+0x52/0x80
[   88.877655]  [<ffffffff813c888c>] ? frontend_changed+0x48c/0x770
[   88.877686]  [<ffffffff813bf7fd>] ? xenbus_otherend_changed+0xdd/0x1b0
[   88.877717]  [<ffffffff810111ef>] ? xen_restore_fl_direct_end+0x0/0x1
[   88.877748]  [<ffffffff810ac8c0>] ? lock_release+0xb0/0x230
[   88.877774]  [<ffffffff813bfb70>] ? frontend_changed+0x10/0x20
[   88.877804]  [<ffffffff813bd585>] ? xenwatch_thread+0x55/0x160
[   88.877836]  [<ffffffff810934a0>] ? autoremove_wake_function+0x0/0x40
[   88.877876]  [<ffffffff813bd530>] ? xenwatch_thread+0x0/0x160
[   88.877922]  [<ffffffff81093126>] ? kthread+0x96/0xb0
[   88.877960]  [<ffffffff8101632a>] ? child_rip+0xa/0x20
[   88.877988]  [<ffffffff81015c90>] ? restore_args+0x0/0x30
[   88.878043]  [<ffffffff81016320>] ? child_rip+0x0/0x20
[   88.878067] ---[ end trace 159d41648bcc43b4 ]---
[   95.254217] vif1.0: no IPv6 routers present



-Bruce

>
>    J
>
>> 0 kaan-18  ~ #> (XEN) tmem: all pools frozen for all domains
>> (XEN) tmem: all pools thawed for all domains
>> (XEN) tmem: all pools frozen for all domains
>> (XEN) tmem: all pools thawed for all domains
>>
>> mapping kernel into physical memory
>> about to get started...
>> [  188.652747] ------------[ cut here ]------------
>> [  188.652800] WARNING: at kernel/lockdep.c:2323
>> trace_hardirqs_on_caller+0x12f/0x190()
>> [  188.652826] Hardware name: ProLiant DL380 G6
>> [  188.652844] Modules linked in: xt_physdev osa_mfgdom0 xenfs
>> xen_gntdev xen_evtchn ipv6 fbcon tileblit font bitblit softcursor
>> xen_pciback radeon tun ttm drm_kms_helper serio_raw ipmi_si bridge drm
>> i2c_algo_bit stp ipmi_msghandler joydev hpilo hpwdt i2c_core llc lp
>> parport usbhid hid cciss usb_storage
>> [  188.653063] Pid: 11, comm: xenwatch Not tainted 2.6.32.21-1 #1
>> [  188.653084] Call Trace:
>> [  188.653094]  <IRQ>  [<ffffffff810aa22f>] ?
>> trace_hardirqs_on_caller+0x12f/0x190
>> [  188.653127]  [<ffffffff8106bf70>] warn_slowpath_common+0x80/0xd0
>> [  188.653153]  [<ffffffff815f3f50>] ? _spin_unlock_irq+0x30/0x40
>> [  188.653177]  [<ffffffff8106bfd4>] warn_slowpath_null+0x14/0x20
>> [  188.653200]  [<ffffffff810aa22f>] trace_hardirqs_on_caller+0x12f/0x190
>> [  188.653224]  [<ffffffff810aa29d>] trace_hardirqs_on+0xd/0x10
>> [  188.653246]  [<ffffffff815f3f50>] _spin_unlock_irq+0x30/0x40
>> [  188.653270]  [<ffffffff813c5055>] add_to_net_schedule_list_tail+0x85/0xd0
>> [  188.653294]  [<ffffffff813c62a6>] netif_be_int+0x36/0x160
>> [  188.653314]  [<ffffffff810e1150>] handle_IRQ_event+0x70/0x180
>> [  188.653338]  [<ffffffff810e3769>] handle_edge_irq+0xc9/0x170
>> [  188.653362]  [<ffffffff813b8dff>] __xen_evtchn_do_upcall+0x1bf/0x1f0
>> [  188.653385]  [<ffffffff813b937d>] xen_evtchn_do_upcall+0x3d/0x60
>> [  188.653409]  [<ffffffff8101647e>] xen_do_hypervisor_callback+0x1e/0x30
>> [  188.653431]  <EOI>  [<ffffffff8100940a>] ? hypercall_page+0x40a/0x1010
>> [  188.653464]  [<ffffffff8100940a>] ? hypercall_page+0x40a/0x1010
>> [  188.653487]  [<ffffffff813bcee4>] ? xb_write+0x1e4/0x290
>> [  188.653507]  [<ffffffff813bd95a>] ? xs_talkv+0x6a/0x1f0
>> [  188.653526]  [<ffffffff813bd968>] ? xs_talkv+0x78/0x1f0
>> [  188.653546]  [<ffffffff813bdc5d>] ? xs_single+0x4d/0x60
>> [  188.653565]  [<ffffffff813be592>] ? xenbus_read+0x52/0x80
>> [  188.653585]  [<ffffffff813c888c>] ? frontend_changed+0x48c/0x770
>> [  188.653609]  [<ffffffff813bf7fd>] ? xenbus_otherend_changed+0xdd/0x1b0
>> [  188.653633]  [<ffffffff810111ef>] ? xen_restore_fl_direct_end+0x0/0x1
>> [  188.653656]  [<ffffffff810ac8c0>] ? lock_release+0xb0/0x230
>> [  188.653676]  [<ffffffff813bfb70>] ? frontend_changed+0x10/0x20
>> [  188.653699]  [<ffffffff813bd585>] ? xenwatch_thread+0x55/0x160
>> [  188.653723]  [<ffffffff810934a0>] ? autoremove_wake_function+0x0/0x40
>> [  188.653747]  [<ffffffff813bd530>] ? xenwatch_thread+0x0/0x160
>> [  188.653770]  [<ffffffff81093126>] ? kthread+0x96/0xb0
>> [  188.653790]  [<ffffffff8101632a>] ? child_rip+0xa/0x20
>> [  188.653809]  [<ffffffff81015c90>] ? restore_args+0x0/0x30
>> [  188.653828]  [<ffffffff81016320>] ? child_rip+0x0/0x20
>> [  188.653846] ---[ end trace ab2eaae7afa5acdb ]---
>> [  195.184915] vif1.0: no IPv6 routers present
>>
>> The domU does start and is functional, complete with PCI passthrough.
>>
>> kern.log repeats the same information, but precedes it with this:
>>
>>
>> ==> kern.log <==
>> 2010-09-15T16:29:40.921979-07:00 kaan-18 [  188.562552] blkback:
>> ring-ref 8, event-channel 73, protocol 1 (x86_64-abi)
>> 2010-09-15T16:29:40.922202-07:00 kaan-18 [  188.562647]   alloc
>> irq_desc for 481 on node 0
>> 2010-09-15T16:29:40.922310-07:00 kaan-18 [  188.562652]   alloc
>> kstat_irqs on node 0
>> 2010-09-15T16:29:40.971943-07:00 kaan-18 [  188.608807]   alloc
>> irq_desc for 480 on node 0
>> 2010-09-15T16:29:40.972387-07:00 kaan-18 [  188.608813]   alloc
>> kstat_irqs on node 0
>> 2010-09-15T16:29:41.013027-07:00 kaan-18 [  188.652651]   alloc
>> irq_desc for 479 on node 0
>> 2010-09-15T16:29:41.013286-07:00 kaan-18 [  188.652657]   alloc
>> kstat_irqs on node 0
>>
>> Let me know if there's anything else I can provide.
>>
>> -Bruce
>>
>> _______________________________________________
>> Xen-devel mailing list
>> Xen-devel@xxxxxxxxxxxxxxxxxxx
>> http://lists.xensource.com/xen-devel
>>
>
>

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel