I switched over to Xen 3.5-unstable (changeset 20303) and pv_ops dom0 2.6.31.1
hoping that this would resolve the IRQ SMP affinity problem. I had to use
pci-stub to hide the PCI devices since pciback wasn't working. With vcpus=16
(APIC routing is physical flat), the interrupts were working in domU and being
routed to CPU0 with the default smp_affinity (ffff) but as soon as I changed it
to any 16-bit one-hot value or even setting it to the same default value
resulted in a complete loss of interrupts (even in the devices that didn't have
any change to smp_affinity). With vcpus=4 (APIC routing is logical flat), I can
see the interrupts being load balanced across all CPUs but as soon as I changed
smp_affinity to any value, the interrupts stopped. This used to work reliably
with the non-pv_ops kernel. I attached the logs in case anyone wants to take a
look.
I did see the MSI message address/data change in both domU and dom0 (using
"lspci -vv"):
vcpus=16:
domU MSI message address/data with default smp_affinity: Address:
00000000fee00000 Data: 40a9
domU MSI message address/data after smp_affinity=0010: Address:
00000000fee08000 Data: 40b1 (8 is APIC ID of CPU4)
dom0 MSI message address/data with default smp_affinity: Address:
00000000fee00000 Data: 4094
dom0 MSI message address/data after smp_affinity=0010: Address:
00000000fee00000 Data: 409c
Aside from "lspci -vv" what other means are there to track down this problem?
Is there some way to print the interrupt vector table? I'm considering adding
printk's to the code that Qing mentioned in his previous email (see below). Any
suggestions on where in the code to add the printk's?
Thanks.
Dante
-----Original Message-----
From: Qing He [mailto:qing.he@xxxxxxxxx]
Sent: Sunday, October 11, 2009 10:55 PM
To: Cinco, Dante
Cc: Keir Fraser; xen-devel@xxxxxxxxxxxxxxxxxxx; xiantao.zhang@xxxxxxxxx
Subject: Re: [Xen-devel] IRQ SMP affinity problems in domU with vcpus > 4 on HP
ProLiant G6 with dual Xeon 5540 (Nehalem)
On Mon, 2009-10-12 at 13:25 +0800, Cinco, Dante wrote:
> With vcpus < 4, logical flat mode works fine (no error message). I can
> change smp_affinity to any value > 0 and < 16 and the interrupts go to
> the proper CPU(s). Could you point me to the code that handles MSI so
> that I can better understand the MSI implementation?
There are two parts:
1) init or changing the data and address of MSI:
1) qemu-xen: hw/passthrough.c: pt_msg.*_write, MSI access are
trapped here first. And then pt_update_msi in
hw/pt-msi.c is called to update the MSI binding.
2) xen: drivers/passthrough/io.c: pt_irq_create_bind_vtd,
where MSI is actually bound to the guest.
2) on MSI reception:
In drivers/passthrough/io.c, hvm_do_IRQ_dpci and hvm_dirq_assist
are the routines responsible for handling all assigned irqs
(including MSI), and if an MSI is received, vmsi_deliver in
arch/x86/vmsi.c get called to deliver MSI to the corresponding
vlapic.
And I just learned from Xiantao Zhang that for the guest Linux kernel, it
enables per-cpu vector if it's in physical mode, and that looks more likely
relevant to this problem. It had problem in the older xen to handle this, and
changeset 20253 is supposed to fix it, although I noticed your xen version is
20270.
Thanks,
Qing
irq_smp_affinity_problem_pv_ops_dom0_2.6.31.1.tar.gz
Description: irq_smp_affinity_problem_pv_ops_dom0_2.6.31.1.tar.gz
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|