WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-devel

Re: [Xen-devel] xen 4 only seeing one keyboard and mouse

On Fri, Sep 03, 2010 at 07:50:06PM +0100, M A Young wrote:
> On Tue, 31 Aug 2010, Konrad Rzeszutek Wilk wrote:
> 
> >If you could, can you instrument it to print the cfg->domain, before the 
> >'vector_allocation_domain'
> >is called, and as well instrument the assign_irq_vector similary to what you 
> >did with Xen?
> >
> >And also instrument the 'dest' value. Basically the idea is to get an idea 
> >of what the
> >per_cpu(vector) gets set during the bootup for legacy IRQs. Similary to what 
> >you did
> >with Xen.
> 
> The kernel code I was working with (2.6.32) doesn't have the
> vector_allocation_domain section . I am attaching the debugging
> output I did get and the patch I used.

OK, so based on the output the IO APIC pins for you two IRQs should have 
destination set
to 1.

.. snip..
_IO_APIC_irq: cfg->domain=-1
setup_IO_APIC_irq: dest=1
IOAPIC[0]: Set routing entry (2-1 -> 0x31 -> IRQ 1 Mode:0 Active:0)
setup_IO_APIC_irq: cfg->domain=-1
setup_IO_APIC_irq: dest=1
IOAPIC[0]: Set routing entry (2-12 -> 0x3c -> IRQ 12 Mode:0 Active:0)

BUT when the IO APIC is being printed:

 01 003 0    0    0   0   0    1    1    31
 0c 003 0    0    0   0   0    1    1    3C


They are set to dest = 3!?

Somehow the IO APIC is being programmed without using the setup_IO_APIC_irq
and its friends. Also the per_cpu(vector_irq, 1)[0x31] = 1 is set.

So, for the second problem, I think the __setup_vector_irq is the one
that sets the vectors on the second CPU to correspond to the right IRQs.

But I am not sure how the IOAPIC pin for all IRQs below 16 get set to '3'.
There is something happening between the initial call to init IO_APIC IRQs
and when it is being printed that sets the destination to a new value.

I've piggybacked on your debug patch and added some extra stuff to see if the 
__setup_vector_irq
is responsible for setting the new per_cpu. Those printk's _might_ not work as
all of that is being run on a secondary CPU that is being initialized..?

For the IO APIC programming I added a printk/debug_stack by the ioapic_write
to see who and when sets those pins on the IOAPIC to 3.

Here is the patch:

diff --git a/arch/x86/kernel/apic/io_apic.c b/arch/x86/kernel/apic/io_apic.c
index ec4e874..37482fe 100644
--- a/arch/x86/kernel/apic/io_apic.c
+++ b/arch/x86/kernel/apic/io_apic.c
@@ -69,7 +69,7 @@
 #include <asm/xen/pci.h>
 
 #include <asm/xen/pci.h>
-
+#include <linux/kernel.h>
 #define __apicdebuginit(type) static type __init
 #define for_each_irq_pin(entry, head) \
        for (entry = head; entry; entry = entry->next)
@@ -486,6 +486,11 @@ __ioapic_write_entry(int apic, int pin, struct 
IO_APIC_route_entry e)
 void ioapic_write_entry(int apic, int pin, struct IO_APIC_route_entry e)
 {
        unsigned long flags;
+       if (pin == 1 || pin == 0xc) {
+               printk(KERN_INFO "Reprogramming PIN%d, dest=%d\n", pin, e.dest);
+               if (e.dest > 1)
+                       dump_stack();
+       }
        spin_lock_irqsave(&ioapic_lock, flags);
        __ioapic_write_entry(apic, pin, e);
        spin_unlock_irqrestore(&ioapic_lock, flags);
@@ -1198,6 +1203,7 @@ __assign_irq_vector(int irq, struct irq_cfg *cfg, const 
struct cpumask *mask)
        if (old_vector) {
                cpumask_and(tmp_mask, mask, cpu_online_mask);
                cpumask_and(tmp_mask, cfg->domain, tmp_mask);
+               printk(KERN_INFO "old_vector: %d  mask: %x\n", old_vector, 
tmp_mask->bits[0]);
                if (!cpumask_empty(tmp_mask)) {
                        free_cpumask_var(tmp_mask);
                        return 0;
@@ -1214,6 +1220,8 @@ __assign_irq_vector(int irq, struct irq_cfg *cfg, const 
struct cpumask *mask)
 
                vector = current_vector;
                offset = current_offset;
+               printk(KERN_INFO "vector: %d, mask: %x, cpu: %d per_cpu:%x\n",
+                       vector, tmp_mask->bits[0], cpu, per_cpu(vector_irq, 
cpu)[vector]);
 next:
                vector += 8;
                if (vector >= first_system_vector) {
@@ -1237,8 +1245,11 @@ next:
                        cfg->move_in_progress = 1;
                        cpumask_copy(cfg->old_domain, cfg->domain);
                }
-               for_each_cpu_and(new_cpu, tmp_mask, cpu_online_mask)
+               for_each_cpu_and(new_cpu, tmp_mask, cpu_online_mask) {
                        per_cpu(vector_irq, new_cpu)[vector] = irq;
+                       printk(KERN_WARNING "__assign_irq_vector: setting 
vector_irq[%d]=%d for cpu=%d\n",
+                               vector, irq, new_cpu);
+                       }
                cfg->vector = vector;
                cpumask_copy(cfg->domain, tmp_mask);
                err = 0;
@@ -1304,6 +1315,8 @@ void __setup_vector_irq(int cpu)
                if (!cpumask_test_cpu(cpu, cfg->domain))
                        continue;
                vector = cfg->vector;
+               printk(KERN_INFO "%s: vector: %d on CPU %d set to IRQ: %d\n",
+                               __FUNCTION__, vector, cpu, irq);
                per_cpu(vector_irq, cpu)[vector] = irq;
        }
        /* Mark the free vectors */
@@ -1313,8 +1326,11 @@ void __setup_vector_irq(int cpu)
                        continue;
 
                cfg = irq_cfg(irq);
-               if (!cpumask_test_cpu(cpu, cfg->domain))
+               if (!cpumask_test_cpu(cpu, cfg->domain)) {
+                       printk(KERN_INFO "%s: vector %d on CPU %d reset b/c not 
in affinity mask (%d)\n",
+                                       __FUNCTION__, vector, cpu, 
cfg->domain->bits[0]);
                        per_cpu(vector_irq, cpu)[vector] = -1;
+               }
        }
 }
 
@@ -1452,7 +1468,20 @@ int setup_ioapic_entry(int apic_id, int irq,
                entry->mask = 1;
        return 0;
 }
-
+static void dump_vectors(const char *prefix) {
+       int cpu;
+       int vector;
+
+       for (vector = 0x30; vector < 0x3f; vector++) {
+               for_each_cpu_and(cpu, 0xff, cpu_online_mask) {
+                       if (per_cpu(vector_irq, cpu)[vector] != -1)
+                               printk(KERN_INFO "%s [vec:%d,cpu:%d] = 
irq:%d\n",
+                                       prefix,
+                                       vector, cpu,
+                                       per_cpu(vector_irq, cpu)[vector]);
+               }
+       }
+}
 static void setup_IO_APIC_irq(int apic_id, int pin, unsigned int irq, struct 
irq_desc *desc,
                              int trigger, int polarity)
 {
@@ -1465,10 +1494,15 @@ static void setup_IO_APIC_irq(int apic_id, int pin, 
unsigned int irq, struct irq
 
        cfg = desc->chip_data;
 
+       printk(KERN_WARNING "setup_IO_APIC_irq: cfg->domain=%d (vector: %d)\n", 
cfg->domain->bits[0], cfg->vector);
+
+       dump_vectors("PRE");
        if (assign_irq_vector(irq, cfg, apic->target_cpus()))
                return;
+       dump_vectors("PAST");
 
        dest = apic->cpu_mask_to_apicid_and(cfg->domain, apic->target_cpus());
+       printk(KERN_WARNING "setup_IO_APIC_irq: dest=%d\n", dest);
 
        apic_printk(APIC_VERBOSE,KERN_DEBUG
                    "IOAPIC[%d]: Set routing entry (%d-%d -> 0x%x -> "

Attachment: debug-ioapic-irq-2-12.patch
Description: Text Data

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel