[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

RE: [Xen-devel] [PATCH 2/2] Add VMDq support to ixgbe



Anna,

Since you have 2 devices (eth0 and eth1) you need to pass a list of values for 
the VMDQ parameter, one for each device.
Try using "VMDQ=8,8".
Regards
Renato  

> -----Original Message-----
> From: xen-devel-bounces@xxxxxxxxxxxxxxxxxxx 
> [mailto:xen-devel-bounces@xxxxxxxxxxxxxxxxxxx] On Behalf Of 
> Fischer, Anna
> Sent: Friday, March 20, 2009 3:02 PM
> To: Mitch Williams
> Cc: xen-devel@xxxxxxxxxxxxxxxxxxx
> Subject: RE: [Xen-devel] [PATCH 2/2] Add VMDq support to ixgbe
> 
> > Subject: [Xen-devel] [PATCH 2/2] Add VMDq support to ixgbe
> > 
> > This patch adds experimental VMDq support (AKA Netchannel2 
> vmq) to the 
> > ixgbe driver.  This applies to the Netchannel2 tree, and 
> should NOT be 
> > applied to the "normal" development tree.
> > 
> > To enable VMDq functionality, load the driver with the command-line 
> > parameter VMDQ=<num queues>, as in:
> > 
> > $ modprobe ixgbe VMDQ=8
> 
> I have installed the latest netchannel2 tree. If I load ixgbe 
> with modprobe VMDQ=x then it seems as if only the first NIC 
> port has VMDQ enabled while the second stays disabled, or 
> only enabled with 2 RX queues and 1 TX queue. Is this 
> expected? Is it not possible to enable 16 queues on both NIC 
> ports? I have listed some logs below.
> 
> 
> ixgbe: Intel(R) 10 Gigabit PCI Express Network Driver - 
> version 1.3.56.5-vmq-NAPI Copyright (c) 1999-2008 Intel Corporation.
> bus pci: add driver ixgbe
> pci: Matched Device 0000:02:00.0 with Driver ixgbe
> PCI: Enabling device 0000:02:00.0 (0100 -> 0103)
> ACPI: PCI Interrupt 0000:02:00.0[A] -> GSI 16 (level, low) -> IRQ 16
> PCI: Enabling bus mastering for device 0000:02:00.0
> PCI: Setting latency timer of device 0000:02:00.0 to 64
> ixgbe: Virtual Machine Device Queues (VMDQ) set to 16
> ixgbe: packet split disabled for Xen VMDQ
> ixgbe: 0000:02:00.0: ixgbe_init_interrupt_scheme: Multiqueue 
> Enabled: Rx Queue count = 16, Tx Queue count = 16
> ixgbe: eth0: ixgbe_probe: (PCI Express:2.5Gb/s:Width x8)
> ixgbe: eth0: ixgbe_probe: MAC: 1, PHY: 0
> ixgbe: eth0: ixgbe_probe: Internal LRO is enabled
> ixgbe: eth0: ixgbe_probe: Intel(R) 10 Gigabit Network 
> Connection bound device '0000:02:00.0' to driver 'ixgbe'
> pci: Bound Device 0000:02:00.0 to Driver ixgbe
> pci: Matched Device 0000:02:00.1 with Driver ixgbe
> PCI: Enabling device 0000:02:00.1 (0100 -> 0103)
> ACPI: PCI Interrupt 0000:02:00.1[B] -> GSI 17 (level, low) -> IRQ 20
> PCI: Enabling bus mastering for device 0000:02:00.1
> PCI: Setting latency timer of device 0000:02:00.1 to 64
> ixgbe: 0000:02:00.1: ixgbe_init_interrupt_scheme: Multiqueue 
> Disabled: Rx Queue count = 1, Tx Queue count = 1
> ixgbe: eth1: ixgbe_probe: (PCI Express:2.5Gb/s:Width x8)
> ixgbe: eth1: ixgbe_probe: MAC: 1, PHY: 0
> ixgbe: eth1: ixgbe_probe: Internal LRO is enabled
> ixgbe: eth1: ixgbe_probe: Intel(R) 10 Gigabit Network 
> Connection bound device '0000:02:00.1' to driver 'ixgbe'
> pci: Bound Device 0000:02:00.1 to Driver ixgbe
> 
> 
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-devel
> 
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.