|
|
|
|
|
|
|
|
|
|
xen-devel
Re: [Xen-devel] [PATCH] Network Checksum Removal
On Monday 23 May 2005 06:59 pm, Ian Pratt wrote:
> > I get the following domU->dom0 throughput on my system (using
> > netperf3 TCP_STREAM testcase):
> > tx on ~1580Mbps
> > tx off ~1230Mbps
> >
> > with my previous patch (on Friday's build), I was seeing the
> > following:
> > with patch ~1610Mbps
> > no patch ~1100Mbps
> >
> > The slight difference between the two might be caused by the
> > changes that were incorporated in xen between those dates.
> > If you think it is worth the time, I can back port the latest
> > patch to Friday's build to see if that makes a difference.
>
> Are you sure these aren't within 'experimental error'? I can't think of
> anything that's changed since Friday that could be effecting this, but
> it would be good to dig a bit further as the difference in 'no patch'
> results is quite significant.
The "tx off" is probably higher because of the offloading for the rx (in both
the netback not checksumming and the physical ethernet checksum verification
being passed to domU).
I'm not sure why "tx on" is lower than my previous tests. It could be
something outside the patch which has been incorporated, or it could be
something in the patch that was committed. The changelog patch diff was
truncated, so I will have to create a diff to apply to my Friday tree to see
if the problem lies in the latter.
> It might be revealing to try running some results on the unpatched
> Fri/Sat/Sun tree.
>
> BTW, dom0<->domU is not that interesting as I'd generally discourage
> people from running services in dom0.
That is why I designed the checksum offload patch the way I did, as there were
otherways which would be significantly better domU->dom0 communication (but
would cause significantly more calculation in dom0).
> I'd be really interested to see
> the following tests:
>
> domU <-> external [dom0 on cpu0; dom1 on cpu1]
> domU <-> external [dom0 on cpu0; dom1 on cpu0]
> domU <-> domU [dom0 on cpu0; dom1 on cpu1; dom2 on cpu2 ** on a 4 way]
> domU <-> domU [dom0 on cpu0; dom1 on cpu0; dom2 on cpu0 ]
> domU <-> domU [dom0 on cpu0; dom1 on cpu1; dom2 on cpu1 ]
> domU <-> domU [dom0 on cpu0; dom1 on cpu0; dom2 on cpu1 ]
> domU <-> domU [dom0 on cpu0; dom1 on cpu1; dom2 on cpu2 ** cpu2
> hyperthread w/ cpu 0]
> domU <-> domU [dom0 on cpu0; dom1 on cpu1; dom2 on cpu3 ** cpu3
> hyperthread w/ cpu 1]
>
> This might help us understand the performance of interdomin networking
> rather better than we do at present. If you could fill a few of these in
> that would be great.
I wish I had all the hardware you describe ;-)
My tests are running on a pentium4 (which has hyperthreading, which shows up
as 2 cpus). dom0 was on cpu0 and domU was on cpu1. I'll be happy to run
netperf on the hardware I have.
Thanks,
Jon
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|
|
|
|
|