On Tue, Feb 23, 2010 at 01:14:47PM +0000, Stefano Stabellini wrote:
> On Mon, 22 Feb 2010, Keith Coleman wrote:
> > On Mon, Feb 22, 2010 at 12:27 PM, Stefano Stabellini
> > <stefano.stabellini@xxxxxxxxxxxxx> wrote:
> > > On Mon, 22 Feb 2010, Keith Coleman wrote:
> > >> On 2/22/10, Stefano Stabellini <stefano.stabellini@xxxxxxxxxxxxx> wrote:
> > >> > On Fri, 19 Feb 2010, Keith Coleman wrote:
> > >> >> I am posting this to xen-devel instead of -users because it paints an
> > >> >> incomplete picture that shouldn't be the basis for deciding how to run
> > >> >> production systems.
> > >> >>
> > >> >> This graph shows the performance under a webserver disk IO workload at
> > >> >> different queue depths. It compares the 4 main IO methods for windows
> > >> >> guests that will be available in the upcoming xen 4.0.0 and 3.4.3
> > >> >> releases: pure HVM, stub domains, gplpv drivers, and xcp winpv
> > >> >> drivers.
> > >> >>
> > >> >> The gplpv and xcp winpv drivers have comparable performance with gplpv
> > >> >> being slightly faster. Both pv drivers are considerably faster than
> > >> >> pure hvm or stub domains. Stub domain performance was about even with
> > >> >> HVM which is lower than we were expecting. We tried a different cpu
> > >> >> pinning in "Stubdom B" with little impact.
> > >> >>
> > >> >
> > >> > What disk backend are you using?
> > >>
> > >> phy, LV
> > >>
> > >
> > > That is strange because in that configuration I get a far better
> > > disk bandwidth with stubdoms compared to qemu running in dom0.
> > >
> >
> > What type of test are you doing?
> >
>
> these are the results I got a while ago running a simple "dd if=/dev/zero
> of=file" for 10 seconds:
>
> qemu in dom0: 25.1 MB/s
> qemu in a stubdom: 56.7 MB/s
>
For dd tests you might want to use "oflag=direct" to make it use direct IO,
and not domU kernel cached.. also longer test would be good.
>
>
> I have just run just now "tiobench with --size 256 --numruns 4 --threads
> 4" using a raw file as a backend:
>
>
> qemu in dom0, using blktap2, best run:
>
> File Blk Num Avg Maximum Lat% Lat% CPU
> Size Size Thr Rate (CPU%) Latency Latency >2s >10s Eff
> ------ ----- --- ------ ------ --------- ----------- -------- -------- -----
> 256 4096 4 85.82 108.6% 0.615 1534.10 0.00000 0.00000 79
>
> qemu in a stubdom, using phy on a loop device, best run:
>
> File Blk Num Avg Maximum Lat% Lat% CPU
> Size Size Thr Rate (CPU%) Latency Latency >2s >10s Eff
> ------ ----- --- ------ ------ --------- ----------- -------- -------- -----
> 256 4096 4 130.49 163.8% 0.345 1459.94 0.00000 0.00000 80
>
>
> These results as for the "sequential reads" test and rate is in
> megabytes per second.
> If I use phy on a loop device with qemu in dom0 unexpectedly I get much
> worse results.
> Same thing happens if I use tap:aio with qemu in a stubdom, but this is
> kind of expected since blktap is never going to be as fast as blkback.
>
Hmm... what's the overall cpu usage difference, measured from the hypervisor?
"xm top" or so..
-- Pasi
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|