On Mon, 22 Feb 2010, Keith Coleman wrote:
> On Mon, Feb 22, 2010 at 12:27 PM, Stefano Stabellini
> <stefano.stabellini@xxxxxxxxxxxxx> wrote:
> > On Mon, 22 Feb 2010, Keith Coleman wrote:
> >> On 2/22/10, Stefano Stabellini <stefano.stabellini@xxxxxxxxxxxxx> wrote:
> >> > On Fri, 19 Feb 2010, Keith Coleman wrote:
> >> >> I am posting this to xen-devel instead of -users because it paints an
> >> >> incomplete picture that shouldn't be the basis for deciding how to run
> >> >> production systems.
> >> >>
> >> >> This graph shows the performance under a webserver disk IO workload at
> >> >> different queue depths. It compares the 4 main IO methods for windows
> >> >> guests that will be available in the upcoming xen 4.0.0 and 3.4.3
> >> >> releases: pure HVM, stub domains, gplpv drivers, and xcp winpv
> >> >> drivers.
> >> >>
> >> >> The gplpv and xcp winpv drivers have comparable performance with gplpv
> >> >> being slightly faster. Both pv drivers are considerably faster than
> >> >> pure hvm or stub domains. Stub domain performance was about even with
> >> >> HVM which is lower than we were expecting. We tried a different cpu
> >> >> pinning in "Stubdom B" with little impact.
> >> >>
> >> >
> >> > What disk backend are you using?
> >>
> >> phy, LV
> >>
> >
> > That is strange because in that configuration I get a far better
> > disk bandwidth with stubdoms compared to qemu running in dom0.
> >
>
> What type of test are you doing?
>
these are the results I got a while ago running a simple "dd if=/dev/zero
of=file" for 10 seconds:
qemu in dom0: 25.1 MB/s
qemu in a stubdom: 56.7 MB/s
I have just run just now "tiobench with --size 256 --numruns 4 --threads
4" using a raw file as a backend:
qemu in dom0, using blktap2, best run:
File Blk Num Avg Maximum Lat% Lat% CPU
Size Size Thr Rate (CPU%) Latency Latency >2s >10s Eff
------ ----- --- ------ ------ --------- ----------- -------- -------- -----
256 4096 4 85.82 108.6% 0.615 1534.10 0.00000 0.00000 79
qemu in a stubdom, using phy on a loop device, best run:
File Blk Num Avg Maximum Lat% Lat% CPU
Size Size Thr Rate (CPU%) Latency Latency >2s >10s Eff
------ ----- --- ------ ------ --------- ----------- -------- -------- -----
256 4096 4 130.49 163.8% 0.345 1459.94 0.00000 0.00000 80
These results as for the "sequential reads" test and rate is in
megabytes per second.
If I use phy on a loop device with qemu in dom0 unexpectedly I get much
worse results.
Same thing happens if I use tap:aio with qemu in a stubdom, but this is
kind of expected since blktap is never going to be as fast as blkback.
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|