* NEW: My 32bit XEN builds are still not PAE enabled. Chasing done
why this is. For now I stopped running the 32b SMP Guests.
Test Configuration:
Dell PowerEdge 430, Dual Core, 2GB, 3 SATA (Intel VT)
32bit XEN PAE Hypervisor on a RHEL4U2 32bit root (/dev/sda)
dom0_mem=256M (required to boot domUs)
32bit fully virtualized (HVM) guest RHEL4U2 256MB (/dev/sdb)
pae=0, acpi=1, apic=1
kernargs clock=pit
32bit fully virtualized (HVM) guest RHEL3U6 256MB (/dev/sdc)
pae=0, acpi=1, apic=1
kernargs clock=pit
Boot Tests:
Boot a fully virtualized (HVM) guest to the login prompt
Results are marked Pass|Fail where (n) points to a failure description
Regression Tests:
851 tests (850 ltp tests and one 30 minute user load test)
Tests are marked #Pass/#Fail where (n) points to a failure description
XEN 32bit PAE 2 CPU Hypervisor (booted smp):
----------------------------------------------------------------------
| XEN | Guest Kernel (SMP kernels booted with 2 CPU) |
| Changeset|-----------------------------------------------------------|
| | RHEL4 UP | RHEL4 SMP | RHEL3 UP | RHEL3 SMP |
| |--------------|--------------|--------------|--------------|
| | Boot | Test | Boot | Test | Boot | Test | Boot | Test |
|----------|------|-------|------|-------|------|-------|------|-------|
| 9960 | Pass | 850/1 | | | Pass |Running| | |
| | | (2) | | | | (2,4) | | |
|----------|------|-------|------|-------|------|-------|------|-------|
| 9925 | Pass | 850/1 | Fail | | Pass | 850/1 | Fail | |
| | | (2) | (1) | | | (2,4) | (1) | |
|----------|------|-------|------|-------|------|-------|------|-------|
| 9920 | Pass | 850/1 | Fail | | Pass | 850/1 | Fail | |
| | | (2) | (1) | | | (2,4) | (1) | |
|----------|------|-------|------|-------|------|-------|------|-------|
| 9913 | Fail | | Fail | | Fail | | Fail | |
| | (3) | | (3) | | (3) | | (3) | |
|----------|------|-------|------|-------|------|-------|------|-------|
| 9903 | Fail | | Fail | | Fail | | Fail | |
| | (3) | | (3) | | (3) | | (3) | |
----------------------------------------------------------------------
Failures:
1. 32 bit SMP guests hang on boot
"Uncompressing Linux... Ok, booting the kernel."
2. 32bit UP guest fail ltp gettimeofday02
"Time is going backwards"
3. [Fixed in 9920] Build broken:
cc1: warnings being treated as errors
mm.c: In function subarch_init_memory:
mm.c:163: warning: format %ld expects type long int,
but argument 2 has type unsigned int
mm.c:163: warning: format %ld expects type long int,
but argument 3 has type unsigned int
4. The RHEL3 UP tests run OK however they take 12 hours
to complete where on RHEL4 UP the run for ~4 hours.
There seems to be some performance issue running a
2.4 kernel fully virtualized guest. One test in
particular ltp's aso-stress025 takes 4 hour to run
on RHEL3U6-32b while it take 4 minutes on RHEL4U2-32b.
File: regression.1
Time Level Message
05:00:50 INFO Reporting status: 'Test Running' for test:
ltp_gettimeofday02
05:00:52 INFO Preparing to run test 'ltp_gettimeofday02' using
profile: /qa/conductor/profiles/ltp/syscalls/gettimeofday02.xml
05:00:52 INFO Starting test 'ltp_gettimeofday02' using profile:
/qa/conductor/profiles/ltp/syscalls/gettimeofday02.xml
05:00:52 INFO Dispatching operation: RemoteShell
05:00:52 FINE Client sequencer got message requesting the start of a
new test: ltp_gettimeofday02
05:00:52 FINER Client sequencer sent message of type: 4 with seq num:
1 of size: 289 bytes
05:00:52 FINER Client sequencer handling new operation from control
sequencer
05:00:52 FINE Client sequencer looking for class:
com.katana.conductor.operations.RemoteShell
05:00:52 INFO Operation RemoteShell running
05:00:52 FINE Client sequencer was told that an operation is now
running
05:00:52 INFO RemoteShell: target node(s) = vs177
05:00:52 INFO ssh: /usr/bin/ssh root@vs177 cd
/qa/conductor/tests/ltp/testcases/bin; gettimeofday02
05:00:52 FINE ssh: waiting for command to finish
05:00:53 INFO ssh: gettimeofday02 0 INFO : checking if
gettimeofday is monotonous, takes 30s
05:00:53 INFO ssh: gettimeofday02 1 FAIL : Time is going backwards
(old 1145696453.61428 vs new 1145696453.60660!
05:00:53 FINE executeShellCmd(ssh): exit value is 1
05:00:53 SEVERE RemoteShell: command failed with error = 1
05:00:53 SEVERE Operation RemoteShell failed
05:00:53 SEVERE Reporting status: 'Test Failed' for test:
ltp_gettimeofday02
05:00:53 FINE Client sequencer detected operation completed with
status of: Fail
05:00:53 FINER Client sequencer sent message of type: 5 with seq num:
2 of size: 429 bytes
05:00:53 SEVERE Crash Collection disabled for queue : RHEL4U2-32b-XEN
05:00:53 INFO Cleaning up after test Queue: RHEL3U6-32b-XEN32
User: QA
Hypervisor Build: 20060505_000 Version: main Config: Release Path:
/repository/trees/main/20060505_000/xen32/xen-unstable/dist Dist:Xen32
Completion status:
test run completed
Start time: May 4, 2006 3:57:19 PM
End time: May 5, 2006 4:11:28 AM
Elapsed time: 12 hours 14 minutes 9 seconds
Tests/Config changes in queue: 851
Tests processed: 851
Tests passed: 850
Tests failed: 1
Tests aborted: 0
Forced reboots: 0
aio-stress025 performance issue:
RHEL4U2-32b native hardware: 6 seconds
RHEL4U2-32b domU guest: 3 minutes 31 seconds
RHEL3U6-32b domU guest: 173 minutes 1 second
Here are the timed runs:
dom0 run on native hardware:
----------------------------
[root@tst122 ~]# time ./aio-stress -I500 -o3 -O -r512 -t8 /test/aiodio/junkfile
/test/aiodio/file2 /test/aiodio/file7 /test/aiodio/file8 /test/aiodio/file3
/test/aiodio/file4 /test/aiodio/file5 /test/aiodio/file6
adding stage random read
starting with random read
file size 1024MB, record size 512KB, depth 64, ios per iteration 8
max io_submit 64, buffer alignment set to 4KB
threads 8 files 8 contexts 1 context offset 2MB verification off
Running multi thread version num_threads:8
random read on /test/aiodio/file3 (182109.19 MB/s) 1024.00 MB in 0.01s
random read on /test/aiodio/file6 (175975.25 MB/s) 1024.00 MB in 0.01s
random read on /test/aiodio/file4 (176887.20 MB/s) 1024.00 MB in 0.01s
random read on /test/aiodio/file2 (40123.82 MB/s) 1024.00 MB in 0.03s
random read on /test/aiodio/file5 (95934.05 MB/s) 1024.00 MB in 0.01s
random read on /test/aiodio/file8 (4322.75 MB/s) 1024.00 MB in 0.24s
random read on /test/aiodio/file7 (4224.42 MB/s) 1024.00 MB in 0.24s
thread 5 random read totals (2815.35 MB/s) 1024.00 MB in 0.36s
thread 7 random read totals (2770.62 MB/s) 1024.00 MB in 0.37s
thread 4 random read totals (2728.51 MB/s) 1024.00 MB in 0.38s
thread 1 random read totals (2720.31 MB/s) 1024.00 MB in 0.38s
thread 6 random read totals (2860.49 MB/s) 1024.00 MB in 0.36s
thread 2 random read totals (1153.87 MB/s) 1024.00 MB in 0.89s
thread 3 random read totals (1153.87 MB/s) 1024.00 MB in 0.89s
random read on /test/aiodio/junkfile (1.63 MB/s) 4.00 MB in 2.46s
thread 0 random read totals (1.62 MB/s) 4.00 MB in 2.47s
random read throughput (2905.81 MB/s) 7172.00 MB in 2.47s min transfer 4.00MB
real 0m5.748s
user 0m0.050s
sys 0m0.620s
[root@tst122 ~]#
RHEL4U2-32b domU guest:
-----------------------
[root@vs162 testcases]# time ./kernel/io/ltp-aiodio/aio-stress -I500 -o3 -O
-r512 -t8 /test/aiodio/junkfile /test/aiodio/file2 /test/aiodio/file7
/test/aiodio/file8 /test/aiodio/file3 /test/aiodio/file4 /test/aiodio/file5
/test/aiodio/file6
adding stage random read
starting with random read
file size 1024MB, record size 512KB, depth 64, ios per iteration 8
max io_submit 64, buffer alignment set to 4KB
threads 8 files 8 contexts 1 context offset 2MB verification off
Running multi thread version num_threads:8
random read on /test/aiodio/file2 (7.30 MB/s) 1024.00 MB in 140.35s
thread 1 random read totals (7.27 MB/s) 1024.00 MB in 140.84s
random read on /test/aiodio/file4 (0.06 MB/s) 8.00 MB in 141.47s
thread 5 random read totals (0.06 MB/s) 8.00 MB in 141.60s
random read on /test/aiodio/file5 (0.03 MB/s) 4.00 MB in 141.83s
thread 6 random read totals (0.03 MB/s) 4.00 MB in 141.94s
random read on /test/aiodio/file8 (0.03 MB/s) 4.00 MB in 142.82s
thread 3 random read totals (0.03 MB/s) 4.00 MB in 142.92s
random read on /test/aiodio/file3 (0.03 MB/s) 4.00 MB in 143.04s
thread 4 random read totals (0.03 MB/s) 4.00 MB in 143.08s
random read on /test/aiodio/file6 (0.03 MB/s) 4.00 MB in 143.45s
thread 7 random read totals (0.03 MB/s) 4.00 MB in 143.52s
random read on /test/aiodio/file7 (0.03 MB/s) 4.00 MB in 143.80s
thread 2 random read totals (0.03 MB/s) 4.00 MB in 143.91s
random read on /test/aiodio/junkfile (0.03 MB/s) 4.00 MB in 145.83s
thread 0 random read totals (0.03 MB/s) 4.00 MB in 145.92s
random read throughput (7.24 MB/s) 1056.00 MB in 145.92s min transfer 4.00MB
real 2m31.215s
user 0m0.724s
sys 0m1.916s
[root@vs162 testcases]#
RHEL3U6 domU guest:
-------------------
[root@vs174 testcases]# time ./kernel/io/ltp-aiodio/aio-stress -I500 -o3 -O
-r512 -t8 /test/aiodio/junkfile /test/aiodio/file2 /test/aiodio/file7
/test/aiodio/file8 /test/aiodio/file3 /test/aiodio/file4 /test/aiodio/file5
/test/aiodio/file6
adding stage random read
starting with random read
file size 1024MB, record size 512KB, depth 64, ios per iteration 8
max io_submit 64, buffer alignment set to 4KB
threads 8 files 8 contexts 1 context offset 2MB verification off
Running multi thread version num_threads:8
random read on /test/aiodio/file7 (0.10 MB/s) 1024.00 MB in 10322.46s
thread 2 random read totals (0.10 MB/s) 1024.00 MB in 10323.40s
random read on /test/aiodio/file4 (0.10 MB/s) 1024.00 MB in 10332.31s
thread 5 random read totals (0.10 MB/s) 1024.00 MB in 10332.59s
random read on /test/aiodio/file5 (0.10 MB/s) 1024.00 MB in 10344.51s
thread 6 random read totals (0.10 MB/s) 1024.00 MB in 10344.56s
random read on /test/aiodio/file6 (0.10 MB/s) 1024.00 MB in 10346.31s
thread 7 random read totals (0.10 MB/s) 1024.00 MB in 10346.88s
random read on /test/aiodio/file3 (0.10 MB/s) 1024.00 MB in 10353.64s
thread 4 random read totals (0.10 MB/s) 1024.00 MB in 10353.68s
random read on /test/aiodio/file8 (0.09 MB/s) 972.00 MB in 10373.37s
random read on /test/aiodio/file2 (0.08 MB/s) 860.00 MB in 10373.57s
random read on /test/aiodio/junkfile (0.08 MB/s) 860.00 MB in 10373.59s
thread 1 random read totals (0.08 MB/s) 860.00 MB in 10373.77s
thread 3 random read totals (0.09 MB/s) 972.00 MB in 10373.76s
thread 0 random read totals (0.08 MB/s) 860.00 MB in 10373.77s
random read throughput (0.75 MB/s) 7812.00 MB in 10373.77s min transfer 860.00MB
real 173m1.057s
user 0m1.140s
sys 0m49.050s
[root@vs174 testcases]#
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|