Hi,
Using an unstable snapshot from today (changeset 6940), I'm trying to
use a lvm2 partition block device as root of domU (in config disk =
['phy:vg/img,hda1,w'])
Try to boot domU I get:
Xen virtual console successfully installed as tty1
Event-channel device installed.
xen_blk: Initialising virtual block device driver
xen_blk: Timeout connecting to device!
xen_net: Initialising virtual ethernet driver.
NET: Registered protocol family 2
.....
VFS: Cannot open root device "hda1" or unknown-block(0,0)
Dom0 spits onto its console:
Oops: 0000 [#1]
SMP
CPU: 1
EIP: 0061:[<c0146282>] Not tainted VLI
EFLAGS: 00010202 (2.6.12.5-xen)
EIP is at generic_page_range+0x23/0x16a
eax: 00000000 ebx: c8c7c000 ecx: 00000000 edx: 00000323
esi: c8c7c000 edi: 05bf3000 ebp: c7ad5684 esp: c7ad5628
ds: 007b es: 007b ss: 0069
Process kxbwatch (pid: 2025, threadinfo=c7ad4000 task=c12d2020)
Stack: aa55aa55 aa55aa55 00000000 00000000 00000000 00000000 77dd77dd c8c7d000
00000000 00001000 c8c7c000 05bf3000 c7ad5684 c0115887 c0115714 c7ad5680
00000000 c7ad5694 00001000 0001056f 00001000 00000000 c7ad5684 f8181818
Call Trace:
[<c0115887>] __direct_remap_pfn_range+0x10c/0x14d
[<c0115714>] direct_remap_area_pte_fn+0x0/0x67
[<c01a4654>] search_by_key+0x41c/0xc48
[<c0158126>] bh_lru_install+0xc8/0x125
[<c0136eb0>] find_get_page+0x39/0x4a
[<c0158230>] __find_get_block+0xad/0x10c
[<c01a4654>] search_by_key+0x41c/0xc48
[<c01a4fd6>] search_for_position_by_key+0x156/0x37a
[<c0204023>] __make_request+0x3c6/0x46a
[<c01a4ee0>] search_for_position_by_key+0x60/0x37a
[<c01a4fd6>] search_for_position_by_key+0x156/0x37a
[<c01900a9>] make_cpu_key+0x4b/0x5b
[<c01a3ecd>] pathrelse+0x1e/0x2c
[<c01907c8>] _get_block_create_0+0x4c6/0x698
[<c013a52f>] mempool_alloc+0x72/0x100
[<c01a0bb0>] leaf_copy_items_entirely+0x191/0x1de
[<c01a2467>] leaf_paste_entries+0x108/0x1b5
[<c018b2b9>] balance_leaf+0xe50/0x2e89
[<c01197ef>] __wake_up+0x38/0x4e
[<c0117c73>] try_to_wake_up+0x2a4/0x2f5
[<c01197ef>] __wake_up+0x38/0x4e
[<c021f18a>] wake_waiting+0x1e/0x27
[<c01357af>] handle_IRQ_event+0x58/0x97
[<c01358f1>] __do_IRQ+0x103/0x132
[<c010af8a>] monotonic_clock+0x55/0x95
[<c02de1d0>] schedule+0x3e4/0xc34
[<c01197ef>] __wake_up+0x38/0x4e
[<c0118b08>] find_busiest_group+0x1bc/0x301
[<c0118e9b>] load_balance_newidle+0x30/0x8c
[<c010af8a>] monotonic_clock+0x55/0x95
[<c02de1d0>] schedule+0x3e4/0xc34
[<c02de203>] schedule+0x417/0xc34
[<c010af8a>] monotonic_clock+0x55/0x95
[<c02de203>] schedule+0x417/0xc34
[<c01197ef>] __wake_up+0x38/0x4e
[<c021f40e>] xs_input_avail+0x28/0x35
[<c021f59f>] xb_read+0x184/0x1c4
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c021f761>] read_reply+0x63/0xa2
[<c014d19d>] __get_vm_area+0x1b7/0x1ec
[<c02249eb>] map_frontend_pages+0x33/0x75
[<c0224a86>] netif_map+0x58/0x133
[<c022429e>] frontend_changed+0x190/0x222
[<c0220456>] watch_thread+0xf2/0x168
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c01198d3>] complete+0x40/0x55
[<c0220364>] watch_thread+0x0/0x168
[<c0130a6e>] kthread+0xa0/0xd4
[<c01309ce>] kthread+0x0/0xd4
[<c0106c81>] kernel_thread_helper+0x5/0xb
Code: b8 2e c0 e8 74 70 fd ff 55 8d 0c 0a 57 56 53 89 d3 83 ec 24 39
ca 89 44 24 20 89 4c 24 1c 0f 83 19 01 00 00 8b 4c 24 20 c1 ea 16 <8b>
41 1c 83 c1 3c 8d 2c 90 89 c8 89 4c 24 10 e8 dd 96 19 00 8b
Following an older post with a similar oops, I increased dom0 memory
from 128mb->256mb and got a slightly different trace:
Oops: 0000 [#1]
SMP
CPU: 1
EIP: 0061:[<c0146282>] Not tainted VLI
EFLAGS: 00010202 (2.6.12.5-xen)
EIP is at generic_page_range+0x23/0x16a
eax: 00000000 ebx: d0c7c000 ecx: 00000000 edx: 00000343
esi: d0c7c000 edi: 0dbf3000 ebp: cfa01684 esp: cfa01628
ds: 007b es: 007b ss: 0069
Process kxbwatch (pid: 2026, threadinfo=cfa00000 task=cf42da20)
Stack: cd29da20 00000000 c120ac20 00000000 cfa01684 c0117c73 00000000 d0c7d000
00000000 00001000 d0c7c000 0dbf3000 cfa01684 c0115887 c0115714 cfa01680
00000000 cfa01694 00001000 00010001 00001000 00000000 cfa01684 cfa016b8
Call Trace:
[<c0117c73>] try_to_wake_up+0x2a4/0x2f5
[<c0115887>] __direct_remap_pfn_range+0x10c/0x14d
[<c0115714>] direct_remap_area_pte_fn+0x0/0x67
[<c0130f21>] autoremove_wake_function+0x1b/0x43
[<c0119797>] __wake_up_common+0x35/0x55
[<c013a63b>] mempool_free+0x7e/0x8e
[<c01197ef>] __wake_up+0x38/0x4e
[<c015a148>] end_bio_bh_io_sync+0x0/0x4f
[<c013a63b>] mempool_free+0x7e/0x8e
[<c0158126>] bh_lru_install+0xc8/0x125
[<c0136eb0>] find_get_page+0x39/0x4a
[<c0158230>] __find_get_block+0xad/0x10c
[<c01a4654>] search_by_key+0x41c/0xc48
[<c013a63b>] mempool_free+0x7e/0x8e
[<c0208384>] as_update_arq+0x1e/0x5e
[<c0209125>] as_add_request+0x1a6/0x22d
[<c0209c78>] as_set_request+0x14/0x63
[<c0209c64>] as_set_request+0x0/0x63
[<c020053e>] __elv_add_request+0x90/0xc7
[<c0204023>] __make_request+0x3c6/0x46a
[<c01a4ee0>] search_for_position_by_key+0x60/0x37a
[<c02043a6>] generic_make_request+0x9f/0x218
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c01900a9>] make_cpu_key+0x4b/0x5b
[<c01a3ecd>] pathrelse+0x1e/0x2c
[<c01907c8>] _get_block_create_0+0x4c6/0x698
[<c0130f0f>] autoremove_wake_function+0x9/0x43
[<c01917e5>] reiserfs_get_block+0xbad/0x11b6
[<c02de1d0>] schedule+0x3e4/0xc34
[<c01197ef>] __wake_up+0x38/0x4e
[<c0108c48>] hypervisor_callback+0x2c/0x34
[<c0117c73>] try_to_wake_up+0x2a4/0x2f5
[<c0104866>] force_evtchn_callback+0xc/0xe
[<c02de203>] schedule+0x417/0xc34
[<c01358f1>] __do_IRQ+0x103/0x132
[<c0130f21>] autoremove_wake_function+0x1b/0x43
[<c0119797>] __wake_up_common+0x35/0x55
[<c010af8a>] monotonic_clock+0x55/0x95
[<c02de1d0>] schedule+0x3e4/0xc34
[<c01358f1>] __do_IRQ+0x103/0x132
[<c01197ef>] __wake_up+0x38/0x4e
[<c0118b08>] find_busiest_group+0x1bc/0x301
[<c0118e9b>] load_balance_newidle+0x30/0x8c
[<c010af8a>] monotonic_clock+0x55/0x95
[<c02de1d0>] schedule+0x3e4/0xc34
[<c02de203>] schedule+0x417/0xc34
[<c010af8a>] monotonic_clock+0x55/0x95
[<c02de203>] schedule+0x417/0xc34
[<c01197ef>] __wake_up+0x38/0x4e
[<c021f40e>] xs_input_avail+0x28/0x35
[<c021f59f>] xb_read+0x184/0x1c4
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c021f761>] read_reply+0x63/0xa2
[<c014d19d>] __get_vm_area+0x1b7/0x1ec
[<c02249eb>] map_frontend_pages+0x33/0x75
[<c0224a86>] netif_map+0x58/0x133
[<c022429e>] frontend_changed+0x190/0x222
[<c0220456>] watch_thread+0xf2/0x168
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c0130f06>] autoremove_wake_function+0x0/0x43
[<c01198d3>] complete+0x40/0x55
[<c0220364>] watch_thread+0x0/0x168
[<c0130a6e>] kthread+0xa0/0xd4
[<c01309ce>] kthread+0x0/0xd4
[<c0106c81>] kernel_thread_helper+0x5/0xb
Code: b8 2e c0 e8 74 70 fd ff 55 8d 0c 0a 57 56 53 89 d3 83 ec 24 39
ca 89 44 24 20 89 4c 24 1c 0f 83 19 01 00 00 8b 4c 24 20 c1 ea 16 <8b>
41 1c 83 c1 3c 8d 2c 90 89 c8 89 4c 24 10 e8 dd 96 19 00 8b
xend-debug.log has:
ERROR: do_evtchn_op: HYPERVISOR_event_channel_op failed: -1
ERROR: do_evtchn_op: HYPERVISOR_event_channel_op failed: -1
xend.log has:
2005-09-19 17:15:14 xend] DEBUG (image:164) initDomain: cpu=-1
mem_kb=131072 ssidref=0 dom=1
[2005-09-19 17:15:14 xend] DEBUG (XendDomainInfo:702) init_domain>
Created domain=1 name=cbc0 memory=128
[2005-09-19 17:15:14 xend] INFO (image:202) buildDomain os=linux dom=1 vcpus=1
[2005-09-19 17:15:14 xend] DEBUG (image:246) dom = 1
[2005-09-19 17:15:14 xend] DEBUG (image:247) image =
/boot_domU/vmlinuz-2.6.12.5-20050919-xenU
[2005-09-19 17:15:14 xend] DEBUG (image:248) store_evtchn = 1
[2005-09-19 17:15:14 xend] DEBUG (image:249) console_evtchn = 2
[2005-09-19 17:15:14 xend] DEBUG (image:250) cmdline =
ip=192.168.1.2:1.2.3.4::::eth0:off root=/dev/hda1 ro
[2005-09-19 17:15:14 xend] DEBUG (image:251) ramdisk =
[2005-09-19 17:15:14 xend] DEBUG (image:252) flags = 0
[2005-09-19 17:15:14 xend] DEBUG (image:253) vcpus = 1
[2005-09-19 17:15:15 xend] DEBUG (blkif:24) exception looking up
device number for hda1: [Errno 2] No such file or directory:
'/dev/hda1'
[2005-09-19 17:15:15 xend] DEBUG (DevController:181) DevController:
writing {'virtual-device': '769', 'backend-id': '0', 'backend':
'/domain/d5de7c20-2e56-4412-aeda-97c8d480c2f1/backend/vbd/0063c554-9062-4268-8dc9-704f4bc26a94/769'}
to /domain/0063c554-9062-4268-8dc9-704f4bc26a94/device/vbd/769.
[2005-09-19 17:15:15 xend] DEBUG (DevController:183) DevController:
writing {'params': 'vg/cbc0', 'domain': 'cbc0', 'type': 'phy',
'frontend': '/domain/0063c554-9062-4268-8dc9-704f4bc26a94/device/vbd/769',
'frontend-id': '1'} to
/domain/d5de7c20-2e56-4412-aeda-97c8d480c2f1/backend/vbd/0063c554-9062-4268-8dc9-704f4bc26a94/769.
[2005-09-19 17:15:15 xend] DEBUG (DevController:181) DevController:
writing {'backend-id': '0', 'mac': 'aa:00:00:04:9e:24', 'handle': '1',
'backend':
'/domain/d5de7c20-2e56-4412-aeda-97c8d480c2f1/backend/vif/0063c554-9062-4268-8dc9-704f4bc26a94/1'}
to /domain/0063c554-9062-4268-8dc9-704f4bc26a94/device/vif/1.
[2005-09-19 17:15:15 xend] DEBUG (DevController:183) DevController:
writing {'bridge': 'xen-br0', 'mac': 'aa:00:00:04:9e:24', 'handle':
'1', 'script': '/etc/xen/scripts/vif-bridge', 'frontend-id': '1',
'domain': 'cbc0', 'frontend':
'/domain/0063c554-9062-4268-8dc9-704f4bc26a94/device/vif/1'} to
/domain/d5de7c20-2e56-4412-aeda-97c8d480c2f1/backend/vif/0063c554-9062-4268-8dc9-704f4bc26a94/1.
[2005-09-19 17:15:15 xend] INFO (XendRoot:141) EVENT>
xend.domain.create ['cbc0', 1]
[2005-09-19 17:15:15 xend] INFO (XendRoot:141) EVENT>
xend.domain.unpause ['cbc0', 1]
[2005-09-19 17:15:25 xend] DEBUG (XendDomain:232) domain died name=cbc0 domid=1
[2005-09-19 17:15:25 xend] INFO (XendRoot:141) EVENT> xend.domain.died
['cbc0', None]
I've seen bugs #220, and #176 which both look kind of similar, though
176 is supposed to be fixed now, which is why I updated to todays
unstable.
Has anyone got a lvm2 backend working with current unstable?
Thanks,
Chris
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|