# HG changeset patch
# User Ian.Campbell@xxxxxxxxxxxxx
# Node ID b16252dbcb1f99a22fcf4c87a015c47f6019cfd2
# Parent c1850c659e40042d4ce2bf006d6f37b9ebda307a
# Parent 01043d5438786a7e023ca9236999d0f5bf919e9d
merge
---
tools/ioemu/patches/series | 1
tools/ioemu/patches/xen-support-buffered-ioreqs | 77 +++++++++++++++++++++
tools/ioemu/target-i386-dm/helper2.c | 88 +++++++++++++++++-------
tools/ioemu/vl.c | 8 +-
tools/libxc/xc_hvm_build.c | 31 ++++++++
xen/arch/x86/hvm/hvm.c | 31 +++++++-
xen/arch/x86/hvm/intercept.c | 66 +++++++++++++++++-
xen/arch/x86/hvm/platform.c | 2
xen/arch/x86/hvm/svm/svm.c | 3
xen/arch/x86/hvm/vmx/vmx.c | 3
xen/include/asm-x86/e820.h | 1
xen/include/asm-x86/hvm/domain.h | 2
xen/include/asm-x86/hvm/hvm.h | 4 -
xen/include/asm-x86/hvm/support.h | 1
xen/include/public/hvm/ioreq.h | 8 ++
15 files changed, 292 insertions(+), 34 deletions(-)
diff -r c1850c659e40 -r b16252dbcb1f tools/ioemu/patches/series
--- a/tools/ioemu/patches/series Wed Aug 09 11:27:28 2006 +0100
+++ b/tools/ioemu/patches/series Wed Aug 09 11:29:06 2006 +0100
@@ -39,3 +39,4 @@ qemu-allow-disable-sdl
qemu-allow-disable-sdl
qemu-fix-memset-args
qemu-fix-write-to-disk-synchronous
+xen-support-buffered-ioreqs
diff -r c1850c659e40 -r b16252dbcb1f tools/ioemu/target-i386-dm/helper2.c
--- a/tools/ioemu/target-i386-dm/helper2.c Wed Aug 09 11:27:28 2006 +0100
+++ b/tools/ioemu/target-i386-dm/helper2.c Wed Aug 09 11:29:06 2006 +0100
@@ -75,6 +75,10 @@ int xc_handle;
int xc_handle;
shared_iopage_t *shared_page = NULL;
+
+#define BUFFER_IO_MAX_DELAY 100
+buffered_iopage_t *buffered_io_page = NULL;
+QEMUTimer *buffered_io_timer;
/* the evtchn fd for polling */
int xce_handle = -1;
@@ -419,36 +423,68 @@ void cpu_ioreq_xor(CPUState *env, ioreq_
req->u.data = tmp1;
}
+void __handle_ioreq(CPUState *env, ioreq_t *req)
+{
+ if (!req->pdata_valid && req->dir == IOREQ_WRITE && req->size != 4)
+ req->u.data &= (1UL << (8 * req->size)) - 1;
+
+ switch (req->type) {
+ case IOREQ_TYPE_PIO:
+ cpu_ioreq_pio(env, req);
+ break;
+ case IOREQ_TYPE_COPY:
+ cpu_ioreq_move(env, req);
+ break;
+ case IOREQ_TYPE_AND:
+ cpu_ioreq_and(env, req);
+ break;
+ case IOREQ_TYPE_OR:
+ cpu_ioreq_or(env, req);
+ break;
+ case IOREQ_TYPE_XOR:
+ cpu_ioreq_xor(env, req);
+ break;
+ default:
+ hw_error("Invalid ioreq type 0x%x\n", req->type);
+ }
+}
+
+void __handle_buffered_iopage(CPUState *env)
+{
+ ioreq_t *req = NULL;
+
+ if (!buffered_io_page)
+ return;
+
+ while (buffered_io_page->read_pointer !=
+ buffered_io_page->write_pointer) {
+ req = &buffered_io_page->ioreq[buffered_io_page->read_pointer %
+ IOREQ_BUFFER_SLOT_NUM];
+
+ __handle_ioreq(env, req);
+
+ mb();
+ buffered_io_page->read_pointer++;
+ }
+}
+
+void handle_buffered_io(void *opaque)
+{
+ CPUState *env = opaque;
+
+ __handle_buffered_iopage(env);
+ qemu_mod_timer(buffered_io_timer, BUFFER_IO_MAX_DELAY +
+ qemu_get_clock(rt_clock));
+}
+
void cpu_handle_ioreq(void *opaque)
{
CPUState *env = opaque;
ioreq_t *req = cpu_get_ioreq();
+ handle_buffered_io(env);
if (req) {
- if ((!req->pdata_valid) && (req->dir == IOREQ_WRITE)) {
- if (req->size != 4)
- req->u.data &= (1UL << (8 * req->size))-1;
- }
-
- switch (req->type) {
- case IOREQ_TYPE_PIO:
- cpu_ioreq_pio(env, req);
- break;
- case IOREQ_TYPE_COPY:
- cpu_ioreq_move(env, req);
- break;
- case IOREQ_TYPE_AND:
- cpu_ioreq_and(env, req);
- break;
- case IOREQ_TYPE_OR:
- cpu_ioreq_or(env, req);
- break;
- case IOREQ_TYPE_XOR:
- cpu_ioreq_xor(env, req);
- break;
- default:
- hw_error("Invalid ioreq type 0x%x\n", req->type);
- }
+ __handle_ioreq(env, req);
/* No state change if state = STATE_IORESP_HOOK */
if (req->state == STATE_IOREQ_INPROCESS) {
@@ -465,6 +501,10 @@ int main_loop(void)
extern int shutdown_requested;
CPUState *env = cpu_single_env;
int evtchn_fd = xc_evtchn_fd(xce_handle);
+
+ buffered_io_timer = qemu_new_timer(rt_clock, handle_buffered_io,
+ cpu_single_env);
+ qemu_mod_timer(buffered_io_timer, qemu_get_clock(rt_clock));
qemu_set_fd_handler(evtchn_fd, cpu_handle_ioreq, NULL, env);
diff -r c1850c659e40 -r b16252dbcb1f tools/ioemu/vl.c
--- a/tools/ioemu/vl.c Wed Aug 09 11:27:28 2006 +0100
+++ b/tools/ioemu/vl.c Wed Aug 09 11:29:06 2006 +0100
@@ -5834,6 +5834,7 @@ int main(int argc, char **argv)
unsigned long nr_pages;
xen_pfn_t *page_array;
extern void *shared_page;
+ extern void *buffered_io_page;
char qemu_dm_logfilename[64];
@@ -6378,11 +6379,16 @@ int main(int argc, char **argv)
phys_ram_base = xc_map_foreign_batch(xc_handle, domid,
PROT_READ|PROT_WRITE, page_array,
- nr_pages - 1);
+ nr_pages - 3);
if (phys_ram_base == 0) {
fprintf(logfile, "xc_map_foreign_batch returned error %d\n", errno);
exit(-1);
}
+
+ /* not yet add for IA64 */
+ buffered_io_page = xc_map_foreign_range(xc_handle, domid, PAGE_SIZE,
+ PROT_READ|PROT_WRITE,
+ page_array[nr_pages - 3]);
shared_page = xc_map_foreign_range(xc_handle, domid, PAGE_SIZE,
PROT_READ|PROT_WRITE,
diff -r c1850c659e40 -r b16252dbcb1f tools/libxc/xc_hvm_build.c
--- a/tools/libxc/xc_hvm_build.c Wed Aug 09 11:27:28 2006 +0100
+++ b/tools/libxc/xc_hvm_build.c Wed Aug 09 11:29:06 2006 +0100
@@ -26,6 +26,7 @@
#define E820_IO 16
#define E820_SHARED_PAGE 17
#define E820_XENSTORE 18
+#define E820_BUFFERED_IO 19
#define E820_MAP_PAGE 0x00090000
#define E820_MAP_NR_OFFSET 0x000001E8
@@ -96,7 +97,13 @@ static void build_e820map(void *e820_pag
e820entry[nr_map].type = E820_RESERVED;
nr_map++;
-#define STATIC_PAGES 2 /* for ioreq_t and store_mfn */
+#define STATIC_PAGES 3
+ /* 3 static pages:
+ * - ioreq buffer.
+ * - xenstore.
+ * - shared_page.
+ */
+
/* Most of the ram goes here */
e820entry[nr_map].addr = 0x100000;
e820entry[nr_map].size = mem_size - 0x100000 - STATIC_PAGES * PAGE_SIZE;
@@ -104,6 +111,12 @@ static void build_e820map(void *e820_pag
nr_map++;
/* Statically allocated special pages */
+
+ /* For buffered IO requests */
+ e820entry[nr_map].addr = mem_size - 3 * PAGE_SIZE;
+ e820entry[nr_map].size = PAGE_SIZE;
+ e820entry[nr_map].type = E820_BUFFERED_IO;
+ nr_map++;
/* For xenstore */
e820entry[nr_map].addr = mem_size - 2 * PAGE_SIZE;
@@ -213,6 +226,9 @@ static int setup_guest(int xc_handle,
unsigned long shared_page_frame = 0;
shared_iopage_t *sp;
+ unsigned long ioreq_buffer_frame = 0;
+ void *ioreq_buffer_page;
+
memset(&dsi, 0, sizeof(struct domain_setup_info));
if ( (parseelfimage(image, image_size, &dsi)) != 0 )
@@ -302,6 +318,19 @@ static int setup_guest(int xc_handle,
goto error_out;
memset(sp, 0, PAGE_SIZE);
munmap(sp, PAGE_SIZE);
+
+ /* clean the buffered IO requests page */
+ ioreq_buffer_frame = page_array[(v_end >> PAGE_SHIFT) - 3];
+ ioreq_buffer_page = xc_map_foreign_range(xc_handle, dom, PAGE_SIZE,
+ PROT_READ | PROT_WRITE,
+ ioreq_buffer_frame);
+
+ if ( ioreq_buffer_page == NULL )
+ goto error_out;
+
+ memset(ioreq_buffer_page, 0, PAGE_SIZE);
+
+ munmap(ioreq_buffer_page, PAGE_SIZE);
xc_set_hvm_param(xc_handle, dom, HVM_PARAM_STORE_PFN, (v_end >>
PAGE_SHIFT) - 2);
xc_set_hvm_param(xc_handle, dom, HVM_PARAM_STORE_EVTCHN, store_evtchn);
diff -r c1850c659e40 -r b16252dbcb1f xen/arch/x86/hvm/hvm.c
--- a/xen/arch/x86/hvm/hvm.c Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/arch/x86/hvm/hvm.c Wed Aug 09 11:29:06 2006 +0100
@@ -134,15 +134,28 @@ static void e820_map_io_shared_callback(
}
}
-void hvm_map_io_shared_page(struct vcpu *v)
-{
- unsigned long mfn = INVALID_MFN;
+static void e820_map_buffered_io_callback(struct domain *d,
+ struct e820entry *e,
+ void *data)
+{
+ unsigned long *mfn = data;
+ if ( e->type == E820_BUFFERED_IO ) {
+ ASSERT(*mfn == INVALID_MFN);
+ *mfn = gmfn_to_mfn(d, e->addr >> PAGE_SHIFT);
+ }
+}
+
+void hvm_map_io_shared_pages(struct vcpu *v)
+{
+ unsigned long mfn;
void *p;
struct domain *d = v->domain;
- if ( d->arch.hvm_domain.shared_page_va )
+ if ( d->arch.hvm_domain.shared_page_va ||
+ d->arch.hvm_domain.buffered_io_va )
return;
+ mfn = INVALID_MFN;
e820_foreach(d, e820_map_io_shared_callback, &mfn);
if ( mfn == INVALID_MFN )
@@ -159,6 +172,14 @@ void hvm_map_io_shared_page(struct vcpu
}
d->arch.hvm_domain.shared_page_va = (unsigned long)p;
+
+ mfn = INVALID_MFN;
+ e820_foreach(d, e820_map_buffered_io_callback, &mfn);
+ if ( mfn != INVALID_MFN ) {
+ p = map_domain_page_global(mfn);
+ if ( p )
+ d->arch.hvm_domain.buffered_io_va = (unsigned long)p;
+ }
}
void hvm_create_event_channels(struct vcpu *v)
@@ -209,6 +230,8 @@ void hvm_setup_platform(struct domain* d
spin_lock_init(&d->arch.hvm_domain.round_robin_lock);
hvm_vioapic_init(d);
}
+
+ spin_lock_init(&d->arch.hvm_domain.buffered_io_lock);
init_timer(&platform->pl_time.periodic_tm.timer,
pt_timer_fn, v, v->processor);
diff -r c1850c659e40 -r b16252dbcb1f xen/arch/x86/hvm/intercept.c
--- a/xen/arch/x86/hvm/intercept.c Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/arch/x86/hvm/intercept.c Wed Aug 09 11:29:06 2006 +0100
@@ -36,10 +36,24 @@ extern struct hvm_mmio_handler vioapic_m
#define HVM_MMIO_HANDLER_NR 2
-struct hvm_mmio_handler *hvm_mmio_handlers[HVM_MMIO_HANDLER_NR] =
+static struct hvm_mmio_handler *hvm_mmio_handlers[HVM_MMIO_HANDLER_NR] =
{
&vlapic_mmio_handler,
&vioapic_mmio_handler
+};
+
+struct hvm_buffered_io_range {
+ unsigned long start_addr;
+ unsigned long length;
+};
+
+#define HVM_BUFFERED_IO_RANGE_NR 1
+
+static struct hvm_buffered_io_range buffered_stdvga_range = {0xA0000, 0x20000};
+static struct hvm_buffered_io_range
+*hvm_buffered_io_ranges[HVM_BUFFERED_IO_RANGE_NR] =
+{
+ &buffered_stdvga_range
};
static inline void hvm_mmio_access(struct vcpu *v,
@@ -140,6 +154,56 @@ static inline void hvm_mmio_access(struc
}
}
+int hvm_buffered_io_intercept(ioreq_t *p)
+{
+ struct vcpu *v = current;
+ spinlock_t *buffered_io_lock;
+ buffered_iopage_t *buffered_iopage =
+ (buffered_iopage_t *)(v->domain->arch.hvm_domain.buffered_io_va);
+ unsigned long tmp_write_pointer = 0;
+ int i;
+
+ /* ignore READ ioreq_t! */
+ if ( p->dir == IOREQ_READ )
+ return 0;
+
+ for ( i = 0; i < HVM_BUFFERED_IO_RANGE_NR; i++ ) {
+ if ( p->addr >= hvm_buffered_io_ranges[i]->start_addr &&
+ p->addr + p->size - 1 < hvm_buffered_io_ranges[i]->start_addr +
+ hvm_buffered_io_ranges[i]->length )
+ break;
+ }
+
+ if ( i == HVM_BUFFERED_IO_RANGE_NR )
+ return 0;
+
+ buffered_io_lock = &v->domain->arch.hvm_domain.buffered_io_lock;
+ spin_lock(buffered_io_lock);
+
+ if ( buffered_iopage->write_pointer - buffered_iopage->read_pointer ==
+ (unsigned long)IOREQ_BUFFER_SLOT_NUM ) {
+ /* the queue is full.
+ * send the iopacket through the normal path.
+ * NOTE: The arithimetic operation could handle the situation for
+ * write_pointer overflow.
+ */
+ spin_unlock(buffered_io_lock);
+ return 0;
+ }
+
+ tmp_write_pointer = buffered_iopage->write_pointer % IOREQ_BUFFER_SLOT_NUM;
+
+ memcpy(&buffered_iopage->ioreq[tmp_write_pointer], p, sizeof(ioreq_t));
+
+ /*make the ioreq_t visible before write_pointer*/
+ wmb();
+ buffered_iopage->write_pointer++;
+
+ spin_unlock(buffered_io_lock);
+
+ return 1;
+}
+
int hvm_mmio_intercept(ioreq_t *p)
{
struct vcpu *v = current;
diff -r c1850c659e40 -r b16252dbcb1f xen/arch/x86/hvm/platform.c
--- a/xen/arch/x86/hvm/platform.c Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/arch/x86/hvm/platform.c Wed Aug 09 11:29:06 2006 +0100
@@ -779,7 +779,7 @@ void send_mmio_req(
} else
p->u.data = value;
- if (hvm_mmio_intercept(p)){
+ if ( hvm_mmio_intercept(p) || hvm_buffered_io_intercept(p) ) {
p->state = STATE_IORESP_READY;
hvm_io_assist(v);
return;
diff -r c1850c659e40 -r b16252dbcb1f xen/arch/x86/hvm/svm/svm.c
--- a/xen/arch/x86/hvm/svm/svm.c Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/arch/x86/hvm/svm/svm.c Wed Aug 09 11:29:06 2006 +0100
@@ -809,6 +809,9 @@ static void svm_relinquish_guest_resourc
if ( d->arch.hvm_domain.shared_page_va )
unmap_domain_page_global(
(void *)d->arch.hvm_domain.shared_page_va);
+
+ if ( d->arch.hvm_domain.buffered_io_va )
+ unmap_domain_page_global((void *)d->arch.hvm_domain.buffered_io_va);
shadow_direct_map_clean(d);
}
diff -r c1850c659e40 -r b16252dbcb1f xen/arch/x86/hvm/vmx/vmx.c
--- a/xen/arch/x86/hvm/vmx/vmx.c Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/arch/x86/hvm/vmx/vmx.c Wed Aug 09 11:29:06 2006 +0100
@@ -150,6 +150,9 @@ static void vmx_relinquish_guest_resourc
if ( d->arch.hvm_domain.shared_page_va )
unmap_domain_page_global(
(void *)d->arch.hvm_domain.shared_page_va);
+
+ if ( d->arch.hvm_domain.buffered_io_va )
+ unmap_domain_page_global((void *)d->arch.hvm_domain.buffered_io_va);
shadow_direct_map_clean(d);
}
diff -r c1850c659e40 -r b16252dbcb1f xen/include/asm-x86/e820.h
--- a/xen/include/asm-x86/e820.h Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/include/asm-x86/e820.h Wed Aug 09 11:29:06 2006 +0100
@@ -12,6 +12,7 @@
#define E820_IO 16
#define E820_SHARED_PAGE 17
#define E820_XENSTORE 18
+#define E820_BUFFERED_IO 19
#define E820_MAP_PAGE 0x00090000
#define E820_MAP_NR_OFFSET 0x000001E8
diff -r c1850c659e40 -r b16252dbcb1f xen/include/asm-x86/hvm/domain.h
--- a/xen/include/asm-x86/hvm/domain.h Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/include/asm-x86/hvm/domain.h Wed Aug 09 11:29:06 2006 +0100
@@ -33,6 +33,8 @@
struct hvm_domain {
unsigned long shared_page_va;
+ unsigned long buffered_io_va;
+ spinlock_t buffered_io_lock;
s64 tsc_frequency;
struct pl_time pl_time;
diff -r c1850c659e40 -r b16252dbcb1f xen/include/asm-x86/hvm/hvm.h
--- a/xen/include/asm-x86/hvm/hvm.h Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/include/asm-x86/hvm/hvm.h Wed Aug 09 11:29:06 2006 +0100
@@ -78,7 +78,7 @@ hvm_disable(void)
}
void hvm_create_event_channels(struct vcpu *v);
-void hvm_map_io_shared_page(struct vcpu *v);
+void hvm_map_io_shared_pages(struct vcpu *v);
static inline int
hvm_initialize_guest_resources(struct vcpu *v)
@@ -87,7 +87,7 @@ hvm_initialize_guest_resources(struct vc
if ( hvm_funcs.initialize_guest_resources )
ret = hvm_funcs.initialize_guest_resources(v);
if ( ret == 1 ) {
- hvm_map_io_shared_page(v);
+ hvm_map_io_shared_pages(v);
hvm_create_event_channels(v);
}
return ret;
diff -r c1850c659e40 -r b16252dbcb1f xen/include/asm-x86/hvm/support.h
--- a/xen/include/asm-x86/hvm/support.h Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/include/asm-x86/hvm/support.h Wed Aug 09 11:29:06 2006 +0100
@@ -139,6 +139,7 @@ extern void hvm_setup_platform(struct do
extern void hvm_setup_platform(struct domain* d);
extern int hvm_mmio_intercept(ioreq_t *p);
extern int hvm_io_intercept(ioreq_t *p, int type);
+extern int hvm_buffered_io_intercept(ioreq_t *p);
extern void hvm_hooks_assist(struct vcpu *v);
extern void hvm_print_line(struct vcpu *v, const char c);
extern void hlt_timer_fn(void *data);
diff -r c1850c659e40 -r b16252dbcb1f xen/include/public/hvm/ioreq.h
--- a/xen/include/public/hvm/ioreq.h Wed Aug 09 11:27:28 2006 +0100
+++ b/xen/include/public/hvm/ioreq.h Wed Aug 09 11:29:06 2006 +0100
@@ -78,6 +78,14 @@ struct shared_iopage {
};
typedef struct shared_iopage shared_iopage_t;
+#define IOREQ_BUFFER_SLOT_NUM 80
+struct buffered_iopage {
+ unsigned long read_pointer;
+ unsigned long write_pointer;
+ ioreq_t ioreq[IOREQ_BUFFER_SLOT_NUM];
+}; /* sizeof this structure must be in one page */
+typedef struct buffered_iopage buffered_iopage_t;
+
#endif /* _IOREQ_H_ */
/*
diff -r c1850c659e40 -r b16252dbcb1f
tools/ioemu/patches/xen-support-buffered-ioreqs
--- /dev/null Thu Jan 01 00:00:00 1970 +0000
+++ b/tools/ioemu/patches/xen-support-buffered-ioreqs Wed Aug 09 11:29:06
2006 +0100
@@ -0,0 +1,150 @@
+Index: ioemu/vl.c
+===================================================================
+--- ioemu.orig/vl.c 2006-08-08 14:33:30.000000000 +0100
++++ ioemu/vl.c 2006-08-08 14:43:34.000000000 +0100
+@@ -5834,6 +5834,7 @@
+ unsigned long nr_pages;
+ xen_pfn_t *page_array;
+ extern void *shared_page;
++ extern void *buffered_io_page;
+
+ char qemu_dm_logfilename[64];
+
+@@ -6378,12 +6379,17 @@
+
+ phys_ram_base = xc_map_foreign_batch(xc_handle, domid,
+ PROT_READ|PROT_WRITE, page_array,
+- nr_pages - 1);
++ nr_pages - 3);
+ if (phys_ram_base == 0) {
+ fprintf(logfile, "xc_map_foreign_batch returned error %d\n", errno);
+ exit(-1);
+ }
+
++ /* not yet add for IA64 */
++ buffered_io_page = xc_map_foreign_range(xc_handle, domid, PAGE_SIZE,
++ PROT_READ|PROT_WRITE,
++ page_array[nr_pages - 3]);
++
+ shared_page = xc_map_foreign_range(xc_handle, domid, PAGE_SIZE,
+ PROT_READ|PROT_WRITE,
+ page_array[nr_pages - 1]);
+Index: ioemu/target-i386-dm/helper2.c
+===================================================================
+--- ioemu.orig/target-i386-dm/helper2.c 2006-08-08 14:33:30.000000000
+0100
++++ ioemu/target-i386-dm/helper2.c 2006-08-09 10:03:40.558744653 +0100
+@@ -76,6 +76,10 @@
+
+ shared_iopage_t *shared_page = NULL;
+
++#define BUFFER_IO_MAX_DELAY 100
++buffered_iopage_t *buffered_io_page = NULL;
++QEMUTimer *buffered_io_timer;
++
+ /* the evtchn fd for polling */
+ int xce_handle = -1;
+
+@@ -419,36 +423,68 @@
+ req->u.data = tmp1;
+ }
+
++void __handle_ioreq(CPUState *env, ioreq_t *req)
++{
++ if (!req->pdata_valid && req->dir == IOREQ_WRITE && req->size != 4)
++ req->u.data &= (1UL << (8 * req->size)) - 1;
++
++ switch (req->type) {
++ case IOREQ_TYPE_PIO:
++ cpu_ioreq_pio(env, req);
++ break;
++ case IOREQ_TYPE_COPY:
++ cpu_ioreq_move(env, req);
++ break;
++ case IOREQ_TYPE_AND:
++ cpu_ioreq_and(env, req);
++ break;
++ case IOREQ_TYPE_OR:
++ cpu_ioreq_or(env, req);
++ break;
++ case IOREQ_TYPE_XOR:
++ cpu_ioreq_xor(env, req);
++ break;
++ default:
++ hw_error("Invalid ioreq type 0x%x\n", req->type);
++ }
++}
++
++void __handle_buffered_iopage(CPUState *env)
++{
++ ioreq_t *req = NULL;
++
++ if (!buffered_io_page)
++ return;
++
++ while (buffered_io_page->read_pointer !=
++ buffered_io_page->write_pointer) {
++ req = &buffered_io_page->ioreq[buffered_io_page->read_pointer %
++ IOREQ_BUFFER_SLOT_NUM];
++
++ __handle_ioreq(env, req);
++
++ mb();
++ buffered_io_page->read_pointer++;
++ }
++}
++
++void handle_buffered_io(void *opaque)
++{
++ CPUState *env = opaque;
++
++ __handle_buffered_iopage(env);
++ qemu_mod_timer(buffered_io_timer, BUFFER_IO_MAX_DELAY +
++ qemu_get_clock(rt_clock));
++}
++
+ void cpu_handle_ioreq(void *opaque)
+ {
+ CPUState *env = opaque;
+ ioreq_t *req = cpu_get_ioreq();
+
++ handle_buffered_io(env);
+ if (req) {
+- if ((!req->pdata_valid) && (req->dir == IOREQ_WRITE)) {
+- if (req->size != 4)
+- req->u.data &= (1UL << (8 * req->size))-1;
+- }
+-
+- switch (req->type) {
+- case IOREQ_TYPE_PIO:
+- cpu_ioreq_pio(env, req);
+- break;
+- case IOREQ_TYPE_COPY:
+- cpu_ioreq_move(env, req);
+- break;
+- case IOREQ_TYPE_AND:
+- cpu_ioreq_and(env, req);
+- break;
+- case IOREQ_TYPE_OR:
+- cpu_ioreq_or(env, req);
+- break;
+- case IOREQ_TYPE_XOR:
+- cpu_ioreq_xor(env, req);
+- break;
+- default:
+- hw_error("Invalid ioreq type 0x%x\n", req->type);
+- }
++ __handle_ioreq(env, req);
+
+ /* No state change if state = STATE_IORESP_HOOK */
+ if (req->state == STATE_IOREQ_INPROCESS) {
+@@ -466,6 +502,10 @@
+ CPUState *env = cpu_single_env;
+ int evtchn_fd = xc_evtchn_fd(xce_handle);
+
++ buffered_io_timer = qemu_new_timer(rt_clock, handle_buffered_io,
++ cpu_single_env);
++ qemu_mod_timer(buffered_io_timer, qemu_get_clock(rt_clock));
++
+ qemu_set_fd_handler(evtchn_fd, cpu_handle_ioreq, NULL, env);
+
+ env->send_event = 0;
_______________________________________________
Xen-changelog mailing list
Xen-changelog@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-changelog
|