On Thu, Sep 30, 2010 at 9:27 AM, George Dunlap
<George.Dunlap@xxxxxxxxxxxxx> wrote:
> On Thu, Sep 30, 2010 at 1:28 PM, Yuehai Xu <yuehaixu@xxxxxxxxx> wrote:
>>> I agree, letting a VM with an interrupt run for a short period of time
>>> makes sense. The challenge is to make sure that it can't simply send
>>> itself interrupts every 50us and get to run 100% of the time. :-)
>>
>> I am afraid I don't really understand the challenge is, or, in another
>> word, this method is good principally, but in practice, it is hard to
>> implement? As I know, the OS should always schedules I/O related
>> processes once they are in runnable queue, so, as long as we give even
>> a very short period of time to the waken up guest VM, the I/O process
>> in it should be scheduled at once. In that case, this problem should
>> be solved. Of course, I don't do experiments, saying is always much
>> easier than doing.
>
> What I mean is that you have to be careful when implementing it. A
> very simple implementation would look like this:
> * Normally, let the VM with the highest credits run. However, if a VM
> is sent an interrupt, give it priority to run for 50us.
>
> Now, suppose, however, that a rogue VM sets up a periodic timer to
> send itself an interrupt every 55us. Then it will get an interrupt,
> get priority for 50us, be preempted for 5us, and then get another
> interrupt, allowing it to run for another 50us. Thus it runs 90% of
> the time, even though it should only run (for example) 50% of the
> time.
>
> We need a way to balance interrupt latency (how long after an
> interrupt is raised before a VM can run) and cpu scheduling fairness.
> That means that if we let a VM run for 50us, and then preempt it, and
> it gets an interrupt 5us later, we need a way to know not to schedule
> it until it's been off the cpu for a reasonable amount of time. It's
> possible, but it will take some experimentation to see what the best
> option is.
>
> -George
>
I'd like try to implement this idea to XEN, even though I am not sure
whether I can do it since I am not an expert. :-D.
The first step for me is to write a very simple scheduler without
considering CPU fairness, I/O performance, etc. Its mechanism is very
simple,
the selection of next VCPU is based on the algorithm of round robin.
The current VCPU is always inserted into the tail of the list while
the next
VCPU of the head is selected to be scheduled. The current test code is
basing on credit scheduler of XEN 4.0.1-rc6-pre, except that I delete
all
the component of credit calculation related, the tick of every 10ms,
30ms is also deleted. The time for the next VCPU which is selected is
set to 30ms.
Here, my pre-assumption is that Dom0 pins to PCPU0, while other DomU
pins to PCPU1 for simplicity.
However, some problems puzzle me a lot. When I start two DomU which
shares PCPU1, and in both of which I run a CPU intensive program,
the trace log from xenalyze is below(I modify some code so that the
format is different from the original):
...
< 0.399300204 -x d1v0> (dom: 1) --> (dom: 2) vruntime : 30000802)
< 0.424239058 -x d2v0> (dom: 2) --> (dom: 1) vruntime : 30000582)
< 0.449177708 -x d1v0> (dom: 1) --> (dom: 2) vruntime : 30000336)
< 0.474116762 -x d2v0> (dom: 2) --> (dom: 1) vruntime : 30000827)
< 0.499055641 -x d1v0> (dom: 1) --> (dom: 2) vruntime : 30000596)
< 0.523972987 |x d2v0> (dom: 2) --> (dom: 1) vruntime : 30001301)
< 0.548911095 -x d1v0> (dom: 1) --> (dom: 2) vruntime : 29999684)
...
I think these results make sense since every domU is using almost 30ms of PCPU1
However, I stop one of the CPU intensive program in a DomU while keep
the other running, the results are:
.....
< 0.327815345 |x d1v0> (dom: 1) --> (dom: 2) vruntime : 1542607)
< 0.327906620 -x d2v0> (dom: 2) --> (dom: 1) vruntime : 109521)
< 0.344349033 |x d1v0> (dom: 1) --> (dom: 2) vruntime : 19779544)
< 0.344377129 -x d2v0> (dom: 2) --> (dom: 1) vruntime : 33528)
< 0.344570662 |x d1v0> (dom: 1) --> (dom: 2) vruntime : 232540)
< 0.344643933 -x d2v0> (dom: 2) --> (dom: 1) vruntime : 87857)
< 0.345009170 |x d1v0> (dom: 1) --> (dom: 2) vruntime : 439081)
< 0.345034387 -x d2v0> (dom: 2) --> (dom: 1) vruntime : 30059)
< 0.369973183 -x d1v0> (dom: 1) --> (dom: 1) vruntime : 30000506)
< 0.392423279 |x d1v0> (dom: 1) --> (dom: 2) vruntime : 27006658)
....
Here I am gotten confusing, since my algorithm of scheduling is very
simple, every VM should have 30ms of PCPU, however, from the results,
the time for
each VCPU to have PCPU is quite unstable. I think somewhere, the
routine of schedule() should be invoked frequently, and from xentop,
the VM with CPU
intensive occupies PCPU almost at 97%.
Thanks,
Yuehai
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|