Re: [PATCH] sched: move enough load to balance average load per task

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Siddha, Suresh B wrote:
On Wed, Apr 12, 2006 at 09:46:32AM +1000, Peter Williams wrote:
Siddha, Suresh B wrote:
On Mon, Apr 10, 2006 at 04:45:32PM +1000, Peter Williams wrote:
Problem:

The current implementation of find_busiest_group() recognizes that approximately equal average loads per task for each group/queue are desirable (e.g. this condition will increase the probability that the top N highest priority tasks on an N CPU system will be on different CPUs) by being slightly more aggressive when *imbalance is small but the average load per task in "busiest" group is more than that in "this" group. Unfortunately, the amount moved from "busiest" to "this" is too small to reduce the average load per task on "busiest" (at best there will be no change and at worst it will get bigger).
Peter, We don't need to reduce the average load per task on "busiest"
always. By moving a "busiest_load_per_task", we will increase the average load per task of lesser busy cpu (there by trying to achieve
the equality with busiest...)
Well, first off, we don't always move busiest_load_per_task we move UP TO busiest_load_per_task so there is no way you can make definitive statements about what will happen to the value "this_load_per_task" as a result of setting *imbalance to busiest_load_per_task. Load balancing is a probabilistic endeavour and we need to take steps that increase the probability that we get the desired result.

I agree with you. But the previous code was more conservative and may slowly
(just from theory pt of view... I don't have an example to show this..)
balance towards the desired state. With this code, I feel we are
aggressive. for example, on a DP system: if I run one high priority
and two low priority processes, they keep hopping from one processor
to another... you may argue it is because of the "top" or some other
process... I agree that it is the case.. But same thing doesn't happen
with the previous version.. I like the conservative approach...

Without this patch there is no chance that busiest_load_per_task will get smaller

Is there an example for this?

Yes, we just take a slight variation of your scenario that prompted the first patch (to which this patch is a minor modification) by adding one normal priority task to each of the CPUs. This gives us a 2 CPU system with CPU-0 having 2 high priority tasks plus 1 normal priority task and CPU-1 having two normal priority tasks. Clearly, the desirable load balancing outcome would be for the two high priority tasks to be on different CPUs otherwise we have a high priority task stuck on a run queue while a normal priority is running on another (less heavily loaded) CPU.

In order to analyze what happens during load balancing, let's use W as the load weight for a normal task and suppose that the load weights of the two high priority tasks are (W + k) and that "this" == CPU-1 in find_busiest_queue(). This will result in "busiest" == CPU-0 and:

this_load = 2W
this_load_per_task = W
max_load = 3W + 2k
busiest_load_per_task = W + 2k / 3
avg_load = 5W / 2 + k
max_pull = W / 2 + k
*imbalance = W / 2 + k

Whenever k < (3W / 2) this will result in *imbalance < busiest_load_per_task and we end up in the small imbalance code.

(max_load - this_load) = W + 2k which is greater than busiest_load_per_task so we decide that we want to move some load from "busiest" to "this".

Without this patch we would set *imbalance to busiest_load_per_task and the only task on "busiest" that has a weighted load less than or equal to this value is the normal task so this is the one that will be moved resulting:

this_load = 3W
this_load_per_task = W
max_load = 2W + 2k
busiest_load_per_task = W + k

Even if you reverse the roles of "busiest" and "this", this will be considered balanced and the system will stabilize in this undesirable state. NB, as predicted, the average load per task on "this" hasn't changed and the average load per task on "busiest" has increased. We still have the situation where a high priority task is stuck on a run queue while a low priority task is running on another CPU -- we've failed :-(.

With this patch, *imbalance will be set to (W + 4k / 3) which is bigger than the weighted load of the high priority tasks so one of them will be moved resulting in:

this_load = 3W + k
this_load_per_task = W + k / 3
max_load = 2W + k
busiest_load_per_task = W + k / 2


and whether this_load_per_task will get bigger is indeterminate. With this patch there IS a chance that busiest_load_per_task will decrease and an INCREASED chance that this_load_per_task will get bigger. Ergo we have increased the probability that the (absolute) difference between this_load_per_task and busiest_load_per_task will decrease. This is a desirable outcome.

All I am saying is we are more aggressive.. I don't have any issue with
the desired outcome..

We need to be more aggressive but not too aggressive and I think this patch achieves the required balance.

NB busiest_load_per_task < *imbalance < (max_load - this_load) is true for this path through the code. To be precise, *imbalance will be half way between busiest_load_per_task and (max_load - this_load).

Peter
--
Peter Williams                                   [email protected]

"Learning, n. The kind of ignorance distinguishing the studious."
 -- Ambrose Bierce
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

[Index of Archives]     [Kernel Newbies]     [Netfilter]     [Bugtraq]     [Photo]     [Stuff]     [Gimp]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Video 4 Linux]     [Linux for the blind]     [Linux Resources]
  Powered by Linux