Re: [rfc][patch] sched: remove smpnice

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Siddha, Suresh B wrote:
On Wed, Feb 15, 2006 at 09:40:32AM +1100, Peter Williams wrote:

Siddha, Suresh B wrote:

On a 4P(8-way with HT), if you run a -20 task(a simple infinite loop)
it hops from one processor to another processor... you can observe it
using top.

How do you get top to display which CPU tasks are running on?


In the interactive mode, you can select the "last used cpu" field to display.
or you can use /proc/pid/stat


find_busiest_group() thinks there is an imbalance and ultimately the
idle cpu kicks active load balance on busy cpu, resulting in the hopping.

I'm still having trouble getting my head around this. A task shouldn't be moved unless there's at least one other task on its current CPU, it


Because of the highest priority task, weighted load of that cpu
will be > SCHED_LOAD_SCALE. Because of this, an idle cpu in find_busiest_group() thinks that there is an imbalance.. This is due to the code near the comment "however we may be able to increase total CPU power used by ...". That piece of code assumes that a unit load
is represented by SCHED_LOAD_SCALE (which is no longer true with smpnice)
and finally results in "pwr_move > pwr_now".. This will make the idle cpu
try to pull that process from busiest cpu and the process will ultimately move
with the help of active load balance...

But as I pointed out, with the code as it is in the recent -mm kernels, the amount of biased load (i.e. NICE_TO_BIAS_PRIO(0)) that find_busiest_group() sets *imbalance to in these circumstances is too small for a task with nice less than zero to be moved i.e. move_tasks() will skip it. Or are you just referring to the vanilla kernels?



I agree with you.. But lets take a DP system with HT, now if there are
only two low priority tasks running, ideally we should be running them
on two different packages. With this patch, we may end up running on the
same logical processor.. leave alone running on the same package..
As these are low priority tasks, it might be ok.. But...

Yes, this is an issue but it's not restricted to HT systems (except for


Agreed.


the same package part). The latest patch that I've posted addresses (part of) this problem by replacing SCHED_LOAD_SCALE with the average load per runnable task in the code at the end of find_busiest_group() which handles the case where imbalance is small. This should enable load balancing to occur even if all runnable tasks are low priority.


Yes. We need to fix the code I mentioned above too.... And please make sure
it doesn't break HT optimizations as this piece of code is mainly used
for implementing HT optimizations..

OK, but I wouldn't call them optimizations (to say something is optimal is a big call :-)) as the HT code is more about enforcing "nice" (at the expense of throughput) than it is anything else. By the way, I think that one avenue that should be considered for HT systems is to have a single run queue for each package and feed the logical CPUs in the package from that. This would greatly improve the likelihood that the tasks selected to run on the logical CPUs within the package are of approximately the same priority and, consequently reduce the need to idle one of them. The downside of this is the extra complexity that it would introduce.

Peter
--
Peter Williams                                   [email protected]

"Learning, n. The kind of ignorance distinguishing the studious."
 -- Ambrose Bierce
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

[Index of Archives]     [Kernel Newbies]     [Netfilter]     [Bugtraq]     [Photo]     [Stuff]     [Gimp]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Video 4 Linux]     [Linux for the blind]     [Linux Resources]
  Powered by Linux