Paul Jackson wrote:
Nick wrote:
Did you send resend the patch to remove sched-domain partitioning?
After clearing up my confusion, IMO that is needed and could probably
go into 2.6.19.
The patch titled
cpuset: remove sched domain hooks from cpusets
went into *-mm on Friday, 20 Oct.
Is that the patch you mean?
Yes.
It's just the first step - unplugging the old.
Now we need the new:
1) Ability at runtime to isolate cpus for real-time and such.
2) Big systems perform better if we can avoid load balancing across
zillions of cpus.
These are both part of the same larger solution, which is to
partition domains. isolated CPUs are just the case of 1 CPU in
its own domain (and that's how they are implemented now).
So we need the interface or some driver to do this partitioning.
A cool option would be to determine the partitions according to the
disjoint set of unions of cpus_allowed masks of all tasks. I see this
getting computationally expensive though, probably O(tasks*CPUs)... I
guess that isn't too bad.
Yeah - if that would work, from the practical perspective of providing
us with a useful partitioning (get those humongous sched domains carved
down to a reasonable size) then that would be cool.
I'm guessing that in practice, it would be annoying to use. One would
end up with stray tasks that happened to be sitting in one of the bigger
cpusets and that did not have their cpus_allowed narrowed, stopping us
from getting a useful partitioning. Perhaps anilliary tasks associated
with the batch scheduler, or some paused tasks in an inactive job that
were it active would need load balancing across a big swath of cpus.
These would be tasks that we really didn't need to load balance, but they
would appear as if they needed it because of their fat cpus_allowed.
But we simply can't make a partition for them because they have asked
to use all CPUs. We can't know if this is something that should be
partitioned or not, can we?
Users (admins) would have to hunt down these tasks that were getting in
the way of a nice partitioning and whack their cpus_allowed down to
size.
In the sense that they get what they ask for, yes. The obvious route
for the big SGI systems is to put them in the right cpuset. The job
managers (or admins) on those things are surely up to the task.
This leaves unbound and transient kernel threads like pdflush as a
remaining problem. Not quite sure what to do about that yet. I see
you have a little hack in there...
So essentially, one would end up with another userspace API, backdoor
again. Like those magic doors in the libraries of wealthy protagonists
in mystery novels, where you have to open a particular book and pull
the lamp cord to get the door to appear and open.
Automatic chokes and transmissions are great - if they work. If not,
give me a knob and a stick.
But your knob is just going to be some mechanism to say that you don't
care about such and such a task, or you want to put task x into domain
y.
===
Another idea for a cpuset-based API to this ...
From our internal perspective, it's all about getting the sched domain
partitions cut down to a reasonable size, for performance reasons.
But from the users perspective, the deal we are asking them to
consider is to trade in fully automatic, all tasks across all cpus,
load balancing, in turn for better performance.
Big system admins would often be quite happy to mark the top cpuset
as "no need to load balance tasks in this cpuset." They would
take responsibility for moving any non-trivial, unpinned tasks into
lower cpusets (or not be upset if something left behind wasn't load
balancing.)
And the batch scheduler would be quite happy to mark its top cpuset as
"no need to load balance". It could mark any cpusets holding inactive
jobs the same way.
This "no need to load balance" flag would be advisory. The kernel
might load balance anyway. For example if the batch scheduler were
running under a top cpuset that was -not- so marked, we'd still have
to load balance everyone. The batch scheduler wouldn't care. It would
have done its duty, to mark which of its cpusets didn't need balancing.
All we need from them is the ok to not load balance certain cpusets,
and the rest is easy enough. If they give us such ok on enough of the
big cpusets, we give back a nice performance improvement.
I think this is much more of an automatic behind your back thing. If
they don't want any load balancing to happen, they could pin the tasks
in that cpuset to the cpu they're currently on, for example.
It would be trivial to make such a script to parse the root cpuset and
do exactly this, wouldn't it?
--
SUSE Labs, Novell Inc.
Send instant messages to your online friends http://au.messenger.yahoo.com
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]