David wrote:
> From a standpoint of the MPOL_PREFERRED memory policy itself, there
> is no documented behavior or standard that specifies its interaction
> with cpusets. Thus, it's "undefined." We are completely free
> to implement an undefined behavior as we choose and change it as
> Linux matures.
You state this point clearly, but I have to disagree.
The Linux documentation is not a legal contract. Anytime we change the
actual behaviour of the code, we have to ask ourselves what will be the
impact of that change on existing users and usages. The burden is on
us to minimize breaking things (by that I mean, what users would
consider breakage, even if we think it is all for the better and that
their code was the real problem.) I didn't say no breakage, but
minimum breakage, doing our best to guide users through changes with
minimum disruption to their work.
Linux is gaining market share rapidly because we co-operate with our
users to give us both the best chance of succeeding.
We don't just play gotcha games with the documentation -- ha ha --
we didn't document that detail, so it's your fault for ever depending
on it. And besides your code sucks. So there! Let's leave that game
for others.
> And you're trying to protect this application that based this
> implementation not on a standard or documentation, but on its observed
> behavior.
Yes, if there is such an application, I'm trying to protect it.
> My bet is that it's going to issue that subsequent
> set_mempolicy(), at least if libnuma returned a numa_preferred() value
> that it wasn't expecting.
Perhaps. Perhaps not. I don't know.
> That still requires a change to the application.
If that were so, then yes much of your subsequent reasoning would follow.
However, let me repeat the following from my previous message:
> However, in deference to the needs of libnuma, if the following
> call was made, this would change the mode for that task to
> Choice A:
>
> get_mempolicy(NULL, NULL, 0, 0, 0);
>
> This last detail above is an admitted hack. *So far as I know*
> it happens that all current infield versions of libnuma issue the
> above call, as their first mempolicy query, to detemine whether
> the active kernel supports mempolicy.
The above is the hack that allows us to support existing libnuma based
applications (the most significant users of memory policy historically)
with a default of Choice A, while other code and future code defaults
to Choice B.
> See? The problem is that you're trying to protect applications that know
> its initial cpuset mems [the only way it could ever send a
> set_mempolicy(MPOL_PREFERRED) for the right node in that range in the
> first place] but then seemingly loses control over its cpuset and intends
> for the kernel to fix it up for it without having the burden of issuing
> another set_mempolicy() call.
That's not the only sort of application I'm trying to protect.
I'm trying to protect almost any application that uses both
set_mempolicy or mbind, while in a cpuset.
If a task is in a cpuset on say nodes 16-23, and it wants to issue
any mbind, or any MPOL_PREFERRED, MPOL_BIND, or MPOL_INTERLEAVE
mempolicy call, then under Choice A it must issue nodemasks offset
by 16, relative to what it would issue under Choice B.
Almost any task using memory policies on a system making active use of
cpusets will be affected, even well written ones doing simple things.
I am more concerned that the above hack for libnuma isn't enough,
rather than it is unnecessary.
I think the above hack covers existing libnuma users rather well,
though I could be wrong even here, as I don't actually work with
most of the existing libnuma users.
And I can cover those using both memory policies and cpusets via
libcpuset, as there I am the expert and know that I can guide
libcpuset and its users through this change. There will be some
breakage, but I know how to manage it.
However, if anyone is deploying product or has important (to them) but
not easy to change software using both memory policies and cpusets that
are not doing so via libnuma or libcpuset, then any change that
changes the default for their memory policy calls from Choice A to
Choice B will probably break them.
Perhaps there are no significant cases like this, using memory policies
on cpuset managed systems, but not via libnuma or libcpuset. It might
be that the only way to smoke out such cases is to ship the change (to
Choice B default) and see what squawks. That kinda sucks.
I'm tempted to think I need to go a bit further:
1) Add a per-system or per-cpuset runtime mode, to enable a system
administrator to revert to the Choice A default.
2) Make sure that both the new libnuma and libcpuset versions dynamically
probe the state of this default, and dynamically adapt to running
on a kernel, or in a cpuset, of either default. If this mode is
per-cpuset, then so long as there are not two applications that
must both run in the same cpuset, both using memory policies,
neither using libnuma or libcpuset, requiring conflicting defaults
(one Choice A, the other Choice B) then this would provide an
administrator sufficient flexibility to adapt.
There would be one key advantage to a per-cpuset mode flag here. It
exposes in a fairly visible way this change in the default numbering of
memory policy node masks, from cpuset relative to system relative (with
the system now automatically making them cpuset relative across any
changes in the number of nodes in the cpuset.) This is a classic way
to guide users to understanding new alternatives and changes; expose it
as a 'button on the dashboard', a feature, not a hidden gotcha.
It's the old "it's a feature, not a bug" approach. It can be a
useful mechanism to educate and enpower customers.
Users end up seeing it, learning about it, enjoying some sense of
control, and usually being able to make it work for them, one way
or the other. That works out better than just hitting some random
subset of users with subtle bugs (from their perspective) over which
they have little or no control and fewer clues.
--
I won't rest till it's the best ...
Programmer, Linux Scalability
Paul Jackson <[email protected]> 1.925.600.0401
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]