On Aug 19 2007 14:39, Paolo Ornati wrote:
>> WT> Under unix, the shell resolves "*" and passes the 10000 file names
>> WT> to the "rm" command. Now, execve() may fail because 10000 names in
>> WT> arguments can require too much memory. That's why find and xargs
>> WT> were invented!
>>
>> It would be very handy if the argument memory space was expanded.
>> Many years ago I hit the limit regularly on Solaris, and going to
>> Linux with its comparatively large limit was a joy. Now it happens to
>> me quite often on Linux as well.
>>
>
>done :)
>
>commit b6a2fea39318e43fee84fa7b0b90d68bed92d2ba
>Author: Ollie Wild <[email protected]>
>Date: Thu Jul 19 01:48:16 2007 -0700
>
> mm: variable length argument support
>
> Remove the arg+env limit of MAX_ARG_PAGES by copying the strings
> directly from the old mm into the new mm.
Me wonders. Will that make the "checking for maximum length of command line
arguments" from autotools run forever since execve() will not fail anymore?
Jan
--
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]