On Thu, 2006-09-14 at 10:21 -0500, Ram Gupta wrote:
> Hi,
> I came across a problem regarding the issue of number of open
> file descriptors.The scenario is as below.
> I have a controlling process which launches other different
> apllication including third party ones. It also enforces various
> resource limits including number of open file descriptors. This
> process forks & reads the resource limits from a configuration files,
> applies the resource limits & then execs for the corresponding
> application. The process has its own various number of open file
> descriptors. If the limit of open file descriptor for the child
> application is less than the number of file descriptors of the parent
> process, then the child application file can not be opened & exec
> fails in this case.
>
> I searched solution for this problem but could not find an existing
> way to solve it. I thought of couple of ways to do it. One idea is to
-) simply close(2) the superfluous file descriptors unconditionally
before exec(2)ing the child process
-) set the close-on-exec flag via fcntl(2) on these file descriptors
after opening them
Bernd
--
Firmix Software GmbH http://www.firmix.at/
mobil: +43 664 4416156 fax: +43 1 7890849-55
Embedded Linux Development and Services
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]