On Mon, 2005-03-28 at 19:28 +0200, Matthieu Castet wrote: > > The memory limits aren't good enough either: if you set them low > > enough that memory-forkbombs are unperilous for > > RLIMIT_NPROC*RLIMIT_DATA, it's probably too low for serious > > applications. > > yes, if you want to run application like openoffice.org you need at > least 200Mo. If you want that your system is usable, you need at least 40 > process per user. So 40*200 = 8Go, and it don't think you have all this > memory... > > I think per user limit could be a solution.
You have /etc/limits and /etc/security/limits.conf. I think it would solve many problems by simply lowering the default max_treads in kernel/fork.c. RLIMIT_NPROC is calculated from this value. --- kernel/fork.c.orig 2005-03-02 08:37:48.000000000 +0100 +++ kernel/fork.c 2005-03-21 15:22:50.000000000 +0100 @@ -119,7 +119,7 @@ * value: the thread structures can take up at most half * of memory. */ - max_threads = mempages / (8 * THREAD_SIZE / PAGE_SIZE); + max_threads = mempages / (16 * THREAD_SIZE / PAGE_SIZE); /* * we need to allow at least 20 threads to boot a system I don't think this will cause much problems for most users. (compare the default maximum process limit in the BSD's and OSX) This will also limit deamons/services started from boot scripts by default. The /etc/limits and /etc/security/limits.conf does not. If it does cause problems for extrem users, they can easily raise the limits in either initrd and/or using /proc/sys/kernel/threads-max (or systctl). BTW... does anyone know *why* the default max number of processes is so high in Linux? -- Natanael Copa - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/