On Jan 20, 2008 5:51 AM, SchC6berle DC!niel <[EMAIL PROTECTED]> wrote:
> > From: Richard Daemon [mailto:[EMAIL PROTECTED] > > > > On Jan 19, 2008 8:31 PM, Schvberle Daniel > > <[EMAIL PROTECTED]> wrote: > > > > > > Hi all! > > > > I've just upgraded my firewall from OpenBSD 4.0 to > > 4.2-stable and ran > > into a small problem regarding mount_mfs. I solved it, > > but in case > > anybody else runs into it, here's something for the archives. > > > > I run the box from a 512MB CF and, originally, with very limited > > memory. The /var, tmp and /dev are mount_mfs and during > > the upgrade I > > had trobule with mounting /dev. > > > > I used to mount /dev with the following line: > > > > swap /dev mfs rw,-P=/proto/dev,-s=700,-i=256 0 0 > > > > It seems that sometime after 4.1 was released (probably > > during ffs2 > > development) mount_mfs was changed in such a way that > > it doesn't allow > > very high density for inodes. This resulted in > > mount_mfs failing on > > replicating the /dev and me getting a readonly /dev, > > which resulted > > in a box that I couldn't login into remotly (with ssh). > > Luckily you I > > could still issue commands with winscp or login > > locally. After couple > > of tests I concluded that mount_mfs simply ignores > > density settings > > lower than 1024, so I changed the /dev to settings to > > the following > > line: > > > > swap /dev mfs rw,-P=/proto/dev,-s=4000,-i=1024 0 0 > > > > Now everything is ok, I'm happy and sice CF is in a new > > box with lots > > of memory I'm not trying to squeeze every byte out of it. > > > > Maybe this maximal density could be documented somehow? > > I glanced at > > the mkfs.c and saw that, in theory, it should warn the user when > > reducing the density but I never got a warning during my tests. > > > > dmesg in case anybody needs it: > > <snip dmesg> > > > > > Wow, very weird that you post this. I just noticed the exact > > same thing yesterday too. Upgraded from 4.0-stable to > > 4.2-stable on a WRAP (pcengines.ch) box with my 512M CF and > > /dev entries failing as well. My previous inode settings used to be: > > > > swap /dev mfs rw,-P=/.devtmp,-s=1200,-i=128 0 0 but that > > crapped out in 4.2. > > > > I changed it to -s=3072, -i=128 just to get it fully working > > properly and I haven't looked into it further yet, but > > wondering if I'm better off maybe trying higher inode (like > > yours) but lower MFS size such as -s=1024 because I'm limited > > in memory (128M total). Other than that, is an MFS /dev size > > bigger than 1M even needed? I'd really like to reduce as much > > as possible. > > > > Thanks for the post! > > > > I'm new to this mailing list and so far, it's great! > > No, I don't think you'd ever need a /dev this big, but in order to > get the needed number of inodes you have to push the size up. > Your line is ok, but maybe you should put i=1024 instead of i=128, > so you know what the real values are - that's what it's using anyway > With 128MB you really shouldn'y worry. I was concerned because I had > only 32MB or 48MB. mount_mfs doesn't really use the memory untill > it's needed, so you could make, say 100GB mfs on a box with 128MB of > RAM and it would work as long as you've got memory to hold the > files. Regardnig /dev, you really don't need much as it's a small > filesystem, but sometimes you can get real files in there. This is > what happend once to my lil' box (I had a _real_ /dev/null) and it > crapped out because it ran out of memory. After that I reduced the > /dev as much as I could, I didn't want another local DoS to happen. > I have 512MB now and couldn't care less if /dev is 0.1 or 1 MB, > and with 128MB you shouldn't either, especially since it gets > allocated only if really needed by the files. > > Thank you very much for the reply! Much appreciate your suggestions and advice.