Quoting Otto Moerbeek <o...@drijf.net>:
On Wed, Jun 16, 2010 at 05:45:40PM -0400, George Morgan wrote:
Quoting Harry Palmer <tumblew...@fast-mail.org>:
>Hi there.
>
>I'm fairly new to openbsd and I'm hoping someone with better
>understanding than me of how its disk handling works can help.
>
>Beginning my effort to encrypt a 300GB drive in a 64bit Ultrasparc,
>I followed these initial steps:
>
>1. used disklabel to create a single slice "a" on the drive
>
>2. made a file system with newfs (is it necessary to have so many
> backup superblocks?)
>
>3. mounted sd2a on "/home/cy" and touched it with an empty file
> "/home/cy/cryptfile"
>
>4. zeroed out the file (and efectively the drive) with
> "dd if=/dev/zero of=/home/cy/cryptfile bs=512"
>
>
>Here's the (eventual!) output of (4):
>
> /home/cy: write failed, file system is full
> dd: /home/cy/cryptfile: No space left on device
> 576520353+0 records in
> 576520352+0 records out
> 295178420224 bytes transferred in 19810.722 secs (14899932 bytes/sec)
>
>
>
>Now I have:
>
> # disklabel sd2a
> # /dev/rsd2a:
> type: SCSI
> disk: SCSI disk
> label: MAW3300NC
> flags: vendor
> bytes/sector: 512
> sectors/track: 930
> tracks/cylinder: 8
> sectors/cylinder: 7440
> cylinders: 13217
> total sectors: 585937500
> rpm: 10025
> interleave: 1
> boundstart: 0
> boundend: 585937500
> drivedata: 0
>
> 16 partitions:
> # size offset fstype [fsize bsize cpg]
> a: 585937200 0 4.2BSD 2048 16384 1
> c: 585937500 0 unused
>
>
>and:
>
> # ls -l /home/cy
> total 576661216
> -rw-r--r-- 1 root wheel 295178420224 Jun 16 03:39 cryptfile
>
>
>and:
>
> # df -h
> Filesystem Size Used Avail Capacity Mounted on
> /dev/sd0a 1007M 44.8M 912M 5% /
> /dev/sd0k 247G 2.0K 235G 0% /home
> /dev/sd0d 3.9G 6.0K 3.7G 0% /tmp
> /dev/sd0f 2.0G 559M 1.3G 29% /usr
> /dev/sd0g 1007M 162M 795M 17% /usr/X11R6
> /dev/sd0h 5.9G 212K 5.6G 0% /usr/local
> /dev/sd0j 2.0G 2.0K 1.9G 0% /usr/obj
> /dev/sd0i 2.0G 2.0K 1.9G 0% /usr/src
> /dev/sd0e 7.9G 7.7M 7.5G 0% /var
> /dev/sd2a 275G 275G -13.7G 105% /home/cy
>
>
>
>I have no understanding of this. I've never seen a df output
>that tells me I'm using 13GB more space than the drive is
>capable of holding.
>
>I ask here because there's obviously potential for me to lose
>data somewhere down the line. I'll be grateful if anyone can
>explain where I've gone wrong.
I've seen the greater than 100% full on a UFS? filesystem before when
you exceed the size of the filesystem. There is space in the
filesystem for "lost+found" and all those superblocks? you were
complaining about that can get overwritten if you write too much to a
partition.
Spoace for superblocks and other metadata is subtracted from available
blocks. lost+found is an ordinary directory.
So setting up your "dd" to actually stop before you overfill the
filesystem is what you need to do. (using bs=# count=# ... info you
can get before you start initializing your file with the df command
without the "-k or -h" to get number of blocks and block size)
I'm sure the fine people on these lists will correct me if I'm wrong
in my assumptions... :-)
You are wrong, there;s no such thing as overfilling a filesystem. It's
just the 5% reserved for root. An ordinary user runs out earlier. It's
in the FAQ.
Sorry for the misinformation. Thanks for the education.
George Morgan