On 8/15/05, Jerry McAllister <[EMAIL PROTECTED]> wrote: > > > As someone mentioned, there is a FAQ on this. You should read it. > > It is going negative because you have used more than the nominal > capacity of the slice. The nominal capacity is the total space > minus the reserved proportion (usually 8%) that is held out. > Root is able to write to that space and you have done something > that got root to write beyond the nominal space.
I'm not sure you are right in this case. I think you need to re-read the post. I've quoted the relevent part here: > > Filesystem Size Used Avail Capacity Mounted on > > /dev/ar0s1e 248M -278K 228M -0% /tmp Looking at how the columns line up I have to state that I too have never seen this behaviour. As an experiment I over-filled a file system and here's the results: Filesystem Size Used Avail Capacity Mounted on /dev/ad0s1f 965M 895M -7.4M 101% /tmp Note capacity is not negative. So that makes three of us in this thread who have not seen negative capacity on UFS. I have seen negative capacity when running an old version of FreeBSD with a very large NFS mount (not enough bits in statfs if I remember correctly). > ////jerry Frem. _______________________________________________ freebsd-questions@freebsd.org mailing list http://lists.freebsd.org/mailman/listinfo/freebsd-questions To unsubscribe, send any mail to "[EMAIL PROTECTED]"