Greetings,

I have Sun Fire 4600 running Solaris 10, running Sun Cluster 3.2.

SunOS hubdb004 5.10 Generic_120012-14 i86pc i386 i86pc

We ran into some space issues in /usr today, so as a quick fix, I created a
slice (c5t0d0s12) with about 25GB of disk in order to create some zfs
filesystems with mount points inside /usr...

# zpool status
  pool: tank01
 state: ONLINE
 scrub: none requested
config:

        NAME         STATE     READ WRITE CKSUM
        tank01       ONLINE       0     0     0
          c5t0d0s12  ONLINE       0     0     0

errors: No known data errors
# zfs list
NAME              USED  AVAIL  REFER  MOUNTPOINT
tank01            251M  24.2G  24.5K  /tank01
tank01/openwin   40.2M  24.2G  40.2M  /usr/openwin
tank01/usrlocal   210M  24.2G   210M  /usr/local

All seems good...except when an unprivileged user runs df -k or df -h, df
core dumps when it hits the zpool.  Running df as root does not produce the
coredump.  Both root and me run /usr/sbin/df.  When me runs df alone it
exits normally.

[EMAIL PROTECTED]:~]$ df -k
Filesystem            kbytes    used   avail capacity  Mounted on
/dev/dsk/c5t0d0s0    2062186  283202 1717119    15%    /
/devices                   0       0       0     0%    /devices
ctfs                       0       0       0     0%    /system/contract
proc                       0       0       0     0%    /proc
mnttab                     0       0       0     0%    /etc/mnttab
swap                 14841484     852 14840632     1%    /etc/svc/volatile
objfs                      0       0       0     0%    /system/object
/dev/dsk/c5t0d0s6    1019856  747846  210819    79%    /usr
/usr/lib/libc/libc_hwcap2.so.1
                     1019856  747846  210819    79%    /lib/libc.so.1
fd                         0       0       0     0%    /dev/fd
/dev/dsk/c5t0d0s1    2062186 1319211  681110    66%    /var
swap                 14840716      84 14840632     1%    /tmp
swap                 14869380   28748 14840632     1%    /var/run
/dev/dsk/c5t0d0s7    29950483 9491964 20159015    33%    /opt
/dev/dsk/c5t0d0s5    1019856  235635  723030    25%    /export/home
/dev/did/dsk/d21s4   2062186   11726 1988595     1%
/global/.devices/[EMAIL PROTECTED]
127.0.0.1:/          113266280 27230964 80372004    26%    /mnt/exp2
127.0.0.1:/          113266280 39484820 68118148    37%    /mnt/exp1
127.0.0.1:/          113266280 49363852 58239116    46%    /mnt/exp3
/dev/did/dsk/d3s4    2062186   11732 1988589     1%
/global/.devices/[EMAIL PROTECTED]
/dev/md/ora-data/dsk/d500
                     1003783871 916489136 77256897    93%
/global/ora-data
/dev/md/ora-log/dsk/d100
                     223071250 176539191 44301347    80%
/global/ora-log/d100
Segmentation Fault (core dumped)

As root... the same as above but with the following instead of the Seg Fault

tank01                  24G    24K    24G     1%    /tank01
tank01/openwin          24G    40M    24G     1%    /usr/openwin
tank01/usrlocal         24G   210M    24G     1%    /usr/local
/export/home/me
                       996M   227M   709M    25%    /home/me


This is slighly problematic as we feel it's improper to run things like
monitors as the root user.

I have been searching Google, SunSolve and OpenSolaris, but either my search
terms are poor (most likely) or this has never been reported.

Any thoughts or suggestions would be most appreciated.  I much prefer the
use of zfs for this purpose over creating a UFS filesystem and using sym
links. from /usr/<whatever> to the UFS filesystem.

-- 
--

Ignorance: America's most abundant and costly commodity.
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to