And the `sudo df -h`? Also a `ceph df` might be helpful to see what's going on.
On Mon, Jun 26, 2017 at 4:41 PM Papp Rudolf Péter <p...@peer.hu> wrote: > Hi David! > > lsblk: > > NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT > sda 8:0 0 931,5G 0 disk > ├─sda1 8:1 0 476M 0 part > ├─sda2 8:2 0 46,6G 0 part > │ └─md0 9:0 0 46,5G 0 raid1 / > └─sda3 8:3 0 884,5G 0 part /var/lib/ceph/osd/ceph-3 > sdb 8:16 0 931,5G 0 disk > ├─sdb1 8:17 0 476M 0 part /boot/efi > ├─sdb2 8:18 0 46,6G 0 part > │ └─md0 9:0 0 46,5G 0 raid1 / > └─sdb3 8:19 0 884,5G 0 part /var/lib/ceph/osd/ceph-0 > sdc 8:32 0 232,9G 0 disk > ├─sdc1 8:33 0 20G 0 part > ├─sdc2 8:34 0 576M 0 part > ├─sdc3 8:35 0 20G 0 part > └─sdc4 8:36 0 576M 0 part > > > 2017-06-26 22:37 keltezéssel, David Turner írta: > > What is the output of `lsblk`? > > On Mon, Jun 26, 2017 at 4:32 PM Papp Rudolf Péter <p...@peer.hu> wrote: > >> Dear cephers, >> >> Could someone show me an url where can I found how ceph calculate the >> available space? >> >> I've installed a small ceph (Kraken) environment with bluestore OSDs. >> The servers contains 2 disks and 1 ssd. The disk 1. part is UEFI (~500 >> MB), 2. raid (~50GB), 3. ceph disk (450-950MB). 1 server with 2 500 GB >> HDDs, 2 with 1 TB HDDs total 3 servers. >> >> For example the HDD parts: >> /dev/sdb1 2048 976895 974848 476M EFI System >> /dev/sdb2 976896 98633727 97656832 46,6G Linux RAID >> /dev/sdb3 98633728 1953525134 1854891407 884,5G Ceph OSD >> info from ceph-disk: >> /dev/sda : >> /dev/sda1 other, vfat >> /dev/sda2 other, linux_raid_member >> /dev/sda3 ceph data, active, cluster ceph, osd.4, block.db /dev/sdc1, >> block.wal /dev/sdc2 >> /dev/sdb : >> /dev/sdb1 other, vfat, mounted on /boot/efi >> /dev/sdb2 other, linux_raid_member >> /dev/sdb3 ceph data, active, cluster ceph, osd.1, block.db /dev/sdc3, >> block.wal /dev/sdc4 >> /dev/sdc : >> /dev/sdc1 ceph block.db, for /dev/sda3 >> /dev/sdc2 ceph block.wal, for /dev/sda3 >> /dev/sdc3 ceph block.db, for /dev/sdb3 >> /dev/sdc4 ceph block.wal, for /dev/sdb3 >> >> The reported size from ceph osd df tree: >> ID WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME >> -1 0.17578 - 179G 104M 179G 0.06 1.00 0 root default >> -2 0.05859 - 61439M 35696k 61405M 0.06 1.00 0 host cl2 >> 0 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.0 >> 3 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.3 >> -3 0.05859 - 61439M 35696k 61405M 0.06 1.00 0 host cl3 >> 1 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.1 >> 4 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.4 >> -4 0.05859 - 61439M 35696k 61405M 0.06 1.00 0 host cl1 >> 2 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.2 >> 5 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.5 >> TOTAL 179G 104M 179G 0.06 >> MIN/MAX VAR: 1.00/1.00 STDDEV: 0 >> >> ~ 30GB each 10 percent of the smallest real size. 3x replication. Could >> be possible that the system using wrong partition (2. in this scenario) >> for usable space calculation? Can I write more data than the calculated? >> >> Another hint? >> >> Thank you! >> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com