and this- after starting the osd manually
root@cephosd01:~# df
Filesystem 1K-blocks Used Available Use% Mounted on
/dev/dm-0 15616412 1583180 13216900 11% /
udev 10240 0 10240 0% /dev
tmpfs 49656 4636 45020 10% /run
tmpfs 124132 0 124132 0% /dev/shm
tmpfs 5120 0 5120 0% /run/lock
tmpfs 124132 0 124132 0% /sys/fs/cgroup
/dev/sda1 240972 33309 195222 15% /boot
/dev/sdb1 47161840 35260 47126580 1% /var/lib/ceph/osd/ceph-0
/dev/sdc1 47161840 34952 47126888 1% /var/lib/ceph/osd/ceph-1
what i did not understand is that i would expect ceph-deploy to work
properly. i just setup all six nodes in a fresh install, and then used
ceph deploy to install them:
All done from a adminvm:
ceph-deploy new cephmon01 cephmon02 cephmon03
ceph-deploy install cephmon01 cephmon02 cephmon03 cephosd01 cephosd02
cephosd03
ceph-deploy mon create cephmon01
ceph-deploy mon create cephmon02
ceph-deploy mon create cephmon03
ceph-deploy osd prepare cephosd01:sdb cephosd01:sdc
ceph-deploy osd prepare cephosd02:sdb cephosd02:sdc
ceph-deploy osd prepare cephosd03:sdb cephosd03:sdc
ceph osd tree
and directly afterwards (after seeing 6 OSDs up)
ssh cephosd01 shutdown -r
root@cephadmin:~# cat /etc/debian_version
8.5
Am 12.07.2016 um 00:05 schrieb Dirk Laurenz:
root@cephosd01:~# fdisk -l /dev/sdb
Disk /dev/sdb: 50 GiB, 53687091200 bytes, 104857600 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: gpt
Disk identifier: 87B152E0-EB5D-4EB0-8FFB-C27096CBB1ED
Device Start End Sectors Size Type
/dev/sdb1 10487808 104857566 94369759 45G unknown
/dev/sdb2 2048 10487807 10485760 5G unknown
Partition table entries are not in disk order.
root@cephosd01:~# fdisk -l /dev/sdc
Disk /dev/sdc: 50 GiB, 53687091200 bytes, 104857600 sectors
Units: sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disklabel type: gpt
Disk identifier: 31B81FCA-9163-4723-B195-97AEC9568AD0
Device Start End Sectors Size Type
/dev/sdc1 10487808 104857566 94369759 45G unknown
/dev/sdc2 2048 10487807 10485760 5G unknown
Partition table entries are not in disk order.
Am 11.07.2016 um 18:01 schrieb George Shuklin:
Check out partition type for data partition for ceph.
fdisk -l /dev/sdc
On 07/11/2016 04:03 PM, Dirk Laurenz wrote:
hmm, helps partially ... running
/usr/sbin/ceph-disk trigger /dev/sdc1 or sdb1 works and brings osd up..
systemctl enable does not help....
Am 11.07.2016 um 14:49 schrieb George Shuklin:
Short story how OSDs are started in systemd environments:
Ceph OSD parittions has specific typecode (partition type
4FBD7E29-9D25-41B8-AFD0-062C0CEFF05D). It handled by udev rules
shipped by ceph package:
/lib/udev/rules.d/95-ceph-osd.rules
It set up proper owner/group for this disk ('ceph' instead 'root')
and calls /usr/sbin/ceph-disk trigger.
ceph-disk triggers creation of instance of ceph-disk@ systemd unit
(to mount disk to /var/lib/ceph/osd/...), and ceph-osd@ (i'm not
sure about all sequence of events).
Basically, to make OSD autostart they NEED to have proper typecode
in their partition. If you using something different (like
'directory based OSD') you should enable OSD autostart:
systemctl enable ceph-osd@42
On 07/11/2016 03:32 PM, Dirk Laurenz wrote:
Hello,
i'm new to ceph an try to do some first steps with ceph to
understand concepts.
my setup is at first completly in vm....
i deployed (with ceph-deploy) three monitors and three osd hosts.
(3+3 vms)
my frist test was to find out, if everything comes back online
after a system restart. this works fine for the monitors, but
fails for the osds. i have to start them manually.
OS is debian jessie, ceph is the current release....
Where can find out, what's going wrong....
Dirk
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com