Srikanth, Try if this helps..
sudo initctl list|grep ceph (should display all ceph daemon) sudo start ceps-mon-all (To start ceph all ceph-monitor) Thanks -Krishna > On May 7, 2015, at 1:35 PM, Srikanth Madugundi <srikanth.madugu...@gmail.com> > wrote: > > Hi, > > I am setting up a local instance of ceph cluster with latest source from git > hub. The build succeeded and installation was successful, But I could not > start the monitor. > > The "ceph start" command returns immediately and does not output anything. > $ sudo /etc/init.d/ceph start mon.monitor1 > > $ > > $ ls -l /var/lib/ceph/mon/ceph-monitor1/ > > total 8 > > -rw-r--r-- 1 root root 0 May 7 20:27 done > > -rw-r--r-- 1 root root 77 May 7 19:12 keyring > > drwxr-xr-x 2 root root 4096 May 7 19:12 store.db > > -rw-r--r-- 1 root root 0 May 7 20:26 sysvinit > > -rw-r--r-- 1 root root 0 May 7 20:09 upstart > > > > > > The log filed does not seem to have any details either > > > > $ cat /var/log/ceph/ceph-mon.monitor1.log > > > 2015-05-07 19:12:13.356389 7f3f06bdb880 -1 did not load config file, using > default settings. > > > > $ cat /etc/ceph/ceph.conf > > [global] > > mon host = 15.43.33.21 > > fsid = 92f859df-8b27-466a-8d44-01af2b7ea7e6 > > mon initial members = monitor1 > > > > # Enable authentication > > auth cluster required = cephx > > auth service required = cephx > > auth client required = cephx > > > > # POOL / PG / CRUSH > > osd pool default size = 3 # Write an object 3 times > > osd pool default min size = 1 # Allow writing one copy in a degraded state > > > > # Ensure you have a realistic number of placement groups. We recommend > > # approximately 200 per OSD. E.g., total number of OSDs multiplied by 200 > > # divided by the number of replicas (i.e., osd pool default size). > > # !! BE CAREFULL !! > > # You properly should never rely on the default numbers when creating pool! > > osd pool default pg num = 32 > > osd pool default pgp num = 32 > > > > #log file = /home/y/logs/ceph/$cluster-$type.$id.log > > > > # Logging > > debug paxos = 0 > > debug throttle = 0 > > > > keyring = /etc/ceph/ceph.client.admin.keyring > > #run dir = /home/y/var/run/ceph > > > > [mon] > > debug mon = 10 > > debug ms = 1 > > # We found that when the disk usage reach to 94%, the disk could not be > written > > # any file (no free space), so that we lower the full ratio and we should > start > > # data migration before it becomes full > > mon osd full ratio = 0.9 > > #mon data = /home/y/var/lib/ceph/mon/$cluster-$id > > mon osd down out interval = 172800 # 2 * 24 * 60 * 60 seconds > > # Ceph monitors need to be told how many reporters must to be seen from > different > > # OSDs before it can be marked offline, this should be greater than the > number of > > # OSDs per OSD host > > mon osd min down reporters = 12 > > #keyring = /home/y/conf/ceph/ceph.mon.keyring > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com