Hi,

I'm trying to setup my first cluster,   (have never manually bootstrapped a 
cluster)

Is ceph-deploy odd activate/prepare supposed to write to the master ceph.conf 
file, specific entries for each OSD along the lines of 
http://ceph.com/docs/master/rados/configuration/osd-config-ref/ ?

I appear to have the OSDs prepared without error, but then.. no OSD entries in 
master cepf.conf nor node /etc/cepf.conf

Am I missing something?

Thanks in advance,

Piers Dawson-Damer
Tasmania


2013-09-28 06:47:00,471 [ceph_deploy.sudo_pushy][DEBUG ] will use a remote 
connection with sudo
2013-09-28 06:47:01,205 [ceph_deploy.osd][INFO  ] Distro info: Ubuntu 12.04 
precise
2013-09-28 06:47:01,205 [ceph_deploy.osd][DEBUG ] Preparing host 
storage03-vs-e2 disk /dev/sdm journal /dev/mapper/ceph_journal-osd_12 activate 
True
2013-09-28 06:47:01,206 [storage03-vs-e2][INFO  ] Running command: 
ceph-disk-prepare --cluster ceph -- /dev/sdm /dev/mapper/ceph_journal-osd_12
2013-09-28 06:47:20,247 [storage03-vs-e2][INFO  ] Information: Moved requested 
sector from 4194338 to 4196352 in
2013-09-28 06:47:20,248 [storage03-vs-e2][INFO  ] order to align on 2048-sector 
boundaries.
2013-09-28 06:47:20,248 [storage03-vs-e2][INFO  ] Warning: The kernel is still 
using the old partition table.
2013-09-28 06:47:20,248 [storage03-vs-e2][INFO  ] The new table will be used at 
the next reboot.
2013-09-28 06:47:20,248 [storage03-vs-e2][INFO  ] The operation has completed 
successfully.
2013-09-28 06:47:20,248 [storage03-vs-e2][INFO  ] Information: Moved requested 
sector from 34 to 2048 in
2013-09-28 06:47:20,249 [storage03-vs-e2][INFO  ] order to align on 2048-sector 
boundaries.
2013-09-28 06:47:20,249 [storage03-vs-e2][INFO  ] The operation has completed 
successfully.
2013-09-28 06:47:20,249 [storage03-vs-e2][INFO  ] meta-data=/dev/sdm1           
   isize=2048   agcount=4, agsize=183105343 blks
2013-09-28 06:47:20,250 [storage03-vs-e2][INFO  ]          =                    
   sectsz=512   attr=2, projid32bit=0
2013-09-28 06:47:20,250 [storage03-vs-e2][INFO  ] data     =                    
   bsize=4096   blocks=732421371, imaxpct=5
2013-09-28 06:47:20,250 [storage03-vs-e2][INFO  ]          =                    
   sunit=0      swidth=0 blks
2013-09-28 06:47:20,250 [storage03-vs-e2][INFO  ] naming   =version 2           
   bsize=4096   ascii-ci=0
2013-09-28 06:47:20,251 [storage03-vs-e2][INFO  ] log      =internal log        
   bsize=4096   blocks=357627, version=2
2013-09-28 06:47:20,251 [storage03-vs-e2][INFO  ]          =                    
   sectsz=512   sunit=0 blks, lazy-count=1
2013-09-28 06:47:20,251 [storage03-vs-e2][INFO  ] realtime =none                
   extsz=4096   blocks=0, rtextents=0
2013-09-28 06:47:20,251 [storage03-vs-e2][INFO  ] The operation has completed 
successfully.
2013-09-28 06:47:20,252 [storage03-vs-e2][ERROR ] WARNING:ceph-disk:OSD will 
not be hot-swappable if journal is not the same device as the osd data
2013-09-28 06:47:20,266 [storage03-vs-e2][INFO  ] Running command: udevadm 
trigger --subsystem-match=block --action=add
2013-09-28 06:47:20,413 [ceph_deploy.osd][DEBUG ] Host storage03-vs-e2 is now 
ready for osd use.





2013-09-27 10:13:25,349 [storage03-vs-e2][DEBUG ] status for monitor: 
mon.storage03-vs-e2
2013-09-27 10:13:25,349 [storage03-vs-e2][DEBUG ] { "name": "storage03-vs-e2",
2013-09-27 10:13:25,350 [storage03-vs-e2][DEBUG ]   "rank": 2,
2013-09-27 10:13:25,350 [storage03-vs-e2][DEBUG ]   "state": "electing",
2013-09-27 10:13:25,350 [storage03-vs-e2][DEBUG ]   "election_epoch": 1,
2013-09-27 10:13:25,351 [storage03-vs-e2][DEBUG ]   "quorum": [],
2013-09-27 10:13:25,351 [storage03-vs-e2][DEBUG ]   "outside_quorum": [],
2013-09-27 10:13:25,351 [storage03-vs-e2][DEBUG ]   "extra_probe_peers": [
2013-09-27 10:13:25,351 [storage03-vs-e2][DEBUG ]         
"172.17.181.47:6789\/0",
2013-09-27 10:13:25,352 [storage03-vs-e2][DEBUG ]         
"172.17.181.48:6789\/0"],
2013-09-27 10:13:25,352 [storage03-vs-e2][DEBUG ]   "sync_provider": [],
2013-09-27 10:13:25,352 [storage03-vs-e2][DEBUG ]   "monmap": { "epoch": 0,
2013-09-27 10:13:25,352 [storage03-vs-e2][DEBUG ]       "fsid": 
"28626c0a-0266-4b80-8c06-0562bf48b793",
2013-09-27 10:13:25,353 [storage03-vs-e2][DEBUG ]       "modified": "0.000000",
2013-09-27 10:13:25,353 [storage03-vs-e2][DEBUG ]       "created": "0.000000",
2013-09-27 10:13:25,353 [storage03-vs-e2][DEBUG ]       "mons": [
2013-09-27 10:13:25,353 [storage03-vs-e2][DEBUG ]             { "rank": 0,
2013-09-27 10:13:25,354 [storage03-vs-e2][DEBUG ]               "name": 
"storage01-vs-e2",
2013-09-27 10:13:25,354 [storage03-vs-e2][DEBUG ]               "addr": 
"172.17.181.47:6789\/0"},
2013-09-27 10:13:25,354 [storage03-vs-e2][DEBUG ]             { "rank": 1,
2013-09-27 10:13:25,354 [storage03-vs-e2][DEBUG ]               "name": 
"storage02-vs-e2",
2013-09-27 10:13:25,355 [storage03-vs-e2][DEBUG ]               "addr": 
"172.17.181.48:6789\/0"},
2013-09-27 10:13:25,355 [storage03-vs-e2][DEBUG ]             { "rank": 2,
2013-09-27 10:13:25,355 [storage03-vs-e2][DEBUG ]               "name": 
"storage03-vs-e2",
2013-09-27 10:13:25,355 [storage03-vs-e2][DEBUG ]               "addr": 
"172.17.181.49:6789\/0"}]}}
2013-09-27 10:13:25,356 [storage03-vs-e2][DEBUG ] 
2013-09-27 10:13:25,356 [storage03-vs-e2][DEBUG ] 
********************************************************************************
2013-09-27 10:13:25,356 [storage03-vs-e2][INFO  ] monitor: mon.storage03-vs-e2 
is running





2013-09-27 10:12:17,384 [storage03-vs-e2.groupthink.cc][INFO  ] ceph-all 
start/running
2013-09-27 10:12:17,384 [storage03-vs-e2.groupthink.cc][INFO  ] Setting up 
ceph-fs-common (0.67.3-1precise) ...
2013-09-27 10:12:17,384 [storage03-vs-e2.groupthink.cc][INFO  ] Setting up 
ceph-mds (0.67.3-1precise) ...
2013-09-27 10:12:17,384 [storage03-vs-e2.groupthink.cc][INFO  ] ceph-mds-all 
start/running
2013-09-27 10:12:17,395 [storage03-vs-e2.groupthink.cc][INFO  ] Running 
command: ceph --version
2013-09-27 10:12:17,547 [storage03-vs-e2.groupthink.cc][INFO  ] ceph version 
0.67.3 (408cd61584c72c0d97b774b3d8f95c6b1b06341a)

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to