To have prometheus plugin working you HAVE to tell it to listen to an IPV4 address ...like this
ceph config set mgr mgr/prometheus/server_addr 0.0.0.0 On Fri, 24 Aug 2018 at 12:44, Jones de Andrade <johanne...@gmail.com> wrote: > Hi all. > > I'm new to ceph, and after having serious problems in ceph stages 0, 1 and > 2 that I could solve myself, now it seems that I have hit a wall harder > than my head. :) > > When I run salt-run state.orch ceph.stage.deploy, i monitor I see it going > up to here: > > ####### > [14/71] ceph.sysctl on > node01....................................... ✓ (0.5s) > node02........................................ ✓ (0.7s) > node03....................................... ✓ (0.6s) > node04......................................... ✓ (0.5s) > node05....................................... ✓ (0.6s) > node06.......................................... ✓ (0.5s) > > [15/71] ceph.osd on > node01...................................... ❌ (0.7s) > node02........................................ ❌ (0.7s) > node03....................................... ❌ (0.7s) > node04......................................... ❌ (0.6s) > node05....................................... ❌ (0.6s) > node06.......................................... ❌ (0.7s) > > Ended stage: ceph.stage.deploy succeeded=14/71 failed=1/71 time=624.7s > > Failures summary: > > ceph.osd (/srv/salt/ceph/osd): > node02: > deploy OSDs: Module function osd.deploy threw an exception. Exception: > Mine on node02 for cephdisks.list > node03: > deploy OSDs: Module function osd.deploy threw an exception. Exception: > Mine on node03 for cephdisks.list > node01: > deploy OSDs: Module function osd.deploy threw an exception. Exception: > Mine on node01 for cephdisks.list > node04: > deploy OSDs: Module function osd.deploy threw an exception. Exception: > Mine on node04 for cephdisks.list > node05: > deploy OSDs: Module function osd.deploy threw an exception. Exception: > Mine on node05 for cephdisks.list > node06: > deploy OSDs: Module function osd.deploy threw an exception. Exception: > Mine on node06 for cephdisks.list > ####### > > Since this is a first attempt in 6 simple test machines, we are going to > put the mon, osds, etc, in all nodes at first. Only the master is left in a > single machine (node01) by now. > > As they are simple machines, they have a single hdd, which is partitioned > as follows (the hda4 partition is unmounted and left for the ceph system): > > ########### > # lsblk > NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT > sda 8:0 0 465,8G 0 disk > ├─sda1 8:1 0 500M 0 part /boot/efi > ├─sda2 8:2 0 16G 0 part [SWAP] > ├─sda3 8:3 0 49,3G 0 part / > └─sda4 8:4 0 400G 0 part > sr0 11:0 1 3,7G 0 rom > > # salt -I 'roles:storage' cephdisks.list > node01: > node02: > node03: > node04: > node05: > node06: > > # salt -I 'roles:storage' pillar.get ceph > node02: > ---------- > storage: > ---------- > osds: > ---------- > /dev/sda4: > ---------- > format: > bluestore > standalone: > True > (and so on for all 6 machines) > ########## > > Finally and just in case, my policy.cfg file reads: > > ######### > #cluster-unassigned/cluster/*.sls > cluster-ceph/cluster/*.sls > profile-default/cluster/*.sls > profile-default/stack/default/ceph/minions/*yml > config/stack/default/global.yml > config/stack/default/ceph/cluster.yml > role-master/cluster/node01.sls > role-admin/cluster/*.sls > role-mon/cluster/*.sls > role-mgr/cluster/*.sls > role-mds/cluster/*.sls > role-ganesha/cluster/*.sls > role-client-nfs/cluster/*.sls > role-client-cephfs/cluster/*.sls > ########## > > Please, could someone help me and shed some light on this issue? > > Thanks a lot in advance, > > Regasrds, > > Jones > > > > On Thu, Aug 23, 2018 at 2:46 PM John Spray <jsp...@redhat.com> wrote: > >> On Thu, Aug 23, 2018 at 5:18 PM Steven Vacaroaia <ste...@gmail.com> >> wrote: >> > >> > Hi All, >> > >> > I am trying to enable prometheus plugin with no success due to "no >> socket could be created" >> > >> > The instructions for enabling the plugin are very straightforward and >> simple >> > >> > Note >> > My ultimate goal is to use Prometheus with Cephmetrics >> > Some of you suggested to deploy ceph-exporter but why do we need to do >> that when there is a plugin already ? >> > >> > >> > How can I troubleshoot this further ? >> > >> > nhandled exception from module 'prometheus' while running on mgr.mon01: >> error('No socket could be created',) >> > Aug 23 12:03:06 mon01 ceph-mgr: 2018-08-23 12:03:06.615 7fadab50e700 -1 >> prometheus.serve: >> > Aug 23 12:03:06 mon01 ceph-mgr: 2018-08-23 12:03:06.615 7fadab50e700 -1 >> Traceback (most recent call last): >> > Aug 23 12:03:06 mon01 ceph-mgr: File >> "/usr/lib64/ceph/mgr/prometheus/module.py", line 720, in serve >> > Aug 23 12:03:06 mon01 ceph-mgr: cherrypy.engine.start() >> > Aug 23 12:03:06 mon01 ceph-mgr: File >> "/usr/lib/python2.7/site-packages/cherrypy/process/wspbus.py", line 250, in >> start >> > Aug 23 12:03:06 mon01 ceph-mgr: raise e_info >> > Aug 23 12:03:06 mon01 ceph-mgr: ChannelFailures: error('No socket could >> be created',) >> >> The things I usually check if a process can't create a socket are: >> - is there anything on the same node already listening on that port? >> - are there security policies (e.g. selinux) that might be preventing it? >> >> John >> >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@lists.ceph.com >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com