ok I will answer by my self: yes gluster daemon is managed by vdms:) and to recover lost config simply one should add "force" keyword gluster volume create GluReplica replica 3 arbiter 1 transport TCP,RDMA 10.10.10.44:/zclei22/01/glu 10.10.10.42:/zclei21/01/glu 10.10.10.41:/zclei26/01/glu force
now everything is up an running ! one annoying thing is epel dependency in the zfs and conflicting ovirt... every time one need to enable and then disable epel. On Wed, Mar 1, 2017 at 5:33 PM, Arman Khalatyan <[email protected]> wrote: > ok Finally by single brick up and running so I can access to data. > Now the question is do we need to run glusterd daemon on startup? or it is > managed by vdsmd? > > > On Wed, Mar 1, 2017 at 2:36 PM, Arman Khalatyan <[email protected]> wrote: > >> all folders /var/lib/glusterd/vols/ are empty >> In the history of one of the servers I found the command how it was >> created: >> >> gluster volume create GluReplica replica 3 arbiter 1 transport TCP,RDMA >> 10.10.10.44:/zclei22/01/glu 10.10.10.42:/zclei21/01/glu 10.10.10.41: >> /zclei26/01/glu >> >> But executing this command it claims that: >> volume create: GluReplica: failed: /zclei22/01/glu is already part of a >> volume >> >> Any chance to force it? >> >> >> >> On Wed, Mar 1, 2017 at 12:13 PM, Ramesh Nachimuthu <[email protected]> >> wrote: >> >>> >>> >>> >>> >>> ----- Original Message ----- >>> > From: "Arman Khalatyan" <[email protected]> >>> > To: "users" <[email protected]> >>> > Sent: Wednesday, March 1, 2017 3:10:38 PM >>> > Subject: Re: [ovirt-users] Gluster setup disappears any chance to >>> recover? >>> > >>> > engine throws following errors: >>> > 2017-03-01 10:39:59,608+01 WARN >>> > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>> > (DefaultQuartzScheduler6) [d7f7d83] EVENT_ID: >>> > GLUSTER_VOLUME_DELETED_FROM_CLI(4,027), Correlation ID: null, Call >>> Stack: >>> > null, Custom Event ID: -1, Message: Detected deletion of volume >>> GluReplica >>> > on cluster HaGLU, and deleted it from engine DB. >>> > 2017-03-01 10:39:59,610+01 ERROR >>> > [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>> (DefaultQuartzScheduler6) >>> > [d7f7d83] Error while removing volumes from database!: >>> > org.springframework.dao.DataIntegrityViolationException: >>> > CallableStatementCallback; SQL [{call deleteglustervolumesbyguids(?) >>> }]; >>> > ERROR: update or delete on table "gluster_volumes" violates foreign key >>> > constraint "fk_storage_connection_to_glustervolume" on table >>> > "storage_server_connections" >>> > Detail: Key (id)=(3d8bfa9d-1c83-46ac-b4e9-bd317623ed2d) is still >>> referenced >>> > from table "storage_server_connections". >>> > Where: SQL statement "DELETE >>> > FROM gluster_volumes >>> > WHERE id IN ( >>> > SELECT * >>> > FROM fnSplitterUuid(v_volume_ids) >>> > )" >>> > PL/pgSQL function deleteglustervolumesbyguids(character varying) line >>> 3 at >>> > SQL statement; nested exception is org.postgresql.util.PSQLException: >>> ERROR: >>> > update or delete on table "gluster_volumes" violates foreign key >>> constraint >>> > "fk_storage_connection_to_glustervolume" on table >>> > "storage_server_connections" >>> > Detail: Key (id)=(3d8bfa9d-1c83-46ac-b4e9-bd317623ed2d) is still >>> referenced >>> > from table "storage_server_connections". >>> > Where: SQL statement "DELETE >>> > FROM gluster_volumes >>> > WHERE id IN ( >>> > SELECT * >>> > FROM fnSplitterUuid(v_volume_ids) >>> > )" >>> > PL/pgSQL function deleteglustervolumesbyguids(character varying) line >>> 3 at >>> > SQL statement >>> > at >>> > org.springframework.jdbc.support.SQLErrorCodeSQLExceptionTra >>> nslator.doTranslate(SQLErrorCodeSQLExceptionTranslator.java:243) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > at >>> > org.springframework.jdbc.support.AbstractFallbackSQLExceptio >>> nTranslator.translate(AbstractFallbackSQLExceptionTranslator.java:73) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > at org.springframework.jdbc.core.JdbcTemplate.execute(JdbcTempl >>> ate.java:1094) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > at org.springframework.jdbc.core.JdbcTemplate.call(JdbcTemplate >>> .java:1130) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > at >>> > org.springframework.jdbc.core.simple.AbstractJdbcCall.execut >>> eCallInternal(AbstractJdbcCall.java:405) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > at >>> > org.springframework.jdbc.core.simple.AbstractJdbcCall.doExec >>> ute(AbstractJdbcCall.java:365) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > at >>> > org.springframework.jdbc.core.simple.SimpleJdbcCall.execute( >>> SimpleJdbcCall.java:198) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > at >>> > org.ovirt.engine.core.dal.dbbroker.SimpleJdbcCallsHandler.ex >>> ecuteImpl(SimpleJdbcCallsHandler.java:135) >>> > [dal.jar:] >>> > at >>> > org.ovirt.engine.core.dal.dbbroker.SimpleJdbcCallsHandler.ex >>> ecuteImpl(SimpleJdbcCallsHandler.java:130) >>> > [dal.jar:] >>> > at >>> > org.ovirt.engine.core.dal.dbbroker.SimpleJdbcCallsHandler.ex >>> ecuteModification(SimpleJdbcCallsHandler.java:76) >>> > [dal.jar:] >>> > at >>> > org.ovirt.engine.core.dao.gluster.GlusterVolumeDaoImpl.remov >>> eAll(GlusterVolumeDaoImpl.java:233) >>> > [dal.jar:] >>> > at >>> > org.ovirt.engine.core.bll.gluster.GlusterSyncJob.removeDelet >>> edVolumes(GlusterSyncJob.java:521) >>> > [bll.jar:] >>> > at >>> > org.ovirt.engine.core.bll.gluster.GlusterSyncJob.refreshVolu >>> meData(GlusterSyncJob.java:465) >>> > [bll.jar:] >>> > at >>> > org.ovirt.engine.core.bll.gluster.GlusterSyncJob.refreshClus >>> terData(GlusterSyncJob.java:133) >>> > [bll.jar:] >>> > at >>> > org.ovirt.engine.core.bll.gluster.GlusterSyncJob.refreshLigh >>> tWeightData(GlusterSyncJob.java:111) >>> > [bll.jar:] >>> > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> > [rt.jar:1.8.0_121] >>> > at >>> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAcce >>> ssorImpl.java:62) >>> > [rt.jar:1.8.0_121] >>> > at >>> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMe >>> thodAccessorImpl.java:43) >>> > [rt.jar:1.8.0_121] >>> > at java.lang.reflect.Method.invoke(Method.java:498) [rt.jar:1.8.0_121] >>> > at >>> > org.ovirt.engine.core.utils.timer.JobWrapper.invokeMethod(Jo >>> bWrapper.java:77) >>> > [scheduler.jar:] >>> > at org.ovirt.engine.core.utils.timer.JobWrapper.execute(JobWrap >>> per.java:51) >>> > [scheduler.jar:] >>> > at org.quartz.core.JobRunShell.run(JobRunShell.java:213) [quartz.jar:] >>> > at java.util.concurrent.Executors$RunnableAdapter.call(Executor >>> s.java:511) >>> > [rt.jar:1.8.0_121] >>> > at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>> > [rt.jar:1.8.0_121] >>> > at >>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool >>> Executor.java:1142) >>> > [rt.jar:1.8.0_121] >>> > at >>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo >>> lExecutor.java:617) >>> > [rt.jar:1.8.0_121] >>> > at java.lang.Thread.run(Thread.java:745) [rt.jar:1.8.0_121] >>> > Caused by: org.postgresql.util.PSQLException: ERROR: update or delete >>> on >>> > table "gluster_volumes" violates foreign key constraint >>> > "fk_storage_connection_to_glustervolume" on table >>> > "storage_server_connections" >>> > Detail: Key (id)=(3d8bfa9d-1c83-46ac-b4e9-bd317623ed2d) is still >>> referenced >>> > from table "storage_server_connections". >>> > Where: SQL statement "DELETE >>> > FROM gluster_volumes >>> > WHERE id IN ( >>> > SELECT * >>> > FROM fnSplitterUuid(v_volume_ids) >>> > )" >>> > PL/pgSQL function deleteglustervolumesbyguids(character varying) line >>> 3 at >>> > SQL statement >>> > at >>> > org.postgresql.core.v3.QueryExecutorImpl.receiveErrorRespons >>> e(QueryExecutorImpl.java:2157) >>> > at >>> > org.postgresql.core.v3.QueryExecutorImpl.processResults(Quer >>> yExecutorImpl.java:1886) >>> > at >>> > org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecut >>> orImpl.java:255) >>> > at >>> > org.postgresql.jdbc2.AbstractJdbc2Statement.execute(Abstract >>> Jdbc2Statement.java:555) >>> > at >>> > org.postgresql.jdbc2.AbstractJdbc2Statement.executeWithFlags >>> (AbstractJdbc2Statement.java:417) >>> > at >>> > org.postgresql.jdbc2.AbstractJdbc2Statement.execute(Abstract >>> Jdbc2Statement.java:410) >>> > at >>> > org.jboss.jca.adapters.jdbc.CachedPreparedStatement.execute( >>> CachedPreparedStatement.java:303) >>> > at >>> > org.jboss.jca.adapters.jdbc.WrappedPreparedStatement.execute >>> (WrappedPreparedStatement.java:442) >>> > at >>> > org.springframework.jdbc.core.JdbcTemplate$6.doInCallableSta >>> tement(JdbcTemplate.java:1133) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > at >>> > org.springframework.jdbc.core.JdbcTemplate$6.doInCallableSta >>> tement(JdbcTemplate.java:1130) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > at org.springframework.jdbc.core.JdbcTemplate.execute(JdbcTempl >>> ate.java:1078) >>> > [spring-jdbc.jar:4.2.4.RELEASE] >>> > ... 24 more >>> > >>> > >>> > >>> >>> This is a side effect volume deletion in the gluster side. Looks like >>> you have storage domains created using those volumes. >>> >>> > On Wed, Mar 1, 2017 at 9:49 AM, Arman Khalatyan < [email protected] > >>> wrote: >>> > >>> > >>> > >>> > Hi, >>> > I just tested power cut on the test system: >>> > >>> > Cluster with 3-Hosts each host has 4TB localdisk with zfs on it >>> /zhost/01/glu >>> > folder as a brick. >>> > >>> > Glusterfs was with replicated to 3 disks with arbiter. So far so good. >>> Vm was >>> > up an running with 5oGB OS disk: dd was showing 100-70MB/s performance >>> with >>> > the Vm disk. >>> > I just simulated disaster powercut: with ipmi power-cycle all 3 hosts >>> same >>> > time. >>> > the result is all hosts are green up and running but bricks are down. >>> > in the processes I can see: >>> > ps aux | grep gluster >>> > root 16156 0.8 0.0 475360 16964 ? Ssl 08:47 0:00 /usr/sbin/glusterd -p >>> > /var/run/glusterd.pid --log-level INFO >>> > >>> > What happened with my volume setup?? >>> > Is it possible to recover it?? >>> > [root@clei21 ~]# gluster peer status >>> > Number of Peers: 2 >>> > >>> > Hostname: clei22.cls >>> > Uuid: 96b52c7e-3526-44fd-af80-14a3073ebac2 >>> > State: Peer in Cluster (Connected) >>> > Other names: >>> > 192.168.101.40 >>> > 10.10.10.44 >>> > >>> > Hostname: clei26.cls >>> > Uuid: c9fab907-5053-41a8-a1fa-d069f34e42dc >>> > State: Peer in Cluster (Connected) >>> > Other names: >>> > 10.10.10.41 >>> > [root@clei21 ~]# gluster volume info >>> > No volumes present >>> > [root@clei21 ~]# >>> >>> I not sure why all volumes are getting deleted after reboot. Do you see >>> any vol files under the directory /var/lib/glusterd/vols/?. Also >>> /var/log/glusterfs/cmd_history.log should have all the gluster commands >>> executed. >>> >>> Regards, >>> Ramesh >>> >>> > >>> > >>> > >>> > _______________________________________________ >>> > Users mailing list >>> > [email protected] >>> > http://lists.ovirt.org/mailman/listinfo/users >>> > >>> >> >> >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

