Could you try "gluster volume start VGSF1 force" to make sure the brick processes are restarted.
From the status output, it looks like the brick processes are not online.

On 04/22/2015 09:14 PM, [email protected] wrote:
Hello dears,
i've got some troubles with reattaching gluster volumes with data.

1) Base on a lot of tests I decided clear oVirt database ( # engine-cleanup ; # yum remove ovirt-engine; # yum -y install ovirt-engine; # engine-setup)
2) clearing sucessfully done and start with empty oVirt envir.
3) then I added networks, nodes and make basic network adjustment = all works fine 4) time to attach volumes/ domains with original data ( a lot of virtuals , ISO files , .... )

So, main question is about HOWTO attach this volumes if I haven't defined any domain and can't clearly import them ??

Current status of nodes are without glusterfs NFS mounted, but bricks are OK

# gluster volume info

Volume Name: VGFS1
Type: Replicate
Volume ID: b9a1c347-6ffd-4122-8756-d513fe3f40b9
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 1kvm2:/FastClass/p1/GFS1
Brick2: 1kvm1:/FastClass/p1/GFS1
Options Reconfigured:
storage.owner-gid: 36
storage.owner-uid: 36

Volume Name: VGFS2
Type: Replicate
Volume ID: b65bb689-ecc8-4c33-a4e7-11dea6028f83
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 1kvm2:/FastClass/p2/GFS1
Brick2: 1kvm1:/FastClass/p2/GFS1
Options Reconfigured:
storage.owner-uid: 36
storage.owner-gid: 36


[root@1kvm1 glusterfs]# gluster volume status
Status of volume: VGFS1
Gluster process Port    Online  Pid
------------------------------------------------------------------------------
Brick 1kvm1:/FastClass/p1/GFS1 N/A     N       N/A
NFS Server on localhost N/A     N       N/A
Self-heal Daemon on localhost N/A     N       N/A

Task Status of Volume VGFS1
------------------------------------------------------------------------------
There are no active volume tasks

Status of volume: VGFS2
Gluster process Port    Online  Pid
------------------------------------------------------------------------------
Brick 1kvm1:/FastClass/p2/GFS1 N/A     N       N/A
NFS Server on localhost N/A     N       N/A
Self-heal Daemon on localhost N/A     N       N/A

Task Status of Volume VGFS2
------------------------------------------------------------------------------
There are no active volume tasks

[root@1kvm1 glusterfs]# gluster volume start VGFS1
volume start: VGFS1: failed: Volume VGFS1 already started



# mount | grep mapper # base XFS mounting
/dev/mapper/3600605b0099f9e601cb1b5bf0e9765e8p1 on /FastClass/p1 type xfs (rw,relatime,seclabel,attr2,inode64,noquota) /dev/mapper/3600605b0099f9e601cb1b5bf0e9765e8p2 on /FastClass/p2 type xfs (rw,relatime,seclabel,attr2,inode64,noquota)


*5)* import screen
/VGFS1 dir exists & iptables flushed


# cat rhev-data-center-mnt-glusterSD-1kvm1:_VGFS1.log
[2015-04-22 15:21:50.204521] I [MSGID: 100030] [glusterfsd.c:2018:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.6.2 (args: /usr/sbin/glusterfs --volfile-server=1kvm1 --volfile-id=/VGFS1 /rhev/data-center/mnt/glusterSD/1kvm1:_VGFS1) [2015-04-22 15:21:50.220383] I [dht-shared.c:337:dht_init_regex] 0-VGFS1-dht: using regex rsync-hash-regex = ^\.(.+)\.[^.]+$ [2015-04-22 15:21:50.222255] I [client.c:2280:notify] 0-VGFS1-client-1: parent translators are ready, attempting connect on transport [2015-04-22 15:21:50.224528] I [client.c:2280:notify] 0-VGFS1-client-2: parent translators are ready, attempting connect on transport
Final graph:
+------------------------------------------------------------------------------+
  1: volume VGFS1-client-1
  2:     type protocol/client
  3:     option ping-timeout 42
  4:     option remote-host 1kvm2
  5:     option remote-subvolume /FastClass/p1/GFS1
  6:     option transport-type socket
  7:     option username 52f1efd1-60dc-4fb1-b94f-572945d6eb66
  8:     option password 34bac9cd-0b4f-41c6-973b-7af568784d7b
  9:     option send-gids true
 10: end-volume
 11:
 12: volume VGFS1-client-2
 13:     type protocol/client
 14:     option ping-timeout 42
 15:     option remote-host 1kvm1
 16:     option remote-subvolume /FastClass/p1/GFS1
 17:     option transport-type socket
 18:     option username 52f1efd1-60dc-4fb1-b94f-572945d6eb66
 19:     option password 34bac9cd-0b4f-41c6-973b-7af568784d7b
 20:     option send-gids true
 21: end-volume
 22:
 23: volume VGFS1-replicate-0
 24:     type cluster/replicate
 25:     subvolumes VGFS1-client-1 VGFS1-client-2
 26: end-volume
 27:
 28: volume VGFS1-dht
 29:     type cluster/distribute
 30:     subvolumes VGFS1-replicate-0
 31: end-volume
 32:
 33: volume VGFS1-write-behind
 34:     type performance/write-behind
 35:     subvolumes VGFS1-dht
 36: end-volume
 37:
 38: volume VGFS1-read-ahead
 39:     type performance/read-ahead
 40:     subvolumes VGFS1-write-behind
 41: end-volume
 42:
 43: volume VGFS1-io-cache
 44:     type performance/io-cache
 45:     subvolumes VGFS1-read-ahead
 46: end-volume
 47:
 48: volume VGFS1-quick-read
 49:     type performance/quick-read
 50:     subvolumes VGFS1-io-cache
 51: end-volume
 52:
 53: volume VGFS1-open-behind
 54:     type performance/open-behind
 55:     subvolumes VGFS1-quick-read
 56: end-volume
 57:
 58: volume VGFS1-md-cache
 59:     type performance/md-cache
 60:     subvolumes VGFS1-open-behind
 61: end-volume
 62:
 63: volume VGFS1
 64:     type debug/io-stats
 65:     option latency-measurement off
 66:     option count-fop-hits off
 67:     subvolumes VGFS1-md-cache
 68: end-volume
 69:
 70: volume meta-autoload
 71:     type meta
 72:     subvolumes VGFS1
 73: end-volume
 74:
+------------------------------------------------------------------------------+
[2015-04-22 15:21:50.227017] E [socket.c:2267:socket_connect_finish] 0-VGFS1-client-1: connection to 172.16.8.161:24007 failed (No route to host) [2015-04-22 15:21:50.227191] E [client-handshake.c:1496:client_query_portmap_cbk] 0-VGFS1-client-2: failed to get the port number for remote subvolume. Please run 'gluster volume status' on server to see if brick process is running. [2015-04-22 15:21:50.227218] I [client.c:2215:client_rpc_notify] 0-VGFS1-client-2: disconnected from VGFS1-client-2. Client process will keep trying to connect to glusterd until brick's port is available [2015-04-22 15:21:50.227227] E [MSGID: 108006] [afr-common.c:3591:afr_notify] 0-VGFS1-replicate-0: All subvolumes are down. Going offline until atleast one of them comes back up. [2015-04-22 15:21:50.229930] I [fuse-bridge.c:5080:fuse_graph_setup] 0-fuse: switched to graph 0 [2015-04-22 15:21:50.233176] I [fuse-bridge.c:4009:fuse_init] 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 kernel 7.22 [2015-04-22 15:21:50.233244] I [afr-common.c:3722:afr_local_init] 0-VGFS1-replicate-0: no subvolumes up [2015-04-22 15:21:50.234996] I [afr-common.c:3722:afr_local_init] 0-VGFS1-replicate-0: no subvolumes up [2015-04-22 15:21:50.235020] W [fuse-bridge.c:779:fuse_attr_cbk] 0-glusterfs-fuse: 2: LOOKUP() / => -1 (Transport endpoint is not connected) [2015-04-22 15:21:50.237342] I [afr-common.c:3722:afr_local_init] 0-VGFS1-replicate-0: no subvolumes up [2015-04-22 15:21:50.237762] I [fuse-bridge.c:4921:fuse_thread_proc] 0-fuse: unmounting /rhev/data-center/mnt/glusterSD/1kvm1:_VGFS1 [2015-04-22 15:21:50.237980] W [glusterfsd.c:1194:cleanup_and_exit] (--> 0-: received signum (15), shutting down [2015-04-22 15:21:50.237993] I [fuse-bridge.c:5599:fini] 0-fuse: Unmounting '/rhev/data-center/mnt/glusterSD/1kvm1:_VGFS1'.
[root@1kvm1 glusterfs]#


THX a lot
Pa.



_______________________________________________
Users mailing list
[email protected]
http://lists.ovirt.org/mailman/listinfo/users

_______________________________________________
Users mailing list
[email protected]
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to