Follow-up Comment #1, sr #106554 (project administration):
Gluster Server Side Configuration [EMAIL PROTECTED] ~# cat /etc/glusterfs/glusterfs-server.vol ## Define the stroage volume fs1-storage type storage/posix # POSIX FS translator option directory /storage # Export this directory end-volume volume iothreads #iothreads can give performance a boost type performance/io-threads option thread-count 16 subvolumes fs1-storage end-volume ## Add network serving capability to above brick. volume server type protocol/server option transport-type tcp/server # For TCP/IP transport option listen-port 6996 # Default is 6996 option client-volume-filename /var/log/glusterfs/client.vol subvolumes iothreads option auth.ip.iothreads.allow * # Allow access to "brick" volume end-volume Gluster Client Side Configuration [EMAIL PROTECTED] ~# cat /etc/glusterfs/glusterfs-client.vol ### Add client feature and attach to remote subvolume volume client_211 type protocol/client option transport-type tcp/client # for TCP/IP transport option remote-host 10.10.1.211 # IP address of the remote brick option remote-port 6996 # default server port is 6996 option remote-subvolume iothreads # name of the remote volume # option transport-timeout 4 end-volume volume afrbricks type cluster/afr subvolumes client_211 option replicate *:1 option self-heal on end-volume volume iothreads #iothreads can give performance a boost type performance/io-threads option thread-count 8 subvolumes afrbricks end-volume _______________________________________________________ Reply to this item at: <http://savannah.gnu.org/support/?106554> _______________________________________________ Message sent via/by Savannah http://savannah.gnu.org/