There should be one port for communication and transport, both of which are 
defined by this option.


Judging from the current log, I suspect that the port is hard-coded.


Is there any other solution?



















在 2020-12-08 14:50:13,"Strahil Nikolov" <[email protected]> 写道:
>This option is only for the transport:
>option transport.socket.listen-port 24017
>
>So it still needs 24007 for communication.
>
>Maybe there is an option for changing the glusterd port , but I can't find it.
>
>Best Regards,
>Strahil Nikolov
>
>
>
>
>
>
>В вторник, 8 декември 2020 г., 07:52:38 Гринуич+2, sky <[email protected]> 
>написа: 
>
>
>
>
>
>Hi:
>     The firewall has been confirmed to be closed. 
>     The following is the startup log,and the ports of both nodes are 24017:
>          Final graph:                                                         
>                                                                               
>                            
>         
> +------------------------------------------------------------------------------+
>                                                                               
>                 
>       1: volume management                                                    
>                                                                               
>                       
>       2:     type mgmt/glusterd                                               
>                                                                               
>                       
>       3:     option rpc-auth.auth-glusterfs on                                
>                                                                               
>                       
>       4:     option rpc-auth.auth-unix on                                     
>                                                                               
>                       
>       5:     option rpc-auth.auth-null on
>       6:     option rpc-auth-allow-insecure on
>       7:     option transport.listen-backlog 1024
>       8:     option max-port 60999
>       9:     option base-port 49252
>      10:     option event-threads 1
>     11:     option ping-timeout 0
>     12:     option transport.rdma.listen-port 24008
>     13:     option transport.socket.listen-port 24017
>     14:     option transport.socket.read-fail-log off
>     15:     option transport.socket.keepalive-interval 2
>     16:     option transport.socket.keepalive-time 10
>     17:     option transport-type rdma
>     18:     option working-directory /var/lib/glusterd
>     19: end-volume                            
>     20:                                       
>     
> +------------------------------------------------------------------------------+
>     [2020-12-08 05:44:37.300602] I [MSGID: 101190] 
> [event-epoll.c:682:event_dispatch_epoll_worker] 0-epoll: Started thread with 
> index 0
>
>    And the following is the log displayed after executing the command 
> “gluster peer probe dev2”. It shows that the connected port is 24007. Is this 
> port hard-coded?
>
> [2020-12-08 05:45:23.136309] I [MSGID: 106487] 
> [glusterd-handler.c:1082:__glusterd_handle_cli_probe] 0-glusterd: Received 
> CLI probe req dev2 24007 
>[2020-12-08 05:45:23.138990] I [MSGID: 106128] 
>[glusterd-handler.c:3541:glusterd_probe_begin] 0-glusterd: Unable to find 
>peerinfo for host: dev2 (24007) 
>[2020-12-08 05:45:23.208257] W [MSGID: 106061] 
>[glusterd-handler.c:3315:glusterd_transport_inet_options_build] 0-glusterd: 
>Failed to get tcp-user-timeout 
>[2020-12-08 05:45:23.208387] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 
>0-management: setting frame-timeout to 600
>[2020-12-08 05:45:23.255375] I [MSGID: 106498] 
>[glusterd-handler.c:3470:glusterd_friend_add] 0-management: connect returned 0 
>[2020-12-08 05:45:23.260180] I [MSGID: 106004] 
>[glusterd-handler.c:6204:__glusterd_peer_rpc_notify] 0-management: Peer <dev2> 
>(<00000000-0000-0000-0000-000000000000>), in state <Establishing Connection>, 
>has disconnected from glusterd. 
>[2020-12-08 05:45:23.260997] I [MSGID: 106599] 
>[glusterd-nfs-svc.c:161:glusterd_nfssvc_reconfigure] 0-management: 
>nfs/server.so xlator is not installed 
>[2020-12-08 05:45:23.261412] I [MSGID: 106544] 
>[glusterd.c:152:glusterd_uuid_init] 0-management: retrieved UUID: 
>769fa9d7-a204-4c44-a15e-b5eac367e322 
>[2020-12-08 05:45:23.261480] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 
>0-quotad: setting frame-timeout to 600
>[2020-12-08 05:45:23.261737] I [MSGID: 106131] 
>[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: quotad already 
>stopped 
>[2020-12-08 05:45:23.261780] I [MSGID: 106568] 
>[glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: quotad service is 
>stopped 
>[2020-12-08 05:45:23.261915] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 
>0-bitd: setting frame-timeout to 600
>[2020-12-08 05:45:23.262039] I [MSGID: 106131] 
>[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: bitd already 
>stopped 
>[2020-12-08 05:45:23.262064] I [MSGID: 106568] 
>[glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: bitd service is 
>stopped 
>[2020-12-08 05:45:23.262110] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 
>0-scrub: setting frame-timeout to 600
>[2020-12-08 05:45:23.262289] I [MSGID: 106131] 
>[glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: scrub already 
>stopped 
>[2020-12-08 05:45:23.262331] I [MSGID: 106568] 
>[glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: scrub service is 
>stopped
>
>
>
>
>
>
>在 2020-12-08 05:01:12,"Strahil Nikolov" <[email protected]> 写道:
>> Are you sure that the firewall is open. Usually , there could be a firewall 
>> inbetween the nodes.
>> 
>> Also, you can run a tcpdump on nodeA and issue a peer probe from NodeB and 
>> see if there is anything hitting it.
>> 
>> What about enabling trace logs and checking for any clues. You can find 
>> details about the log level at : 
>> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3/html/administration_guide/configuring_the_log_level
>>  Best Regards,
>> Strahil Nikolov
>> 
>> В 10:07 +0800 на 07.12.2020 (пн), sky написа:
>>> Hi: 
>>> 1. The ports of both nodes are changed to 24017
>>> 2. After installation, change the port and start directly (non-restart)
>>> 3. I tried, but the nodes in the pool cannot connect after restarting. The 
>>> log shows an attempt to connect to the default port
>>> 
>>> Steps(both node1 and node2)
>>>    # yum install glusterfs glusterfs-server glusterfs-fuse
>>>    # vim /etc/glusterfs/glusterd.vol
>>>  ....
>>>  option transport.socket.listen-port 24017
>>>  ....
>>>  # systemctl start glusterd
>>> 
>>>  # gluster peer probe node2
>>>  Probe returned with Transport endpoint is not connected
>>> 
>>> After changing the port back to 24007 and restarting, the connection is 
>>> normal
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 在 2020-12-06 19:27:01,"Strahil Nikolov" <[email protected]> 写道:>Did you 
>>> change the port on both nodes ?>Did you restart glusterd on both nodes (the 
>>> one you do peer probe and the other one that is being probed) ?>>Have you 
>>> tried to first peer probe and then change the port on all nodes in the pool 
>>> ?>>>Best Regards,>Strahil Nikolov>>>>>>В петък, 4 декември 2020 г., 
>>> 12:18:42 Гринуич+2, sky <[email protected]> написа: >>>>>>linux version: 
>>> cenots 7.5 >gluster verison: 7.5.1>/etc/glusterfs/glusterd.vol: >volume 
>>> management>    type mgmt/glusterd>    option working-directory 
>>> /var/lib/glusterd>    option transport-type socket,rdma>    option 
>>> transport.socket.keepalive-time 10>    option 
>>> transport.socket.keepalive-interval 2>    option 
>>> transport.socket.read-fail-log off>    option transport.socket.listen-port 
>>> 24017>    option transport.rdma.listen-port 24008>    option ping-timeout 
>>> 0>    option event-threads 1>#   option lock-timer 180>#   option 
>>> transport.address-family inet6>    option base-port 49252>    option 
>>> max-port  60999>end-volume>>I started normally after changing the ports on 
>>> both nodes(port from 24007 to 24017), but I cannot add nodes through the 
>>> ‘gluster peer probe node2‘ command,>It always prompts me: >   Probe 
>>> returned with Transport endpoint is not connected>>>>>>>>>>>> >>> 
>>> >________>>>>Community Meeting Calendar:>>Schedule ->Every 2nd and 4th 
>>> Tuesday at 14:30 IST / 09:00 UTC>Bridge: 
>>> https://meet.google.com/cpu-eiue-hvk>Gluster-users mailing 
>>> list>[email protected]>https://lists.gluster.org/mailman/listinfo/gluster-users
>>> 
>>> 
>>> 
>>>  
>> 
>
>
>
> 
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
[email protected]
https://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to