On March 27, 2020 2:49:13 PM GMT+02:00, Jorick Astrego <[email protected]> 
wrote:
>
>On 3/24/20 7:25 PM, Alex McWhirter wrote:
>> Red hat also recommends a shard size of 512mb, it's actually the only
>> shard size they support. Also check the chunk size on the LVM thin
>> pools running the bricks, should be at least 2mb. Note that changing
>> the shard size only applies to new VM disks after the change.
>Changing
>> the chunk size requires making a new brick.
>>
>Regarding the chunk size, red hat tells me it depends on RAID or JBOD
>
>https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.5/html/administration_guide/Brick_Configuration
>
>    chunksize
>        An important parameter to be specified while creating a thin
>       pool is the chunk size,which is the unit of allocation. For good
>       performance, the chunk size for the thin pool and the parameters
>       of the underlying hardware RAID storage should be chosen so that
>        they work well together.
>
>And regarding the shard size, you can fix that with storage live
>migration right? Use two volumes and domains and move them so they will
>adopt the new shard size...
>
>Am I correct that when you change the sharding on a running volume, it
>only applies for new disks? Or does it also apply to extensions to a
>current disk?
>
>
>
>
>
>
>
>Met vriendelijke groet, With kind regards,
>
>Jorick Astrego
>
>Netbulae Virtualization Experts 
>
>----------------
>
>       Tel: 053 20 30 270      [email protected]        Staalsteden 4-3A        
> KvK 08198180
>       Fax: 053 20 30 271      www.netbulae.eu         7547 TA Enschede        
> BTW
>NL821234584B01
>
>----------------

Shard size change is valid for new images,  but this can be fixed either via 
storage migration between volumes or via creating new disk and migrating within 
the OS (if possible).

Still, MTU is important and you can use
'ping -s  <size_of_data>  -c 1 -M do destination ' to test.

Keep in mind that VLANs also  take some data in the packet (I think around 8 
bytes). Today I have set MTU 9100 on some servers in order to guarantee  that 
the app will be able to transfer 9000 bytes of data,  but this depends on the 
switches between the nodes and the NICs of the servers.

You can use  tracepath to detect if there  is a switch that doesn't support 
Jumbo Frames.

Actually setting ctdb with NFS Ganesha is quite easy  .  You will be able to 
get all 'goodies' from oVirt  (snapshots, live migration, etc) while using 
Higher performance via NFS Ganesha - which is like a gateway for the clients  
(while accessing all servers simultaneously),  so it will be better situated 
outside the Gluster servers.

Best Regards,
Strahil Nikolov
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/V2XENQHC5XZUDXV3OKNGOHSD4CDQVPFU/

Reply via email to