Well...i'm pretty much certain that at my job i faced something similar..
We had a server with 2 raidz2 groups each with 3 drives, and one drive
has failed and replaced by a hot spare. However, the balance of data
between the 2 groups of raidz2 start to be imbalance.
I assume this is due the fact that during the resilvering process (due
to the hot spare thing) the vdev was somehow "halted" of writting new
data..therefore during that time only 1 group was getting new data, and
that lead to an imbalance data across the 2 raidz2 groups..

Maybe you have something similar, or maybe i'm talking a huge mistake.
If someone with more knowledge about zfs would like to comment, please
do so.. It's always a learning experience.

Bruno


On 25-3-2010 11:53, Ian Collins wrote:
> On 03/25/10 11:23 PM, Bruno Sousa wrote:
>> On 25-3-2010 9:46, Ian Collins wrote:
>>   
>>> On 03/25/10 09:32 PM, Bruno Sousa wrote:
>>>     
>>>> On 24-3-2010 22:29, Ian Collins wrote:
>>>>
>>>>       
>>>>> On 02/28/10 08:09 PM, Ian Collins wrote:
>>>>>
>>>>>         
>>>>>> I was running zpool iostat on a pool comprising a stripe of raidz2
>>>>>> vdevs that appears to be writing slowly and I notice a considerable
>>>>>> imbalance of both free space and write operations.  The pool is
>>>>>> currently feeding a tape backup while receiving a large filesystem.
>>>>>>
>>>>>> Is this imbalance normal?  I would expect a more even
>>>>>> distribution as
>>>>>> the poll configuration hasn't been changed since creation.
>>>>>>
>>>>>> The system is running Solaris 10 update 7.
>>>>>>
>>>>>>                   capacity     operations    bandwidth
>>>>>> pool           used  avail   read  write   read  write
>>>>>> ------------  -----  -----  -----  -----  -----  -----
>>>>>> tank          15.9T  2.19T     87    119  2.34M  1.88M
>>>>>>    raidz2      2.90T   740G     24     27   762K  95.5K
>>>>>>    raidz2      3.59T  37.8G     20      0   546K      0
>>>>>>    raidz2      3.58T  44.1G     27      0  1.01M      0
>>>>>>    raidz2      3.05T   587G      7     47  24.9K  1.07M
>>>>>>    raidz2      2.81T   835G      8     45  30.9K   733K
>>>>>> ------------  -----  -----  -----  -----  -----  -----
>>>>>>
>>>>>>
>>>>>>            
>>>>> This system has since been upgraded, but the imbalance in getting
>>>>> worse:
>>>>>
>>>>> zpool iostat -v tank | grep raid
>>>>>     raidz2      3.60T  28.5G    166     41  6.97M   764K
>>>>>     raidz2      3.59T  33.3G    170     35  7.35M   709K
>>>>>     raidz2      3.60T  26.1G    173     35  7.36M   658K
>>>>>     raidz2      1.69T  1.93T    129     46  6.70M   610K
>>>>>     raidz2      2.25T  1.38T    124     54  5.77M   967K
>>>>>
>>>>> Is there any way to determine how this is happening?
>>>>>
>>>>> I may have to resort to destroying and recreating some large
>>>>> filesystems, but there's no way to determine which ones to target..
>>>>>          
>>>> Hi,
>>>>
>>>> As far as i know this is a "normal" behaviour in ZFS...
>>>> So what we need is somesort of "rebalance" task what moves data around
>>>> multiple vdevs in order to achieve the best performance possible...
>>>> Take a look to
>>>> http://bugs.opensolaris.org/bugdatabase/view_bug.do?bug_id=6855425
>>>>
>>>>
>>>>        
>>> It would be if drives had been added, but this pool hasn't been
>>> changed since it was created
>> Hi,
>>
>> You never experienced any faulted drives, or something similar? So far i
>> only saw imbalance if the vdevs add changed, if a hotspare is used and i
>> think even during a replacement of one disk of a raidz2 group.
>>    
> There has been one faulted drive, but a hot spare kicked in.  A the
> moment, I have another replaced by a spare, but I/O to that vdev is
> unaffected:
>
>   raidz2      1.69T  1.94T    126     46  6.54M   598K
>     spare         -      -    121     34  1.09M  31.4K
>       c0t4d0      -      -     34     23  1.37M  98.7K
>       c7t5d0      -      -      0     78      0   786K
>     c1t2d0        -      -     34     23  1.37M  98.8K
>     c4t3d0        -      -     37     24  1.44M  98.9K
>     c5t3d0        -      -     36     24  1.44M  98.8K
>     c6t2d0        -      -     36     23  1.42M  98.9K
>     c7t1d0        -      -     36     24  1.42M  98.8K
>     c6t1d0        -      -     36     23  1.44M  98.7K
>     c4t1d0        -      -     36     23  1.43M  98.6K
>
> The vdev with the most free space has never lost a drive.
>
> Cheers,
>


Attachment: smime.p7s
Description: S/MIME Cryptographic Signature

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to