> On 17 May 2016, at 11:09, Steven Hartland <kill...@multiplay.co.uk> wrote:
> 
>> I understand that, but I don’t think it’s a good that ZFS depends blindly on 
>> a driver feature such
>> as that. Of course, it’s great to exploit it.
>> 
>> I have also noticed that ZFS has a good throttling mechanism for write 
>> operations. A similar
>> mechanism should throttle trim requests so that trim requests don’t clog the 
>> whole system.
> It already does.

I see that there’s a limit to the number of active TRIM requests, but not an 
explicit delay such
as it’s applied to write requests. So, even with a single maximum active TRIM 
request,
it seems that TRIM wins. 


>> 
>>> I’d be extremely hesitant to tossing away TRIMs. They are actually quite 
>>> important for
>>> the FTL in the drive’s firmware to proper manage the NAND wear. More free 
>>> space always
>>> reduces write amplification. It tends to go as 1 / freespace, so simply 
>>> dropping them on
>>> the floor should be done with great reluctance.
>> I understand. I was wondering about choosing the lesser between two evils. A 
>> 15 minute
>> I/O stall (I deleted 2 TB of data, that’s a lot, but not so unrealistic) or 
>> settings trims aside
>> during the peak activity.
>> 
>> I see that I was wrong on that, as a throttling mechanism would be more than 
>> enough probably,
>> unless the system is close to running out of space.
>> 
>> I’ve filed a bug report anyway. And copying to -stable.
>> 
>> 
>> https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=209571
>> 
> TBH it sounds like you may have badly behaved HW, we've used ZFS + TRIM and 
> for years on large production boxes and while we're seen slow down we haven't 
> experienced the total lockups you're describing.

I am using ZFS+TRIM on SATA SSD disks for a very long time. Actually, a single 
SSD I tried at home can TRIM at around 2 GB/s. 
Warner Losh told me that the nvd driver is not currently coalescing the TRIMs, 
which is a disadvantage compared to the ada driver, which does.


> The graphs on you're ticket seem to indicate peak throughput of 250MB/s which 
> is extremely slow for standard SSD's let alone NVMe ones and when you add in 
> the fact you have 10 well it seems like something is VERY wrong.

The pool is a raidz2 vdev with 10 P3500 NVMe disks. That graph is the 
throughput of just one of the disks (the other 9 graphs are identical). Bonnie++
reports around 1.7 Gigabytes/s writing “intelligently”, 1 GB/s “rewriting” and 
almost 2 GB/s “reading intelligently” which, as far as I know, is more or less
reasonable.

The really slow part are the TRIM requests. When destroying the files (it’s 
four concurrent bonnie++ tasks writing a total of 2 Terabytes) 

> I just did a quick test on our DB box here creating and then deleting a 2G 
> file as you describe and I couldn't even spot the delete in the general noise 
> it was so quick to process and that's a 6 disk machine with P3700’s.

Totalling 2 TB? In my case it was FOUR files, 512 GB each.

I’m realy puzzled, 




Borja.



_______________________________________________
freebsd-stable@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"

Reply via email to