On Sun, Jul 31, 2011 at 6:37 PM, Volker Armin Hemmann
<volkerar...@googlemail.com> wrote:
> Am Sonntag 31 Juli 2011, 10:44:28 schrieb Michael Mol:
>> While I take your point about write-cycle limitations, and I would
>> *assume* you're familiar with the various improvements on
>> wear-leveling technique that have happened over the past *ten years*
>
> yeah, I am. Or let it phrase it differently:
> I know what is claimed.
>
> The problem is, the best wear leveling does not help you if your disk is
> pretty filled up and you still do a lot of writing. 1 000 000 write cycles
> aren't much.

Ok; I wasn't certain, but it sounded like you'd had your head in the
sand (if you'll pardon the expression). It's clear you didn't. I'm
sorry.

>
>> since those concerns were first raised, I could probably raise an
>> argument that a fresh SSD is likely to last longer as a swap device
>> than as a filesystem.
>
> depends - because thanks to wear leveling that 'swap partition' is just
> something the firmware makes the kernel believe to be there.
>
>
>>
>> Swap is only touched as-needed, while there's been an explosion in
>> programs and user software which demands synchronous writes to disk
>> for data integrity purposes. (Firefox uses sqlite in such a way, for
>> example; I discovered this when I was using sqlite heavily in my *own*
>> application, and Firefox hung for a couple minutes during every batch
>> insert.)
>
> which is another goof reason not to use firefox - but
>             total       used       free     shared    buffers     cached
> Mem:       8182556    7373736     808820          0      56252    2197064
> -/+ buffers/cache:    5120420    3062136
> Swap:     23446848      82868   23363980
>
> even with lots of ram, you will hit swap. And since you are using the wear-
> leveling of the drive's firmware it does not matter that your swap resides on
> its own partition - every page written means a block-rewrite somewhere. Really
> not good for your ssd.

Fair enough.

It Would Be Nice(tm) if the SSD's block size and alignment matched
that of the kernel's pagesize. Not certain if it's possible to tune
those settings (reliably) in the kernel.

Also, my stats, from three different systems (they appear to be using
trivial amounts of swap, though my Gentoo box doesn't appear to be
using any)

(Desktop box)
shortcircuit:1@serenity~
Sun Jul 31 07:03 PM
!499 #1 j0 ?0 $ free -m
             total       used       free     shared    buffers     cached
Mem:          5975       3718       2256          0        617       1106
-/+ buffers/cache:       1994       3980
Swap:         9993          0       9993

(laptop)
shortcircuit@saffron:~$ free -m
             total       used       free     shared    buffers     cached
Mem:          1995       1732        263          0        169        913
-/+ buffers/cache:        648       1347
Swap:         3921          3       3918

(server)
shortcirc...@rosettacode.xen.prgmr.com~
23:05:34 $ free -m
             total       used       free     shared    buffers     cached
Mem:          2048       2000         47          0        285        488
-/+ buffers/cache:       1225        822
Swap:          511          1        510

>> Also, despite the MBTF data provided by the manufacturers, there's
>> more empirical evidence that the drives expire faster than expected,
>> anyway. I'm aware of this, and not particularly concerned about it.
>
> well, it is your money to burn.

Best evidence I've read lately is that the drives last about a year
under heavy use. I was going to include a reference in the last email,
but I can't find a link to the post. I thought it was something Joel
Spolsky (or *someone* at StackOverflow) wrote, but I was unable to
find it quickly.

My parts usually last 3-5 years, so that's pretty low. Still, having
my swap partition drop (and the entire system halt) would be generally
less damaging to me than having real data on the drive.

>> False dichotomy. Yes, it increases the wear on the device. That says
>> nothing of its impact on system performance, which was the nature of
>> my point.
>
> if you are so concerned of swap performance you should probably go with a
> smaller ssd, get more ram and let that few mb of swap you need been handled by
> several swap partitions.

This is where I get back to my original, 'prohibitively expensive'
bit. I can get 16GB of RAM into my system for about $200. The use
cases where I've been contemplating this have been where I wanted to
have 60GB to 80GB of data quickly accessible in a random-access
fashion, but where that type of load wasn't what I normally spent my
time doing. (Hence the idea to have a broader improvement from
something such as the file cache)

And, really, the whole point of the thread was for thought
experiments. Posits are occasionally required.

>> As for a filecache not being that important, that's only the case if
>> your data of interest exists on the filesystem you put on the SSD.
>>
>> Let's say you're someone like me, who would tend to go with 60GB for /
>> and 3TB for /home. At various times, I'll be doing HDR photo
>> processing, some video transcoding, some random non-portage compile
>> jobs, web browsing, coding, etc.
>
> 60gb for /, 75gb for /var, and 2.5tb data...
> my current setup.

Handy; we'll have common frames of reference.

>> If I take a 160GB SSD, I could put / (or, at least, /var/ and /usr),
>> and have some space left over for scratch--but it's going to be a pain
>> trying to figure out which of my 3TB of /home data I want in that fast
>> scratch.
>>
>> File cache is great, because it takes caches your most-used data from
>> *anywhere* and keeps it in a fast-access datastore. I could have a 3
>> *petabyte* volume, not be particularly concerned about data
>> distribution, and have just as response from the filecache as if I had
>> a mere 30GB volume. Putting a filesystem on an SSD simply cannot scale
>> that way.
>
> true, but all those microseconds saved with swap on ssd won't offset the pain
> when the ssd dies earlier.

It really depends on the quantity and nature of the pain. When the
things I'm toying around with have projected completion times of a
*week* rather than an hour or two, and when I don't normally need so
much memory, it wouldn't be too much of a hassle to remove the dead
drive from fstab and boot back up. (after fsck, etc, natch). In the
words of the Architect, "There are levels of existence we are prepared
to accept..."

>> Actually, this conversation reminds me of another idea I'd had at one
>> point...putting ext3/ext4's journal on an SSD, while keeping the bulk
>> of the data on large, dense spinning platters.
>
> which sounds nice in theory.

Yet would potentially run afoul of the SSD's write block resolution.
And, of course, having the journal fail out from under me would be a
fair bit worse than the kernel panicking during a swap operation.

>> Did you miss the last week's worth of discussion of memory limits on tmpfs?
>
> probably. Because I am using tempfs for /var/tmp/portage for ages and the only
> problematic packet is openoffice/libreoffice.

I ran into trouble with Thunderbird a couple months ago, which is why
I had to drop from using tmpfs. (Also, I compile with -ggdb in CFLAGS,
so I expect my build sizes bloat a bit more than most)

Anyway, the edge cases and caveats like the ones discussed are why I
ask about what people have tried, and what mitigators, workarounds and
technological improvements people have been working on.

-- 
:wq

Reply via email to