On 24 February, 2010 - Bob Friesenhahn sent me these 1,0K bytes:

> On Wed, 24 Feb 2010, Steve wrote:
>>
>> The overhead I was thinking of was more in the pointer structures...  
>> (bearing in mind this is a 128 bit file system), I would guess that  
>> memory requirements would be HUGE for all these files...otherwise arc 
>> is gonna struggle, and paging system is going mental....?
>
> It is not reasonable to assume that zfs has to retain everything in  
> memory.
>
> I have a directory here containing a million files and it has not caused 
> any strain for zfs at all although it can cause considerable stress on 
> applications.
>
> 400 million tiny files is quite a lot and I would hate to use anything  
> but mirrors with so many tiny files.

Another tought is "am I using the correct storage model for this data"?

/Tomas
-- 
Tomas Ögren, st...@acc.umu.se, http://www.acc.umu.se/~stric/
|- Student at Computing Science, University of Umeå
`- Sysadmin at {cs,acc}.umu.se
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to