----- Original Message -----
From: Min Yuan
Sent: Tuesday, May 08, 2001 4:41 PM
Subject: Re: Speed problem in scanning a directory with 500,000 files

 
----- Original Message -----
From: Julie
Sent: Tuesday, April 24, 2001 11:09 AM
Subject: Re: Speed problem in scanning a directory with 500,000 files

From: Min Yuan
 
Hello,
 
We have a directory on Redhat 6.2 with 500, 000 files. In our code we open and read the directory and for each entry in the directory we use lstat() to check for some information. The whole scanning takes more than eight hours which is terribly long.
 
Is there any way we could reduce this length of time? If the answer is NO, then is there any official documents about it and where can we find it?
>>Yes.  Stop putting so many files into a single directory.
 
Besides this, there is no other ways? This is not the right solution for us. Because we are developing an application which needs to handle large directories on client sites and these 500000 files have to be put in a single directory.
 
Min
 
>>Directory search time becomes linear in the number of entries
>>once the size exceeds any directory name cache capacity.
>>Repeated directory searches then becomes quadratic in the
>>number of entries. 500,000 ^ 2 isn't a small number ...
 
-- Julie.

Reply via email to