Hi all:

    First, thanks for the multiple suggestions.  I'm pretty new at PHP 
programming, so all suggestions are great learning opportunities.

    Now, some observations: I've tried placing the "ignore_user_abort(TRUE);" 
in the code.  It seems to have made little, if any, impact -- the page still 
appears to time out.  I've also tried placing "set_time_limit(0);" both before 
and after the "ignore_user_abort(TRUE);" call.  Still no improvement.

    I'm now wondering if some error is occurring that, for some reason, is 
silently ending the routine.  I'm building what may be a very long SQL INSERT 
statement for each line in the CSV file that I'm reading; could I be hitting 
some upper limit for the length of the SQL code?  I'd think that an error would 
be presented in this case, but maybe I have to do something explicitly to force 
all errors to display?  Even warnings?

    Another thing I've noticed is that the "timeout" (I'm not even certain the 
problem IS a timeout any longer, hence the quotation marks) doesn't happen at 
the same record every time.  That's why I thought it was a timeout problem at 
first, and assumed that the varying load on the server would account for the 
different record numbers processed.  If I were hitting some problem with the 
SQL statement, I'd expect it to stop at the same record every time.  Or is that 
misguided thinking, too?

    More info: I've activated a session (have to be logged into the application 
to update the data), so is it possible that something with the session module 
could be causing the problem?  I have adjusted the session.gc_maxlifetime value 
(per an example I saw in the PHP manual comments elsewhere), but is there some 
other value I should adjust, too?

    Thanks for all of your continued help and assistance!  And please don't 
argue over "best" ways to solve my problem -- ALL suggestions are welcome!  
(Even if I don't know thing one about CLI or even how to access it. <g>)

    Jon

Reply via email to