Maybe you want to optimize your script first, and I don't think read
entire data set into array would save you much time.
Don't create new variables when its unnecessary, cache data when its
necessary, try memcached,
and I think heavy php cli scripts are always likely to resume a lot of
resource
What I usually do is to pull a limited set of records ( like 10 or 50
) and the do the operations on them, update a column in that table to
mark them completed and use JavaScript to reload the page and pull the
next set out where that flag field is null.
No memory issue, no need to large timeouts
I have a script that connects to an external database to process about 4,000
records. Each record needs to be operated on pretty heavily, and the script
overall takes about half an hour to execute. We've hit a wall where the
script's memory usage exceeds the amount allocated to PHP. I've increased
3 matches
Mail list logo