Lionel MARTIN wrote:

- Don't load large amounts of data into scalars.
Fine. Now I know why. But sometimes, you don't have the choice.

I'd like to know what situations you encounter where you are forced to load
large amounts of data into scalars. I can't think of any.

I don't have any clear situations right here in mind, but we could imagine many: -for example, a bulletin board system, where you are retrieving posted message from a DB. Each message could weigh several dozens of kilo. (especially if you store HTML formatting in the DB) -another example hat comes to my mind is a project (implemented in PHP) where I had to work. Part of the process was retrieving cached HTML template pages from the server and do BB tags parsing before serving the client. Of course, you would tell me that I could retrieve the data chunk by chunk, but this is not as obvious, as some BB tags could spread over several lines. So, this would need to devise an algorithm to be sure we are not cutting in the middle of a tag. In this kind of situation and if the file to be retrieved don't grow too large, I would prefer to retrieve the file all at once, do the processing, serve it to the client, and then, undef the buffer. Far easier than doing chunk by chunk.

Lionel.


Reply via email to