On Fri, Feb 10, 2012 at 6:43 PM, Milan Bouchet-Valat <nalimi...@club.fr> wrote: > Doesn't sound so bad to me. I don't think you'll find a use case where > 3s will really be a problem.
Actually I have exactly such a case because I have to produce many such documents and a difference of two orders of magnitude in processing matters greatly. > From what Rprof() says, xmlNode() doesn't seem to do anything obviously > wrong. It's just that you're calling it 500 times, so there's some > overhead. You'd need a vectorized version that would handle all the data > in one go My code that patches together strings is also not vectorized and does something in 500 iteration. Still it is 300 times faster than the solution that uses the XML package. Also, I could only bring down the processing time of the code that uses XML package to 3 seconds because (i) the amount of data that was very small and (ii) because the XML document had a very repetitive structure. Documents with a more varied structure don't allow the kind of optimization that you proposed and if they have a non-trivial size the processing time can easily become prohibitive. Anyway, my problem is solved because I'm switched to my own homebrewed xml generation and it works fine. Titus ______________________________________________ R-devel@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-devel