How optimized are the Commons math3 methods that showed up in profiling?
Are there any higher performance alternatives to these?
marko
Hi,
I did some profiling for my LDA prototype code that requests topic
distributions from a model.
According to Java Mission Control more than 80 % of execution time during
sample interval is spent in the following methods:
org.apache.commons.math3.util.FastMath.log(double); count: 337; 47.07%
or
Hi Feynman,
I just tried that, but there wasn't a noticeable change in training
performance. On the other hand model loading time was reduced to ~ 5
seconds from ~ 2 minutes (now persisted as LocalLDAModel).
However, query / prediction time was unchanged.
Unfortunately, this is the critical perfo
end-user operation execution flow, so a ~4 second
execution time is very problematic. Am I using the MLlib LDA API correctly
or is this just reflecting the current performance characteristics of the
LDA implementation? My code can be found here:
https://github.com/marko-asplund/tech-protos/blob
ng can be found here:
https://github.com/marko-asplund/tech-protos/blob/master/mllib-lda/src/main/scala/fi/markoa/proto/mllib/LDADemo.scala#L33-L47
marko
On 2015-09-11 at 16:08:44 +0200 Carsten Schnober wrote:
Hi,
> I don't have practical experience with the MLlib LDA implementation, but
> regarding the variations in the topic matrix: LDA make use of stochastic
> processes. If you use setSeed(seed) with the same value for seed during
> initializati
the LDA algorithm?
Any caveats to be aware of with the LDA implementation?
For reference, my prototype code can be found here:
https://github.com/marko-asplund/tech-protos/blob/master/mllib-lda/src/main/scala/fi/markoa/proto/mllib/LDADemo.scala
thanks,
marko