I have copied my code in the below link (Julia user group), let me know if you cant access that https://discourse.julialang.org/t/julia-call-from-python3-running-in-single-core/508/5
On Wed, Nov 23, 2016 at 2:03 PM, Douglas Bates <dmba...@gmail.com> wrote: > On Tuesday, November 22, 2016 at 1:12:26 PM UTC-6, Harish Kumar wrote: > > I found the cause for this ... When i run julia 0.3.2 or 0.5 as > standalone (mix model) it uses all the available cores from my server, so > it was fast. > > Fitting a linear mixed effects model only uses multiple threads for the > BLAS (Basic Linear Algebra Subroutine) calls and a few LAPACK calls. In > Julia v0.5 you may be able to set the number of threads for the BLAS by > calling, say, > > BLAS.set_num_threads(4) > > (or some other number) before trying to fit a model. Be aware that > increasing the number of threads doesn't always make things faster. You > may need to do a bit of experimentation to determine a suitable number of > threads. > > Can you describe the formula you are using? If you are trying to fit a > "maximal" model with large-dimensional vector-valued random effects for > crossed grouping factors you should be aware that most of the time fitting > such models is just a convenient way of burning up a lot of computing time. > > > > If i call Julia from Python (Pyjulia), i see only one core is busy with > python process (100% cpu) and all other cores are free. Can you help me > how can i force Pyjulia/python to use available cores from my server? > > > > > > Regards, > > Harish > > > > > > > > > > > > > > > > > > On Sat, Nov 19, 2016 at 8:32 PM, Mauro <maur...@runbox.com> wrote: > > On Sat, 2016-11-19 at 20:48, Harish Kumar <harish....@gmail.com> wrote: > > > > > Thank you. I agree on python.. but my question was did they update the > > > > > Pyjulia libraries for latest Julia version? . We tried with 0.4.3 which > > > > > failed 6 months back. So we revered to 0.3.4. Or is this library remain > > > > > same for all Julia versions? > > > > > > > > > > Any suggestion on this? > > > > > > > > They are testing against the latest release, i.e. 0.5: > > > > https://github.com/JuliaPy/pyjulia/blob/master/.travis.yml > > > > > > > > You should try and file an issue if it doesn't work. 6 months are a > > > > long time at the current julia development pace. > > > > > > > > > > > > > > > > > > On Sat, Nov 19, 2016 at 7:38 PM, Mauro <maur...@runbox.com> wrote: > > > > > > > > > >> On Sat, 2016-11-19 at 18:36, Harish Kumar <harish....@gmail.com> > > > > >> wrote: > > > > >> > Will it support Python 3.4 ? I am calling this from pyjulia > interface > > > > >> > > > > >> https://github.com/JuliaPy/pyjulia says that it is tested against > 3.5, > > > > >> but it doesn't say that 3.4 is not supported. So you should try. > > > > >> > > > > >> > On Nov 19, 2016 4:58 PM, "Mauro" <maur...@runbox.com> wrote: > > > > >> > > > > > >> >> Julia 0.3.12, that's a stone-age version of Julia. You should > move to > > > > >> 0.5! > > > > >> >> > > > > >> >> On Sat, 2016-11-19 at 16:42, Harish Kumar <harish....@gmail.com> > > > > >> >> wrote: > > > > >> >> > I am using Version 0.3.12 calling from python (pyjulia). I do > LME fit > > > > >> >> with > > > > >> >> > 2.8 M rows and 60-70 Variables. It is taking 2 hours just to > model (+ > > > > >> >> data > > > > >> >> > transfer time). Any tips? > > > > >> >> > using MixedModels > > > > >> >> > modelREML = lmm({formula}, dataset) > > > > >> >> > reml!(modelREML,true) > > > > >> >> > lmeModel = fit(modelREML) > > > > >> >> > fixedDF = DataFrame(fixedEffVar = > coeftable(lmeModel).rownms, > > > > >> >> estimate > > > > >> >> > = coeftable(lmeModel).mat[:,1], > > > > >> >> > stdError = coeftable(lmeModel).mat[:,2],zVal > = > > > > >> >> > coeftable(lmeModel).mat[:,3]) > > > > >> >> > > > > > >> >> > On Tuesday, February 23, 2016 at 9:16:47 AM UTC-6, Stefan > Karpinski > > > > >> >> wrote: > > > > >> >> >> > > > > >> >> >> I'm glad that particular slow case got faster! If you want to > submit > > > > >> >> some > > > > >> >> >> reduced version of it as a performance test, we could still > include > > > > >> it > > > > >> >> in > > > > >> >> >> our perf suite. And of course, if you find that anything else > has > > > > >> ever > > > > >> >> >> slowed down, please don't hesitate to file an issue. > > > > >> >> >> > > > > >> >> >> On Tue, Feb 23, 2016 at 9:55 AM, Jonathan Goldfarb < > > > > >> jgol...@gmail.com > > > > >> >> >> <javascript:>> wrote: > > > > >> >> >> > > > > >> >> >>> Yes, understood about difficulty keeping track of regressions. > I was > > > > >> >> >>> originally going to send a message relating up to 2x longer > test > > > > >> time > > > > >> >> on > > > > >> >> >>> the same code on Travis, but it appears as though something has > > > > >> >> changed in > > > > >> >> >>> the nightly build available to CI that now gives significantly > > > > >> faster > > > > >> >> >>> builds, even though the previous poor performance had been > > > > >> >> dependable... > > > > >> >> >>> Evidently that build is not as up-to-date as I thought. Our > code is > > > > >> >> >>> currently not open source, but should be soon after which I can > > > > >> share > > > > >> >> an > > > > >> >> >>> example. > > > > >> >> >>> > > > > >> >> >>> Thanks for your comments, and thanks again for your work on > Julia. > > > > >> >> >>> > > > > >> >> >>> -Max > > > > >> >> >>> > > > > >> >> >>> > > > > >> >> >>> On Monday, February 22, 2016 at 11:12:58 AM UTC-5, Stefan > Karpinski > > > > >> >> wrote: > > > > >> >> >>>> > > > > >> >> >>>> Yes, ideally code should not get slower with new releases – > > > > >> >> >>>> unfortunately, keeping track of performance regressions can > be a > > > > >> bit > > > > >> >> of a > > > > >> >> >>>> game of whack-a-mole. Having examples of code whose speed has > > > > >> >> regressed is > > > > >> >> >>>> very helpful. Thanks to Jarrett Revels excellent work, we now > have > > > > >> >> some > > > > >> >> >>>> great performance regression tracking infrastructure, but of > > > > >> course we > > > > >> >> >>>> always need more things to test! > > > > >> >> >>>> > > > > >> >> >>>> On Mon, Feb 22, 2016 at 9:58 AM, Milan Bouchet-Valat < > > > > >> nali...@club.fr > > > > >> >> > > > > > >> >> >>>> wrote: > > > > >> >> >>>> > > > > >> >> >>>>> Le lundi 22 février 2016 à 06:27 -0800, Jonathan Goldfarb a > écrit > > > > >> : > > > > >> >> >>>>> > I've really been enjoying writing Julia code as a user, and > > > > >> >> following > > > > >> >> >>>>> > the language as it develops, but I have noticed that over > time, > > > > >> >> >>>>> > previously fast code sometimes gets slower, and > (impressively) > > > > >> >> >>>>> > previously slow code will sometimes get faster, with > updates to > > > > >> the > > > > >> >> >>>>> > Julia codebase. > > > > >> >> >>>>> Code is not supposed to get slower with newer releases. If > this > > > > >> >> >>>>> happens, please report the problem here or on GitHub (if > possible > > > > >> >> with > > > > >> >> >>>>> a reproducible example). This will be very helpful to help > > > > >> avoiding > > > > >> >> >>>>> regressions. > > > > >> >> >>>>> > > > > >> >> >>>>> > No complaint here in general; I really appreciate the work > all > > > > >> of > > > > >> >> the > > > > >> >> >>>>> > Core and package developers do, and variations in > performance of > > > > >> >> >>>>> > different codes it to be expected. > > > > >> >> >>>>> > My question is this: has anyone in the Julia community > thought > > > > >> >> about > > > > >> >> >>>>> > updated performance tips for writing high performance code? > > > > >> >> >>>>> > Obviously, using the profiler, along with many of the tips > > > > >> >> >>>>> > at https://github.com/JuliaLang/julia/commits/master/doc/ > > > > >> >> manual/perfo > > > > >> >> >>>>> > rmance-tips.rst still apply, but I am wondering more about > > > > >> >> >>>>> > general/structural ideas to keep in mind in Julia v0.4, as > well > > > > >> as > > > > >> >> >>>>> > guidance on how best to take advantage of recent changes on > > > > >> >> master. I > > > > >> >> >>>>> > know that document hasn't been stagnant in any sense, but > > > > >> >> relatively > > > > >> >> >>>>> > "big in any case, I'd be happy to help make some updates > in a > > > > >> PR if > > > > >> >> >>>>> > there's anything we come up with. > > > > >> >> >>>>> I've just skimmed through this page, and I don't think any > of the > > > > >> >> >>>>> advice given there is outdated. What's new in master is that > > > > >> >> anonymous > > > > >> >> >>>>> functions (and therefore map) are now fast, but that wasn't > > > > >> >> previously > > > > >> >> >>>>> mentioned in the tips as a performance issue anyway. > > > > >> >> >>>>> > > > > >> >> >>>>> The only small sentence which should likely be removed is > "for > > > > >> >> example, > > > > >> >> >>>>> currently it’s not possible to infer the return type of an > > > > >> anonymous > > > > >> >> >>>>> function". Type inference seems to work fine now on master > with > > > > >> >> >>>>> anonymous functions. I'll leave others confirm this. > > > > >> >> >>>>> > > > > >> >> >>>>> Anyway, do you have any specific points in mind? > > > > >> >> >>>>> > > > > >> >> >>>>> > > > > >> >> >>>>> Regards > > > > >> >> >>>>> > > > > >> >> >>>> > > > > >> >> >>>> > > > > >> >> >> > > > > >> >> > > > > >> > >