In terms of Mahout DSL it means implementing a bunch of physical operators such as transpose, A'B or B'A on large row or column partitioned matrices.
Mahout optimizer takes care of simplifying algebraic expressions such as 1+ exp(drm) => drm.apply-unary(1+exp(x)) and tracking things like identical partitioning of datasets where applicable. Adding a back turned out to be pretty trivial for h20 back. I don't think there's a need to bridge flink operations to any of existing backs of DSL for support of Flink. instead i was hoping flink could have its own native back. Also keep in mind that I am not aware of any actual applications with Mahout + h2o back whereas i have built dozens on Spark in past 2 years (albeit on a quite intensely hacked version of the bindings). With Flink it is a bit less trivial i guess because mahout optimizer sometimes tries to run quick summary things such as matrix geometry detection as a part of optimizer action itself. And Flink, last time i checked, did not support running multiple computational actions on a cache-pinned dataset. So that was the main difficulty last time. Perhaps no more. Or perhaps there's a clever way to work around this. not sure. At the same time Flink back for algebraic optimizer also may be more trivial than e.g H20 back was since H20 insisted on marrying their own in-core Matrix api so they created a bridge between Mahout (former Colt) Matrix api and one of their own. Whereas if distributed system just takes care of serializing Writables of Matrices and Vectors (such as it was done for Spark backend) then there's practically nothing to do. Which, my understanding is, Flink is doing good. Now, on the part of current shortcomings, currently there is a fairly long list of performance problems, many of which i fixed internally but not publicly. But hopefully some or all of that will eventually be pushed sooner rather than later. The biggest sticking issue is in-core performance of the Colt apis. Even jvm-based matrices could do much better if they took a bit of cost-based approach doing so, and perhaps integrating a bit more sophisticated Winograd type of gemm approaches. Not to mention integration of things like Magma, jcuda or even simply netlib.org native dense blas stuff (as in Breeze). Good thing though it does not affect backend work since once it is fixed for in-core algebra, it is fixed elsewhere and this is hidden behind scalabindings (in-core algebra DSL). These in-core and out-of-core performance optimization issues are probably the only thing between code base and a good well rounded release. Out of core bad stuff is mostly fixed internally though. Anyone interested in working on any of the issues I mentioned please throw a note to Mahout list to me or Suneel Marthi. On the issue of optimized ML stuff vs. general algebra, here's food for thought. I found that 70% of algorithms are not purely algebraic in R-like operation set sense. But close to 95% could contain significant elements of simplification thru algebra. Even probabilistic things that just use stuff like MVN or Wishart sampling, or Gaussian processes. A lot easier to read and maintain. Even LLoyd iteration has a simple algebraic expression (turns out). Close to 80% could not avoid using some algebra. Only 5% could get away with not doing algebra at all. Thus, 95% of things i ever worked on are either purely or quasi algebraic. Even when they are probabilisitic at core. What it means is that any ML library would benefit enormously if it acts in terms of common algebraic data structures. It will create opportunity for pipelines to seamlessly connect elements of learning such as standardization, dimensionality reduction, MDS/visualization methods, recommender methods as well as clustering methods. My general criticism for MLLib has been that until recently, they did not work on creating such common math structure standards, algebraic data in particular, and every new method's input/output came in its own form an shape. Still does. So dilemma of separating efforts on ones using and not using algebra is a bit false. Most methods are quasi-algebraic (meaning they have at least some need for R-like matrix manipulations). Of course there's need for specific distributed primitives from time to time, there's no arguing about it (like i said, 70% of all methods cannot be purely algebraic, only about 25% are). There's an issue of algebra physical ops performance on spark and in-core, but there's no reason (for me) to believe that effort to fix it in POJO based implementations would require any less effort. There were also some discussion whether and how it is possible to create quasi algebraic methods such that their non-algebraic part is easy to port to another platform (provided algebraic part is already compatible), but that's completely another topic. But what i am saying additional benefit might be that moving ML methodology to platform-agnostic package in Mahout (if such interest ever appears) would also be much easier even for quasi-algebraic approaches if the solutions gravitated to using algebra as opposed to not using it . just some food for thought. thanks. -D On Sun, Jan 4, 2015 at 10:14 AM, Till Rohrmann <trohrm...@apache.org> wrote: > The idea to work with H2O sounds really interesting. > > In terms of the Mahout DSL this would mean that we have to translate a > Flink dataset into H2O's basic abstraction of distributed data and vice > versa. Everything other than writing to disk with one system and reading > from there with the other is probably non-trivial and hard to realize. > On Jan 4, 2015 9:18 AM, "Henry Saputra" <henry.sapu...@gmail.com> wrote: > > > Happy new year all! > > > > Like the idea to add ML module with Flink. > > > > As I have mentioned to Kostas, Stephan, and Robert before, I would > > love to see if we could work with H20 project [1], and it seemed like > > the community has added support for it for Apache Mahout backend > > binding [2]. > > > > So we might get some additional scale ML algos like deep learning. > > > > Definitely would love to help with this initiative =) > > > > - Henry > > > > [1] https://github.com/h2oai/h2o-dev > > [2] https://issues.apache.org/jira/browse/MAHOUT-1500 > > > > On Fri, Jan 2, 2015 at 6:46 AM, Stephan Ewen <se...@apache.org> wrote: > > > Hi everyone! > > > > > > Happy new year, first of all and I hope you had a nice end-of-the-year > > > season. > > > > > > I thought that it is a good time now to officially kick off the > creation > > of > > > a library of machine learning algorithms. There are a lot of individual > > > artifacts and algorithms floating around which we should consolidate. > > > > > > The machine-learning library in Flink would stand on two legs: > > > > > > - A collection of efficient implementations for common problems and > > > algorithms, e.g., Regression (logistic), clustering (k-Means, Canopy), > > > Matrix Factorization (ALS), ... > > > > > > - An adapter to the linear algebra DSL in Apache Mahout. > > > > > > In the long run, it would be the goal to be able to mix and match code > > from > > > both parts. > > > The linear algebra DSL is very convenient when it comes to quickly > > > composing an algorithm, or some custom pre- and post-processing steps. > > > For some complex algorithms, however, a low level system specific > > > implementation is necessary to make the algorithm efficient. > > > Being able to call the tailored algorithms from the DSL would combine > the > > > benefits. > > > > > > > > > As a concrete initial step, I suggest to do the following: > > > > > > 1) We create a dedicated maven sub-project for that ML library > > > (flink-lib-ml). The project gets two sub-projects, one for the > collection > > > of specialized algorithms, one for the Mahout DSL > > > > > > 2) We add the code for the existing specialized algorithms. As followup > > > work, we need to consolidate data types between those algorithms, to > > ensure > > > that they can easily be combined/chained. > > > > > > 3) The code for the Flink bindings to the Mahout DSL will actually > reside > > > in the Mahout project, which we need to add as a dependency to > > flink-lib-ml. > > > > > > 4) We add some examples of Mahout DSL algorithms, and a template how to > > use > > > them within Flink programs. > > > > > > 5) Create a good introductory readme.md, outlining this structure. The > > > readme can also track the implemented algorithms and the ones we put on > > the > > > roadmap. > > > > > > > > > Comments welcome :-) > > > > > > > > > Greetings, > > > Stephan > > >