Re: Including data nucleus tools

2014-12-13 Thread spark.dubovsky.jakub
So to answer my own question. It is a bug and there is unmerged PR for that already. https://issues.apache.org/jira/browse/SPARK-2624 https://github.com/apache/spark/pull/3238 Jakub -- Původní zpráva -- Od: spark.dubovsky.ja...@seznam.cz Komu: spark.dubovsky.ja...@seznam.cz Dat

Re: Including data nucleus tools

2014-12-12 Thread spark.dubovsky.jakub
Hi,   I had time to try it again. I submited my app by the same command with these additional options:   --jars lib/datanucleus-api-jdo-3.2.6.jar,lib/datanucleus-core-3.2.10.jar, lib/datanucleus-rdbms-3.2.9.jar   Now an app successfully creates hive context. So my question remains: Is "classpa

Re: Including data nucleus tools

2014-12-06 Thread spark.dubovsky.jakub
Next try. I copied whole dist directory created by make-distribution script to cluster not just assembly jar. Then I used ./bin/spark-submit --num-executors 200 --master yarn-cluster --class org. apache.spark.mllib.CreateGuidDomainDictionary ../spark/root-0.1.jar ${args}  ...to run app again. St

Re: Including data nucleus tools

2014-12-06 Thread spark.dubovsky.jakub
Hi again, I have tried to recompile and run this again with new assembly created by ./make-distribution.sh -Phadoop-2.3 -Dhadoop.version=2.3.0-cdh5.1.3 -Pyarn - Phive -DskipTests It results in exactly the same error. Any other hints? Bonus question: Should the class org.datanucleus.api.jdo. JDOP

Including data nucleus tools

2014-12-05 Thread spark.dubovsky.jakub
Hi all,   I have created assembly jar from 1.2 snapshot source by running [1] which sets correct version of hadoop for our cluster and uses hive profile. I also have written relatively simple test program which starts by reading data from parquet using hive context. I compile the code against as