okay, but how i can compile my app to run this without -Dconfig.file=alt_
reference1.conf?

2015-06-02 15:43 GMT+02:00 Yana Kadiyska <yana.kadiy...@gmail.com>:

> This looks like your app is not finding your Typesafe config. The config
> should usually be placed in particular folder under your app to be seen
> correctly. If it's in a non-standard location you can
> pass  -Dconfig.file=alt_reference1.conf to java to tell it where to look.
> If this is a config that belogs to Spark and not your app, I'd recommend
> running your jar via spark submit (that should run) and dump out the
> classpath/variables that spark submit sets up...
>
> On Tue, Jun 2, 2015 at 6:58 AM, Pa Rö <paul.roewer1...@googlemail.com>
> wrote:
>
>> hello community,
>>
>> i have build a jar file from my spark app with maven (mvn clean compile
>> assembly:single) and the following pom file:
>>
>> <project xmlns="http://maven.apache.org/POM/4.0.0"; xmlns:xsi="
>> http://www.w3.org/2001/XMLSchema-instance";
>>   xsi:schemaLocation="http://maven.apache.org/POM/4.0.0
>> http://maven.apache.org/xsd/maven-4.0.0.xsd";>
>>   <modelVersion>4.0.0</modelVersion>
>>
>>   <groupId>mgm.tp.bigdata</groupId>
>>   <artifactId>ma-spark</artifactId>
>>   <version>0.0.1-SNAPSHOT</version>
>>   <packaging>jar</packaging>
>>
>>   <name>ma-spark</name>
>>   <url>http://maven.apache.org</url>
>>
>>   <properties>
>>     <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
>>   </properties>
>>
>>   <repositories>
>>     <repository>
>>       <id>cloudera</id>
>>       <url>https://repository.cloudera.com/artifactory/cloudera-repos/
>> </url>
>>     </repository>
>>   </repositories>
>>
>>   <dependencies>
>>     <dependency>
>>       <groupId>junit</groupId>
>>       <artifactId>junit</artifactId>
>>       <version>3.8.1</version>
>>       <scope>test</scope>
>>     </dependency>
>>     <dependency>
>>         <groupId>org.apache.spark</groupId>
>>         <artifactId>spark-core_2.10</artifactId>
>>         <version>1.1.0-cdh5.2.5</version>
>>     </dependency>
>>     <dependency>
>>         <groupId>mgm.tp.bigdata</groupId>
>>         <artifactId>ma-commons</artifactId>
>>         <version>0.0.1-SNAPSHOT</version>
>>     </dependency>
>>   </dependencies>
>>
>>   <build>
>>   <plugins>
>>     <plugin>
>>       <artifactId>maven-assembly-plugin</artifactId>
>>       <configuration>
>>         <archive>
>>           <manifest>
>>             <mainClass>mgm.tp.bigdata.ma_spark.SparkMain</mainClass>
>>           </manifest>
>>         </archive>
>>         <descriptorRefs>
>>           <descriptorRef>jar-with-dependencies</descriptorRef>
>>         </descriptorRefs>
>>       </configuration>
>>     </plugin>
>>   </plugins>
>> </build>
>> </project>
>>
>> if i run my app with  java -jar
>> ma-spark-0.0.1-SNAPSHOT-jar-with-dependencies.jar on terminal, i get the
>> following error message:
>>
>> proewer@proewer-VirtualBox:~/Schreibtisch$ java -jar
>> ma-spark-0.0.1-SNAPSHOT-jar-with-dependencies.jar
>> 2015-Jun-02 12:53:36,348 [main] org.apache.spark.util.Utils
>>  WARN  - Your hostname, proewer-VirtualBox resolves to a loopback
>> address: 127.0.1.1; using 10.0.2.15 instead (on interface eth0)
>> 2015-Jun-02 12:53:36,350 [main] org.apache.spark.util.Utils
>>  WARN  - Set SPARK_LOCAL_IP if you need to bind to another address
>> 2015-Jun-02 12:53:36,401 [main] org.apache.spark.SecurityManager
>>  INFO  - Changing view acls to: proewer
>> 2015-Jun-02 12:53:36,402 [main] org.apache.spark.SecurityManager
>>  INFO  - Changing modify acls to: proewer
>> 2015-Jun-02 12:53:36,403 [main] org.apache.spark.SecurityManager
>>  INFO  - SecurityManager: authentication disabled; ui acls disabled;
>> users with view permissions: Set(proewer); users with modify permissions:
>> Set(proewer)
>> Exception in thread "main" com.typesafe.config.ConfigException$Missing:
>> No configuration setting found for key 'akka.version'
>>     at
>> com.typesafe.config.impl.SimpleConfig.findKey(SimpleConfig.java:115)
>>     at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:136)
>>     at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:142)
>>     at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:150)
>>     at com.typesafe.config.impl.SimpleConfig.find(SimpleConfig.java:155)
>>     at
>> com.typesafe.config.impl.SimpleConfig.getString(SimpleConfig.java:197)
>>     at akka.actor.ActorSystem$Settings.<init>(ActorSystem.scala:136)
>>     at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:470)
>>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:111)
>>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:104)
>>     at
>> org.apache.spark.util.AkkaUtils$.org$apache$spark$util$AkkaUtils$$doCreateActorSystem(AkkaUtils.scala:121)
>>     at
>> org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:54)
>>     at
>> org.apache.spark.util.AkkaUtils$$anonfun$1.apply(AkkaUtils.scala:53)
>>     at
>> org.apache.spark.util.Utils$$anonfun$startServiceOnPort$1.apply$mcVI$sp(Utils.scala:1454)
>>     at scala.collection.immutable.Range.foreach$mVc$sp(Range.scala:141)
>>     at org.apache.spark.util.Utils$.startServiceOnPort(Utils.scala:1450)
>>     at
>> org.apache.spark.util.AkkaUtils$.createActorSystem(AkkaUtils.scala:56)
>>     at org.apache.spark.SparkEnv$.create(SparkEnv.scala:156)
>>     at org.apache.spark.SparkContext.<init>(SparkContext.scala:203)
>>     at
>> org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:53)
>>     at mgm.tp.bigdata.ma_spark.SparkMain.main(SparkMain.java:38)
>>
>> what i do wrong?
>>
>> best regards,
>> paul
>>
>>
>

Reply via email to