Works for me too....you are a life-saver :)

But the question: should/how we report this to Azure team?

On Fri, May 12, 2017 at 10:32 AM, Denny Lee <denny.g....@gmail.com> wrote:

> I was able to repro your issue when I had downloaded the jars via blob but
> when I downloaded them as raw, I was able to get everything up and
> running.  For example:
>
> wget https://github.com/Azure/azure-documentdb-spark/*blob*/
> master/releases/azure-documentdb-spark-0.0.3_2.0.2_
> 2.11/azure-documentdb-1.10.0.jar
> wget https://github.com/Azure/azure-documentdb-spark/*blob*/
> master/releases/azure-documentdb-spark-0.0.3_2.0.2_
> 2.11/azure-documentdb-spark-0.0.3-SNAPSHOT.jar
> spark-shell --master yarn --jars azure-documentdb-spark-0.0.3-
> SNAPSHOT.jar,azure-documentdb-1.10.0.jar
>
> resulted in the error:
> SPARK_MAJOR_VERSION is set to 2, using Spark2
> Setting default log level to "WARN".
> To adjust logging level use sc.setLogLevel(newLevel).
> [init] error: error while loading <root>, Error accessing
> /home/sshuser/jars/test/azure-documentdb-spark-0.0.3-SNAPSHOT.jar
>
> Failed to initialize compiler: object java.lang.Object in compiler mirror
> not found.
> ** Note that as of 2.8 scala does not assume use of the java classpath.
> ** For the old behavior pass -usejavacp to scala, or if using a Settings
> ** object programmatically, settings.usejavacp.value = true.
>
> But when running:
> wget https://github.com/Azure/azure-documentdb-spark/raw/
> master/releases/azure-documentdb-spark-0.0.3_2.0.2_
> 2.11/azure-documentdb-1.10.0.jar
> wget https://github.com/Azure/azure-documentdb-spark/raw/
> master/releases/azure-documentdb-spark-0.0.3_2.0.2_
> 2.11/azure-documentdb-spark-0.0.3-SNAPSHOT.jar
> spark-shell --master yarn --jars azure-documentdb-spark-0.0.3-
> SNAPSHOT.jar,azure-documentdb-1.10.0.jar
>
> it was up and running:
> spark-shell --master yarn --jars azure-documentdb-spark-0.0.3-
> SNAPSHOT.jar,azure-documentdb-1.10.0.jar
> SPARK_MAJOR_VERSION is set to 2, using Spark2
> Setting default log level to "WARN".
> To adjust logging level use sc.setLogLevel(newLevel).
> 17/05/11 22:54:06 WARN SparkContext: Use an existing SparkContext, some
> configuration may not take effect.
> Spark context Web UI available at http://10.0.0.22:4040
> Spark context available as 'sc' (master = yarn, app id =
> application_1494248502247_0013).
> Spark session available as 'spark'.
> Welcome to
>       ____              __
>      / __/__  ___ _____/ /__
>     _\ \/ _ \/ _ `/ __/  '_/
>    /___/ .__/\_,_/_/ /_/\_\   version 2.0.2.2.5.4.0-121
>       /_/
>
> Using Scala version 2.11.8 (OpenJDK 64-Bit Server VM, Java 1.8.0_121)
> Type in expressions to have them evaluated.
> Type :help for more information.
>
> scala>
>
> HTH!
>
>
> On Wed, May 10, 2017 at 11:49 PM ayan guha <guha.a...@gmail.com> wrote:
>
>> Hi
>>
>> Thanks for reply, but unfortunately did not work. I am getting same error.
>>
>> sshuser@ed0-svochd:~/azure-spark-docdb-test$ spark-shell --jars
>> azure-documentdb-spark-0.0.3-SNAPSHOT.jar,azure-documentdb-1.10.0.jar
>> SPARK_MAJOR_VERSION is set to 2, using Spark2
>> Setting default log level to "WARN".
>> To adjust logging level use sc.setLogLevel(newLevel).
>> [init] error: error while loading <root>, Error accessing
>> /home/sshuser/azure-spark-docdb-test/azure-documentdb-
>> spark-0.0.3-SNAPSHOT.jar
>>
>> Failed to initialize compiler: object java.lang.Object in compiler mirror
>> not found.
>> ** Note that as of 2.8 scala does not assume use of the java classpath.
>> ** For the old behavior pass -usejavacp to scala, or if using a Settings
>> ** object programmatically, settings.usejavacp.value = true.
>>
>> Failed to initialize compiler: object java.lang.Object in compiler mirror
>> not found.
>> ** Note that as of 2.8 scala does not assume use of the java classpath.
>> ** For the old behavior pass -usejavacp to scala, or if using a Settings
>> ** object programmatically, settings.usejavacp.value = true.
>> Exception in thread "main" java.lang.NullPointerException
>>         at scala.reflect.internal.SymbolTable.exitingPhase(
>> SymbolTable.scala:256)
>>         at scala.tools.nsc.interpreter.IMain$Request.x$20$lzycompute(
>> IMain.scala:896)
>>         at scala.tools.nsc.interpreter.IMain$Request.x$20(IMain.
>> scala:895)
>>         at scala.tools.nsc.interpreter.IMain$Request.headerPreamble$
>> lzycompute(IMain.scala:895)
>>         at scala.tools.nsc.interpreter.IMain$Request.headerPreamble(
>> IMain.scala:895)
>>         at scala.tools.nsc.interpreter.IMain$Request$Wrapper.
>> preamble(IMain.scala:918)
>>         at scala.tools.nsc.interpreter.IMain$CodeAssembler$$anonfun$
>> apply$23.apply(IMain.scala:1337)
>>         at scala.tools.nsc.interpreter.IMain$CodeAssembler$$anonfun$
>> apply$23.apply(IMain.scala:1336)
>>         at scala.tools.nsc.util.package$.stringFromWriter(package.
>> scala:64)
>>         at scala.tools.nsc.interpreter.IMain$CodeAssembler$class.
>> apply(IMain.scala:1336)
>>         at scala.tools.nsc.interpreter.IMain$Request$Wrapper.apply(
>> IMain.scala:908)
>>         at scala.tools.nsc.interpreter.IMain$Request.compile$
>> lzycompute(IMain.scala:1002)
>>         at scala.tools.nsc.interpreter.IMain$Request.compile(IMain.
>> scala:997)
>>         at scala.tools.nsc.interpreter.IMain.compile(IMain.scala:579)
>>         at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:567)
>>         at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:565)
>>         at scala.tools.nsc.interpreter.ILoop.interpretStartingWith(
>> ILoop.scala:807)
>>         at scala.tools.nsc.interpreter.ILoop.command(ILoop.scala:681)
>>         at scala.tools.nsc.interpreter.ILoop.processLine(ILoop.scala:395)
>>         at org.apache.spark.repl.SparkILoop$$anonfun$
>> initializeSpark$1.apply$mcV$sp(SparkILoop.scala:38)
>>         at org.apache.spark.repl.SparkILoop$$anonfun$
>> initializeSpark$1.apply(SparkILoop.scala:37)
>>         at org.apache.spark.repl.SparkILoop$$anonfun$
>> initializeSpark$1.apply(SparkILoop.scala:37)
>>         at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.
>> scala:214)
>>         at org.apache.spark.repl.SparkILoop.initializeSpark(
>> SparkILoop.scala:37)
>>         at org.apache.spark.repl.SparkILoop.loadFiles(
>> SparkILoop.scala:94)
>>         at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.
>> apply$mcZ$sp(ILoop.scala:920)
>>         at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.
>> apply(ILoop.scala:909)
>>         at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.
>> apply(ILoop.scala:909)
>>         at scala.reflect.internal.util.ScalaClassLoader$.
>> savingContextLoader(ScalaClassLoader.scala:97)
>>         at scala.tools.nsc.interpreter.ILoop.process(ILoop.scala:909)
>>         at org.apache.spark.repl.Main$.doMain(Main.scala:68)
>>         at org.apache.spark.repl.Main$.main(Main.scala:51)
>>         at org.apache.spark.repl.Main.main(Main.scala)
>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>         at sun.reflect.NativeMethodAccessorImpl.invoke(
>> NativeMethodAccessorImpl.java:62)
>>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>> DelegatingMethodAccessorImpl.java:43)
>>         at java.lang.reflect.Method.invoke(Method.java:498)
>>         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
>> deploy$SparkSubmit$$runMain(SparkSubmit.scala:736)
>>         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
>> SparkSubmit.scala:185)
>>         at org.apache.spark.deploy.SparkSubmit$.submit(
>> SparkSubmit.scala:210)
>>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
>> scala:124)
>>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>> sshuser@ed0-svochd:~/azure-spark-docdb-test$
>>
>>
>> On Mon, May 8, 2017 at 11:50 PM, Denny Lee <denny.g....@gmail.com> wrote:
>>
>>> This appears to be an issue with the Spark to DocumentDB connector,
>>> specifically version 0.0.1. Could you run the 0.0.3 version of the jar and
>>> see if you're still getting the same error?  i.e.
>>>
>>>     spark-shell --master yarn --jars azure-documentdb-spark-0.0.3-
>>> SNAPSHOT.jar,azure-documentdb-1.10.0.jar
>>>
>>>
>>> On Mon, May 8, 2017 at 5:01 AM ayan guha <guha.a...@gmail.com> wrote:
>>>
>>>> Hi
>>>>
>>>> I am facing an issue while trying to use azure-document-db connector
>>>> from Microsoft. Instructions/Github
>>>> <https://github.com/Azure/azure-documentdb-spark/wiki/Azure-DocumentDB-Spark-Connector-User-Guide>
>>>> .
>>>>
>>>> Error while trying to add jar in spark-shell:
>>>>
>>>> spark-shell --jars azure-documentdb-spark-0.0.1.
>>>> jar,azure-documentdb-1.9.6.jar
>>>> SPARK_MAJOR_VERSION is set to 2, using Spark2
>>>> Setting default log level to "WARN".
>>>> To adjust logging level use sc.setLogLevel(newLevel).
>>>> [init] error: error while loading <root>, Error accessing
>>>> /home/sshuser/azure-spark-docdb-test/v1/azure-
>>>> documentdb-spark-0.0.1.jar
>>>>
>>>> Failed to initialize compiler: object java.lang.Object in compiler
>>>> mirror not found.
>>>> ** Note that as of 2.8 scala does not assume use of the java classpath.
>>>> ** For the old behavior pass -usejavacp to scala, or if using a Settings
>>>> ** object programmatically, settings.usejavacp.value = true.
>>>>
>>>> Failed to initialize compiler: object java.lang.Object in compiler
>>>> mirror not found.
>>>> ** Note that as of 2.8 scala does not assume use of the java classpath.
>>>> ** For the old behavior pass -usejavacp to scala, or if using a Settings
>>>> ** object programmatically, settings.usejavacp.value = true.
>>>> Exception in thread "main" java.lang.NullPointerException
>>>>         at scala.reflect.internal.SymbolTable.exitingPhase(
>>>> SymbolTable.scala:256)
>>>>         at scala.tools.nsc.interpreter.IMain$Request.x$20$lzycompute(
>>>> IMain.scala:896)
>>>>         at scala.tools.nsc.interpreter.IMain$Request.x$20(IMain.
>>>> scala:895)
>>>>         at scala.tools.nsc.interpreter.IMain$Request.headerPreamble$
>>>> lzycompute(IMain.scala:895)
>>>>         at scala.tools.nsc.interpreter.IMain$Request.headerPreamble(
>>>> IMain.scala:895)
>>>>         at scala.tools.nsc.interpreter.IMain$Request$Wrapper.
>>>> preamble(IMain.scala:918)
>>>>         at scala.tools.nsc.interpreter.IMain$CodeAssembler$$anonfun$
>>>> apply$23.apply(IMain.scala:1337)
>>>>         at scala.tools.nsc.interpreter.IMain$CodeAssembler$$anonfun$
>>>> apply$23.apply(IMain.scala:1336)
>>>>         at scala.tools.nsc.util.package$.stringFromWriter(package.
>>>> scala:64)
>>>>         at scala.tools.nsc.interpreter.IMain$CodeAssembler$class.
>>>> apply(IMain.scala:1336)
>>>>         at scala.tools.nsc.interpreter.IMain$Request$Wrapper.apply(
>>>> IMain.scala:908)
>>>>         at scala.tools.nsc.interpreter.IMain$Request.compile$
>>>> lzycompute(IMain.scala:1002)
>>>>         at scala.tools.nsc.interpreter.IMain$Request.compile(IMain.
>>>> scala:997)
>>>>         at scala.tools.nsc.interpreter.IMain.compile(IMain.scala:579)
>>>>         at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:567)
>>>>         at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:565)
>>>>         at scala.tools.nsc.interpreter.ILoop.interpretStartingWith(
>>>> ILoop.scala:807)
>>>>         at scala.tools.nsc.interpreter.ILoop.command(ILoop.scala:681)
>>>>         at scala.tools.nsc.interpreter.ILoop.processLine(ILoop.scala:
>>>> 395)
>>>>         at org.apache.spark.repl.SparkILoop$$anonfun$
>>>> initializeSpark$1.apply$mcV$sp(SparkILoop.scala:38)
>>>>         at org.apache.spark.repl.SparkILoop$$anonfun$
>>>> initializeSpark$1.apply(SparkILoop.scala:37)
>>>>         at org.apache.spark.repl.SparkILoop$$anonfun$
>>>> initializeSpark$1.apply(SparkILoop.scala:37)
>>>>         at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.
>>>> scala:214)
>>>>         at org.apache.spark.repl.SparkILoop.initializeSpark(
>>>> SparkILoop.scala:37)
>>>>         at org.apache.spark.repl.SparkILoop.loadFiles(
>>>> SparkILoop.scala:94)
>>>>         at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.
>>>> apply$mcZ$sp(ILoop.scala:920)
>>>>         at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.
>>>> apply(ILoop.scala:909)
>>>>         at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.
>>>> apply(ILoop.scala:909)
>>>>         at scala.reflect.internal.util.ScalaClassLoader$.
>>>> savingContextLoader(ScalaClassLoader.scala:97)
>>>>         at scala.tools.nsc.interpreter.ILoop.process(ILoop.scala:909)
>>>>         at org.apache.spark.repl.Main$.doMain(Main.scala:68)
>>>>         at org.apache.spark.repl.Main$.main(Main.scala:51)
>>>>         at org.apache.spark.repl.Main.main(Main.scala)
>>>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>>         at sun.reflect.NativeMethodAccessorImpl.invoke(
>>>> NativeMethodAccessorImpl.java:62)
>>>>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>>> DelegatingMethodAccessorImpl.java:43)
>>>>         at java.lang.reflect.Method.invoke(Method.java:498)
>>>>         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
>>>> deploy$SparkSubmit$$runMain(SparkSubmit.scala:736)
>>>>         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
>>>> SparkSubmit.scala:185)
>>>>         at org.apache.spark.deploy.SparkSubmit$.submit(
>>>> SparkSubmit.scala:210)
>>>>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
>>>> scala:124)
>>>>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>>>> sshuser@ed0-svochd:~/azure-spark-docdb-test/v1$
>>>>
>>>> I think I am missing some basic configuration here or there is
>>>> classpath related issue. Can anyone help?
>>>>
>>>> Additional info:
>>>> Environment: HDInsight 3.5, based on HDP 2.5
>>>>
>>>> sshuser@ed0-svochd:~/azure-spark-docdb-test/v1$ echo $JAVA_HOME
>>>> /usr/lib/jvm/java-8-openjdk-amd64
>>>>
>>>> sshuser@ed0-svochd:~/azure-spark-docdb-test/v1$ echo $SPARK_HOME
>>>> /usr/hdp/current/spark2-client
>>>>
>>>> sshuser@ed0-svochd:~/azure-spark-docdb-test/v1$ java -version
>>>> openjdk version "1.8.0_121"
>>>> OpenJDK Runtime Environment (build 1.8.0_121-8u121-b13-0ubuntu1.
>>>> 16.04.2-b13)
>>>> OpenJDK 64-Bit Server VM (build 25.121-b13, mixed mode)
>>>>
>>>> sshuser@ed0-svochd:~/azure-spark-docdb-test/v1$ uname -a
>>>> Linux ed0-svochd 4.4.0-72-generic #93-Ubuntu SMP Fri Mar 31 14:07:41
>>>> UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
>>>> sshuser@ed0-svochd:~/azure-spark-docdb-test/v1$
>>>>
>>>> --
>>>> Best Regards,
>>>> Ayan Guha
>>>>
>>>
>>
>>
>> --
>> Best Regards,
>> Ayan Guha
>>
>


-- 
Best Regards,
Ayan Guha

Reply via email to