If you can test it quickly, an option would be to try with the exact same
version that Sean used (1.7.0_51) ?

Maybe it was a bug fixed in 51 and a regression has been introduced in 55
:-D

Andy
On Thu, Apr 17, 2014 at 9:36 PM, Steven Cox <s...@renci.org> wrote:

>  FYI, I've tried older versions (jdk6.x), openjdk. Also here's a fresh
> core dump on jdk7u55-b13:
>
>  # A fatal error has been detected by the Java Runtime Environment:
>
> #
>
> #  SIGSEGV (0xb) at pc=0x00007f7c6b718d39, pid=7708, tid=140171900581632
>
> #
>
> # JRE version: Java(TM) SE Runtime Environment (7.0_55-b13) (build
> 1.7.0_55-b13)
>
> # Java VM: Java HotSpot(TM) 64-Bit Server VM (24.55-b03 mixed mode
> linux-amd64 compressed oops)
>
> # Problematic frame:
>
> # V  [libjvm.so+0x632d39]  jni_GetByteArrayElements+0x89
>
> #
>
> # Failed to write core dump. Core dumps have been disabled. To enable core
> dumping, try "ulimit -c unlimited" before starting Java again
>
> #
>
> # An error report file with more information is saved as:
>
> # /home/scox/skylr/skylr-analytics/hs_err_pid7708.log
>
> #
>
> # If you would like to submit a bug report, please visit:
>
> #   http://bugreport.sun.com/bugreport/crash.jsp
>
>
>  Steve
>
>
>  ------------------------------
> *From:* andy petrella [andy.petre...@gmail.com]
> *Sent:* Thursday, April 17, 2014 3:21 PM
> *To:* user@spark.apache.org
> *Subject:* Re: Spark 0.9.1 core dumps on Mesos 0.18.0
>
>   No of course, but I was guessing some native libs imported (to
> communicate with Mesos) in the project that... could miserably crash the
> JVM.
>
>  Anyway, so you tell us that using this oracle version, you don't have
> any issues when using spark on mesos 0.18.0, that's interesting 'cause
> AFAIR, my last test (done by night, which means floating and eventual
> memory) I was using this particular version as well.
>
>  Just to make thing clear, Sean, you're using spark 0.9.1 on Mesos 0.18.0
> with Hadoop 2.x (x >= 2) without any modification than just specifying
> against which version of hadoop you had run make-distribution?
>
>  Thanks for your help,
>
>  Andy
>
> On Thu, Apr 17, 2014 at 9:11 PM, Sean Owen <so...@cloudera.com> wrote:
>
>> I don't know if it's anything you or the project is missing... that's
>> just a JDK bug.
>> FWIW I am on 1.7.0_51 and have not seen anything like that.
>>
>> I don't think it's a protobuf issue -- you don't crash the JVM with
>> simple version incompatibilities :)
>> --
>> Sean Owen | Director, Data Science | London
>>
>>
>> On Thu, Apr 17, 2014 at 7:29 PM, Steven Cox <s...@renci.org> wrote:
>>  > So I tried a fix found on the list...
>> >
>> >    "The issue was due to meos version mismatch as I am using latest
>> mesos
>> > 0.17.0, but spark uses 0.13.0.
>> >     Fixed by updating the SparkBuild.scala to latest version."
>> >
>> > I changed this line in SparkBuild.scala
>> >         "org.apache.mesos"         % "mesos"            % "0.13.0",
>> > to
>> >         "org.apache.mesos"         % "mesos"            % "0.18.0",
>> >
>> > ...ran make-distribution.sh, repackaged and redeployed the tar.gz to
>> HDFS.
>> >
>> > It still core dumps like this:
>> > https://gist.github.com/stevencox/11002498
>> >
>> > In this environment:
>> >   Ubuntu 13.10
>> >   Mesos 0.18.0
>> >   Spark 0.9.1
>> >   JDK 1.7.0_45
>> >   Scala 2.10.1
>> >
>> > What am I missing?
>>
>
>

Reply via email to