This makes no sense. There is no difference between v1.2.0-rc2 and v1.2.0: https://github.com/apache/spark/compare/v1.2.0-rc2...v1.2.0
On Sat, Dec 20, 2014 at 12:44 PM, Matt Mead <m...@matthewcmead.com> wrote: > First, thanks for the efforts and contribution to such a useful software > stack! Spark is great! > > I have been using the git tags for v1.2.0-rc1 and v1.2.0-rc2 built as > follows: > > ./make-distribution.sh -Dhadoop.version=2.5.0-cdh5.2.0 >> -Dyarn.version=2.5.0-cdh5.2.0 -Phadoop-2.4 -Phive -Pyarn -Phive-thriftserver > > > I have been starting the thriftserver as follows: > > HADOOP_CONF_DIR=/etc/hadoop/conf ./sbin/start-thriftserver.sh --master >> yarn --num-executors 16 > > > Under v1.2.0-rc1 and v1.2.0-rc2, this has worked properly, where the > thriftserver starts up and I am able to interact with it and execute > queries as expected using the JDBC driver. > > I have updated to git tag v1.2.0, built identically and started the > thriftserver identically, but am now running into the following issue on > startup: > > Exception in thread "main" java.lang.IllegalArgumentException: Wrong FS: >> hdfs://myhdfs/user/user/.sparkStaging/application_1416150945509_0055/datanucleus-api-jdo-3.2.6.jar, >> expected: file:/// >> at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:645) >> at >> org.apache.hadoop.fs.RawLocalFileSystem.pathToFile(RawLocalFileSystem.java:80) >> at >> org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:519) >> at >> org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:737) >> at >> org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:514) >> at >> org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:398) >> at >> org.apache.spark.deploy.yarn.ClientDistributedCacheManager.addResource(ClientDistributedCacheManager.scala:67) >> at >> org.apache.spark.deploy.yarn.ClientBase$$anonfun$prepareLocalResources$5.apply(ClientBase.scala:257) >> at >> org.apache.spark.deploy.yarn.ClientBase$$anonfun$prepareLocalResources$5.apply(ClientBase.scala:242) >> at scala.Option.foreach(Option.scala:236) >> at >> org.apache.spark.deploy.yarn.ClientBase$class.prepareLocalResources(ClientBase.scala:242) >> at >> org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:35) >> at >> org.apache.spark.deploy.yarn.ClientBase$class.createContainerLaunchContext(ClientBase.scala:350) >> at >> org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:35) >> at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:80) >> at >> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) >> at >> org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:140) >> at org.apache.spark.SparkContext.<init>(SparkContext.scala:335) >> at >> org.apache.spark.sql.hive.thriftserver.SparkSQLEnv$.init(SparkSQLEnv.scala:38) >> at >> org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$.main(HiveThriftServer2.scala:56) >> at >> org.apache.spark.sql.hive.thriftserver.HiveThriftServer2.main(HiveThriftServer2.scala) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:358) >> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:75) >> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > > > Looking at SPARK-4757, it appears others were seeing this behavior in > earlier releases and it is fixed in v1.2.0, whereas I did not see the > behavior in earlier releases and now am seeing it in v1.2.0. > > I have tested this with the exact same build/launch commands on two > separate CDH5.2.0 clusters with identical results. Both machines where the > build and execution take place have a proper HDFS/YARN client configuration > in /etc/hadoop/conf and other hadoop tools like MR2 on YARN function as > expected. > > Any ideas on what to do to resolve this issue? > > Thanks! > > > > > -matt > >