Although that should be fixed if it's incorrect, it's not something that would nearly block a release. The question here is whether this artifact can be released as 1.4.1, or whether it has a blocking regression from 1.4.0.
On Wed, Jul 8, 2015 at 5:35 PM, Pradeep Bashyal <prad...@bashyal.com> wrote: > Here's one thing I ran into: > > The SparkR documentation example in > http://people.apache.org/~pwendell/spark-releases/latest/sparkr.html is > incorrect. > > sc <- sparkR.init(packages="com.databricks:spark-csv_2.11:1.0.3") > > should be > > sc <- sparkR.init(sparkPackages="com.databricks:spark-csv_2.11:1.0.3") > > > Thanks > Pradeep > > > On Wed, Jul 8, 2015 at 6:18 AM, Sean Owen <so...@cloudera.com> wrote: >> >> The POM issue is resolved and the build succeeds. The license and sigs >> still work. The tests pass for me with "-Pyarn -Phadoop-2.6", with the >> following two exceptions. Is anyone else seeing these? this is >> consistent on Ubuntu 14 with Java 7/8: >> >> DataFrameStatSuite: >> ... >> - special crosstab elements (., '', null, ``) *** FAILED *** >> java.lang.NullPointerException: >> at >> org.apache.spark.sql.execution.stat.StatFunctions$$anonfun$4.apply(StatFunctions.scala:131) >> at >> org.apache.spark.sql.execution.stat.StatFunctions$$anonfun$4.apply(StatFunctions.scala:121) >> at >> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) >> at >> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:244) >> at scala.collection.immutable.Map$Map4.foreach(Map.scala:181) >> at scala.collection.TraversableLike$class.map(TraversableLike.scala:244) >> at scala.collection.AbstractTraversable.map(Traversable.scala:105) >> at >> org.apache.spark.sql.execution.stat.StatFunctions$.crossTabulate(StatFunctions.scala:121) >> at >> org.apache.spark.sql.DataFrameStatFunctions.crosstab(DataFrameStatFunctions.scala:94) >> at >> org.apache.spark.sql.DataFrameStatSuite$$anonfun$5.apply$mcV$sp(DataFrameStatSuite.scala:97) >> ... >> >> HiveSparkSubmitSuite: >> - SPARK-8368: includes jars passed in through --jars *** FAILED *** >> Process returned with exit code 1. See the log4j logs for more >> detail. (HiveSparkSubmitSuite.scala:92) >> - SPARK-8020: set sql conf in spark conf *** FAILED *** >> Process returned with exit code 1. See the log4j logs for more >> detail. (HiveSparkSubmitSuite.scala:92) >> - SPARK-8489: MissingRequirementError during reflection *** FAILED *** >> Process returned with exit code 1. See the log4j logs for more >> detail. (HiveSparkSubmitSuite.scala:92) >> >> On Tue, Jul 7, 2015 at 8:06 PM, Patrick Wendell <pwend...@gmail.com> >> wrote: >> > Please vote on releasing the following candidate as Apache Spark version >> > 1.4.1! >> > >> > This release fixes a handful of known issues in Spark 1.4.0, listed >> > here: >> > http://s.apache.org/spark-1.4.1 >> > >> > The tag to be voted on is v1.4.1-rc3 (commit 3e8ae38): >> > https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h= >> > 3e8ae38944f13895daf328555c1ad22cd590b089 >> > >> > The release files, including signatures, digests, etc. can be found at: >> > http://people.apache.org/~pwendell/spark-releases/spark-1.4.1-rc3-bin/ >> > >> > Release artifacts are signed with the following key: >> > https://people.apache.org/keys/committer/pwendell.asc >> > >> > The staging repository for this release can be found at: >> > [published as version: 1.4.1] >> > https://repository.apache.org/content/repositories/orgapachespark-1123/ >> > [published as version: 1.4.1-rc3] >> > https://repository.apache.org/content/repositories/orgapachespark-1124/ >> > >> > The documentation corresponding to this release can be found at: >> > http://people.apache.org/~pwendell/spark-releases/spark-1.4.1-rc3-docs/ >> > >> > Please vote on releasing this package as Apache Spark 1.4.1! >> > >> > The vote is open until Friday, July 10, at 20:00 UTC and passes >> > if a majority of at least 3 +1 PMC votes are cast. >> > >> > [ ] +1 Release this package as Apache Spark 1.4.1 >> > [ ] -1 Do not release this package because ... >> > >> > To learn more about Apache Spark, please see >> > http://spark.apache.org/ >> > >> > --------------------------------------------------------------------- >> > To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >> > For additional commands, e-mail: dev-h...@spark.apache.org >> > >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org >> For additional commands, e-mail: dev-h...@spark.apache.org >> > --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org For additional commands, e-mail: dev-h...@spark.apache.org