I take it there will be another RC due to some blockers and as there were
no +1 votes anyway.

FWIW, I cannot run python tests using "./python/run-tests".

I'd be -1 for this reason (see https://github.com/apache/spark/pull/13737 /
http://issues.apache.org/jira/browse/SPARK-15954) - does anyone else
encounter this?

./python/run-tests --python-executables=python2.7
Running PySpark tests. Output is in
/Users/nick/workspace/scala/spark-rcs/spark-2.0.0/python/unit-tests.log
Will test against the following Python executables: ['python2.7']
Will test the following Python modules: ['pyspark-core', 'pyspark-ml',
'pyspark-mllib', 'pyspark-sql', 'pyspark-streaming']
....Using Spark's default log4j profile:
org/apache/spark/log4j-defaults.properties
Setting default log level to "WARN".
To adjust logging level use sc.setLogLevel(newLevel).
======================================================================
ERROR: setUpClass (pyspark.sql.tests.HiveContextSQLTests)
----------------------------------------------------------------------
Traceback (most recent call last):
  File
"/Users/nick/workspace/scala/spark-rcs/spark-2.0.0/python/pyspark/sql/tests.py",
line 1620, in setUpClass
    cls.spark = HiveContext._createForTesting(cls.sc)
  File
"/Users/nick/workspace/scala/spark-rcs/spark-2.0.0/python/pyspark/sql/context.py",
line 490, in _createForTesting
    jtestHive =
sparkContext._jvm.org.apache.spark.sql.hive.test.TestHiveContext(jsc)
  File
"/Users/nick/workspace/scala/spark-rcs/spark-2.0.0/python/lib/py4j-0.10.1-src.zip/py4j/java_gateway.py",
line 1183, in __call__
    answer, self._gateway_client, None, self._fqn)
  File
"/Users/nick/workspace/scala/spark-rcs/spark-2.0.0/python/lib/py4j-0.10.1-src.zip/py4j/protocol.py",
line 312, in get_return_value
    format(target_id, ".", name), value)
Py4JJavaError: An error occurred while calling
None.org.apache.spark.sql.hive.test.TestHiveContext.
: java.lang.NullPointerException
at
org.apache.spark.sql.hive.test.TestHiveSparkSession.getHiveFile(TestHive.scala:183)
at
org.apache.spark.sql.hive.test.TestHiveSparkSession.<init>(TestHive.scala:214)
at
org.apache.spark.sql.hive.test.TestHiveSparkSession.<init>(TestHive.scala:122)
at org.apache.spark.sql.hive.test.TestHiveContext.<init>(TestHive.scala:77)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:240)
at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
at py4j.Gateway.invoke(Gateway.java:236)
at
py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)
at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)
at py4j.GatewayConnection.run(GatewayConnection.java:211)
at java.lang.Thread.run(Thread.java:745)


======================================================================
ERROR: setUpClass (pyspark.sql.tests.SQLTests)
----------------------------------------------------------------------
Traceback (most recent call last):
  File
"/Users/nick/workspace/scala/spark-rcs/spark-2.0.0/python/pyspark/sql/tests.py",
line 189, in setUpClass
    ReusedPySparkTestCase.setUpClass()
  File
"/Users/nick/workspace/scala/spark-rcs/spark-2.0.0/python/pyspark/tests.py",
line 344, in setUpClass
    cls.sc = SparkContext('local[4]', cls.__name__)
  File
"/Users/nick/workspace/scala/spark-rcs/spark-2.0.0/python/pyspark/context.py",
line 112, in __init__
    SparkContext._ensure_initialized(self, gateway=gateway)
  File
"/Users/nick/workspace/scala/spark-rcs/spark-2.0.0/python/pyspark/context.py",
line 261, in _ensure_initialized
    callsite.function, callsite.file, callsite.linenum))
ValueError: Cannot run multiple SparkContexts at once; existing
SparkContext(app=ReusedPySparkTestCase, master=local[4]) created by
<module> at /Users/nick/miniconda2/lib/python2.7/runpy.py:72

----------------------------------------------------------------------
Ran 4 tests in 4.800s

FAILED (errors=2)

Had test failures in pyspark.sql.tests with python2.7; see logs.


On Mon, 27 Jun 2016 at 20:13 Egor Pahomov <pahomov.e...@gmail.com> wrote:

> -1 : SPARK-16228 [SQL]  - "Percentile" needs explicit cast to double,
> otherwise it throws an error. I can not move my existing 100500 quires to
> 2.0 transparently.
>
> 2016-06-24 11:52 GMT-07:00 Matt Cheah <mch...@palantir.com>:
>
>> -1 because of SPARK-16181 which is a correctness regression from 1.6.
>> Looks like the patch is ready though:
>> https://github.com/apache/spark/pull/13884 – it would be ideal for this
>> patch to make it into the release.
>>
>> -Matt Cheah
>>
>> From: Nick Pentreath <nick.pentre...@gmail.com>
>> Date: Friday, June 24, 2016 at 4:37 AM
>> To: "dev@spark.apache.org" <dev@spark.apache.org>
>> Subject: Re: [VOTE] Release Apache Spark 2.0.0 (RC1)
>>
>> I'm getting the following when trying to run ./dev/run-tests (not
>> happening on master) from the extracted source tar. Anyone else seeing
>> this?
>>
>> error: Could not access 'fc0a1475ef'
>> **********************************************************************
>> File "./dev/run-tests.py", line 69, in
>> __main__.identify_changed_files_from_git_commits
>> Failed example:
>>     [x.name
>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__x.name&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wx5Qjw-efxMVvKXnjUsSkkQcEF6zQHQLQaGtAK9pxIw&e=>
>> for x in determine_modules_for_files(
>> identify_changed_files_from_git_commits("fc0a1475ef",
>> target_ref="5da21f07"))]
>> Exception raised:
>>     Traceback (most recent call last):
>>       File "/Users/nick/miniconda2/lib/python2.7/doctest.py", line 1315,
>> in __run
>>         compileflags, 1) in test.globs
>>       File "<doctest
>> __main__.identify_changed_files_from_git_commits[0]>", line 1, in <module>
>>         [x.name
>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__x.name&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wx5Qjw-efxMVvKXnjUsSkkQcEF6zQHQLQaGtAK9pxIw&e=>
>> for x in determine_modules_for_files(
>> identify_changed_files_from_git_commits("fc0a1475ef",
>> target_ref="5da21f07"))]
>>       File "./dev/run-tests.py", line 86, in
>> identify_changed_files_from_git_commits
>>         universal_newlines=True)
>>       File "/Users/nick/miniconda2/lib/python2.7/subprocess.py", line
>> 573, in check_output
>>         raise CalledProcessError(retcode, cmd, output=output)
>>     CalledProcessError: Command '['git', 'diff', '--name-only',
>> 'fc0a1475ef', '5da21f07']' returned non-zero exit status 1
>> error: Could not access '50a0496a43'
>> **********************************************************************
>> File "./dev/run-tests.py", line 71, in
>> __main__.identify_changed_files_from_git_commits
>> Failed example:
>>     'root' in [x.name
>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__x.name&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wx5Qjw-efxMVvKXnjUsSkkQcEF6zQHQLQaGtAK9pxIw&e=>
>> for x in determine_modules_for_files(
>>  identify_changed_files_from_git_commits("50a0496a43",
>> target_ref="6765ef9"))]
>> Exception raised:
>>     Traceback (most recent call last):
>>       File "/Users/nick/miniconda2/lib/python2.7/doctest.py", line 1315,
>> in __run
>>         compileflags, 1) in test.globs
>>       File "<doctest
>> __main__.identify_changed_files_from_git_commits[1]>", line 1, in <module>
>>         'root' in [x.name
>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__x.name&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wx5Qjw-efxMVvKXnjUsSkkQcEF6zQHQLQaGtAK9pxIw&e=>
>> for x in determine_modules_for_files(
>>  identify_changed_files_from_git_commits("50a0496a43",
>> target_ref="6765ef9"))]
>>       File "./dev/run-tests.py", line 86, in
>> identify_changed_files_from_git_commits
>>         universal_newlines=True)
>>       File "/Users/nick/miniconda2/lib/python2.7/subprocess.py", line
>> 573, in check_output
>>         raise CalledProcessError(retcode, cmd, output=output)
>>     CalledProcessError: Command '['git', 'diff', '--name-only',
>> '50a0496a43', '6765ef9']' returned non-zero exit status 1
>> **********************************************************************
>> 1 items had failures:
>>    2 of   2 in __main__.identify_changed_files_from_git_commits
>> ***Test Failed*** 2 failures.
>>
>>
>>
>> On Fri, 24 Jun 2016 at 06:59 Yin Huai <yh...@databricks.com> wrote:
>>
>>> -1 because of https://issues.apache.org/jira/browse/SPARK-16121
>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__issues.apache.org_jira_browse_SPARK-2D16121&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=9200NP4SpeJSUNrSrlWWEC7vFvjWSyCHnx5LD7Sj9u4&e=>.
>>>
>>>
>>> This jira was resolved after 2.0.0-RC1 was cut. Without the fix, Spark
>>> SQL effectively only uses the driver to list files when loading datasets
>>> and the driver-side file listing is very slow for datasets having many
>>> files and partitions. Since this bug causes a serious performance
>>> regression, I am giving -1.
>>>
>>> On Thu, Jun 23, 2016 at 1:25 AM, Pete Robbins <robbin...@gmail.com>
>>> wrote:
>>>
>>>> I'm also seeing some of these same failures:
>>>>
>>>> - spilling with compression *** FAILED ***
>>>> I have seen this occassionaly
>>>>
>>>> - to UTC timestamp *** FAILED ***
>>>> This was fixed yesterday in branch-2.0 (
>>>> https://issues.apache.org/jira/browse/SPARK-16078
>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__issues.apache.org_jira_browse_SPARK-2D16078&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=SuVdXUNGdAhYgtA2fMLe5vZ2PFrPOaeO3i3cbhYU4tc&e=>)
>>>>
>>>>
>>>> - offset recovery *** FAILED ***
>>>> Haven't seen this for a while and thought the flaky test was fixed but
>>>> it popped up again in one of our builds.
>>>>
>>>> StateStoreSuite:
>>>> - maintenance *** FAILED ***
>>>> Just seen this has been failing for last 2 days on one build machine
>>>> (linux amd64)
>>>>
>>>>
>>>> On 23 June 2016 at 08:51, Sean Owen <so...@cloudera.com> wrote:
>>>>
>>>>> First pass of feedback on the RC: all the sigs, hashes, etc are fine.
>>>>> Licensing is up to date to the best of my knowledge.
>>>>>
>>>>> I'm hitting test failures, some of which may be spurious. Just putting
>>>>> them out there to see if they ring bells. This is Java 8 on Ubuntu 16.
>>>>>
>>>>>
>>>>> - spilling with compression *** FAILED ***
>>>>>   java.lang.Exception: Test failed with compression using codec
>>>>> org.apache.spark.io.SnappyCompressionCodec:
>>>>> assertion failed: expected cogroup to spill, but did not
>>>>>   at scala.Predef$.assert(Predef.scala:170)
>>>>>   at org.apache.spark.TestUtils$.assertSpilled(TestUtils.scala:170)
>>>>>   at org.apache.spark.util.collection.ExternalAppendOnlyMapSuite.org
>>>>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__org.apache.spark.util.collection.ExternalAppendOnlyMapSuite.org&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=goarAptcJYfLg44f7BAwhbipqJlRFKz9Y6Z36HItiKg&e=>
>>>>> $apache$spark$util$collection$ExternalAppendOnlyMapSuite$$testSimpleSpilling(ExternalAppendOnlyMapSuite.scala:263)
>>>>> ...
>>>>>
>>>>> I feel like I've seen this before, and see some possibly relevant
>>>>> fixes, but they're in 2.0.0 already:
>>>>> https://github.com/apache/spark/pull/10990
>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_apache_spark_pull_10990&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=dFymYD9NRVHIJ5MKpmzPcH_NYwLjOWcZd7FUuQBpTUU&e=>
>>>>> Is this something where a native library needs to be installed or
>>>>> something?
>>>>>
>>>>>
>>>>> - to UTC timestamp *** FAILED ***
>>>>>   "2016-03-13 [02]:00:00.0" did not equal "2016-03-13 [10]:00:00.0"
>>>>> (DateTimeUtilsSuite.scala:506)
>>>>>
>>>>> I know, we talked about this for the 1.6.2 RC, but I reproduced this
>>>>> locally too. I will investigate, could still be spurious.
>>>>>
>>>>>
>>>>> StateStoreSuite:
>>>>> - maintenance *** FAILED ***
>>>>>   The code passed to eventually never returned normally. Attempted 627
>>>>> times over 10.000180116 seconds. Last failure message:
>>>>> StateStoreSuite.this.fileExists(provider, 1L, false) was true earliest
>>>>> file not deleted. (StateStoreSuite.scala:395)
>>>>>
>>>>> No idea.
>>>>>
>>>>>
>>>>> - offset recovery *** FAILED ***
>>>>>   The code passed to eventually never returned normally. Attempted 197
>>>>> times over 10.040864806 seconds. Last failure message:
>>>>> strings.forall({
>>>>>     ((x$1: Any) => DirectKafkaStreamSuite.collectedData.contains(x$1))
>>>>>   }) was false. (DirectKafkaStreamSuite.scala:250)
>>>>>
>>>>> Also something that was possibly fixed already for 2.0.0 and that I
>>>>> just back-ported into 1.6. Could be just a very similar failure.
>>>>>
>>>>> On Wed, Jun 22, 2016 at 2:26 AM, Reynold Xin <r...@databricks.com>
>>>>> wrote:
>>>>> > Please vote on releasing the following candidate as Apache Spark
>>>>> version
>>>>> > 2.0.0. The vote is open until Friday, June 24, 2016 at 19:00 PDT and
>>>>> passes
>>>>> > if a majority of at least 3+1 PMC votes are cast.
>>>>> >
>>>>> > [ ] +1 Release this package as Apache Spark 2.0.0
>>>>> > [ ] -1 Do not release this package because ...
>>>>> >
>>>>> >
>>>>> > The tag to be voted on is v2.0.0-rc1
>>>>> > (0c66ca41afade6db73c9aeddd5aed6e5dcea90df).
>>>>> >
>>>>> > This release candidate resolves ~2400 issues:
>>>>> > https://s.apache.org/spark-2.0.0-rc1-jira
>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__s.apache.org_spark-2D2.0.0-2Drc1-2Djira&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=ZD_PezvsJ1GyDhv7MhaeUrVba_uhED5mPkqKpfenKEE&e=>
>>>>> >
>>>>> > The release files, including signatures, digests, etc. can be found
>>>>> at:
>>>>> >
>>>>> http://people.apache.org/~pwendell/spark-releases/spark-2.0.0-rc1-bin/
>>>>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__people.apache.org_-7Epwendell_spark-2Dreleases_spark-2D2.0.0-2Drc1-2Dbin_&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=wSbzZ2LyuDcNKaCijEPdt9rokQ0R9w66tn2jMfjKN2I&e=>
>>>>> >
>>>>> > Release artifacts are signed with the following key:
>>>>> > https://people.apache.org/keys/committer/pwendell.asc
>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__people.apache.org_keys_committer_pwendell.asc&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=i1Uxw1NyUf2iuA3CXbyiEODD1RR24rAXUvkc42ut8Ao&e=>
>>>>> >
>>>>> > The staging repository for this release can be found at:
>>>>> >
>>>>> https://repository.apache.org/content/repositories/orgapachespark-1187/
>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__repository.apache.org_content_repositories_orgapachespark-2D1187_&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=QjsvnxXe6JBQqXwKw6r-fIIHI9E0ugeeICAqjRXRNwc&e=>
>>>>> >
>>>>> > The documentation corresponding to this release can be found at:
>>>>> >
>>>>> http://people.apache.org/~pwendell/spark-releases/spark-2.0.0-rc1-docs/
>>>>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__people.apache.org_-7Epwendell_spark-2Dreleases_spark-2D2.0.0-2Drc1-2Ddocs_&d=DQMFaQ&c=izlc9mHr637UR4lpLEZLFFS3Vn2UXBrZ4tFb6oOnmz8&r=hzwIMNQ9E99EMYGuqHI0kXhVbvX3nU3OSDadUnJxjAs&m=Y3d-oJvw2gK_2KXYjXY8_yzfAosPOqqaV4wtMg6ZPwM&s=_6IZExLgc8WoxW0kft_weR7AvELgbFXnHZdezQ_IYGk&e=>
>>>>> >
>>>>> >
>>>>> > =======================================
>>>>> > == How can I help test this release? ==
>>>>> > =======================================
>>>>> > If you are a Spark user, you can help us test this release by taking
>>>>> an
>>>>> > existing Spark workload and running on this release candidate, then
>>>>> > reporting any regressions from 1.x.
>>>>> >
>>>>> > ================================================
>>>>> > == What justifies a -1 vote for this release? ==
>>>>> > ================================================
>>>>> > Critical bugs impacting major functionalities.
>>>>> >
>>>>> > Bugs already present in 1.x, missing features, or bugs related to new
>>>>> > features will not necessarily block this release. Note that
>>>>> historically
>>>>> > Spark documentation has been published on the website separately
>>>>> from the
>>>>> > main release so we do not need to block the release due to
>>>>> documentation
>>>>> > errors either.
>>>>> >
>>>>> >
>>>>>
>>>>> ---------------------------------------------------------------------
>>>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
>>>>> For additional commands, e-mail: dev-h...@spark.apache.org
>>>>>
>>>>>
>>>>
>>>
>
>
> --
>
>
> *Sincerely yoursEgor Pakhomov*
>

Reply via email to