Otherwise we could just set the legacy config as 'true' by default, until
SPARK-31257 (unification of create table syntax) is introduced. That would
be smallest and safest approach - no ambiguity by default, and if anyone is
adventurous or fully understanding the ambiguity they can set it to "false"
and deal with it. WDYT?

On Tue, Mar 31, 2020 at 7:48 AM Jungtaek Lim <kabhwan.opensou...@gmail.com>
wrote:

> I'm not sure I understand the direction of resolution. I'm not saying it's
> just a confusion - it's "ambiguous" and "indeterministic".
>
> Two syntaxes were at least distinguished before Spark 3.0 by "USING ...",
> and it's simply broken. I understand we are lack of time and we may need
> some workaround but I'm not sure I agree with leaving it worse than before.
>
> The source code made it clear that the Hive create table syntax is the
> legacy syntax for Hive compatibility, and we recommend users to use native
> create table syntax. What makes us fear to put legacy syntax to be a second
> class of support, even we put legacy config to let end users revert it back?
>
> Single line of code change (excluding tests) would fix everything until we
> reach to unify the two syntax - add a marker to Hive create table syntax.
> That would make the Hive create table syntax to become a second class. Yes
> it requires end users to add a marker if they need to run create table for
> Hive syntax, but wouldn't we want to recommend end users to use native
> create table syntax? Is it a thing we should support natively as well, with
> such ambiguity?
>
>
> On Mon, Mar 30, 2020 at 12:49 PM Wenchen Fan <cloud0...@gmail.com> wrote:
>
>> I agree that we can cut the RC anyway even if there are blockers, to move
>> us to a more official "code freeze" status.
>>
>> About the CREATE TABLE unification, it's still WIP and not close-to-merge
>> yet. Can we fix some specific problems like CREATE EXTERNAL TABLE
>> surgically and leave the unification to 3.1 or 3.0.1? Spark has 2 CREATE
>> TABLE syntaxes for a long time and the confusion is already there.
>>
>> On Sun, Mar 29, 2020 at 9:29 AM Nicholas Chammas <
>> nicholas.cham...@gmail.com> wrote:
>>
>>> I don't have a dog in this race, but: Would it be OK to ship 3.0 with
>>> some release notes and/or prominent documentation calling out this issue,
>>> and then fixing it in 3.0.1?
>>>
>>> On Sat, Mar 28, 2020 at 8:45 PM Jungtaek Lim <
>>> kabhwan.opensou...@gmail.com> wrote:
>>>
>>>> I'd say SPARK-31257 as open blocker, because the change in upcoming
>>>> Spark 3.0 made the create table be ambiguous, and once it's shipped it will
>>>> be harder to correct again.
>>>>
>>>> On Sun, Mar 29, 2020 at 4:53 AM Reynold Xin <r...@databricks.com>
>>>> wrote:
>>>>
>>>>> Let's start cutting RC next week.
>>>>>
>>>>>
>>>>> On Sat, Mar 28, 2020 at 11:51 AM, Sean Owen <sro...@gmail.com> wrote:
>>>>>
>>>>>> I'm also curious - there no open blockers for 3.0 but I know a few
>>>>>> are still floating around open to revert changes. What is the status 
>>>>>> there?
>>>>>> From my field of view I'm not aware of other blocking issues.
>>>>>>
>>>>>> On Fri, Mar 27, 2020 at 10:56 PM Jungtaek Lim <
>>>>>> kabhwan.opensou...@gmail.com> wrote:
>>>>>>
>>>>>>> Now the end of March is just around the corner. I'm not qualified to
>>>>>>> say (and honestly don't know) where we are, but if we were intended to 
>>>>>>> be
>>>>>>> in blocker mode it doesn't seem to work; lots of developments still 
>>>>>>> happen,
>>>>>>> and priority/urgency doesn't seem to be applied to the sequence of
>>>>>>> reviewing.
>>>>>>>
>>>>>>> How about listing (or linking to epic, or labelling) JIRA issues/PRs
>>>>>>> which are blockers (either from priority or technically) for Spark 3.0
>>>>>>> release, and make clear we should try to review these blockers
>>>>>>> first? Github PR label may help here to filter out other PRs and
>>>>>>> concentrate these things.
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Jungtaek Lim (HeartSaVioR)
>>>>>>>
>>>>>>>
>>>>>>> On Wed, Mar 25, 2020 at 1:52 PM Xiao Li <lix...@databricks.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Let us try to finish the remaining major blockers in the next few
>>>>>>>> days. For example,
>>>>>>>> https://issues.apache.org/jira/browse/SPARK-31085
>>>>>>>>
>>>>>>>> +1 to cut the RC even if we still have the blockers that will fail
>>>>>>>> the RCs.
>>>>>>>>
>>>>>>>> Cheers,
>>>>>>>>
>>>>>>>> Xiao
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, Mar 24, 2020 at 6:56 PM Dongjoon Hyun <
>>>>>>>> dongjoon.h...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> +1
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Dongjoon.
>>>>>>>>>
>>>>>>>>> On Tue, Mar 24, 2020 at 14:49 Reynold Xin <r...@databricks.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> I actually think we should start cutting RCs. We can cut RCs even
>>>>>>>>>> with blockers.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Tue, Mar 24, 2020 at 12:51 PM, Dongjoon Hyun <
>>>>>>>>>> dongjoon.h...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi, All.
>>>>>>>>>>>
>>>>>>>>>>> First of all, always "Community Over Code"!
>>>>>>>>>>> I wish you the best health and happiness.
>>>>>>>>>>>
>>>>>>>>>>> As we know, we are still working on QA period, we didn't reach
>>>>>>>>>>> RC stage. It seems that we need to make website up-to-date once 
>>>>>>>>>>> more.
>>>>>>>>>>>
>>>>>>>>>>>     https://spark.apache.org/versioning-policy.html
>>>>>>>>>>>
>>>>>>>>>>> If possible, it would be really great if we can get `3.0.0`
>>>>>>>>>>> release manager's official `branch-3.0` assessment because we have 
>>>>>>>>>>> only 1
>>>>>>>>>>> week before the end of March.
>>>>>>>>>>>
>>>>>>>>>>> Cloud you, the 3.0.0 release manager, share your thought and
>>>>>>>>>>> update the website, please?
>>>>>>>>>>>
>>>>>>>>>>> Bests
>>>>>>>>>>> Dongjoon.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> <https://databricks.com/sparkaisummit/north-america>
>>>>>>>>
>>>>>>>
>>>>>

Reply via email to