Oops - I actually should have added v0.13.0 (i.e. to match whatever I
did in the profile).

On Tue, Oct 28, 2014 at 10:05 PM, Patrick Wendell <pwend...@gmail.com> wrote:
> Cheng - to make it recognize the new HiveShim for 0.12 I had to click
> on spark-hive under "packages" in the left pane, then go to "Open
> Module Settings" - then explicitly add the v0.12.0/src/main/scala
> folder to the sources by navigating to it and then <ctrl>+click to add
> it as a source. Did you have to do this?
>
> On Tue, Oct 28, 2014 at 9:57 PM, Patrick Wendell <pwend...@gmail.com> wrote:
>> I just started a totally fresh IntelliJ project importing from our
>> root pom. I used all the default options and I added "hadoop-2.4,
>> hive, hive-0.13.1" profiles. I was able to run spark core tests from
>> within IntelliJ. Didn't try anything beyond that, but FWIW this
>> worked.
>>
>> - Patrick
>>
>> On Tue, Oct 28, 2014 at 9:54 PM, Cheng Lian <lian.cs....@gmail.com> wrote:
>>> You may first open the root pom.xml file in IDEA, and then go for menu View
>>> / Tool Windows / Maven Projects, then choose desired Maven profile
>>> combination under the "Profiles" node (e.g. I usually use hadoop-2.4 + hive
>>> + hive-0.12.0). IDEA will ask you to re-import the Maven projects, confirm,
>>> then it should be OK.
>>>
>>> I can debug within IDEA with this approach. However, you have to clean the
>>> whole project before debugging Spark within IDEA if you compiled the project
>>> outside IDEA. Haven't got time to investigate this annoying issue.
>>>
>>> Also, you can remove sub projects unrelated to your tasks to accelerate
>>> compilation and/or avoid other IDEA build issues (e.g. Avro related Spark
>>> streaming build failure in IDEA).
>>>
>>>
>>> On 10/29/14 12:42 PM, Stephen Boesch wrote:
>>>
>>> I am interested specifically in how to build (and hopefully run/debug..)
>>> under Intellij.  Your posts sound like command line maven - which has always
>>> been working already.
>>>
>>> Do you have instructions for building in IJ?
>>>
>>> 2014-10-28 21:38 GMT-07:00 Cheng Lian <lian.cs....@gmail.com>:
>>>>
>>>> Yes, these two combinations work for me.
>>>>
>>>>
>>>> On 10/29/14 12:32 PM, Zhan Zhang wrote:
>>>>>
>>>>> -Phive is to enable hive-0.13.1 and "-Phive -Phive-0.12.0" is to enable
>>>>> hive-0.12.0. Note that the thrift-server is not supported yet in 
>>>>> hive-0.13,
>>>>> but expected to go to upstream soon (Spark-3720).
>>>>>
>>>>> Thanks.
>>>>>
>>>>> Zhan Zhang
>>>>>
>>>>>
>>>>>   On Oct 28, 2014, at 9:09 PM, Stephen Boesch <java...@gmail.com> wrote:
>>>>>
>>>>>> Thanks Patrick for the heads up.
>>>>>>
>>>>>> I have not been successful to discover a combination of profiles (i.e.
>>>>>> enabling hive or hive-0.12.0 or hive-13.0) that works in Intellij with
>>>>>> maven. Anyone who knows how to handle this - a quick note here would be
>>>>>> appreciated.
>>>>>>
>>>>>>
>>>>>>
>>>>>> 2014-10-28 20:20 GMT-07:00 Patrick Wendell <pwend...@gmail.com>:
>>>>>>
>>>>>>> Hey Stephen,
>>>>>>>
>>>>>>> In some cases in the maven build we now have pluggable source
>>>>>>> directories based on profiles using the maven build helper plug-in.
>>>>>>> This is necessary to support cross building against different Hive
>>>>>>> versions, and there will be additional instances of this due to
>>>>>>> supporting scala 2.11 and 2.10.
>>>>>>>
>>>>>>> In these cases, you may need to add source locations explicitly to
>>>>>>> intellij if you want the entire project to compile there.
>>>>>>>
>>>>>>> Unfortunately as long as we support cross-building like this, it will
>>>>>>> be an issue. Intellij's maven support does not correctly detect our
>>>>>>> use of the maven-build-plugin to add source directories.
>>>>>>>
>>>>>>> We should come up with a good set of instructions on how to import the
>>>>>>> pom files + add the few extra source directories. Off hand I am not
>>>>>>> sure exactly what the correct sequence is.
>>>>>>>
>>>>>>> - Patrick
>>>>>>>
>>>>>>> On Tue, Oct 28, 2014 at 7:57 PM, Stephen Boesch <java...@gmail.com>
>>>>>>> wrote:
>>>>>>>>
>>>>>>>> Hi Matei,
>>>>>>>>   Until my latest pull from upstream/master it had not been necessary
>>>>>>>> to
>>>>>>>> add the hive profile: is it now??
>>>>>>>>
>>>>>>>> I am not using sbt gen-idea. The way to open in intellij has been to
>>>>>>>> Open
>>>>>>>> the parent directory. IJ recognizes it as a maven project.
>>>>>>>>
>>>>>>>> There are several steps to do surgery on the yarn-parent / yarn
>>>>>>>> projects
>>>>>>>
>>>>>>> ,
>>>>>>>>
>>>>>>>> then do a full rebuild.  That was working until one week ago.
>>>>>>>> Intellij/maven is presently broken in  two ways:  this hive shim
>>>>>>>> (which
>>>>>>>
>>>>>>> may
>>>>>>>>
>>>>>>>> yet hopefully be a small/simple fix - let us see) and  (2) the
>>>>>>>> "NoClassDefFoundError
>>>>>>>> on ThreadFactoryBuilder" from my prior emails -and which is quite a
>>>>>>>
>>>>>>> serious
>>>>>>>>
>>>>>>>> problem .
>>>>>>>>
>>>>>>>> 2014-10-28 19:46 GMT-07:00 Matei Zaharia <matei.zaha...@gmail.com>:
>>>>>>>>
>>>>>>>>> Hi Stephen,
>>>>>>>>>
>>>>>>>>> How did you generate your Maven workspace? You need to make sure the
>>>>>>>
>>>>>>> Hive
>>>>>>>>>
>>>>>>>>> profile is enabled for it. For example sbt/sbt -Phive gen-idea.
>>>>>>>>>
>>>>>>>>> Matei
>>>>>>>>>
>>>>>>>>>> On Oct 28, 2014, at 7:42 PM, Stephen Boesch <java...@gmail.com>
>>>>>>>
>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>> I have run on the command line via maven and it is fine:
>>>>>>>>>>
>>>>>>>>>> mvn   -Dscalastyle.failOnViolation=false -DskipTests -Pyarn
>>>>>>>
>>>>>>> -Phadoop-2.3
>>>>>>>>>>
>>>>>>>>>> compile package install
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> But with the latest code Intellij builds do not work. Following is
>>>>>>>
>>>>>>> one of
>>>>>>>>>>
>>>>>>>>>> 26 similar errors:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Error:(173, 38) not found: value HiveShim
>>>>>>>>>>
>>>>>>>>> Option(tableParameters.get(HiveShim.getStatsSetupConstTotalSize))
>>>>>>>>>>
>>>>>>>>>>                                     ^
>>>>>>>>>
>>>>>>>>>
>>>>>
>>>>
>>>
>>>

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to