When I run spark-shell and hive cml, everything is good.
One more thing I find is that when I run the hive CML, there is single hive job 
running (seen from hue job browser) all the time until I exit from hive cml.
No matter how many HQL I submit, it is still single job there. I think it is 
because of using Tez. Is Tez having conflict with zeppelin?

> On Dec 3, 2015, at 2:15 AM, Felix Cheung <felixcheun...@hotmail.com> wrote:
> 
> I don't know enough about HDP, but there should be a way to check user queue 
> in YARN?
> Spark job shouldn't affect Hive job though. Have you tried running 
> spark-shell (--master yarn-client) and a Hive job at the same time?
>  
> From: will...@gmail.com
> Subject: Re: zeppelin job is running all the time
> Date: Tue, 1 Dec 2015 22:56:55 -0500
> To: users@zeppelin.incubator.apache.org
> 
> Do I have to stop zeppelin to make hive job running?
> I am thinking to have one notebook have both spark and hive coding running.
> 
> In addition, I am quite sure the first job in the picture is hive job. The 
> 2nd one is zeppelin running spark. I wonder how to unblock hive job.
> 
> On Dec 1, 2015, at 10:45 PM, Jeff Zhang <zjf...@gmail.com 
> <mailto:zjf...@gmail.com>> wrote:
> 
> >>> the status of hive query becomes from ACCEPT to RUNNING
> The status you see on the hue is the yarn app status which is used for spark. 
> It should be running forever until you shutdown it in zeppelin  
> 
> On Wed, Dec 2, 2015 at 11:42 AM, Will Du <will...@gmail.com 
> <mailto:will...@gmail.com>> wrote:
> The latest HDP sandbox 2.3.2
> 
> On Dec 1, 2015, at 10:38 PM, Jeff Zhang <zjf...@gmail.com 
> <mailto:zjf...@gmail.com>> wrote:
> 
> Which version of HDP do you use ?
> 
> On Wed, Dec 2, 2015 at 11:23 AM, Will Du <will...@gmail.com 
> <mailto:will...@gmail.com>> wrote:
> I have assigned an dedicate yarn queue to spark, the status of hive query 
> becomes from ACCEPT to RUNNING. However, it sees running forever still. 
> Everything is else is in default config of HDP sandbox. Do I need to set 
> something else? The browser I saw status is from hue.
> 
> Thanks,
> wd
> 
> On Nov 30, 2015, at 12:40 AM, Rick Moritz <rah...@gmail.com 
> <mailto:rah...@gmail.com>> wrote:
> 
> To explain the previous reply: the SparkContext created by Zeppelin is 
> persistent and independent of whether it's currently processing a paragraph 
> or not. Therefore the Zeppelin job will claim all ressources assigned to it 
> until Zeppelin is stopped.
> 
> On Mon, Nov 30, 2015 at 4:20 AM, Jeff Zhang <zjf...@gmail.com 
> <mailto:zjf...@gmail.com>> wrote:
> I assume this is yarn app Job Browser. How many executors do you specify for 
> your zeppelin yarn app ? It seems your zeppelin yarn app consume all the 
> resources so that it block other applications.
> 
> 
> 
> On Mon, Nov 30, 2015 at 11:10 AM, Will Du <will...@gmail.com 
> <mailto:will...@gmail.com>> wrote:
> Hi folks,
> I am running a simple scala word count from zeppelin in HDP sandbox. The job 
> is successful with expected result. But the job of zeppelin is shown in job 
> status of hue forever. It seems block the my hive job. Does anyone know why?
> thanks,
> wd
> <PastedGraphic-1.tiff>
> 
> 
> 
> -- 
> Best Regards
> 
> Jeff Zhang
> 
> 
> 
> 
> 
> -- 
> Best Regards
> 
> Jeff Zhang
> 
> 
> 
> 
> -- 
> Best Regards
> 
> Jeff Zhang
> 

Reply via email to