Re: PySpark Logs location

2015-05-21 Thread Ruslan Dautkhanov
https://spark.apache.org/docs/latest/running-on-yarn.html#debugging-your-application When log aggregation isn’t turned on, logs are retained locally on each machine under YARN_APP_LOGS_DIR, which is usually configured to/tmp/logs or $HADOOP_HOME/logs/userlogs depending on the Hadoop version and in

Re: PySpark Logs location

2015-05-21 Thread Oleg Ruchovets
Doesn't work for me so far , using command but got such output. What should I check to fix the issue? Any configuration parameters ... [root@sdo-hdp-bd-master1 ~]# yarn logs -applicationId application_1426424283508_0048 15/05/21 13:25:09 INFO impl.TimelineClientImpl: Timeline service address:

Re: PySpark Logs location

2015-05-20 Thread Ruslan Dautkhanov
Oleg, You can see applicationId in your Spark History Server. Go to http://historyserver:18088/ Also check https://spark.apache.org/docs/1.1.0/running-on-yarn.html#debugging-your-application It should be no different with PySpark. -- Ruslan Dautkhanov On Wed, May 20, 2015 at 2:12 PM, Oleg Ru

Re: PySpark Logs location

2015-05-20 Thread Oleg Ruchovets
Hi Ruslan. Could you add more details please. Where do I get applicationId? In case I have a lot of log files would it make sense to view it from single point. How actually I can configure / manage log location of PySpark? Thanks Oleg. On Wed, May 20, 2015 at 10:24 PM, Ruslan Dautkhanov wrote:

Re: PySpark Logs location

2015-05-20 Thread Ruslan Dautkhanov
You could use yarn logs -applicationId application_1383601692319_0008 -- Ruslan Dautkhanov On Wed, May 20, 2015 at 5:37 AM, Oleg Ruchovets wrote: > Hi , > > I am executing PySpark job on yarn ( hortonworks distribution). > > Could someone pointing me where is the log locations? > > Thanks

PySpark Logs location

2015-05-20 Thread Oleg Ruchovets
Hi , I am executing PySpark job on yarn ( hortonworks distribution). Could someone pointing me where is the log locations? Thanks Oleg.