https://spark.apache.org/docs/latest/running-on-yarn.html#debugging-your-application
When log aggregation isn’t turned on, logs are retained locally on each
machine under YARN_APP_LOGS_DIR, which is usually configured to/tmp/logs or
$HADOOP_HOME/logs/userlogs depending on the Hadoop version and
in
Doesn't work for me so far ,
using command but got such output. What should I check to fix the issue?
Any configuration parameters ...
[root@sdo-hdp-bd-master1 ~]# yarn logs -applicationId
application_1426424283508_0048
15/05/21 13:25:09 INFO impl.TimelineClientImpl: Timeline service address:
Oleg,
You can see applicationId in your Spark History Server.
Go to http://historyserver:18088/
Also check
https://spark.apache.org/docs/1.1.0/running-on-yarn.html#debugging-your-application
It should be no different with PySpark.
--
Ruslan Dautkhanov
On Wed, May 20, 2015 at 2:12 PM, Oleg Ru
Hi Ruslan.
Could you add more details please.
Where do I get applicationId? In case I have a lot of log files would it
make sense to view it from single point.
How actually I can configure / manage log location of PySpark?
Thanks
Oleg.
On Wed, May 20, 2015 at 10:24 PM, Ruslan Dautkhanov
wrote:
You could use
yarn logs -applicationId application_1383601692319_0008
--
Ruslan Dautkhanov
On Wed, May 20, 2015 at 5:37 AM, Oleg Ruchovets
wrote:
> Hi ,
>
> I am executing PySpark job on yarn ( hortonworks distribution).
>
> Could someone pointing me where is the log locations?
>
> Thanks
Hi ,
I am executing PySpark job on yarn ( hortonworks distribution).
Could someone pointing me where is the log locations?
Thanks
Oleg.