Please use the `yarn logs -applicationId <id>` to retrieve the logs. If you
have enabled log aggregation, this will give you all container logs
concatenated.

Cheers,
Max

On Wed, Jun 15, 2016 at 12:24 AM, Theofilos Kakantousis <t...@kth.se> wrote:

> Hi Max,
>
> The runBlocking(..) problem was due to a Netty dependency issue on my
> project, it works fine now :)
>
> To pinpoint the logging issue, I just ran a single flink job on yarn as
> per the documentation "./bin/flink run -m yarn-cluster -yn 2
> ./examples/streaming/Iteration.jar" and I have the same issue.During the
> job I can see in the containers the taskmanager logs, and a sample output
> from the taskmanager.out is the following:
> "cat
> /srv/hadoop/logs/userlogs/application_1465901188070_0037/container_1465901188070_0037_01_000002/taskmanager.out
>
> 2> ((49,1),3)
> 2> ((25,11),4)
> 2> ((46,44),2
> .."
>
> However, the yarn aggregated log contains only the jobmanager output. Is
> this expected or could it indicate a problem with my hadoop logging
> configuration not picking up taskmanager logs?
>
> Cheers,
> Theofilos
>
>
> On 6/13/2016 12:13 PM, Maximilian Michels wrote:
>
>> Hi Theofilos,
>>
>> Flink doesn't send the local client output to the Yarn cluster. I
>> think this will only change once we move the entire execution of the
>> Job to the cluster framework. All output of the actual Flink job
>> should be within the JobManager or TaskManager logs.
>>
>> There is something wrong with the network communication if the Client
>> doesn't return from `runBlocking(..)`. Would be interesting to take a
>> look at the logs to find out why that could be.
>>
>> Cheers,
>> Max
>>
>>
>> On Sat, Jun 11, 2016 at 1:53 PM, Theofilos Kakantousis <t...@kth.se>
>> wrote:
>>
>>> Hi Robert,
>>>
>>> Thanks for the prompt reply. I'm using the IterateExample from Flink
>>> examples. In the yarn log I get entries for the YarnJobManager and
>>> ExecutionGraph, but I was wondering if there is a way to push all the
>>> logging that the client produces into the yarn log. Including the
>>> System.out
>>> calls. Is there a way to modify the example to use a logging framework to
>>> achieve it?
>>>
>>> Also when I submit the program using the Client runBlocking method,
>>> although
>>> I see in the taskmanager and jobmanager log that the application has
>>> finished, the runBlocking method does not return. Should I call it in a
>>> separate thread?
>>>
>>> Cheers,
>>> Theofilos
>>>
>>> On 2016-06-10 22:12, Robert Metzger wrote:
>>>
>>> Hi Theofilos,
>>>
>>> how exactly are you writing the application output?
>>> Are you using a logging framework?
>>> Are you writing the log statements from the open(), map(), invoke()
>>> methods
>>> or from some constructors? (I'm asking since different parts are
>>> executed on
>>> the cluster and locally).
>>>
>>> On Fri, Jun 10, 2016 at 4:00 PM, Theofilos Kakantousis <t...@kth.se>
>>> wrote:
>>>
>>>> Hi all,
>>>>
>>>> Flink 1.0.3
>>>> Hadoop 2.4.0
>>>>
>>>> When running a job on a Flink Cluster on Yarn, the application output is
>>>> not included in the Yarn log. Instead, it is only printed in the stdout
>>>> from
>>>> where I run my program.  For the jobmanager, I'm using the
>>>> log4j.properties
>>>> file from the flink/conf directory. Yarn log aggregation is enabled and
>>>> the
>>>> YarnJobManager log is printed in the yarn log. The application is
>>>> submitted
>>>> by a Flink Client to the FlinkYarnCluster using a PackagedProgram.
>>>>
>>>> Is this expected behavior and if so, is there a way to include the
>>>> application output in the Yarn aggregated log? Thanks!
>>>>
>>>> Cheers,
>>>> Theofilos
>>>>
>>>>
>>>
>

Reply via email to