Hi Sidhant,

when using Flink's Docker image, then the cluster won't create the out
files. Instead the components will directly write to STDOUT which is
captured by Kubernetes and can be viewed using `kubectl logs POD_NAME`. The
error which appears in your logs is not a problem. It is simply the REST
handler which tries to serve the out files.

Cheers,
Till

On Wed, Oct 7, 2020 at 5:11 PM 大森林 <appleyu...@foxmail.com> wrote:

> what's your running mode?
> if your flink cluster is on yarn mode,then the output you need has no
> relation to $FLINK_HOME/logs/*.out
>
>
> ------------------ 原始邮件 ------------------
> *发件人:* "sidhant gupta" <sidhan...@gmail.com>;
> *发送时间:* 2020年10月7日(星期三) 晚上11:33
> *收件人:* "大森林"<appleyu...@foxmail.com>;"user"<user@flink.apache.org>;
> *主题:* Re: The file STDOUT does not exist on the TaskExecutor
>
> Hi,
>
> I'm running flink cluster in ecs. There is a pipeline which creates the
> job manager and then the task manager using the docker image.
>
> Not sure if we would want to restart the cluster in production.
>
> Is there any way we can make sure the .out files will be created without
> restart ?
>
> I am able to see the logs in the logs tab but not the stdout logs in the
> web ui and getting the below mentioned error after running the job.
>
> Thanks
> Sidhant Gupta
>
>
> On Wed, Oct 7, 2020, 8:00 PM 大森林 <appleyu...@foxmail.com> wrote:
>
>> it's easy,
>> just restart your flink cluster(standalone mode)
>>
>> if you run flink in yarn mode,then the result will display on
>> $HADOOP/logs/*.out files
>>
>> ------------------ 原始邮件 ------------------
>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>;
>> *发送时间:* 2020年10月7日(星期三) 晚上9:52
>> *收件人:* "大森林"<appleyu...@foxmail.com>;
>> *抄送:* "user"<user@flink.apache.org>;
>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor
>>
>> ++ user
>>
>> On Wed, Oct 7, 2020, 6:47 PM sidhant gupta <sidhan...@gmail.com> wrote:
>>
>>> Hi
>>>
>>> I checked in the $FLINK_HOME/logs. The .out file was not there. Can you
>>> suggest what should be the action item ?
>>>
>>> Thanks
>>> Sidhant Gupta
>>>
>>>
>>> On Wed, Oct 7, 2020, 7:17 AM 大森林 <appleyu...@foxmail.com> wrote:
>>>
>>>>
>>>> check if the .out file is in $FLINK_HOME/logs  please.
>>>>
>>>> ------------------ 原始邮件 ------------------
>>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>;
>>>> *发送时间:* 2020年10月7日(星期三) 凌晨1:52
>>>> *收件人:* "大森林"<appleyu...@foxmail.com>;
>>>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor
>>>>
>>>> Hi,
>>>>
>>>> I am just running the docker container as it is by adding just the
>>>> conf/flink.yaml .
>>>> I am not sure if the .out file got deleted. Do we need to expose some
>>>> ports ?
>>>>
>>>> Thanks
>>>> Sidhant Gupta
>>>>
>>>>
>>>>
>>>> On Tue, Oct 6, 2020, 8:51 PM 大森林 <appleyu...@foxmail.com> wrote:
>>>>
>>>>>
>>>>> Hi,I guess you may deleted .out file in $FLINK_HOME/logs.
>>>>> you can just use your default log settings.
>>>>> ------------------ 原始邮件 ------------------
>>>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>;
>>>>> *发送时间:* 2020年10月6日(星期二) 晚上10:59
>>>>> *收件人:* "user"<user@flink.apache.org>;
>>>>> *主题:* The file STDOUT does not exist on the TaskExecutor
>>>>>
>>>>> Hi,
>>>>>
>>>>> I am running dockerized flink:1.11.0-scala_2.11 container in ecs. I
>>>>> am getting the following error after the job runs:
>>>>>
>>>>> ERROR org.apache.flink.runtime.rest.handler.taskmanager.
>>>>> TaskManagerStdoutFileHandler [] - Unhandled exception.
>>>>> org.apache.flink.util.FlinkException: The file STDOUT does not exist
>>>>> on the TaskExecutor.
>>>>>     at org.apache.flink.runtime.taskexecutor.TaskExecutor
>>>>> .lambda$requestFileUploadByFilePath$25(TaskExecutor.java:1742)
>>>>> ~[flink-dist_2.11-1.11.0.jar:1.11.0]
>>>>>     at java.util.concurrent.CompletableFuture$AsyncSupply.run(
>>>>> CompletableFuture.java:1604) ~[?:1.8.0_262]
>>>>>     at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>> ThreadPoolExecutor.java:1149) ~[?:1.8.0_262]
>>>>>     at java.util.concurrent.ThreadPoolExecutor$Worker.run(
>>>>> ThreadPoolExecutor.java:624) ~[?:1.8.0_262]
>>>>>     at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_262]
>>>>>
>>>>> I guess "file" needs to be added in log4j.properties in the docker
>>>>> container e.g. log4j.rootLogger=INFO, file
>>>>> Are there any other properties which needs to be configured in any of
>>>>> the other property files or any jar needs to be added in the */opt/flink
>>>>> *path ?
>>>>> Thanks
>>>>> Sidhant Gupta
>>>>>
>>>>>

Reply via email to