The easiest way to suppress this error would be to disable the logging for TaskManagerStdoutFileHandler in your log4j.properties file.
Cheers, Till On Wed, Oct 7, 2020 at 8:48 PM sidhant gupta <sidhan...@gmail.com> wrote: > Hi Till, > > I understand the errors which appears in my logs are not stopping me from > running the job. I am running flink session cluster in ECS and also > configured graylog to get the container logs. So getting the docker logs is > also not an issue. > But is there a way to suppress this error or any work around ? > > Thanks > Sidhant Gupta > > On Wed, Oct 7, 2020, 9:15 PM Till Rohrmann <trohrm...@apache.org> wrote: > >> Hi Sidhant, >> >> when using Flink's Docker image, then the cluster won't create the out >> files. Instead the components will directly write to STDOUT which is >> captured by Kubernetes and can be viewed using `kubectl logs POD_NAME`. The >> error which appears in your logs is not a problem. It is simply the REST >> handler which tries to serve the out files. >> >> Cheers, >> Till >> >> On Wed, Oct 7, 2020 at 5:11 PM 大森林 <appleyu...@foxmail.com> wrote: >> >>> what's your running mode? >>> if your flink cluster is on yarn mode,then the output you need has no >>> relation to $FLINK_HOME/logs/*.out >>> >>> >>> ------------------ 原始邮件 ------------------ >>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>; >>> *发送时间:* 2020年10月7日(星期三) 晚上11:33 >>> *收件人:* "大森林"<appleyu...@foxmail.com>;"user"<user@flink.apache.org>; >>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor >>> >>> Hi, >>> >>> I'm running flink cluster in ecs. There is a pipeline which creates the >>> job manager and then the task manager using the docker image. >>> >>> Not sure if we would want to restart the cluster in production. >>> >>> Is there any way we can make sure the .out files will be created without >>> restart ? >>> >>> I am able to see the logs in the logs tab but not the stdout logs in the >>> web ui and getting the below mentioned error after running the job. >>> >>> Thanks >>> Sidhant Gupta >>> >>> >>> On Wed, Oct 7, 2020, 8:00 PM 大森林 <appleyu...@foxmail.com> wrote: >>> >>>> it's easy, >>>> just restart your flink cluster(standalone mode) >>>> >>>> if you run flink in yarn mode,then the result will display on >>>> $HADOOP/logs/*.out files >>>> >>>> ------------------ 原始邮件 ------------------ >>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>; >>>> *发送时间:* 2020年10月7日(星期三) 晚上9:52 >>>> *收件人:* "大森林"<appleyu...@foxmail.com>; >>>> *抄送:* "user"<user@flink.apache.org>; >>>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor >>>> >>>> ++ user >>>> >>>> On Wed, Oct 7, 2020, 6:47 PM sidhant gupta <sidhan...@gmail.com> wrote: >>>> >>>>> Hi >>>>> >>>>> I checked in the $FLINK_HOME/logs. The .out file was not there. Can >>>>> you suggest what should be the action item ? >>>>> >>>>> Thanks >>>>> Sidhant Gupta >>>>> >>>>> >>>>> On Wed, Oct 7, 2020, 7:17 AM 大森林 <appleyu...@foxmail.com> wrote: >>>>> >>>>>> >>>>>> check if the .out file is in $FLINK_HOME/logs please. >>>>>> >>>>>> ------------------ 原始邮件 ------------------ >>>>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>; >>>>>> *发送时间:* 2020年10月7日(星期三) 凌晨1:52 >>>>>> *收件人:* "大森林"<appleyu...@foxmail.com>; >>>>>> *主题:* Re: The file STDOUT does not exist on the TaskExecutor >>>>>> >>>>>> Hi, >>>>>> >>>>>> I am just running the docker container as it is by adding just the >>>>>> conf/flink.yaml . >>>>>> I am not sure if the .out file got deleted. Do we need to expose some >>>>>> ports ? >>>>>> >>>>>> Thanks >>>>>> Sidhant Gupta >>>>>> >>>>>> >>>>>> >>>>>> On Tue, Oct 6, 2020, 8:51 PM 大森林 <appleyu...@foxmail.com> wrote: >>>>>> >>>>>>> >>>>>>> Hi,I guess you may deleted .out file in $FLINK_HOME/logs. >>>>>>> you can just use your default log settings. >>>>>>> ------------------ 原始邮件 ------------------ >>>>>>> *发件人:* "sidhant gupta" <sidhan...@gmail.com>; >>>>>>> *发送时间:* 2020年10月6日(星期二) 晚上10:59 >>>>>>> *收件人:* "user"<user@flink.apache.org>; >>>>>>> *主题:* The file STDOUT does not exist on the TaskExecutor >>>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> I am running dockerized flink:1.11.0-scala_2.11 container in ecs. I >>>>>>> am getting the following error after the job runs: >>>>>>> >>>>>>> ERROR org.apache.flink.runtime.rest.handler.taskmanager. >>>>>>> TaskManagerStdoutFileHandler [] - Unhandled exception. >>>>>>> org.apache.flink.util.FlinkException: The file STDOUT does not >>>>>>> exist on the TaskExecutor. >>>>>>> at org.apache.flink.runtime.taskexecutor.TaskExecutor >>>>>>> .lambda$requestFileUploadByFilePath$25(TaskExecutor.java:1742) >>>>>>> ~[flink-dist_2.11-1.11.0.jar:1.11.0] >>>>>>> at java.util.concurrent.CompletableFuture$AsyncSupply.run( >>>>>>> CompletableFuture.java:1604) ~[?:1.8.0_262] >>>>>>> at java.util.concurrent.ThreadPoolExecutor.runWorker( >>>>>>> ThreadPoolExecutor.java:1149) ~[?:1.8.0_262] >>>>>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run( >>>>>>> ThreadPoolExecutor.java:624) ~[?:1.8.0_262] >>>>>>> at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_262] >>>>>>> >>>>>>> I guess "file" needs to be added in log4j.properties in the docker >>>>>>> container e.g. log4j.rootLogger=INFO, file >>>>>>> Are there any other properties which needs to be configured in any >>>>>>> of the other property files or any jar needs to be added in the >>>>>>> */opt/flink >>>>>>> *path ? >>>>>>> Thanks >>>>>>> Sidhant Gupta >>>>>>> >>>>>>>