Sure. I will check that and get back to you. could you please share how to check java dynamic options?
Best, Diwakar On Mon, Nov 2, 2020 at 1:33 AM Yang Wang <danrtsey...@gmail.com> wrote: > If you have already updated the log4j.properties, and it still could not > work, then I > suggest to log in the Yarn NodeManager machine and check the > log4j.properties > in the container workdir is correct. Also you could have a look at the > java dynamic > options are correctly set. > > I think it should work if the log4j.properties and java dynamic options > are set correctly. > > BTW, could you share the new yarn logs? > > Best, > Yang > > Diwakar Jha <diwakar.n...@gmail.com> 于2020年11月2日周一 下午4:32写道: > >> >> >> Hi Yang, >> >> Thank you so much for taking a look at the log files. I changed my >> log4j.properties. Below is the actual file that I got from EMR 6.1.0 >> distribution of flink 1.11. I observed that it is different from Flink 1.11 >> that i downloaded so i changed it. Still I didn't see any logs. >> >> *Actual* >> log4j.rootLogger=INFO,file >> >> # Log all infos in the given file >> log4j.appender.file=org.apache.log4j.FileAppender >> log4j.appender.file.file=${log.file} >> log4j.appender.file.append=false >> log4j.appender.file.layout=org.apache.log4j.PatternLayout >> log4j.appender.file.layout.ConversionPattern=%d{yyyy-MM-dd HH:mm:ss,SSS} >> %-5p %-60c %x - %m%n >> >> # suppress the irrelevant (wrong) warnings from the netty channel handler >> log4j.logger.org.jboss.netty.channel.DefaultChannelPipeline=ERROR,file >> >> >> *modified : *commented the above and added new logging from actual flink >> application log4.properties file >> >> #log4j.rootLogger=INFO,file >> >> # Log all infos in the given file >> #log4j.appender.file=org.apache.log4j.FileAppender >> #log4j.appender.file.file=${log.file} >> #log4j.appender.file.append=false >> #log4j.appender.file.layout=org.apache.log4j.PatternLayout >> #log4j.appender.file.layout.ConversionPattern=%d{yyyy-MM-dd HH:mm:ss,SSS} >> %-5p %-60c %x - %m%n >> >> # suppress the irrelevant (wrong) warnings from the netty channel handler >> #log4j.logger.org.jboss.netty.channel.DefaultChannelPipeline=ERROR,file >> >> # This affects logging for both user code and Flink >> rootLogger.level = INFO >> rootLogger.appenderRef.file.ref = MainAppender >> >> # Uncomment this if you want to _only_ change Flink's logging >> #logger.flink.name = org.apache.flink >> #logger.flink.level = INFO >> >> # The following lines keep the log level of common libraries/connectors on >> # log level INFO. The root logger does not override this. You have to >> manually >> # change the log levels here. >> logger.akka.name = akka >> logger.akka.level = INFO >> logger.kafka.name= org.apache.kafka >> logger.kafka.level = INFO >> logger.hadoop.name = org.apache.hadoop >> logger.hadoop.level = INFO >> logger.zookeeper.name = org.apache.zookeeper >> logger.zookeeper.level = INFO >> >> # Log all infos in the given file >> appender.main.name = MainAppender >> appender.main.type = File >> appender.main.append = false >> appender.main.fileName = ${sys:log.file} >> appender.main.layout.type = PatternLayout >> appender.main.layout.pattern = %d{yyyy-MM-dd HH:mm:ss,SSS} %-5p %-60c %x >> - %m%n >> >> # Suppress the irrelevant (wrong) warnings from the Netty channel handler >> logger.netty.name = >> org.apache.flink.shaded.akka.org.jboss.netty.channel.DefaultChannelPipeline >> logger.netty.level = OFF >> >> ********************************** >> I also think its related to the log4j setting but I'm not able to figure >> it out. >> Please let me know if you want any other log files or configuration. >> >> Thanks. >> >> On Sun, Nov 1, 2020 at 10:06 PM Yang Wang <danrtsey...@gmail.com> wrote: >> >>> Hi Diwakar Jha, >>> >>> From the logs you have provided, everything seems working as expected. >>> The JobManager and TaskManager >>> java processes have been started with correct dynamic options, >>> especially for the logging. >>> >>> Could you share the content of $FLINK_HOME/conf/log4j.properties? I >>> think there's something wrong with the >>> log4j config file. For example, it is a log4j1 format. But we are using >>> log4j2 in Flink 1.11. >>> >>> >>> Best, >>> Yang >>> >>> Diwakar Jha <diwakar.n...@gmail.com> 于2020年11月2日周一 上午1:57写道: >>> >>>> Hi >>>> I'm running Flink 1.11 on EMR 6.1.0. I can see my job is running fine >>>> but i'm not seeing any taskmanager/jobmanager logs. >>>> I see the below error in stdout. >>>> 18:29:19.834 [flink-akka.actor.default-dispatcher-28] ERROR >>>> org.apache.flink.runtime.rest.handler.taskmanager. >>>> TaskManagerLogFileHandler - Failed to transfer file from TaskExecutor >>>> container_1604033334508_0001_01_000004. >>>> java.util.concurrent.CompletionException: org.apache.flink.util. >>>> FlinkException: The file LOG does not exist on the TaskExecutor. >>>> >>>> I'm stuck at this step for a couple of days now and not able to migrate >>>> to Flink 1.11. I would appreciate it if anyone can help me. >>>> i have the following setup : >>>> a) i'm deploying flink using yarn. I have attached yarn application id >>>> logs. >>>> c) stsd setup >>>> >>>> metrics.reporters: stsd >>>> metrics.reporter.stsd.factory.class: >>>> org.apache.flink.metrics.statsd.StatsDReporterFactory >>>> metrics.reporter.stsd.host: localhost >>>> metrics.reporter.stsd.port: 8125 >>>> >>>>