Sure. I will check that and get back to you. could you please share how to
check java dynamic options?

Best,
Diwakar

On Mon, Nov 2, 2020 at 1:33 AM Yang Wang <danrtsey...@gmail.com> wrote:

> If you have already updated the log4j.properties, and it still could not
> work, then I
> suggest to log in the Yarn NodeManager machine and check the
> log4j.properties
> in the container workdir is correct. Also you could have a look at the
> java dynamic
> options are correctly set.
>
> I think it should work if the log4j.properties and java dynamic options
> are set correctly.
>
> BTW, could you share the new yarn logs?
>
> Best,
> Yang
>
> Diwakar Jha <diwakar.n...@gmail.com> 于2020年11月2日周一 下午4:32写道:
>
>>
>>
>> Hi Yang,
>>
>> Thank you so much for taking a look at the log files. I changed my
>> log4j.properties. Below is the actual file that I got from EMR 6.1.0
>> distribution of flink 1.11. I observed that it is different from Flink 1.11
>> that i downloaded so i changed it. Still I didn't see any logs.
>>
>> *Actual*
>> log4j.rootLogger=INFO,file
>>
>> # Log all infos in the given file
>> log4j.appender.file=org.apache.log4j.FileAppender
>> log4j.appender.file.file=${log.file}
>> log4j.appender.file.append=false
>> log4j.appender.file.layout=org.apache.log4j.PatternLayout
>> log4j.appender.file.layout.ConversionPattern=%d{yyyy-MM-dd HH:mm:ss,SSS}
>> %-5p %-60c %x - %m%n
>>
>> # suppress the irrelevant (wrong) warnings from the netty channel handler
>> log4j.logger.org.jboss.netty.channel.DefaultChannelPipeline=ERROR,file
>>
>>
>> *modified : *commented the above and added new logging from actual flink
>> application log4.properties file
>>
>> #log4j.rootLogger=INFO,file
>>
>> # Log all infos in the given file
>> #log4j.appender.file=org.apache.log4j.FileAppender
>> #log4j.appender.file.file=${log.file}
>> #log4j.appender.file.append=false
>> #log4j.appender.file.layout=org.apache.log4j.PatternLayout
>> #log4j.appender.file.layout.ConversionPattern=%d{yyyy-MM-dd HH:mm:ss,SSS}
>> %-5p %-60c %x - %m%n
>>
>> # suppress the irrelevant (wrong) warnings from the netty channel handler
>> #log4j.logger.org.jboss.netty.channel.DefaultChannelPipeline=ERROR,file
>>
>> # This affects logging for both user code and Flink
>> rootLogger.level = INFO
>> rootLogger.appenderRef.file.ref = MainAppender
>>
>> # Uncomment this if you want to _only_ change Flink's logging
>> #logger.flink.name = org.apache.flink
>> #logger.flink.level = INFO
>>
>> # The following lines keep the log level of common libraries/connectors on
>> # log level INFO. The root logger does not override this. You have to
>> manually
>> # change the log levels here.
>> logger.akka.name = akka
>> logger.akka.level = INFO
>> logger.kafka.name= org.apache.kafka
>> logger.kafka.level = INFO
>> logger.hadoop.name = org.apache.hadoop
>> logger.hadoop.level = INFO
>> logger.zookeeper.name = org.apache.zookeeper
>> logger.zookeeper.level = INFO
>>
>> # Log all infos in the given file
>> appender.main.name = MainAppender
>> appender.main.type = File
>> appender.main.append = false
>> appender.main.fileName = ${sys:log.file}
>> appender.main.layout.type = PatternLayout
>> appender.main.layout.pattern = %d{yyyy-MM-dd HH:mm:ss,SSS} %-5p %-60c %x
>> - %m%n
>>
>> # Suppress the irrelevant (wrong) warnings from the Netty channel handler
>> logger.netty.name =
>> org.apache.flink.shaded.akka.org.jboss.netty.channel.DefaultChannelPipeline
>> logger.netty.level = OFF
>>
>> **********************************
>> I also think its related to the log4j setting but I'm not able to figure
>> it out.
>> Please let me know if you want any other log files or configuration.
>>
>> Thanks.
>>
>> On Sun, Nov 1, 2020 at 10:06 PM Yang Wang <danrtsey...@gmail.com> wrote:
>>
>>> Hi Diwakar Jha,
>>>
>>> From the logs you have provided, everything seems working as expected.
>>> The JobManager and TaskManager
>>> java processes have been started with correct dynamic options,
>>> especially for the logging.
>>>
>>> Could you share the content of $FLINK_HOME/conf/log4j.properties? I
>>> think there's something wrong with the
>>> log4j config file. For example, it is a log4j1 format. But we are using
>>> log4j2 in Flink 1.11.
>>>
>>>
>>> Best,
>>> Yang
>>>
>>> Diwakar Jha <diwakar.n...@gmail.com> 于2020年11月2日周一 上午1:57写道:
>>>
>>>> Hi
>>>> I'm running Flink 1.11 on EMR 6.1.0. I can see my job is running fine
>>>> but i'm not seeing any taskmanager/jobmanager logs.
>>>> I see the below error in stdout.
>>>> 18:29:19.834 [flink-akka.actor.default-dispatcher-28] ERROR
>>>> org.apache.flink.runtime.rest.handler.taskmanager.
>>>> TaskManagerLogFileHandler - Failed to transfer file from TaskExecutor
>>>> container_1604033334508_0001_01_000004.
>>>> java.util.concurrent.CompletionException: org.apache.flink.util.
>>>> FlinkException: The file LOG does not exist on the TaskExecutor.
>>>>
>>>> I'm stuck at this step for a couple of days now and not able to migrate
>>>> to Flink 1.11. I would appreciate it if anyone can help me.
>>>> i have the following setup :
>>>> a) i'm deploying flink using yarn. I have attached yarn application id
>>>> logs.
>>>> c) stsd setup
>>>>
>>>> metrics.reporters: stsd
>>>> metrics.reporter.stsd.factory.class: 
>>>> org.apache.flink.metrics.statsd.StatsDReporterFactory
>>>> metrics.reporter.stsd.host: localhost
>>>> metrics.reporter.stsd.port: 8125
>>>>
>>>>

Reply via email to