You could issue "ps -ef | grep container_id_for_some_tm". And then you will
find the
following java options about log4j.

-Dlog.file=/var/log/hadoop-yarn/containers/application_xx/container_xx/taskmanager.log
-Dlog4j.configuration=file:./log4j.properties
-Dlog4j.configurationFile=file:./log4j.properties

Best,
Yang

Diwakar Jha <diwakar.n...@gmail.com> 于2020年11月2日周一 下午11:37写道:

> Sure. I will check that and get back to you. could you please share how to
> check java dynamic options?
>
> Best,
> Diwakar
>
> On Mon, Nov 2, 2020 at 1:33 AM Yang Wang <danrtsey...@gmail.com> wrote:
>
>> If you have already updated the log4j.properties, and it still could not
>> work, then I
>> suggest to log in the Yarn NodeManager machine and check the
>> log4j.properties
>> in the container workdir is correct. Also you could have a look at the
>> java dynamic
>> options are correctly set.
>>
>> I think it should work if the log4j.properties and java dynamic options
>> are set correctly.
>>
>> BTW, could you share the new yarn logs?
>>
>> Best,
>> Yang
>>
>> Diwakar Jha <diwakar.n...@gmail.com> 于2020年11月2日周一 下午4:32写道:
>>
>>>
>>>
>>> Hi Yang,
>>>
>>> Thank you so much for taking a look at the log files. I changed my
>>> log4j.properties. Below is the actual file that I got from EMR 6.1.0
>>> distribution of flink 1.11. I observed that it is different from Flink 1.11
>>> that i downloaded so i changed it. Still I didn't see any logs.
>>>
>>> *Actual*
>>> log4j.rootLogger=INFO,file
>>>
>>> # Log all infos in the given file
>>> log4j.appender.file=org.apache.log4j.FileAppender
>>> log4j.appender.file.file=${log.file}
>>> log4j.appender.file.append=false
>>> log4j.appender.file.layout=org.apache.log4j.PatternLayout
>>> log4j.appender.file.layout.ConversionPattern=%d{yyyy-MM-dd HH:mm:ss,SSS}
>>> %-5p %-60c %x - %m%n
>>>
>>> # suppress the irrelevant (wrong) warnings from the netty channel handler
>>> log4j.logger.org.jboss.netty.channel.DefaultChannelPipeline=ERROR,file
>>>
>>>
>>> *modified : *commented the above and added new logging from
>>> actual flink application log4.properties file
>>>
>>> #log4j.rootLogger=INFO,file
>>>
>>> # Log all infos in the given file
>>> #log4j.appender.file=org.apache.log4j.FileAppender
>>> #log4j.appender.file.file=${log.file}
>>> #log4j.appender.file.append=false
>>> #log4j.appender.file.layout=org.apache.log4j.PatternLayout
>>> #log4j.appender.file.layout.ConversionPattern=%d{yyyy-MM-dd
>>> HH:mm:ss,SSS} %-5p %-60c %x - %m%n
>>>
>>> # suppress the irrelevant (wrong) warnings from the netty channel handler
>>> #log4j.logger.org.jboss.netty.channel.DefaultChannelPipeline=ERROR,file
>>>
>>> # This affects logging for both user code and Flink
>>> rootLogger.level = INFO
>>> rootLogger.appenderRef.file.ref = MainAppender
>>>
>>> # Uncomment this if you want to _only_ change Flink's logging
>>> #logger.flink.name = org.apache.flink
>>> #logger.flink.level = INFO
>>>
>>> # The following lines keep the log level of common libraries/connectors
>>> on
>>> # log level INFO. The root logger does not override this. You have to
>>> manually
>>> # change the log levels here.
>>> logger.akka.name = akka
>>> logger.akka.level = INFO
>>> logger.kafka.name= org.apache.kafka
>>> logger.kafka.level = INFO
>>> logger.hadoop.name = org.apache.hadoop
>>> logger.hadoop.level = INFO
>>> logger.zookeeper.name = org.apache.zookeeper
>>> logger.zookeeper.level = INFO
>>>
>>> # Log all infos in the given file
>>> appender.main.name = MainAppender
>>> appender.main.type = File
>>> appender.main.append = false
>>> appender.main.fileName = ${sys:log.file}
>>> appender.main.layout.type = PatternLayout
>>> appender.main.layout.pattern = %d{yyyy-MM-dd HH:mm:ss,SSS} %-5p %-60c %x
>>> - %m%n
>>>
>>> # Suppress the irrelevant (wrong) warnings from the Netty channel handler
>>> logger.netty.name =
>>> org.apache.flink.shaded.akka.org.jboss.netty.channel.DefaultChannelPipeline
>>> logger.netty.level = OFF
>>>
>>> **********************************
>>> I also think its related to the log4j setting but I'm not able to figure
>>> it out.
>>> Please let me know if you want any other log files or configuration.
>>>
>>> Thanks.
>>>
>>> On Sun, Nov 1, 2020 at 10:06 PM Yang Wang <danrtsey...@gmail.com> wrote:
>>>
>>>> Hi Diwakar Jha,
>>>>
>>>> From the logs you have provided, everything seems working as expected.
>>>> The JobManager and TaskManager
>>>> java processes have been started with correct dynamic options,
>>>> especially for the logging.
>>>>
>>>> Could you share the content of $FLINK_HOME/conf/log4j.properties? I
>>>> think there's something wrong with the
>>>> log4j config file. For example, it is a log4j1 format. But we are using
>>>> log4j2 in Flink 1.11.
>>>>
>>>>
>>>> Best,
>>>> Yang
>>>>
>>>> Diwakar Jha <diwakar.n...@gmail.com> 于2020年11月2日周一 上午1:57写道:
>>>>
>>>>> Hi
>>>>> I'm running Flink 1.11 on EMR 6.1.0. I can see my job is running fine
>>>>> but i'm not seeing any taskmanager/jobmanager logs.
>>>>> I see the below error in stdout.
>>>>> 18:29:19.834 [flink-akka.actor.default-dispatcher-28] ERROR
>>>>> org.apache.flink.runtime.rest.handler.taskmanager.
>>>>> TaskManagerLogFileHandler - Failed to transfer file from TaskExecutor
>>>>> container_1604033334508_0001_01_000004.
>>>>> java.util.concurrent.CompletionException: org.apache.flink.util.
>>>>> FlinkException: The file LOG does not exist on the TaskExecutor.
>>>>>
>>>>> I'm stuck at this step for a couple of days now and not able to
>>>>> migrate to Flink 1.11. I would appreciate it if anyone can help me.
>>>>> i have the following setup :
>>>>> a) i'm deploying flink using yarn. I have attached yarn application
>>>>> id logs.
>>>>> c) stsd setup
>>>>>
>>>>> metrics.reporters: stsd
>>>>> metrics.reporter.stsd.factory.class: 
>>>>> org.apache.flink.metrics.statsd.StatsDReporterFactory
>>>>> metrics.reporter.stsd.host: localhost
>>>>> metrics.reporter.stsd.port: 8125
>>>>>
>>>>>

Reply via email to