Hi Robert,
Could please correct me. I'm not able to stop the app. Also, i
stopped flink job already.

sh-4.2$ yarn app -stop application_1603649952937_0002
2020-10-27 20:04:25,543 INFO client.RMProxy: Connecting to ResourceManager
at ip-10-0-55-50.ec2.internal/10.0.55.50:8032
2020-10-27 20:04:25,717 INFO client.AHSProxy: Connecting to Application
History server at ip-10-0-55-50.ec2.internal/10.0.55.50:10200
Exception in thread "main" java.lang.IllegalArgumentException: App admin
client class name not specified for type Apache Flink
        at
org.apache.hadoop.yarn.client.api.AppAdminClient.createAppAdminClient(AppAdminClient.java:76)
        at
org.apache.hadoop.yarn.client.cli.ApplicationCLI.run(ApplicationCLI.java:597)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
        at
org.apache.hadoop.yarn.client.cli.ApplicationCLI.main(ApplicationCLI.java:126)
sh-4.2$

On Tue, Oct 27, 2020 at 9:34 AM Robert Metzger <rmetz...@apache.org> wrote:

> Hi,
> are you intentionally not posting this response to the mailing list?
>
> As you can see from the yarn logs, log aggregation only works for finished
> applications ("End of LogType:prelaunch.out.This log file belongs to a
> running container (container_1603649952937_0002_01_000002) and so may not
> be complete.")
>
> Please stop the app, then provide the logs.
>
>
> On Tue, Oct 27, 2020 at 5:11 PM Diwakar Jha <diwakar.n...@gmail.com>
> wrote:
>
>> Hi Robert,
>>
>> Yes, i'm using Flink on EMR using YARN. Please find attached the yarn
>> logs -applicationId. I also attached haddop-yarn-nodemanager logs.
>> Also, I followed this link below which has the same problem :
>> http://mail-archives.apache.org/mod_mbox/flink-user/202009.mbox/%3CCAGDv3o5WyJTrXs9Pg+Vy-b+LwgEE26iN54iqE0=f5t+m8vw...@mail.gmail.com%3E
>>
>> https://www.talkend.net/post/75078.html
>> Based on this I changed the log4j.properties.
>> Let me know what you think. Please also let me know if you need some
>> specific logs.  Appreciate your help.
>>
>> Best,
>> Diwakar
>>
>> On Tue, Oct 27, 2020 at 12:26 AM Robert Metzger <rmetz...@apache.org>
>> wrote:
>>
>>> Hey Diwakar,
>>>
>>> how are you deploying Flink on EMR? Are you using YARN?
>>> If so, you could also use log aggregation to see all the logs at once
>>> (from both JobManager and TaskManagers). (yarn logs -applicationId
>>> <Application ID>)
>>>
>>> Could you post (or upload somewhere) all logs you have of one run? It is
>>> much easier for us to debug something if we have the full logs (the logs
>>> show for example the classpath that you are using, we would see how you are
>>> deploying Flink, etc.)
>>>
>>> From the information available, my guess is that you have modified your
>>> deployment in some way (use of a custom logging version, custom deployment
>>> method, version mixup with jars from both Flink 1.8 and 1.11, ...).
>>>
>>> Best,
>>> Robert
>>>
>>>
>>> On Tue, Oct 27, 2020 at 12:41 AM Diwakar Jha <diwakar.n...@gmail.com>
>>> wrote:
>>>
>>>> This is what I see on the WebUI.
>>>>
>>>> 23:19:24.263 [flink-akka.actor.default-dispatcher-1865] ERROR
>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>> - Failed to transfer file from TaskExecutor
>>>> container_1603649952937_0002_01_000004.
>>>> java.util.concurrent.CompletionException:
>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>> TaskExecutor. at
>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
>>>> Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252] Caused by:
>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>> TaskExecutor. ... 5 more 23:19:24.275
>>>> [flink-akka.actor.default-dispatcher-1865] ERROR
>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>> - Unhandled exception. org.apache.flink.util.FlinkException: The file LOG
>>>> does not exist on the TaskExecutor. at
>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
>>>> Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252]
>>>>
>>>> Appreciate if anyone has any pointer for this.
>>>>
>>>> On Mon, Oct 26, 2020 at 10:45 AM Chesnay Schepler <ches...@apache.org>
>>>> wrote:
>>>>
>>>>> Flink 1.11 uses slf4j 1.7.15; the easiest way to check the log files
>>>>> is usually via the WebUI.
>>>>>
>>>>> On 10/26/2020 5:30 PM, Diwakar Jha wrote:
>>>>>
>>>>> I think my problem is with Sl4j library. I'm using sl4j 1.7 with Flink
>>>>> 1.11. If that's correct then i appreciate if someone can point me to the
>>>>> exact Slf4j library that i should use with Flink 1.11
>>>>>
>>>>> Flink = 1.11.x;
>>>>> Slf4j = 1.7;
>>>>>
>>>>>
>>>>> On Sun, Oct 25, 2020 at 8:00 PM Diwakar Jha <diwakar.n...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Thanks for checking my configurations. Could you also point me where
>>>>>> I can see the log files? Just to give more details. I'm trying to access
>>>>>> these logs in AWS cloudwatch.
>>>>>>
>>>>>> Best,
>>>>>> Diwakar
>>>>>>
>>>>>> On Sun, Oct 25, 2020 at 2:16 PM Chesnay Schepler <ches...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> With Flink 1.11 reporters were refactored to plugins, and are now
>>>>>>> accessible by default (so you no longer have to bother with copying jars
>>>>>>> around).
>>>>>>>
>>>>>>> Your configuration appears to be correct, so I suggest to take a
>>>>>>> look at the log files.
>>>>>>>
>>>>>>> On 10/25/2020 9:52 PM, Diwakar Jha wrote:
>>>>>>>
>>>>>>> Hello Everyone,
>>>>>>>
>>>>>>> I'm new to flink and i'm trying to upgrade from flink 1.8 to flink
>>>>>>> 1.11 on an emr cluster. after upgrading to flink1.11 One of the 
>>>>>>> differences
>>>>>>> that i see is i don't get any metrics. I found out that flink 1.11 does 
>>>>>>> not
>>>>>>> have *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar in
>>>>>>> /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have 
>>>>>>> any
>>>>>>> pointer to locate
>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or how
>>>>>>> to use metrics in flink.1.11?
>>>>>>>
>>>>>>> Things i tried :
>>>>>>> a) the below setup
>>>>>>>
>>>>>>> metrics.reporters: stsdmetrics.reporter.stsd.factory.class: 
>>>>>>> org.apache.flink.metrics.statsd.StatsDReporterFactorymetrics.reporter.stsd.host:
>>>>>>>  localhostmetrics.reporter.stsd.port: 8125
>>>>>>>
>>>>>>> b) I tried downloading the statsd jar from
>>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
>>>>>>> putting it inside plugins/statsd directory.
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best,
>>>>>>> Diwakar Jha.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best,
>>>>>> Diwakar Jha.
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Best,
>>>>> Diwakar Jha.
>>>>>
>>>>>
>>>>>
>>>>
>>>> --
>>>> Best,
>>>> Diwakar Jha.
>>>>
>>>
>>
>> --
>> Best,
>> Diwakar Jha.
>>
>

-- 
Best,
Diwakar Jha.

Reply via email to