You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@flink.apache.org by Diwakar Jha <di...@gmail.com> on 2020/10/25 20:52:39 UTC

how to enable metrics in Flink 1.11

Hello Everyone,

I'm new to flink and i'm trying to upgrade from flink 1.8 to flink 1.11 on
an emr cluster. after upgrading to flink1.11 One of the differences that i
see is i don't get any metrics. I found out that flink 1.11 does not have
*org.apache.flink.metrics.statsd.StatsDReporterFactory* jar in
/usr/lib/flink/opt which was the case for flink 1.8. Could anyone have any
pointer to locate *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar
or how to use metrics in flink.1.11?

Things i tried :
a) the below setup

metrics.reporters: stsd
metrics.reporter.stsd.factory.class:
org.apache.flink.metrics.statsd.StatsDReporterFactory
metrics.reporter.stsd.host: localhost
metrics.reporter.stsd.port: 8125

b) I tried downloading the statsd jar from
https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
putting it inside plugins/statsd directory.


-- 
Best,
Diwakar Jha.

Re: how to enable metrics in Flink 1.11

Posted by Diwakar Jha <di...@gmail.com>.
HI Robert,

I'm able to see taskmanage and jobmanager logs after I changed the
log4j.properties file (/usr/lib/flink/conf).
It seems to be a problem with EMR 6.1 distribution. the
log4j.properties files is different in the Flink package that I downloaded
and the one that comes with EMR 6.1. I replaced the log4j.properties and
it's working.
Thanks for helping me debug the issue.

Best,
Diwakar

On Tue, Nov 3, 2020 at 11:36 AM Robert Metzger <rm...@apache.org> wrote:

> Hey Diwakar,
>
> the logs you are providing still don't contain the full Flink logs.
>
> You can not stop the Flink on YARN using "yarn app -stop
> application_1603649952937_0002". To stop Flink on YARN, use: "yarn
> application -kill <appId>".
>
>
>
> On Sat, Oct 31, 2020 at 6:26 PM Diwakar Jha <di...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I wanted to check if anyone can help me with the logs. I have sent
>> several emails but not getting any response.
>>
>> I'm running Flink 1.11 on EMR 6.1. I don't see any logs though I get this
>> stdout error. I'm trying to upgrade Flink 1.8 to Flink 1.11
>>
>> 18:29:19.834 [flink-akka.actor.default-dispatcher-28] ERROR
>> org.apache.flink.runtime.rest.handler.taskmanager.
>> TaskManagerLogFileHandler - Failed to transfer file from TaskExecutor
>> container_1604033334508_0001_01_000004.
>> java.util.concurrent.CompletionException: org.apache.flink.util.
>> FlinkException: The file LOG does not exist on the TaskExecutor.
>>
>> Thanks!
>>
>>
>> On Fri, Oct 30, 2020 at 9:04 AM Diwakar Jha <di...@gmail.com>
>> wrote:
>>
>>> Hello,
>>>
>>> I see that in my class path (below) I have both log4j-1 and lo4j-api-2.
>>> is this because of which i'm not seeing any logs. If so, could someone
>>> suggest how to fix it?
>>>
>>> export
>>> CLASSPATH=":lib/flink-csv-1.11.0.jar:lib/flink-json-1.11.0.jar:lib/flink-shaded-zookeeper-3.4.14.jar:lib/flink-table-blink_2.12-1.11.0.jar:lib/flink-table_2.12-1.11.0.jar:
>>> *lib/log4j-1.2-api-2.12.1.jar:lib/log4j-api-2.12.1.jar*
>>> :lib/log4j-core-2.12.1.jar:lib/
>>>
>>> export
>>> _FLINK_CLASSPATH=":lib/flink-csv-1.11.0.jar:lib/flink-json-1.11.0.jar:lib/flink-shaded-zookeeper-3.4.14.jar:lib/flink-table-blink_2.12-1.11.0.jar:lib/flink-table_2.12-1.11.0.jar:
>>> *lib/log4j-1.2-api-2.12.1.jar:lib/log4j-api-2.12.1.jar*
>>> :lib/log4j-core-2.12.1.jar:lib/log4j-slf4j-impl-2.12.1.jar:flink-dist_2.12-1.11.0.jar:flink-conf.yaml:"
>>>
>>> thanks.
>>>
>>> On Thu, Oct 29, 2020 at 6:21 PM Diwakar Jha <di...@gmail.com>
>>> wrote:
>>>
>>>> Hello Everyone,
>>>>
>>>> I'm able to get my Flink UI up and running (it was related to the
>>>> session manager plugin on my local laptop) but I'm not seeing any
>>>> taskmanager/jobmanager logs in my Flink application. I have attached some
>>>> yarn application logs while it's running but am not able to figure out how
>>>> to stop and get more logs. Could someone please help me figure this out?
>>>> I'm running Flink 1.11 on the EMR 6.1 cluster.
>>>>
>>>> On Tue, Oct 27, 2020 at 1:06 PM Diwakar Jha <di...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi Robert,
>>>>> Could please correct me. I'm not able to stop the app. Also, i
>>>>> stopped flink job already.
>>>>>
>>>>> sh-4.2$ yarn app -stop application_1603649952937_0002
>>>>> 2020-10-27 20:04:25,543 INFO client.RMProxy: Connecting to
>>>>> ResourceManager at ip-10-0-55-50.ec2.internal/10.0.55.50:8032
>>>>> 2020-10-27 20:04:25,717 INFO client.AHSProxy: Connecting to
>>>>> Application History server at ip-10-0-55-50.ec2.internal/
>>>>> 10.0.55.50:10200
>>>>> Exception in thread "main" java.lang.IllegalArgumentException: App
>>>>> admin client class name not specified for type Apache Flink
>>>>>         at
>>>>> org.apache.hadoop.yarn.client.api.AppAdminClient.createAppAdminClient(AppAdminClient.java:76)
>>>>>         at
>>>>> org.apache.hadoop.yarn.client.cli.ApplicationCLI.run(ApplicationCLI.java:597)
>>>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
>>>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
>>>>>         at
>>>>> org.apache.hadoop.yarn.client.cli.ApplicationCLI.main(ApplicationCLI.java:126)
>>>>> sh-4.2$
>>>>>
>>>>> On Tue, Oct 27, 2020 at 9:34 AM Robert Metzger <rm...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> Hi,
>>>>>> are you intentionally not posting this response to the mailing list?
>>>>>>
>>>>>> As you can see from the yarn logs, log aggregation only works for
>>>>>> finished applications ("End of LogType:prelaunch.out.This log file belongs
>>>>>> to a running container (container_1603649952937_0002_01_000002) and so may
>>>>>> not be complete.")
>>>>>>
>>>>>> Please stop the app, then provide the logs.
>>>>>>
>>>>>>
>>>>>> On Tue, Oct 27, 2020 at 5:11 PM Diwakar Jha <di...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Robert,
>>>>>>>
>>>>>>> Yes, i'm using Flink on EMR using YARN. Please find attached the
>>>>>>> yarn logs -applicationId. I also attached haddop-yarn-nodemanager logs.
>>>>>>> Also, I followed this link below which has the same problem :
>>>>>>> http://mail-archives.apache.org/mod_mbox/flink-user/202009.mbox/%3CCAGDv3o5WyJTrXs9Pg+Vy-b+LwgEE26iN54iqE0=f5T+M8VWm_w@mail.gmail.com%3E
>>>>>>>
>>>>>>> https://www.talkend.net/post/75078.html
>>>>>>> Based on this I changed the log4j.properties.
>>>>>>> Let me know what you think. Please also let me know if you need some
>>>>>>> specific logs.  Appreciate your help.
>>>>>>>
>>>>>>> Best,
>>>>>>> Diwakar
>>>>>>>
>>>>>>> On Tue, Oct 27, 2020 at 12:26 AM Robert Metzger <rm...@apache.org>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hey Diwakar,
>>>>>>>>
>>>>>>>> how are you deploying Flink on EMR? Are you using YARN?
>>>>>>>> If so, you could also use log aggregation to see all the logs at
>>>>>>>> once (from both JobManager and TaskManagers). (yarn logs -applicationId
>>>>>>>> <Application ID>)
>>>>>>>>
>>>>>>>> Could you post (or upload somewhere) all logs you have of one run?
>>>>>>>> It is much easier for us to debug something if we have the full logs (the
>>>>>>>> logs show for example the classpath that you are using, we would see how
>>>>>>>> you are deploying Flink, etc.)
>>>>>>>>
>>>>>>>> From the information available, my guess is that you have modified
>>>>>>>> your deployment in some way (use of a custom logging version, custom
>>>>>>>> deployment method, version mixup with jars from both Flink 1.8 and 1.11,
>>>>>>>> ...).
>>>>>>>>
>>>>>>>> Best,
>>>>>>>> Robert
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, Oct 27, 2020 at 12:41 AM Diwakar Jha <
>>>>>>>> diwakar.nith@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> This is what I see on the WebUI.
>>>>>>>>>
>>>>>>>>> 23:19:24.263 [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>>>>>> - Failed to transfer file from TaskExecutor
>>>>>>>>> container_1603649952937_0002_01_000004.
>>>>>>>>> java.util.concurrent.CompletionException:
>>>>>>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>>>>>>> TaskExecutor. at
>>>>>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>>>>> ThreadPoolExecutor.java:1149
>>>>>>>>> <http://threadpoolexecutor.java:1149/>) ~[?:1.8.0_252] at
>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>>>>>> ~[?:1.8.0_252] at java.lang.Thread.run
>>>>>>>>> <http://java.lang.thread.run/>(Thread.java:748
>>>>>>>>> <http://thread.java:748/>) ~[?:1.8.0_252] Caused by:
>>>>>>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>>>>>>> TaskExecutor. ... 5 more 23:19:24.275
>>>>>>>>> [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>>>>>> - Unhandled exception. org.apache.flink.util.FlinkException: The file LOG
>>>>>>>>> does not exist on the TaskExecutor. at
>>>>>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>>>>> ThreadPoolExecutor.java:1149
>>>>>>>>> <http://threadpoolexecutor.java:1149/>) ~[?:1.8.0_252] at
>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>>>>>> ~[?:1.8.0_252] at java.lang.Thread.run
>>>>>>>>> <http://java.lang.thread.run/>(Thread.java:748
>>>>>>>>> <http://thread.java:748/>) ~[?:1.8.0_252]
>>>>>>>>>
>>>>>>>>> Appreciate if anyone has any pointer for this.
>>>>>>>>>
>>>>>>>>> On Mon, Oct 26, 2020 at 10:45 AM Chesnay Schepler <
>>>>>>>>> chesnay@apache.org> wrote:
>>>>>>>>>
>>>>>>>>>> Flink 1.11 uses slf4j 1.7.15; the easiest way to check the log
>>>>>>>>>> files is usually via the WebUI.
>>>>>>>>>>
>>>>>>>>>> On 10/26/2020 5:30 PM, Diwakar Jha wrote:
>>>>>>>>>>
>>>>>>>>>> I think my problem is with Sl4j library. I'm using sl4j 1.7 with
>>>>>>>>>> Flink 1.11. If that's correct then i appreciate if someone can point me to
>>>>>>>>>> the exact Slf4j library that i should use with Flink 1.11
>>>>>>>>>>
>>>>>>>>>> Flink = 1.11.x;
>>>>>>>>>> Slf4j = 1.7;
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Sun, Oct 25, 2020 at 8:00 PM Diwakar Jha <
>>>>>>>>>> diwakar.nith@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Thanks for checking my configurations. Could you also point me
>>>>>>>>>>> where I can see the log files? Just to give more details. I'm trying to
>>>>>>>>>>> access these logs in AWS cloudwatch.
>>>>>>>>>>>
>>>>>>>>>>> Best,
>>>>>>>>>>> Diwakar
>>>>>>>>>>>
>>>>>>>>>>> On Sun, Oct 25, 2020 at 2:16 PM Chesnay Schepler <
>>>>>>>>>>> chesnay@apache.org> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> With Flink 1.11 reporters were refactored to plugins, and are
>>>>>>>>>>>> now accessible by default (so you no longer have to bother with copying
>>>>>>>>>>>> jars around).
>>>>>>>>>>>>
>>>>>>>>>>>> Your configuration appears to be correct, so I suggest to take
>>>>>>>>>>>> a look at the log files.
>>>>>>>>>>>>
>>>>>>>>>>>> On 10/25/2020 9:52 PM, Diwakar Jha wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Hello Everyone,
>>>>>>>>>>>>
>>>>>>>>>>>> I'm new to flink and i'm trying to upgrade from flink 1.8 to
>>>>>>>>>>>> flink 1.11 on an emr cluster. after upgrading to flink1.11 One of the
>>>>>>>>>>>> differences that i see is i don't get any metrics. I found out that flink
>>>>>>>>>>>> 1.11 does not have
>>>>>>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar in
>>>>>>>>>>>> /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have any
>>>>>>>>>>>> pointer to locate
>>>>>>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or
>>>>>>>>>>>> how to use metrics in flink.1.11?
>>>>>>>>>>>>
>>>>>>>>>>>> Things i tried :
>>>>>>>>>>>> a) the below setup
>>>>>>>>>>>>
>>>>>>>>>>>> metrics.reporters: stsdmetrics.reporter.stsd.factory.class: org.apache.flink.metrics.statsd.StatsDReporterFactorymetrics.reporter.stsd.host: localhostmetrics.reporter.stsd.port: 8125
>>>>>>>>>>>>
>>>>>>>>>>>> b) I tried downloading the statsd jar from
>>>>>>>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
>>>>>>>>>>>> putting it inside plugins/statsd directory.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> Best,
>>>>>>>>>>>> Diwakar Jha.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Best,
>>>>>>>>>>> Diwakar Jha.
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Best,
>>>>>>>>>> Diwakar Jha.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Best,
>>>>>>>>> Diwakar Jha.
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best,
>>>>>>> Diwakar Jha.
>>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> Best,
>>>>> Diwakar Jha.
>>>>>
>>>>

Re: how to enable metrics in Flink 1.11

Posted by Robert Metzger <rm...@apache.org>.
Hey Diwakar,

the logs you are providing still don't contain the full Flink logs.

You can not stop the Flink on YARN using "yarn app -stop
application_1603649952937_0002". To stop Flink on YARN, use: "yarn
application -kill <appId>".



On Sat, Oct 31, 2020 at 6:26 PM Diwakar Jha <di...@gmail.com> wrote:

> Hi,
>
> I wanted to check if anyone can help me with the logs. I have sent several
> emails but not getting any response.
>
> I'm running Flink 1.11 on EMR 6.1. I don't see any logs though I get this
> stdout error. I'm trying to upgrade Flink 1.8 to Flink 1.11
>
> 18:29:19.834 [flink-akka.actor.default-dispatcher-28] ERROR
> org.apache.flink.runtime.rest.handler.taskmanager.
> TaskManagerLogFileHandler - Failed to transfer file from TaskExecutor
> container_1604033334508_0001_01_000004.
> java.util.concurrent.CompletionException: org.apache.flink.util.
> FlinkException: The file LOG does not exist on the TaskExecutor.
>
> Thanks!
>
>
> On Fri, Oct 30, 2020 at 9:04 AM Diwakar Jha <di...@gmail.com>
> wrote:
>
>> Hello,
>>
>> I see that in my class path (below) I have both log4j-1 and lo4j-api-2.
>> is this because of which i'm not seeing any logs. If so, could someone
>> suggest how to fix it?
>>
>> export
>> CLASSPATH=":lib/flink-csv-1.11.0.jar:lib/flink-json-1.11.0.jar:lib/flink-shaded-zookeeper-3.4.14.jar:lib/flink-table-blink_2.12-1.11.0.jar:lib/flink-table_2.12-1.11.0.jar:
>> *lib/log4j-1.2-api-2.12.1.jar:lib/log4j-api-2.12.1.jar*
>> :lib/log4j-core-2.12.1.jar:lib/
>>
>> export
>> _FLINK_CLASSPATH=":lib/flink-csv-1.11.0.jar:lib/flink-json-1.11.0.jar:lib/flink-shaded-zookeeper-3.4.14.jar:lib/flink-table-blink_2.12-1.11.0.jar:lib/flink-table_2.12-1.11.0.jar:
>> *lib/log4j-1.2-api-2.12.1.jar:lib/log4j-api-2.12.1.jar*
>> :lib/log4j-core-2.12.1.jar:lib/log4j-slf4j-impl-2.12.1.jar:flink-dist_2.12-1.11.0.jar:flink-conf.yaml:"
>>
>> thanks.
>>
>> On Thu, Oct 29, 2020 at 6:21 PM Diwakar Jha <di...@gmail.com>
>> wrote:
>>
>>> Hello Everyone,
>>>
>>> I'm able to get my Flink UI up and running (it was related to the
>>> session manager plugin on my local laptop) but I'm not seeing any
>>> taskmanager/jobmanager logs in my Flink application. I have attached some
>>> yarn application logs while it's running but am not able to figure out how
>>> to stop and get more logs. Could someone please help me figure this out?
>>> I'm running Flink 1.11 on the EMR 6.1 cluster.
>>>
>>> On Tue, Oct 27, 2020 at 1:06 PM Diwakar Jha <di...@gmail.com>
>>> wrote:
>>>
>>>> Hi Robert,
>>>> Could please correct me. I'm not able to stop the app. Also, i
>>>> stopped flink job already.
>>>>
>>>> sh-4.2$ yarn app -stop application_1603649952937_0002
>>>> 2020-10-27 20:04:25,543 INFO client.RMProxy: Connecting to
>>>> ResourceManager at ip-10-0-55-50.ec2.internal/10.0.55.50:8032
>>>> 2020-10-27 20:04:25,717 INFO client.AHSProxy: Connecting to Application
>>>> History server at ip-10-0-55-50.ec2.internal/10.0.55.50:10200
>>>> Exception in thread "main" java.lang.IllegalArgumentException: App
>>>> admin client class name not specified for type Apache Flink
>>>>         at
>>>> org.apache.hadoop.yarn.client.api.AppAdminClient.createAppAdminClient(AppAdminClient.java:76)
>>>>         at
>>>> org.apache.hadoop.yarn.client.cli.ApplicationCLI.run(ApplicationCLI.java:597)
>>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
>>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
>>>>         at
>>>> org.apache.hadoop.yarn.client.cli.ApplicationCLI.main(ApplicationCLI.java:126)
>>>> sh-4.2$
>>>>
>>>> On Tue, Oct 27, 2020 at 9:34 AM Robert Metzger <rm...@apache.org>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>> are you intentionally not posting this response to the mailing list?
>>>>>
>>>>> As you can see from the yarn logs, log aggregation only works for
>>>>> finished applications ("End of LogType:prelaunch.out.This log file belongs
>>>>> to a running container (container_1603649952937_0002_01_000002) and so may
>>>>> not be complete.")
>>>>>
>>>>> Please stop the app, then provide the logs.
>>>>>
>>>>>
>>>>> On Tue, Oct 27, 2020 at 5:11 PM Diwakar Jha <di...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi Robert,
>>>>>>
>>>>>> Yes, i'm using Flink on EMR using YARN. Please find attached the yarn
>>>>>> logs -applicationId. I also attached haddop-yarn-nodemanager logs.
>>>>>> Also, I followed this link below which has the same problem :
>>>>>> http://mail-archives.apache.org/mod_mbox/flink-user/202009.mbox/%3CCAGDv3o5WyJTrXs9Pg+Vy-b+LwgEE26iN54iqE0=f5T+M8VWm_w@mail.gmail.com%3E
>>>>>>
>>>>>> https://www.talkend.net/post/75078.html
>>>>>> Based on this I changed the log4j.properties.
>>>>>> Let me know what you think. Please also let me know if you need some
>>>>>> specific logs.  Appreciate your help.
>>>>>>
>>>>>> Best,
>>>>>> Diwakar
>>>>>>
>>>>>> On Tue, Oct 27, 2020 at 12:26 AM Robert Metzger <rm...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> Hey Diwakar,
>>>>>>>
>>>>>>> how are you deploying Flink on EMR? Are you using YARN?
>>>>>>> If so, you could also use log aggregation to see all the logs at
>>>>>>> once (from both JobManager and TaskManagers). (yarn logs -applicationId
>>>>>>> <Application ID>)
>>>>>>>
>>>>>>> Could you post (or upload somewhere) all logs you have of one run?
>>>>>>> It is much easier for us to debug something if we have the full logs (the
>>>>>>> logs show for example the classpath that you are using, we would see how
>>>>>>> you are deploying Flink, etc.)
>>>>>>>
>>>>>>> From the information available, my guess is that you have modified
>>>>>>> your deployment in some way (use of a custom logging version, custom
>>>>>>> deployment method, version mixup with jars from both Flink 1.8 and 1.11,
>>>>>>> ...).
>>>>>>>
>>>>>>> Best,
>>>>>>> Robert
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Oct 27, 2020 at 12:41 AM Diwakar Jha <di...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> This is what I see on the WebUI.
>>>>>>>>
>>>>>>>> 23:19:24.263 [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>>>>> - Failed to transfer file from TaskExecutor
>>>>>>>> container_1603649952937_0002_01_000004.
>>>>>>>> java.util.concurrent.CompletionException:
>>>>>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>>>>>> TaskExecutor. at
>>>>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>>>>>> ~[?:1.8.0_252] at
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>>>>> ~[?:1.8.0_252] at java.lang.Thread.run
>>>>>>>> <http://java.lang.thread.run/>(Thread.java:748
>>>>>>>> <http://thread.java:748/>) ~[?:1.8.0_252] Caused by:
>>>>>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>>>>>> TaskExecutor. ... 5 more 23:19:24.275
>>>>>>>> [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>>>>> - Unhandled exception. org.apache.flink.util.FlinkException: The file LOG
>>>>>>>> does not exist on the TaskExecutor. at
>>>>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>>>>>> ~[?:1.8.0_252] at
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>>>>> ~[?:1.8.0_252] at java.lang.Thread.run
>>>>>>>> <http://java.lang.thread.run/>(Thread.java:748
>>>>>>>> <http://thread.java:748/>) ~[?:1.8.0_252]
>>>>>>>>
>>>>>>>> Appreciate if anyone has any pointer for this.
>>>>>>>>
>>>>>>>> On Mon, Oct 26, 2020 at 10:45 AM Chesnay Schepler <
>>>>>>>> chesnay@apache.org> wrote:
>>>>>>>>
>>>>>>>>> Flink 1.11 uses slf4j 1.7.15; the easiest way to check the log
>>>>>>>>> files is usually via the WebUI.
>>>>>>>>>
>>>>>>>>> On 10/26/2020 5:30 PM, Diwakar Jha wrote:
>>>>>>>>>
>>>>>>>>> I think my problem is with Sl4j library. I'm using sl4j 1.7 with
>>>>>>>>> Flink 1.11. If that's correct then i appreciate if someone can point me to
>>>>>>>>> the exact Slf4j library that i should use with Flink 1.11
>>>>>>>>>
>>>>>>>>> Flink = 1.11.x;
>>>>>>>>> Slf4j = 1.7;
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Sun, Oct 25, 2020 at 8:00 PM Diwakar Jha <
>>>>>>>>> diwakar.nith@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Thanks for checking my configurations. Could you also point me
>>>>>>>>>> where I can see the log files? Just to give more details. I'm trying to
>>>>>>>>>> access these logs in AWS cloudwatch.
>>>>>>>>>>
>>>>>>>>>> Best,
>>>>>>>>>> Diwakar
>>>>>>>>>>
>>>>>>>>>> On Sun, Oct 25, 2020 at 2:16 PM Chesnay Schepler <
>>>>>>>>>> chesnay@apache.org> wrote:
>>>>>>>>>>
>>>>>>>>>>> With Flink 1.11 reporters were refactored to plugins, and are
>>>>>>>>>>> now accessible by default (so you no longer have to bother with copying
>>>>>>>>>>> jars around).
>>>>>>>>>>>
>>>>>>>>>>> Your configuration appears to be correct, so I suggest to take a
>>>>>>>>>>> look at the log files.
>>>>>>>>>>>
>>>>>>>>>>> On 10/25/2020 9:52 PM, Diwakar Jha wrote:
>>>>>>>>>>>
>>>>>>>>>>> Hello Everyone,
>>>>>>>>>>>
>>>>>>>>>>> I'm new to flink and i'm trying to upgrade from flink 1.8 to
>>>>>>>>>>> flink 1.11 on an emr cluster. after upgrading to flink1.11 One of the
>>>>>>>>>>> differences that i see is i don't get any metrics. I found out that flink
>>>>>>>>>>> 1.11 does not have
>>>>>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar in
>>>>>>>>>>> /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have any
>>>>>>>>>>> pointer to locate
>>>>>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or
>>>>>>>>>>> how to use metrics in flink.1.11?
>>>>>>>>>>>
>>>>>>>>>>> Things i tried :
>>>>>>>>>>> a) the below setup
>>>>>>>>>>>
>>>>>>>>>>> metrics.reporters: stsdmetrics.reporter.stsd.factory.class: org.apache.flink.metrics.statsd.StatsDReporterFactorymetrics.reporter.stsd.host: localhostmetrics.reporter.stsd.port: 8125
>>>>>>>>>>>
>>>>>>>>>>> b) I tried downloading the statsd jar from
>>>>>>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
>>>>>>>>>>> putting it inside plugins/statsd directory.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Best,
>>>>>>>>>>> Diwakar Jha.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Best,
>>>>>>>>>> Diwakar Jha.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Best,
>>>>>>>>> Diwakar Jha.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best,
>>>>>>>> Diwakar Jha.
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best,
>>>>>> Diwakar Jha.
>>>>>>
>>>>>
>>>>
>>>> --
>>>> Best,
>>>> Diwakar Jha.
>>>>
>>>

Re: how to enable metrics in Flink 1.11

Posted by Diwakar Jha <di...@gmail.com>.
Hi,

I wanted to check if anyone can help me with the logs. I have sent several
emails but not getting any response.

I'm running Flink 1.11 on EMR 6.1. I don't see any logs though I get this
stdout error. I'm trying to upgrade Flink 1.8 to Flink 1.11

18:29:19.834 [flink-akka.actor.default-dispatcher-28] ERROR
org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
- Failed to transfer file from TaskExecutor
container_1604033334508_0001_01_000004.
java.util.concurrent.CompletionException: org.apache.flink.util.
FlinkException: The file LOG does not exist on the TaskExecutor.

Thanks!


On Fri, Oct 30, 2020 at 9:04 AM Diwakar Jha <di...@gmail.com> wrote:

> Hello,
>
> I see that in my class path (below) I have both log4j-1 and lo4j-api-2. is
> this because of which i'm not seeing any logs. If so, could someone suggest
> how to fix it?
>
> export
> CLASSPATH=":lib/flink-csv-1.11.0.jar:lib/flink-json-1.11.0.jar:lib/flink-shaded-zookeeper-3.4.14.jar:lib/flink-table-blink_2.12-1.11.0.jar:lib/flink-table_2.12-1.11.0.jar:
> *lib/log4j-1.2-api-2.12.1.jar:lib/log4j-api-2.12.1.jar*
> :lib/log4j-core-2.12.1.jar:lib/
>
> export
> _FLINK_CLASSPATH=":lib/flink-csv-1.11.0.jar:lib/flink-json-1.11.0.jar:lib/flink-shaded-zookeeper-3.4.14.jar:lib/flink-table-blink_2.12-1.11.0.jar:lib/flink-table_2.12-1.11.0.jar:
> *lib/log4j-1.2-api-2.12.1.jar:lib/log4j-api-2.12.1.jar*
> :lib/log4j-core-2.12.1.jar:lib/log4j-slf4j-impl-2.12.1.jar:flink-dist_2.12-1.11.0.jar:flink-conf.yaml:"
>
> thanks.
>
> On Thu, Oct 29, 2020 at 6:21 PM Diwakar Jha <di...@gmail.com>
> wrote:
>
>> Hello Everyone,
>>
>> I'm able to get my Flink UI up and running (it was related to the session
>> manager plugin on my local laptop) but I'm not seeing any
>> taskmanager/jobmanager logs in my Flink application. I have attached some
>> yarn application logs while it's running but am not able to figure out how
>> to stop and get more logs. Could someone please help me figure this out?
>> I'm running Flink 1.11 on the EMR 6.1 cluster.
>>
>> On Tue, Oct 27, 2020 at 1:06 PM Diwakar Jha <di...@gmail.com>
>> wrote:
>>
>>> Hi Robert,
>>> Could please correct me. I'm not able to stop the app. Also, i
>>> stopped flink job already.
>>>
>>> sh-4.2$ yarn app -stop application_1603649952937_0002
>>> 2020-10-27 20:04:25,543 INFO client.RMProxy: Connecting to
>>> ResourceManager at ip-10-0-55-50.ec2.internal/10.0.55.50:8032
>>> 2020-10-27 20:04:25,717 INFO client.AHSProxy: Connecting to Application
>>> History server at ip-10-0-55-50.ec2.internal/10.0.55.50:10200
>>> Exception in thread "main" java.lang.IllegalArgumentException: App admin
>>> client class name not specified for type Apache Flink
>>>         at
>>> org.apache.hadoop.yarn.client.api.AppAdminClient.createAppAdminClient(AppAdminClient.java:76)
>>>         at
>>> org.apache.hadoop.yarn.client.cli.ApplicationCLI.run(ApplicationCLI.java:597)
>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
>>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
>>>         at
>>> org.apache.hadoop.yarn.client.cli.ApplicationCLI.main(ApplicationCLI.java:126)
>>> sh-4.2$
>>>
>>> On Tue, Oct 27, 2020 at 9:34 AM Robert Metzger <rm...@apache.org>
>>> wrote:
>>>
>>>> Hi,
>>>> are you intentionally not posting this response to the mailing list?
>>>>
>>>> As you can see from the yarn logs, log aggregation only works for
>>>> finished applications ("End of LogType:prelaunch.out.This log file belongs
>>>> to a running container (container_1603649952937_0002_01_000002) and so may
>>>> not be complete.")
>>>>
>>>> Please stop the app, then provide the logs.
>>>>
>>>>
>>>> On Tue, Oct 27, 2020 at 5:11 PM Diwakar Jha <di...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi Robert,
>>>>>
>>>>> Yes, i'm using Flink on EMR using YARN. Please find attached the yarn
>>>>> logs -applicationId. I also attached haddop-yarn-nodemanager logs.
>>>>> Also, I followed this link below which has the same problem :
>>>>> http://mail-archives.apache.org/mod_mbox/flink-user/202009.mbox/%3CCAGDv3o5WyJTrXs9Pg+Vy-b+LwgEE26iN54iqE0=f5T+M8VWm_w@mail.gmail.com%3E
>>>>>
>>>>> https://www.talkend.net/post/75078.html
>>>>> Based on this I changed the log4j.properties.
>>>>> Let me know what you think. Please also let me know if you need some
>>>>> specific logs.  Appreciate your help.
>>>>>
>>>>> Best,
>>>>> Diwakar
>>>>>
>>>>> On Tue, Oct 27, 2020 at 12:26 AM Robert Metzger <rm...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> Hey Diwakar,
>>>>>>
>>>>>> how are you deploying Flink on EMR? Are you using YARN?
>>>>>> If so, you could also use log aggregation to see all the logs at once
>>>>>> (from both JobManager and TaskManagers). (yarn logs -applicationId
>>>>>> <Application ID>)
>>>>>>
>>>>>> Could you post (or upload somewhere) all logs you have of one run? It
>>>>>> is much easier for us to debug something if we have the full logs (the logs
>>>>>> show for example the classpath that you are using, we would see how you are
>>>>>> deploying Flink, etc.)
>>>>>>
>>>>>> From the information available, my guess is that you have modified
>>>>>> your deployment in some way (use of a custom logging version, custom
>>>>>> deployment method, version mixup with jars from both Flink 1.8 and 1.11,
>>>>>> ...).
>>>>>>
>>>>>> Best,
>>>>>> Robert
>>>>>>
>>>>>>
>>>>>> On Tue, Oct 27, 2020 at 12:41 AM Diwakar Jha <di...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> This is what I see on the WebUI.
>>>>>>>
>>>>>>> 23:19:24.263 [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>>>> - Failed to transfer file from TaskExecutor
>>>>>>> container_1603649952937_0002_01_000004.
>>>>>>> java.util.concurrent.CompletionException:
>>>>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>>>>> TaskExecutor. at
>>>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>>>> ~[?:1.8.0_252] at java.lang.Thread.run
>>>>>>> <http://java.lang.thread.run/>(Thread.java:748
>>>>>>> <http://thread.java:748/>) ~[?:1.8.0_252] Caused by:
>>>>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>>>>> TaskExecutor. ... 5 more 23:19:24.275
>>>>>>> [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>>>> - Unhandled exception. org.apache.flink.util.FlinkException: The file LOG
>>>>>>> does not exist on the TaskExecutor. at
>>>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>>>> ~[?:1.8.0_252] at java.lang.Thread.run
>>>>>>> <http://java.lang.thread.run/>(Thread.java:748
>>>>>>> <http://thread.java:748/>) ~[?:1.8.0_252]
>>>>>>>
>>>>>>> Appreciate if anyone has any pointer for this.
>>>>>>>
>>>>>>> On Mon, Oct 26, 2020 at 10:45 AM Chesnay Schepler <
>>>>>>> chesnay@apache.org> wrote:
>>>>>>>
>>>>>>>> Flink 1.11 uses slf4j 1.7.15; the easiest way to check the log
>>>>>>>> files is usually via the WebUI.
>>>>>>>>
>>>>>>>> On 10/26/2020 5:30 PM, Diwakar Jha wrote:
>>>>>>>>
>>>>>>>> I think my problem is with Sl4j library. I'm using sl4j 1.7 with
>>>>>>>> Flink 1.11. If that's correct then i appreciate if someone can point me to
>>>>>>>> the exact Slf4j library that i should use with Flink 1.11
>>>>>>>>
>>>>>>>> Flink = 1.11.x;
>>>>>>>> Slf4j = 1.7;
>>>>>>>>
>>>>>>>>
>>>>>>>> On Sun, Oct 25, 2020 at 8:00 PM Diwakar Jha <di...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Thanks for checking my configurations. Could you also point me
>>>>>>>>> where I can see the log files? Just to give more details. I'm trying to
>>>>>>>>> access these logs in AWS cloudwatch.
>>>>>>>>>
>>>>>>>>> Best,
>>>>>>>>> Diwakar
>>>>>>>>>
>>>>>>>>> On Sun, Oct 25, 2020 at 2:16 PM Chesnay Schepler <
>>>>>>>>> chesnay@apache.org> wrote:
>>>>>>>>>
>>>>>>>>>> With Flink 1.11 reporters were refactored to plugins, and are now
>>>>>>>>>> accessible by default (so you no longer have to bother with copying jars
>>>>>>>>>> around).
>>>>>>>>>>
>>>>>>>>>> Your configuration appears to be correct, so I suggest to take a
>>>>>>>>>> look at the log files.
>>>>>>>>>>
>>>>>>>>>> On 10/25/2020 9:52 PM, Diwakar Jha wrote:
>>>>>>>>>>
>>>>>>>>>> Hello Everyone,
>>>>>>>>>>
>>>>>>>>>> I'm new to flink and i'm trying to upgrade from flink 1.8 to
>>>>>>>>>> flink 1.11 on an emr cluster. after upgrading to flink1.11 One of the
>>>>>>>>>> differences that i see is i don't get any metrics. I found out that flink
>>>>>>>>>> 1.11 does not have
>>>>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar in
>>>>>>>>>> /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have any
>>>>>>>>>> pointer to locate
>>>>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or
>>>>>>>>>> how to use metrics in flink.1.11?
>>>>>>>>>>
>>>>>>>>>> Things i tried :
>>>>>>>>>> a) the below setup
>>>>>>>>>>
>>>>>>>>>> metrics.reporters: stsdmetrics.reporter.stsd.factory.class: org.apache.flink.metrics.statsd.StatsDReporterFactorymetrics.reporter.stsd.host: localhostmetrics.reporter.stsd.port: 8125
>>>>>>>>>>
>>>>>>>>>> b) I tried downloading the statsd jar from
>>>>>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
>>>>>>>>>> putting it inside plugins/statsd directory.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Best,
>>>>>>>>>> Diwakar Jha.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Best,
>>>>>>>>> Diwakar Jha.
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best,
>>>>>>>> Diwakar Jha.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best,
>>>>>>> Diwakar Jha.
>>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> Best,
>>>>> Diwakar Jha.
>>>>>
>>>>
>>>
>>> --
>>> Best,
>>> Diwakar Jha.
>>>
>>

Re: how to enable metrics in Flink 1.11

Posted by Diwakar Jha <di...@gmail.com>.
Hello,

I see that in my class path (below) I have both log4j-1 and lo4j-api-2. is
this because of which i'm not seeing any logs. If so, could someone suggest
how to fix it?

export
CLASSPATH=":lib/flink-csv-1.11.0.jar:lib/flink-json-1.11.0.jar:lib/flink-shaded-zookeeper-3.4.14.jar:lib/flink-table-blink_2.12-1.11.0.jar:lib/flink-table_2.12-1.11.0.jar:
*lib/log4j-1.2-api-2.12.1.jar:lib/log4j-api-2.12.1.jar*
:lib/log4j-core-2.12.1.jar:lib/

export
_FLINK_CLASSPATH=":lib/flink-csv-1.11.0.jar:lib/flink-json-1.11.0.jar:lib/flink-shaded-zookeeper-3.4.14.jar:lib/flink-table-blink_2.12-1.11.0.jar:lib/flink-table_2.12-1.11.0.jar:
*lib/log4j-1.2-api-2.12.1.jar:lib/log4j-api-2.12.1.jar*
:lib/log4j-core-2.12.1.jar:lib/log4j-slf4j-impl-2.12.1.jar:flink-dist_2.12-1.11.0.jar:flink-conf.yaml:"

thanks.

On Thu, Oct 29, 2020 at 6:21 PM Diwakar Jha <di...@gmail.com> wrote:

> Hello Everyone,
>
> I'm able to get my Flink UI up and running (it was related to the session
> manager plugin on my local laptop) but I'm not seeing any
> taskmanager/jobmanager logs in my Flink application. I have attached some
> yarn application logs while it's running but am not able to figure out how
> to stop and get more logs. Could someone please help me figure this out?
> I'm running Flink 1.11 on the EMR 6.1 cluster.
>
> On Tue, Oct 27, 2020 at 1:06 PM Diwakar Jha <di...@gmail.com>
> wrote:
>
>> Hi Robert,
>> Could please correct me. I'm not able to stop the app. Also, i
>> stopped flink job already.
>>
>> sh-4.2$ yarn app -stop application_1603649952937_0002
>> 2020-10-27 20:04:25,543 INFO client.RMProxy: Connecting to
>> ResourceManager at ip-10-0-55-50.ec2.internal/10.0.55.50:8032
>> 2020-10-27 20:04:25,717 INFO client.AHSProxy: Connecting to Application
>> History server at ip-10-0-55-50.ec2.internal/10.0.55.50:10200
>> Exception in thread "main" java.lang.IllegalArgumentException: App admin
>> client class name not specified for type Apache Flink
>>         at
>> org.apache.hadoop.yarn.client.api.AppAdminClient.createAppAdminClient(AppAdminClient.java:76)
>>         at
>> org.apache.hadoop.yarn.client.cli.ApplicationCLI.run(ApplicationCLI.java:597)
>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
>>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
>>         at
>> org.apache.hadoop.yarn.client.cli.ApplicationCLI.main(ApplicationCLI.java:126)
>> sh-4.2$
>>
>> On Tue, Oct 27, 2020 at 9:34 AM Robert Metzger <rm...@apache.org>
>> wrote:
>>
>>> Hi,
>>> are you intentionally not posting this response to the mailing list?
>>>
>>> As you can see from the yarn logs, log aggregation only works for
>>> finished applications ("End of LogType:prelaunch.out.This log file belongs
>>> to a running container (container_1603649952937_0002_01_000002) and so may
>>> not be complete.")
>>>
>>> Please stop the app, then provide the logs.
>>>
>>>
>>> On Tue, Oct 27, 2020 at 5:11 PM Diwakar Jha <di...@gmail.com>
>>> wrote:
>>>
>>>> Hi Robert,
>>>>
>>>> Yes, i'm using Flink on EMR using YARN. Please find attached the yarn
>>>> logs -applicationId. I also attached haddop-yarn-nodemanager logs.
>>>> Also, I followed this link below which has the same problem :
>>>> http://mail-archives.apache.org/mod_mbox/flink-user/202009.mbox/%3CCAGDv3o5WyJTrXs9Pg+Vy-b+LwgEE26iN54iqE0=f5T+M8VWm_w@mail.gmail.com%3E
>>>>
>>>> https://www.talkend.net/post/75078.html
>>>> Based on this I changed the log4j.properties.
>>>> Let me know what you think. Please also let me know if you need some
>>>> specific logs.  Appreciate your help.
>>>>
>>>> Best,
>>>> Diwakar
>>>>
>>>> On Tue, Oct 27, 2020 at 12:26 AM Robert Metzger <rm...@apache.org>
>>>> wrote:
>>>>
>>>>> Hey Diwakar,
>>>>>
>>>>> how are you deploying Flink on EMR? Are you using YARN?
>>>>> If so, you could also use log aggregation to see all the logs at once
>>>>> (from both JobManager and TaskManagers). (yarn logs -applicationId
>>>>> <Application ID>)
>>>>>
>>>>> Could you post (or upload somewhere) all logs you have of one run? It
>>>>> is much easier for us to debug something if we have the full logs (the logs
>>>>> show for example the classpath that you are using, we would see how you are
>>>>> deploying Flink, etc.)
>>>>>
>>>>> From the information available, my guess is that you have modified
>>>>> your deployment in some way (use of a custom logging version, custom
>>>>> deployment method, version mixup with jars from both Flink 1.8 and 1.11,
>>>>> ...).
>>>>>
>>>>> Best,
>>>>> Robert
>>>>>
>>>>>
>>>>> On Tue, Oct 27, 2020 at 12:41 AM Diwakar Jha <di...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> This is what I see on the WebUI.
>>>>>>
>>>>>> 23:19:24.263 [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>>> - Failed to transfer file from TaskExecutor
>>>>>> container_1603649952937_0002_01_000004.
>>>>>> java.util.concurrent.CompletionException:
>>>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>>>> TaskExecutor. at
>>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>>> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>
>>>>>> (Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252] Caused
>>>>>> by: org.apache.flink.util.FlinkException: The file LOG does not exist on
>>>>>> the TaskExecutor. ... 5 more 23:19:24.275
>>>>>> [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>>> - Unhandled exception. org.apache.flink.util.FlinkException: The file LOG
>>>>>> does not exist on the TaskExecutor. at
>>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>>> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>
>>>>>> (Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252]
>>>>>>
>>>>>> Appreciate if anyone has any pointer for this.
>>>>>>
>>>>>> On Mon, Oct 26, 2020 at 10:45 AM Chesnay Schepler <ch...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> Flink 1.11 uses slf4j 1.7.15; the easiest way to check the log files
>>>>>>> is usually via the WebUI.
>>>>>>>
>>>>>>> On 10/26/2020 5:30 PM, Diwakar Jha wrote:
>>>>>>>
>>>>>>> I think my problem is with Sl4j library. I'm using sl4j 1.7 with
>>>>>>> Flink 1.11. If that's correct then i appreciate if someone can point me to
>>>>>>> the exact Slf4j library that i should use with Flink 1.11
>>>>>>>
>>>>>>> Flink = 1.11.x;
>>>>>>> Slf4j = 1.7;
>>>>>>>
>>>>>>>
>>>>>>> On Sun, Oct 25, 2020 at 8:00 PM Diwakar Jha <di...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Thanks for checking my configurations. Could you also point me
>>>>>>>> where I can see the log files? Just to give more details. I'm trying to
>>>>>>>> access these logs in AWS cloudwatch.
>>>>>>>>
>>>>>>>> Best,
>>>>>>>> Diwakar
>>>>>>>>
>>>>>>>> On Sun, Oct 25, 2020 at 2:16 PM Chesnay Schepler <
>>>>>>>> chesnay@apache.org> wrote:
>>>>>>>>
>>>>>>>>> With Flink 1.11 reporters were refactored to plugins, and are now
>>>>>>>>> accessible by default (so you no longer have to bother with copying jars
>>>>>>>>> around).
>>>>>>>>>
>>>>>>>>> Your configuration appears to be correct, so I suggest to take a
>>>>>>>>> look at the log files.
>>>>>>>>>
>>>>>>>>> On 10/25/2020 9:52 PM, Diwakar Jha wrote:
>>>>>>>>>
>>>>>>>>> Hello Everyone,
>>>>>>>>>
>>>>>>>>> I'm new to flink and i'm trying to upgrade from flink 1.8 to flink
>>>>>>>>> 1.11 on an emr cluster. after upgrading to flink1.11 One of the differences
>>>>>>>>> that i see is i don't get any metrics. I found out that flink 1.11 does not
>>>>>>>>> have *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar
>>>>>>>>> in /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have
>>>>>>>>> any pointer to locate
>>>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or
>>>>>>>>> how to use metrics in flink.1.11?
>>>>>>>>>
>>>>>>>>> Things i tried :
>>>>>>>>> a) the below setup
>>>>>>>>>
>>>>>>>>> metrics.reporters: stsdmetrics.reporter.stsd.factory.class: org.apache.flink.metrics.statsd.StatsDReporterFactorymetrics.reporter.stsd.host: localhostmetrics.reporter.stsd.port: 8125
>>>>>>>>>
>>>>>>>>> b) I tried downloading the statsd jar from
>>>>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
>>>>>>>>> putting it inside plugins/statsd directory.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Best,
>>>>>>>>> Diwakar Jha.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best,
>>>>>>>> Diwakar Jha.
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best,
>>>>>>> Diwakar Jha.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best,
>>>>>> Diwakar Jha.
>>>>>>
>>>>>
>>>>
>>>> --
>>>> Best,
>>>> Diwakar Jha.
>>>>
>>>
>>
>> --
>> Best,
>> Diwakar Jha.
>>
>

Re: how to enable metrics in Flink 1.11

Posted by Diwakar Jha <di...@gmail.com>.
Hello Everyone,

I'm able to get my Flink UI up and running (it was related to the session
manager plugin on my local laptop) but I'm not seeing any
taskmanager/jobmanager logs in my Flink application. I have attached some
yarn application logs while it's running but am not able to figure out how
to stop and get more logs. Could someone please help me figure this out?
I'm running Flink 1.11 on the EMR 6.1 cluster.

On Tue, Oct 27, 2020 at 1:06 PM Diwakar Jha <di...@gmail.com> wrote:

> Hi Robert,
> Could please correct me. I'm not able to stop the app. Also, i
> stopped flink job already.
>
> sh-4.2$ yarn app -stop application_1603649952937_0002
> 2020-10-27 20:04:25,543 INFO client.RMProxy: Connecting to ResourceManager
> at ip-10-0-55-50.ec2.internal/10.0.55.50:8032
> 2020-10-27 20:04:25,717 INFO client.AHSProxy: Connecting to Application
> History server at ip-10-0-55-50.ec2.internal/10.0.55.50:10200
> Exception in thread "main" java.lang.IllegalArgumentException: App admin
> client class name not specified for type Apache Flink
>         at
> org.apache.hadoop.yarn.client.api.AppAdminClient.createAppAdminClient(AppAdminClient.java:76)
>         at
> org.apache.hadoop.yarn.client.cli.ApplicationCLI.run(ApplicationCLI.java:597)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
>         at
> org.apache.hadoop.yarn.client.cli.ApplicationCLI.main(ApplicationCLI.java:126)
> sh-4.2$
>
> On Tue, Oct 27, 2020 at 9:34 AM Robert Metzger <rm...@apache.org>
> wrote:
>
>> Hi,
>> are you intentionally not posting this response to the mailing list?
>>
>> As you can see from the yarn logs, log aggregation only works for
>> finished applications ("End of LogType:prelaunch.out.This log file belongs
>> to a running container (container_1603649952937_0002_01_000002) and so may
>> not be complete.")
>>
>> Please stop the app, then provide the logs.
>>
>>
>> On Tue, Oct 27, 2020 at 5:11 PM Diwakar Jha <di...@gmail.com>
>> wrote:
>>
>>> Hi Robert,
>>>
>>> Yes, i'm using Flink on EMR using YARN. Please find attached the yarn
>>> logs -applicationId. I also attached haddop-yarn-nodemanager logs.
>>> Also, I followed this link below which has the same problem :
>>> http://mail-archives.apache.org/mod_mbox/flink-user/202009.mbox/%3CCAGDv3o5WyJTrXs9Pg+Vy-b+LwgEE26iN54iqE0=f5T+M8VWm_w@mail.gmail.com%3E
>>>
>>> https://www.talkend.net/post/75078.html
>>> Based on this I changed the log4j.properties.
>>> Let me know what you think. Please also let me know if you need some
>>> specific logs.  Appreciate your help.
>>>
>>> Best,
>>> Diwakar
>>>
>>> On Tue, Oct 27, 2020 at 12:26 AM Robert Metzger <rm...@apache.org>
>>> wrote:
>>>
>>>> Hey Diwakar,
>>>>
>>>> how are you deploying Flink on EMR? Are you using YARN?
>>>> If so, you could also use log aggregation to see all the logs at once
>>>> (from both JobManager and TaskManagers). (yarn logs -applicationId
>>>> <Application ID>)
>>>>
>>>> Could you post (or upload somewhere) all logs you have of one run? It
>>>> is much easier for us to debug something if we have the full logs (the logs
>>>> show for example the classpath that you are using, we would see how you are
>>>> deploying Flink, etc.)
>>>>
>>>> From the information available, my guess is that you have modified your
>>>> deployment in some way (use of a custom logging version, custom deployment
>>>> method, version mixup with jars from both Flink 1.8 and 1.11, ...).
>>>>
>>>> Best,
>>>> Robert
>>>>
>>>>
>>>> On Tue, Oct 27, 2020 at 12:41 AM Diwakar Jha <di...@gmail.com>
>>>> wrote:
>>>>
>>>>> This is what I see on the WebUI.
>>>>>
>>>>> 23:19:24.263 [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>> - Failed to transfer file from TaskExecutor
>>>>> container_1603649952937_0002_01_000004.
>>>>> java.util.concurrent.CompletionException:
>>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>>> TaskExecutor. at
>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
>>>>> Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252] Caused by:
>>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>>> TaskExecutor. ... 5 more 23:19:24.275
>>>>> [flink-akka.actor.default-dispatcher-1865] ERROR
>>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>>> - Unhandled exception. org.apache.flink.util.FlinkException: The file LOG
>>>>> does not exist on the TaskExecutor. at
>>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>>> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
>>>>> Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252]
>>>>>
>>>>> Appreciate if anyone has any pointer for this.
>>>>>
>>>>> On Mon, Oct 26, 2020 at 10:45 AM Chesnay Schepler <ch...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> Flink 1.11 uses slf4j 1.7.15; the easiest way to check the log files
>>>>>> is usually via the WebUI.
>>>>>>
>>>>>> On 10/26/2020 5:30 PM, Diwakar Jha wrote:
>>>>>>
>>>>>> I think my problem is with Sl4j library. I'm using sl4j 1.7 with
>>>>>> Flink 1.11. If that's correct then i appreciate if someone can point me to
>>>>>> the exact Slf4j library that i should use with Flink 1.11
>>>>>>
>>>>>> Flink = 1.11.x;
>>>>>> Slf4j = 1.7;
>>>>>>
>>>>>>
>>>>>> On Sun, Oct 25, 2020 at 8:00 PM Diwakar Jha <di...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Thanks for checking my configurations. Could you also point me where
>>>>>>> I can see the log files? Just to give more details. I'm trying to access
>>>>>>> these logs in AWS cloudwatch.
>>>>>>>
>>>>>>> Best,
>>>>>>> Diwakar
>>>>>>>
>>>>>>> On Sun, Oct 25, 2020 at 2:16 PM Chesnay Schepler <ch...@apache.org>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> With Flink 1.11 reporters were refactored to plugins, and are now
>>>>>>>> accessible by default (so you no longer have to bother with copying jars
>>>>>>>> around).
>>>>>>>>
>>>>>>>> Your configuration appears to be correct, so I suggest to take a
>>>>>>>> look at the log files.
>>>>>>>>
>>>>>>>> On 10/25/2020 9:52 PM, Diwakar Jha wrote:
>>>>>>>>
>>>>>>>> Hello Everyone,
>>>>>>>>
>>>>>>>> I'm new to flink and i'm trying to upgrade from flink 1.8 to flink
>>>>>>>> 1.11 on an emr cluster. after upgrading to flink1.11 One of the differences
>>>>>>>> that i see is i don't get any metrics. I found out that flink 1.11 does not
>>>>>>>> have *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar
>>>>>>>> in /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have
>>>>>>>> any pointer to locate
>>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or how
>>>>>>>> to use metrics in flink.1.11?
>>>>>>>>
>>>>>>>> Things i tried :
>>>>>>>> a) the below setup
>>>>>>>>
>>>>>>>> metrics.reporters: stsdmetrics.reporter.stsd.factory.class: org.apache.flink.metrics.statsd.StatsDReporterFactorymetrics.reporter.stsd.host: localhostmetrics.reporter.stsd.port: 8125
>>>>>>>>
>>>>>>>> b) I tried downloading the statsd jar from
>>>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
>>>>>>>> putting it inside plugins/statsd directory.
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best,
>>>>>>>> Diwakar Jha.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best,
>>>>>>> Diwakar Jha.
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best,
>>>>>> Diwakar Jha.
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> Best,
>>>>> Diwakar Jha.
>>>>>
>>>>
>>>
>>> --
>>> Best,
>>> Diwakar Jha.
>>>
>>
>
> --
> Best,
> Diwakar Jha.
>

Re: how to enable metrics in Flink 1.11

Posted by Diwakar Jha <di...@gmail.com>.
Hi Robert,
Could please correct me. I'm not able to stop the app. Also, i
stopped flink job already.

sh-4.2$ yarn app -stop application_1603649952937_0002
2020-10-27 20:04:25,543 INFO client.RMProxy: Connecting to ResourceManager
at ip-10-0-55-50.ec2.internal/10.0.55.50:8032
2020-10-27 20:04:25,717 INFO client.AHSProxy: Connecting to Application
History server at ip-10-0-55-50.ec2.internal/10.0.55.50:10200
Exception in thread "main" java.lang.IllegalArgumentException: App admin
client class name not specified for type Apache Flink
        at
org.apache.hadoop.yarn.client.api.AppAdminClient.createAppAdminClient(AppAdminClient.java:76)
        at
org.apache.hadoop.yarn.client.cli.ApplicationCLI.run(ApplicationCLI.java:597)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:90)
        at
org.apache.hadoop.yarn.client.cli.ApplicationCLI.main(ApplicationCLI.java:126)
sh-4.2$

On Tue, Oct 27, 2020 at 9:34 AM Robert Metzger <rm...@apache.org> wrote:

> Hi,
> are you intentionally not posting this response to the mailing list?
>
> As you can see from the yarn logs, log aggregation only works for finished
> applications ("End of LogType:prelaunch.out.This log file belongs to a
> running container (container_1603649952937_0002_01_000002) and so may not
> be complete.")
>
> Please stop the app, then provide the logs.
>
>
> On Tue, Oct 27, 2020 at 5:11 PM Diwakar Jha <di...@gmail.com>
> wrote:
>
>> Hi Robert,
>>
>> Yes, i'm using Flink on EMR using YARN. Please find attached the yarn
>> logs -applicationId. I also attached haddop-yarn-nodemanager logs.
>> Also, I followed this link below which has the same problem :
>> http://mail-archives.apache.org/mod_mbox/flink-user/202009.mbox/%3CCAGDv3o5WyJTrXs9Pg+Vy-b+LwgEE26iN54iqE0=f5T+M8VWm_w@mail.gmail.com%3E
>>
>> https://www.talkend.net/post/75078.html
>> Based on this I changed the log4j.properties.
>> Let me know what you think. Please also let me know if you need some
>> specific logs.  Appreciate your help.
>>
>> Best,
>> Diwakar
>>
>> On Tue, Oct 27, 2020 at 12:26 AM Robert Metzger <rm...@apache.org>
>> wrote:
>>
>>> Hey Diwakar,
>>>
>>> how are you deploying Flink on EMR? Are you using YARN?
>>> If so, you could also use log aggregation to see all the logs at once
>>> (from both JobManager and TaskManagers). (yarn logs -applicationId
>>> <Application ID>)
>>>
>>> Could you post (or upload somewhere) all logs you have of one run? It is
>>> much easier for us to debug something if we have the full logs (the logs
>>> show for example the classpath that you are using, we would see how you are
>>> deploying Flink, etc.)
>>>
>>> From the information available, my guess is that you have modified your
>>> deployment in some way (use of a custom logging version, custom deployment
>>> method, version mixup with jars from both Flink 1.8 and 1.11, ...).
>>>
>>> Best,
>>> Robert
>>>
>>>
>>> On Tue, Oct 27, 2020 at 12:41 AM Diwakar Jha <di...@gmail.com>
>>> wrote:
>>>
>>>> This is what I see on the WebUI.
>>>>
>>>> 23:19:24.263 [flink-akka.actor.default-dispatcher-1865] ERROR
>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>> - Failed to transfer file from TaskExecutor
>>>> container_1603649952937_0002_01_000004.
>>>> java.util.concurrent.CompletionException:
>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>> TaskExecutor. at
>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
>>>> Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252] Caused by:
>>>> org.apache.flink.util.FlinkException: The file LOG does not exist on the
>>>> TaskExecutor. ... 5 more 23:19:24.275
>>>> [flink-akka.actor.default-dispatcher-1865] ERROR
>>>> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
>>>> - Unhandled exception. org.apache.flink.util.FlinkException: The file LOG
>>>> does not exist on the TaskExecutor. at
>>>> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
>>>> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
>>>> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
>>>> java.util.concurrent.CompletableFuture$AsyncSupply.run
>>>> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
>>>> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
>>>> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
>>>> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
>>>> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
>>>> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
>>>> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
>>>> Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252]
>>>>
>>>> Appreciate if anyone has any pointer for this.
>>>>
>>>> On Mon, Oct 26, 2020 at 10:45 AM Chesnay Schepler <ch...@apache.org>
>>>> wrote:
>>>>
>>>>> Flink 1.11 uses slf4j 1.7.15; the easiest way to check the log files
>>>>> is usually via the WebUI.
>>>>>
>>>>> On 10/26/2020 5:30 PM, Diwakar Jha wrote:
>>>>>
>>>>> I think my problem is with Sl4j library. I'm using sl4j 1.7 with Flink
>>>>> 1.11. If that's correct then i appreciate if someone can point me to the
>>>>> exact Slf4j library that i should use with Flink 1.11
>>>>>
>>>>> Flink = 1.11.x;
>>>>> Slf4j = 1.7;
>>>>>
>>>>>
>>>>> On Sun, Oct 25, 2020 at 8:00 PM Diwakar Jha <di...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Thanks for checking my configurations. Could you also point me where
>>>>>> I can see the log files? Just to give more details. I'm trying to access
>>>>>> these logs in AWS cloudwatch.
>>>>>>
>>>>>> Best,
>>>>>> Diwakar
>>>>>>
>>>>>> On Sun, Oct 25, 2020 at 2:16 PM Chesnay Schepler <ch...@apache.org>
>>>>>> wrote:
>>>>>>
>>>>>>> With Flink 1.11 reporters were refactored to plugins, and are now
>>>>>>> accessible by default (so you no longer have to bother with copying jars
>>>>>>> around).
>>>>>>>
>>>>>>> Your configuration appears to be correct, so I suggest to take a
>>>>>>> look at the log files.
>>>>>>>
>>>>>>> On 10/25/2020 9:52 PM, Diwakar Jha wrote:
>>>>>>>
>>>>>>> Hello Everyone,
>>>>>>>
>>>>>>> I'm new to flink and i'm trying to upgrade from flink 1.8 to flink
>>>>>>> 1.11 on an emr cluster. after upgrading to flink1.11 One of the differences
>>>>>>> that i see is i don't get any metrics. I found out that flink 1.11 does not
>>>>>>> have *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar in
>>>>>>> /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have any
>>>>>>> pointer to locate
>>>>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or how
>>>>>>> to use metrics in flink.1.11?
>>>>>>>
>>>>>>> Things i tried :
>>>>>>> a) the below setup
>>>>>>>
>>>>>>> metrics.reporters: stsdmetrics.reporter.stsd.factory.class: org.apache.flink.metrics.statsd.StatsDReporterFactorymetrics.reporter.stsd.host: localhostmetrics.reporter.stsd.port: 8125
>>>>>>>
>>>>>>> b) I tried downloading the statsd jar from
>>>>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
>>>>>>> putting it inside plugins/statsd directory.
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best,
>>>>>>> Diwakar Jha.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best,
>>>>>> Diwakar Jha.
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Best,
>>>>> Diwakar Jha.
>>>>>
>>>>>
>>>>>
>>>>
>>>> --
>>>> Best,
>>>> Diwakar Jha.
>>>>
>>>
>>
>> --
>> Best,
>> Diwakar Jha.
>>
>

-- 
Best,
Diwakar Jha.

Re: how to enable metrics in Flink 1.11

Posted by Robert Metzger <rm...@apache.org>.
Hey Diwakar,

how are you deploying Flink on EMR? Are you using YARN?
If so, you could also use log aggregation to see all the logs at once (from
both JobManager and TaskManagers). (yarn logs -applicationId <Application
ID>)

Could you post (or upload somewhere) all logs you have of one run? It is
much easier for us to debug something if we have the full logs (the logs
show for example the classpath that you are using, we would see how you are
deploying Flink, etc.)

From the information available, my guess is that you have modified your
deployment in some way (use of a custom logging version, custom deployment
method, version mixup with jars from both Flink 1.8 and 1.11, ...).

Best,
Robert


On Tue, Oct 27, 2020 at 12:41 AM Diwakar Jha <di...@gmail.com> wrote:

> This is what I see on the WebUI.
>
> 23:19:24.263 [flink-akka.actor.default-dispatcher-1865] ERROR
> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
> - Failed to transfer file from TaskExecutor
> container_1603649952937_0002_01_000004.
> java.util.concurrent.CompletionException:
> org.apache.flink.util.FlinkException: The file LOG does not exist on the
> TaskExecutor. at
> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
> java.util.concurrent.CompletableFuture$AsyncSupply.run
> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
> Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252] Caused by:
> org.apache.flink.util.FlinkException: The file LOG does not exist on the
> TaskExecutor. ... 5 more 23:19:24.275
> [flink-akka.actor.default-dispatcher-1865] ERROR
> org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
> - Unhandled exception. org.apache.flink.util.FlinkException: The file LOG
> does not exist on the TaskExecutor. at
> org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
> TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
> ~[flink-dist_2.12-1.11.0.jar:1.11.0] at
> java.util.concurrent.CompletableFuture$AsyncSupply.run
> <http://java.util.concurrent.completablefuture$asyncsupply.run/>(
> CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
> ~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
> <http://java.util.concurrent.threadpoolexecutor$worker.run/>(
> ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
> ~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
> Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252]
>
> Appreciate if anyone has any pointer for this.
>
> On Mon, Oct 26, 2020 at 10:45 AM Chesnay Schepler <ch...@apache.org>
> wrote:
>
>> Flink 1.11 uses slf4j 1.7.15; the easiest way to check the log files is
>> usually via the WebUI.
>>
>> On 10/26/2020 5:30 PM, Diwakar Jha wrote:
>>
>> I think my problem is with Sl4j library. I'm using sl4j 1.7 with Flink
>> 1.11. If that's correct then i appreciate if someone can point me to the
>> exact Slf4j library that i should use with Flink 1.11
>>
>> Flink = 1.11.x;
>> Slf4j = 1.7;
>>
>>
>> On Sun, Oct 25, 2020 at 8:00 PM Diwakar Jha <di...@gmail.com>
>> wrote:
>>
>>> Thanks for checking my configurations. Could you also point me where I
>>> can see the log files? Just to give more details. I'm trying to access
>>> these logs in AWS cloudwatch.
>>>
>>> Best,
>>> Diwakar
>>>
>>> On Sun, Oct 25, 2020 at 2:16 PM Chesnay Schepler <ch...@apache.org>
>>> wrote:
>>>
>>>> With Flink 1.11 reporters were refactored to plugins, and are now
>>>> accessible by default (so you no longer have to bother with copying jars
>>>> around).
>>>>
>>>> Your configuration appears to be correct, so I suggest to take a look
>>>> at the log files.
>>>>
>>>> On 10/25/2020 9:52 PM, Diwakar Jha wrote:
>>>>
>>>> Hello Everyone,
>>>>
>>>> I'm new to flink and i'm trying to upgrade from flink 1.8 to flink 1.11
>>>> on an emr cluster. after upgrading to flink1.11 One of the differences that
>>>> i see is i don't get any metrics. I found out that flink 1.11 does not have
>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar in
>>>> /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have any
>>>> pointer to locate
>>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or how to
>>>> use metrics in flink.1.11?
>>>>
>>>> Things i tried :
>>>> a) the below setup
>>>>
>>>> metrics.reporters: stsdmetrics.reporter.stsd.factory.class: org.apache.flink.metrics.statsd.StatsDReporterFactorymetrics.reporter.stsd.host: localhostmetrics.reporter.stsd.port: 8125
>>>>
>>>> b) I tried downloading the statsd jar from
>>>> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
>>>> putting it inside plugins/statsd directory.
>>>>
>>>>
>>>> --
>>>> Best,
>>>> Diwakar Jha.
>>>>
>>>>
>>>>
>>>
>>> --
>>> Best,
>>> Diwakar Jha.
>>>
>>
>>
>> --
>> Best,
>> Diwakar Jha.
>>
>>
>>
>
> --
> Best,
> Diwakar Jha.
>

Re: how to enable metrics in Flink 1.11

Posted by Diwakar Jha <di...@gmail.com>.
This is what I see on the WebUI.

23:19:24.263 [flink-akka.actor.default-dispatcher-1865] ERROR
org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
- Failed to transfer file from TaskExecutor
container_1603649952937_0002_01_000004.
java.util.concurrent.CompletionException:
org.apache.flink.util.FlinkException: The file LOG does not exist on the
TaskExecutor. at
org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
~[flink-dist_2.12-1.11.0.jar:1.11.0] at
java.util.concurrent.CompletableFuture$AsyncSupply.run
<http://java.util.concurrent.completablefuture$asyncsupply.run/>(
CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
<http://java.util.concurrent.threadpoolexecutor$worker.run/>(
ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252] Caused by:
org.apache.flink.util.FlinkException: The file LOG does not exist on the
TaskExecutor. ... 5 more 23:19:24.275
[flink-akka.actor.default-dispatcher-1865] ERROR
org.apache.flink.runtime.rest.handler.taskmanager.TaskManagerLogFileHandler
- Unhandled exception. org.apache.flink.util.FlinkException: The file LOG
does not exist on the TaskExecutor. at
org.apache.flink.runtime.taskexecutor.TaskExecutor.lambda$requestFileUploadByFilePath$25(
TaskExecutor.java:1742 <http://taskexecutor.java:1742/>)
~[flink-dist_2.12-1.11.0.jar:1.11.0] at
java.util.concurrent.CompletableFuture$AsyncSupply.run
<http://java.util.concurrent.completablefuture$asyncsupply.run/>(
CompletableFuture.java:1604 <http://completablefuture.java:1604/>)
~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor.runWorker(
ThreadPoolExecutor.java:1149 <http://threadpoolexecutor.java:1149/>)
~[?:1.8.0_252] at java.util.concurrent.ThreadPoolExecutor$Worker.run
<http://java.util.concurrent.threadpoolexecutor$worker.run/>(
ThreadPoolExecutor.java:624 <http://threadpoolexecutor.java:624/>)
~[?:1.8.0_252] at java.lang.Thread.run <http://java.lang.thread.run/>(
Thread.java:748 <http://thread.java:748/>) ~[?:1.8.0_252]

Appreciate if anyone has any pointer for this.

On Mon, Oct 26, 2020 at 10:45 AM Chesnay Schepler <ch...@apache.org>
wrote:

> Flink 1.11 uses slf4j 1.7.15; the easiest way to check the log files is
> usually via the WebUI.
>
> On 10/26/2020 5:30 PM, Diwakar Jha wrote:
>
> I think my problem is with Sl4j library. I'm using sl4j 1.7 with Flink
> 1.11. If that's correct then i appreciate if someone can point me to the
> exact Slf4j library that i should use with Flink 1.11
>
> Flink = 1.11.x;
> Slf4j = 1.7;
>
>
> On Sun, Oct 25, 2020 at 8:00 PM Diwakar Jha <di...@gmail.com>
> wrote:
>
>> Thanks for checking my configurations. Could you also point me where I
>> can see the log files? Just to give more details. I'm trying to access
>> these logs in AWS cloudwatch.
>>
>> Best,
>> Diwakar
>>
>> On Sun, Oct 25, 2020 at 2:16 PM Chesnay Schepler <ch...@apache.org>
>> wrote:
>>
>>> With Flink 1.11 reporters were refactored to plugins, and are now
>>> accessible by default (so you no longer have to bother with copying jars
>>> around).
>>>
>>> Your configuration appears to be correct, so I suggest to take a look at
>>> the log files.
>>>
>>> On 10/25/2020 9:52 PM, Diwakar Jha wrote:
>>>
>>> Hello Everyone,
>>>
>>> I'm new to flink and i'm trying to upgrade from flink 1.8 to flink 1.11
>>> on an emr cluster. after upgrading to flink1.11 One of the differences that
>>> i see is i don't get any metrics. I found out that flink 1.11 does not have
>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar in
>>> /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have any
>>> pointer to locate
>>> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or how to
>>> use metrics in flink.1.11?
>>>
>>> Things i tried :
>>> a) the below setup
>>>
>>> metrics.reporters: stsdmetrics.reporter.stsd.factory.class: org.apache.flink.metrics.statsd.StatsDReporterFactorymetrics.reporter.stsd.host: localhostmetrics.reporter.stsd.port: 8125
>>>
>>> b) I tried downloading the statsd jar from
>>> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd
>>> putting it inside plugins/statsd directory.
>>>
>>>
>>> --
>>> Best,
>>> Diwakar Jha.
>>>
>>>
>>>
>>
>> --
>> Best,
>> Diwakar Jha.
>>
>
>
> --
> Best,
> Diwakar Jha.
>
>
>

-- 
Best,
Diwakar Jha.

Re: how to enable metrics in Flink 1.11

Posted by Chesnay Schepler <ch...@apache.org>.
With Flink 1.11 reporters were refactored to plugins, and are now 
accessible by default (so you no longer have to bother with copying jars 
around).

Your configuration appears to be correct, so I suggest to take a look at 
the log files.

On 10/25/2020 9:52 PM, Diwakar Jha wrote:
> Hello Everyone,
>
> I'm new to flink and i'm trying to upgrade from flink 1.8 to flink 
> 1.11 on an emr cluster. after upgrading to flink1.11 One of the 
> differences that i see is i don't get any metrics. I found out that 
> flink 1.11 does not have 
> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar in 
> /usr/lib/flink/opt which was the case for flink 1.8. Could anyone have 
> any pointer to locate 
> *org.apache.flink.metrics.statsd.StatsDReporterFactory* jar or how to 
> use metrics in flink.1.11?
>
> Things i tried :
> a) the below setup
> metrics.reporters: stsd
> metrics.reporter.stsd.factory.class: org.apache.flink.metrics.statsd.StatsDReporterFactory
> metrics.reporter.stsd.host: localhost
> metrics.reporter.stsd.port: 8125
> b) I tried downloading the statsd jar from 
> https://mvnrepository.com/artifact/org.apache.flink/flink-metrics-statsd 
> putting it inside plugins/statsd directory.
>
>
> -- 
> Best,
> Diwakar Jha.