You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@flink.apache.org by aj <aj...@gmail.com> on 2020/05/01 10:35:51 UTC

Re: upgrade flink from 1.9.1 to 1.10.0 on EMR

Hi Yang,

I am attaching the logs for your reference, please help me what i am doing
wrong.

Thanks,
Anuj

On Wed, Apr 29, 2020 at 9:06 AM Yang Wang <da...@gmail.com> wrote:

> Hi Anuj,
>
> I think the exception you come across still because the hadoop version
> is 2.4.1. I have checked the hadoop code, the code line are exactly same.
> For 2.8.1, i also have checked the ruleParse. It could work.
>
> /**
>  * A pattern for parsing a auth_to_local rule.
>  */
> private static final Pattern ruleParser =
>   Pattern.compile("\\s*((DEFAULT)|(RULE:\\[(\\d*):([^\\]]*)](\\(([^)]*)\\))?"+
>                   "(s/([^/]*)/([^/]*)/(g)?)?))/?(L)?");
>
>
> Could you share the jobmanager logs so that i could check the classpath
> and hadoop version?
>
> Best,
> Yang
>
> aj <aj...@gmail.com> 于2020年4月28日周二 上午1:01写道:
>
>> Hello Yang,
>> My Hadoop version is Hadoop 3.2.1-amzn-0
>> and I have put in flink/lib.   flink-shaded-hadoop-2-uber-2.8.3-10.0.jar
>>
>> then I am getting below error :
>>
>> SLF4J: Class path contains multiple SLF4J bindings.
>> SLF4J: Found binding in
>> [jar:file:/mnt/yarn/usercache/hadoop/appcache/application_1587983834922_0002/filecache/10/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: Found binding in
>> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>> explanation.
>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>> Exception in thread "main" java.lang.IllegalArgumentException: Invalid
>> rule: /L
>>       RULE:[2:$1@$0](.*@)s/@.*///L
>>       DEFAULT
>>         at
>> org.apache.hadoop.security.authentication.util.KerberosName.parseRules(KerberosName.java:321)
>>         at
>> org.apache.hadoop.security.authentication.util.KerberosName.setRules(KerberosName.java:386)
>>         at
>> org.apache.hadoop.security.HadoopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>         at
>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:247)
>>         at
>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232)
>>         at
>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:718)
>>         at
>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:703)
>>         at
>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:605)
>>         at
>> org.apache.flink.yarn.entrypoint.YarnEntrypointUtils.logYarnEnvironmentInformation(YarnEntrypointUtils.java:136)
>>         at
>> org.apache.flink.yarn.entrypoint.YarnJobClusterEntrypoint.main(YarnJobClusterEntrypoint.java:109)
>>
>>
>> if I remove the  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar  from lib
>> then i get below error:
>>
>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>                   -  Classpath:
>> /usr/lib/flink/lib/flink-table-blink_2.11-1.10.0.jar:/usr/lib/flink/lib/flink-table_2.11-1.10.0.jar:/usr/lib/flink/lib/log4j-1.2.17.jar:/usr/lib/flink/lib/slf4j-log4j12-1.7.15.jar:/usr/lib/flink/lib/flink-dist_2.11-1.10.0.jar::/etc/hadoop/conf:/etc/hadoop/conf
>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>                   -
>> --------------------------------------------------------------------------------
>> 2020-04-27 16:59:37,300 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: jobmanager.heap.size, 1024m
>> 2020-04-27 16:59:37,300 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: taskmanager.memory.process.size, 1568m
>> 2020-04-27 16:59:37,300 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: taskmanager.numberOfTaskSlots, 1
>> 2020-04-27 16:59:37,300 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: parallelism.default, 1
>> 2020-04-27 16:59:37,300 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: env.yarn.conf.dir, /etc/hadoop/conf
>> 2020-04-27 16:59:37,300 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: env.hadoop.conf.dir, /etc/hadoop/conf
>> 2020-04-27 16:59:37,301 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: jobmanager.execution.failover-strategy, region
>> 2020-04-27 16:59:37,301 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: classloader.resolve-order, parent-first
>> 2020-04-27 16:59:37,301 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: s3.access-key, AKIA52DD5QA5FC7HPKXG
>> 2020-04-27 16:59:37,301 INFO
>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>> configuration property: s3.secret-key, ******
>> 2020-04-27 16:59:37,305 WARN  org.apache.flink.client.cli.CliFrontend
>>                   - Could not load CLI class
>> org.apache.flink.yarn.cli.FlinkYarnSessionCli.
>> java.lang.NoClassDefFoundError:
>> org/apache/hadoop/yarn/exceptions/YarnException
>>         at java.lang.Class.forName0(Native Method)
>>         at java.lang.Class.forName(Class.java:264)
>>         at
>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1076)
>>         at
>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1030)
>>         at
>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:957)
>> Caused by: java.lang.ClassNotFoundException:
>> org.apache.hadoop.yarn.exceptions.YarnException
>>         at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:419)
>>         at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:352)
>>         ... 5 more
>> 2020-04-27 16:59:37,406 INFO  org.apache.flink.core.fs.FileSystem
>>                   - Hadoop is not in the classpath/dependencies. The
>> extended set of supported File Systems via Hadoop is not available.
>> 2020-04-27 16:59:37,458 INFO
>>  org.apache.flink.runtime.security.modules.HadoopModuleFactory  - Cannot
>> create Hadoop Security Module because Hadoop cannot be found in the
>> Classpath.
>> 2020-04-27 16:59:37,476 INFO
>>  org.apache.flink.runtime.security.modules.JaasModule          - Jaas file
>> will be created as /tmp/jaas-7054453135321774613.conf.
>> 2020-04-27 16:59:37,480 INFO
>>  org.apache.flink.runtime.security.SecurityUtils               - Cannot
>> install HadoopSecurityContext because Hadoop cannot be found in the
>> Classpath.
>> 2020-04-27 16:59:37,481 INFO  org.apache.flink.client.cli.CliFrontend
>>                   - Running 'run' command.
>> 2020-04-27 16:59:37,488 INFO  org.apache.flink.client.cli.CliFrontend
>>                   - Building program from JAR file
>> 2020-04-27 16:59:37,488 ERROR org.apache.flink.client.cli.CliFrontend
>>                   - Invalid command line arguments.
>> org.apache.flink.client.cli.CliArgsException: Could not build the program
>> from JAR file.
>>         at
>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:203)
>>         at
>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>>         at
>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>>         at
>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
>>         at
>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
>> Caused by: java.io.FileNotFoundException: JAR file does not exist: -ynm
>>         at
>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:719)
>>         at
>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:695)
>>         at
>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:200)
>>         ... 4 more
>>
>>
>> Can you please help.
>>
>> Thanks,
>> Anuj
>>
>>
>> On Mon, Apr 13, 2020 at 7:43 AM Yang Wang <da...@gmail.com> wrote:
>>
>>> Hi Anuj,
>>>
>>> It seems that you are using hadoop version 2.4.1. I think "L" could not
>>> be supported in
>>> this version. Could you upgrade your hadoop version to 2.8 and have a
>>> try? If your
>>> YARN cluster version is 2.8+, then you could directly remove the
>>> flink-shaded-hadoop
>>> in your lib directory. Otherwise, you need to download the
>>> flink-shaded-hadoop with
>>> version 2.8 here[1].
>>>
>>>
>>> [1]. https://flink.apache.org/downloads.html#additional-components
>>>
>>> Best,
>>> Yang
>>>
>>> aj <aj...@gmail.com> 于2020年4月11日周六 上午4:21写道:
>>>
>>>> Hi Robert,
>>>> attached the full application log file.
>>>>
>>>> Thanks,
>>>> Anuj
>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>
>>>
>>
>> --
>> Thanks & Regards,
>> Anuj Jain
>> Mob. : +91- 8588817877
>> Skype : anuj.jain07
>> <http://www.oracle.com/>
>>
>>
>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>
>

-- 
Thanks & Regards,
Anuj Jain
Mob. : +91- 8588817877
Skype : anuj.jain07
<http://www.oracle.com/>


<http://www.cse.iitm.ac.in/%7Eanujjain/>

Re: upgrade flink from 1.9.1 to 1.10.0 on EMR

Posted by Robert Metzger <rm...@apache.org>.
Flink 1.11 will support Hadoop 3. EMR 6 requires Hadoop 3, that's why Flink
was not included anymore. Amazon will add Flink back to EMR 6.0 soon.

On Thu, May 14, 2020 at 7:11 PM aj <aj...@gmail.com> wrote:

> Hi Yang,
>
> I am able to resolve the issue by removing Hadoop dependency as you
> mentioned.
>
> 1. Removed hadoop-common dependency and
>
>     <groupId>org.apache.flink</groupId>
>     <artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
>     <version>${flink.version}</version>
>     <exclusions>
>         <exclusion>
>                 <groupId>org.apache.flink</groupId>
>                 <artifactId>flink-hadoop-fs</artifactId>
>         </exclusion>
>     </exclusions>
>
> </dependency>
>
>
> 2. After the above changes, I am able to submit job on yarn but facing
> issues with s3 plugin that I resolved by putting *
> flink-s3-fs-hadoop-1.9.0.jar *i n the plugins/s3-fs-hadoop  directory.
>
> Thanks for your support.
>
> Any update when will flink.10 officially supported in EMR.  Even in new
> EMR version(emr 6.0)  flink has been removed.
>
>
>
> On Sat, May 9, 2020 at 1:36 PM aj <aj...@gmail.com> wrote:
>
>> Hello Yang,
>>
>> I have attached my pom file and I did not see that I am using any Hadoop
>> dependency. Can you please help me.
>>
>> On Wed, May 6, 2020 at 1:22 PM Yang Wang <da...@gmail.com> wrote:
>>
>>> Hi aj,
>>>
>>> From the logs you have provided, the hadoop version is still 2.4.1.
>>> Could you check the user jar(i.e. events-processor-1.0-SNAPSHOT.jar)
>>> have some
>>> hadoop classes? If it is, you need to exclude the hadoop dependency.
>>>
>>>
>>> Best,
>>> Yang
>>>
>>> aj <aj...@gmail.com> 于2020年5月6日周三 下午3:38写道:
>>>
>>>> Hello,
>>>>
>>>> Please help me upgrade to 1.10 in AWS EMR.
>>>>
>>>> On Fri, May 1, 2020 at 4:05 PM aj <aj...@gmail.com> wrote:
>>>>
>>>>> Hi Yang,
>>>>>
>>>>> I am attaching the logs for your reference, please help me what i am
>>>>> doing wrong.
>>>>>
>>>>> Thanks,
>>>>> Anuj
>>>>>
>>>>> On Wed, Apr 29, 2020 at 9:06 AM Yang Wang <da...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi Anuj,
>>>>>>
>>>>>> I think the exception you come across still because the hadoop version
>>>>>> is 2.4.1. I have checked the hadoop code, the code line are exactly
>>>>>> same.
>>>>>> For 2.8.1, i also have checked the ruleParse. It could work.
>>>>>>
>>>>>> /**
>>>>>>  * A pattern for parsing a auth_to_local rule.
>>>>>>  */
>>>>>> private static final Pattern ruleParser =
>>>>>>   Pattern.compile("\\s*((DEFAULT)|(RULE:\\[(\\d*):([^\\]]*)](\\(([^)]*)\\))?"+
>>>>>>                   "(s/([^/]*)/([^/]*)/(g)?)?))/?(L)?");
>>>>>>
>>>>>>
>>>>>> Could you share the jobmanager logs so that i could check the
>>>>>> classpath and hadoop version?
>>>>>>
>>>>>> Best,
>>>>>> Yang
>>>>>>
>>>>>> aj <aj...@gmail.com> 于2020年4月28日周二 上午1:01写道:
>>>>>>
>>>>>>> Hello Yang,
>>>>>>> My Hadoop version is Hadoop 3.2.1-amzn-0
>>>>>>> and I have put in flink/lib.
>>>>>>>  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar
>>>>>>>
>>>>>>> then I am getting below error :
>>>>>>>
>>>>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>>>>> SLF4J: Found binding in
>>>>>>> [jar:file:/mnt/yarn/usercache/hadoop/appcache/application_1587983834922_0002/filecache/10/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>> SLF4J: Found binding in
>>>>>>> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>>>>> explanation.
>>>>>>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>>>>>>> Exception in thread "main" java.lang.IllegalArgumentException:
>>>>>>> Invalid rule: /L
>>>>>>>       RULE:[2:$1@$0](.*@)s/@.*///L
>>>>>>>       DEFAULT
>>>>>>>         at
>>>>>>> org.apache.hadoop.security.authentication.util.KerberosName.parseRules(KerberosName.java:321)
>>>>>>>         at
>>>>>>> org.apache.hadoop.security.authentication.util.KerberosName.setRules(KerberosName.java:386)
>>>>>>>         at
>>>>>>> org.apache.hadoop.security.HadoopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>>>>>         at
>>>>>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:247)
>>>>>>>         at
>>>>>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232)
>>>>>>>         at
>>>>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:718)
>>>>>>>         at
>>>>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:703)
>>>>>>>         at
>>>>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:605)
>>>>>>>         at
>>>>>>> org.apache.flink.yarn.entrypoint.YarnEntrypointUtils.logYarnEnvironmentInformation(YarnEntrypointUtils.java:136)
>>>>>>>         at
>>>>>>> org.apache.flink.yarn.entrypoint.YarnJobClusterEntrypoint.main(YarnJobClusterEntrypoint.java:109)
>>>>>>>
>>>>>>>
>>>>>>> if I remove the  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar  from
>>>>>>> lib  then i get below error:
>>>>>>>
>>>>>>> 2020-04-27 16:59:37,293 INFO
>>>>>>>  org.apache.flink.client.cli.CliFrontend                       -
>>>>>>>  Classpath:
>>>>>>> /usr/lib/flink/lib/flink-table-blink_2.11-1.10.0.jar:/usr/lib/flink/lib/flink-table_2.11-1.10.0.jar:/usr/lib/flink/lib/log4j-1.2.17.jar:/usr/lib/flink/lib/slf4j-log4j12-1.7.15.jar:/usr/lib/flink/lib/flink-dist_2.11-1.10.0.jar::/etc/hadoop/conf:/etc/hadoop/conf
>>>>>>> 2020-04-27 16:59:37,293 INFO
>>>>>>>  org.apache.flink.client.cli.CliFrontend                       -
>>>>>>> --------------------------------------------------------------------------------
>>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: jobmanager.heap.size, 1024m
>>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: taskmanager.memory.process.size, 1568m
>>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: taskmanager.numberOfTaskSlots, 1
>>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: parallelism.default, 1
>>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: env.yarn.conf.dir, /etc/hadoop/conf
>>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: env.hadoop.conf.dir, /etc/hadoop/conf
>>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: jobmanager.execution.failover-strategy, region
>>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: classloader.resolve-order, parent-first
>>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: s3.access-key, AKIA52DD5QA5FC7HPKXG
>>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>>> configuration property: s3.secret-key, ******
>>>>>>> 2020-04-27 16:59:37,305 WARN
>>>>>>>  org.apache.flink.client.cli.CliFrontend                       - Could not
>>>>>>> load CLI class org.apache.flink.yarn.cli.FlinkYarnSessionCli.
>>>>>>> java.lang.NoClassDefFoundError:
>>>>>>> org/apache/hadoop/yarn/exceptions/YarnException
>>>>>>>         at java.lang.Class.forName0(Native Method)
>>>>>>>         at java.lang.Class.forName(Class.java:264)
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1076)
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1030)
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:957)
>>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>>> org.apache.hadoop.yarn.exceptions.YarnException
>>>>>>>         at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
>>>>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:419)
>>>>>>>         at
>>>>>>> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
>>>>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:352)
>>>>>>>         ... 5 more
>>>>>>> 2020-04-27 16:59:37,406 INFO  org.apache.flink.core.fs.FileSystem
>>>>>>>                         - Hadoop is not in the classpath/dependencies. The
>>>>>>> extended set of supported File Systems via Hadoop is not available.
>>>>>>> 2020-04-27 16:59:37,458 INFO
>>>>>>>  org.apache.flink.runtime.security.modules.HadoopModuleFactory  - Cannot
>>>>>>> create Hadoop Security Module because Hadoop cannot be found in the
>>>>>>> Classpath.
>>>>>>> 2020-04-27 16:59:37,476 INFO
>>>>>>>  org.apache.flink.runtime.security.modules.JaasModule          - Jaas file
>>>>>>> will be created as /tmp/jaas-7054453135321774613.conf.
>>>>>>> 2020-04-27 16:59:37,480 INFO
>>>>>>>  org.apache.flink.runtime.security.SecurityUtils               - Cannot
>>>>>>> install HadoopSecurityContext because Hadoop cannot be found in the
>>>>>>> Classpath.
>>>>>>> 2020-04-27 16:59:37,481 INFO
>>>>>>>  org.apache.flink.client.cli.CliFrontend                       - Running
>>>>>>> 'run' command.
>>>>>>> 2020-04-27 16:59:37,488 INFO
>>>>>>>  org.apache.flink.client.cli.CliFrontend                       - Building
>>>>>>> program from JAR file
>>>>>>> 2020-04-27 16:59:37,488 ERROR
>>>>>>> org.apache.flink.client.cli.CliFrontend                       - Invalid
>>>>>>> command line arguments.
>>>>>>> org.apache.flink.client.cli.CliArgsException: Could not build the
>>>>>>> program from JAR file.
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:203)
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>>>>>>>         at
>>>>>>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
>>>>>>> Caused by: java.io.FileNotFoundException: JAR file does not exist:
>>>>>>> -ynm
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:719)
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:695)
>>>>>>>         at
>>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:200)
>>>>>>>         ... 4 more
>>>>>>>
>>>>>>>
>>>>>>> Can you please help.
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Anuj
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Apr 13, 2020 at 7:43 AM Yang Wang <da...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi Anuj,
>>>>>>>>
>>>>>>>> It seems that you are using hadoop version 2.4.1. I think "L" could
>>>>>>>> not be supported in
>>>>>>>> this version. Could you upgrade your hadoop version to 2.8 and have
>>>>>>>> a try? If your
>>>>>>>> YARN cluster version is 2.8+, then you could directly remove the
>>>>>>>> flink-shaded-hadoop
>>>>>>>> in your lib directory. Otherwise, you need to download the
>>>>>>>> flink-shaded-hadoop with
>>>>>>>> version 2.8 here[1].
>>>>>>>>
>>>>>>>>
>>>>>>>> [1]. https://flink.apache.org/downloads.html#additional-components
>>>>>>>>
>>>>>>>> Best,
>>>>>>>> Yang
>>>>>>>>
>>>>>>>> aj <aj...@gmail.com> 于2020年4月11日周六 上午4:21写道:
>>>>>>>>
>>>>>>>>> Hi Robert,
>>>>>>>>> attached the full application log file.
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Anuj
>>>>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Thanks & Regards,
>>>>>>> Anuj Jain
>>>>>>> Mob. : +91- 8588817877
>>>>>>> Skype : anuj.jain07
>>>>>>> <http://www.oracle.com/>
>>>>>>>
>>>>>>>
>>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> Thanks & Regards,
>>>>> Anuj Jain
>>>>> Mob. : +91- 8588817877
>>>>> Skype : anuj.jain07
>>>>> <http://www.oracle.com/>
>>>>>
>>>>>
>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>
>>>>
>>>>
>>>> --
>>>> Thanks & Regards,
>>>> Anuj Jain
>>>> Mob. : +91- 8588817877
>>>> Skype : anuj.jain07
>>>> <http://www.oracle.com/>
>>>>
>>>>
>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>
>>>
>>
>> --
>> Thanks & Regards,
>> Anuj Jain
>> Mob. : +91- 8588817877
>> Skype : anuj.jain07
>> <http://www.oracle.com/>
>>
>>
>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>
>
>
> --
> Thanks & Regards,
> Anuj Jain
> Mob. : +91- 8588817877
> Skype : anuj.jain07
> <http://www.oracle.com/>
>
>
> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>

Re: upgrade flink from 1.9.1 to 1.10.0 on EMR

Posted by aj <aj...@gmail.com>.
Hi Yang,

I am able to resolve the issue by removing Hadoop dependency as you
mentioned.

1. Removed hadoop-common dependency and

    <groupId>org.apache.flink</groupId>
    <artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
    <version>${flink.version}</version>
    <exclusions>
        <exclusion>
                <groupId>org.apache.flink</groupId>
                <artifactId>flink-hadoop-fs</artifactId>
        </exclusion>
    </exclusions>

</dependency>


2. After the above changes, I am able to submit job on yarn but facing
issues with s3 plugin that I resolved by putting *
flink-s3-fs-hadoop-1.9.0.jar *i n the plugins/s3-fs-hadoop  directory.

Thanks for your support.

Any update when will flink.10 officially supported in EMR.  Even in new EMR
version(emr 6.0)  flink has been removed.



On Sat, May 9, 2020 at 1:36 PM aj <aj...@gmail.com> wrote:

> Hello Yang,
>
> I have attached my pom file and I did not see that I am using any Hadoop
> dependency. Can you please help me.
>
> On Wed, May 6, 2020 at 1:22 PM Yang Wang <da...@gmail.com> wrote:
>
>> Hi aj,
>>
>> From the logs you have provided, the hadoop version is still 2.4.1.
>> Could you check the user jar(i.e. events-processor-1.0-SNAPSHOT.jar) have
>> some
>> hadoop classes? If it is, you need to exclude the hadoop dependency.
>>
>>
>> Best,
>> Yang
>>
>> aj <aj...@gmail.com> 于2020年5月6日周三 下午3:38写道:
>>
>>> Hello,
>>>
>>> Please help me upgrade to 1.10 in AWS EMR.
>>>
>>> On Fri, May 1, 2020 at 4:05 PM aj <aj...@gmail.com> wrote:
>>>
>>>> Hi Yang,
>>>>
>>>> I am attaching the logs for your reference, please help me what i am
>>>> doing wrong.
>>>>
>>>> Thanks,
>>>> Anuj
>>>>
>>>> On Wed, Apr 29, 2020 at 9:06 AM Yang Wang <da...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi Anuj,
>>>>>
>>>>> I think the exception you come across still because the hadoop version
>>>>> is 2.4.1. I have checked the hadoop code, the code line are exactly
>>>>> same.
>>>>> For 2.8.1, i also have checked the ruleParse. It could work.
>>>>>
>>>>> /**
>>>>>  * A pattern for parsing a auth_to_local rule.
>>>>>  */
>>>>> private static final Pattern ruleParser =
>>>>>   Pattern.compile("\\s*((DEFAULT)|(RULE:\\[(\\d*):([^\\]]*)](\\(([^)]*)\\))?"+
>>>>>                   "(s/([^/]*)/([^/]*)/(g)?)?))/?(L)?");
>>>>>
>>>>>
>>>>> Could you share the jobmanager logs so that i could check the
>>>>> classpath and hadoop version?
>>>>>
>>>>> Best,
>>>>> Yang
>>>>>
>>>>> aj <aj...@gmail.com> 于2020年4月28日周二 上午1:01写道:
>>>>>
>>>>>> Hello Yang,
>>>>>> My Hadoop version is Hadoop 3.2.1-amzn-0
>>>>>> and I have put in flink/lib.
>>>>>>  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar
>>>>>>
>>>>>> then I am getting below error :
>>>>>>
>>>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>>>> SLF4J: Found binding in
>>>>>> [jar:file:/mnt/yarn/usercache/hadoop/appcache/application_1587983834922_0002/filecache/10/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>> SLF4J: Found binding in
>>>>>> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>>>> explanation.
>>>>>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>>>>>> Exception in thread "main" java.lang.IllegalArgumentException:
>>>>>> Invalid rule: /L
>>>>>>       RULE:[2:$1@$0](.*@)s/@.*///L
>>>>>>       DEFAULT
>>>>>>         at
>>>>>> org.apache.hadoop.security.authentication.util.KerberosName.parseRules(KerberosName.java:321)
>>>>>>         at
>>>>>> org.apache.hadoop.security.authentication.util.KerberosName.setRules(KerberosName.java:386)
>>>>>>         at
>>>>>> org.apache.hadoop.security.HadoopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:247)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:718)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:703)
>>>>>>         at
>>>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:605)
>>>>>>         at
>>>>>> org.apache.flink.yarn.entrypoint.YarnEntrypointUtils.logYarnEnvironmentInformation(YarnEntrypointUtils.java:136)
>>>>>>         at
>>>>>> org.apache.flink.yarn.entrypoint.YarnJobClusterEntrypoint.main(YarnJobClusterEntrypoint.java:109)
>>>>>>
>>>>>>
>>>>>> if I remove the  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar  from lib
>>>>>> then i get below error:
>>>>>>
>>>>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                       -  Classpath:
>>>>>> /usr/lib/flink/lib/flink-table-blink_2.11-1.10.0.jar:/usr/lib/flink/lib/flink-table_2.11-1.10.0.jar:/usr/lib/flink/lib/log4j-1.2.17.jar:/usr/lib/flink/lib/slf4j-log4j12-1.7.15.jar:/usr/lib/flink/lib/flink-dist_2.11-1.10.0.jar::/etc/hadoop/conf:/etc/hadoop/conf
>>>>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                       -
>>>>>> --------------------------------------------------------------------------------
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: jobmanager.heap.size, 1024m
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: taskmanager.memory.process.size, 1568m
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: taskmanager.numberOfTaskSlots, 1
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: parallelism.default, 1
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: env.yarn.conf.dir, /etc/hadoop/conf
>>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: env.hadoop.conf.dir, /etc/hadoop/conf
>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: jobmanager.execution.failover-strategy, region
>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: classloader.resolve-order, parent-first
>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: s3.access-key, AKIA52DD5QA5FC7HPKXG
>>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>>> configuration property: s3.secret-key, ******
>>>>>> 2020-04-27 16:59:37,305 WARN  org.apache.flink.client.cli.CliFrontend
>>>>>>                       - Could not load CLI class
>>>>>> org.apache.flink.yarn.cli.FlinkYarnSessionCli.
>>>>>> java.lang.NoClassDefFoundError:
>>>>>> org/apache/hadoop/yarn/exceptions/YarnException
>>>>>>         at java.lang.Class.forName0(Native Method)
>>>>>>         at java.lang.Class.forName(Class.java:264)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1076)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1030)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:957)
>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>> org.apache.hadoop.yarn.exceptions.YarnException
>>>>>>         at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
>>>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:419)
>>>>>>         at
>>>>>> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
>>>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:352)
>>>>>>         ... 5 more
>>>>>> 2020-04-27 16:59:37,406 INFO  org.apache.flink.core.fs.FileSystem
>>>>>>                       - Hadoop is not in the classpath/dependencies. The
>>>>>> extended set of supported File Systems via Hadoop is not available.
>>>>>> 2020-04-27 16:59:37,458 INFO
>>>>>>  org.apache.flink.runtime.security.modules.HadoopModuleFactory  - Cannot
>>>>>> create Hadoop Security Module because Hadoop cannot be found in the
>>>>>> Classpath.
>>>>>> 2020-04-27 16:59:37,476 INFO
>>>>>>  org.apache.flink.runtime.security.modules.JaasModule          - Jaas file
>>>>>> will be created as /tmp/jaas-7054453135321774613.conf.
>>>>>> 2020-04-27 16:59:37,480 INFO
>>>>>>  org.apache.flink.runtime.security.SecurityUtils               - Cannot
>>>>>> install HadoopSecurityContext because Hadoop cannot be found in the
>>>>>> Classpath.
>>>>>> 2020-04-27 16:59:37,481 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                       - Running 'run' command.
>>>>>> 2020-04-27 16:59:37,488 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>>                       - Building program from JAR file
>>>>>> 2020-04-27 16:59:37,488 ERROR org.apache.flink.client.cli.CliFrontend
>>>>>>                       - Invalid command line arguments.
>>>>>> org.apache.flink.client.cli.CliArgsException: Could not build the
>>>>>> program from JAR file.
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:203)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>>>>>>         at
>>>>>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
>>>>>> Caused by: java.io.FileNotFoundException: JAR file does not exist:
>>>>>> -ynm
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:719)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:695)
>>>>>>         at
>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:200)
>>>>>>         ... 4 more
>>>>>>
>>>>>>
>>>>>> Can you please help.
>>>>>>
>>>>>> Thanks,
>>>>>> Anuj
>>>>>>
>>>>>>
>>>>>> On Mon, Apr 13, 2020 at 7:43 AM Yang Wang <da...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi Anuj,
>>>>>>>
>>>>>>> It seems that you are using hadoop version 2.4.1. I think "L" could
>>>>>>> not be supported in
>>>>>>> this version. Could you upgrade your hadoop version to 2.8 and have
>>>>>>> a try? If your
>>>>>>> YARN cluster version is 2.8+, then you could directly remove the
>>>>>>> flink-shaded-hadoop
>>>>>>> in your lib directory. Otherwise, you need to download the
>>>>>>> flink-shaded-hadoop with
>>>>>>> version 2.8 here[1].
>>>>>>>
>>>>>>>
>>>>>>> [1]. https://flink.apache.org/downloads.html#additional-components
>>>>>>>
>>>>>>> Best,
>>>>>>> Yang
>>>>>>>
>>>>>>> aj <aj...@gmail.com> 于2020年4月11日周六 上午4:21写道:
>>>>>>>
>>>>>>>> Hi Robert,
>>>>>>>> attached the full application log file.
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Anuj
>>>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>> --
>>>>>> Thanks & Regards,
>>>>>> Anuj Jain
>>>>>> Mob. : +91- 8588817877
>>>>>> Skype : anuj.jain07
>>>>>> <http://www.oracle.com/>
>>>>>>
>>>>>>
>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>>
>>>>>
>>>>
>>>> --
>>>> Thanks & Regards,
>>>> Anuj Jain
>>>> Mob. : +91- 8588817877
>>>> Skype : anuj.jain07
>>>> <http://www.oracle.com/>
>>>>
>>>>
>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>
>>>
>>>
>>> --
>>> Thanks & Regards,
>>> Anuj Jain
>>> Mob. : +91- 8588817877
>>> Skype : anuj.jain07
>>> <http://www.oracle.com/>
>>>
>>>
>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>
>>
>
> --
> Thanks & Regards,
> Anuj Jain
> Mob. : +91- 8588817877
> Skype : anuj.jain07
> <http://www.oracle.com/>
>
>
> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>


-- 
Thanks & Regards,
Anuj Jain
Mob. : +91- 8588817877
Skype : anuj.jain07
<http://www.oracle.com/>


<http://www.cse.iitm.ac.in/%7Eanujjain/>

Re: upgrade flink from 1.9.1 to 1.10.0 on EMR

Posted by aj <aj...@gmail.com>.
Hello Yang,

I have attached my pom file and I did not see that I am using any Hadoop
dependency. Can you please help me.

On Wed, May 6, 2020 at 1:22 PM Yang Wang <da...@gmail.com> wrote:

> Hi aj,
>
> From the logs you have provided, the hadoop version is still 2.4.1.
> Could you check the user jar(i.e. events-processor-1.0-SNAPSHOT.jar) have
> some
> hadoop classes? If it is, you need to exclude the hadoop dependency.
>
>
> Best,
> Yang
>
> aj <aj...@gmail.com> 于2020年5月6日周三 下午3:38写道:
>
>> Hello,
>>
>> Please help me upgrade to 1.10 in AWS EMR.
>>
>> On Fri, May 1, 2020 at 4:05 PM aj <aj...@gmail.com> wrote:
>>
>>> Hi Yang,
>>>
>>> I am attaching the logs for your reference, please help me what i am
>>> doing wrong.
>>>
>>> Thanks,
>>> Anuj
>>>
>>> On Wed, Apr 29, 2020 at 9:06 AM Yang Wang <da...@gmail.com> wrote:
>>>
>>>> Hi Anuj,
>>>>
>>>> I think the exception you come across still because the hadoop version
>>>> is 2.4.1. I have checked the hadoop code, the code line are exactly
>>>> same.
>>>> For 2.8.1, i also have checked the ruleParse. It could work.
>>>>
>>>> /**
>>>>  * A pattern for parsing a auth_to_local rule.
>>>>  */
>>>> private static final Pattern ruleParser =
>>>>   Pattern.compile("\\s*((DEFAULT)|(RULE:\\[(\\d*):([^\\]]*)](\\(([^)]*)\\))?"+
>>>>                   "(s/([^/]*)/([^/]*)/(g)?)?))/?(L)?");
>>>>
>>>>
>>>> Could you share the jobmanager logs so that i could check the classpath
>>>> and hadoop version?
>>>>
>>>> Best,
>>>> Yang
>>>>
>>>> aj <aj...@gmail.com> 于2020年4月28日周二 上午1:01写道:
>>>>
>>>>> Hello Yang,
>>>>> My Hadoop version is Hadoop 3.2.1-amzn-0
>>>>> and I have put in flink/lib.
>>>>>  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar
>>>>>
>>>>> then I am getting below error :
>>>>>
>>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>>> SLF4J: Found binding in
>>>>> [jar:file:/mnt/yarn/usercache/hadoop/appcache/application_1587983834922_0002/filecache/10/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>> SLF4J: Found binding in
>>>>> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>>> explanation.
>>>>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>>>>> Exception in thread "main" java.lang.IllegalArgumentException: Invalid
>>>>> rule: /L
>>>>>       RULE:[2:$1@$0](.*@)s/@.*///L
>>>>>       DEFAULT
>>>>>         at
>>>>> org.apache.hadoop.security.authentication.util.KerberosName.parseRules(KerberosName.java:321)
>>>>>         at
>>>>> org.apache.hadoop.security.authentication.util.KerberosName.setRules(KerberosName.java:386)
>>>>>         at
>>>>> org.apache.hadoop.security.HadoopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>>>         at
>>>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:247)
>>>>>         at
>>>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232)
>>>>>         at
>>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:718)
>>>>>         at
>>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:703)
>>>>>         at
>>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:605)
>>>>>         at
>>>>> org.apache.flink.yarn.entrypoint.YarnEntrypointUtils.logYarnEnvironmentInformation(YarnEntrypointUtils.java:136)
>>>>>         at
>>>>> org.apache.flink.yarn.entrypoint.YarnJobClusterEntrypoint.main(YarnJobClusterEntrypoint.java:109)
>>>>>
>>>>>
>>>>> if I remove the  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar  from lib
>>>>> then i get below error:
>>>>>
>>>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>                       -  Classpath:
>>>>> /usr/lib/flink/lib/flink-table-blink_2.11-1.10.0.jar:/usr/lib/flink/lib/flink-table_2.11-1.10.0.jar:/usr/lib/flink/lib/log4j-1.2.17.jar:/usr/lib/flink/lib/slf4j-log4j12-1.7.15.jar:/usr/lib/flink/lib/flink-dist_2.11-1.10.0.jar::/etc/hadoop/conf:/etc/hadoop/conf
>>>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>                       -
>>>>> --------------------------------------------------------------------------------
>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: jobmanager.heap.size, 1024m
>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: taskmanager.memory.process.size, 1568m
>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: taskmanager.numberOfTaskSlots, 1
>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: parallelism.default, 1
>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: env.yarn.conf.dir, /etc/hadoop/conf
>>>>> 2020-04-27 16:59:37,300 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: env.hadoop.conf.dir, /etc/hadoop/conf
>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: jobmanager.execution.failover-strategy, region
>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: classloader.resolve-order, parent-first
>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: s3.access-key, AKIA52DD5QA5FC7HPKXG
>>>>> 2020-04-27 16:59:37,301 INFO
>>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>>> configuration property: s3.secret-key, ******
>>>>> 2020-04-27 16:59:37,305 WARN  org.apache.flink.client.cli.CliFrontend
>>>>>                       - Could not load CLI class
>>>>> org.apache.flink.yarn.cli.FlinkYarnSessionCli.
>>>>> java.lang.NoClassDefFoundError:
>>>>> org/apache/hadoop/yarn/exceptions/YarnException
>>>>>         at java.lang.Class.forName0(Native Method)
>>>>>         at java.lang.Class.forName(Class.java:264)
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1076)
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1030)
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:957)
>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>> org.apache.hadoop.yarn.exceptions.YarnException
>>>>>         at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
>>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:419)
>>>>>         at
>>>>> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
>>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:352)
>>>>>         ... 5 more
>>>>> 2020-04-27 16:59:37,406 INFO  org.apache.flink.core.fs.FileSystem
>>>>>                       - Hadoop is not in the classpath/dependencies. The
>>>>> extended set of supported File Systems via Hadoop is not available.
>>>>> 2020-04-27 16:59:37,458 INFO
>>>>>  org.apache.flink.runtime.security.modules.HadoopModuleFactory  - Cannot
>>>>> create Hadoop Security Module because Hadoop cannot be found in the
>>>>> Classpath.
>>>>> 2020-04-27 16:59:37,476 INFO
>>>>>  org.apache.flink.runtime.security.modules.JaasModule          - Jaas file
>>>>> will be created as /tmp/jaas-7054453135321774613.conf.
>>>>> 2020-04-27 16:59:37,480 INFO
>>>>>  org.apache.flink.runtime.security.SecurityUtils               - Cannot
>>>>> install HadoopSecurityContext because Hadoop cannot be found in the
>>>>> Classpath.
>>>>> 2020-04-27 16:59:37,481 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>                       - Running 'run' command.
>>>>> 2020-04-27 16:59:37,488 INFO  org.apache.flink.client.cli.CliFrontend
>>>>>                       - Building program from JAR file
>>>>> 2020-04-27 16:59:37,488 ERROR org.apache.flink.client.cli.CliFrontend
>>>>>                       - Invalid command line arguments.
>>>>> org.apache.flink.client.cli.CliArgsException: Could not build the
>>>>> program from JAR file.
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:203)
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>>>>>         at
>>>>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
>>>>> Caused by: java.io.FileNotFoundException: JAR file does not exist: -ynm
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:719)
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:695)
>>>>>         at
>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:200)
>>>>>         ... 4 more
>>>>>
>>>>>
>>>>> Can you please help.
>>>>>
>>>>> Thanks,
>>>>> Anuj
>>>>>
>>>>>
>>>>> On Mon, Apr 13, 2020 at 7:43 AM Yang Wang <da...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi Anuj,
>>>>>>
>>>>>> It seems that you are using hadoop version 2.4.1. I think "L" could
>>>>>> not be supported in
>>>>>> this version. Could you upgrade your hadoop version to 2.8 and have a
>>>>>> try? If your
>>>>>> YARN cluster version is 2.8+, then you could directly remove the
>>>>>> flink-shaded-hadoop
>>>>>> in your lib directory. Otherwise, you need to download the
>>>>>> flink-shaded-hadoop with
>>>>>> version 2.8 here[1].
>>>>>>
>>>>>>
>>>>>> [1]. https://flink.apache.org/downloads.html#additional-components
>>>>>>
>>>>>> Best,
>>>>>> Yang
>>>>>>
>>>>>> aj <aj...@gmail.com> 于2020年4月11日周六 上午4:21写道:
>>>>>>
>>>>>>> Hi Robert,
>>>>>>> attached the full application log file.
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Anuj
>>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> Thanks & Regards,
>>>>> Anuj Jain
>>>>> Mob. : +91- 8588817877
>>>>> Skype : anuj.jain07
>>>>> <http://www.oracle.com/>
>>>>>
>>>>>
>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>
>>>>
>>>
>>> --
>>> Thanks & Regards,
>>> Anuj Jain
>>> Mob. : +91- 8588817877
>>> Skype : anuj.jain07
>>> <http://www.oracle.com/>
>>>
>>>
>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>
>>
>>
>> --
>> Thanks & Regards,
>> Anuj Jain
>> Mob. : +91- 8588817877
>> Skype : anuj.jain07
>> <http://www.oracle.com/>
>>
>>
>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>
>

-- 
Thanks & Regards,
Anuj Jain
Mob. : +91- 8588817877
Skype : anuj.jain07
<http://www.oracle.com/>


<http://www.cse.iitm.ac.in/%7Eanujjain/>

Re: upgrade flink from 1.9.1 to 1.10.0 on EMR

Posted by Yang Wang <da...@gmail.com>.
Hi aj,

From the logs you have provided, the hadoop version is still 2.4.1.
Could you check the user jar(i.e. events-processor-1.0-SNAPSHOT.jar) have
some
hadoop classes? If it is, you need to exclude the hadoop dependency.


Best,
Yang

aj <aj...@gmail.com> 于2020年5月6日周三 下午3:38写道:

> Hello,
>
> Please help me upgrade to 1.10 in AWS EMR.
>
> On Fri, May 1, 2020 at 4:05 PM aj <aj...@gmail.com> wrote:
>
>> Hi Yang,
>>
>> I am attaching the logs for your reference, please help me what i am
>> doing wrong.
>>
>> Thanks,
>> Anuj
>>
>> On Wed, Apr 29, 2020 at 9:06 AM Yang Wang <da...@gmail.com> wrote:
>>
>>> Hi Anuj,
>>>
>>> I think the exception you come across still because the hadoop version
>>> is 2.4.1. I have checked the hadoop code, the code line are exactly same.
>>> For 2.8.1, i also have checked the ruleParse. It could work.
>>>
>>> /**
>>>  * A pattern for parsing a auth_to_local rule.
>>>  */
>>> private static final Pattern ruleParser =
>>>   Pattern.compile("\\s*((DEFAULT)|(RULE:\\[(\\d*):([^\\]]*)](\\(([^)]*)\\))?"+
>>>                   "(s/([^/]*)/([^/]*)/(g)?)?))/?(L)?");
>>>
>>>
>>> Could you share the jobmanager logs so that i could check the classpath
>>> and hadoop version?
>>>
>>> Best,
>>> Yang
>>>
>>> aj <aj...@gmail.com> 于2020年4月28日周二 上午1:01写道:
>>>
>>>> Hello Yang,
>>>> My Hadoop version is Hadoop 3.2.1-amzn-0
>>>> and I have put in flink/lib.
>>>>  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar
>>>>
>>>> then I am getting below error :
>>>>
>>>> SLF4J: Class path contains multiple SLF4J bindings.
>>>> SLF4J: Found binding in
>>>> [jar:file:/mnt/yarn/usercache/hadoop/appcache/application_1587983834922_0002/filecache/10/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>> SLF4J: Found binding in
>>>> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>>> explanation.
>>>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>>>> Exception in thread "main" java.lang.IllegalArgumentException: Invalid
>>>> rule: /L
>>>>       RULE:[2:$1@$0](.*@)s/@.*///L
>>>>       DEFAULT
>>>>         at
>>>> org.apache.hadoop.security.authentication.util.KerberosName.parseRules(KerberosName.java:321)
>>>>         at
>>>> org.apache.hadoop.security.authentication.util.KerberosName.setRules(KerberosName.java:386)
>>>>         at
>>>> org.apache.hadoop.security.HadoopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>>         at
>>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:247)
>>>>         at
>>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232)
>>>>         at
>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:718)
>>>>         at
>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:703)
>>>>         at
>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:605)
>>>>         at
>>>> org.apache.flink.yarn.entrypoint.YarnEntrypointUtils.logYarnEnvironmentInformation(YarnEntrypointUtils.java:136)
>>>>         at
>>>> org.apache.flink.yarn.entrypoint.YarnJobClusterEntrypoint.main(YarnJobClusterEntrypoint.java:109)
>>>>
>>>>
>>>> if I remove the  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar  from lib
>>>> then i get below error:
>>>>
>>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>>                     -  Classpath:
>>>> /usr/lib/flink/lib/flink-table-blink_2.11-1.10.0.jar:/usr/lib/flink/lib/flink-table_2.11-1.10.0.jar:/usr/lib/flink/lib/log4j-1.2.17.jar:/usr/lib/flink/lib/slf4j-log4j12-1.7.15.jar:/usr/lib/flink/lib/flink-dist_2.11-1.10.0.jar::/etc/hadoop/conf:/etc/hadoop/conf
>>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>>                     -
>>>> --------------------------------------------------------------------------------
>>>> 2020-04-27 16:59:37,300 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: jobmanager.heap.size, 1024m
>>>> 2020-04-27 16:59:37,300 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: taskmanager.memory.process.size, 1568m
>>>> 2020-04-27 16:59:37,300 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: taskmanager.numberOfTaskSlots, 1
>>>> 2020-04-27 16:59:37,300 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: parallelism.default, 1
>>>> 2020-04-27 16:59:37,300 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: env.yarn.conf.dir, /etc/hadoop/conf
>>>> 2020-04-27 16:59:37,300 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: env.hadoop.conf.dir, /etc/hadoop/conf
>>>> 2020-04-27 16:59:37,301 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: jobmanager.execution.failover-strategy, region
>>>> 2020-04-27 16:59:37,301 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: classloader.resolve-order, parent-first
>>>> 2020-04-27 16:59:37,301 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: s3.access-key, AKIA52DD5QA5FC7HPKXG
>>>> 2020-04-27 16:59:37,301 INFO
>>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>>> configuration property: s3.secret-key, ******
>>>> 2020-04-27 16:59:37,305 WARN  org.apache.flink.client.cli.CliFrontend
>>>>                     - Could not load CLI class
>>>> org.apache.flink.yarn.cli.FlinkYarnSessionCli.
>>>> java.lang.NoClassDefFoundError:
>>>> org/apache/hadoop/yarn/exceptions/YarnException
>>>>         at java.lang.Class.forName0(Native Method)
>>>>         at java.lang.Class.forName(Class.java:264)
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1076)
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1030)
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:957)
>>>> Caused by: java.lang.ClassNotFoundException:
>>>> org.apache.hadoop.yarn.exceptions.YarnException
>>>>         at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:419)
>>>>         at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
>>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:352)
>>>>         ... 5 more
>>>> 2020-04-27 16:59:37,406 INFO  org.apache.flink.core.fs.FileSystem
>>>>                     - Hadoop is not in the classpath/dependencies. The
>>>> extended set of supported File Systems via Hadoop is not available.
>>>> 2020-04-27 16:59:37,458 INFO
>>>>  org.apache.flink.runtime.security.modules.HadoopModuleFactory  - Cannot
>>>> create Hadoop Security Module because Hadoop cannot be found in the
>>>> Classpath.
>>>> 2020-04-27 16:59:37,476 INFO
>>>>  org.apache.flink.runtime.security.modules.JaasModule          - Jaas file
>>>> will be created as /tmp/jaas-7054453135321774613.conf.
>>>> 2020-04-27 16:59:37,480 INFO
>>>>  org.apache.flink.runtime.security.SecurityUtils               - Cannot
>>>> install HadoopSecurityContext because Hadoop cannot be found in the
>>>> Classpath.
>>>> 2020-04-27 16:59:37,481 INFO  org.apache.flink.client.cli.CliFrontend
>>>>                     - Running 'run' command.
>>>> 2020-04-27 16:59:37,488 INFO  org.apache.flink.client.cli.CliFrontend
>>>>                     - Building program from JAR file
>>>> 2020-04-27 16:59:37,488 ERROR org.apache.flink.client.cli.CliFrontend
>>>>                     - Invalid command line arguments.
>>>> org.apache.flink.client.cli.CliArgsException: Could not build the
>>>> program from JAR file.
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:203)
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>>>>         at
>>>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
>>>> Caused by: java.io.FileNotFoundException: JAR file does not exist: -ynm
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:719)
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:695)
>>>>         at
>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:200)
>>>>         ... 4 more
>>>>
>>>>
>>>> Can you please help.
>>>>
>>>> Thanks,
>>>> Anuj
>>>>
>>>>
>>>> On Mon, Apr 13, 2020 at 7:43 AM Yang Wang <da...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hi Anuj,
>>>>>
>>>>> It seems that you are using hadoop version 2.4.1. I think "L" could
>>>>> not be supported in
>>>>> this version. Could you upgrade your hadoop version to 2.8 and have a
>>>>> try? If your
>>>>> YARN cluster version is 2.8+, then you could directly remove the
>>>>> flink-shaded-hadoop
>>>>> in your lib directory. Otherwise, you need to download the
>>>>> flink-shaded-hadoop with
>>>>> version 2.8 here[1].
>>>>>
>>>>>
>>>>> [1]. https://flink.apache.org/downloads.html#additional-components
>>>>>
>>>>> Best,
>>>>> Yang
>>>>>
>>>>> aj <aj...@gmail.com> 于2020年4月11日周六 上午4:21写道:
>>>>>
>>>>>> Hi Robert,
>>>>>> attached the full application log file.
>>>>>>
>>>>>> Thanks,
>>>>>> Anuj
>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>>
>>>>>
>>>>
>>>> --
>>>> Thanks & Regards,
>>>> Anuj Jain
>>>> Mob. : +91- 8588817877
>>>> Skype : anuj.jain07
>>>> <http://www.oracle.com/>
>>>>
>>>>
>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>
>>>
>>
>> --
>> Thanks & Regards,
>> Anuj Jain
>> Mob. : +91- 8588817877
>> Skype : anuj.jain07
>> <http://www.oracle.com/>
>>
>>
>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>
>
>
> --
> Thanks & Regards,
> Anuj Jain
> Mob. : +91- 8588817877
> Skype : anuj.jain07
> <http://www.oracle.com/>
>
>
> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>

Re: upgrade flink from 1.9.1 to 1.10.0 on EMR

Posted by aj <aj...@gmail.com>.
Hello,

Please help me upgrade to 1.10 in AWS EMR.

On Fri, May 1, 2020 at 4:05 PM aj <aj...@gmail.com> wrote:

> Hi Yang,
>
> I am attaching the logs for your reference, please help me what i am doing
> wrong.
>
> Thanks,
> Anuj
>
> On Wed, Apr 29, 2020 at 9:06 AM Yang Wang <da...@gmail.com> wrote:
>
>> Hi Anuj,
>>
>> I think the exception you come across still because the hadoop version
>> is 2.4.1. I have checked the hadoop code, the code line are exactly same.
>> For 2.8.1, i also have checked the ruleParse. It could work.
>>
>> /**
>>  * A pattern for parsing a auth_to_local rule.
>>  */
>> private static final Pattern ruleParser =
>>   Pattern.compile("\\s*((DEFAULT)|(RULE:\\[(\\d*):([^\\]]*)](\\(([^)]*)\\))?"+
>>                   "(s/([^/]*)/([^/]*)/(g)?)?))/?(L)?");
>>
>>
>> Could you share the jobmanager logs so that i could check the classpath
>> and hadoop version?
>>
>> Best,
>> Yang
>>
>> aj <aj...@gmail.com> 于2020年4月28日周二 上午1:01写道:
>>
>>> Hello Yang,
>>> My Hadoop version is Hadoop 3.2.1-amzn-0
>>> and I have put in flink/lib.   flink-shaded-hadoop-2-uber-2.8.3-10.0.jar
>>>
>>> then I am getting below error :
>>>
>>> SLF4J: Class path contains multiple SLF4J bindings.
>>> SLF4J: Found binding in
>>> [jar:file:/mnt/yarn/usercache/hadoop/appcache/application_1587983834922_0002/filecache/10/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> SLF4J: Found binding in
>>> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
>>> explanation.
>>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
>>> Exception in thread "main" java.lang.IllegalArgumentException: Invalid
>>> rule: /L
>>>       RULE:[2:$1@$0](.*@)s/@.*///L
>>>       DEFAULT
>>>         at
>>> org.apache.hadoop.security.authentication.util.KerberosName.parseRules(KerberosName.java:321)
>>>         at
>>> org.apache.hadoop.security.authentication.util.KerberosName.setRules(KerberosName.java:386)
>>>         at
>>> org.apache.hadoop.security.HadoopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:247)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:718)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:703)
>>>         at
>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:605)
>>>         at
>>> org.apache.flink.yarn.entrypoint.YarnEntrypointUtils.logYarnEnvironmentInformation(YarnEntrypointUtils.java:136)
>>>         at
>>> org.apache.flink.yarn.entrypoint.YarnJobClusterEntrypoint.main(YarnJobClusterEntrypoint.java:109)
>>>
>>>
>>> if I remove the  flink-shaded-hadoop-2-uber-2.8.3-10.0.jar  from lib
>>> then i get below error:
>>>
>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>                     -  Classpath:
>>> /usr/lib/flink/lib/flink-table-blink_2.11-1.10.0.jar:/usr/lib/flink/lib/flink-table_2.11-1.10.0.jar:/usr/lib/flink/lib/log4j-1.2.17.jar:/usr/lib/flink/lib/slf4j-log4j12-1.7.15.jar:/usr/lib/flink/lib/flink-dist_2.11-1.10.0.jar::/etc/hadoop/conf:/etc/hadoop/conf
>>> 2020-04-27 16:59:37,293 INFO  org.apache.flink.client.cli.CliFrontend
>>>                     -
>>> --------------------------------------------------------------------------------
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: jobmanager.heap.size, 1024m
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: taskmanager.memory.process.size, 1568m
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: taskmanager.numberOfTaskSlots, 1
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: parallelism.default, 1
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: env.yarn.conf.dir, /etc/hadoop/conf
>>> 2020-04-27 16:59:37,300 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: env.hadoop.conf.dir, /etc/hadoop/conf
>>> 2020-04-27 16:59:37,301 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: jobmanager.execution.failover-strategy, region
>>> 2020-04-27 16:59:37,301 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: classloader.resolve-order, parent-first
>>> 2020-04-27 16:59:37,301 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: s3.access-key, AKIA52DD5QA5FC7HPKXG
>>> 2020-04-27 16:59:37,301 INFO
>>>  org.apache.flink.configuration.GlobalConfiguration            - Loading
>>> configuration property: s3.secret-key, ******
>>> 2020-04-27 16:59:37,305 WARN  org.apache.flink.client.cli.CliFrontend
>>>                     - Could not load CLI class
>>> org.apache.flink.yarn.cli.FlinkYarnSessionCli.
>>> java.lang.NoClassDefFoundError:
>>> org/apache/hadoop/yarn/exceptions/YarnException
>>>         at java.lang.Class.forName0(Native Method)
>>>         at java.lang.Class.forName(Class.java:264)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1076)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1030)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:957)
>>> Caused by: java.lang.ClassNotFoundException:
>>> org.apache.hadoop.yarn.exceptions.YarnException
>>>         at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:419)
>>>         at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
>>>         at java.lang.ClassLoader.loadClass(ClassLoader.java:352)
>>>         ... 5 more
>>> 2020-04-27 16:59:37,406 INFO  org.apache.flink.core.fs.FileSystem
>>>                     - Hadoop is not in the classpath/dependencies. The
>>> extended set of supported File Systems via Hadoop is not available.
>>> 2020-04-27 16:59:37,458 INFO
>>>  org.apache.flink.runtime.security.modules.HadoopModuleFactory  - Cannot
>>> create Hadoop Security Module because Hadoop cannot be found in the
>>> Classpath.
>>> 2020-04-27 16:59:37,476 INFO
>>>  org.apache.flink.runtime.security.modules.JaasModule          - Jaas file
>>> will be created as /tmp/jaas-7054453135321774613.conf.
>>> 2020-04-27 16:59:37,480 INFO
>>>  org.apache.flink.runtime.security.SecurityUtils               - Cannot
>>> install HadoopSecurityContext because Hadoop cannot be found in the
>>> Classpath.
>>> 2020-04-27 16:59:37,481 INFO  org.apache.flink.client.cli.CliFrontend
>>>                     - Running 'run' command.
>>> 2020-04-27 16:59:37,488 INFO  org.apache.flink.client.cli.CliFrontend
>>>                     - Building program from JAR file
>>> 2020-04-27 16:59:37,488 ERROR org.apache.flink.client.cli.CliFrontend
>>>                     - Invalid command line arguments.
>>> org.apache.flink.client.cli.CliArgsException: Could not build the
>>> program from JAR file.
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:203)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968)
>>>         at
>>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968)
>>> Caused by: java.io.FileNotFoundException: JAR file does not exist: -ynm
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:719)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:695)
>>>         at
>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:200)
>>>         ... 4 more
>>>
>>>
>>> Can you please help.
>>>
>>> Thanks,
>>> Anuj
>>>
>>>
>>> On Mon, Apr 13, 2020 at 7:43 AM Yang Wang <da...@gmail.com> wrote:
>>>
>>>> Hi Anuj,
>>>>
>>>> It seems that you are using hadoop version 2.4.1. I think "L" could not
>>>> be supported in
>>>> this version. Could you upgrade your hadoop version to 2.8 and have a
>>>> try? If your
>>>> YARN cluster version is 2.8+, then you could directly remove the
>>>> flink-shaded-hadoop
>>>> in your lib directory. Otherwise, you need to download the
>>>> flink-shaded-hadoop with
>>>> version 2.8 here[1].
>>>>
>>>>
>>>> [1]. https://flink.apache.org/downloads.html#additional-components
>>>>
>>>> Best,
>>>> Yang
>>>>
>>>> aj <aj...@gmail.com> 于2020年4月11日周六 上午4:21写道:
>>>>
>>>>> Hi Robert,
>>>>> attached the full application log file.
>>>>>
>>>>> Thanks,
>>>>> Anuj
>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>>>
>>>>
>>>
>>> --
>>> Thanks & Regards,
>>> Anuj Jain
>>> Mob. : +91- 8588817877
>>> Skype : anuj.jain07
>>> <http://www.oracle.com/>
>>>
>>>
>>> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>>>
>>
>
> --
> Thanks & Regards,
> Anuj Jain
> Mob. : +91- 8588817877
> Skype : anuj.jain07
> <http://www.oracle.com/>
>
>
> <http://www.cse.iitm.ac.in/%7Eanujjain/>
>


-- 
Thanks & Regards,
Anuj Jain
Mob. : +91- 8588817877
Skype : anuj.jain07
<http://www.oracle.com/>


<http://www.cse.iitm.ac.in/%7Eanujjain/>