Flink on yarn:如何在本地客户端

时间:2018-03-22 08:31:55

标签: apache-flink

我尝试将我的应用程序部署到使用cli的flink上,不幸的是,它在以下情况下失败

java.lang.NoClassDefFoundError: Lredis/clients/jedis/JedisCluster;
    at java.lang.Class.getDeclaredFields0(Native Method)
    at java.lang.Class.privateGetDeclaredFields(Class.java:2583)
    at java.lang.Class.getDeclaredFields(Class.java:1916)
    at org.apache.flink.api.java.ClosureCleaner.clean(ClosureCleaner.java:72)
    at org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.clean(StreamExecutionEnvironment.java:1548)
    at org.apache.flink.streaming.api.datastream.DataStream.clean(DataStream.java:183)
    at org.apache.flink.streaming.api.datastream.DataStream.flatMap(DataStream.java:551)
    at org.apache.flink.streaming.api.scala.DataStream.flatMap(DataStream.scala:594)
    at com.hypers.hwt.realtime.top.HwtRealTimeTopRunner.executeLateStream(HwtRealTimeTop.scala:138)
    at com.hypers.hwt.realtime.top.HwtRealTimeTopRunner.run(HwtRealTimeTop.scala:72)
    at com.hypers.hwt.realtime.top.HwtRealTimeTop$.main(HwtRealTimeTop.scala:265)
    at com.hypers.hwt.realtime.top.HwtRealTimeTop.main(HwtRealTimeTop.scala)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:528)
    at org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:419)
    at org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:381)
    at org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:838)
    at org.apache.flink.client.CliFrontend.run(CliFrontend.java:259)
    at org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1086)
    at org.apache.flink.client.CliFrontend$2.call(CliFrontend.java:1133)
    at org.apache.flink.client.CliFrontend$2.call(CliFrontend.java:1130)
    at org.apache.flink.runtime.security.HadoopSecurityContext$1.run(HadoopSecurityContext.java:43)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:422)
    at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1709)
    at org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:40)
    at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1130)

我已经使用-yt参数来分发我的外部jar,但仍然失败。 实际上,flink提交工作有3个步骤:

  1. 在客户端
  2. 中包装代码并构建图表
  3. 客户向jobmanager提交工作
  4. jobmanager将工作分配给taskmanager
  5. 问题

    在长时间测试中,我发现此异常发生在step1中。并且step1由YarnClusterClient在本地运行。我知道这个问题将通过在$ FLINK_HOME / lib中添加我的外部jar来解决,但它会导致与其他应用程序的冲突

    期望

    所以我想知道是否有办法在本地添加外部jar类路径?

    Addtion

    class LateFlatMap(conf: FlinkJedisClusterConfig) extends RichFlatMapFunction[(PvAccBean, UvAccBean), Iterable[(String, Array[Byte])]] {
      var jedisCluster: JedisCluster = null
    
      override def open(properties: Configuration): Unit = {
       val genericObjectPoolConfig = new GenericObjectPoolConfig()
           genericObjectPoolConfig.setMaxIdle(conf.getMaxIdle())
           genericObjectPoolConfig.setMaxTotal(conf.getMaxTotal())
           genericObjectPoolConfig.setMinIdle(conf.getMinIdle())
       jedisCluster = new JedisCluster(conf.getNodes(), conf.getConnectionTimeout(),
       conf.getMaxRedirections(), genericObjectPoolConfig)
     }
    
     @Override
      override def close(): Unit = {
       jedisCluster.close()
     }
     ...
    } 
    

3 个答案:

答案 0 :(得分:0)

基本上我看到两种可能性:

  1. 通过构建一个胖罐子将第三方库添加到您的工作罐中。每个主要构建系统都可以这样做(例如Maven Assembly PluginSBT Assembly Plugin)。这将是我的首选解决方案
  2. 如果要将第三方库用于所有flink作业,可以在启动集群之前将它们添加到flinks jars目录。这也可以,但为您提供的灵活性较低。
  3. 希望有所帮助

答案 1 :(得分:0)

尝试将-C和-yt组合在一起的所有选项,将外部jar添加到类路径中,并添加到yarn.ship.directories中,但是在初始化mq连接工厂时失败。当放在flink lib中时,同样的工作。 仍然想知道这在2020年底是否行得通

答案 2 :(得分:0)

尝试使用

bin/start-scala-shell.sh local -a <full_external_jar_path>