使用Kerberos安全性连接到Mapper中的HIVE

时间:2015-10-21 09:16:24

标签: hadoop mapreduce hive kerberos oozie

我的目标是在安全集群(kerberos)HDP2.3上的Oozie工作流程调度程序中连接到Hive运行MapReduce。

我可以直接连接到配置单元,或者当我使用以下连接字符串作为java app(yarn jar)运行它时:

DriverManager.getConnection("jdbc:hive2://host:10000/;principal=hive/_HOST@REALM", "", "");

但是当我在Mapper中运行它时,它失败了。

 ERROR [main] org.apache.thrift.transport.TSaslTransport: SASL negotiation failure
    javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]
        at com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:212)
        at org.apache.thrift.transport.TSaslClientTransport.handleSaslStartMessage(TSaslClientTransport.java:94)
        at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:253)
        at org.apache.thrift.transport.TSaslClientTransport.open(TSaslClientTransport.java:37)
        at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:52)
        at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:49)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
        at org.apache.hadoop.hive.thrift.client.TUGIAssumingTransport.open(TUGIAssumingTransport.java:49)
        at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:190)
        ...
Caused by: GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)
        at sun.security.jgss.krb5.Krb5InitCredential.getInstance(Krb5InitCredential.java:147)
        at sun.security.jgss.krb5.Krb5MechFactory.getCredentialElement(Krb5MechFactory.java:121)
        at sun.security.jgss.krb5.Krb5MechFactory.getMechanismContext(Krb5MechFactory.java:187)
        at sun.security.jgss.GSSManagerImpl.getMechanismContext(GSSManagerImpl.java:223)
        at sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:212)
        at sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:179)
        at com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:193)

如何让它在Mapper中运行?

2 个答案:

答案 0 :(得分:1)

它与hive委托令牌一起使用:

  1. Oozie的

    • 添加属性:

      hive2.server.principal=hive/_HOST@REALM
      hive2.jdbc.url=jdbc:hive2://{host}:10000/default
      
    • 将凭据设置为 hive2

    • Mapper示例:

      public class HiveMapperExample extends Mapper<LongWritable, Text, Text, Text> {
      
          @Override
          protected void map(LongWritable key, Text value, Context context) throws IOException, InterruptedException {
              try {
                  Class.forName("org.apache.hive.jdbc.HiveDriver");
                  Connection connect = DriverManager.getConnection("jdbc:hive2://{host}:10000/;auth=delegationToken", "", "");
                  Statement state = connect.createStatement();
                  ResultSet resultSet = state.executeQuery("select * from some_table");
                  while (resultSet.next()) {
                      ...
                  }
              } catch (Exception e) {
                  ...
              }
          }
       }
      
  2. ToolRunner

    public class HiveTestApplication extends Configured implements Tool {
    
        public static void main(String[] args) throws Exception {
            System.exit(ToolRunner.run(new HiveTestApplication(), args));
        }
    
        @Override
        public int run(String[] args) throws Exception {
            Configuration conf = new Configuration();
            //set your conf
            Job job = Job.getInstance(conf);
            job.setMapperClass(HiveMapperExample.class);
    
            addHiveDelegationToken(job.getCredentials(), "jdbc:hive2://{host}:10000/", "hive/_HOST@REALM");
    
            job.waitForCompletion(true);
    
            return 0;
        }
    
    
        public void addHiveDelegationToken(Credentials creds, String url, String principal) throws Exception {
            Class.forName("org.apache.hive.jdbc.HiveDriver");
    
            Connection con = DriverManager.getConnection(url + ";principal=" + principal);
            // get delegation token for the given proxy user
            String tokenStr = ((HiveConnection) con).getDelegationToken(UserGroupInformation.getCurrentUser().getShortUserName(), principal);
            con.close();
    
            Token<DelegationTokenIdentifier> hive2Token = new Token<>();
            hive2Token.decodeFromUrlString(tokenStr);
            creds.addToken(new Text("hive.server2.delegation.token"), hive2Token);
            creds.addToken(new Text(HiveAuthFactory.HS2_CLIENT_TOKEN), hive2Token);
        }
    }
    

答案 1 :(得分:0)

错误“无法找到任何Kerberos tgt”非常简单。在连接到正在执行映射器的节点上的Hive之前,您需要获取kerberos票证。