连接到 Kerberrized HDFS ,java.lang.IllegalArgumentException:无法指定服务器的 Kerberos 主体名称;

avi*_*til 3 java hadoop kerberos cloudera keytab

我正在尝试使用下面的代码连接到 Kerberized hdfs 集群,使用下面相同的代码我当然可以使用 HBaseConfiguration 访问 hbase,

Configuration config = new Configuration();
config.set("hadoop.security.authentication", "Kerberos");

UserGroupInformation.setConfiguration(config);
UserGroupInformation ugi = null;
ugi = UserGroupInformation.loginUserFromKeytabAndReturnUGI("me@EXAMPLE>COM","me.keytab");
model = ugi.doAs((PrivilegedExceptionAction<Map<String,Object>>) () -> { 
  testHadoop(hcb.gethDFSConfigBean());
  return null;
});
Run Code Online (Sandbox Code Playgroud)

我已经能够使用相同的密钥表和主体成功访问 Solr、Impala,但我遇到了这个奇怪的问题:无法找到 hdfs 的服务名称。

请查看下面的堆栈跟踪

java.io.IOException: Failed on local exception: java.io.IOException: java.lang.IllegalArgumentException: Failed to specify server's Kerberos principal name; Host Details : local host is: "Securonix-int3.local/10.0.4.36"; destination host is: "sobd189.securonix.com":8020; 
    at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:772)
    at org.apache.hadoop.ipc.Client.call(Client.java:1472)
    at org.apache.hadoop.ipc.Client.call(Client.java:1399)
    at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
    at com.sun.proxy.$Proxy9.getFileInfo(Unknown Source)
    at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:752)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:497)
    at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
    at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
    at com.sun.proxy.$Proxy10.getFileInfo(Unknown Source)
    at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1988)
    at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1118)
    at org.apache.hadoop.hdfs.DistributedFileSystem$18.doCall(DistributedFileSystem.java:1114)
    at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
    at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1114)
    at org.apache.hadoop.fs.FileSystem.exists(FileSystem.java:1400)
    at com.securonix.application.ui.uiUtil.SnyperUIUtil.lambda$main$4(SnyperUIUtil.java:1226)
    at com.securonix.application.ui.uiUtil.SnyperUIUtil$$Lambda$6/1620890840.run(Unknown Source)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:422)
    at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
    at com.securonix.application.ui.uiUtil.SnyperUIUtil.main(SnyperUIUtil.java:1216)
Caused by: java.io.IOException: java.lang.IllegalArgumentException: Failed to specify server's Kerberos principal name
    at org.apache.hadoop.ipc.Client$Connection$1.run(Client.java:680)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:422)
    at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
    at org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:643)
    at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:730)
    at org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:368)
    at org.apache.hadoop.ipc.Client.getConnection(Client.java:1521)
    at org.apache.hadoop.ipc.Client.call(Client.java:1438)
    ... 23 more
Caused by: java.lang.IllegalArgumentException: Failed to specify server's Kerberos principal name
    at org.apache.hadoop.security.SaslRpcClient.getServerPrincipal(SaslRpcClient.java:322)
    at org.apache.hadoop.security.SaslRpcClient.createSaslClient(SaslRpcClient.java:231)
    at org.apache.hadoop.security.SaslRpcClient.selectSaslClient(SaslRpcClient.java:159)
    at org.apache.hadoop.security.SaslRpcClient.saslConnect(SaslRpcClient.java:396)
    at org.apache.hadoop.ipc.Client$Connection.setupSaslConnection(Client.java:553)
    at org.apache.hadoop.ipc.Client$Connection.access$1800(Client.java:368)
    at org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:722)
    at org.apache.hadoop.ipc.Client$Connection$2.run(Client.java:718)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:422)
    at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
    at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:717)
Run Code Online (Sandbox Code Playgroud)

在我启用 Kerberos 的调试代码后,当我调用 FileSystem.get(); 时,我得到了以下调试日志;Kerberor 调试日志:

Java 配置名称:null Java 配置名称:null 本机配置名称:/etc/krb5.conf 本机配置名称:/etc/krb5.conf 从本机配置加载 从本机配置加载 16/02/22 15:53:14 WARN util .NativeCodeLoader:无法为您的平台加载本机hadoop库...在适用的情况下使用内置java类Java
配置名称:null Java配置名称:null本机配置名称:/etc/krb5.conf本机配置名称:/etc/ krb5.conf 从本机配置加载 从本机配置加载

KdcAccessibility:重置 >>> KdcAccessibility:重置 KdcAccessibility:重置 >>> KdcAccessibility:重置 KeyTabInputStream,readName():EXAMPLE.COM >>> KeyTabInputStream,readName():EXAMPLE.COM KeyTabInputStream,readName():securonix >>> KeyTabInputStream , readName(): securonix KeyTab: load() 条目长度: 55; 类型:23 >>> KeyTab:load()条目长度:55;类型:23 KeyTabInputStream,readName():EXAMPLE.COM >>> KeyTabInputStream,readName():EXAMPLE.COM KeyTabInputStream,readName():securonix >>> KeyTabInputStream,readName():securonix KeyTab:load()条目长度:71 ; 类型:18 >>> KeyTab:load()条目长度:71;类型: 18 正在寻找密钥: securonix@EXAMPLE.COM 正在寻找密钥: securonix@EXAMPLE.COM 已添加密钥: 18version: 1 已添加密钥: 18version: 1 已添加密钥: 23version: 1 已添加密钥: 23version: 1 正在查找密钥for: securonix@EXAMPLE.COM 寻找密钥: securonix@EXAMPLE.COM 添加密钥: 18version: 1 添加密钥: 18version: 1 添加密钥: 23version: 1 添加密钥: 23version: 1 default_tkt_enctypes 的默认 etypes: 18 18 16。 default_tkt_enctypes 的默认 etypes:18 18 16. KrbAsReq 创建消息 >>> KrbAsReq 创建消息 KrbKdcReq 发送:kdc=sobd189.securonix.com TCP:88,超时=30000,重试次数 =3,#bytes=139 >>> KrbKdcReq发送:kdc=sobd189.securonix.com TCP:88,超时=30000,重试次数=3,#bytes=139 KDCCommunication:kdc=sobd189.securonix.com TCP:88,超时=30000,尝试=1,#bytes =139 >>> KDCCommunication:kdc=sobd189.securonix.com TCP:88,超时=30000,尝试=1,#bytes=139 调试:TCPClient 读取 639 字节 >>>DEBUG:TCPClient 读取 639 字节 KrbKdcReq 发送:#bytes read=639 >>> KrbKdcReq 发送:#bytes read=639 KdcAccessibility:删除 sobd189.securonix.com >>> KdcAccessibility:删除 sobd189.securonix.com 寻找密钥:securonix@EXAMPLE.COM 寻找密钥:securonix@ EXAMPLE.COM 添加密钥:18版本:1 添加密钥:18版本:1 添加密钥:23版本:1 添加密钥:23版本:1 EType:sun.security.krb5.internal.crypto.Aes256CtsHmacSha1EType >>> EType:sun.security.krb5 KrbAsReq.getReply securonix 中的 .internal.crypto.Aes256CtsHmacSha1EType KrbAsRep 缺点

有趣的是,当我使用 hdfs.exists() 等文件系统的 api 时

 >>>KinitOptions cache name is /tmp/krb5cc_501
 >> Acquire default native Credentials
 default etypes for default_tkt_enctypes: 18 18 16.
 >>> Found no TGT's in LSA
Run Code Online (Sandbox Code Playgroud)

Bry*_*nde 5

我认为问题在于 HDFS 期望配置具有 dfs.datanode.kerberos.principal 的值,这是数据节点的主体,但在本例中缺少该值。

当我仅从 core-site.xml 创建配置实例并忘记添加 hdfs-site.xml 时,我遇到了同样的问题。一旦我添加 hdfs-site.xml 它就开始工作,并且 hdfs-site.xml 具有:

 <property>
      <name>dfs.datanode.kerberos.principal</name>
      <value>....</value>
 </property>
Run Code Online (Sandbox Code Playgroud)

希望这可以帮助。