我正在尝试在Oracle Linux 7上设置HDP 3.1.0。
Ambari,HDFS和HIVE Metastore服务已经在运行,但是HiveServer2没有启动。
当我尝试手动启动它时:
# hive --service hiveserver2
等待几分钟后我得到了:
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/hdp/3.1.0.0-78/phoenix/phoenix-5.0.0.3.1.0.0-78-server.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/hdp/3.1.0.0-78/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
2018-12-15 14:15:28: Starting HiveServer2
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/usr/hdp/3.1.0.0-78/hive/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/usr/hdp/3.1.0.0-78/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
Hive Session ID = 99822aa9-957a-439e-904e-d9adce9a7893
Hive Session ID = fa34c442-4598-4f85-9493-daaf93804164
Hive Session ID = ed8de700-4ebf-4985-ae13-830b306be0e7
Hive Session ID = 6093d16b-53f0-4e21-9429-8046d3f3917a
Hive Session ID = a4fc572d-d56f-4c8a-97a0-8fc8bc115233
Hive Session ID = 02fdb753-45f7-4009-8283-bf3d5eef00b2
Hive Session ID = 47be06ad-42d2-4281-83f3-7e9b4cac1690
Hive Session ID = dae77692-3296-464f-995b-cb45a98d2e09
Hive Session ID = c4d49aa0-f829-4765-adbc-9afd5414775b
Hive Session ID = 8e26f8d8-bb01-4384-bfa2-8cb5ea66d1e8
这是netstat报告的内容:
# netstat -ntpl | egrep "10000|10001|10002"
tcp 0 0 192.168.1.100:10001 0.0.0.0:* LISTEN 422/java
tcp 0 0 192.168.1.100:10002 0.0.0.0:* LISTEN 26918/java
没人在端口10000上监听:(
这是我在/hive-site.xml中拥有的内容:
<property>
<name>hive.server2.thrift.port</name>
<value>10000</value>
</property>
<property>
<name>hive.server2.thrift.http.port</name>
<value>10001</value>
</property>
<property>
<name>hive.server2.webui.port</name>
<value>10002</value>
</property>
我认为我可以忽略SLF4J警告,对吗?我还应该检查什么?
答案 0 :(得分:0)
事实证明,此服务试图获取的内存超过默认 /etc/hadoop/conf/yarn-site.xml 设置允许的1024 MB:
<property>
<name>yarn.nodemanager.resource.memory-mb</name>
<value>1024</value>
</property>
我将该限制提高到1792,问题得到解决。