MapReduce作业未在HADOOP 2.6.0(多节点群集)上运行

时间:2018-07-31 11:09:31

标签: hadoop mapreduce yarn hadoop2

我已经在4台机器(1个主节点和3个从属节点)上成功完成了Hadoop 2.6.0多节点集群设置。但是,当我尝试在群集上运行简单的字数统计作业时,它就会卡住。

它被卡在这里:

:~$ hadoop jar ~/MY_MAP_JARS/wordcount_f2.jar /input/crime /output/cc1
18/07/31 02:25:04 INFO client.RMProxy: Connecting to ResourceManager at master/192.168.10.150:8032
18/07/31 02:25:06 WARN mapreduce.JobSubmitter: Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this.
18/07/31 02:25:10 INFO input.FileInputFormat: Total input paths to process : 1
18/07/31 02:25:11 INFO mapreduce.JobSubmitter: number of splits:1
18/07/31 02:25:11 INFO Configuration.deprecation: mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
18/07/31 02:25:12 INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1533029038340_0001
18/07/31 02:25:13 INFO impl.YarnClientImpl: Submitted application application_1533029038340_0001
18/07/31 02:25:14 INFO mapreduce.Job: The url to track the job: http://master:8088/proxy/application_1533029038340_0001/
18/07/31 02:25:14 INFO mapreduce.Job: Running job: job_1533029038340_0001

有时它停留在地图阶段:

18/07/31 02:27:53 INFO mapreduce.Job: Running job: job_1533029038340_0003
18/07/31 02:28:00 INFO mapreduce.Job: Job job_1533029038340_0003 running in uber mode : false
18/07/31 02:28:00 INFO mapreduce.Job:  map 0% reduce 0% 

有时地图阶段已完成,但reduce阶段未运行:

18/07/31 02:26:01 INFO mapreduce.Job: Running job: job_1533029038340_0002
18/07/31 02:26:12 INFO mapreduce.Job: Job job_1533029038340_0002 running in uber mode : false
18/07/31 02:26:12 INFO mapreduce.Job:  map 0% reduce 0%
18/07/31 02:26:22 INFO mapreduce.Job:  map 100% reduce 0%

集群详细信息:

在母版上:

RAM:6 GB | 硬盘:20 GB | 核心:4 | 守护进程:NameNode,资源管理器,辅助NN(默认)

在从站上:(所有3个从站节点都具有相同的配置)

RAM:4 GB | 硬盘:20 GB | 得分:4 | 守护程序:数据节点和节点管理器

我的hadoop conf文件:(我将下面提到的配置复制到了我所有的节点(主节点和从节点)。

core-site.xml:

<configuration>
 <property>
 <name>fs.default.name</name>
  <value>hdfs://master:50000</value>
 </property>
</configuration>

hdfs-site.xml:

<property>
  <name>dfs.replication</name>
  <value>2</value>
</property>

<property>
  <name>dfs.name.dir</name>
  <value>file:/home/aroon/hadoop_store260/name-dir</value>
</property>

<property>
 <name>dfs.data.dir</name>
 <value>file:/home/aroon/hadoop_store260/data-dir</value>
</property>

mapred-site.xml:

<property>
  <name>mapreduce.framework.name</name>
  <value>yarn</value>
</property>

yarn-site.xml:

<property>
  <name>yarn.resourcemanager.hostname</name>
  <value>master</value>
</property>

<property>
  <name>yarn.nodemanager.aux-services</name>
  <value>mapreduce_shuffle</value>
</property>

<property>
  <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
  <value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>

奴隶

slave1
slave2
slave3

任何人都可以清楚地解决此问题,因为提出和回答了类似的问题,并且大多数问题难以理解。如果有人可以清楚地回答这个问题,那么对像我这样的新手可能会有帮助。

预先感谢您的回答。

PS:集装箱日志(stderr)

请在此处找到容器错误日志:https://pastebin.com/VpzfU7Dr

0 个答案:

没有答案