尝试初始化集群以使用DistCp时,Hadoop API引发错误

时间:2018-07-24 13:20:56

标签: java hadoop hdfs bigdata cloudera

我正在尝试使用Hadoop API通过DistCp类进行分布式副本,但是在尝试连接到集群时会引发错误。我试图更改hadoop和hdfs的配置文件,但似乎不起作用。我正在最新的Cloudera快速入门上测试该应用程序

我运行此命令以执行该类。 java -cp myjar com.keedio.hadoop.Mover

package com.keedio.hadoop;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.*;
import org.apache.hadoop.tools.DistCp;

import org.apache.hadoop.tools.DistCpOptionSwitch;
import org.apache.hadoop.tools.DistCpOptions;
import org.apache.hadoop.util.ToolRunner;

import java.util.Collections;


public class Mover {

public static void main(String args []) {

 try {
  FileSystem fs = FileSystem.getLocal(new Configuration());
  FileSystem fs2= FileSystem.get(java.net.URI.create("file:///"),new 
  Configuration());
  DistCpOptions distCpOptions=new DistCpOptions(new Path("file:/Users/jvasquez/Desktop/ficheros1"),new Path("file:/Users/jvasquez/Desktop/ficheros2"));
String argumentos [] = {"file:////Users/jvasquez/Desktop/ficheros1","file:///Users/jvasquez/Desktop/ficheros2"};


Configuration conf=new Configuration();

conf.addResource(new org.apache.hadoop.fs.Path("/etc/hadoop/conf/yarn-site.xml"));
conf.addResource(new org.apache.hadoop.fs.Path("/etc/hadoop/conf/core-site.xml"));
conf.addResource(new org.apache.hadoop.fs.Path("/etc/hadoop/conf/mapred-site.xml"));
conf.addResource(new Path("/etc/hadoop/conf/hdfs-site.xml"));

DistCp distCp= new DistCp(conf,distCpOptions);

ToolRunner.run(distCp,argumentos);




}
 catch (Exception e) {
 e.printStackTrace();

}


}

}

这是错误

log4j:WARN No appenders could be found for logger 
(org.apache.hadoop.metrics2.lib.MutableMetricsFactory).
log4j:WARN Please initialize the log4j system properly.
log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
java.io.IOException: Cannot initialize Cluster. Please check your configuration for mapreduce.framework.name and the correspond server addresses.
at org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:143)
at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:108)
at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:101)
at org.apache.hadoop.tools.DistCp.createMetaFolderPath(DistCp.java:471)
at org.apache.hadoop.tools.DistCp.<init>(DistCp.java:107)
at com.keedio.hadoop.Mover.main(Mover.java:48)

1 个答案:

答案 0 :(得分:0)

最后我解决了它,结果是我只需要使用命令hadoop jar将应用程序作为mapreduce应用程序启动。