Sqoop从couchbase导入到hadoop

时间:2014-04-11 11:49:31

标签: hadoop couchbase sqoop cloudera-cdh

在Ubuntu上,使用couchbase 2.5.1,cloudera cdh4,couchbase和oracle jdk的hadoop插件6.一切安装都很好(貌似),我可以独立使用hadoop和couchbase没有任何问题,但是当我尝试使用插件如下

sqoop import --connect http://127.0.0.1:8091/ --table DUMP

我收到以下错误

Please set $ACCUMULO_HOME to the root of your Accumulo installation.
14/04/11 11:44:08 INFO sqoop.Sqoop: Running Sqoop version: 1.4.3-cdh4.6.0
14/04/11 11:44:08 INFO tool.CodeGenTool: Beginning code generation
14/04/11 11:44:08 INFO orm.CompilationManager: HADOOP_MAPRED_HOME is /usr/lib/hadoop-0.20-mapreduce
Note: /tmp/sqoop-vagrant/compile/30e6774902d338663db059706cde5b12/DUMP.java uses or overrides a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
14/04/11 11:44:09 INFO orm.CompilationManager: Writing jar file: /tmp/sqoop-vagrant/compile/30e6774902d338663db059706cde5b12/DUMP.jar
14/04/11 11:44:09 INFO mapreduce.ImportJobBase: Beginning import of DUMP
14/04/11 11:44:09 WARN util.Jars: No such class couchbase doesn't use a jdbc driver available.
14/04/11 11:44:11 INFO ipc.Client: Retrying connect to server: localhost/127.0.0.1:8021. Already tried 0 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
14/04/11 11:44:12 INFO ipc.Client: Retrying connect to server: localhost/127.0.0.1:8021. Already tried 1 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)
14/04/11 11:44:13 INFO ipc.Client: Retrying connect to server: localhost/127.0.0.1:8021. Already tried 2 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1 SECONDS)

知道我哪里出错了吗?或者我能做些什么才能找到答案?

2 个答案:

答案 0 :(得分:1)

似乎我使用的语法错了。假设我们要将beer-sample存储桶从couchbase导入HDFS,下面是正确的语法,其中存储桶名称实际上是username传递的。

sqoop import --connect http://localhost:8091/pools --password password --username beer-sample --table DUMP

答案 1 :(得分:1)

我不认为您可以使用couchbase hadoop插件连接到带密码的couchbase存储桶。我曾经获得身份验证异常,但从未能解决它。我编辑了源代码,然后我就能使它工作。