我在一组单独的主机上有5个节点火花簇。我在一个单独的主机上安装了zeppeline,并连接了spark解释器来执行针对spark集群的查询。
Zeppeline 1.6版 - 安装在桌面上
我试过了两个。
添加了“export MASTER = spark:// sparkmasterhost:7077” 并将设置spark解释器主变量设置为“spark:// sparkmasterhost:7077”
当我运行“sc.version”时,我收到此错误。
org.apache.thrift.transport.TTransportException
at org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86)
at org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:429)
at org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:318)
at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:219)
at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69)
at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Client.recv_interpret(RemoteInterpreterService.java:220)
at org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Client.interpret(RemoteInterpreterService.java:205)
at org.apache.zeppelin.interpreter.remote.RemoteInterpreter.interpret(RemoteInterpreter.java:208)
at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93)
at org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:211)
at org.apache.zeppelin.scheduler.Job.run(Job.java:169)
at org.apache.zeppelin.scheduler.RemoteScheduler$JobRunner.run(RemoteScheduler.java:322)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:745)
我不确定发生了什么。
答案 0 :(得分:2)
Zeppelin中嵌入的spark版本很可能与spark群集的版本不同。打开http://<spark-master.url>:8080/
并检查左上角的版本 - 1.6.0
。然后下载Zeppelin Source并使用Spark版本标志zeppelin@<remote-host>:~/incubator-zeppelin$ mvn clean package -DskipTests -Pspark-1.6 -Dspark.version=1.6.0
在本地构建它。我刚刚克隆了git https://github.com/apache/incubator-zeppelin
。另一个非常相似的问题here