Spark / Zeppelin 0.8.1-为什么pyspark解释器无法正常使用?

时间:2019-03-15 09:09:03

标签: apache-spark kubernetes pyspark apache-zeppelin

我正在尝试在默认的Zeppelin教程中运行%spark.pyspark解释器,但在下面出现错误。我在做什么错了?

# In the Zeppelin 0.8.1 Tutorial this is %pyspark as written
%spark.pyspark
import os


java.lang.RuntimeException: No pyspark.zip found under /opt/spark/python/lib
    at org.apache.zeppelin.spark.PythonUtils.sparkPythonPath(PythonUtils.java:51)
    at org.apache.zeppelin.spark.PySparkInterpreter.setupPySparkEnv(PySparkInterpreter.java:218)
    at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:291)
    at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:194)
    at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
    at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:616)
    at org.apache.zeppelin.scheduler.Job.run(Job.java:188)
    at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:140)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.run(FutureTask.java:266)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
    at java.lang.Thread.run(Thread.java:748)

这有点像pyspark interpreter not found in apache zeppelin,它的答案对我不起作用。

此外,我正在通过https://github.com/SnappyDataInc/spark-on-k8s/tree/chart_upgrade_2.4/charts/zeppelin-with-spark在Kubernetes上运行:\

0 个答案:

没有答案