我正在尝试在pycharm上运行pyspark。我已连接所有内容并设置环境变量。我可以读取sc.textFile但是当我尝试从pyspark.sql读取csv文件时出错了。
以下是代码:
import os
import sys
from pyspark import SparkContext
from pyspark import SparkConf
from pyspark.sql import SQLContext
from pyspark.sql import SparkSession
# Path for spark source folder
os.environ['SPARK_HOME']="E:/spark-2.0.0-bin-hadoop2.7/spark-2.0.0-bin-hadoop2.7"
# Append pyspark to Python Path
sys.path.append("E:/spark-2.0.0-bin-hadoop2.7/spark-2.0.0-bin-hadoop2.7/python")
sys.path.append("E:/spark-2.0.0-bin-hadoop2.7/spark-2.0.0-bin-hadoop2.7/python/lib/py4j-0.10.1.zip")
conf = SparkConf().setAppName('Simple App')
sc = SparkContext("local", "Simple App")
spark = SparkSession.builder.config(conf=SparkConf()).getOrCreate()
accounts_rdd = spark.read.csv('test.csv')
print accounts_rdd.show()
这是错误:
Traceback (most recent call last):
File "C:/Users/bjlinmanna/PycharmProjects/untitled1/spark.py", line 25, in <module>
accounts_rdd = spark.read.csv('pmec_close_position_order.csv')
File "E:\spark-2.0.0-bin-hadoop2.7\spark-2.0.0-bin-hadoop2.7\python\pyspark\sql\readwriter.py", line 363, in csv
return self._df(self._jreader.csv(self._spark._sc._jvm.PythonUtils.toSeq(path)))
File "E:\spark-2.0.0-bin-hadoop2.7\spark-2.0.0-bin-hadoop2.7\python\lib\py4j-0.10.1-src.zip\py4j\java_gateway.py", line 933, in __call__
File "E:\spark-2.0.0-bin-hadoop2.7\spark-2.0.0-bin-hadoop2.7\python\pyspark\sql\utils.py", line 79, in deco
raise IllegalArgumentException(s.split(': ', 1)[1], stackTrace)
pyspark.sql.utils.IllegalArgumentException: u'java.net.URISyntaxException: Relative path in absolute URI: file:C:/the/path/to/myfile/spark-warehouse'
感谢@Hyunsoo Park,我解决了我的问题如下:
spark = SparkSession.builder\
.master('local[*]')\
.appName('My App')\
.config('spark.sql.warehouse.dir', 'file:///C:/the/path/to/myfile')\
.getOrCreate()
accounts_rdd = spark.read\
.format('csv')\
.option('header', 'true')\
.load('file.csv')
设置配置时,请注意文件路径中的“//”。我不知道为什么当我设置'file:C:/ the / path / to / myfile'时,它不起作用
答案 0 :(得分:3)
这个链接可能很有帮助。
http://quabr.com/38669206/spark-2-0-relative-path-in-absolute-uri-spark-warehouse
简而言之,有一个配置选项spark.sql.warehouse.dir
来设置仓库文件夹。如果手动设置仓库文件夹,错误消息将消失。
我今天遇到了同样的问题。我在ubuntu 16.04中没有问题但是,当我在Windows 10中运行相同的代码时,火花显示错误信息就像你一样。可能无法在窗户中找到火花或制作仓库文件夹。