无法读取pyspark中的csv文件

时间:2018-10-25 09:46:07

标签: pyspark

我是pyspark的新手,我做了一些初始教程。当我尝试使用Jupyter Notebook在Spark框架中的本地主机上加载CSV文件时,弹出以下错误。我的Java版本是8.0

int port = 48569;
                        try {
                            DatagramSocket s = new DatagramSocket();
                            InetAddress local = InetAddress.getByName("10.3.22.218");
                            int msg_length = msg.length();
                            byte[] message = msg.getBytes();
                            DatagramPacket p = new DatagramPacket(message, msg_length, local, port);
                            s.send(p);
                        }catch (SocketException e) {
                            e.printStackTrace();
                        }catch (UnknownHostException e) {
                            e.printStackTrace();
                        }catch(IOException e) {
                            e.printStackTrace();
                        }

我的错误如下:

from pyspark.sql import SparkSession
spark = SparkSession \
    .builder \
    .appName('sql based spark data analysis') \
    .config('spark.some.config.option', 'some-value') \
    .getOrCreate()
df = spark.read.csv('C:/Users/sitaram/Downloads/creditcardfraud/creditcard.csv')

1 个答案:

答案 0 :(得分:0)

请尝试C://Users//sitaram//Downloads//creditcardfraud//creditcard.csv