工作站节点无法在Windows

时间:2015-04-22 13:16:00

标签: apache-spark cluster-computing

步骤3:使用以下命令提取此文件                          Tar -xf spark-1.3.1-bin-hadoop2.4.tgz

步骤4:使用以下命令设置环境变量以创建SPARK环境

   SET HADOOP_HOME=C:\Hadoop
   SET SCALA_HOME =C:\scala
   SET SPARK_EXECUTOR_MEMORY =512m
   SET SPARK_HOME=F:\spark-1.3.1-bin-hadoop2.4
   SET SPARK_MASTER_IP =synclapn2881
   SET SPARK_WORKER_CORES =2 
   SET SPARK_WORKER_DIR=F:\work\sparkdata
   SET SPARK_WORKER_INSTANCES =4 
   SET SPARK_WORKER_MEMORY =1g
   SET Path=%SPARK_HOME%\bin;%Path%;

步骤5:使用以下命令启动主节点        spark-class org.apache.spark.deploy.master.Master

步骤6:使用以下命令启动从属节点        spark-class org.apache.spark.deploy.worker.Worker spark:// masternode:7077

注意:masternode是localhostname

实际结果

只启动了1个工作节点,但我设置了4个实例,

         SET SPARK_WORKER_INSTANCES =4 

参考下面的截图

SPARK WEB UI

   Expected Result
   Create 4 Worker nodes, as I had SET SPARK_WORKER_INSTANCES to 4 

感谢Advanvce

1 个答案:

答案 0 :(得分:0)

windows 中, 如果您需要4个工作节点,则必须在同一台机器或4台不同的机器上执行 STEP 6 4 times 使用以下命令spark-class org.apache.spark.deploy.worker.Worker spark://masternode:7077

启动从属节点

在Windows中不完全支持Spark。我试图在Windows上运行spark时遇到很多问题。 即使是start-slaves,start-master,start-slave脚本也无法在windows中运行。我认为最好切换到linux。

来自[spark documentation] [1]
Note: The launch scripts do not currently support Windows. To run a Spark cluster on Windows, start the master and workers by hand.

[1]:https://spark.apache.org/docs/latest/spark-standalone.html#Cluster启动脚本