警告会话:创建到/xxx.xxx.xxx.xxx:28730的池时出错

时间:2018-08-23 10:32:31

标签: apache-spark ibm-cloud compose scylla analytics-engine

我试图从运行在IBM Analytics Engine上的Spark 2.3连接到运行在IBM Cloud上的ScyllaDB数据库。

我正在像这样启动火花壳...

$ spark-shell --master local[1] \
       --files jaas.conf \
       --packages org.apache.spark:spark-sql-kafka-0-10_2.11:2.3.0,datastax:spark-cassandra-connector:2.3.0-s_2.11,commons-configuration:commons-configuration:1.10 \
       --conf "spark.driver.extraJavaOptions=-Djava.security.auth.login.config=jaas.conf" \
       --conf "spark.executor.extraJavaOptions=-Djava.security.auth.login.config=jaas.conf" \
       --conf spark.cassandra.connection.host=xxx1.composedb.com,xxx2.composedb.com,xxx3.composedb.com \
       --conf spark.cassandra.connection.port=28730 \
       --conf spark.cassandra.auth.username=scylla \
       --conf spark.cassandra.auth.password=SECRET \
       --conf spark.cassandra.connection.ssl.enabled=true \
       --num-executors 1  \
       --executor-cores 1 

然后执行以下spark scala代码:

import com.datastax.spark.connector._
import org.apache.spark.sql.cassandra._

val stocksRdd = sc.cassandraTable("stocks", "stocks")

stocksRdd.count()

但是,我看到了一堆警告:

18/08/23 10:11:01 WARN Cluster: You listed xxx1.composedb.com/xxx.xxx.xxx.xxx:28730 in your contact points, but it wasn't found in the control host's system.peers at startup
18/08/23 10:11:01 WARN Cluster: You listed xxx1.composedb.com/xxx.xxx.xxx.xxx:28730 in your contact points, but it wasn't found in the control host's system.peers at startup
18/08/23 10:11:06 WARN Session: Error creating pool to /xxx.xxx.xxx.xxx:28730
com.datastax.driver.core.exceptions.ConnectionException: [/xxx.xxx.xxx.xxx:28730] Pool was closed during initialization
...

但是,在警告中的stacktrace之后,我看到了期望的输出:

res2: Long = 4 

如果我导航到撰写用户界面,则会看到一个地图json:

[
  {"xxx.xxx.xxx.xxx:9042":"xxx1.composedb.com:28730"},
  {"xxx.xxx.xxx.xxx:9042":"xxx2.composedb.com:28730"},
  {"xxx.xxx.xxx.xxx:9042":"xxx3.composedb.com:28730"}
]

警告似乎与地图文件有关。

警告的含义是什么?我可以忽略它吗?


注意::我见过类似的question,但是由于映射文件,我相信这个问题有所不同,而且我无法控制scylladb群集的设置方式撰写。

1 个答案:

答案 0 :(得分:5)

这只是警告。之所以会发出警告,是因为Scylla本身并不知道火花试图到达的IP。显然,Spark正在连接到群集并正在检索预期的信息,所以您应该没事。