使用Spark

时间:2015-11-21 02:09:37

标签: scala apache-spark apache-spark-sql

Spark 1.4.0及更高版本的源代码似乎表明此帖子的主题不可能(Spark特定格式除外)。

def saveAsTable(tableName: String): Unit = {

  * When the DataFrame is created from a non-partitioned [[HadoopFsRelation]] with a single input
   * path, and the data source provider can be mapped to an existing Hive builtin SerDe (i.e. ORC
   * and Parquet), the table is persisted in a Hive compatible format, which means other systems
   * like Hive will be able to read this table. Otherwise, the table is persisted in a Spark SQL
   * specific format.

我想知道是否有系统的解决方法。由于可伸缩性和性能原因,任何值得使用的hive表都将进行分区。因此,这是普通用例而非角落案例。

0 个答案:

没有答案