Spark 1.4.0及更高版本的源代码似乎表明此帖子的主题不可能(Spark特定格式除外)。
def saveAsTable(tableName: String): Unit = {
* When the DataFrame is created from a non-partitioned [[HadoopFsRelation]] with a single input
* path, and the data source provider can be mapped to an existing Hive builtin SerDe (i.e. ORC
* and Parquet), the table is persisted in a Hive compatible format, which means other systems
* like Hive will be able to read this table. Otherwise, the table is persisted in a Spark SQL
* specific format.
我想知道是否有系统的解决方法。由于可伸缩性和性能原因,任何值得使用的hive表都将进行分区。因此,这是普通用例而非角落案例。