Apache spark 插入覆盖vs删除HDFS分区并在spark中写入拼花文件
拼花地板在使用过程中是否存在重大性能问题Apache spark 插入覆盖vs删除HDFS分区并在spark中写入拼花文件,apache-spark,hive,partitioning,Apache Spark,Hive,Partitioning,拼花地板在使用过程中是否存在重大性能问题 hiveContext.sql("INSERT OVERWRITE TABLE table_name PARTITION(dt='2016-06-08', country) , select x,y,z,country from DFTmpTable"); VS DeleteHDFSfile(/table/dt='2016-06-08') DF.write().mode("append").partitionBy("dt","country").pa
hiveContext.sql("INSERT OVERWRITE TABLE table_name PARTITION(dt='2016-06-08', country) , select x,y,z,country from DFTmpTable");
VS
DeleteHDFSfile(/table/dt='2016-06-08')
DF.write().mode("append").partitionBy("dt","country").parquet("/table/dt='2016-06-08'")