Spark中使用SparkSql进行shuffle操作,默认分区数是200个;参数配置是
--conf spark.sql.shuffle.partitions
Spark进行RDD操作,默认分区数看机器的配置,一般是cores*executors
--conf spark.default.parallelis
引自原文:https://blog.youkuaiyun.com/yolohohohoho/article/details/87967783
Spark中使用SparkSql进行shuffle操作,默认分区数是200个;参数配置是
--conf spark.sql.shuffle.partitions
Spark进行RDD操作,默认分区数看机器的配置,一般是cores*executors
--conf spark.default.parallelis
引自原文:https://blog.youkuaiyun.com/yolohohohoho/article/details/87967783