-bash-4.2$ spark-shell --conf spark.sql.sources.partitionOverwriteMode=dynamic \
> --conf spark.hadoop.hive.exec.dynamic.partition.mode=nonstrict \
> --conf spark.sql.orc.impl=native \
> --conf spark.sql.orc.enableVectorizedReader=true \
> --conf spark.hadoop.hive.exec.max.dynamic.partitions=2550
Observe spark.hadoop.* is placed for all the hadoop configurations
If we pass only --conf hive.exec.max.dynamic.partitions=2550 then this property will be ignored by spark-shell in 2.3.1 version
We can also pass memory config as below
> --conf spark.hadoop.hive.exec.dynamic.partition.mode=nonstrict \
> --conf spark.sql.orc.impl=native \
> --conf spark.sql.orc.enableVectorizedReader=true \
> --conf spark.hadoop.hive.exec.max.dynamic.partitions=2550
Observe spark.hadoop.* is placed for all the hadoop configurations
If we pass only --conf hive.exec.max.dynamic.partitions=2550 then this property will be ignored by spark-shell in 2.3.1 version
We can also pass memory config as below
spark-shell --driver-memory 10G --executor-memory 15G --executor-cores 8
No comments:
Post a Comment