You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@spark.apache.org by Chang Ya-Hsuan <su...@gmail.com> on 2015/12/23 17:09:38 UTC
value of sc.defaultParallelism
python version: 2.7.9
os: ubuntu 14.04
spark: 1.5.2
I run a standalone spark on localhost, and use the following code to access
sc.defaultParallism
# a.py
import pyspark
sc = pyspark.SparkContext()
print(sc.defaultParallelism)
and use the following command to submit
$ spark-submit --master spark://yahsuan-vm:7077 a.py
it prints 2, however, my spark web page shows I got 4 cores
according to http://spark.apache.org/docs/latest/configuration.html
spark.default.parallelismFor distributed shuffle operations likereduceByKey
and join, the largest number of partitions in a parent RDD. For operations
likeparallelize with no parent RDDs, it depends on the cluster manager:
- Local mode: number of cores on the local machine
- Mesos fine grained mode: 8
- Others: total number of cores on all executor nodes or 2, whichever is
larger
Default number of partitions in RDDs returned by transformations like join,
reduceByKey, andparallelize when not set by user.
It seems I should get 4 rather than 2.
Am I misunderstood the document?
--
-- 張雅軒