You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Tao Li (JIRA)" <ji...@apache.org> on 2015/12/07 13:44:11 UTC

[jira] [Created] (SPARK-12179) Spark SQL get different result with the same code

Tao Li created SPARK-12179:
------------------------------

             Summary: Spark SQL get different result with the same code
                 Key: SPARK-12179
                 URL: https://issues.apache.org/jira/browse/SPARK-12179
             Project: Spark
          Issue Type: Bug
          Components: Spark Core, SQL
    Affects Versions: 1.5.2, 1.5.1, 1.5.0, 1.4.1, 1.4.0, 1.3.1, 1.3.0, 1.3.2, 1.4.2, 1.5.3
         Environment: hadoop version: 2.5.0-cdh5.3.2
spark version: 1.5.3
run mode: yarn-client
            Reporter: Tao Li


I run the sql in yarn-client mode, but get different result each time.

As you can see the example, I get the different shuffle write with the same shuffle read in two jobs.

Some of my spark app runs well, but some always met this problem. And I met this problem on spark 1.3, 1.4 and 1.5 version.

Can you git me some suggestion about possible causes or how to figure out the problem?

1. First Run
Details for Stage 9 (Attempt 0)
Total Time Across All Tasks: 5.8 min
Shuffle Read: 24.4 MB / 205399
Shuffle Write: 6.8 MB / 54934

2. Second Run
Details for Stage 9 (Attempt 0)
Total Time Across All Tasks: 5.6 min
Shuffle Read: 24.4 MB / 205399
Shuffle Write: 6.8 MB / 54905



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org