You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by pw...@apache.org on 2014/01/24 04:12:10 UTC

[4/4] git commit: Merge pull request #502 from pwendell/clone-1

Merge pull request #502 from pwendell/clone-1

Remove Hadoop object cloning and warn users making Hadoop RDD's.

The code introduced in #359 used Hadoop's WritableUtils.clone() to
duplicate objects when reading from Hadoop files. Some users have
reported exceptions when cloning data in various file formats,
including Avro and another custom format.

This patch removes that functionality to ensure stability for the
0.9 release. Instead, it puts a clear warning in the documentation
that copying may be necessary for Hadoop data sets.


Project: http://git-wip-us.apache.org/repos/asf/incubator-spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-spark/commit/c3196171
Tree: http://git-wip-us.apache.org/repos/asf/incubator-spark/tree/c3196171
Diff: http://git-wip-us.apache.org/repos/asf/incubator-spark/diff/c3196171

Branch: refs/heads/master
Commit: c3196171f3dffde6c9e67e3d35c398a01fbba846
Parents: cad3002 268ecbd
Author: Patrick Wendell <pw...@gmail.com>
Authored: Thu Jan 23 19:11:59 2014 -0800
Committer: Patrick Wendell <pw...@gmail.com>
Committed: Thu Jan 23 19:11:59 2014 -0800

----------------------------------------------------------------------
 .../scala/org/apache/spark/SparkContext.scala   | 127 ++++++++------
 .../spark/api/java/JavaSparkContext.scala       | 165 ++++++-------------
 .../scala/org/apache/spark/rdd/HadoopRDD.scala  |  28 +---
 .../org/apache/spark/rdd/NewHadoopRDD.scala     |  24 +--
 .../scala/org/apache/spark/util/Utils.scala     |  22 ---
 5 files changed, 137 insertions(+), 229 deletions(-)
----------------------------------------------------------------------