You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Apache Spark (JIRA)" <ji...@apache.org> on 2016/09/16 04:50:20 UTC
[jira] [Commented] (SPARK-17559) PeriodicGraphCheckpointer didnot
persist edges as expected in some cases
[ https://issues.apache.org/jira/browse/SPARK-17559?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15495370#comment-15495370 ]
Apache Spark commented on SPARK-17559:
--------------------------------------
User 'dding3' has created a pull request for this issue:
https://github.com/apache/spark/pull/15116
> PeriodicGraphCheckpointer didnot persist edges as expected in some cases
> ------------------------------------------------------------------------
>
> Key: SPARK-17559
> URL: https://issues.apache.org/jira/browse/SPARK-17559
> Project: Spark
> Issue Type: Bug
> Components: MLlib
> Reporter: ding
> Priority: Minor
> Original Estimate: 1h
> Remaining Estimate: 1h
>
> When use PeriodicGraphCheckpointer to persist graph, sometimes the edge isn't persisted. As currently only when vertices's storage level is none, graph is persisted. However there is a chance vertices's storage level is not none while edges's is none. Eg. graph created by a outerJoinVertices operation, vertices is automatically cached while edges is not. In this way, edges will not be persisted if we use PeriodicGraphCheckpointer do persist.
> See below minimum example:
> val graphCheckpointer = new PeriodicGraphCheckpointer[Array[String], Int](2, sc)
> val users = sc.textFile("data/graphx/users.txt")
> .map(line => line.split(",")).map(parts => (parts.head.toLong, parts.tail))
> val followerGraph = GraphLoader.edgeListFile(sc, "data/graphx/followers.txt")
> val graph = followerGraph.outerJoinVertices(users) {
> case (uid, deg, Some(attrList)) => attrList
> case (uid, deg, None) => Array.empty[String]
> }
> graphCheckpointer.update(graph)
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org