You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by me...@apache.org on 2015/07/09 01:21:31 UTC
spark git commit: [SPARK-8068] [MLLIB] Add confusionMatrix method at
class MulticlassMetrics in pyspark/mllib
Repository: spark
Updated Branches:
refs/heads/master 4ffc27caa -> 381cb161b
[SPARK-8068] [MLLIB] Add confusionMatrix method at class MulticlassMetrics in pyspark/mllib
Add confusionMatrix method at class MulticlassMetrics in pyspark/mllib
Author: Yanbo Liang <yb...@gmail.com>
Closes #7286 from yanboliang/spark-8068 and squashes the following commits:
6109fe1 [Yanbo Liang] Add confusionMatrix method at class MulticlassMetrics in pyspark/mllib
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/381cb161
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/381cb161
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/381cb161
Branch: refs/heads/master
Commit: 381cb161ba4e3a30f2da3c4ef4ee19869d51f101
Parents: 4ffc27c
Author: Yanbo Liang <yb...@gmail.com>
Authored: Wed Jul 8 16:21:28 2015 -0700
Committer: Xiangrui Meng <me...@databricks.com>
Committed: Wed Jul 8 16:21:28 2015 -0700
----------------------------------------------------------------------
python/pyspark/mllib/evaluation.py | 11 +++++++++++
1 file changed, 11 insertions(+)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/381cb161/python/pyspark/mllib/evaluation.py
----------------------------------------------------------------------
diff --git a/python/pyspark/mllib/evaluation.py b/python/pyspark/mllib/evaluation.py
index c5cf3a4..f214037 100644
--- a/python/pyspark/mllib/evaluation.py
+++ b/python/pyspark/mllib/evaluation.py
@@ -152,6 +152,10 @@ class MulticlassMetrics(JavaModelWrapper):
>>> predictionAndLabels = sc.parallelize([(0.0, 0.0), (0.0, 1.0), (0.0, 0.0),
... (1.0, 0.0), (1.0, 1.0), (1.0, 1.0), (1.0, 1.0), (2.0, 2.0), (2.0, 0.0)])
>>> metrics = MulticlassMetrics(predictionAndLabels)
+ >>> metrics.confusionMatrix().toArray()
+ array([[ 2., 1., 1.],
+ [ 1., 3., 0.],
+ [ 0., 0., 1.]])
>>> metrics.falsePositiveRate(0.0)
0.2...
>>> metrics.precision(1.0)
@@ -186,6 +190,13 @@ class MulticlassMetrics(JavaModelWrapper):
java_model = java_class(df._jdf)
super(MulticlassMetrics, self).__init__(java_model)
+ def confusionMatrix(self):
+ """
+ Returns confusion matrix: predicted classes are in columns,
+ they are ordered by class label ascending, as in "labels".
+ """
+ return self.call("confusionMatrix")
+
def truePositiveRate(self, label):
"""
Returns true positive rate for a given label (category).
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org