You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by sr...@apache.org on 2018/01/03 17:31:36 UTC
spark git commit: [SPARK-22896] Improvement in String interpolation
Repository: spark
Updated Branches:
refs/heads/master a66fe36ce -> 9a2b65a3c
[SPARK-22896] Improvement in String interpolation
## What changes were proposed in this pull request?
* String interpolation in ml pipeline example has been corrected as per scala standard.
## How was this patch tested?
* manually tested.
Author: chetkhatri <ck...@gmail.com>
Closes #20070 from chetkhatri/mllib-chetan-contrib.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/9a2b65a3
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/9a2b65a3
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/9a2b65a3
Branch: refs/heads/master
Commit: 9a2b65a3c0c36316aae0a53aa0f61c5044c2ceff
Parents: a66fe36
Author: chetkhatri <ck...@gmail.com>
Authored: Wed Jan 3 11:31:32 2018 -0600
Committer: Sean Owen <so...@cloudera.com>
Committed: Wed Jan 3 11:31:32 2018 -0600
----------------------------------------------------------------------
.../spark/examples/ml/JavaQuantileDiscretizerExample.java | 2 +-
.../org/apache/spark/examples/SimpleSkewedGroupByTest.scala | 4 ----
.../scala/org/apache/spark/examples/graphx/Analytics.scala | 6 ++++--
.../org/apache/spark/examples/graphx/SynthBenchmark.scala | 6 +++---
.../org/apache/spark/examples/ml/ChiSquareTestExample.scala | 6 +++---
.../org/apache/spark/examples/ml/CorrelationExample.scala | 4 ++--
.../org/apache/spark/examples/ml/DataFrameExample.scala | 4 ++--
.../examples/ml/DecisionTreeClassificationExample.scala | 4 ++--
.../spark/examples/ml/DecisionTreeRegressionExample.scala | 4 ++--
.../org/apache/spark/examples/ml/DeveloperApiExample.scala | 6 +++---
.../spark/examples/ml/EstimatorTransformerParamExample.scala | 6 +++---
.../examples/ml/GradientBoostedTreeClassifierExample.scala | 4 ++--
.../examples/ml/GradientBoostedTreeRegressorExample.scala | 4 ++--
.../MulticlassLogisticRegressionWithElasticNetExample.scala | 2 +-
.../examples/ml/MultilayerPerceptronClassifierExample.scala | 2 +-
.../org/apache/spark/examples/ml/NaiveBayesExample.scala | 2 +-
.../spark/examples/ml/QuantileDiscretizerExample.scala | 4 ++--
.../spark/examples/ml/RandomForestClassifierExample.scala | 4 ++--
.../spark/examples/ml/RandomForestRegressorExample.scala | 4 ++--
.../org/apache/spark/examples/ml/VectorIndexerExample.scala | 4 ++--
.../spark/examples/mllib/AssociationRulesExample.scala | 6 +++---
.../examples/mllib/BinaryClassificationMetricsExample.scala | 4 ++--
.../examples/mllib/DecisionTreeClassificationExample.scala | 4 ++--
.../spark/examples/mllib/DecisionTreeRegressionExample.scala | 4 ++--
.../org/apache/spark/examples/mllib/FPGrowthExample.scala | 2 +-
.../mllib/GradientBoostingClassificationExample.scala | 4 ++--
.../examples/mllib/GradientBoostingRegressionExample.scala | 4 ++--
.../spark/examples/mllib/HypothesisTestingExample.scala | 2 +-
.../spark/examples/mllib/IsotonicRegressionExample.scala | 2 +-
.../org/apache/spark/examples/mllib/KMeansExample.scala | 2 +-
.../scala/org/apache/spark/examples/mllib/LBFGSExample.scala | 2 +-
.../examples/mllib/LatentDirichletAllocationExample.scala | 8 +++++---
.../examples/mllib/LinearRegressionWithSGDExample.scala | 2 +-
.../scala/org/apache/spark/examples/mllib/PCAExample.scala | 4 ++--
.../apache/spark/examples/mllib/PMMLModelExportExample.scala | 2 +-
.../org/apache/spark/examples/mllib/PrefixSpanExample.scala | 4 ++--
.../examples/mllib/RandomForestClassificationExample.scala | 4 ++--
.../spark/examples/mllib/RandomForestRegressionExample.scala | 4 ++--
.../apache/spark/examples/mllib/RecommendationExample.scala | 2 +-
.../org/apache/spark/examples/mllib/SVMWithSGDExample.scala | 2 +-
.../org/apache/spark/examples/mllib/SimpleFPGrowth.scala | 8 +++-----
.../spark/examples/mllib/StratifiedSamplingExample.scala | 4 ++--
.../org/apache/spark/examples/mllib/TallSkinnyPCA.scala | 2 +-
.../org/apache/spark/examples/mllib/TallSkinnySVD.scala | 2 +-
.../org/apache/spark/examples/streaming/CustomReceiver.scala | 6 +++---
.../org/apache/spark/examples/streaming/RawNetworkGrep.scala | 2 +-
.../examples/streaming/RecoverableNetworkWordCount.scala | 8 ++++----
.../examples/streaming/clickstream/PageViewGenerator.scala | 4 ++--
.../examples/streaming/clickstream/PageViewStream.scala | 4 ++--
49 files changed, 94 insertions(+), 96 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/java/org/apache/spark/examples/ml/JavaQuantileDiscretizerExample.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaQuantileDiscretizerExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaQuantileDiscretizerExample.java
index dd20cac..43cc30c1 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaQuantileDiscretizerExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaQuantileDiscretizerExample.java
@@ -66,7 +66,7 @@ public class JavaQuantileDiscretizerExample {
.setNumBuckets(3);
Dataset<Row> result = discretizer.fit(df).transform(df);
- result.show();
+ result.show(false);
// $example off$
spark.stop();
}
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/SimpleSkewedGroupByTest.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/SimpleSkewedGroupByTest.scala b/examples/src/main/scala/org/apache/spark/examples/SimpleSkewedGroupByTest.scala
index e64dcbd..2332a66 100644
--- a/examples/src/main/scala/org/apache/spark/examples/SimpleSkewedGroupByTest.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/SimpleSkewedGroupByTest.scala
@@ -60,10 +60,6 @@ object SimpleSkewedGroupByTest {
pairs1.count
println(s"RESULT: ${pairs1.groupByKey(numReducers).count}")
- // Print how many keys each reducer got (for debugging)
- // println("RESULT: " + pairs1.groupByKey(numReducers)
- // .map{case (k,v) => (k, v.size)}
- // .collectAsMap)
spark.stop()
}
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala b/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala
index 92936bd..815404d 100644
--- a/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala
@@ -145,9 +145,11 @@ object Analytics extends Logging {
// TriangleCount requires the graph to be partitioned
.partitionBy(partitionStrategy.getOrElse(RandomVertexCut)).cache()
val triangles = TriangleCount.run(graph)
- println("Triangles: " + triangles.vertices.map {
+ val triangleTypes = triangles.vertices.map {
case (vid, data) => data.toLong
- }.reduce(_ + _) / 3)
+ }.reduce(_ + _) / 3
+
+ println(s"Triangles: ${triangleTypes}")
sc.stop()
case _ =>
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala b/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala
index 6d2228c..57b2edf 100644
--- a/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala
@@ -52,7 +52,7 @@ object SynthBenchmark {
arg =>
arg.dropWhile(_ == '-').split('=') match {
case Array(opt, v) => (opt -> v)
- case _ => throw new IllegalArgumentException("Invalid argument: " + arg)
+ case _ => throw new IllegalArgumentException(s"Invalid argument: $arg")
}
}
@@ -76,7 +76,7 @@ object SynthBenchmark {
case ("sigma", v) => sigma = v.toDouble
case ("degFile", v) => degFile = v
case ("seed", v) => seed = v.toInt
- case (opt, _) => throw new IllegalArgumentException("Invalid option: " + opt)
+ case (opt, _) => throw new IllegalArgumentException(s"Invalid option: $opt")
}
val conf = new SparkConf()
@@ -86,7 +86,7 @@ object SynthBenchmark {
val sc = new SparkContext(conf)
// Create the graph
- println(s"Creating graph...")
+ println("Creating graph...")
val unpartitionedGraph = GraphGenerators.logNormalGraph(sc, numVertices,
numEPart.getOrElse(sc.defaultParallelism), mu, sigma, seed)
// Repartition the graph
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/ChiSquareTestExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/ChiSquareTestExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/ChiSquareTestExample.scala
index dcee1e4..5146fd0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/ChiSquareTestExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/ChiSquareTestExample.scala
@@ -52,9 +52,9 @@ object ChiSquareTestExample {
val df = data.toDF("label", "features")
val chi = ChiSquareTest.test(df, "features", "label").head
- println("pValues = " + chi.getAs[Vector](0))
- println("degreesOfFreedom = " + chi.getSeq[Int](1).mkString("[", ",", "]"))
- println("statistics = " + chi.getAs[Vector](2))
+ println(s"pValues = ${chi.getAs[Vector](0)}")
+ println(s"degreesOfFreedom ${chi.getSeq[Int](1).mkString("[", ",", "]")}")
+ println(s"statistics ${chi.getAs[Vector](2)}")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/CorrelationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/CorrelationExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/CorrelationExample.scala
index 3f57dc3..d7f1fc8 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/CorrelationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/CorrelationExample.scala
@@ -51,10 +51,10 @@ object CorrelationExample {
val df = data.map(Tuple1.apply).toDF("features")
val Row(coeff1: Matrix) = Correlation.corr(df, "features").head
- println("Pearson correlation matrix:\n" + coeff1.toString)
+ println(s"Pearson correlation matrix:\n $coeff1")
val Row(coeff2: Matrix) = Correlation.corr(df, "features", "spearman").head
- println("Spearman correlation matrix:\n" + coeff2.toString)
+ println(s"Spearman correlation matrix:\n $coeff2")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala
index 0658bdd..ee4469f 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DataFrameExample.scala
@@ -47,7 +47,7 @@ object DataFrameExample {
val parser = new OptionParser[Params]("DataFrameExample") {
head("DataFrameExample: an example app using DataFrame for ML.")
opt[String]("input")
- .text(s"input path to dataframe")
+ .text("input path to dataframe")
.action((x, c) => c.copy(input = x))
checkConfig { params =>
success
@@ -93,7 +93,7 @@ object DataFrameExample {
// Load the records back.
println(s"Loading Parquet file with UDT from $outputDir.")
val newDF = spark.read.parquet(outputDir)
- println(s"Schema from Parquet:")
+ println("Schema from Parquet:")
newDF.printSchema()
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
index bc6d327..276ceda 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
@@ -83,10 +83,10 @@ object DecisionTreeClassificationExample {
.setPredictionCol("prediction")
.setMetricName("accuracy")
val accuracy = evaluator.evaluate(predictions)
- println("Test Error = " + (1.0 - accuracy))
+ println(s"Test Error = ${(1.0 - accuracy)}")
val treeModel = model.stages(2).asInstanceOf[DecisionTreeClassificationModel]
- println("Learned classification tree model:\n" + treeModel.toDebugString)
+ println(s"Learned classification tree model:\n ${treeModel.toDebugString}")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
index ee61200..aaaecae 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
@@ -73,10 +73,10 @@ object DecisionTreeRegressionExample {
.setPredictionCol("prediction")
.setMetricName("rmse")
val rmse = evaluator.evaluate(predictions)
- println("Root Mean Squared Error (RMSE) on test data = " + rmse)
+ println(s"Root Mean Squared Error (RMSE) on test data = $rmse")
val treeModel = model.stages(1).asInstanceOf[DecisionTreeRegressionModel]
- println("Learned regression tree model:\n" + treeModel.toDebugString)
+ println(s"Learned regression tree model:\n ${treeModel.toDebugString}")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala
index d94d837..2dc11b0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DeveloperApiExample.scala
@@ -53,7 +53,7 @@ object DeveloperApiExample {
// Create a LogisticRegression instance. This instance is an Estimator.
val lr = new MyLogisticRegression()
// Print out the parameters, documentation, and any default values.
- println("MyLogisticRegression parameters:\n" + lr.explainParams() + "\n")
+ println(s"MyLogisticRegression parameters:\n ${lr.explainParams()}")
// We may set parameters using setter methods.
lr.setMaxIter(10)
@@ -169,10 +169,10 @@ private class MyLogisticRegressionModel(
Vectors.dense(-margin, margin)
}
- /** Number of classes the label can take. 2 indicates binary classification. */
+ // Number of classes the label can take. 2 indicates binary classification.
override val numClasses: Int = 2
- /** Number of features the model was trained on. */
+ // Number of features the model was trained on.
override val numFeatures: Int = coefficients.size
/**
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala
index f18d86e..e5d91f1 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/EstimatorTransformerParamExample.scala
@@ -46,7 +46,7 @@ object EstimatorTransformerParamExample {
// Create a LogisticRegression instance. This instance is an Estimator.
val lr = new LogisticRegression()
// Print out the parameters, documentation, and any default values.
- println("LogisticRegression parameters:\n" + lr.explainParams() + "\n")
+ println(s"LogisticRegression parameters:\n ${lr.explainParams()}\n")
// We may set parameters using setter methods.
lr.setMaxIter(10)
@@ -58,7 +58,7 @@ object EstimatorTransformerParamExample {
// we can view the parameters it used during fit().
// This prints the parameter (name: value) pairs, where names are unique IDs for this
// LogisticRegression instance.
- println("Model 1 was fit using parameters: " + model1.parent.extractParamMap)
+ println(s"Model 1 was fit using parameters: ${model1.parent.extractParamMap}")
// We may alternatively specify parameters using a ParamMap,
// which supports several methods for specifying parameters.
@@ -73,7 +73,7 @@ object EstimatorTransformerParamExample {
// Now learn a new model using the paramMapCombined parameters.
// paramMapCombined overrides all parameters set earlier via lr.set* methods.
val model2 = lr.fit(training, paramMapCombined)
- println("Model 2 was fit using parameters: " + model2.parent.extractParamMap)
+ println(s"Model 2 was fit using parameters: ${model2.parent.extractParamMap}")
// Prepare test data.
val test = spark.createDataFrame(Seq(
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
index 3656773..ef78c0a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
@@ -86,10 +86,10 @@ object GradientBoostedTreeClassifierExample {
.setPredictionCol("prediction")
.setMetricName("accuracy")
val accuracy = evaluator.evaluate(predictions)
- println("Test Error = " + (1.0 - accuracy))
+ println(s"Test Error = ${1.0 - accuracy}")
val gbtModel = model.stages(2).asInstanceOf[GBTClassificationModel]
- println("Learned classification GBT model:\n" + gbtModel.toDebugString)
+ println(s"Learned classification GBT model:\n ${gbtModel.toDebugString}")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
index e53aab7..3feb234 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
@@ -73,10 +73,10 @@ object GradientBoostedTreeRegressorExample {
.setPredictionCol("prediction")
.setMetricName("rmse")
val rmse = evaluator.evaluate(predictions)
- println("Root Mean Squared Error (RMSE) on test data = " + rmse)
+ println(s"Root Mean Squared Error (RMSE) on test data = $rmse")
val gbtModel = model.stages(1).asInstanceOf[GBTRegressionModel]
- println("Learned regression GBT model:\n" + gbtModel.toDebugString)
+ println(s"Learned regression GBT model:\n ${gbtModel.toDebugString}")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/MulticlassLogisticRegressionWithElasticNetExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/MulticlassLogisticRegressionWithElasticNetExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/MulticlassLogisticRegressionWithElasticNetExample.scala
index 42f0ace..3e61dbe 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/MulticlassLogisticRegressionWithElasticNetExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/MulticlassLogisticRegressionWithElasticNetExample.scala
@@ -48,7 +48,7 @@ object MulticlassLogisticRegressionWithElasticNetExample {
// Print the coefficients and intercept for multinomial logistic regression
println(s"Coefficients: \n${lrModel.coefficientMatrix}")
- println(s"Intercepts: ${lrModel.interceptVector}")
+ println(s"Intercepts: \n${lrModel.interceptVector}")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
index 6fce82d..646f46a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
@@ -66,7 +66,7 @@ object MultilayerPerceptronClassifierExample {
val evaluator = new MulticlassClassificationEvaluator()
.setMetricName("accuracy")
- println("Test set accuracy = " + evaluator.evaluate(predictionAndLabels))
+ println(s"Test set accuracy = ${evaluator.evaluate(predictionAndLabels)}")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala
index bd9fcc4..50c70c6 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/NaiveBayesExample.scala
@@ -52,7 +52,7 @@ object NaiveBayesExample {
.setPredictionCol("prediction")
.setMetricName("accuracy")
val accuracy = evaluator.evaluate(predictions)
- println("Test set accuracy = " + accuracy)
+ println(s"Test set accuracy = $accuracy")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
index aedb9e7..0fe16fb 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
@@ -36,7 +36,7 @@ object QuantileDiscretizerExample {
// Output of QuantileDiscretizer for such small datasets can depend on the number of
// partitions. Here we force a single partition to ensure consistent results.
// Note this is not necessary for normal use cases
- .repartition(1)
+ .repartition(1)
// $example on$
val discretizer = new QuantileDiscretizer()
@@ -45,7 +45,7 @@ object QuantileDiscretizerExample {
.setNumBuckets(3)
val result = discretizer.fit(df).transform(df)
- result.show()
+ result.show(false)
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
index 5eafda8..6265f83 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
@@ -85,10 +85,10 @@ object RandomForestClassifierExample {
.setPredictionCol("prediction")
.setMetricName("accuracy")
val accuracy = evaluator.evaluate(predictions)
- println("Test Error = " + (1.0 - accuracy))
+ println(s"Test Error = ${(1.0 - accuracy)}")
val rfModel = model.stages(2).asInstanceOf[RandomForestClassificationModel]
- println("Learned classification forest model:\n" + rfModel.toDebugString)
+ println(s"Learned classification forest model:\n ${rfModel.toDebugString}")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
index 9a0a001..2679fcb 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
@@ -72,10 +72,10 @@ object RandomForestRegressorExample {
.setPredictionCol("prediction")
.setMetricName("rmse")
val rmse = evaluator.evaluate(predictions)
- println("Root Mean Squared Error (RMSE) on test data = " + rmse)
+ println(s"Root Mean Squared Error (RMSE) on test data = $rmse")
val rfModel = model.stages(1).asInstanceOf[RandomForestRegressionModel]
- println("Learned regression forest model:\n" + rfModel.toDebugString)
+ println(s"Learned regression forest model:\n ${rfModel.toDebugString}")
// $example off$
spark.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
index afa761a..96bb8ea 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
@@ -41,8 +41,8 @@ object VectorIndexerExample {
val indexerModel = indexer.fit(data)
val categoricalFeatures: Set[Int] = indexerModel.categoryMaps.keys.toSet
- println(s"Chose ${categoricalFeatures.size} categorical features: " +
- categoricalFeatures.mkString(", "))
+ println(s"Chose ${categoricalFeatures.size} " +
+ s"categorical features: ${categoricalFeatures.mkString(", ")}")
// Create new column "indexed" with categorical values transformed to indices
val indexedData = indexerModel.transform(data)
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala
index ff44de5..a07535b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala
@@ -42,9 +42,8 @@ object AssociationRulesExample {
val results = ar.run(freqItemsets)
results.collect().foreach { rule =>
- println("[" + rule.antecedent.mkString(",")
- + "=>"
- + rule.consequent.mkString(",") + "]," + rule.confidence)
+ println(s"[${rule.antecedent.mkString(",")}=>${rule.consequent.mkString(",")} ]" +
+ s" ${rule.confidence}")
}
// $example off$
@@ -53,3 +52,4 @@ object AssociationRulesExample {
}
// scalastyle:on println
+
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala
index b9263ac..c6312d7 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala
@@ -86,7 +86,7 @@ object BinaryClassificationMetricsExample {
// AUPRC
val auPRC = metrics.areaUnderPR
- println("Area under precision-recall curve = " + auPRC)
+ println(s"Area under precision-recall curve = $auPRC")
// Compute thresholds used in ROC and PR curves
val thresholds = precision.map(_._1)
@@ -96,7 +96,7 @@ object BinaryClassificationMetricsExample {
// AUROC
val auROC = metrics.areaUnderROC
- println("Area under ROC = " + auROC)
+ println(s"Area under ROC = $auROC")
// $example off$
sc.stop()
}
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala
index b50b459..c2f89b7 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala
@@ -55,8 +55,8 @@ object DecisionTreeClassificationExample {
(point.label, prediction)
}
val testErr = labelAndPreds.filter(r => r._1 != r._2).count().toDouble / testData.count()
- println("Test Error = " + testErr)
- println("Learned classification tree model:\n" + model.toDebugString)
+ println(s"Test Error = $testErr")
+ println(s"Learned classification tree model:\n ${model.toDebugString}")
// Save and load model
model.save(sc, "target/tmp/myDecisionTreeClassificationModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala
index 2af45af..1ecf642 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala
@@ -54,8 +54,8 @@ object DecisionTreeRegressionExample {
(point.label, prediction)
}
val testMSE = labelsAndPredictions.map{ case (v, p) => math.pow(v - p, 2) }.mean()
- println("Test Mean Squared Error = " + testMSE)
- println("Learned regression tree model:\n" + model.toDebugString)
+ println(s"Test Mean Squared Error = $testMSE")
+ println(s"Learned regression tree model:\n ${model.toDebugString}")
// Save and load model
model.save(sc, "target/tmp/myDecisionTreeRegressionModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala
index 6435abc..f724ee1 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala
@@ -74,7 +74,7 @@ object FPGrowthExample {
println(s"Number of frequent itemsets: ${model.freqItemsets.count()}")
model.freqItemsets.collect().foreach { itemset =>
- println(itemset.items.mkString("[", ",", "]") + ", " + itemset.freq)
+ println(s"${itemset.items.mkString("[", ",", "]")}, ${itemset.freq}")
}
sc.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala
index 00bb334..3c56e19 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala
@@ -54,8 +54,8 @@ object GradientBoostingClassificationExample {
(point.label, prediction)
}
val testErr = labelAndPreds.filter(r => r._1 != r._2).count.toDouble / testData.count()
- println("Test Error = " + testErr)
- println("Learned classification GBT model:\n" + model.toDebugString)
+ println(s"Test Error = $testErr")
+ println(s"Learned classification GBT model:\n ${model.toDebugString}")
// Save and load model
model.save(sc, "target/tmp/myGradientBoostingClassificationModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala
index d8c2634..c288bf2 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala
@@ -53,8 +53,8 @@ object GradientBoostingRegressionExample {
(point.label, prediction)
}
val testMSE = labelsAndPredictions.map{ case(v, p) => math.pow((v - p), 2)}.mean()
- println("Test Mean Squared Error = " + testMSE)
- println("Learned regression GBT model:\n" + model.toDebugString)
+ println(s"Test Mean Squared Error = $testMSE")
+ println(s"Learned regression GBT model:\n ${model.toDebugString}")
// Save and load model
model.save(sc, "target/tmp/myGradientBoostingRegressionModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/HypothesisTestingExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/HypothesisTestingExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/HypothesisTestingExample.scala
index 0d391a3..add1719 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/HypothesisTestingExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/HypothesisTestingExample.scala
@@ -68,7 +68,7 @@ object HypothesisTestingExample {
// against the label.
val featureTestResults: Array[ChiSqTestResult] = Statistics.chiSqTest(obs)
featureTestResults.zipWithIndex.foreach { case (k, v) =>
- println("Column " + (v + 1).toString + ":")
+ println(s"Column ${(v + 1)} :")
println(k)
} // summary of the test
// $example off$
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala
index 4aee951..a10d6f0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala
@@ -56,7 +56,7 @@ object IsotonicRegressionExample {
// Calculate mean squared error between predicted and real labels.
val meanSquaredError = predictionAndLabel.map { case (p, l) => math.pow((p - l), 2) }.mean()
- println("Mean Squared Error = " + meanSquaredError)
+ println(s"Mean Squared Error = $meanSquaredError")
// Save and load model
model.save(sc, "target/tmp/myIsotonicRegressionModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/KMeansExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/KMeansExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/KMeansExample.scala
index c4d71d8..b0a6f16 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/KMeansExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/KMeansExample.scala
@@ -43,7 +43,7 @@ object KMeansExample {
// Evaluate clustering by computing Within Set Sum of Squared Errors
val WSSSE = clusters.computeCost(parsedData)
- println("Within Set Sum of Squared Errors = " + WSSSE)
+ println(s"Within Set Sum of Squared Errors = $WSSSE")
// Save and load model
clusters.save(sc, "target/org/apache/spark/KMeansExample/KMeansModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala
index fedcefa..123782f 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala
@@ -82,7 +82,7 @@ object LBFGSExample {
println("Loss of each step in training process")
loss.foreach(println)
- println("Area under ROC = " + auROC)
+ println(s"Area under ROC = $auROC")
// $example off$
sc.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/LatentDirichletAllocationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/LatentDirichletAllocationExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/LatentDirichletAllocationExample.scala
index f2c8ec0..d25962c 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/LatentDirichletAllocationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/LatentDirichletAllocationExample.scala
@@ -42,11 +42,13 @@ object LatentDirichletAllocationExample {
val ldaModel = new LDA().setK(3).run(corpus)
// Output topics. Each is a distribution over words (matching word count vectors)
- println("Learned topics (as distributions over vocab of " + ldaModel.vocabSize + " words):")
+ println(s"Learned topics (as distributions over vocab of ${ldaModel.vocabSize} words):")
val topics = ldaModel.topicsMatrix
for (topic <- Range(0, 3)) {
- print("Topic " + topic + ":")
- for (word <- Range(0, ldaModel.vocabSize)) { print(" " + topics(word, topic)); }
+ print(s"Topic $topic :")
+ for (word <- Range(0, ldaModel.vocabSize)) {
+ print(s"${topics(word, topic)}")
+ }
println()
}
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegressionWithSGDExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegressionWithSGDExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegressionWithSGDExample.scala
index d399618..449b725 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegressionWithSGDExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegressionWithSGDExample.scala
@@ -52,7 +52,7 @@ object LinearRegressionWithSGDExample {
(point.label, prediction)
}
val MSE = valuesAndPreds.map{ case(v, p) => math.pow((v - p), 2) }.mean()
- println("training Mean Squared Error = " + MSE)
+ println(s"training Mean Squared Error $MSE")
// Save and load model
model.save(sc, "target/tmp/scalaLinearRegressionWithSGDModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/PCAExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/PCAExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/PCAExample.scala
index eb36697..eff2393 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/PCAExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/PCAExample.scala
@@ -65,8 +65,8 @@ object PCAExample {
val MSE = valuesAndPreds.map { case (v, p) => math.pow((v - p), 2) }.mean()
val MSE_pca = valuesAndPreds_pca.map { case (v, p) => math.pow((v - p), 2) }.mean()
- println("Mean Squared Error = " + MSE)
- println("PCA Mean Squared Error = " + MSE_pca)
+ println(s"Mean Squared Error = $MSE")
+ println(s"PCA Mean Squared Error = $MSE_pca")
// $example off$
sc.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/PMMLModelExportExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/PMMLModelExportExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/PMMLModelExportExample.scala
index d74d74a..96deafd 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/PMMLModelExportExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/PMMLModelExportExample.scala
@@ -41,7 +41,7 @@ object PMMLModelExportExample {
val clusters = KMeans.train(parsedData, numClusters, numIterations)
// Export to PMML to a String in PMML format
- println("PMML Model:\n" + clusters.toPMML)
+ println(s"PMML Model:\n ${clusters.toPMML}")
// Export the model to a local file in PMML format
clusters.toPMML("/tmp/kmeans.xml")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala
index 69c72c4..8b78927 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala
@@ -42,8 +42,8 @@ object PrefixSpanExample {
val model = prefixSpan.run(sequences)
model.freqSequences.collect().foreach { freqSequence =>
println(
- freqSequence.sequence.map(_.mkString("[", ", ", "]")).mkString("[", ", ", "]") +
- ", " + freqSequence.freq)
+ s"${freqSequence.sequence.map(_.mkString("[", ", ", "]")).mkString("[", ", ", "]")}," +
+ s" ${freqSequence.freq}")
}
// $example off$
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala
index f1ebdf1..246e71d 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala
@@ -55,8 +55,8 @@ object RandomForestClassificationExample {
(point.label, prediction)
}
val testErr = labelAndPreds.filter(r => r._1 != r._2).count.toDouble / testData.count()
- println("Test Error = " + testErr)
- println("Learned classification forest model:\n" + model.toDebugString)
+ println(s"Test Error = $testErr")
+ println(s"Learned classification forest model:\n ${model.toDebugString}")
// Save and load model
model.save(sc, "target/tmp/myRandomForestClassificationModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala
index 11d612e..770e302 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala
@@ -55,8 +55,8 @@ object RandomForestRegressionExample {
(point.label, prediction)
}
val testMSE = labelsAndPredictions.map{ case(v, p) => math.pow((v - p), 2)}.mean()
- println("Test Mean Squared Error = " + testMSE)
- println("Learned regression forest model:\n" + model.toDebugString)
+ println(s"Test Mean Squared Error = $testMSE")
+ println(s"Learned regression forest model:\n ${model.toDebugString}")
// Save and load model
model.save(sc, "target/tmp/myRandomForestRegressionModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala
index 6df742d..0bb2b8c 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala
@@ -56,7 +56,7 @@ object RecommendationExample {
val err = (r1 - r2)
err * err
}.mean()
- println("Mean Squared Error = " + MSE)
+ println(s"Mean Squared Error = $MSE")
// Save and load model
model.save(sc, "target/tmp/myCollaborativeFilter")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/SVMWithSGDExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/SVMWithSGDExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/SVMWithSGDExample.scala
index b73fe9b..285e2ce 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/SVMWithSGDExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/SVMWithSGDExample.scala
@@ -57,7 +57,7 @@ object SVMWithSGDExample {
val metrics = new BinaryClassificationMetrics(scoreAndLabels)
val auROC = metrics.areaUnderROC()
- println("Area under ROC = " + auROC)
+ println(s"Area under ROC = $auROC")
// Save and load model
model.save(sc, "target/tmp/scalaSVMWithSGDModel")
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala
index b5c3033..694c3bb 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala
@@ -42,15 +42,13 @@ object SimpleFPGrowth {
val model = fpg.run(transactions)
model.freqItemsets.collect().foreach { itemset =>
- println(itemset.items.mkString("[", ",", "]") + ", " + itemset.freq)
+ println(s"${itemset.items.mkString("[", ",", "]")},${itemset.freq}")
}
val minConfidence = 0.8
model.generateAssociationRules(minConfidence).collect().foreach { rule =>
- println(
- rule.antecedent.mkString("[", ",", "]")
- + " => " + rule.consequent .mkString("[", ",", "]")
- + ", " + rule.confidence)
+ println(s"${rule.antecedent.mkString("[", ",", "]")}=> " +
+ s"${rule.consequent .mkString("[", ",", "]")},${rule.confidence}")
}
// $example off$
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/StratifiedSamplingExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/StratifiedSamplingExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/StratifiedSamplingExample.scala
index 16b074e..3d41bef 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/StratifiedSamplingExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/StratifiedSamplingExample.scala
@@ -41,10 +41,10 @@ object StratifiedSamplingExample {
val exactSample = data.sampleByKeyExact(withReplacement = false, fractions = fractions)
// $example off$
- println("approxSample size is " + approxSample.collect().size.toString)
+ println(s"approxSample size is ${approxSample.collect().size}")
approxSample.collect().foreach(println)
- println("exactSample its size is " + exactSample.collect().size.toString)
+ println(s"exactSample its size is ${exactSample.collect().size}")
exactSample.collect().foreach(println)
sc.stop()
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnyPCA.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnyPCA.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnyPCA.scala
index 03bc675..071d341 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnyPCA.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnyPCA.scala
@@ -54,7 +54,7 @@ object TallSkinnyPCA {
// Compute principal components.
val pc = mat.computePrincipalComponents(mat.numCols().toInt)
- println("Principal components are:\n" + pc)
+ println(s"Principal components are:\n $pc")
sc.stop()
}
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnySVD.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnySVD.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnySVD.scala
index 067e49b..8ae6de1 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnySVD.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/TallSkinnySVD.scala
@@ -54,7 +54,7 @@ object TallSkinnySVD {
// Compute SVD.
val svd = mat.computeSVD(mat.numCols().toInt)
- println("Singular values are " + svd.s)
+ println(s"Singular values are ${svd.s}")
sc.stop()
}
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala
index 43044d0..25c7bf2 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala
@@ -82,9 +82,9 @@ class CustomReceiver(host: String, port: Int)
var socket: Socket = null
var userInput: String = null
try {
- logInfo("Connecting to " + host + ":" + port)
+ logInfo(s"Connecting to $host : $port")
socket = new Socket(host, port)
- logInfo("Connected to " + host + ":" + port)
+ logInfo(s"Connected to $host : $port")
val reader = new BufferedReader(
new InputStreamReader(socket.getInputStream(), StandardCharsets.UTF_8))
userInput = reader.readLine()
@@ -98,7 +98,7 @@ class CustomReceiver(host: String, port: Int)
restart("Trying to connect again")
} catch {
case e: java.net.ConnectException =>
- restart("Error connecting to " + host + ":" + port, e)
+ restart(s"Error connecting to $host : $port", e)
case t: Throwable =>
restart("Error receiving data", t)
}
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/streaming/RawNetworkGrep.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/RawNetworkGrep.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/RawNetworkGrep.scala
index 5322929..437ccf0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/RawNetworkGrep.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/RawNetworkGrep.scala
@@ -54,7 +54,7 @@ object RawNetworkGrep {
ssc.rawSocketStream[String](host, port, StorageLevel.MEMORY_ONLY_SER_2)).toArray
val union = ssc.union(rawStreams)
union.filter(_.contains("the")).count().foreachRDD(r =>
- println("Grep count: " + r.collect().mkString))
+ println(s"Grep count: ${r.collect().mkString}"))
ssc.start()
ssc.awaitTermination()
}
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala
index 49c0427..f018f3a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala
@@ -130,10 +130,10 @@ object RecoverableNetworkWordCount {
true
}
}.collect().mkString("[", ", ", "]")
- val output = "Counts at time " + time + " " + counts
+ val output = s"Counts at time $time $counts"
println(output)
- println("Dropped " + droppedWordsCounter.value + " word(s) totally")
- println("Appending to " + outputFile.getAbsolutePath)
+ println(s"Dropped ${droppedWordsCounter.value} word(s) totally")
+ println(s"Appending to ${outputFile.getAbsolutePath}")
Files.append(output + "\n", outputFile, Charset.defaultCharset())
}
ssc
@@ -141,7 +141,7 @@ object RecoverableNetworkWordCount {
def main(args: Array[String]) {
if (args.length != 4) {
- System.err.println("Your arguments were " + args.mkString("[", ", ", "]"))
+ System.err.println(s"Your arguments were ${args.mkString("[", ", ", "]")}")
System.err.println(
"""
|Usage: RecoverableNetworkWordCount <hostname> <port> <checkpoint-directory>
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala
index 0ddd065..2108bc6 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala
@@ -90,13 +90,13 @@ object PageViewGenerator {
val viewsPerSecond = args(1).toFloat
val sleepDelayMs = (1000.0 / viewsPerSecond).toInt
val listener = new ServerSocket(port)
- println("Listening on port: " + port)
+ println(s"Listening on port: $port")
while (true) {
val socket = listener.accept()
new Thread() {
override def run(): Unit = {
- println("Got client connected from: " + socket.getInetAddress)
+ println(s"Got client connected from: ${socket.getInetAddress}")
val out = new PrintWriter(socket.getOutputStream(), true)
while (true) {
http://git-wip-us.apache.org/repos/asf/spark/blob/9a2b65a3/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
index 1ba093f..b8e7c7e 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
@@ -104,8 +104,8 @@ object PageViewStream {
.foreachRDD((rdd, time) => rdd.join(userList)
.map(_._2._2)
.take(10)
- .foreach(u => println("Saw user %s at time %s".format(u, time))))
- case _ => println("Invalid metric entered: " + metric)
+ .foreach(u => println(s"Saw user $u at time $time")))
+ case _ => println(s"Invalid metric entered: $metric")
}
ssc.start()
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org