You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by rx...@apache.org on 2016/01/05 07:42:58 UTC
[1/2] spark git commit: [SPARK-3873][EXAMPLES] Import ordering fixes.
Repository: spark
Updated Branches:
refs/heads/master cc4d5229c -> 7058dc115
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala
index d237232..ef86eab 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/PrefixSpanExample.scala
@@ -18,12 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.fpm.PrefixSpan
// $example off$
-import org.apache.spark.{SparkConf, SparkContext}
-
object PrefixSpanExample {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala
index 5e55abd..7805153 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestClassificationExample.scala
@@ -18,7 +18,7 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
-import org.apache.spark.{SparkContext, SparkConf}
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.tree.RandomForest
import org.apache.spark.mllib.tree.model.RandomForestModel
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala
index a54fb3a..655a277 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomForestRegressionExample.scala
@@ -18,7 +18,7 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
-import org.apache.spark.{SparkContext, SparkConf}
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.tree.RandomForest
import org.apache.spark.mllib.tree.model.RandomForestModel
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/RandomRDDGeneration.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomRDDGeneration.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomRDDGeneration.scala
index bee85ba..7ccbb5a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RandomRDDGeneration.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RandomRDDGeneration.scala
@@ -18,11 +18,10 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.mllib.random.RandomRDDs
import org.apache.spark.rdd.RDD
-import org.apache.spark.{SparkConf, SparkContext}
-
/**
* An example app for randomly generated RDDs. Run with
* {{{
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala
index cffa03d..fdb01b8 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RankingMetricsExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
-import org.apache.spark.mllib.evaluation.{RegressionMetrics, RankingMetrics}
+import org.apache.spark.mllib.evaluation.{RankingMetrics, RegressionMetrics}
import org.apache.spark.mllib.recommendation.{ALS, Rating}
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkContext, SparkConf}
object RankingMetricsExample {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala
index 64e4602..bc94695 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RecommendationExample.scala
@@ -18,7 +18,7 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
-import org.apache.spark.{SparkContext, SparkConf}
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.recommendation.ALS
import org.apache.spark.mllib.recommendation.MatrixFactorizationModel
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala
index 47d4453..ace16ff 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/RegressionMetricsExample.scala
@@ -18,13 +18,13 @@
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
-import org.apache.spark.mllib.regression.LinearRegressionWithSGD
import org.apache.spark.mllib.evaluation.RegressionMetrics
+import org.apache.spark.mllib.regression.LinearRegressionWithSGD
import org.apache.spark.mllib.util.MLUtils
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object RegressionMetricsExample {
def main(args: Array[String]) : Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/SampledRDDs.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/SampledRDDs.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/SampledRDDs.scala
index 6963f43..c4e5e96 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/SampledRDDs.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/SampledRDDs.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
-import org.apache.spark.mllib.util.MLUtils
import scopt.OptionParser
import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.SparkContext._
+import org.apache.spark.mllib.util.MLUtils
/**
* An example app for randomly generated and sampled RDDs. Run with
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala
index b4e06af..ab15ac2 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/SimpleFPGrowth.scala
@@ -18,13 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.fpm.FPGrowth
import org.apache.spark.rdd.RDD
// $example off$
-import org.apache.spark.{SparkContext, SparkConf}
-
object SimpleFPGrowth {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLinearRegression.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLinearRegression.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLinearRegression.scala
index b4a5dca..e559296 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLinearRegression.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLinearRegression.scala
@@ -18,9 +18,9 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.SparkConf
import org.apache.spark.mllib.linalg.Vectors
import org.apache.spark.mllib.regression.{LabeledPoint, StreamingLinearRegressionWithSGD}
-import org.apache.spark.SparkConf
import org.apache.spark.streaming.{Seconds, StreamingContext}
/**
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLogisticRegression.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLogisticRegression.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLogisticRegression.scala
index b42f4cb..a8b144a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLogisticRegression.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingLogisticRegression.scala
@@ -18,10 +18,10 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.SparkConf
+import org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
import org.apache.spark.mllib.linalg.Vectors
import org.apache.spark.mllib.regression.LabeledPoint
-import org.apache.spark.mllib.classification.StreamingLogisticRegressionWithSGD
-import org.apache.spark.SparkConf
import org.apache.spark.streaming.{Seconds, StreamingContext}
/**
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/pythonconverters/HBaseConverters.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/pythonconverters/HBaseConverters.scala b/examples/src/main/scala/org/apache/spark/examples/pythonconverters/HBaseConverters.scala
index 0a25ee7..e252ca8 100644
--- a/examples/src/main/scala/org/apache/spark/examples/pythonconverters/HBaseConverters.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/pythonconverters/HBaseConverters.scala
@@ -20,12 +20,13 @@ package org.apache.spark.examples.pythonconverters
import scala.collection.JavaConverters._
import scala.util.parsing.json.JSONObject
-import org.apache.spark.api.python.Converter
+import org.apache.hadoop.hbase.CellUtil
+import org.apache.hadoop.hbase.KeyValue.Type
import org.apache.hadoop.hbase.client.{Put, Result}
import org.apache.hadoop.hbase.io.ImmutableBytesWritable
import org.apache.hadoop.hbase.util.Bytes
-import org.apache.hadoop.hbase.KeyValue.Type
-import org.apache.hadoop.hbase.CellUtil
+
+import org.apache.spark.api.python.Converter
/**
* Implementation of [[org.apache.spark.api.python.Converter]] that converts all
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala b/examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala
index bf40bd1..4e427f5 100644
--- a/examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala
@@ -18,10 +18,10 @@
// scalastyle:off println
package org.apache.spark.examples.sql.hive
-import com.google.common.io.{ByteStreams, Files}
-
import java.io.File
+import com.google.common.io.{ByteStreams, Files}
+
import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.sql._
import org.apache.spark.sql.hive.HiveContext
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/ActorWordCount.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/ActorWordCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/ActorWordCount.scala
index e9c9907..8b8dae0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/ActorWordCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/ActorWordCount.scala
@@ -22,13 +22,13 @@ import scala.collection.mutable.LinkedList
import scala.reflect.ClassTag
import scala.util.Random
-import akka.actor.{Actor, ActorRef, Props, actorRef2Scala}
+import akka.actor.{actorRef2Scala, Actor, ActorRef, Props}
-import org.apache.spark.{SparkConf, SecurityManager}
+import org.apache.spark.{SecurityManager, SparkConf}
import org.apache.spark.streaming.{Seconds, StreamingContext}
import org.apache.spark.streaming.StreamingContext.toPairDStreamFunctions
-import org.apache.spark.util.AkkaUtils
import org.apache.spark.streaming.receiver.ActorHelper
+import org.apache.spark.util.AkkaUtils
case class SubscribeReceiver(receiverActor: ActorRef)
case class UnsubscribeReceiver(receiverActor: ActorRef)
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala
index 28e9bf5..ad13d43 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/CustomReceiver.scala
@@ -18,10 +18,10 @@
// scalastyle:off println
package org.apache.spark.examples.streaming
-import java.io.{InputStreamReader, BufferedReader, InputStream}
+import java.io.{BufferedReader, InputStream, InputStreamReader}
import java.net.Socket
-import org.apache.spark.{SparkConf, Logging}
+import org.apache.spark.{Logging, SparkConf}
import org.apache.spark.storage.StorageLevel
import org.apache.spark.streaming.{Seconds, StreamingContext}
import org.apache.spark.streaming.receiver.Receiver
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/FlumePollingEventCount.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/FlumePollingEventCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/FlumePollingEventCount.scala
index 2bdbc37..fe3b79e 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/FlumePollingEventCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/FlumePollingEventCount.scala
@@ -18,12 +18,13 @@
// scalastyle:off println
package org.apache.spark.examples.streaming
+import java.net.InetSocketAddress
+
import org.apache.spark.SparkConf
import org.apache.spark.storage.StorageLevel
import org.apache.spark.streaming._
import org.apache.spark.streaming.flume._
import org.apache.spark.util.IntParam
-import java.net.InetSocketAddress
/**
* Produces a count of events received from Flume.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/KafkaWordCount.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/KafkaWordCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/KafkaWordCount.scala
index b40d17e..e7f9bf3 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/KafkaWordCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/KafkaWordCount.scala
@@ -20,11 +20,11 @@ package org.apache.spark.examples.streaming
import java.util.HashMap
-import org.apache.kafka.clients.producer.{ProducerConfig, KafkaProducer, ProducerRecord}
+import org.apache.kafka.clients.producer.{KafkaProducer, ProducerConfig, ProducerRecord}
+import org.apache.spark.SparkConf
import org.apache.spark.streaming._
import org.apache.spark.streaming.kafka._
-import org.apache.spark.SparkConf
/**
* Consumes messages from one or more topics in Kafka and does wordcount.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/NetworkWordCount.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/NetworkWordCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/NetworkWordCount.scala
index 9a57fe2..15b57fc 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/NetworkWordCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/NetworkWordCount.scala
@@ -19,8 +19,8 @@
package org.apache.spark.examples.streaming
import org.apache.spark.SparkConf
-import org.apache.spark.streaming.{Seconds, StreamingContext}
import org.apache.spark.storage.StorageLevel
+import org.apache.spark.streaming.{Seconds, StreamingContext}
/**
* Counts words in UTF8 encoded, '\n' delimited text received from the network every second.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala
index 38d4fd1..05f8e65 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/RecoverableNetworkWordCount.scala
@@ -26,7 +26,7 @@ import com.google.common.io.Files
import org.apache.spark.{Accumulator, SparkConf, SparkContext}
import org.apache.spark.broadcast.Broadcast
import org.apache.spark.rdd.RDD
-import org.apache.spark.streaming.{Time, Seconds, StreamingContext}
+import org.apache.spark.streaming.{Seconds, StreamingContext, Time}
import org.apache.spark.util.IntParam
/**
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala
index ed61775..9aa0f54 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/SqlNetworkWordCount.scala
@@ -21,10 +21,10 @@ package org.apache.spark.examples.streaming
import org.apache.spark.SparkConf
import org.apache.spark.SparkContext
import org.apache.spark.rdd.RDD
-import org.apache.spark.streaming.{Time, Seconds, StreamingContext}
-import org.apache.spark.util.IntParam
import org.apache.spark.sql.SQLContext
import org.apache.spark.storage.StorageLevel
+import org.apache.spark.streaming.{Seconds, StreamingContext, Time}
+import org.apache.spark.util.IntParam
/**
* Use DataFrames and SQL to count words in UTF8 encoded, '\n' delimited text received from the
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala
index 2dce182..c85d684 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/StatefulNetworkWordCount.scala
@@ -18,8 +18,8 @@
// scalastyle:off println
package org.apache.spark.examples.streaming
-import org.apache.spark.SparkConf
import org.apache.spark.HashPartitioner
+import org.apache.spark.SparkConf
import org.apache.spark.streaming._
/**
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/StreamingExamples.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/StreamingExamples.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/StreamingExamples.scala
index 8396e65..22a5654 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/StreamingExamples.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/StreamingExamples.scala
@@ -17,10 +17,10 @@
package org.apache.spark.examples.streaming
-import org.apache.spark.Logging
-
import org.apache.log4j.{Level, Logger}
+import org.apache.spark.Logging
+
/** Utility functions for Spark Streaming examples. */
object StreamingExamples extends Logging {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterAlgebirdHLL.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterAlgebirdHLL.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterAlgebirdHLL.scala
index 49826ed..0ec6214 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterAlgebirdHLL.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterAlgebirdHLL.scala
@@ -18,13 +18,13 @@
// scalastyle:off println
package org.apache.spark.examples.streaming
-import com.twitter.algebird.HyperLogLogMonoid
import com.twitter.algebird.HyperLogLog._
+import com.twitter.algebird.HyperLogLogMonoid
+import org.apache.spark.SparkConf
import org.apache.spark.storage.StorageLevel
import org.apache.spark.streaming.{Seconds, StreamingContext}
import org.apache.spark.streaming.twitter._
-import org.apache.spark.SparkConf
// scalastyle:off
/**
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterHashTagJoinSentiments.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterHashTagJoinSentiments.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterHashTagJoinSentiments.scala
index 0328fa8..edf0e0b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterHashTagJoinSentiments.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/TwitterHashTagJoinSentiments.scala
@@ -19,8 +19,8 @@
package org.apache.spark.examples.streaming
import org.apache.spark.SparkConf
-import org.apache.spark.streaming.twitter.TwitterUtils
import org.apache.spark.streaming.{Seconds, StreamingContext}
+import org.apache.spark.streaming.twitter.TwitterUtils
/**
* Displays the most positive hash tags by joining the streaming Twitter data with a static RDD of
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/ZeroMQWordCount.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/ZeroMQWordCount.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/ZeroMQWordCount.scala
index 6ac9a72..9644890 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/ZeroMQWordCount.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/ZeroMQWordCount.scala
@@ -18,18 +18,18 @@
// scalastyle:off println
package org.apache.spark.examples.streaming
+import scala.language.implicitConversions
+
import akka.actor.ActorSystem
import akka.actor.actorRef2Scala
+import akka.util.ByteString
import akka.zeromq._
import akka.zeromq.Subscribe
-import akka.util.ByteString
+import org.apache.spark.SparkConf
import org.apache.spark.streaming.{Seconds, StreamingContext}
import org.apache.spark.streaming.zeromq._
-import scala.language.implicitConversions
-import org.apache.spark.SparkConf
-
/**
* A simple publisher for demonstration purposes, repeatedly publishes random Messages
* every one second.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala
index 2fcccb2..ce1a620 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewGenerator.scala
@@ -18,9 +18,9 @@
// scalastyle:off println
package org.apache.spark.examples.streaming.clickstream
-import java.net.ServerSocket
import java.io.PrintWriter
-import util.Random
+import java.net.ServerSocket
+import java.util.Random
/** Represents a page view on a website with associated dimension data. */
class PageView(val url : String, val status : Int, val zipCode : Int, val userID : Int)
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
index 7236168..4b43550 100644
--- a/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/streaming/clickstream/PageViewStream.scala
@@ -18,8 +18,9 @@
// scalastyle:off println
package org.apache.spark.examples.streaming.clickstream
-import org.apache.spark.streaming.{Seconds, StreamingContext}
import org.apache.spark.examples.streaming.StreamingExamples
+import org.apache.spark.streaming.{Seconds, StreamingContext}
+
// scalastyle:off
/** Analyses a streaming dataset of web page views. This class demonstrates several types of
* operators available in Spark streaming.
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org
[2/2] spark git commit: [SPARK-3873][EXAMPLES] Import ordering fixes.
Posted by rx...@apache.org.
[SPARK-3873][EXAMPLES] Import ordering fixes.
Author: Marcelo Vanzin <va...@cloudera.com>
Closes #10575 from vanzin/SPARK-3873-examples.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/7058dc11
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/7058dc11
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/7058dc11
Branch: refs/heads/master
Commit: 7058dc115047258197f6c09eee404f1ccf41038d
Parents: cc4d522
Author: Marcelo Vanzin <va...@cloudera.com>
Authored: Mon Jan 4 22:42:54 2016 -0800
Committer: Reynold Xin <rx...@databricks.com>
Committed: Mon Jan 4 22:42:54 2016 -0800
----------------------------------------------------------------------
.../org/apache/spark/examples/CassandraCQLTest.scala | 3 +--
.../scala/org/apache/spark/examples/CassandraTest.scala | 2 +-
.../org/apache/spark/examples/DFSReadWriteTest.scala | 2 +-
.../scala/org/apache/spark/examples/HBaseTest.scala | 3 +--
.../scala/org/apache/spark/examples/LocalFileLR.scala | 2 +-
.../scala/org/apache/spark/examples/LocalKMeans.scala | 2 +-
.../main/scala/org/apache/spark/examples/LocalLR.scala | 2 +-
.../org/apache/spark/examples/MultiBroadcastTest.scala | 2 +-
.../scala/org/apache/spark/examples/SparkHdfsLR.scala | 2 +-
.../scala/org/apache/spark/examples/SparkKMeans.scala | 2 +-
.../main/scala/org/apache/spark/examples/SparkLR.scala | 2 +-
.../scala/org/apache/spark/examples/SparkPageRank.scala | 2 +-
.../main/scala/org/apache/spark/examples/SparkTC.scala | 2 +-
.../org/apache/spark/examples/SparkTachyonHdfsLR.scala | 3 +--
.../org/apache/spark/examples/graphx/Analytics.scala | 5 +++--
.../apache/spark/examples/graphx/SynthBenchmark.scala | 5 +++--
.../examples/ml/AFTSurvivalRegressionExample.scala | 4 ++--
.../org/apache/spark/examples/ml/BinarizerExample.scala | 2 +-
.../apache/spark/examples/ml/BucketizerExample.scala | 2 +-
.../apache/spark/examples/ml/ChiSqSelectorExample.scala | 2 +-
.../spark/examples/ml/CountVectorizerExample.scala | 3 +--
.../spark/examples/ml/CrossValidatorExample.scala | 2 +-
.../scala/org/apache/spark/examples/ml/DCTExample.scala | 2 +-
.../examples/ml/DecisionTreeClassificationExample.scala | 8 ++++----
.../apache/spark/examples/ml/DecisionTreeExample.scala | 7 +++----
.../examples/ml/DecisionTreeRegressionExample.scala | 12 +++++++-----
.../spark/examples/ml/ElementwiseProductExample.scala | 2 +-
.../ml/GradientBoostedTreeClassifierExample.scala | 2 +-
.../ml/GradientBoostedTreeRegressorExample.scala | 2 +-
.../apache/spark/examples/ml/IndexToStringExample.scala | 4 ++--
.../scala/org/apache/spark/examples/ml/LDAExample.scala | 4 ++--
.../ml/LinearRegressionWithElasticNetExample.scala | 2 +-
.../examples/ml/LogisticRegressionSummaryExample.scala | 2 +-
.../ml/LogisticRegressionWithElasticNetExample.scala | 2 +-
.../apache/spark/examples/ml/MinMaxScalerExample.scala | 2 +-
.../ml/MultilayerPerceptronClassifierExample.scala | 4 ++--
.../org/apache/spark/examples/ml/NGramExample.scala | 2 +-
.../apache/spark/examples/ml/NormalizerExample.scala | 2 +-
.../apache/spark/examples/ml/OneHotEncoderExample.scala | 2 +-
.../org/apache/spark/examples/ml/OneVsRestExample.scala | 4 ++--
.../scala/org/apache/spark/examples/ml/PCAExample.scala | 2 +-
.../spark/examples/ml/PolynomialExpansionExample.scala | 2 +-
.../spark/examples/ml/QuantileDiscretizerExample.scala | 2 +-
.../org/apache/spark/examples/ml/RFormulaExample.scala | 2 +-
.../examples/ml/RandomForestClassifierExample.scala | 2 +-
.../examples/ml/RandomForestRegressorExample.scala | 2 +-
.../spark/examples/ml/SQLTransformerExample.scala | 3 +--
.../spark/examples/ml/StandardScalerExample.scala | 2 +-
.../spark/examples/ml/StopWordsRemoverExample.scala | 2 +-
.../apache/spark/examples/ml/StringIndexerExample.scala | 2 +-
.../org/apache/spark/examples/ml/TfIdfExample.scala | 2 +-
.../org/apache/spark/examples/ml/TokenizerExample.scala | 2 +-
.../spark/examples/ml/TrainValidationSplitExample.scala | 2 +-
.../spark/examples/ml/VectorAssemblerExample.scala | 2 +-
.../apache/spark/examples/ml/VectorIndexerExample.scala | 2 +-
.../apache/spark/examples/ml/VectorSlicerExample.scala | 2 +-
.../org/apache/spark/examples/ml/Word2VecExample.scala | 2 +-
.../spark/examples/mllib/AssociationRulesExample.scala | 3 +--
.../spark/examples/mllib/BinaryClassification.scala | 2 +-
.../mllib/BinaryClassificationMetricsExample.scala | 2 +-
.../spark/examples/mllib/BisectingKMeansExample.scala | 2 +-
.../org/apache/spark/examples/mllib/Correlations.scala | 3 +--
.../apache/spark/examples/mllib/CosineSimilarity.scala | 2 +-
.../mllib/DecisionTreeClassificationExample.scala | 2 +-
.../examples/mllib/DecisionTreeRegressionExample.scala | 2 +-
.../spark/examples/mllib/DecisionTreeRunner.scala | 2 +-
.../apache/spark/examples/mllib/FPGrowthExample.scala | 2 +-
.../examples/mllib/GradientBoostedTreesRunner.scala | 3 +--
.../mllib/GradientBoostingClassificationExample.scala | 2 +-
.../mllib/GradientBoostingRegressionExample.scala | 2 +-
.../examples/mllib/IsotonicRegressionExample.scala | 2 +-
.../org/apache/spark/examples/mllib/LBFGSExample.scala | 3 +--
.../org/apache/spark/examples/mllib/LDAExample.scala | 4 ++--
.../apache/spark/examples/mllib/LinearRegression.scala | 2 +-
.../spark/examples/mllib/MultiLabelMetricsExample.scala | 2 +-
.../spark/examples/mllib/MulticlassMetricsExample.scala | 2 +-
.../spark/examples/mllib/MultivariateSummarizer.scala | 3 +--
.../apache/spark/examples/mllib/NaiveBayesExample.scala | 2 +-
.../mllib/PowerIterationClusteringExample.scala | 2 +-
.../apache/spark/examples/mllib/PrefixSpanExample.scala | 3 +--
.../mllib/RandomForestClassificationExample.scala | 2 +-
.../examples/mllib/RandomForestRegressionExample.scala | 2 +-
.../spark/examples/mllib/RandomRDDGeneration.scala | 3 +--
.../spark/examples/mllib/RankingMetricsExample.scala | 4 ++--
.../spark/examples/mllib/RecommendationExample.scala | 2 +-
.../spark/examples/mllib/RegressionMetricsExample.scala | 4 ++--
.../org/apache/spark/examples/mllib/SampledRDDs.scala | 2 +-
.../apache/spark/examples/mllib/SimpleFPGrowth.scala | 3 +--
.../examples/mllib/StreamingLinearRegression.scala | 2 +-
.../examples/mllib/StreamingLogisticRegression.scala | 4 ++--
.../examples/pythonconverters/HBaseConverters.scala | 7 ++++---
.../apache/spark/examples/sql/hive/HiveFromSpark.scala | 4 ++--
.../spark/examples/streaming/ActorWordCount.scala | 6 +++---
.../spark/examples/streaming/CustomReceiver.scala | 4 ++--
.../examples/streaming/FlumePollingEventCount.scala | 3 ++-
.../spark/examples/streaming/KafkaWordCount.scala | 4 ++--
.../spark/examples/streaming/NetworkWordCount.scala | 2 +-
.../streaming/RecoverableNetworkWordCount.scala | 2 +-
.../spark/examples/streaming/SqlNetworkWordCount.scala | 4 ++--
.../examples/streaming/StatefulNetworkWordCount.scala | 2 +-
.../spark/examples/streaming/StreamingExamples.scala | 4 ++--
.../spark/examples/streaming/TwitterAlgebirdHLL.scala | 4 ++--
.../streaming/TwitterHashTagJoinSentiments.scala | 2 +-
.../spark/examples/streaming/ZeroMQWordCount.scala | 8 ++++----
.../streaming/clickstream/PageViewGenerator.scala | 4 ++--
.../examples/streaming/clickstream/PageViewStream.scala | 3 ++-
106 files changed, 147 insertions(+), 154 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/CassandraCQLTest.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/CassandraCQLTest.scala b/examples/src/main/scala/org/apache/spark/examples/CassandraCQLTest.scala
index 5a80985..973b005 100644
--- a/examples/src/main/scala/org/apache/spark/examples/CassandraCQLTest.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/CassandraCQLTest.scala
@@ -22,15 +22,14 @@ import java.nio.ByteBuffer
import java.util.Collections
import org.apache.cassandra.hadoop.ConfigHelper
-import org.apache.cassandra.hadoop.cql3.CqlPagingInputFormat
import org.apache.cassandra.hadoop.cql3.CqlConfigHelper
import org.apache.cassandra.hadoop.cql3.CqlOutputFormat
+import org.apache.cassandra.hadoop.cql3.CqlPagingInputFormat
import org.apache.cassandra.utils.ByteBufferUtil
import org.apache.hadoop.mapreduce.Job
import org.apache.spark.{SparkConf, SparkContext}
-
/*
Need to create following keyspace and column family in cassandra before running this example
Start CQL shell using ./bin/cqlsh and execute following commands
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/CassandraTest.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/CassandraTest.scala b/examples/src/main/scala/org/apache/spark/examples/CassandraTest.scala
index ad39a01..6a8f73a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/CassandraTest.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/CassandraTest.scala
@@ -23,9 +23,9 @@ import java.util.Arrays
import java.util.SortedMap
import org.apache.cassandra.db.IColumn
+import org.apache.cassandra.hadoop.ColumnFamilyInputFormat
import org.apache.cassandra.hadoop.ColumnFamilyOutputFormat
import org.apache.cassandra.hadoop.ConfigHelper
-import org.apache.cassandra.hadoop.ColumnFamilyInputFormat
import org.apache.cassandra.thrift._
import org.apache.cassandra.utils.ByteBufferUtil
import org.apache.hadoop.mapreduce.Job
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/DFSReadWriteTest.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/DFSReadWriteTest.scala b/examples/src/main/scala/org/apache/spark/examples/DFSReadWriteTest.scala
index d651fe4..b26db0b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/DFSReadWriteTest.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/DFSReadWriteTest.scala
@@ -22,7 +22,7 @@ import java.io.File
import scala.io.Source._
-import org.apache.spark.{SparkContext, SparkConf}
+import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.SparkContext._
/**
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/HBaseTest.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/HBaseTest.scala b/examples/src/main/scala/org/apache/spark/examples/HBaseTest.scala
index 2447423..65d7489 100644
--- a/examples/src/main/scala/org/apache/spark/examples/HBaseTest.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/HBaseTest.scala
@@ -18,13 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples
-import org.apache.hadoop.hbase.client.HBaseAdmin
import org.apache.hadoop.hbase.{HBaseConfiguration, HTableDescriptor, TableName}
+import org.apache.hadoop.hbase.client.HBaseAdmin
import org.apache.hadoop.hbase.mapreduce.TableInputFormat
import org.apache.spark._
-
object HBaseTest {
def main(args: Array[String]) {
val sparkConf = new SparkConf().setAppName("HBaseTest")
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/LocalFileLR.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/LocalFileLR.scala b/examples/src/main/scala/org/apache/spark/examples/LocalFileLR.scala
index 9c8aae5..a390185 100644
--- a/examples/src/main/scala/org/apache/spark/examples/LocalFileLR.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/LocalFileLR.scala
@@ -20,7 +20,7 @@ package org.apache.spark.examples
import java.util.Random
-import breeze.linalg.{Vector, DenseVector}
+import breeze.linalg.{DenseVector, Vector}
/**
* Logistic regression based classification.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/LocalKMeans.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/LocalKMeans.scala b/examples/src/main/scala/org/apache/spark/examples/LocalKMeans.scala
index e7b28d3..407e3e0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/LocalKMeans.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/LocalKMeans.scala
@@ -23,7 +23,7 @@ import java.util.Random
import scala.collection.mutable.HashMap
import scala.collection.mutable.HashSet
-import breeze.linalg.{Vector, DenseVector, squaredDistance}
+import breeze.linalg.{squaredDistance, DenseVector, Vector}
import org.apache.spark.SparkContext._
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/LocalLR.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/LocalLR.scala b/examples/src/main/scala/org/apache/spark/examples/LocalLR.scala
index 4f6b092..58adbab 100644
--- a/examples/src/main/scala/org/apache/spark/examples/LocalLR.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/LocalLR.scala
@@ -20,7 +20,7 @@ package org.apache.spark.examples
import java.util.Random
-import breeze.linalg.{Vector, DenseVector}
+import breeze.linalg.{DenseVector, Vector}
/**
* Logistic regression based classification.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/MultiBroadcastTest.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/MultiBroadcastTest.scala b/examples/src/main/scala/org/apache/spark/examples/MultiBroadcastTest.scala
index 61ce9db..a797111 100644
--- a/examples/src/main/scala/org/apache/spark/examples/MultiBroadcastTest.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/MultiBroadcastTest.scala
@@ -18,8 +18,8 @@
// scalastyle:off println
package org.apache.spark.examples
-import org.apache.spark.rdd.RDD
import org.apache.spark.{SparkConf, SparkContext}
+import org.apache.spark.rdd.RDD
/**
* Usage: MultiBroadcastTest [slices] [numElem]
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/SparkHdfsLR.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/SparkHdfsLR.scala b/examples/src/main/scala/org/apache/spark/examples/SparkHdfsLR.scala
index 505ea5a..6c90dbe 100644
--- a/examples/src/main/scala/org/apache/spark/examples/SparkHdfsLR.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/SparkHdfsLR.scala
@@ -22,7 +22,7 @@ import java.util.Random
import scala.math.exp
-import breeze.linalg.{Vector, DenseVector}
+import breeze.linalg.{DenseVector, Vector}
import org.apache.hadoop.conf.Configuration
import org.apache.spark._
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/SparkKMeans.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/SparkKMeans.scala b/examples/src/main/scala/org/apache/spark/examples/SparkKMeans.scala
index c56e112..1ea9121 100644
--- a/examples/src/main/scala/org/apache/spark/examples/SparkKMeans.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/SparkKMeans.scala
@@ -18,7 +18,7 @@
// scalastyle:off println
package org.apache.spark.examples
-import breeze.linalg.{Vector, DenseVector, squaredDistance}
+import breeze.linalg.{squaredDistance, DenseVector, Vector}
import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.SparkContext._
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/SparkLR.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/SparkLR.scala b/examples/src/main/scala/org/apache/spark/examples/SparkLR.scala
index d265c22..132800e 100644
--- a/examples/src/main/scala/org/apache/spark/examples/SparkLR.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/SparkLR.scala
@@ -22,7 +22,7 @@ import java.util.Random
import scala.math.exp
-import breeze.linalg.{Vector, DenseVector}
+import breeze.linalg.{DenseVector, Vector}
import org.apache.spark._
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/SparkPageRank.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/SparkPageRank.scala b/examples/src/main/scala/org/apache/spark/examples/SparkPageRank.scala
index 0fd7966..018bdf6 100644
--- a/examples/src/main/scala/org/apache/spark/examples/SparkPageRank.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/SparkPageRank.scala
@@ -18,8 +18,8 @@
// scalastyle:off println
package org.apache.spark.examples
-import org.apache.spark.SparkContext._
import org.apache.spark.{SparkConf, SparkContext}
+import org.apache.spark.SparkContext._
/**
* Computes the PageRank of URLs from an input file. Input file should
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/SparkTC.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/SparkTC.scala b/examples/src/main/scala/org/apache/spark/examples/SparkTC.scala
index 9507207..b92740f 100644
--- a/examples/src/main/scala/org/apache/spark/examples/SparkTC.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/SparkTC.scala
@@ -18,8 +18,8 @@
// scalastyle:off println
package org.apache.spark.examples
-import scala.util.Random
import scala.collection.mutable
+import scala.util.Random
import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.SparkContext._
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/SparkTachyonHdfsLR.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/SparkTachyonHdfsLR.scala b/examples/src/main/scala/org/apache/spark/examples/SparkTachyonHdfsLR.scala
index cfbdae0..e492582 100644
--- a/examples/src/main/scala/org/apache/spark/examples/SparkTachyonHdfsLR.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/SparkTachyonHdfsLR.scala
@@ -22,14 +22,13 @@ import java.util.Random
import scala.math.exp
-import breeze.linalg.{Vector, DenseVector}
+import breeze.linalg.{DenseVector, Vector}
import org.apache.hadoop.conf.Configuration
import org.apache.spark._
import org.apache.spark.scheduler.InputFormatInfo
import org.apache.spark.storage.StorageLevel
-
/**
* Logistic regression based classification.
* This example uses Tachyon to persist rdds during computation.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala b/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala
index 8dd6c97..39cb83d 100644
--- a/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/graphx/Analytics.scala
@@ -19,11 +19,12 @@
package org.apache.spark.examples.graphx
import scala.collection.mutable
+
import org.apache.spark._
-import org.apache.spark.storage.StorageLevel
import org.apache.spark.graphx._
-import org.apache.spark.graphx.lib._
import org.apache.spark.graphx.PartitionStrategy._
+import org.apache.spark.graphx.lib._
+import org.apache.spark.storage.StorageLevel
/**
* Driver program for running graph algorithms.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala b/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala
index 46e52aa..41ca5cb 100644
--- a/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/graphx/SynthBenchmark.scala
@@ -18,11 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.graphx
+import java.io.{FileOutputStream, PrintWriter}
+
+import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.SparkContext._
import org.apache.spark.graphx.{GraphXUtils, PartitionStrategy}
-import org.apache.spark.{SparkContext, SparkConf}
import org.apache.spark.graphx.util.GraphGenerators
-import java.io.{PrintWriter, FileOutputStream}
/**
* The SynthBenchmark application can be used to run various GraphX algorithms on
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala
index f4b3613..21f58dd 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/AFTSurvivalRegressionExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.ml
-import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkContext, SparkConf}
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.regression.AFTSurvivalRegression
import org.apache.spark.mllib.linalg.Vectors
// $example off$
+import org.apache.spark.sql.SQLContext
/**
* An example for AFTSurvivalRegression.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala
index e724aa5..2ed8101 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/BinarizerExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.Binarizer
// $example off$
import org.apache.spark.sql.{DataFrame, SQLContext}
-import org.apache.spark.{SparkConf, SparkContext}
object BinarizerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala
index 7c75e3d..6f6236a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/BucketizerExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.Bucketizer
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object BucketizerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala
index a8d2bc4..2be6153 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/ChiSqSelectorExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.ChiSqSelector
import org.apache.spark.mllib.linalg.Vectors
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object ChiSqSelectorExample {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala
index ba916f6..7d07fc7 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/CountVectorizerExample.scala
@@ -18,12 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.{CountVectorizer, CountVectorizerModel}
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
-
object CountVectorizerExample {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/CrossValidatorExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/CrossValidatorExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/CrossValidatorExample.scala
index 14b358d..bca301d 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/CrossValidatorExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/CrossValidatorExample.scala
@@ -23,7 +23,7 @@ import org.apache.spark.ml.Pipeline
import org.apache.spark.ml.classification.LogisticRegression
import org.apache.spark.ml.evaluation.BinaryClassificationEvaluator
import org.apache.spark.ml.feature.{HashingTF, Tokenizer}
-import org.apache.spark.ml.tuning.{ParamGridBuilder, CrossValidator}
+import org.apache.spark.ml.tuning.{CrossValidator, ParamGridBuilder}
import org.apache.spark.mllib.linalg.Vector
import org.apache.spark.sql.{Row, SQLContext}
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala
index 314c2c2..dc26b55 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DCTExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.DCT
import org.apache.spark.mllib.linalg.Vectors
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object DCTExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
index db024b5..224d8da 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeClassificationExample.scala
@@ -18,15 +18,15 @@
// scalastyle:off println
package org.apache.spark.examples.ml
-import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkContext, SparkConf}
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.Pipeline
-import org.apache.spark.ml.classification.DecisionTreeClassifier
import org.apache.spark.ml.classification.DecisionTreeClassificationModel
-import org.apache.spark.ml.feature.{StringIndexer, IndexToString, VectorIndexer}
+import org.apache.spark.ml.classification.DecisionTreeClassifier
import org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
+import org.apache.spark.ml.feature.{IndexToString, StringIndexer, VectorIndexer}
// $example off$
+import org.apache.spark.sql.SQLContext
object DecisionTreeClassificationExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala
index c4e98df..a37d12a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeExample.scala
@@ -27,14 +27,13 @@ import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.examples.mllib.AbstractParams
import org.apache.spark.ml.{Pipeline, PipelineStage, Transformer}
import org.apache.spark.ml.classification.{DecisionTreeClassificationModel, DecisionTreeClassifier}
-import org.apache.spark.ml.feature.{VectorIndexer, StringIndexer}
+import org.apache.spark.ml.feature.{StringIndexer, VectorIndexer}
import org.apache.spark.ml.regression.{DecisionTreeRegressionModel, DecisionTreeRegressor}
import org.apache.spark.ml.util.MetadataUtils
-import org.apache.spark.mllib.evaluation.{RegressionMetrics, MulticlassMetrics}
+import org.apache.spark.mllib.evaluation.{MulticlassMetrics, RegressionMetrics}
import org.apache.spark.mllib.linalg.Vector
import org.apache.spark.mllib.util.MLUtils
-import org.apache.spark.sql.{SQLContext, DataFrame}
-
+import org.apache.spark.sql.{DataFrame, SQLContext}
/**
* An example runner for decision trees. Run with
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
index ad01f55..ad32e56 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/DecisionTreeRegressionExample.scala
@@ -17,15 +17,17 @@
// scalastyle:off println
package org.apache.spark.examples.ml
-import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkContext, SparkConf}
+
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.Pipeline
-import org.apache.spark.ml.regression.DecisionTreeRegressor
-import org.apache.spark.ml.regression.DecisionTreeRegressionModel
-import org.apache.spark.ml.feature.VectorIndexer
import org.apache.spark.ml.evaluation.RegressionEvaluator
+import org.apache.spark.ml.feature.VectorIndexer
+import org.apache.spark.ml.regression.DecisionTreeRegressionModel
+import org.apache.spark.ml.regression.DecisionTreeRegressor
// $example off$
+import org.apache.spark.sql.SQLContext
+
object DecisionTreeRegressionExample {
def main(args: Array[String]): Unit = {
val conf = new SparkConf().setAppName("DecisionTreeRegressionExample")
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala
index 872de51..629d322 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/ElementwiseProductExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.ElementwiseProduct
import org.apache.spark.mllib.linalg.Vectors
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object ElementwiseProductExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
index 474af7d..cd62a80 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeClassifierExample.scala
@@ -18,7 +18,6 @@
// scalastyle:off println
package org.apache.spark.examples.ml
-import org.apache.spark.sql.SQLContext
import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.Pipeline
@@ -26,6 +25,7 @@ import org.apache.spark.ml.classification.{GBTClassificationModel, GBTClassifier
import org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
import org.apache.spark.ml.feature.{IndexToString, StringIndexer, VectorIndexer}
// $example off$
+import org.apache.spark.sql.SQLContext
object GradientBoostedTreeClassifierExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
index da1cd9c..b8cf962 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/GradientBoostedTreeRegressorExample.scala
@@ -18,7 +18,6 @@
// scalastyle:off println
package org.apache.spark.examples.ml
-import org.apache.spark.sql.SQLContext
import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.Pipeline
@@ -26,6 +25,7 @@ import org.apache.spark.ml.evaluation.RegressionEvaluator
import org.apache.spark.ml.feature.VectorIndexer
import org.apache.spark.ml.regression.{GBTRegressionModel, GBTRegressor}
// $example off$
+import org.apache.spark.sql.SQLContext
object GradientBoostedTreeRegressorExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala
index 52537e5..4cea09b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/IndexToStringExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
-import org.apache.spark.sql.SQLContext
import org.apache.spark.{SparkConf, SparkContext}
// $example on$
-import org.apache.spark.ml.feature.{StringIndexer, IndexToString}
+import org.apache.spark.ml.feature.{IndexToString, StringIndexer}
// $example off$
+import org.apache.spark.sql.SQLContext
object IndexToStringExample {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala
index 419ce3d..f9ddac7 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/LDAExample.scala
@@ -18,10 +18,10 @@
package org.apache.spark.examples.ml
// scalastyle:off println
-import org.apache.spark.{SparkContext, SparkConf}
-import org.apache.spark.mllib.linalg.{VectorUDT, Vectors}
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.clustering.LDA
+import org.apache.spark.mllib.linalg.{Vectors, VectorUDT}
import org.apache.spark.sql.{Row, SQLContext}
import org.apache.spark.sql.types.{StructField, StructType}
// $example off$
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala
index 22c824c..c7352b3 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/LinearRegressionWithElasticNetExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.regression.LinearRegression
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object LinearRegressionWithElasticNetExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala
index 4c42042..04c60c0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionSummaryExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.classification.{BinaryLogisticRegressionSummary, LogisticRegression}
// $example off$
import org.apache.spark.sql.SQLContext
import org.apache.spark.sql.functions.max
-import org.apache.spark.{SparkConf, SparkContext}
object LogisticRegressionSummaryExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionWithElasticNetExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionWithElasticNetExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionWithElasticNetExample.scala
index 9ee995b..f632960 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionWithElasticNetExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/LogisticRegressionWithElasticNetExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.classification.LogisticRegression
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object LogisticRegressionWithElasticNetExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala
index fb7f28c..9a03f69 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/MinMaxScalerExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.MinMaxScaler
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object MinMaxScalerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
index 9c98076..d7d1e82 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/MultilayerPerceptronClassifierExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.ml
-import org.apache.spark.{SparkContext, SparkConf}
-import org.apache.spark.sql.SQLContext
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.classification.MultilayerPerceptronClassifier
import org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
// $example off$
+import org.apache.spark.sql.SQLContext
/**
* An example for Multilayer Perceptron Classification.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala
index 8a85f71..77b913a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/NGramExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.NGram
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object NGramExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala
index 1990b55..6b33c16 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/NormalizerExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.Normalizer
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object NormalizerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala
index 66602e2..cb9fe65 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/OneHotEncoderExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.{OneHotEncoder, StringIndexer}
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object OneHotEncoderExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala
index b46faea..ccee3b2 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/OneVsRestExample.scala
@@ -22,10 +22,10 @@ import java.util.concurrent.TimeUnit.{NANOSECONDS => NANO}
import scopt.OptionParser
-import org.apache.spark.{SparkContext, SparkConf}
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.examples.mllib.AbstractParams
-import org.apache.spark.ml.classification.{OneVsRest, LogisticRegression}
+import org.apache.spark.ml.classification.{LogisticRegression, OneVsRest}
import org.apache.spark.ml.util.MetadataUtils
import org.apache.spark.mllib.evaluation.MulticlassMetrics
import org.apache.spark.mllib.linalg.Vector
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala
index 4c806f7..535652e 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/PCAExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.PCA
import org.apache.spark.mllib.linalg.Vectors
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object PCAExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala
index 39fb79a..3014008 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/PolynomialExpansionExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.PolynomialExpansion
import org.apache.spark.mllib.linalg.Vectors
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object PolynomialExpansionExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
index 8f29b7e..e64e673 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/QuantileDiscretizerExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.QuantileDiscretizer
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object QuantileDiscretizerExample {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala
index 286866e..bec831d 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/RFormulaExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.RFormula
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object RFormulaExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
index e79176c..6c9b52c 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestClassifierExample.scala
@@ -18,7 +18,6 @@
// scalastyle:off println
package org.apache.spark.examples.ml
-import org.apache.spark.sql.SQLContext
import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.Pipeline
@@ -26,6 +25,7 @@ import org.apache.spark.ml.classification.{RandomForestClassificationModel, Rand
import org.apache.spark.ml.evaluation.MulticlassClassificationEvaluator
import org.apache.spark.ml.feature.{IndexToString, StringIndexer, VectorIndexer}
// $example off$
+import org.apache.spark.sql.SQLContext
object RandomForestClassifierExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
index acec143..4d2db01 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/RandomForestRegressorExample.scala
@@ -18,7 +18,6 @@
// scalastyle:off println
package org.apache.spark.examples.ml
-import org.apache.spark.sql.SQLContext
import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.Pipeline
@@ -26,6 +25,7 @@ import org.apache.spark.ml.evaluation.RegressionEvaluator
import org.apache.spark.ml.feature.VectorIndexer
import org.apache.spark.ml.regression.{RandomForestRegressionModel, RandomForestRegressor}
// $example off$
+import org.apache.spark.sql.SQLContext
object RandomForestRegressorExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala
index 014abd1..202925a 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/SQLTransformerExample.scala
@@ -18,12 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.SQLTransformer
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
-
object SQLTransformerExample {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala
index e0a41e3..e343967 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/StandardScalerExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.StandardScaler
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object StandardScalerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala
index 655ffce..8199be1 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/StopWordsRemoverExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.StopWordsRemover
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object StopWordsRemoverExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala
index 9fa494c..3f0e870 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/StringIndexerExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.StringIndexer
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object StringIndexerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala
index 40c33e4..28115f9 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/TfIdfExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.{HashingTF, IDF, Tokenizer}
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object TfIdfExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala
index 01e0d13..c667728 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/TokenizerExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.{RegexTokenizer, Tokenizer}
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object TokenizerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/TrainValidationSplitExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/TrainValidationSplitExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/TrainValidationSplitExample.scala
index cd1b0e9..fbba17e 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/TrainValidationSplitExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/TrainValidationSplitExample.scala
@@ -17,11 +17,11 @@
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.ml.evaluation.RegressionEvaluator
import org.apache.spark.ml.regression.LinearRegression
import org.apache.spark.ml.tuning.{ParamGridBuilder, TrainValidationSplit}
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
/**
* A simple example demonstrating model selection using TrainValidationSplit.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala
index d527924..768a8c0 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/VectorAssemblerExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.VectorAssembler
import org.apache.spark.mllib.linalg.Vectors
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object VectorAssemblerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
index 685891c..3bef37b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/VectorIndexerExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.VectorIndexer
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object VectorIndexerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala
index 04f1982..01377d8 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/VectorSlicerExample.scala
@@ -18,6 +18,7 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.attribute.{Attribute, AttributeGroup, NumericAttribute}
import org.apache.spark.ml.feature.VectorSlicer
@@ -26,7 +27,6 @@ import org.apache.spark.sql.Row
import org.apache.spark.sql.types.StructType
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object VectorSlicerExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala b/examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala
index 631ab4c..e77aa59 100644
--- a/examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/ml/Word2VecExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.ml
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.ml.feature.Word2Vec
// $example off$
import org.apache.spark.sql.SQLContext
-import org.apache.spark.{SparkConf, SparkContext}
object Word2VecExample {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala
index ca22dda..11e18c9 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/AssociationRulesExample.scala
@@ -18,13 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.fpm.AssociationRules
import org.apache.spark.mllib.fpm.FPGrowth.FreqItemset
// $example off$
-import org.apache.spark.{SparkConf, SparkContext}
-
object AssociationRulesExample {
def main(args: Array[String]) {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassification.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassification.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassification.scala
index 1a4016f..2282bd2 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassification.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassification.scala
@@ -24,8 +24,8 @@ import scopt.OptionParser
import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.mllib.classification.{LogisticRegressionWithLBFGS, SVMWithSGD}
import org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
+import org.apache.spark.mllib.optimization.{L1Updater, SquaredL2Updater}
import org.apache.spark.mllib.util.MLUtils
-import org.apache.spark.mllib.optimization.{SquaredL2Updater, L1Updater}
/**
* An example app for binary classification. Run with
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala
index 13a3782..ade33fc 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/BinaryClassificationMetricsExample.scala
@@ -18,13 +18,13 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.classification.LogisticRegressionWithLBFGS
import org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
import org.apache.spark.mllib.regression.LabeledPoint
import org.apache.spark.mllib.util.MLUtils
// $example off$
-import org.apache.spark.{SparkContext, SparkConf}
object BinaryClassificationMetricsExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/BisectingKMeansExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/BisectingKMeansExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/BisectingKMeansExample.scala
index 3a596cc..53d0b8f 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/BisectingKMeansExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/BisectingKMeansExample.scala
@@ -18,11 +18,11 @@
package org.apache.spark.examples.mllib
// scalastyle:off println
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.clustering.BisectingKMeans
import org.apache.spark.mllib.linalg.{Vector, Vectors}
// $example off$
-import org.apache.spark.{SparkConf, SparkContext}
/**
* An example demonstrating a bisecting k-means clustering in spark.mllib.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/Correlations.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/Correlations.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/Correlations.scala
index 026d4ec..e003f35 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/Correlations.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/Correlations.scala
@@ -20,10 +20,9 @@ package org.apache.spark.examples.mllib
import scopt.OptionParser
+import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.mllib.stat.Statistics
import org.apache.spark.mllib.util.MLUtils
-import org.apache.spark.{SparkConf, SparkContext}
-
/**
* An example app for summarizing multivariate data from a file. Run with
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/CosineSimilarity.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/CosineSimilarity.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/CosineSimilarity.scala
index 69988cc..eda211b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/CosineSimilarity.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/CosineSimilarity.scala
@@ -20,10 +20,10 @@ package org.apache.spark.examples.mllib
import scopt.OptionParser
+import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.SparkContext._
import org.apache.spark.mllib.linalg.Vectors
import org.apache.spark.mllib.linalg.distributed.{MatrixEntry, RowMatrix}
-import org.apache.spark.{SparkConf, SparkContext}
/**
* Compute the similar columns of a matrix, using cosine similarity.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala
index d427bba..c6c7c6f 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeClassificationExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.tree.DecisionTree
import org.apache.spark.mllib.tree.model.DecisionTreeModel
import org.apache.spark.mllib.util.MLUtils
// $example off$
-import org.apache.spark.{SparkConf, SparkContext}
object DecisionTreeClassificationExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala
index fb05e7d..9c8baed 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRegressionExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.tree.DecisionTree
import org.apache.spark.mllib.tree.model.DecisionTreeModel
import org.apache.spark.mllib.util.MLUtils
// $example off$
-import org.apache.spark.{SparkConf, SparkContext}
object DecisionTreeRegressionExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRunner.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRunner.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRunner.scala
index cc6bce3..c263f4f 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRunner.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/DecisionTreeRunner.scala
@@ -26,7 +26,7 @@ import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.mllib.evaluation.MulticlassMetrics
import org.apache.spark.mllib.linalg.Vector
import org.apache.spark.mllib.regression.LabeledPoint
-import org.apache.spark.mllib.tree.{DecisionTree, RandomForest, impurity}
+import org.apache.spark.mllib.tree.{impurity, DecisionTree, RandomForest}
import org.apache.spark.mllib.tree.configuration.{Algo, Strategy}
import org.apache.spark.mllib.tree.configuration.Algo._
import org.apache.spark.mllib.util.MLUtils
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala
index 14b9305..a7a3ead 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/FPGrowthExample.scala
@@ -20,8 +20,8 @@ package org.apache.spark.examples.mllib
import scopt.OptionParser
-import org.apache.spark.mllib.fpm.FPGrowth
import org.apache.spark.{SparkConf, SparkContext}
+import org.apache.spark.mllib.fpm.FPGrowth
/**
* Example for mining frequent itemsets using FP-growth.
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostedTreesRunner.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostedTreesRunner.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostedTreesRunner.scala
index e16a6bf..b0144ef 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostedTreesRunner.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostedTreesRunner.scala
@@ -23,10 +23,9 @@ import scopt.OptionParser
import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.mllib.evaluation.MulticlassMetrics
import org.apache.spark.mllib.tree.GradientBoostedTrees
-import org.apache.spark.mllib.tree.configuration.{BoostingStrategy, Algo}
+import org.apache.spark.mllib.tree.configuration.{Algo, BoostingStrategy}
import org.apache.spark.util.Utils
-
/**
* An example runner for Gradient Boosting using decision trees as weak learners. Run with
* {{{
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala
index 139e1f9..0ec2e11 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingClassificationExample.scala
@@ -18,7 +18,7 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
-import org.apache.spark.{SparkContext, SparkConf}
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.tree.GradientBoostedTrees
import org.apache.spark.mllib.tree.configuration.BoostingStrategy
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala
index 3dc86da..b87ba0d 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/GradientBoostingRegressionExample.scala
@@ -18,7 +18,7 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
-import org.apache.spark.{SparkContext, SparkConf}
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.tree.GradientBoostedTrees
import org.apache.spark.mllib.tree.configuration.BoostingStrategy
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala
index 52ac9ae..3834ea8 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/IsotonicRegressionExample.scala
@@ -18,10 +18,10 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.regression.{IsotonicRegression, IsotonicRegressionModel}
// $example off$
-import org.apache.spark.{SparkConf, SparkContext}
object IsotonicRegressionExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala
index 61d2e77..75a0419 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/LBFGSExample.scala
@@ -18,6 +18,7 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.classification.LogisticRegressionModel
import org.apache.spark.mllib.evaluation.BinaryClassificationMetrics
@@ -26,8 +27,6 @@ import org.apache.spark.mllib.optimization.{LBFGS, LogisticGradient, SquaredL2Up
import org.apache.spark.mllib.util.MLUtils
// $example off$
-import org.apache.spark.{SparkConf, SparkContext}
-
object LBFGSExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala
index 70010b0..d283235 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/LDAExample.scala
@@ -18,16 +18,16 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.log4j.{Level, Logger}
import scopt.OptionParser
-import org.apache.log4j.{Level, Logger}
+import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.ml.Pipeline
import org.apache.spark.ml.feature.{CountVectorizer, CountVectorizerModel, RegexTokenizer, StopWordsRemover}
import org.apache.spark.mllib.clustering.{DistributedLDAModel, EMLDAOptimizer, LDA, OnlineLDAOptimizer}
import org.apache.spark.mllib.linalg.Vector
import org.apache.spark.rdd.RDD
import org.apache.spark.sql.{Row, SQLContext}
-import org.apache.spark.{SparkConf, SparkContext}
/**
* An example Latent Dirichlet Allocation (LDA) app. Run with
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegression.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegression.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegression.scala
index 8878061..f87611f 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegression.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/LinearRegression.scala
@@ -22,9 +22,9 @@ import org.apache.log4j.{Level, Logger}
import scopt.OptionParser
import org.apache.spark.{SparkConf, SparkContext}
+import org.apache.spark.mllib.optimization.{L1Updater, SimpleUpdater, SquaredL2Updater}
import org.apache.spark.mllib.regression.LinearRegressionWithSGD
import org.apache.spark.mllib.util.MLUtils
-import org.apache.spark.mllib.optimization.{SimpleUpdater, SquaredL2Updater, L1Updater}
/**
* An example app for linear regression. Run with
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/MultiLabelMetricsExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/MultiLabelMetricsExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/MultiLabelMetricsExample.scala
index 4503c15..c0d447b 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/MultiLabelMetricsExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/MultiLabelMetricsExample.scala
@@ -18,11 +18,11 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.evaluation.MultilabelMetrics
import org.apache.spark.rdd.RDD
// $example off$
-import org.apache.spark.{SparkContext, SparkConf}
object MultiLabelMetricsExample {
def main(args: Array[String]): Unit = {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/MulticlassMetricsExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/MulticlassMetricsExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/MulticlassMetricsExample.scala
index 0904449..4f925ed 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/MulticlassMetricsExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/MulticlassMetricsExample.scala
@@ -18,13 +18,13 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.classification.LogisticRegressionWithLBFGS
import org.apache.spark.mllib.evaluation.MulticlassMetrics
import org.apache.spark.mllib.regression.LabeledPoint
import org.apache.spark.mllib.util.MLUtils
// $example off$
-import org.apache.spark.{SparkContext, SparkConf}
object MulticlassMetricsExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/MultivariateSummarizer.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/MultivariateSummarizer.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/MultivariateSummarizer.scala
index 5f839c7..3c59817 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/MultivariateSummarizer.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/MultivariateSummarizer.scala
@@ -20,11 +20,10 @@ package org.apache.spark.examples.mllib
import scopt.OptionParser
+import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.mllib.linalg.Vectors
import org.apache.spark.mllib.stat.MultivariateOnlineSummarizer
import org.apache.spark.mllib.util.MLUtils
-import org.apache.spark.{SparkConf, SparkContext}
-
/**
* An example app for summarizing multivariate data from a file. Run with
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/NaiveBayesExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/NaiveBayesExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/NaiveBayesExample.scala
index a7a47c2..8bae1b9 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/NaiveBayesExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/NaiveBayesExample.scala
@@ -18,12 +18,12 @@
// scalastyle:off println
package org.apache.spark.examples.mllib
+import org.apache.spark.{SparkConf, SparkContext}
// $example on$
import org.apache.spark.mllib.classification.{NaiveBayes, NaiveBayesModel}
import org.apache.spark.mllib.linalg.Vectors
import org.apache.spark.mllib.regression.LabeledPoint
// $example off$
-import org.apache.spark.{SparkConf, SparkContext}
object NaiveBayesExample {
http://git-wip-us.apache.org/repos/asf/spark/blob/7058dc11/examples/src/main/scala/org/apache/spark/examples/mllib/PowerIterationClusteringExample.scala
----------------------------------------------------------------------
diff --git a/examples/src/main/scala/org/apache/spark/examples/mllib/PowerIterationClusteringExample.scala b/examples/src/main/scala/org/apache/spark/examples/mllib/PowerIterationClusteringExample.scala
index 0723223..9208d8e 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/PowerIterationClusteringExample.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/PowerIterationClusteringExample.scala
@@ -21,9 +21,9 @@ package org.apache.spark.examples.mllib
import org.apache.log4j.{Level, Logger}
import scopt.OptionParser
+import org.apache.spark.{SparkConf, SparkContext}
import org.apache.spark.mllib.clustering.PowerIterationClustering
import org.apache.spark.rdd.RDD
-import org.apache.spark.{SparkConf, SparkContext}
/**
* An example Power Iteration Clustering http://www.icml2010.org/papers/387.pdf app.
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org