You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by sr...@apache.org on 2016/03/21 08:58:20 UTC
[2/3] spark git commit: [SPARK-14011][CORE][SQL] Enable `LineLength`
Java checkstyle rule
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/core/src/test/java/org/apache/spark/JavaAPISuite.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/spark/JavaAPISuite.java b/core/src/test/java/org/apache/spark/JavaAPISuite.java
index c1036b8..0f65554 100644
--- a/core/src/test/java/org/apache/spark/JavaAPISuite.java
+++ b/core/src/test/java/org/apache/spark/JavaAPISuite.java
@@ -54,7 +54,7 @@ import org.apache.hadoop.mapred.SequenceFileInputFormat;
import org.apache.hadoop.mapred.SequenceFileOutputFormat;
import org.apache.hadoop.mapreduce.Job;
import org.junit.After;
-import org.junit.Assert;
+import static org.junit.Assert.*;
import org.junit.Before;
import org.junit.Test;
@@ -102,19 +102,19 @@ public class JavaAPISuite implements Serializable {
JavaRDD<String> s2 = sc.parallelize(strings);
// Varargs
JavaRDD<String> sUnion = sc.union(s1, s2);
- Assert.assertEquals(4, sUnion.count());
+ assertEquals(4, sUnion.count());
// List
List<JavaRDD<String>> list = new ArrayList<>();
list.add(s2);
sUnion = sc.union(s1, list);
- Assert.assertEquals(4, sUnion.count());
+ assertEquals(4, sUnion.count());
// Union of JavaDoubleRDDs
List<Double> doubles = Arrays.asList(1.0, 2.0);
JavaDoubleRDD d1 = sc.parallelizeDoubles(doubles);
JavaDoubleRDD d2 = sc.parallelizeDoubles(doubles);
JavaDoubleRDD dUnion = sc.union(d1, d2);
- Assert.assertEquals(4, dUnion.count());
+ assertEquals(4, dUnion.count());
// Union of JavaPairRDDs
List<Tuple2<Integer, Integer>> pairs = new ArrayList<>();
@@ -123,7 +123,7 @@ public class JavaAPISuite implements Serializable {
JavaPairRDD<Integer, Integer> p1 = sc.parallelizePairs(pairs);
JavaPairRDD<Integer, Integer> p2 = sc.parallelizePairs(pairs);
JavaPairRDD<Integer, Integer> pUnion = sc.union(p1, p2);
- Assert.assertEquals(4, pUnion.count());
+ assertEquals(4, pUnion.count());
}
@SuppressWarnings("unchecked")
@@ -135,17 +135,17 @@ public class JavaAPISuite implements Serializable {
JavaRDD<Integer> s2 = sc.parallelize(ints2);
JavaRDD<Integer> intersections = s1.intersection(s2);
- Assert.assertEquals(3, intersections.count());
+ assertEquals(3, intersections.count());
JavaRDD<Integer> empty = sc.emptyRDD();
JavaRDD<Integer> emptyIntersection = empty.intersection(s2);
- Assert.assertEquals(0, emptyIntersection.count());
+ assertEquals(0, emptyIntersection.count());
List<Double> doubles = Arrays.asList(1.0, 2.0);
JavaDoubleRDD d1 = sc.parallelizeDoubles(doubles);
JavaDoubleRDD d2 = sc.parallelizeDoubles(doubles);
JavaDoubleRDD dIntersection = d1.intersection(d2);
- Assert.assertEquals(2, dIntersection.count());
+ assertEquals(2, dIntersection.count());
List<Tuple2<Integer, Integer>> pairs = new ArrayList<>();
pairs.add(new Tuple2<>(1, 2));
@@ -153,7 +153,7 @@ public class JavaAPISuite implements Serializable {
JavaPairRDD<Integer, Integer> p1 = sc.parallelizePairs(pairs);
JavaPairRDD<Integer, Integer> p2 = sc.parallelizePairs(pairs);
JavaPairRDD<Integer, Integer> pIntersection = p1.intersection(p2);
- Assert.assertEquals(2, pIntersection.count());
+ assertEquals(2, pIntersection.count());
}
@Test
@@ -162,9 +162,9 @@ public class JavaAPISuite implements Serializable {
JavaRDD<Integer> rdd = sc.parallelize(ints);
// the seeds here are "magic" to make this work out nicely
JavaRDD<Integer> sample20 = rdd.sample(true, 0.2, 8);
- Assert.assertEquals(2, sample20.count());
+ assertEquals(2, sample20.count());
JavaRDD<Integer> sample20WithoutReplacement = rdd.sample(false, 0.2, 2);
- Assert.assertEquals(2, sample20WithoutReplacement.count());
+ assertEquals(2, sample20WithoutReplacement.count());
}
@Test
@@ -176,13 +176,13 @@ public class JavaAPISuite implements Serializable {
JavaRDD<Integer> rdd = sc.parallelize(ints);
JavaRDD<Integer>[] splits = rdd.randomSplit(new double[] { 0.4, 0.6, 1.0 }, 31);
// the splits aren't perfect -- not enough data for them to be -- just check they're about right
- Assert.assertEquals(3, splits.length);
+ assertEquals(3, splits.length);
long s0 = splits[0].count();
long s1 = splits[1].count();
long s2 = splits[2].count();
- Assert.assertTrue(s0 + " not within expected range", s0 > 150 && s0 < 250);
- Assert.assertTrue(s1 + " not within expected range", s1 > 250 && s0 < 350);
- Assert.assertTrue(s2 + " not within expected range", s2 > 430 && s2 < 570);
+ assertTrue(s0 + " not within expected range", s0 > 150 && s0 < 250);
+ assertTrue(s1 + " not within expected range", s1 > 250 && s0 < 350);
+ assertTrue(s2 + " not within expected range", s2 > 430 && s2 < 570);
}
@Test
@@ -196,17 +196,17 @@ public class JavaAPISuite implements Serializable {
// Default comparator
JavaPairRDD<Integer, Integer> sortedRDD = rdd.sortByKey();
- Assert.assertEquals(new Tuple2<>(-1, 1), sortedRDD.first());
+ assertEquals(new Tuple2<>(-1, 1), sortedRDD.first());
List<Tuple2<Integer, Integer>> sortedPairs = sortedRDD.collect();
- Assert.assertEquals(new Tuple2<>(0, 4), sortedPairs.get(1));
- Assert.assertEquals(new Tuple2<>(3, 2), sortedPairs.get(2));
+ assertEquals(new Tuple2<>(0, 4), sortedPairs.get(1));
+ assertEquals(new Tuple2<>(3, 2), sortedPairs.get(2));
// Custom comparator
sortedRDD = rdd.sortByKey(Collections.<Integer>reverseOrder(), false);
- Assert.assertEquals(new Tuple2<>(-1, 1), sortedRDD.first());
+ assertEquals(new Tuple2<>(-1, 1), sortedRDD.first());
sortedPairs = sortedRDD.collect();
- Assert.assertEquals(new Tuple2<>(0, 4), sortedPairs.get(1));
- Assert.assertEquals(new Tuple2<>(3, 2), sortedPairs.get(2));
+ assertEquals(new Tuple2<>(0, 4), sortedPairs.get(1));
+ assertEquals(new Tuple2<>(3, 2), sortedPairs.get(2));
}
@SuppressWarnings("unchecked")
@@ -235,19 +235,19 @@ public class JavaAPISuite implements Serializable {
JavaPairRDD<Integer, Integer> repartitioned =
rdd.repartitionAndSortWithinPartitions(partitioner);
- Assert.assertTrue(repartitioned.partitioner().isPresent());
- Assert.assertEquals(repartitioned.partitioner().get(), partitioner);
+ assertTrue(repartitioned.partitioner().isPresent());
+ assertEquals(repartitioned.partitioner().get(), partitioner);
List<List<Tuple2<Integer, Integer>>> partitions = repartitioned.glom().collect();
- Assert.assertEquals(partitions.get(0),
+ assertEquals(partitions.get(0),
Arrays.asList(new Tuple2<>(0, 5), new Tuple2<>(0, 8), new Tuple2<>(2, 6)));
- Assert.assertEquals(partitions.get(1),
+ assertEquals(partitions.get(1),
Arrays.asList(new Tuple2<>(1, 3), new Tuple2<>(3, 8), new Tuple2<>(3, 8)));
}
@Test
public void emptyRDD() {
JavaRDD<String> rdd = sc.emptyRDD();
- Assert.assertEquals("Empty RDD shouldn't have any values", 0, rdd.count());
+ assertEquals("Empty RDD shouldn't have any values", 0, rdd.count());
}
@Test
@@ -260,17 +260,18 @@ public class JavaAPISuite implements Serializable {
JavaRDD<Tuple2<Integer, Integer>> rdd = sc.parallelize(pairs);
// compare on first value
- JavaRDD<Tuple2<Integer, Integer>> sortedRDD = rdd.sortBy(new Function<Tuple2<Integer, Integer>, Integer>() {
+ JavaRDD<Tuple2<Integer, Integer>> sortedRDD =
+ rdd.sortBy(new Function<Tuple2<Integer, Integer>, Integer>() {
@Override
public Integer call(Tuple2<Integer, Integer> t) {
return t._1();
}
}, true, 2);
- Assert.assertEquals(new Tuple2<>(-1, 1), sortedRDD.first());
+ assertEquals(new Tuple2<>(-1, 1), sortedRDD.first());
List<Tuple2<Integer, Integer>> sortedPairs = sortedRDD.collect();
- Assert.assertEquals(new Tuple2<>(0, 4), sortedPairs.get(1));
- Assert.assertEquals(new Tuple2<>(3, 2), sortedPairs.get(2));
+ assertEquals(new Tuple2<>(0, 4), sortedPairs.get(1));
+ assertEquals(new Tuple2<>(3, 2), sortedPairs.get(2));
// compare on second value
sortedRDD = rdd.sortBy(new Function<Tuple2<Integer, Integer>, Integer>() {
@@ -279,10 +280,10 @@ public class JavaAPISuite implements Serializable {
return t._2();
}
}, true, 2);
- Assert.assertEquals(new Tuple2<>(-1, 1), sortedRDD.first());
+ assertEquals(new Tuple2<>(-1, 1), sortedRDD.first());
sortedPairs = sortedRDD.collect();
- Assert.assertEquals(new Tuple2<>(3, 2), sortedPairs.get(1));
- Assert.assertEquals(new Tuple2<>(0, 4), sortedPairs.get(2));
+ assertEquals(new Tuple2<>(3, 2), sortedPairs.get(1));
+ assertEquals(new Tuple2<>(0, 4), sortedPairs.get(2));
}
@Test
@@ -295,7 +296,7 @@ public class JavaAPISuite implements Serializable {
accum.add(1);
}
});
- Assert.assertEquals(2, accum.value().intValue());
+ assertEquals(2, accum.value().intValue());
}
@Test
@@ -311,7 +312,7 @@ public class JavaAPISuite implements Serializable {
}
}
});
- Assert.assertEquals(2, accum.value().intValue());
+ assertEquals(2, accum.value().intValue());
}
@Test
@@ -319,7 +320,7 @@ public class JavaAPISuite implements Serializable {
List<Integer> correct = Arrays.asList(1, 2, 3, 4);
JavaRDD<Integer> rdd = sc.parallelize(correct);
List<Integer> result = Lists.newArrayList(rdd.toLocalIterator());
- Assert.assertEquals(correct, result);
+ assertEquals(correct, result);
}
@Test
@@ -327,7 +328,7 @@ public class JavaAPISuite implements Serializable {
List<Integer> dataArray = Arrays.asList(1, 2, 3, 4);
JavaPairRDD<Integer, Long> zip = sc.parallelize(dataArray).zipWithUniqueId();
JavaRDD<Long> indexes = zip.values();
- Assert.assertEquals(4, new HashSet<>(indexes.collect()).size());
+ assertEquals(4, new HashSet<>(indexes.collect()).size());
}
@Test
@@ -336,7 +337,7 @@ public class JavaAPISuite implements Serializable {
JavaPairRDD<Integer, Long> zip = sc.parallelize(dataArray).zipWithIndex();
JavaRDD<Long> indexes = zip.values();
List<Long> correctIndexes = Arrays.asList(0L, 1L, 2L, 3L);
- Assert.assertEquals(correctIndexes, indexes.collect());
+ assertEquals(correctIndexes, indexes.collect());
}
@SuppressWarnings("unchecked")
@@ -347,8 +348,8 @@ public class JavaAPISuite implements Serializable {
new Tuple2<>("Oranges", "Fruit"),
new Tuple2<>("Oranges", "Citrus")
));
- Assert.assertEquals(2, categories.lookup("Oranges").size());
- Assert.assertEquals(2, Iterables.size(categories.groupByKey().lookup("Oranges").get(0)));
+ assertEquals(2, categories.lookup("Oranges").size());
+ assertEquals(2, Iterables.size(categories.groupByKey().lookup("Oranges").get(0)));
}
@Test
@@ -361,14 +362,14 @@ public class JavaAPISuite implements Serializable {
}
};
JavaPairRDD<Boolean, Iterable<Integer>> oddsAndEvens = rdd.groupBy(isOdd);
- Assert.assertEquals(2, oddsAndEvens.count());
- Assert.assertEquals(2, Iterables.size(oddsAndEvens.lookup(true).get(0))); // Evens
- Assert.assertEquals(5, Iterables.size(oddsAndEvens.lookup(false).get(0))); // Odds
+ assertEquals(2, oddsAndEvens.count());
+ assertEquals(2, Iterables.size(oddsAndEvens.lookup(true).get(0))); // Evens
+ assertEquals(5, Iterables.size(oddsAndEvens.lookup(false).get(0))); // Odds
oddsAndEvens = rdd.groupBy(isOdd, 1);
- Assert.assertEquals(2, oddsAndEvens.count());
- Assert.assertEquals(2, Iterables.size(oddsAndEvens.lookup(true).get(0))); // Evens
- Assert.assertEquals(5, Iterables.size(oddsAndEvens.lookup(false).get(0))); // Odds
+ assertEquals(2, oddsAndEvens.count());
+ assertEquals(2, Iterables.size(oddsAndEvens.lookup(true).get(0))); // Evens
+ assertEquals(5, Iterables.size(oddsAndEvens.lookup(false).get(0))); // Odds
}
@Test
@@ -384,14 +385,14 @@ public class JavaAPISuite implements Serializable {
};
JavaPairRDD<Integer, Integer> pairRDD = rdd.zip(rdd);
JavaPairRDD<Boolean, Iterable<Tuple2<Integer, Integer>>> oddsAndEvens = pairRDD.groupBy(areOdd);
- Assert.assertEquals(2, oddsAndEvens.count());
- Assert.assertEquals(2, Iterables.size(oddsAndEvens.lookup(true).get(0))); // Evens
- Assert.assertEquals(5, Iterables.size(oddsAndEvens.lookup(false).get(0))); // Odds
+ assertEquals(2, oddsAndEvens.count());
+ assertEquals(2, Iterables.size(oddsAndEvens.lookup(true).get(0))); // Evens
+ assertEquals(5, Iterables.size(oddsAndEvens.lookup(false).get(0))); // Odds
oddsAndEvens = pairRDD.groupBy(areOdd, 1);
- Assert.assertEquals(2, oddsAndEvens.count());
- Assert.assertEquals(2, Iterables.size(oddsAndEvens.lookup(true).get(0))); // Evens
- Assert.assertEquals(5, Iterables.size(oddsAndEvens.lookup(false).get(0))); // Odds
+ assertEquals(2, oddsAndEvens.count());
+ assertEquals(2, Iterables.size(oddsAndEvens.lookup(true).get(0))); // Evens
+ assertEquals(5, Iterables.size(oddsAndEvens.lookup(false).get(0))); // Odds
}
@SuppressWarnings("unchecked")
@@ -408,8 +409,8 @@ public class JavaAPISuite implements Serializable {
};
JavaPairRDD<Integer, Integer> pairRDD = rdd.zip(rdd);
JavaPairRDD<String, Tuple2<Integer, Integer>> keyed = pairRDD.keyBy(sumToString);
- Assert.assertEquals(7, keyed.count());
- Assert.assertEquals(1, (long) keyed.lookup("2").get(0)._1());
+ assertEquals(7, keyed.count());
+ assertEquals(1, (long) keyed.lookup("2").get(0)._1());
}
@SuppressWarnings("unchecked")
@@ -426,8 +427,8 @@ public class JavaAPISuite implements Serializable {
));
JavaPairRDD<String, Tuple2<Iterable<String>, Iterable<Integer>>> cogrouped =
categories.cogroup(prices);
- Assert.assertEquals("[Fruit, Citrus]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._1()));
- Assert.assertEquals("[2]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._2()));
+ assertEquals("[Fruit, Citrus]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._1()));
+ assertEquals("[2]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._2()));
cogrouped.collect();
}
@@ -451,9 +452,9 @@ public class JavaAPISuite implements Serializable {
JavaPairRDD<String, Tuple3<Iterable<String>, Iterable<Integer>, Iterable<Integer>>> cogrouped =
categories.cogroup(prices, quantities);
- Assert.assertEquals("[Fruit, Citrus]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._1()));
- Assert.assertEquals("[2]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._2()));
- Assert.assertEquals("[42]", Iterables.toString(cogrouped.lookup("Apples").get(0)._3()));
+ assertEquals("[Fruit, Citrus]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._1()));
+ assertEquals("[2]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._2()));
+ assertEquals("[42]", Iterables.toString(cogrouped.lookup("Apples").get(0)._3()));
cogrouped.collect();
@@ -480,12 +481,12 @@ public class JavaAPISuite implements Serializable {
new Tuple2<>("Apples", "US")
));
- JavaPairRDD<String, Tuple4<Iterable<String>, Iterable<Integer>, Iterable<Integer>, Iterable<String>>> cogrouped =
- categories.cogroup(prices, quantities, countries);
- Assert.assertEquals("[Fruit, Citrus]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._1()));
- Assert.assertEquals("[2]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._2()));
- Assert.assertEquals("[42]", Iterables.toString(cogrouped.lookup("Apples").get(0)._3()));
- Assert.assertEquals("[BR]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._4()));
+ JavaPairRDD<String, Tuple4<Iterable<String>, Iterable<Integer>, Iterable<Integer>,
+ Iterable<String>>> cogrouped = categories.cogroup(prices, quantities, countries);
+ assertEquals("[Fruit, Citrus]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._1()));
+ assertEquals("[2]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._2()));
+ assertEquals("[42]", Iterables.toString(cogrouped.lookup("Apples").get(0)._3()));
+ assertEquals("[BR]", Iterables.toString(cogrouped.lookup("Oranges").get(0)._4()));
cogrouped.collect();
}
@@ -507,7 +508,7 @@ public class JavaAPISuite implements Serializable {
));
List<Tuple2<Integer,Tuple2<Integer,Optional<Character>>>> joined =
rdd1.leftOuterJoin(rdd2).collect();
- Assert.assertEquals(5, joined.size());
+ assertEquals(5, joined.size());
Tuple2<Integer,Tuple2<Integer,Optional<Character>>> firstUnmatched =
rdd1.leftOuterJoin(rdd2).filter(
new Function<Tuple2<Integer, Tuple2<Integer, Optional<Character>>>, Boolean>() {
@@ -516,7 +517,7 @@ public class JavaAPISuite implements Serializable {
return !tup._2()._2().isPresent();
}
}).first();
- Assert.assertEquals(3, firstUnmatched._1().intValue());
+ assertEquals(3, firstUnmatched._1().intValue());
}
@Test
@@ -530,10 +531,10 @@ public class JavaAPISuite implements Serializable {
};
int sum = rdd.fold(0, add);
- Assert.assertEquals(33, sum);
+ assertEquals(33, sum);
sum = rdd.reduce(add);
- Assert.assertEquals(33, sum);
+ assertEquals(33, sum);
}
@Test
@@ -547,7 +548,7 @@ public class JavaAPISuite implements Serializable {
};
for (int depth = 1; depth <= 10; depth++) {
int sum = rdd.treeReduce(add, depth);
- Assert.assertEquals(-5, sum);
+ assertEquals(-5, sum);
}
}
@@ -562,7 +563,7 @@ public class JavaAPISuite implements Serializable {
};
for (int depth = 1; depth <= 10; depth++) {
int sum = rdd.treeAggregate(0, add, add, depth);
- Assert.assertEquals(-5, sum);
+ assertEquals(-5, sum);
}
}
@@ -592,10 +593,10 @@ public class JavaAPISuite implements Serializable {
return a;
}
}).collectAsMap();
- Assert.assertEquals(3, sets.size());
- Assert.assertEquals(new HashSet<>(Arrays.asList(1)), sets.get(1));
- Assert.assertEquals(new HashSet<>(Arrays.asList(2)), sets.get(3));
- Assert.assertEquals(new HashSet<>(Arrays.asList(1, 3)), sets.get(5));
+ assertEquals(3, sets.size());
+ assertEquals(new HashSet<>(Arrays.asList(1)), sets.get(1));
+ assertEquals(new HashSet<>(Arrays.asList(2)), sets.get(3));
+ assertEquals(new HashSet<>(Arrays.asList(1, 3)), sets.get(5));
}
@SuppressWarnings("unchecked")
@@ -616,9 +617,9 @@ public class JavaAPISuite implements Serializable {
return a + b;
}
});
- Assert.assertEquals(1, sums.lookup(1).get(0).intValue());
- Assert.assertEquals(2, sums.lookup(2).get(0).intValue());
- Assert.assertEquals(3, sums.lookup(3).get(0).intValue());
+ assertEquals(1, sums.lookup(1).get(0).intValue());
+ assertEquals(2, sums.lookup(2).get(0).intValue());
+ assertEquals(3, sums.lookup(3).get(0).intValue());
}
@SuppressWarnings("unchecked")
@@ -639,14 +640,14 @@ public class JavaAPISuite implements Serializable {
return a + b;
}
});
- Assert.assertEquals(1, counts.lookup(1).get(0).intValue());
- Assert.assertEquals(2, counts.lookup(2).get(0).intValue());
- Assert.assertEquals(3, counts.lookup(3).get(0).intValue());
+ assertEquals(1, counts.lookup(1).get(0).intValue());
+ assertEquals(2, counts.lookup(2).get(0).intValue());
+ assertEquals(3, counts.lookup(3).get(0).intValue());
Map<Integer, Integer> localCounts = counts.collectAsMap();
- Assert.assertEquals(1, localCounts.get(1).intValue());
- Assert.assertEquals(2, localCounts.get(2).intValue());
- Assert.assertEquals(3, localCounts.get(3).intValue());
+ assertEquals(1, localCounts.get(1).intValue());
+ assertEquals(2, localCounts.get(2).intValue());
+ assertEquals(3, localCounts.get(3).intValue());
localCounts = rdd.reduceByKeyLocally(new Function2<Integer, Integer, Integer>() {
@Override
@@ -654,45 +655,45 @@ public class JavaAPISuite implements Serializable {
return a + b;
}
});
- Assert.assertEquals(1, localCounts.get(1).intValue());
- Assert.assertEquals(2, localCounts.get(2).intValue());
- Assert.assertEquals(3, localCounts.get(3).intValue());
+ assertEquals(1, localCounts.get(1).intValue());
+ assertEquals(2, localCounts.get(2).intValue());
+ assertEquals(3, localCounts.get(3).intValue());
}
@Test
public void approximateResults() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 1, 2, 3, 5, 8, 13));
Map<Integer, Long> countsByValue = rdd.countByValue();
- Assert.assertEquals(2, countsByValue.get(1).longValue());
- Assert.assertEquals(1, countsByValue.get(13).longValue());
+ assertEquals(2, countsByValue.get(1).longValue());
+ assertEquals(1, countsByValue.get(13).longValue());
PartialResult<Map<Integer, BoundedDouble>> approx = rdd.countByValueApprox(1);
Map<Integer, BoundedDouble> finalValue = approx.getFinalValue();
- Assert.assertEquals(2.0, finalValue.get(1).mean(), 0.01);
- Assert.assertEquals(1.0, finalValue.get(13).mean(), 0.01);
+ assertEquals(2.0, finalValue.get(1).mean(), 0.01);
+ assertEquals(1.0, finalValue.get(13).mean(), 0.01);
}
@Test
public void take() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 1, 2, 3, 5, 8, 13));
- Assert.assertEquals(1, rdd.first().intValue());
+ assertEquals(1, rdd.first().intValue());
rdd.take(2);
rdd.takeSample(false, 2, 42);
}
@Test
public void isEmpty() {
- Assert.assertTrue(sc.emptyRDD().isEmpty());
- Assert.assertTrue(sc.parallelize(new ArrayList<Integer>()).isEmpty());
- Assert.assertFalse(sc.parallelize(Arrays.asList(1)).isEmpty());
- Assert.assertTrue(sc.parallelize(Arrays.asList(1, 2, 3), 3).filter(
+ assertTrue(sc.emptyRDD().isEmpty());
+ assertTrue(sc.parallelize(new ArrayList<Integer>()).isEmpty());
+ assertFalse(sc.parallelize(Arrays.asList(1)).isEmpty());
+ assertTrue(sc.parallelize(Arrays.asList(1, 2, 3), 3).filter(
new Function<Integer,Boolean>() {
@Override
public Boolean call(Integer i) {
return i < 0;
}
}).isEmpty());
- Assert.assertFalse(sc.parallelize(Arrays.asList(1, 2, 3)).filter(
+ assertFalse(sc.parallelize(Arrays.asList(1, 2, 3)).filter(
new Function<Integer, Boolean>() {
@Override
public Boolean call(Integer i) {
@@ -706,35 +707,35 @@ public class JavaAPISuite implements Serializable {
JavaDoubleRDD doubleRDD = sc.parallelizeDoubles(Arrays.asList(1.0, 1.0, 2.0, 3.0, 5.0, 8.0));
JavaRDD<String> stringRDD = sc.parallelize(Arrays.asList("Hello", "World"));
JavaPairRDD<String, Double> cartesian = stringRDD.cartesian(doubleRDD);
- Assert.assertEquals(new Tuple2<>("Hello", 1.0), cartesian.first());
+ assertEquals(new Tuple2<>("Hello", 1.0), cartesian.first());
}
@Test
public void javaDoubleRDD() {
JavaDoubleRDD rdd = sc.parallelizeDoubles(Arrays.asList(1.0, 1.0, 2.0, 3.0, 5.0, 8.0));
JavaDoubleRDD distinct = rdd.distinct();
- Assert.assertEquals(5, distinct.count());
+ assertEquals(5, distinct.count());
JavaDoubleRDD filter = rdd.filter(new Function<Double, Boolean>() {
@Override
public Boolean call(Double x) {
return x > 2.0;
}
});
- Assert.assertEquals(3, filter.count());
+ assertEquals(3, filter.count());
JavaDoubleRDD union = rdd.union(rdd);
- Assert.assertEquals(12, union.count());
+ assertEquals(12, union.count());
union = union.cache();
- Assert.assertEquals(12, union.count());
+ assertEquals(12, union.count());
- Assert.assertEquals(20, rdd.sum(), 0.01);
+ assertEquals(20, rdd.sum(), 0.01);
StatCounter stats = rdd.stats();
- Assert.assertEquals(20, stats.sum(), 0.01);
- Assert.assertEquals(20/6.0, rdd.mean(), 0.01);
- Assert.assertEquals(20/6.0, rdd.mean(), 0.01);
- Assert.assertEquals(6.22222, rdd.variance(), 0.01);
- Assert.assertEquals(7.46667, rdd.sampleVariance(), 0.01);
- Assert.assertEquals(2.49444, rdd.stdev(), 0.01);
- Assert.assertEquals(2.73252, rdd.sampleStdev(), 0.01);
+ assertEquals(20, stats.sum(), 0.01);
+ assertEquals(20/6.0, rdd.mean(), 0.01);
+ assertEquals(20/6.0, rdd.mean(), 0.01);
+ assertEquals(6.22222, rdd.variance(), 0.01);
+ assertEquals(7.46667, rdd.sampleVariance(), 0.01);
+ assertEquals(2.49444, rdd.stdev(), 0.01);
+ assertEquals(2.73252, rdd.sampleStdev(), 0.01);
rdd.first();
rdd.take(5);
@@ -747,13 +748,13 @@ public class JavaAPISuite implements Serializable {
Tuple2<double[], long[]> results = rdd.histogram(2);
double[] expected_buckets = {1.0, 2.5, 4.0};
long[] expected_counts = {2, 2};
- Assert.assertArrayEquals(expected_buckets, results._1(), 0.1);
- Assert.assertArrayEquals(expected_counts, results._2());
+ assertArrayEquals(expected_buckets, results._1(), 0.1);
+ assertArrayEquals(expected_counts, results._2());
// Test with provided buckets
long[] histogram = rdd.histogram(expected_buckets);
- Assert.assertArrayEquals(expected_counts, histogram);
+ assertArrayEquals(expected_counts, histogram);
// SPARK-5744
- Assert.assertArrayEquals(
+ assertArrayEquals(
new long[] {0},
sc.parallelizeDoubles(new ArrayList<Double>(0), 1).histogram(new double[]{0.0, 1.0}));
}
@@ -769,42 +770,42 @@ public class JavaAPISuite implements Serializable {
public void max() {
JavaDoubleRDD rdd = sc.parallelizeDoubles(Arrays.asList(1.0, 2.0, 3.0, 4.0));
double max = rdd.max(new DoubleComparator());
- Assert.assertEquals(4.0, max, 0.001);
+ assertEquals(4.0, max, 0.001);
}
@Test
public void min() {
JavaDoubleRDD rdd = sc.parallelizeDoubles(Arrays.asList(1.0, 2.0, 3.0, 4.0));
double max = rdd.min(new DoubleComparator());
- Assert.assertEquals(1.0, max, 0.001);
+ assertEquals(1.0, max, 0.001);
}
@Test
public void naturalMax() {
JavaDoubleRDD rdd = sc.parallelizeDoubles(Arrays.asList(1.0, 2.0, 3.0, 4.0));
double max = rdd.max();
- Assert.assertEquals(4.0, max, 0.0);
+ assertEquals(4.0, max, 0.0);
}
@Test
public void naturalMin() {
JavaDoubleRDD rdd = sc.parallelizeDoubles(Arrays.asList(1.0, 2.0, 3.0, 4.0));
double max = rdd.min();
- Assert.assertEquals(1.0, max, 0.0);
+ assertEquals(1.0, max, 0.0);
}
@Test
public void takeOrdered() {
JavaDoubleRDD rdd = sc.parallelizeDoubles(Arrays.asList(1.0, 2.0, 3.0, 4.0));
- Assert.assertEquals(Arrays.asList(1.0, 2.0), rdd.takeOrdered(2, new DoubleComparator()));
- Assert.assertEquals(Arrays.asList(1.0, 2.0), rdd.takeOrdered(2));
+ assertEquals(Arrays.asList(1.0, 2.0), rdd.takeOrdered(2, new DoubleComparator()));
+ assertEquals(Arrays.asList(1.0, 2.0), rdd.takeOrdered(2));
}
@Test
public void top() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
List<Integer> top2 = rdd.top(2);
- Assert.assertEquals(Arrays.asList(4, 3), top2);
+ assertEquals(Arrays.asList(4, 3), top2);
}
private static class AddInts implements Function2<Integer, Integer, Integer> {
@@ -818,7 +819,7 @@ public class JavaAPISuite implements Serializable {
public void reduce() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
int sum = rdd.reduce(new AddInts());
- Assert.assertEquals(10, sum);
+ assertEquals(10, sum);
}
@Test
@@ -830,21 +831,21 @@ public class JavaAPISuite implements Serializable {
return v1 + v2;
}
});
- Assert.assertEquals(10.0, sum, 0.001);
+ assertEquals(10.0, sum, 0.001);
}
@Test
public void fold() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
int sum = rdd.fold(0, new AddInts());
- Assert.assertEquals(10, sum);
+ assertEquals(10, sum);
}
@Test
public void aggregate() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4));
int sum = rdd.aggregate(0, new AddInts(), new AddInts());
- Assert.assertEquals(10, sum);
+ assertEquals(10, sum);
}
@Test
@@ -884,8 +885,8 @@ public class JavaAPISuite implements Serializable {
return Arrays.asList(x.split(" ")).iterator();
}
});
- Assert.assertEquals("Hello", words.first());
- Assert.assertEquals(11, words.count());
+ assertEquals("Hello", words.first());
+ assertEquals(11, words.count());
JavaPairRDD<String, String> pairsRDD = rdd.flatMapToPair(
new PairFlatMapFunction<String, String, String>() {
@@ -899,8 +900,8 @@ public class JavaAPISuite implements Serializable {
}
}
);
- Assert.assertEquals(new Tuple2<>("Hello", "Hello"), pairsRDD.first());
- Assert.assertEquals(11, pairsRDD.count());
+ assertEquals(new Tuple2<>("Hello", "Hello"), pairsRDD.first());
+ assertEquals(11, pairsRDD.count());
JavaDoubleRDD doubles = rdd.flatMapToDouble(new DoubleFlatMapFunction<String>() {
@Override
@@ -912,8 +913,8 @@ public class JavaAPISuite implements Serializable {
return lengths.iterator();
}
});
- Assert.assertEquals(5.0, doubles.first(), 0.01);
- Assert.assertEquals(11, pairsRDD.count());
+ assertEquals(5.0, doubles.first(), 0.01);
+ assertEquals(11, pairsRDD.count());
}
@SuppressWarnings("unchecked")
@@ -959,7 +960,7 @@ public class JavaAPISuite implements Serializable {
return Collections.singletonList(sum).iterator();
}
});
- Assert.assertEquals("[3, 7]", partitionSums.collect().toString());
+ assertEquals("[3, 7]", partitionSums.collect().toString());
}
@@ -977,7 +978,7 @@ public class JavaAPISuite implements Serializable {
return Collections.singletonList(sum).iterator();
}
}, false);
- Assert.assertEquals("[3, 7]", partitionSums.collect().toString());
+ assertEquals("[3, 7]", partitionSums.collect().toString());
}
@Test
@@ -989,9 +990,9 @@ public class JavaAPISuite implements Serializable {
new Tuple2<>("aa", 2),
new Tuple2<>("aaa", 3)
), 2);
- Assert.assertEquals(3, rdd1.getNumPartitions());
- Assert.assertEquals(2, rdd2.getNumPartitions());
- Assert.assertEquals(2, rdd3.getNumPartitions());
+ assertEquals(3, rdd1.getNumPartitions());
+ assertEquals(2, rdd2.getNumPartitions());
+ assertEquals(2, rdd3.getNumPartitions());
}
@Test
@@ -1000,18 +1001,18 @@ public class JavaAPISuite implements Serializable {
JavaRDD<Integer> in1 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6, 7, 8), 2);
JavaRDD<Integer> repartitioned1 = in1.repartition(4);
List<List<Integer>> result1 = repartitioned1.glom().collect();
- Assert.assertEquals(4, result1.size());
+ assertEquals(4, result1.size());
for (List<Integer> l : result1) {
- Assert.assertFalse(l.isEmpty());
+ assertFalse(l.isEmpty());
}
// Growing number of partitions
JavaRDD<Integer> in2 = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5, 6, 7, 8), 4);
JavaRDD<Integer> repartitioned2 = in2.repartition(2);
List<List<Integer>> result2 = repartitioned2.glom().collect();
- Assert.assertEquals(2, result2.size());
+ assertEquals(2, result2.size());
for (List<Integer> l: result2) {
- Assert.assertFalse(l.isEmpty());
+ assertFalse(l.isEmpty());
}
}
@@ -1020,7 +1021,7 @@ public class JavaAPISuite implements Serializable {
public void persist() {
JavaDoubleRDD doubleRDD = sc.parallelizeDoubles(Arrays.asList(1.0, 1.0, 2.0, 3.0, 5.0, 8.0));
doubleRDD = doubleRDD.persist(StorageLevel.DISK_ONLY());
- Assert.assertEquals(20, doubleRDD.sum(), 0.1);
+ assertEquals(20, doubleRDD.sum(), 0.1);
List<Tuple2<Integer, String>> pairs = Arrays.asList(
new Tuple2<>(1, "a"),
@@ -1029,24 +1030,24 @@ public class JavaAPISuite implements Serializable {
);
JavaPairRDD<Integer, String> pairRDD = sc.parallelizePairs(pairs);
pairRDD = pairRDD.persist(StorageLevel.DISK_ONLY());
- Assert.assertEquals("a", pairRDD.first()._2());
+ assertEquals("a", pairRDD.first()._2());
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5));
rdd = rdd.persist(StorageLevel.DISK_ONLY());
- Assert.assertEquals(1, rdd.first().intValue());
+ assertEquals(1, rdd.first().intValue());
}
@Test
public void iterator() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5), 2);
TaskContext context = TaskContext$.MODULE$.empty();
- Assert.assertEquals(1, rdd.iterator(rdd.partitions().get(0), context).next().intValue());
+ assertEquals(1, rdd.iterator(rdd.partitions().get(0), context).next().intValue());
}
@Test
public void glom() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4), 2);
- Assert.assertEquals("[1, 2]", rdd.glom().first().toString());
+ assertEquals("[1, 2]", rdd.glom().first().toString());
}
// File input / output tests are largely adapted from FileSuite:
@@ -1059,11 +1060,11 @@ public class JavaAPISuite implements Serializable {
// Read the plain text file and check it's OK
File outputFile = new File(outputDir, "part-00000");
String content = Files.toString(outputFile, StandardCharsets.UTF_8);
- Assert.assertEquals("1\n2\n3\n4\n", content);
+ assertEquals("1\n2\n3\n4\n", content);
// Also try reading it in as a text file RDD
List<String> expected = Arrays.asList("1", "2", "3", "4");
JavaRDD<String> readRDD = sc.textFile(outputDir);
- Assert.assertEquals(expected, readRDD.collect());
+ assertEquals(expected, readRDD.collect());
}
@Test
@@ -1083,7 +1084,7 @@ public class JavaAPISuite implements Serializable {
List<Tuple2<String, String>> result = readRDD.collect();
for (Tuple2<String, String> res : result) {
- Assert.assertEquals(res._2(), container.get(new URI(res._1()).getPath()));
+ assertEquals(res._2(), container.get(new URI(res._1()).getPath()));
}
}
@@ -1096,7 +1097,7 @@ public class JavaAPISuite implements Serializable {
// Try reading it in as a text file RDD
List<String> expected = Arrays.asList("1", "2", "3", "4");
JavaRDD<String> readRDD = sc.textFile(outputDir);
- Assert.assertEquals(expected, readRDD.collect());
+ assertEquals(expected, readRDD.collect());
}
@SuppressWarnings("unchecked")
@@ -1125,7 +1126,7 @@ public class JavaAPISuite implements Serializable {
return new Tuple2<>(pair._1().get(), pair._2().toString());
}
});
- Assert.assertEquals(pairs, readRDD.collect());
+ assertEquals(pairs, readRDD.collect());
}
@Test
@@ -1145,7 +1146,7 @@ public class JavaAPISuite implements Serializable {
JavaPairRDD<String, PortableDataStream> readRDD = sc.binaryFiles(tempDirName, 3);
List<Tuple2<String, PortableDataStream>> result = readRDD.collect();
for (Tuple2<String, PortableDataStream> res : result) {
- Assert.assertArrayEquals(content1, res._2().toArray());
+ assertArrayEquals(content1, res._2().toArray());
}
}
@@ -1174,7 +1175,7 @@ public class JavaAPISuite implements Serializable {
List<Tuple2<String, PortableDataStream>> result = readRDD.collect();
for (Tuple2<String, PortableDataStream> res : result) {
- Assert.assertArrayEquals(content1, res._2().toArray());
+ assertArrayEquals(content1, res._2().toArray());
}
}
@@ -1197,10 +1198,10 @@ public class JavaAPISuite implements Serializable {
channel1.close();
JavaRDD<byte[]> readRDD = sc.binaryRecords(tempDirName, content1.length);
- Assert.assertEquals(numOfCopies,readRDD.count());
+ assertEquals(numOfCopies,readRDD.count());
List<byte[]> result = readRDD.collect();
for (byte[] res : result) {
- Assert.assertArrayEquals(content1, res);
+ assertArrayEquals(content1, res);
}
}
@@ -1224,8 +1225,9 @@ public class JavaAPISuite implements Serializable {
outputDir, IntWritable.class, Text.class,
org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat.class);
- JavaPairRDD<IntWritable, Text> output = sc.sequenceFile(outputDir, IntWritable.class, Text.class);
- Assert.assertEquals(pairs.toString(), output.map(new Function<Tuple2<IntWritable, Text>, String>() {
+ JavaPairRDD<IntWritable, Text> output =
+ sc.sequenceFile(outputDir, IntWritable.class, Text.class);
+ assertEquals(pairs.toString(), output.map(new Function<Tuple2<IntWritable, Text>, String>() {
@Override
public String call(Tuple2<IntWritable, Text> x) {
return x.toString();
@@ -1254,7 +1256,7 @@ public class JavaAPISuite implements Serializable {
JavaPairRDD<IntWritable, Text> output = sc.newAPIHadoopFile(outputDir,
org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat.class,
IntWritable.class, Text.class, Job.getInstance().getConfiguration());
- Assert.assertEquals(pairs.toString(), output.map(new Function<Tuple2<IntWritable, Text>, String>() {
+ assertEquals(pairs.toString(), output.map(new Function<Tuple2<IntWritable, Text>, String>() {
@Override
public String call(Tuple2<IntWritable, Text> x) {
return x.toString();
@@ -1270,7 +1272,7 @@ public class JavaAPISuite implements Serializable {
// Try reading the output back as an object file
List<Integer> expected = Arrays.asList(1, 2, 3, 4);
JavaRDD<Integer> readRDD = sc.objectFile(outputDir);
- Assert.assertEquals(expected, readRDD.collect());
+ assertEquals(expected, readRDD.collect());
}
@SuppressWarnings("unchecked")
@@ -1286,7 +1288,7 @@ public class JavaAPISuite implements Serializable {
rdd.saveAsObjectFile(outputDir);
// Try reading the output back as an object file
JavaRDD<Tuple2<Integer, String>> readRDD = sc.objectFile(outputDir);
- Assert.assertEquals(pairs, readRDD.collect());
+ assertEquals(pairs, readRDD.collect());
}
@SuppressWarnings("unchecked")
@@ -1309,7 +1311,7 @@ public class JavaAPISuite implements Serializable {
JavaPairRDD<IntWritable, Text> output = sc.hadoopFile(outputDir,
SequenceFileInputFormat.class, IntWritable.class, Text.class);
- Assert.assertEquals(pairs.toString(), output.map(new Function<Tuple2<IntWritable, Text>, String>() {
+ assertEquals(pairs.toString(), output.map(new Function<Tuple2<IntWritable, Text>, String>() {
@Override
public String call(Tuple2<IntWritable, Text> x) {
return x.toString();
@@ -1339,7 +1341,7 @@ public class JavaAPISuite implements Serializable {
JavaPairRDD<IntWritable, Text> output = sc.hadoopFile(outputDir,
SequenceFileInputFormat.class, IntWritable.class, Text.class);
- Assert.assertEquals(pairs.toString(), output.map(new Function<Tuple2<IntWritable, Text>, String>() {
+ assertEquals(pairs.toString(), output.map(new Function<Tuple2<IntWritable, Text>, String>() {
@Override
public String call(Tuple2<IntWritable, Text> x) {
return x.toString();
@@ -1373,7 +1375,7 @@ public class JavaAPISuite implements Serializable {
};
JavaRDD<Integer> sizes = rdd1.zipPartitions(rdd2, sizesFn);
- Assert.assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
+ assertEquals("[3, 2, 3, 2]", sizes.collect().toString());
}
@Test
@@ -1387,7 +1389,7 @@ public class JavaAPISuite implements Serializable {
intAccum.add(x);
}
});
- Assert.assertEquals((Integer) 25, intAccum.value());
+ assertEquals((Integer) 25, intAccum.value());
final Accumulator<Double> doubleAccum = sc.doubleAccumulator(10.0);
rdd.foreach(new VoidFunction<Integer>() {
@@ -1396,7 +1398,7 @@ public class JavaAPISuite implements Serializable {
doubleAccum.add((double) x);
}
});
- Assert.assertEquals((Double) 25.0, doubleAccum.value());
+ assertEquals((Double) 25.0, doubleAccum.value());
// Try a custom accumulator type
AccumulatorParam<Float> floatAccumulatorParam = new AccumulatorParam<Float>() {
@@ -1423,11 +1425,11 @@ public class JavaAPISuite implements Serializable {
floatAccum.add((float) x);
}
});
- Assert.assertEquals((Float) 25.0f, floatAccum.value());
+ assertEquals((Float) 25.0f, floatAccum.value());
// Test the setValue method
floatAccum.setValue(5.0f);
- Assert.assertEquals((Float) 5.0f, floatAccum.value());
+ assertEquals((Float) 5.0f, floatAccum.value());
}
@Test
@@ -1439,33 +1441,33 @@ public class JavaAPISuite implements Serializable {
return t.toString();
}
}).collect();
- Assert.assertEquals(new Tuple2<>("1", 1), s.get(0));
- Assert.assertEquals(new Tuple2<>("2", 2), s.get(1));
+ assertEquals(new Tuple2<>("1", 1), s.get(0));
+ assertEquals(new Tuple2<>("2", 2), s.get(1));
}
@Test
public void checkpointAndComputation() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5));
sc.setCheckpointDir(tempDir.getAbsolutePath());
- Assert.assertFalse(rdd.isCheckpointed());
+ assertFalse(rdd.isCheckpointed());
rdd.checkpoint();
rdd.count(); // Forces the DAG to cause a checkpoint
- Assert.assertTrue(rdd.isCheckpointed());
- Assert.assertEquals(Arrays.asList(1, 2, 3, 4, 5), rdd.collect());
+ assertTrue(rdd.isCheckpointed());
+ assertEquals(Arrays.asList(1, 2, 3, 4, 5), rdd.collect());
}
@Test
public void checkpointAndRestore() {
JavaRDD<Integer> rdd = sc.parallelize(Arrays.asList(1, 2, 3, 4, 5));
sc.setCheckpointDir(tempDir.getAbsolutePath());
- Assert.assertFalse(rdd.isCheckpointed());
+ assertFalse(rdd.isCheckpointed());
rdd.checkpoint();
rdd.count(); // Forces the DAG to cause a checkpoint
- Assert.assertTrue(rdd.isCheckpointed());
+ assertTrue(rdd.isCheckpointed());
- Assert.assertTrue(rdd.getCheckpointFile().isPresent());
+ assertTrue(rdd.getCheckpointFile().isPresent());
JavaRDD<Integer> recovered = sc.checkpointFile(rdd.getCheckpointFile().get());
- Assert.assertEquals(Arrays.asList(1, 2, 3, 4, 5), recovered.collect());
+ assertEquals(Arrays.asList(1, 2, 3, 4, 5), recovered.collect());
}
@Test
@@ -1484,7 +1486,8 @@ public class JavaAPISuite implements Serializable {
}
};
- Function2<Integer, Integer, Integer> mergeValueFunction = new Function2<Integer, Integer, Integer>() {
+ Function2<Integer, Integer, Integer> mergeValueFunction =
+ new Function2<Integer, Integer, Integer>() {
@Override
public Integer call(Integer v1, Integer v2) {
return v1 + v2;
@@ -1495,7 +1498,7 @@ public class JavaAPISuite implements Serializable {
.combineByKey(createCombinerFunction, mergeValueFunction, mergeValueFunction);
Map<Integer, Integer> results = combinedRDD.collectAsMap();
ImmutableMap<Integer, Integer> expected = ImmutableMap.of(0, 9, 1, 5, 2, 7);
- Assert.assertEquals(expected, results);
+ assertEquals(expected, results);
Partitioner defaultPartitioner = Partitioner.defaultPartitioner(
combinedRDD.rdd(),
@@ -1510,7 +1513,7 @@ public class JavaAPISuite implements Serializable {
false,
new KryoSerializer(new SparkConf()));
results = combinedRDD.collectAsMap();
- Assert.assertEquals(expected, results);
+ assertEquals(expected, results);
}
@SuppressWarnings("unchecked")
@@ -1531,7 +1534,7 @@ public class JavaAPISuite implements Serializable {
return new Tuple2<>(in._2(), in._1());
}
});
- Assert.assertEquals(Arrays.asList(
+ assertEquals(Arrays.asList(
new Tuple2<>(1, 1),
new Tuple2<>(0, 2),
new Tuple2<>(1, 3),
@@ -1553,21 +1556,19 @@ public class JavaAPISuite implements Serializable {
});
List<Integer>[] parts = rdd1.collectPartitions(new int[] {0});
- Assert.assertEquals(Arrays.asList(1, 2), parts[0]);
+ assertEquals(Arrays.asList(1, 2), parts[0]);
parts = rdd1.collectPartitions(new int[] {1, 2});
- Assert.assertEquals(Arrays.asList(3, 4), parts[0]);
- Assert.assertEquals(Arrays.asList(5, 6, 7), parts[1]);
+ assertEquals(Arrays.asList(3, 4), parts[0]);
+ assertEquals(Arrays.asList(5, 6, 7), parts[1]);
- Assert.assertEquals(Arrays.asList(new Tuple2<>(1, 1),
+ assertEquals(Arrays.asList(new Tuple2<>(1, 1),
new Tuple2<>(2, 0)),
rdd2.collectPartitions(new int[] {0})[0]);
List<Tuple2<Integer,Integer>>[] parts2 = rdd2.collectPartitions(new int[] {1, 2});
- Assert.assertEquals(Arrays.asList(new Tuple2<>(3, 1),
- new Tuple2<>(4, 0)),
- parts2[0]);
- Assert.assertEquals(Arrays.asList(new Tuple2<>(5, 1),
+ assertEquals(Arrays.asList(new Tuple2<>(3, 1), new Tuple2<>(4, 0)), parts2[0]);
+ assertEquals(Arrays.asList(new Tuple2<>(5, 1),
new Tuple2<>(6, 0),
new Tuple2<>(7, 1)),
parts2[1]);
@@ -1581,7 +1582,7 @@ public class JavaAPISuite implements Serializable {
arrayData.add(i % size);
}
JavaRDD<Integer> simpleRdd = sc.parallelize(arrayData, 10);
- Assert.assertTrue(Math.abs((simpleRdd.countApproxDistinct(0.05) - size) / (size * 1.0)) <= 0.1);
+ assertTrue(Math.abs((simpleRdd.countApproxDistinct(0.05) - size) / (size * 1.0)) <= 0.1);
}
@Test
@@ -1599,7 +1600,7 @@ public class JavaAPISuite implements Serializable {
double count = resItem._1();
long resCount = resItem._2();
double error = Math.abs((resCount - count) / count);
- Assert.assertTrue(error < 0.1);
+ assertTrue(error < 0.1);
}
}
@@ -1629,7 +1630,7 @@ public class JavaAPISuite implements Serializable {
new ObjectOutputStream(bytes).writeObject(map);
Map<String,Integer> deserializedMap = (Map<String,Integer>)
new ObjectInputStream(new ByteArrayInputStream(bytes.toByteArray())).readObject();
- Assert.assertEquals(1, deserializedMap.get("foo").intValue());
+ assertEquals(1, deserializedMap.get("foo").intValue());
}
@Test
@@ -1648,14 +1649,14 @@ public class JavaAPISuite implements Serializable {
fractions.put(1, 1.0);
JavaPairRDD<Integer, Integer> wr = rdd2.sampleByKey(true, fractions, 1L);
Map<Integer, Long> wrCounts = wr.countByKey();
- Assert.assertEquals(2, wrCounts.size());
- Assert.assertTrue(wrCounts.get(0) > 0);
- Assert.assertTrue(wrCounts.get(1) > 0);
+ assertEquals(2, wrCounts.size());
+ assertTrue(wrCounts.get(0) > 0);
+ assertTrue(wrCounts.get(1) > 0);
JavaPairRDD<Integer, Integer> wor = rdd2.sampleByKey(false, fractions, 1L);
Map<Integer, Long> worCounts = wor.countByKey();
- Assert.assertEquals(2, worCounts.size());
- Assert.assertTrue(worCounts.get(0) > 0);
- Assert.assertTrue(worCounts.get(1) > 0);
+ assertEquals(2, worCounts.size());
+ assertTrue(worCounts.get(0) > 0);
+ assertTrue(worCounts.get(1) > 0);
}
@Test
@@ -1674,14 +1675,14 @@ public class JavaAPISuite implements Serializable {
fractions.put(1, 1.0);
JavaPairRDD<Integer, Integer> wrExact = rdd2.sampleByKeyExact(true, fractions, 1L);
Map<Integer, Long> wrExactCounts = wrExact.countByKey();
- Assert.assertEquals(2, wrExactCounts.size());
- Assert.assertTrue(wrExactCounts.get(0) == 2);
- Assert.assertTrue(wrExactCounts.get(1) == 4);
+ assertEquals(2, wrExactCounts.size());
+ assertTrue(wrExactCounts.get(0) == 2);
+ assertTrue(wrExactCounts.get(1) == 4);
JavaPairRDD<Integer, Integer> worExact = rdd2.sampleByKeyExact(false, fractions, 1L);
Map<Integer, Long> worExactCounts = worExact.countByKey();
- Assert.assertEquals(2, worExactCounts.size());
- Assert.assertTrue(worExactCounts.get(0) == 2);
- Assert.assertTrue(worExactCounts.get(1) == 4);
+ assertEquals(2, worExactCounts.size());
+ assertTrue(worExactCounts.get(0) == 2);
+ assertTrue(worExactCounts.get(1) == 4);
}
private static class SomeCustomClass implements Serializable {
@@ -1697,8 +1698,9 @@ public class JavaAPISuite implements Serializable {
data.add(new SomeCustomClass());
}
JavaRDD<SomeCustomClass> rdd = sc.parallelize(data);
- SomeCustomClass[] collected = (SomeCustomClass[]) rdd.rdd().retag(SomeCustomClass.class).collect();
- Assert.assertEquals(data.size(), collected.length);
+ SomeCustomClass[] collected =
+ (SomeCustomClass[]) rdd.rdd().retag(SomeCustomClass.class).collect();
+ assertEquals(data.size(), collected.length);
}
private static final class BuggyMapFunction<T> implements Function<T, T> {
@@ -1715,10 +1717,10 @@ public class JavaAPISuite implements Serializable {
JavaRDD<Integer> rdd = sc.parallelize(data, 1);
JavaFutureAction<List<Integer>> future = rdd.collectAsync();
List<Integer> result = future.get();
- Assert.assertEquals(data, result);
- Assert.assertFalse(future.isCancelled());
- Assert.assertTrue(future.isDone());
- Assert.assertEquals(1, future.jobIds().size());
+ assertEquals(data, result);
+ assertFalse(future.isCancelled());
+ assertTrue(future.isDone());
+ assertEquals(1, future.jobIds().size());
}
@Test
@@ -1727,11 +1729,11 @@ public class JavaAPISuite implements Serializable {
JavaRDD<Integer> rdd = sc.parallelize(data, 1);
JavaFutureAction<List<Integer>> future = rdd.takeAsync(1);
List<Integer> result = future.get();
- Assert.assertEquals(1, result.size());
- Assert.assertEquals((Integer) 1, result.get(0));
- Assert.assertFalse(future.isCancelled());
- Assert.assertTrue(future.isDone());
- Assert.assertEquals(1, future.jobIds().size());
+ assertEquals(1, result.size());
+ assertEquals((Integer) 1, result.get(0));
+ assertFalse(future.isCancelled());
+ assertTrue(future.isDone());
+ assertEquals(1, future.jobIds().size());
}
@Test
@@ -1747,9 +1749,9 @@ public class JavaAPISuite implements Serializable {
}
);
future.get();
- Assert.assertFalse(future.isCancelled());
- Assert.assertTrue(future.isDone());
- Assert.assertEquals(1, future.jobIds().size());
+ assertFalse(future.isCancelled());
+ assertTrue(future.isDone());
+ assertEquals(1, future.jobIds().size());
}
@Test
@@ -1758,10 +1760,10 @@ public class JavaAPISuite implements Serializable {
JavaRDD<Integer> rdd = sc.parallelize(data, 1);
JavaFutureAction<Long> future = rdd.countAsync();
long count = future.get();
- Assert.assertEquals(data.size(), count);
- Assert.assertFalse(future.isCancelled());
- Assert.assertTrue(future.isDone());
- Assert.assertEquals(1, future.jobIds().size());
+ assertEquals(data.size(), count);
+ assertFalse(future.isCancelled());
+ assertTrue(future.isDone());
+ assertEquals(1, future.jobIds().size());
}
@Test
@@ -1775,11 +1777,11 @@ public class JavaAPISuite implements Serializable {
}
});
future.cancel(true);
- Assert.assertTrue(future.isCancelled());
- Assert.assertTrue(future.isDone());
+ assertTrue(future.isCancelled());
+ assertTrue(future.isDone());
try {
future.get(2000, TimeUnit.MILLISECONDS);
- Assert.fail("Expected future.get() for cancelled job to throw CancellationException");
+ fail("Expected future.get() for cancelled job to throw CancellationException");
} catch (CancellationException ignored) {
// pass
}
@@ -1792,11 +1794,11 @@ public class JavaAPISuite implements Serializable {
JavaFutureAction<Long> future = rdd.map(new BuggyMapFunction<Integer>()).countAsync();
try {
future.get(2, TimeUnit.SECONDS);
- Assert.fail("Expected future.get() for failed job to throw ExcecutionException");
+ fail("Expected future.get() for failed job to throw ExcecutionException");
} catch (ExecutionException ee) {
- Assert.assertTrue(Throwables.getStackTraceAsString(ee).contains("Custom exception!"));
+ assertTrue(Throwables.getStackTraceAsString(ee).contains("Custom exception!"));
}
- Assert.assertTrue(future.isDone());
+ assertTrue(future.isDone());
}
static class Class1 {}
@@ -1806,7 +1808,7 @@ public class JavaAPISuite implements Serializable {
public void testRegisterKryoClasses() {
SparkConf conf = new SparkConf();
conf.registerKryoClasses(new Class<?>[]{ Class1.class, Class2.class });
- Assert.assertEquals(
+ assertEquals(
Class1.class.getName() + "," + Class2.class.getName(),
conf.get("spark.kryo.classesToRegister"));
}
@@ -1814,13 +1816,13 @@ public class JavaAPISuite implements Serializable {
@Test
public void testGetPersistentRDDs() {
java.util.Map<Integer, JavaRDD<?>> cachedRddsMap = sc.getPersistentRDDs();
- Assert.assertTrue(cachedRddsMap.isEmpty());
+ assertTrue(cachedRddsMap.isEmpty());
JavaRDD<String> rdd1 = sc.parallelize(Arrays.asList("a", "b")).setName("RDD1").cache();
JavaRDD<String> rdd2 = sc.parallelize(Arrays.asList("c", "d")).setName("RDD2").cache();
cachedRddsMap = sc.getPersistentRDDs();
- Assert.assertEquals(2, cachedRddsMap.size());
- Assert.assertEquals("RDD1", cachedRddsMap.get(0).name());
- Assert.assertEquals("RDD2", cachedRddsMap.get(1).name());
+ assertEquals(2, cachedRddsMap.size());
+ assertEquals("RDD1", cachedRddsMap.get(0).name());
+ assertEquals("RDD2", cachedRddsMap.get(1).name());
}
}
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/core/src/test/java/org/apache/spark/unsafe/map/AbstractBytesToBytesMapSuite.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/spark/unsafe/map/AbstractBytesToBytesMapSuite.java b/core/src/test/java/org/apache/spark/unsafe/map/AbstractBytesToBytesMapSuite.java
index 9aab226..6667179 100644
--- a/core/src/test/java/org/apache/spark/unsafe/map/AbstractBytesToBytesMapSuite.java
+++ b/core/src/test/java/org/apache/spark/unsafe/map/AbstractBytesToBytesMapSuite.java
@@ -92,9 +92,11 @@ public abstract class AbstractBytesToBytesMapSuite {
spillFilesCreated.clear();
MockitoAnnotations.initMocks(this);
when(blockManager.diskBlockManager()).thenReturn(diskBlockManager);
- when(diskBlockManager.createTempLocalBlock()).thenAnswer(new Answer<Tuple2<TempLocalBlockId, File>>() {
+ when(diskBlockManager.createTempLocalBlock()).thenAnswer(
+ new Answer<Tuple2<TempLocalBlockId, File>>() {
@Override
- public Tuple2<TempLocalBlockId, File> answer(InvocationOnMock invocationOnMock) throws Throwable {
+ public Tuple2<TempLocalBlockId, File> answer(InvocationOnMock invocationOnMock)
+ throws Throwable {
TempLocalBlockId blockId = new TempLocalBlockId(UUID.randomUUID());
File file = File.createTempFile("spillFile", ".spill", tempDir);
spillFilesCreated.add(file);
@@ -544,7 +546,8 @@ public abstract class AbstractBytesToBytesMapSuite {
@Test
public void spillInIterator() throws IOException {
- BytesToBytesMap map = new BytesToBytesMap(taskMemoryManager, blockManager, 1, 0.75, 1024, false);
+ BytesToBytesMap map =
+ new BytesToBytesMap(taskMemoryManager, blockManager, 1, 0.75, 1024, false);
try {
int i;
for (i = 0; i < 1024; i++) {
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeExternalSorterSuite.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeExternalSorterSuite.java b/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeExternalSorterSuite.java
index a79ed58..db50e55 100644
--- a/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeExternalSorterSuite.java
+++ b/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeExternalSorterSuite.java
@@ -103,9 +103,11 @@ public class UnsafeExternalSorterSuite {
taskContext = mock(TaskContext.class);
when(taskContext.taskMetrics()).thenReturn(new TaskMetrics());
when(blockManager.diskBlockManager()).thenReturn(diskBlockManager);
- when(diskBlockManager.createTempLocalBlock()).thenAnswer(new Answer<Tuple2<TempLocalBlockId, File>>() {
+ when(diskBlockManager.createTempLocalBlock()).thenAnswer(
+ new Answer<Tuple2<TempLocalBlockId, File>>() {
@Override
- public Tuple2<TempLocalBlockId, File> answer(InvocationOnMock invocationOnMock) throws Throwable {
+ public Tuple2<TempLocalBlockId, File> answer(InvocationOnMock invocationOnMock)
+ throws Throwable {
TempLocalBlockId blockId = new TempLocalBlockId(UUID.randomUUID());
File file = File.createTempFile("spillFile", ".spill", tempDir);
spillFilesCreated.add(file);
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeInMemorySorterSuite.java
----------------------------------------------------------------------
diff --git a/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeInMemorySorterSuite.java b/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeInMemorySorterSuite.java
index 4833194..f90214f 100644
--- a/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeInMemorySorterSuite.java
+++ b/core/src/test/java/org/apache/spark/util/collection/unsafe/sort/UnsafeInMemorySorterSuite.java
@@ -108,8 +108,8 @@ public class UnsafeInMemorySorterSuite {
return (int) prefix1 - (int) prefix2;
}
};
- UnsafeInMemorySorter sorter = new UnsafeInMemorySorter(consumer, memoryManager, recordComparator,
- prefixComparator, dataToSort.length);
+ UnsafeInMemorySorter sorter = new UnsafeInMemorySorter(consumer, memoryManager,
+ recordComparator, prefixComparator, dataToSort.length);
// Given a page of records, insert those records into the sorter one-by-one:
position = dataPage.getBaseOffset();
for (int i = 0; i < dataToSort.length; i++) {
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/dev/checkstyle-suppressions.xml
----------------------------------------------------------------------
diff --git a/dev/checkstyle-suppressions.xml b/dev/checkstyle-suppressions.xml
index 9242be3..a1a88ac 100644
--- a/dev/checkstyle-suppressions.xml
+++ b/dev/checkstyle-suppressions.xml
@@ -28,6 +28,12 @@
-->
<suppressions>
-<suppress checks=".*"
- files="core/src/main/java/org/apache/spark/util/collection/TimSort.java"/>
+ <suppress checks=".*"
+ files="core/src/main/java/org/apache/spark/util/collection/TimSort.java"/>
+ <suppress checks=".*"
+ files="sql/core/src/main/java/org/apache/spark/sql/api.java/*"/>
+ <suppress checks="LineLength"
+ files="src/test/java/org/apache/spark/sql/hive/test/Complex.java"/>
+ <suppress checks="LineLength"
+ files="src/main/java/org/apache/spark/examples/JavaLogQuery.java"/>
</suppressions>
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/dev/checkstyle.xml
----------------------------------------------------------------------
diff --git a/dev/checkstyle.xml b/dev/checkstyle.xml
index 2261cc9..b66dca9 100644
--- a/dev/checkstyle.xml
+++ b/dev/checkstyle.xml
@@ -76,13 +76,10 @@
<property name="allowByTailComment" value="true"/>
<property name="allowNonPrintableEscapes" value="true"/>
</module>
- <!-- TODO: 11/09/15 disabled - the lengths are currently > 100 in many places -->
- <!--
<module name="LineLength">
<property name="max" value="100"/>
<property name="ignorePattern" value="^package.*|^import.*|a href|href|http://|https://|ftp://"/>
</module>
- -->
<module name="NoLineWrap"/>
<module name="EmptyBlock">
<property name="option" value="TEXT"/>
@@ -167,5 +164,7 @@
</module>
<module name="CommentsIndentation"/>
<module name="UnusedImports"/>
+ <module name="RedundantImport"/>
+ <module name="RedundantModifier"/>
</module>
</module>
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/JavaPageRank.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/JavaPageRank.java b/examples/src/main/java/org/apache/spark/examples/JavaPageRank.java
index c3ef93c..229d123 100644
--- a/examples/src/main/java/org/apache/spark/examples/JavaPageRank.java
+++ b/examples/src/main/java/org/apache/spark/examples/JavaPageRank.java
@@ -84,13 +84,14 @@ public final class JavaPageRank {
JavaRDD<String> lines = ctx.textFile(args[0], 1);
// Loads all URLs from input file and initialize their neighbors.
- JavaPairRDD<String, Iterable<String>> links = lines.mapToPair(new PairFunction<String, String, String>() {
- @Override
- public Tuple2<String, String> call(String s) {
- String[] parts = SPACES.split(s);
- return new Tuple2<>(parts[0], parts[1]);
- }
- }).distinct().groupByKey().cache();
+ JavaPairRDD<String, Iterable<String>> links = lines.mapToPair(
+ new PairFunction<String, String, String>() {
+ @Override
+ public Tuple2<String, String> call(String s) {
+ String[] parts = SPACES.split(s);
+ return new Tuple2<>(parts[0], parts[1]);
+ }
+ }).distinct().groupByKey().cache();
// Loads all URLs with other URL(s) link to from input file and initialize ranks of them to one.
JavaPairRDD<String, Double> ranks = links.mapValues(new Function<Iterable<String>, Double>() {
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/JavaWordCount.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/JavaWordCount.java b/examples/src/main/java/org/apache/spark/examples/JavaWordCount.java
index 84dbea5..3ff5412 100644
--- a/examples/src/main/java/org/apache/spark/examples/JavaWordCount.java
+++ b/examples/src/main/java/org/apache/spark/examples/JavaWordCount.java
@@ -52,19 +52,21 @@ public final class JavaWordCount {
}
});
- JavaPairRDD<String, Integer> ones = words.mapToPair(new PairFunction<String, String, Integer>() {
- @Override
- public Tuple2<String, Integer> call(String s) {
- return new Tuple2<>(s, 1);
- }
- });
+ JavaPairRDD<String, Integer> ones = words.mapToPair(
+ new PairFunction<String, String, Integer>() {
+ @Override
+ public Tuple2<String, Integer> call(String s) {
+ return new Tuple2<>(s, 1);
+ }
+ });
- JavaPairRDD<String, Integer> counts = ones.reduceByKey(new Function2<Integer, Integer, Integer>() {
- @Override
- public Integer call(Integer i1, Integer i2) {
- return i1 + i2;
- }
- });
+ JavaPairRDD<String, Integer> counts = ones.reduceByKey(
+ new Function2<Integer, Integer, Integer>() {
+ @Override
+ public Integer call(Integer i1, Integer i2) {
+ return i1 + i2;
+ }
+ });
List<Tuple2<String, Integer>> output = counts.collect();
for (Tuple2<?,?> tuple : output) {
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/ml/JavaDecisionTreeClassificationExample.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaDecisionTreeClassificationExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaDecisionTreeClassificationExample.java
index 5bd61fe..8214952 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaDecisionTreeClassificationExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaDecisionTreeClassificationExample.java
@@ -39,7 +39,10 @@ public class JavaDecisionTreeClassificationExample {
// $example on$
// Load the data stored in LIBSVM format as a DataFrame.
- Dataset<Row> data = sqlContext.read().format("libsvm").load("data/mllib/sample_libsvm_data.txt");
+ Dataset<Row> data = sqlContext
+ .read()
+ .format("libsvm")
+ .load("data/mllib/sample_libsvm_data.txt");
// Index labels, adding metadata to the label column.
// Fit on whole dataset to include all labels in index.
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
index 8a10dd4..fbd8817 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaDeveloperApiExample.java
@@ -107,11 +107,11 @@ public class JavaDeveloperApiExample {
class MyJavaLogisticRegression
extends Classifier<Vector, MyJavaLogisticRegression, MyJavaLogisticRegressionModel> {
- public MyJavaLogisticRegression() {
+ MyJavaLogisticRegression() {
init();
}
- public MyJavaLogisticRegression(String uid) {
+ MyJavaLogisticRegression(String uid) {
this.uid_ = uid;
init();
}
@@ -177,7 +177,7 @@ class MyJavaLogisticRegressionModel
private Vector coefficients_;
public Vector coefficients() { return coefficients_; }
- public MyJavaLogisticRegressionModel(String uid, Vector coefficients) {
+ MyJavaLogisticRegressionModel(String uid, Vector coefficients) {
this.uid_ = uid;
this.coefficients_ = coefficients;
}
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java b/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
index c2cb955..553070d 100644
--- a/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/ml/JavaGradientBoostedTreeClassifierExample.java
@@ -40,7 +40,8 @@ public class JavaGradientBoostedTreeClassifierExample {
// $example on$
// Load and parse the data file, converting it to a DataFrame.
- Dataset<Row> data = sqlContext.read().format("libsvm").load("data/mllib/sample_libsvm_data.txt");
+ Dataset<Row> data = sqlContext.read().format("libsvm")
+ .load("data/mllib/sample_libsvm_data.txt");
// Index labels, adding metadata to the label column.
// Fit on whole dataset to include all labels in index.
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/mllib/JavaBinaryClassificationMetricsExample.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/mllib/JavaBinaryClassificationMetricsExample.java b/examples/src/main/java/org/apache/spark/examples/mllib/JavaBinaryClassificationMetricsExample.java
index 3d8babb..7561a1f 100644
--- a/examples/src/main/java/org/apache/spark/examples/mllib/JavaBinaryClassificationMetricsExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/mllib/JavaBinaryClassificationMetricsExample.java
@@ -65,7 +65,8 @@ public class JavaBinaryClassificationMetricsExample {
);
// Get evaluation metrics.
- BinaryClassificationMetrics metrics = new BinaryClassificationMetrics(predictionAndLabels.rdd());
+ BinaryClassificationMetrics metrics =
+ new BinaryClassificationMetrics(predictionAndLabels.rdd());
// Precision by threshold
JavaRDD<Tuple2<Object, Object>> precision = metrics.precisionByThreshold().toJavaRDD();
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/mllib/JavaIsotonicRegressionExample.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/mllib/JavaIsotonicRegressionExample.java b/examples/src/main/java/org/apache/spark/examples/mllib/JavaIsotonicRegressionExample.java
index 0e15f75..c6361a3 100644
--- a/examples/src/main/java/org/apache/spark/examples/mllib/JavaIsotonicRegressionExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/mllib/JavaIsotonicRegressionExample.java
@@ -48,7 +48,8 @@ public class JavaIsotonicRegressionExample {
);
// Split data into training (60%) and test (40%) sets.
- JavaRDD<Tuple3<Double, Double, Double>>[] splits = parsedData.randomSplit(new double[]{0.6, 0.4}, 11L);
+ JavaRDD<Tuple3<Double, Double, Double>>[] splits =
+ parsedData.randomSplit(new double[]{0.6, 0.4}, 11L);
JavaRDD<Tuple3<Double, Double, Double>> training = splits[0];
JavaRDD<Tuple3<Double, Double, Double>> test = splits[1];
@@ -80,7 +81,8 @@ public class JavaIsotonicRegressionExample {
// Save and load model
model.save(jsc.sc(), "target/tmp/myIsotonicRegressionModel");
- IsotonicRegressionModel sameModel = IsotonicRegressionModel.load(jsc.sc(), "target/tmp/myIsotonicRegressionModel");
+ IsotonicRegressionModel sameModel =
+ IsotonicRegressionModel.load(jsc.sc(), "target/tmp/myIsotonicRegressionModel");
// $example off$
jsc.stop();
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/mllib/JavaStreamingTestExample.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/mllib/JavaStreamingTestExample.java b/examples/src/main/java/org/apache/spark/examples/mllib/JavaStreamingTestExample.java
index 4c87559..984909c 100644
--- a/examples/src/main/java/org/apache/spark/examples/mllib/JavaStreamingTestExample.java
+++ b/examples/src/main/java/org/apache/spark/examples/mllib/JavaStreamingTestExample.java
@@ -18,7 +18,6 @@
package org.apache.spark.examples.mllib;
-import org.apache.spark.Accumulator;
import org.apache.spark.api.java.function.VoidFunction;
import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.api.java.function.Function;
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/streaming/JavaDirectKafkaWordCount.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/streaming/JavaDirectKafkaWordCount.java b/examples/src/main/java/org/apache/spark/examples/streaming/JavaDirectKafkaWordCount.java
index bfbad91..769b21c 100644
--- a/examples/src/main/java/org/apache/spark/examples/streaming/JavaDirectKafkaWordCount.java
+++ b/examples/src/main/java/org/apache/spark/examples/streaming/JavaDirectKafkaWordCount.java
@@ -40,7 +40,8 @@ import org.apache.spark.streaming.Durations;
* <topics> is a list of one or more kafka topics to consume from
*
* Example:
- * $ bin/run-example streaming.JavaDirectKafkaWordCount broker1-host:port,broker2-host:port topic1,topic2
+ * $ bin/run-example streaming.JavaDirectKafkaWordCount broker1-host:port,broker2-host:port \
+ * topic1,topic2
*/
public final class JavaDirectKafkaWordCount {
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/streaming/JavaQueueStream.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/streaming/JavaQueueStream.java b/examples/src/main/java/org/apache/spark/examples/streaming/JavaQueueStream.java
index 426eaa5..62413b4 100644
--- a/examples/src/main/java/org/apache/spark/examples/streaming/JavaQueueStream.java
+++ b/examples/src/main/java/org/apache/spark/examples/streaming/JavaQueueStream.java
@@ -30,7 +30,6 @@ import org.apache.spark.SparkConf;
import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.api.java.function.Function2;
import org.apache.spark.api.java.function.PairFunction;
-import org.apache.spark.examples.streaming.StreamingExamples;
import org.apache.spark.streaming.Duration;
import org.apache.spark.streaming.api.java.JavaDStream;
import org.apache.spark.streaming.api.java.JavaPairDStream;
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/streaming/JavaRecoverableNetworkWordCount.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/streaming/JavaRecoverableNetworkWordCount.java b/examples/src/main/java/org/apache/spark/examples/streaming/JavaRecoverableNetworkWordCount.java
index a597ecb..e5fb2bf 100644
--- a/examples/src/main/java/org/apache/spark/examples/streaming/JavaRecoverableNetworkWordCount.java
+++ b/examples/src/main/java/org/apache/spark/examples/streaming/JavaRecoverableNetworkWordCount.java
@@ -155,9 +155,11 @@ public final class JavaRecoverableNetworkWordCount {
@Override
public void call(JavaPairRDD<String, Integer> rdd, Time time) throws IOException {
// Get or register the blacklist Broadcast
- final Broadcast<List<String>> blacklist = JavaWordBlacklist.getInstance(new JavaSparkContext(rdd.context()));
+ final Broadcast<List<String>> blacklist =
+ JavaWordBlacklist.getInstance(new JavaSparkContext(rdd.context()));
// Get or register the droppedWordsCounter Accumulator
- final Accumulator<Integer> droppedWordsCounter = JavaDroppedWordsCounter.getInstance(new JavaSparkContext(rdd.context()));
+ final Accumulator<Integer> droppedWordsCounter =
+ JavaDroppedWordsCounter.getInstance(new JavaSparkContext(rdd.context()));
// Use blacklist to drop words and use droppedWordsCounter to count them
String counts = rdd.filter(new Function<Tuple2<String, Integer>, Boolean>() {
@Override
@@ -210,7 +212,8 @@ public final class JavaRecoverableNetworkWordCount {
}
};
- JavaStreamingContext ssc = JavaStreamingContext.getOrCreate(checkpointDirectory, createContextFunc);
+ JavaStreamingContext ssc =
+ JavaStreamingContext.getOrCreate(checkpointDirectory, createContextFunc);
ssc.start();
ssc.awaitTermination();
}
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/examples/src/main/java/org/apache/spark/examples/streaming/JavaStatefulNetworkWordCount.java
----------------------------------------------------------------------
diff --git a/examples/src/main/java/org/apache/spark/examples/streaming/JavaStatefulNetworkWordCount.java b/examples/src/main/java/org/apache/spark/examples/streaming/JavaStatefulNetworkWordCount.java
index 6beab90..4230dab 100644
--- a/examples/src/main/java/org/apache/spark/examples/streaming/JavaStatefulNetworkWordCount.java
+++ b/examples/src/main/java/org/apache/spark/examples/streaming/JavaStatefulNetworkWordCount.java
@@ -91,7 +91,8 @@ public class JavaStatefulNetworkWordCount {
Function3<String, Optional<Integer>, State<Integer>, Tuple2<String, Integer>> mappingFunc =
new Function3<String, Optional<Integer>, State<Integer>, Tuple2<String, Integer>>() {
@Override
- public Tuple2<String, Integer> call(String word, Optional<Integer> one, State<Integer> state) {
+ public Tuple2<String, Integer> call(String word, Optional<Integer> one,
+ State<Integer> state) {
int sum = one.orElse(0) + (state.exists() ? state.get() : 0);
Tuple2<String, Integer> output = new Tuple2<>(word, sum);
state.update(sum);
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/external/kinesis-asl/src/main/java/org/apache/spark/examples/streaming/JavaKinesisWordCountASL.java
----------------------------------------------------------------------
diff --git a/external/kinesis-asl/src/main/java/org/apache/spark/examples/streaming/JavaKinesisWordCountASL.java b/external/kinesis-asl/src/main/java/org/apache/spark/examples/streaming/JavaKinesisWordCountASL.java
index 5dc825d..0e43e92 100644
--- a/external/kinesis-asl/src/main/java/org/apache/spark/examples/streaming/JavaKinesisWordCountASL.java
+++ b/external/kinesis-asl/src/main/java/org/apache/spark/examples/streaming/JavaKinesisWordCountASL.java
@@ -140,7 +140,8 @@ public final class JavaKinesisWordCountASL { // needs to be public for access fr
for (int i = 0; i < numStreams; i++) {
streamsList.add(
KinesisUtils.createStream(jssc, kinesisAppName, streamName, endpointUrl, regionName,
- InitialPositionInStream.LATEST, kinesisCheckpointInterval, StorageLevel.MEMORY_AND_DISK_2())
+ InitialPositionInStream.LATEST, kinesisCheckpointInterval,
+ StorageLevel.MEMORY_AND_DISK_2())
);
}
@@ -167,7 +168,7 @@ public final class JavaKinesisWordCountASL { // needs to be public for access fr
new PairFunction<String, String, Integer>() {
@Override
public Tuple2<String, Integer> call(String s) {
- return new Tuple2<String, Integer>(s, 1);
+ return new Tuple2<>(s, 1);
}
}
).reduceByKey(
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/launcher/src/main/java/org/apache/spark/launcher/AbstractCommandBuilder.java
----------------------------------------------------------------------
diff --git a/launcher/src/main/java/org/apache/spark/launcher/AbstractCommandBuilder.java b/launcher/src/main/java/org/apache/spark/launcher/AbstractCommandBuilder.java
index f6c7e07..587fda7 100644
--- a/launcher/src/main/java/org/apache/spark/launcher/AbstractCommandBuilder.java
+++ b/launcher/src/main/java/org/apache/spark/launcher/AbstractCommandBuilder.java
@@ -57,7 +57,7 @@ abstract class AbstractCommandBuilder {
// properties files multiple times.
private Map<String, String> effectiveConfig;
- public AbstractCommandBuilder() {
+ AbstractCommandBuilder() {
this.appArgs = new ArrayList<>();
this.childEnv = new HashMap<>();
this.conf = new HashMap<>();
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/launcher/src/main/java/org/apache/spark/launcher/CommandBuilderUtils.java
----------------------------------------------------------------------
diff --git a/launcher/src/main/java/org/apache/spark/launcher/CommandBuilderUtils.java b/launcher/src/main/java/org/apache/spark/launcher/CommandBuilderUtils.java
index 37afafe..39fdf30 100644
--- a/launcher/src/main/java/org/apache/spark/launcher/CommandBuilderUtils.java
+++ b/launcher/src/main/java/org/apache/spark/launcher/CommandBuilderUtils.java
@@ -32,7 +32,7 @@ class CommandBuilderUtils {
static final String ENV_SPARK_HOME = "SPARK_HOME";
/** The set of known JVM vendors. */
- static enum JavaVendor {
+ enum JavaVendor {
Oracle, IBM, OpenJDK, Unknown
};
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/launcher/src/main/java/org/apache/spark/launcher/SparkAppHandle.java
----------------------------------------------------------------------
diff --git a/launcher/src/main/java/org/apache/spark/launcher/SparkAppHandle.java b/launcher/src/main/java/org/apache/spark/launcher/SparkAppHandle.java
index e9caf0b..625d026 100644
--- a/launcher/src/main/java/org/apache/spark/launcher/SparkAppHandle.java
+++ b/launcher/src/main/java/org/apache/spark/launcher/SparkAppHandle.java
@@ -32,7 +32,7 @@ public interface SparkAppHandle {
*
* @since 1.6.0
*/
- public enum State {
+ enum State {
/** The application has not reported back yet. */
UNKNOWN(false),
/** The application has connected to the handle. */
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/mllib/src/test/java/org/apache/spark/ml/classification/JavaLogisticRegressionSuite.java
----------------------------------------------------------------------
diff --git a/mllib/src/test/java/org/apache/spark/ml/classification/JavaLogisticRegressionSuite.java b/mllib/src/test/java/org/apache/spark/ml/classification/JavaLogisticRegressionSuite.java
index 536f0dc..e160a5a 100644
--- a/mllib/src/test/java/org/apache/spark/ml/classification/JavaLogisticRegressionSuite.java
+++ b/mllib/src/test/java/org/apache/spark/ml/classification/JavaLogisticRegressionSuite.java
@@ -18,7 +18,6 @@
package org.apache.spark.ml.classification;
import java.io.Serializable;
-import java.lang.Math;
import java.util.List;
import org.junit.After;
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/mllib/src/test/java/org/apache/spark/ml/classification/JavaOneVsRestSuite.java
----------------------------------------------------------------------
diff --git a/mllib/src/test/java/org/apache/spark/ml/classification/JavaOneVsRestSuite.java b/mllib/src/test/java/org/apache/spark/ml/classification/JavaOneVsRestSuite.java
index d493a7f..00f4476 100644
--- a/mllib/src/test/java/org/apache/spark/ml/classification/JavaOneVsRestSuite.java
+++ b/mllib/src/test/java/org/apache/spark/ml/classification/JavaOneVsRestSuite.java
@@ -48,7 +48,8 @@ public class JavaOneVsRestSuite implements Serializable {
jsql = new SQLContext(jsc);
int nPoints = 3;
- // The following coefficients and xMean/xVariance are computed from iris dataset with lambda=0.2.
+ // The following coefficients and xMean/xVariance are computed from iris dataset with
+ // lambda=0.2.
// As a result, we are drawing samples from probability distribution of an actual model.
double[] coefficients = {
-0.57997, 0.912083, -0.371077, -0.819866, 2.688191,
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/mllib/src/test/java/org/apache/spark/ml/feature/JavaStopWordsRemoverSuite.java
----------------------------------------------------------------------
diff --git a/mllib/src/test/java/org/apache/spark/ml/feature/JavaStopWordsRemoverSuite.java b/mllib/src/test/java/org/apache/spark/ml/feature/JavaStopWordsRemoverSuite.java
index 5812037..bdcbde5 100644
--- a/mllib/src/test/java/org/apache/spark/ml/feature/JavaStopWordsRemoverSuite.java
+++ b/mllib/src/test/java/org/apache/spark/ml/feature/JavaStopWordsRemoverSuite.java
@@ -63,7 +63,8 @@ public class JavaStopWordsRemoverSuite {
RowFactory.create(Arrays.asList("Mary", "had", "a", "little", "lamb"))
);
StructType schema = new StructType(new StructField[] {
- new StructField("raw", DataTypes.createArrayType(DataTypes.StringType), false, Metadata.empty())
+ new StructField("raw", DataTypes.createArrayType(DataTypes.StringType), false,
+ Metadata.empty())
});
Dataset<Row> dataset = jsql.createDataFrame(data, schema);
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/sql/catalyst/src/main/java/org/apache/spark/sql/execution/UnsafeExternalRowSorter.java
----------------------------------------------------------------------
diff --git a/sql/catalyst/src/main/java/org/apache/spark/sql/execution/UnsafeExternalRowSorter.java b/sql/catalyst/src/main/java/org/apache/spark/sql/execution/UnsafeExternalRowSorter.java
index 0ad0f49..d85147e 100644
--- a/sql/catalyst/src/main/java/org/apache/spark/sql/execution/UnsafeExternalRowSorter.java
+++ b/sql/catalyst/src/main/java/org/apache/spark/sql/execution/UnsafeExternalRowSorter.java
@@ -171,7 +171,7 @@ public final class UnsafeExternalRowSorter {
private final UnsafeRow row1;
private final UnsafeRow row2;
- public RowComparator(Ordering<InternalRow> ordering, int numFields) {
+ RowComparator(Ordering<InternalRow> ordering, int numFields) {
this.numFields = numFields;
this.row1 = new UnsafeRow(numFields);
this.row2 = new UnsafeRow(numFields);
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeFixedWidthAggregationMap.java
----------------------------------------------------------------------
diff --git a/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeFixedWidthAggregationMap.java b/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeFixedWidthAggregationMap.java
index 57e8218..acf6c58 100644
--- a/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeFixedWidthAggregationMap.java
+++ b/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeFixedWidthAggregationMap.java
@@ -236,8 +236,8 @@ public final class UnsafeFixedWidthAggregationMap {
/**
* Sorts the map's records in place, spill them to disk, and returns an [[UnsafeKVExternalSorter]]
*
- * Note that the map will be reset for inserting new records, and the returned sorter can NOT be used
- * to insert records.
+ * Note that the map will be reset for inserting new records, and the returned sorter can NOT be
+ * used to insert records.
*/
public UnsafeKVExternalSorter destructAndCreateExternalSorter() throws IOException {
return new UnsafeKVExternalSorter(
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeKVExternalSorter.java
----------------------------------------------------------------------
diff --git a/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeKVExternalSorter.java b/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeKVExternalSorter.java
index 51e10b0..9e08675 100644
--- a/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeKVExternalSorter.java
+++ b/sql/core/src/main/java/org/apache/spark/sql/execution/UnsafeKVExternalSorter.java
@@ -198,7 +198,7 @@ public final class UnsafeKVExternalSorter {
private final UnsafeRow row2;
private final int numKeyFields;
- public KVComparator(BaseOrdering ordering, int numKeyFields) {
+ KVComparator(BaseOrdering ordering, int numKeyFields) {
this.numKeyFields = numKeyFields;
this.row1 = new UnsafeRow(numKeyFields);
this.row2 = new UnsafeRow(numKeyFields);
http://git-wip-us.apache.org/repos/asf/spark/blob/20fd2541/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/ColumnVector.java
----------------------------------------------------------------------
diff --git a/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/ColumnVector.java b/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/ColumnVector.java
index ffcc9c2..04adf1f 100644
--- a/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/ColumnVector.java
+++ b/sql/core/src/main/java/org/apache/spark/sql/execution/vectorized/ColumnVector.java
@@ -94,7 +94,7 @@ public abstract class ColumnVector {
}
@Override
- public final int numElements() { return length; }
+ public int numElements() { return length; }
@Override
public ArrayData copy() {
@@ -175,10 +175,10 @@ public abstract class ColumnVector {
}
@Override
- public final boolean isNullAt(int ordinal) { return data.getIsNull(offset + ordinal); }
+ public boolean isNullAt(int ordinal) { return data.getIsNull(offset + ordinal); }
@Override
- public final boolean getBoolean(int ordinal) {
+ public boolean getBoolean(int ordinal) {
throw new NotImplementedException();
}
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org