You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by matd <ma...@gmail.com> on 2016/07/06 13:23:50 UTC
spark 2.0 bloom filters
A question for Spark developers
I see that Bloom filters have been integrated in Spark 2.0
<https://spark.apache.org/docs/2.0.0-preview/api/scala/index.html#org.apache.spark.util.sketch.BloomFilter>
.
Hadoop already has some Bloom filter implementations, especially a dynamic
one
<https://hadoop.apache.org/docs/r2.7.2/api/org/apache/hadoop/util/bloom/DynamicBloomFilter.html>
, very interesting when the number of keys largely exceed what was imagined.
Is there any rationale (performance, implem...) for this implem in Spark
instead of re-using the one from Hadoop ?
Thanks !
--
View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/spark-2-0-bloom-filters-tp27297.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.
---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscribe@spark.apache.org