You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by Apache Wiki <wi...@apache.org> on 2008/11/24 22:10:56 UTC
[Hadoop Wiki] Update of "PoweredBy" by JamesOmdahl
Dear Wiki user,
You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.
The following page has been changed by JamesOmdahl:
http://wiki.apache.org/hadoop/PoweredBy
------------------------------------------------------------------------------
* [http://www.weblab.infosci.cornell.edu/ Cornell University Web Lab]
* Generating web graphs on 100 nodes (dual 2.4GHz Xeon Processor, 2 GB RAM, 72GB Hard Drive)
+
+ * [http://www.enormo.com/ Enormo]
+ * 4 nodes cluster (32 cores, 1TB).
+ * We use Hadoop to filter and index our listings, removing exact duplicates and grouping similar ones.
+ * We plan to use Pig very shortly to produce statistics.
* [http://www.systems.ethz.ch/education/courses/hs08/map-reduce/ ETH Zurich Systems Group]
* We are using Hadoop in a course that we are currently teaching: "Massively Parallel Data Analysis with MapReduce". The course projects are based on real use-cases from biological data analysis.
@@ -156, +161 @@
* Using Hadoop on EC2 to process documents from a continuous web crawl and distributed training of support vector machines
* Using HDFS for large archival data storage
- * [http://www.properazzi.com/ Properazzi]
- * 4 nodes cluster (32 cores, 1TB).
- * We use Hadoop to filter and index our listings, removing exact duplicates and grouping similar ones.
- * We plan to use Pig very shortly to produce statistics.
-
* [http://www.quantcast.com/ Quantcast]
* 3000 cores, 3500TB. 1PB+ processing each day.
* Hadoop scheduler with fully custom data path / sorter