You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by Apache Wiki <wi...@apache.org> on 2008/11/24 22:10:56 UTC

[Hadoop Wiki] Update of "PoweredBy" by JamesOmdahl

Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change notification.

The following page has been changed by JamesOmdahl:
http://wiki.apache.org/hadoop/PoweredBy

------------------------------------------------------------------------------
  
   * [http://www.weblab.infosci.cornell.edu/ Cornell University Web Lab]
    * Generating web graphs on 100 nodes (dual 2.4GHz Xeon Processor, 2 GB RAM, 72GB Hard Drive)
+ 
+  * [http://www.enormo.com/ Enormo]
+   * 4 nodes cluster (32 cores, 1TB).
+   * We use Hadoop to filter and index our listings, removing exact duplicates and grouping similar ones.
+   * We plan to use Pig very shortly to produce statistics.
  
   * [http://www.systems.ethz.ch/education/courses/hs08/map-reduce/ ETH Zurich Systems Group]
    * We are using Hadoop in a course that we are currently teaching: "Massively Parallel Data Analysis with MapReduce". The course projects are based on real use-cases from biological data analysis.
@@ -156, +161 @@

    * Using Hadoop on EC2 to process documents from a continuous web crawl and distributed training of support vector machines
    * Using HDFS for large archival data storage
  
-  * [http://www.properazzi.com/ Properazzi]
-   * 4 nodes cluster (32 cores, 1TB).
-   * We use Hadoop to filter and index our listings, removing exact duplicates and grouping similar ones.
-   * We plan to use Pig very shortly to produce statistics.
- 
   * [http://www.quantcast.com/ Quantcast]
    * 3000 cores, 3500TB. 1PB+ processing each day.
    * Hadoop scheduler with fully custom data path / sorter