You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@bigtop.apache.org by rv...@apache.org on 2011/11/23 22:26:19 UTC

svn commit: r1205607 [3/3] - in /incubator/bigtop/branches/hadoop-0.23: ./ bigtop-packages/src/common/hive/ bigtop-packages/src/common/sqoop/ bigtop-packages/src/rpm/hive/SPECS/ bigtop-packages/src/rpm/mahout/SPECS/ bigtop-packages/src/rpm/sqoop/SPECS/

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/sqoop/patch
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/sqoop/patch?rev=1205607&r1=1205606&r2=1205607&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/sqoop/patch (original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/common/sqoop/patch Wed Nov 23 21:26:18 2011
@@ -1,8 +1,20 @@
-diff --git ivy.xml ivy.xml
-index d686c76..a882dae 100644
---- ivy.xml
-+++ ivy.xml
-@@ -67,12 +67,19 @@ under the License.
+Index: src/test/com/cloudera/sqoop/TestCompression.java
+===================================================================
+--- src/test/com/cloudera/sqoop/TestCompression.java	(revision 1205535)
++++ src/test/com/cloudera/sqoop/TestCompression.java	(working copy)
+@@ -163,6 +163,7 @@
+ 
+     if (codec == null) {
+       codec = new GzipCodec();
++      ReflectionUtils.setConf(codec, getConf());
+     }
+     Path p = new Path(getDataFilePath().toString()
+         + codec.getDefaultExtension());
+Index: ivy.xml
+===================================================================
+--- ivy.xml	(revision 1205535)
++++ ivy.xml	(working copy)
+@@ -67,11 +67,20 @@
      <artifact conf="master"/>
    </publications>
    <dependencies>
@@ -11,24 +23,24 @@ index d686c76..a882dae 100644
 -      rev="${hadoop-core.cloudera.version}" conf="cloudera->default"/>
 -    <dependency org="org.apache.hadoop" name="hadoop-test"
 -      rev="${hadoop-core.cloudera.version}" conf="clouderatest->default"/>
--
 +    <dependency org="org.apache.hadoop" name="hadoop-common" rev="${hadoop.version}"
-+      conf="cloudera->default"/> 
-+    <dependency org="org.apache.hadoop" name="hadoop-common-test" rev="${hadoop.version}"
-+                conf="clouderatest->default"/> 
++      conf="cloudera->default">
++      <artifact name="hadoop-common" type="jar" />
++      <artifact name="hadoop-common" type="jar" m:classifier="tests"/>
++    </dependency>
 +    <dependency org="org.apache.hadoop" name="hadoop-hdfs" rev="${hadoop.version}"
-+      conf="cloudera->default"/>
-+    <dependency org="org.apache.hadoop" name="hadoop-hdfs-test" rev="${hadoop.version}"
-+      conf="clouderatest->default"/>
-+    <dependency org="org.apache.hadoop" name="hadoop-mapred" rev="${hadoop.version}"
++      conf="cloudera->default">
++      <artifact name="hadoop-hdfs" type="jar" />
++      <artifact name="hadoop-hdfs" type="jar" m:classifier="tests"/>
++    </dependency>
++    <dependency org="org.apache.hadoop" name="hadoop-mapreduce-client-common" rev="${hadoop.version}"
 +       conf="cloudera->default"/>
-+    <dependency org="org.apache.hadoop" name="hadoop-mapred-test" rev="${hadoop.version}"
-+      conf="clouderatest->default"/>
-+    
++    <dependency org="org.apache.hadoop" name="hadoop-mapreduce-client-core" rev="${hadoop.version}"
++       conf="cloudera->default"/>
+ 
      <!-- Common dependencies for Sqoop -->
      <dependency org="commons-cli" name="commons-cli"
-       rev="${commons-cli.version}" conf="common->default"/>
-@@ -121,9 +128,13 @@ under the License.
+@@ -121,9 +130,12 @@
        <exclude org="com.sun.jersey" module="jersey-server"/>
        <exclude org="org.apache.thrift" module="thrift"/>
        <exclude org="log4j" module="log4j"/>
@@ -37,19 +49,19 @@ index d686c76..a882dae 100644
      </dependency>
  
      <exclude org="org.apache.hadoop" module="avro"/>
-+    <exclude org="commons-configuration" module="commons-configuration" />
 +    <exclude org="commons-daemon" module="commons-daemon" />
  
    </dependencies>
  </ivy-module>
-diff --git ivy/libraries.properties b/ivy/libraries.properties
-index dbbcb3b..8ba9a8b 100644
---- ivy/libraries.properties
-+++ ivy/libraries.properties
-@@ -29,14 +29,14 @@ commons-lang.version=2.4
+Index: ivy/libraries.properties
+===================================================================
+--- ivy/libraries.properties	(revision 1205535)
++++ ivy/libraries.properties	(working copy)
+@@ -28,15 +28,14 @@
+ commons-lang.version=2.4
  commons-logging.version=1.0.4
  
- # Cloudera Distribution dependency version
+-# Cloudera Distribution dependency version
 -hadoop-core.cloudera.version=0.20.2-cdh3u1
 +hadoop.version=0.23.0-SNAPSHOT
  
@@ -63,190 +75,3 @@ index dbbcb3b..8ba9a8b 100644
  
  junit.version=4.5
  
-Index: src/test/com/cloudera/sqoop/lib/TestLargeObjectLoader.java
-===================================================================
---- src/test/com/cloudera/sqoop/lib/TestLargeObjectLoader.java	(revision 1198954)
-+++ src/test/com/cloudera/sqoop/lib/TestLargeObjectLoader.java	(working copy)
-@@ -31,10 +31,7 @@
- import org.apache.hadoop.conf.Configuration;
- import org.apache.hadoop.fs.FileSystem;
- import org.apache.hadoop.fs.Path;
--import org.apache.hadoop.mapreduce.MapContext;
--import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
- 
--import com.cloudera.sqoop.testutil.MockObjectFactory;
- import com.cloudera.sqoop.testutil.MockResultSet;
- 
- /**
-@@ -45,7 +42,6 @@
-   protected Configuration conf;
-   protected LargeObjectLoader loader;
-   protected Path outDir;
--  protected MapContext mapContext;
- 
-   public void setUp() throws IOException, InterruptedException {
-     conf = new Configuration();
-@@ -60,17 +56,7 @@
-     }
-     fs.mkdirs(outDir);
- 
--    /* A mock MapContext that uses FileOutputCommitter.
--     * This MapContext is actually serving two roles here; when writing the
--     * CLOB files, its OutputCommitter is used to determine where to write
--     * the CLOB data, as these are placed in the task output work directory.
--     * When reading the CLOB data back for verification, we use the
--     * getInputSplit() to determine where to read our source data from--the same
--     * directory. We are repurposing the same context for both output and input.
--     */
--    mapContext = MockObjectFactory.getMapContextForIOPath(conf, outDir);
--    loader = new LargeObjectLoader(mapContext.getConfiguration(),
--        FileOutputFormat.getWorkOutputPath(mapContext));
-+    loader = new LargeObjectLoader(conf, outDir);
-   }
- 
-   public void testReadClobRef()
-@@ -88,7 +74,6 @@
-     assertNotNull(clob);
-     assertTrue(clob.isExternal());
-     loader.close();
--    mapContext.getOutputCommitter().commitTask(mapContext);
-     Reader r = clob.getDataStream(conf, outDir);
-     char [] buf = new char[4096];
-     int chars = r.read(buf, 0, 4096);
-@@ -117,7 +102,6 @@
-     assertNotNull(blob);
-     assertTrue(blob.isExternal());
-     loader.close();
--    mapContext.getOutputCommitter().commitTask(mapContext);
-     InputStream is = blob.getDataStream(conf, outDir);
-     byte [] buf = new byte[4096];
-     int bytes = is.read(buf, 0, 4096);
-Index: src/test/com/cloudera/sqoop/testutil/MockObjectFactory.java
-===================================================================
---- src/test/com/cloudera/sqoop/testutil/MockObjectFactory.java	(revision 1198954)
-+++ src/test/com/cloudera/sqoop/testutil/MockObjectFactory.java	(working copy)
-@@ -1,84 +0,0 @@
--/**
-- * Licensed to the Apache Software Foundation (ASF) under one
-- * or more contributor license agreements.  See the NOTICE file
-- * distributed with this work for additional information
-- * regarding copyright ownership.  The ASF licenses this file
-- * to you under the Apache License, Version 2.0 (the
-- * "License"); you may not use this file except in compliance
-- * with the License.  You may obtain a copy of the License at
-- *
-- *     http://www.apache.org/licenses/LICENSE-2.0
-- *
-- * Unless required by applicable law or agreed to in writing, software
-- * distributed under the License is distributed on an "AS IS" BASIS,
-- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-- * See the License for the specific language governing permissions and
-- * limitations under the License.
-- */
--
--package com.cloudera.sqoop.testutil;
--
--import java.io.IOException;
--
--import org.apache.hadoop.conf.Configuration;
--import org.apache.hadoop.fs.Path;
--import org.apache.hadoop.mapreduce.Counters;
--import org.apache.hadoop.mapreduce.InputSplit;
--import org.apache.hadoop.mapreduce.MapContext;
--import org.apache.hadoop.mapreduce.OutputCommitter;
--import org.apache.hadoop.mapreduce.TaskAttemptID;
--import org.apache.hadoop.mapreduce.lib.input.FileSplit;
--import org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter;
--import org.apache.hadoop.mrunit.mapreduce.mock.MockReporter;
--
--/**
-- * Allows the creation of various mock objects for testing purposes.
-- */
--public final class MockObjectFactory {
--
--  /**
--   * Returns a mock MapContext that has both an OutputCommitter and an
--   * InputSplit wired to the specified path.
--   * Used for testing LargeObjectLoader.
--   */
--  public static MapContext getMapContextForIOPath(Configuration conf, Path p) {
--    return new MockMapContextWithCommitter(conf, p);
--  }
--
--  private static class MockMapContextWithCommitter
--      extends MapContext<Object, Object, Object, Object> {
--    private Path path;
--    private Configuration conf;
--
--    public MockMapContextWithCommitter(Configuration c, Path p) {
--      super(c, new TaskAttemptID("jt", 0, true, 0, 0),
--            null, null, null, new MockReporter(new Counters()), null);
--
--      this.path = p;
--      this.conf = c;
--    }
--
--    @Override
--    public OutputCommitter getOutputCommitter() {
--      try {
--        return new FileOutputCommitter(path, this);
--      } catch (IOException ioe) {
--        return null;
--      }
--    }
--
--    @Override
--    public InputSplit getInputSplit() {
--      return new FileSplit(new Path(path, "inputFile"), 0, 0, new String[0]);
--    }
--
--    @Override
--    public Configuration getConfiguration() {
--      return conf;
--    }
--  }
--
--  private MockObjectFactory() {
--    // Disable explicity object creation
--  }
--}
-Index: src/test/com/cloudera/sqoop/mapreduce/db/TestDataDrivenDBInputFormat.java
-===================================================================
---- src/test/com/cloudera/sqoop/mapreduce/db/TestDataDrivenDBInputFormat.java	(revision 1198954)
-+++ src/test/com/cloudera/sqoop/mapreduce/db/TestDataDrivenDBInputFormat.java	(working copy)
-@@ -23,13 +23,14 @@
- import java.io.DataOutput;
- import java.io.IOException;
- 
-+import junit.framework.TestCase;
-+
- import org.apache.commons.logging.Log;
- import org.apache.commons.logging.LogFactory;
- import org.apache.hadoop.fs.*;
- import org.apache.hadoop.conf.Configuration;
- import org.apache.hadoop.io.NullWritable;
- import org.apache.hadoop.io.WritableComparable;
--import org.apache.hadoop.mapred.HadoopTestCase;
- import org.apache.hadoop.mapreduce.*;
- import org.apache.hadoop.mapreduce.lib.db.*;
- import org.apache.hadoop.mapreduce.lib.output.*;
-@@ -38,7 +39,7 @@
- /**
-  * Test aspects of DataDrivenDBInputFormat.
-  */
--public class TestDataDrivenDBInputFormat extends HadoopTestCase {
-+public class TestDataDrivenDBInputFormat extends TestCase {
- 
-   private static final Log LOG = LogFactory.getLog(
-       TestDataDrivenDBInputFormat.class);
-@@ -52,10 +53,6 @@
- 
-   private static final String OUT_DIR;
- 
--  public TestDataDrivenDBInputFormat() throws IOException {
--    super(LOCAL_MR, LOCAL_FS, 1, 1);
--  }
--
-   static {
-     OUT_DIR = System.getProperty("test.build.data", "/tmp") + "/dddbifout";
-   }

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/hive/SPECS/hive.spec
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/hive/SPECS/hive.spec?rev=1205607&r1=1205606&r2=1205607&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/hive/SPECS/hive.spec (original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/hive/SPECS/hive.spec Wed Nov 23 21:26:18 2011
@@ -115,7 +115,7 @@ This optional package hosts a metadata s
 
 
 %prep
-%setup -n apache-hive-f412000
+%setup -n apache-hive-4910f33
 %patch0 -p0 
 
 %build

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec?rev=1205607&r1=1205606&r2=1205607&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec (original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/mahout/SPECS/mahout.spec Wed Nov 23 21:26:18 2011
@@ -66,7 +66,7 @@ diverse community to facilitate discussi
 also on potential use cases. Come to the mailing lists to find out more.
     
 %prep
-%setup -n apache-mahout-464be41
+%setup -n apache-mahout-dc3dcf5
 %patch0 -p0
 
 %build

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/sqoop/SPECS/sqoop.spec
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/sqoop/SPECS/sqoop.spec?rev=1205607&r1=1205606&r2=1205607&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/sqoop/SPECS/sqoop.spec (original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop-packages/src/rpm/sqoop/SPECS/sqoop.spec Wed Nov 23 21:26:18 2011
@@ -79,7 +79,7 @@ Shared metadata repository for Sqoop. Th
 server for Sqoop clients across a network to use.
 
 %prep
-%setup -n apache-sqoop-e9d0bed
+%setup -n apache-sqoop-1bf7cbf
 %patch0 -p0
 
 %build

Modified: incubator/bigtop/branches/hadoop-0.23/bigtop.mk
URL: http://svn.apache.org/viewvc/incubator/bigtop/branches/hadoop-0.23/bigtop.mk?rev=1205607&r1=1205606&r2=1205607&view=diff
==============================================================================
--- incubator/bigtop/branches/hadoop-0.23/bigtop.mk (original)
+++ incubator/bigtop/branches/hadoop-0.23/bigtop.mk Wed Nov 23 21:26:18 2011
@@ -79,7 +79,7 @@ HIVE_TARBALL_DST=hive-$(HIVE_BASE_VERSIO
 #HIVE_TARBALL_SRC=$(HIVE_TARBALL_DST)
 #HIVE_SITE=$(APACHE_MIRROR)/hive/hive-$(HIVE_BASE_VERSION)/
 HIVE_SITE=https://github.com/apache/hive/tarball
-HIVE_TARBALL_SRC=f412000
+HIVE_TARBALL_SRC=4910f33
 $(eval $(call PACKAGE,hive,HIVE))
 
 # Sqoop
@@ -93,7 +93,7 @@ SQOOP_TARBALL_DST=sqoop-$(SQOOP_BASE_VER
 #SQOOP_TARBALL_SRC=$(SQOOP_TARBALL_DST)
 #SQOOP_SITE=http://github.com/downloads/cloudera/sqoop/
 SQOOP_SITE=https://github.com/apache/sqoop/tarball
-SQOOP_TARBALL_SRC=e9d0bed
+SQOOP_TARBALL_SRC=1bf7cbf
 $(eval $(call PACKAGE,sqoop,SQOOP))
 
 # Oozie
@@ -132,7 +132,7 @@ MAHOUT_RELEASE_VERSION=1
 MAHOUT_TARBALL_DST=mahout-distribution-$(MAHOUT_BASE_VERSION)-src.tar.gz
 #MAHOUT_TARBALL_SRC=$(MAHOUT_TARBALL_DST)
 #MAHOUT_SITE=$(APACHE_MIRROR)/mahout/0.5/
-MAHOUT_TARBALL_SRC=464be41
+MAHOUT_TARBALL_SRC=dc3dcf5
 MAHOUT_SITE=https://github.com/apache/mahout/tarball
 $(eval $(call PACKAGE,mahout,MAHOUT))