You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by kh...@apache.org on 2013/09/05 05:09:48 UTC

svn commit: r1520202 [4/6] - in /hive/trunk/hcatalog: ./ bin/ build-support/ant/ conf/ core/ core/src/main/java/org/apache/hive/hcatalog/cli/ core/src/main/java/org/apache/hive/hcatalog/cli/SemanticAnalysis/ core/src/main/java/org/apache/hive/hcatalog/...

Modified: hive/trunk/hcatalog/src/packages/templates/conf/hive-site.xml.template
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/packages/templates/conf/hive-site.xml.template?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/packages/templates/conf/hive-site.xml.template (original)
+++ hive/trunk/hcatalog/src/packages/templates/conf/hive-site.xml.template Thu Sep  5 03:09:42 2013
@@ -87,7 +87,7 @@
 
 <property>
   <name>hive.semantic.analyzer.factory.impl</name>
-  <value>org.apache.hcatalog.cli.HCatSemanticAnalyzerFactory</value>
+  <value>org.apache.hive.hcatalog.cli.HCatSemanticAnalyzerFactory</value>
   <description>controls which SemanticAnalyzerFactory implemenation class is used by CLI</description>
 </property>
 

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/drivers/TestDriverHadoop.pm
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/drivers/TestDriverHadoop.pm?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/drivers/TestDriverHadoop.pm (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/drivers/TestDriverHadoop.pm Thu Sep  5 03:09:42 2013
@@ -145,7 +145,7 @@ sub runTest
 	               $testCmd->{'group'} .  "_" .  $testCmd->{'num'} . ".$i.out";
                    $tableName = $results[$i];
 	           $modifiedTestCmd{'num'} = $testCmd->{'num'} . "_" . $i . "_benchmark";
-                   $modifiedTestCmd{'pig'} = "a = load '$tableName' using org.apache.hcatalog.pig.HCatLoader(); store a into ':OUTPATH:';";
+                   $modifiedTestCmd{'pig'} = "a = load '$tableName' using org.apache.hive.hcatalog.pig.HCatLoader(); store a into ':OUTPATH:';";
                    my $r = $self->runPig(\%modifiedTestCmd, $log, 1);
 	           $outputs[$i] = $r->{'output'};
                } else {
@@ -185,7 +185,7 @@ sub dumpPigTable
     my $outfile = $testCmd->{'outpath'} . $testCmd->{'group'} . "_" . $testCmd->{'num'}  . $id . "dump.out";
 
     open(FH, "> $pigfile") or die "Unable to open file $pigfile to write pig script, $ERRNO\n";
-    print FH "a = load '$table' using org.apache.hcatalog.pig.HCatLoader(); store a into '$outfile';\n";
+    print FH "a = load '$table' using org.apache.hive.hcatalog.pig.HCatLoader(); store a into '$outfile';\n";
     close(FH);
 
 

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/drivers/TestDriverPig.pm
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/drivers/TestDriverPig.pm?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/drivers/TestDriverPig.pm (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/drivers/TestDriverPig.pm Thu Sep  5 03:09:42 2013
@@ -184,7 +184,7 @@ sub runTest
 	               $testCmd->{'group'} .  "_" .  $testCmd->{'num'} . ".$i.out";
                    $tableName = $results[$i];
 	           $modifiedTestCmd{'num'} = $testCmd->{'num'} . "_" . $i . "_benchmark";
-                   $modifiedTestCmd{'pig'} = "a = load '$tableName' using org.apache.hcatalog.pig.HCatLoader(); store a into ':OUTPATH:';";
+                   $modifiedTestCmd{'pig'} = "a = load '$tableName' using org.apache.hive.hcatalog.pig.HCatLoader(); store a into ':OUTPATH:';";
                    my $r = $self->runPig(\%modifiedTestCmd, $log, 1, 1);
 	           $outputs[$i] = $r->{'output'};
                } else {

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/hadoop.conf
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/hadoop.conf?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/hadoop.conf (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/hadoop.conf Thu Sep  5 03:09:42 2013
@@ -40,7 +40,7 @@ $cfg = {
 				{
 				 'num' => 1
 				,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.SimpleRead -libjars :HCAT_JAR: :THRIFTSERVER: studenttab10k :OUTPATH:
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.SimpleRead -libjars :HCAT_JAR: :THRIFTSERVER: studenttab10k :OUTPATH:
 \,
                                 ,'sql' => q\select name, age from studenttab10k;\
                                 ,'floatpostprocess' => 1
@@ -51,7 +51,7 @@ jar :FUNCPATH:/testudf.jar org.apache.hc
                                 ,'hcat_prep'=>q\drop table if exists hadoop_checkin_2;
 create table hadoop_checkin_2 (name string, age int, gpa double) STORED AS TEXTFILE;\
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.ReadWrite -libjars :HCAT_JAR: :THRIFTSERVER: studenttab10k hadoop_checkin_2
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.ReadWrite -libjars :HCAT_JAR: :THRIFTSERVER: studenttab10k hadoop_checkin_2
 \,
                                 ,'result_table' => 'hadoop_checkin_2'
                                 ,'sql' => q\select * from studenttab10k;\
@@ -63,7 +63,7 @@ jar :FUNCPATH:/testudf.jar org.apache.hc
                                 ,'hcat_prep'=>q\drop table if exists hadoop_checkin_3;
 create table hadoop_checkin_3 (age int, cnt int) STORED AS TEXTFILE;\
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.GroupByAge -libjars :HCAT_JAR: :THRIFTSERVER: studenttab10k hadoop_checkin_3
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.GroupByAge -libjars :HCAT_JAR: :THRIFTSERVER: studenttab10k hadoop_checkin_3
 \,
                                 ,'result_table' => 'hadoop_checkin_3'
                                 ,'sql' => q\select age, count(*) from studenttab10k group by age;\
@@ -71,7 +71,7 @@ jar :FUNCPATH:/testudf.jar org.apache.hc
                                  # Read from a partitioned table
                                  'num' => 4
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.SimpleRead -libjars :HCAT_JAR: :THRIFTSERVER: studentparttab30k :OUTPATH:
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.SimpleRead -libjars :HCAT_JAR: :THRIFTSERVER: studentparttab30k :OUTPATH:
 \,
                                 ,'sql' => q\select name, age from studentparttab30k;\
                                 ,'floatpostprocess' => 1
@@ -82,7 +82,7 @@ jar :FUNCPATH:/testudf.jar org.apache.hc
                                 ,'hcat_prep'=>q\drop table if exists hadoop_checkin_5;
 create table hadoop_checkin_5 (name string, age int) partitioned by (ds string) STORED AS TEXTFILE;\
                                 ,'hadoop' => q?
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.WriteTextPartitioned -libjars :HCAT_JAR: :THRIFTSERVER: studentparttab30k hadoop_checkin_5 ds=\"20110924\"
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.WriteTextPartitioned -libjars :HCAT_JAR: :THRIFTSERVER: studentparttab30k hadoop_checkin_5 ds=\"20110924\"
 ?,
                                 ,'result_table' => 'hadoop_checkin_5'
                                 ,'sql' => q\select name, age, ds from studentparttab30k where ds='20110924';\
@@ -94,7 +94,7 @@ jar :FUNCPATH:/testudf.jar org.apache.hc
                                 ,'hcat_prep'=>q\drop table if exists hadoop_checkin_6;
 create table hadoop_checkin_6 (name string, age int) partitioned by (ds string) STORED AS TEXTFILE;\
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.WriteTextPartitioned -libjars :HCAT_JAR: :THRIFTSERVER: studentparttab30k hadoop_checkin_6
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.WriteTextPartitioned -libjars :HCAT_JAR: :THRIFTSERVER: studentparttab30k hadoop_checkin_6
 \,
                                 ,'result_table' => 'hadoop_checkin_6'
                                 ,'sql' => q\select name, age, ds from studentparttab30k;\
@@ -109,7 +109,7 @@ jar :FUNCPATH:/testudf.jar org.apache.hc
                                 {
                                  'num' => 1
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.ReadText -libjars :HCAT_JAR: :THRIFTSERVER: all100k :OUTPATH:
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.ReadText -libjars :HCAT_JAR: :THRIFTSERVER: all100k :OUTPATH:
 \,
                                 ,'sql' => q\select * from all100k;\
                                 ,'floatpostprocess' => 1
@@ -118,7 +118,7 @@ jar :FUNCPATH:/testudf.jar org.apache.hc
                                 {
                                  'num' => 2
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.ReadJson -libjars :HCAT_JAR: :THRIFTSERVER: all100kjson :OUTPATH:
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.ReadJson -libjars :HCAT_JAR: :THRIFTSERVER: all100kjson :OUTPATH:
 \,
                                 ,'sql' => q\select s, i, d from all100kjson;\
                                 ,'floatpostprocess' => 1
@@ -127,7 +127,7 @@ jar :FUNCPATH:/testudf.jar org.apache.hc
                                 {
                                  'num' => 3
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.ReadRC -libjars :HCAT_JAR: :THRIFTSERVER: all100krc :OUTPATH:
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.ReadRC -libjars :HCAT_JAR: :THRIFTSERVER: all100krc :OUTPATH:
 \,
                                 ,'sql' => q\select name, age, floor(gpa) + 0.1 from all100krc;\
                                 ,'floatpostprocess' => 1
@@ -155,7 +155,7 @@ create table hadoop_write_1(
         fields terminated by ':'
         stored as textfile;\
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.WriteText -libjars :HCAT_JAR: :THRIFTSERVER: all100k hadoop_write_1
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.WriteText -libjars :HCAT_JAR: :THRIFTSERVER: all100k hadoop_write_1
 \,
                                 ,'result_table' => 'hadoop_write_1'
                                 ,'sql' => q\select * from all100k;\
@@ -172,10 +172,10 @@ create table hadoop_write_2(
             d double,
             m map<string, string>,
             bb array<struct<a: int, b: string>>)
-            row format serde 'org.apache.hcatalog.data.JsonSerDe'
+            row format serde 'org.apache.hive.hcatalog.data.JsonSerDe'
             stored as textfile;\
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.WriteJson -libjars :HCAT_JAR: :THRIFTSERVER: all100kjson hadoop_write_2
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.WriteJson -libjars :HCAT_JAR: :THRIFTSERVER: all100kjson hadoop_write_2
 \,
                                 ,'result_table' => 'hadoop_write_2'
                                 ,'sql' => q\select s, i, d, '', '' from all100kjson;\
@@ -193,7 +193,7 @@ create table hadoop_write_3(
 stored as rcfile;
 \,
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.WriteRC -libjars :HCAT_JAR: :THRIFTSERVER: all100krc hadoop_write_3
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.WriteRC -libjars :HCAT_JAR: :THRIFTSERVER: all100krc hadoop_write_3
 \,
                                 ,'result_table' => 'hadoop_write_3'
                                 ,'sql' => q\select name, age, floor(gpa) + 0.1 from all100krc;\
@@ -210,7 +210,7 @@ create table hadoop_write_4(
 stored as sequencefile;
 \,
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.WriteRC -libjars :HCAT_JAR: :THRIFTSERVER: all100krc hadoop_write_4
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.WriteRC -libjars :HCAT_JAR: :THRIFTSERVER: all100krc hadoop_write_4
 \,
                                 ,'result_table' => 'hadoop_write_4'
                                 ,'sql' => q\select name, age, floor(gpa) + 0.1 from all100krc;\
@@ -225,9 +225,9 @@ jar :FUNCPATH:/testudf.jar org.apache.hc
                                 {
                                  'num' => 1
                                 ,'hcat_prep'=>q\drop table if exists hadoop_hbase_1;
-create table hadoop_hbase_1(key string, gpa string) STORED BY 'org.apache.hcatalog.hbase.HBaseHCatStorageHandler' TBLPROPERTIES ('hbase.columns.mapping'=':key,info:gpa');\
+create table hadoop_hbase_1(key string, gpa string) STORED BY 'org.apache.hive.hcatalog.hbase.HBaseHCatStorageHandler' TBLPROPERTIES ('hbase.columns.mapping'=':key,info:gpa');\
                                 ,'hadoop' => q\
-jar :FUNCPATH:/testudf.jar org.apache.hcatalog.utils.HBaseReadWrite -libjars :HCAT_JAR: :THRIFTSERVER: :INPATH:/studenttab10k hadoop_hbase_1 :OUTPATH:
+jar :FUNCPATH:/testudf.jar org.apache.hive.hcatalog.utils.HBaseReadWrite -libjars :HCAT_JAR: :THRIFTSERVER: :INPATH:/studenttab10k hadoop_hbase_1 :OUTPATH:
 \,
                                 ,'sql' => q\select name, sum(gpa) from studenttab10k group by name;\
                                 ,'floatpostprocess' => 1

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/hive.conf
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/hive.conf?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/hive.conf (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/hive.conf Thu Sep  5 03:09:42 2013
@@ -167,7 +167,7 @@ insert into TABLE hive_write_1 select t,
                         'num' => 2,
                          'sql' => q\
 drop table if exists hive_write_2;
-create table hive_write_2 (name string, age int, gpa double) row format serde 'org.apache.hcatalog.data.JsonSerDe' stored as textfile;
+create table hive_write_2 (name string, age int, gpa double) row format serde 'org.apache.hive.hcatalog.data.JsonSerDe' stored as textfile;
 insert into TABLE hive_write_2 select s, i, 0.1 from all100kjson;\,
                          'result_table' => 'hive_write_2',
                          'verify_sql' =>"select s, i, 0.1 from all100kjson;",

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/pig.conf
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/pig.conf?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/pig.conf (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/tests/pig.conf Thu Sep  5 03:09:42 2013
@@ -42,8 +42,8 @@ $cfg = {
 				 'num' => 1
                                 ,'hcat_prep'=>q\drop table if exists pig_checkin_1;
 create table pig_checkin_1 (name string, age int, gpa double) STORED AS TEXTFILE;\
-				,'pig' => q\a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
-store a into 'pig_checkin_1' using org.apache.hcatalog.pig.HCatStorer();\,
+				,'pig' => q\a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
+store a into 'pig_checkin_1' using org.apache.hive.hcatalog.pig.HCatStorer();\,
                                 ,'result_table' => 'pig_checkin_1'
                                 ,'sql' => q\select * from studenttab10k;\
                                 ,'floatpostprocess' => 1
@@ -51,8 +51,8 @@ store a into 'pig_checkin_1' using org.a
 				}, 
 				{
 				 'num' => 2
-                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
-b = load 'votertab10k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
+b = load 'votertab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 c = join a by name, b by name;
 store c into ':OUTPATH:';\,
 				,'sql'   => [ 'select s.name, s.age, gpa, v.name, v.age, registration, contributions from studenttab10k s join votertab10k v on (s.name = v.name);']
@@ -61,7 +61,7 @@ store c into ':OUTPATH:';\,
 				}, 
 				{
 				 'num' => 3
-                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = load ':INPATH:/votertab10k' as (name:chararray, age:int, registration:chararray, contributions:float);
 c = join a by name, b by name;
 store c into ':OUTPATH:';\
@@ -75,10 +75,10 @@ store c into ':OUTPATH:';\
 drop table if exists pig_checkin_4_2;
 create table pig_checkin_4_1 (name string, age int, gpa double) STORED AS TEXTFILE;
 create table pig_checkin_4_2 (name string, age int, gpa double) STORED AS TEXTFILE;\
-                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 split a into b if age <=40, c if age > 40;
-store b into 'pig_checkin_4_1' using org.apache.hcatalog.pig.HCatStorer();
-store c into 'pig_checkin_4_2' using org.apache.hcatalog.pig.HCatStorer();\,
+store b into 'pig_checkin_4_1' using org.apache.hive.hcatalog.pig.HCatStorer();
+store c into 'pig_checkin_4_2' using org.apache.hive.hcatalog.pig.HCatStorer();\,
                                 ,'result_table' => ['pig_checkin_4_1','pig_checkin_4_2']
 				,'sql'   => [ 'select * from studenttab10k where age<=40;', 'select * from studenttab10k where age>40;']
                                 ,'floatpostprocess' => 1
@@ -88,9 +88,9 @@ store c into 'pig_checkin_4_2' using org
 				 'num' => 5
                                 ,'hcat_prep'=>q\drop table if exists pig_checkin_5;
 create table pig_checkin_5 (name string, age int, gpa double) STORED AS TEXTFILE;\
-                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 split a into b if age <=40, c if age > 40;
-store b into 'pig_checkin_5' using org.apache.hcatalog.pig.HCatStorer();
+store b into 'pig_checkin_5' using org.apache.hive.hcatalog.pig.HCatStorer();
 store c into ':OUTPATH:';\,
                                 ,'result_table' => ['pig_checkin_5','?']
 				,'sql'   => [ 'select * from studenttab10k where age<=40;', 'select * from studenttab10k where age>40;']
@@ -101,10 +101,10 @@ store c into ':OUTPATH:';\,
 				 'num' => 6
                                 ,'hcat_prep'=>q\drop table if exists pig_checkin_6;
 create table pig_checkin_6 (name string, age int) partitioned by (ds string) STORED AS TEXTFILE;\
-                                ,'pig' => q\a = load 'studentparttab30k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studentparttab30k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = filter a by ds == '20110924';
 c = foreach b generate name, age;
-store c into 'pig_checkin_6' using org.apache.hcatalog.pig.HCatStorer('ds=20110924');\,
+store c into 'pig_checkin_6' using org.apache.hive.hcatalog.pig.HCatStorer('ds=20110924');\,
 #dump a;\,
                                 ,'result_table' => 'pig_checkin_6',
 				,'sql'   => "select name, age, ds from studentparttab30k where ds='20110924';",
@@ -115,9 +115,9 @@ store c into 'pig_checkin_6' using org.a
 				 'num' => 7
                                 ,'hcat_prep'=>q\drop table if exists pig_checkin_7;
 create table pig_checkin_7 (name string, age int) partitioned by (ds string) STORED AS TEXTFILE;\
-                                ,'pig' => q\a = load 'studentparttab30k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studentparttab30k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate name, age, ds;
-store b into 'pig_checkin_7' using org.apache.hcatalog.pig.HCatStorer();\,
+store b into 'pig_checkin_7' using org.apache.hive.hcatalog.pig.HCatStorer();\,
                                 ,'result_table' => 'pig_checkin_7',
 				,'sql'   => "select name, age, ds from studentparttab30k;",
                                 ,'floatpostprocess' => 1
@@ -133,7 +133,7 @@ store b into 'pig_checkin_7' using org.a
                                 {
                                  'ignore' => 1, # Need to checkin HCATALOG-168.
                                  'num' => 1
-                                ,'pig' => q\a = load 'all100k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'all100k' using org.apache.hive.hcatalog.pig.HCatLoader();
 store a into ':OUTPATH:';\,
 				,'sql' => q\select * from all100k;\
                                 ,'floatpostprocess' => 1
@@ -141,7 +141,7 @@ store a into ':OUTPATH:';\,
                                 },
                                 {
                                  'num' => 2
-                                ,'pig' => q\a = load 'all100kjson' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'all100kjson' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate s, i, d;
 store b into ':OUTPATH:';\,
 				,'sql' => q\select s, i, d from all100kjson;\
@@ -150,7 +150,7 @@ store b into ':OUTPATH:';\,
                                 },
                                 {
                                  'num' => 3
-                                ,'pig' => q\a = load 'all100krc' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'all100krc' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate name, age;
 store b into ':OUTPATH:';\,
 				,'sql' => q\select name, age from all100krc;\
@@ -165,7 +165,7 @@ create external table pig_read_4 (name s
 alter table pig_read_4 add partition (b='1') location '/user/hcat/tests/data/studenttab10k';
 alter table pig_read_4 set fileformat rcfile;
 alter table pig_read_4 add partition (b='2') location '/user/hcat/tests/data/all100krc';?
-                                ,'pig' => q\a = load 'pig_read_4' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'pig_read_4' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate name, age, b;
 store b into ':OUTPATH:';\,
 				,'sql' => q\(select name, age, 1 from studenttab10k)
@@ -180,7 +180,7 @@ drop table if exists pig_db_1.pig_read_5
 create external table pig_db_1.pig_read_5 (name string, age int, gpa double) partitioned by (b string) row format delimited fields terminated by '\t' stored as textfile;
 use pig_db_1;
 alter table pig_read_5 add partition (b='1') location '/user/hcat/tests/data/studenttab10k';?
-                                ,'pig' => q\a = load 'pig_db_1.pig_read_5' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'pig_db_1.pig_read_5' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate name, age, b;
 store b into ':OUTPATH:';\,
 				,'sql' => q\select name, age, 1 from studenttab10k;\
@@ -196,7 +196,7 @@ store b into ':OUTPATH:';\,
                                 ,'hcat_prep'=>q\drop table if exists pig_write_1;
 create table pig_write_1(t tinyint,si smallint,i int,b bigint,bool boolean,f float,d double,s string) stored as rcfile;\
                                 ,'pig' => q\a = load ':INPATH:/all100k' using PigStorage(':') as (t:int,si:int,i:int,b:int,bo:boolean,f:float,d:double,s:chararray);
-store a into 'pig_write_1' using org.apache.hcatalog.pig.HCatStorer();\,
+store a into 'pig_write_1' using org.apache.hive.hcatalog.pig.HCatStorer();\,
                                 ,'result_table' => 'pig_write_1'
 				,'sql' => q\select * from all100k;\
                                 ,'floatpostprocess' => 1
@@ -211,12 +211,12 @@ create table pig_write_2(
             d double,
             m map<string, string>,
             bb array<struct<a: int, b: string>>)
-            row format serde 'org.apache.hcatalog.data.JsonSerDe'
+            row format serde 'org.apache.hive.hcatalog.data.JsonSerDe'
             STORED AS TEXTFILE;
 \
-                                ,'pig' => q\a = load 'all100kjson' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'all100kjson' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate s, i, d;
-store b into 'pig_write_2' using org.apache.hcatalog.pig.HCatStorer();\,
+store b into 'pig_write_2' using org.apache.hive.hcatalog.pig.HCatStorer();\,
 				,'sql' => q\select IFNULL(s, ""), IFNULL(i, ""), IFNULL(d, "") from all100kjson;\
                                 ,'result_table' => 'pig_write_2'
                                 ,'floatpostprocess' => 1
@@ -231,9 +231,9 @@ create table pig_write_3(
             gpa double)
 stored as rcfile;
 \
-                                ,'pig' => q\a = load 'all100krc' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'all100krc' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate name, age;
-store b into 'pig_write_3' using org.apache.hcatalog.pig.HCatStorer();\,
+store b into 'pig_write_3' using org.apache.hive.hcatalog.pig.HCatStorer();\,
 				,'sql' => q\select name, age from all100krc;\
                                 ,'result_table' => 'pig_write_3'
                                 ,'floatpostprocess' => 1
@@ -249,10 +249,10 @@ create table pig_write_4(
             gpa double)
 stored as sequencefile;
 \
-                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate name, age, 0.1;
 c = foreach b generate name, age, $2 as gpa;
-store c into 'pig_write_4' using org.apache.hcatalog.pig.HCatStorer();\,
+store c into 'pig_write_4' using org.apache.hive.hcatalog.pig.HCatStorer();\,
 				,'sql' => q\select name, age, 0.1 from studenttab10k;\
                                 ,'result_table' => 'pig_write_4'
                                 ,'floatpostprocess' => 1
@@ -263,9 +263,9 @@ store c into 'pig_write_4' using org.apa
                                  'num' => 5
                                 ,'hcat_prep'=>q?create database if not exists pig_db_1;
 create table if not exists pig_db_1.pig_write_5 (name string, age int) row format delimited fields terminated by '\t' stored as textfile;?
-                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate name, age;
-store b into 'pig_db_1.pig_write_5' using org.apache.hcatalog.pig.HCatStorer();\,
+store b into 'pig_db_1.pig_write_5' using org.apache.hive.hcatalog.pig.HCatStorer();\,
 				                ,'sql' => q\select name, age from studenttab10k;\
                                 ,'result_table' => 'pig_db_1.pig_write_5'
                                 }
@@ -279,10 +279,10 @@ store b into 'pig_db_1.pig_write_5' usin
                                  'num' => 1
                                 ,'hcat_prep'=>q\drop table if exists pig_change_schema_1;
 create table pig_change_schema_1 (name string) partitioned by (ds string) STORED AS TEXTFILE;\
-                                ,'pig' => q\a = load 'studentparttab30k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studentparttab30k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = filter a by ds == '20110924';
 c = foreach b generate name;
-store c into 'pig_change_schema_1' using org.apache.hcatalog.pig.HCatStorer('ds=20110924');\,
+store c into 'pig_change_schema_1' using org.apache.hive.hcatalog.pig.HCatStorer('ds=20110924');\,
                                 ,'result_table' => 'pig_change_schema_1'
 				,'sql' => q\select name, ds from studentparttab30k where ds='20110924';\
                                 },
@@ -291,10 +291,10 @@ store c into 'pig_change_schema_1' using
                                  'num' => 2
                                 ,'depends_on' => 'Pig_Change_Schema_1'
                                 ,'hcat_prep'=>q\alter table pig_change_schema_1 add columns (age int);\
-                                ,'pig' => q\a = load 'studentparttab30k' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'studentparttab30k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = filter a by ds == '20110925';
 c = foreach b generate name, age;
-store c into 'pig_change_schema_1' using org.apache.hcatalog.pig.HCatStorer('ds=20110925');\,
+store c into 'pig_change_schema_1' using org.apache.hive.hcatalog.pig.HCatStorer('ds=20110925');\,
                                 ,'result_table' => 'pig_change_schema_1'
 				                ,'sql' => q\(select name, '', ds from studentparttab30k where ds='20110924')
                                             union all
@@ -304,7 +304,7 @@ store c into 'pig_change_schema_1' using
                                  # I don't like this, I'm using one test to setup for the next.  But I don't know how else to do this.
                                  'num' => 3
                                 , 'depends_on' => 'Pig_Change_Schema_2'
-                                ,'pig' => q\a = load 'pig_change_schema_1' using org.apache.hcatalog.pig.HCatLoader();
+                                ,'pig' => q\a = load 'pig_change_schema_1' using org.apache.hive.hcatalog.pig.HCatLoader();
 c = foreach a generate name, age, ds;
 store c into ':OUTPATH:';\
 				                ,'sql' => q\(select name, '', ds from studentparttab30k where ds='20110924')
@@ -319,15 +319,15 @@ store c into ':OUTPATH:';\
                                 {
                                  'num' => 1
                                 ,'hcat_prep'=>q\drop table if exists pig_hbase_1;
-create table pig_hbase_1(key string, age string, gpa string) STORED BY 'org.apache.hcatalog.hbase.HBaseHCatStorageHandler' TBLPROPERTIES ('hbase.columns.mapping'=':key,info:age,info:gpa');\
+create table pig_hbase_1(key string, age string, gpa string) STORED BY 'org.apache.hive.hcatalog.hbase.HBaseHCatStorageHandler' TBLPROPERTIES ('hbase.columns.mapping'=':key,info:age,info:gpa');\
                                 ,'pig' => q\set hcat.hbase.output.bulkMode 'false'
 a = load ':INPATH:/studenttab10k' as (name:chararray, age:int, gpa:float);
 b = group a by name;
 c = foreach b generate group as name, AVG(a.age) as age, AVG(a.gpa) as gpa;
 d = foreach c generate name as key, (chararray)age, (chararray)gpa as gpa;
-store d into 'pig_hbase_1' using org.apache.hcatalog.pig.HCatStorer();
+store d into 'pig_hbase_1' using org.apache.hive.hcatalog.pig.HCatStorer();
 exec
-e = load 'pig_hbase_1' using org.apache.hcatalog.pig.HCatLoader();
+e = load 'pig_hbase_1' using org.apache.hive.hcatalog.pig.HCatLoader();
 store e into ':OUTPATH:';\,
                                 ,'result_table' => ['pig_hbase_1','?']
 				,'sql'   => [ 'select name, avg(cast(age as decimal(10,5))), avg(gpa) from studenttab10k group by name;', 'select name, avg(cast(age as decimal(10,5))), avg(gpa) from studenttab10k group by name;' ]
@@ -338,17 +338,17 @@ store e into ':OUTPATH:';\,
                                  # multiquery
                                  'num' => 2
                                 ,'hcat_prep'=>q\drop table if exists pig_hbase_2_1;
-create table pig_hbase_2_1(key string, age string, gpa string) STORED BY 'org.apache.hcatalog.hbase.HBaseHCatStorageHandler' TBLPROPERTIES ('hbase.columns.mapping'=':key,info:age,info:gpa');
+create table pig_hbase_2_1(key string, age string, gpa string) STORED BY 'org.apache.hive.hcatalog.hbase.HBaseHCatStorageHandler' TBLPROPERTIES ('hbase.columns.mapping'=':key,info:age,info:gpa');
 drop table if exists pig_hbase_2_2;
-create table pig_hbase_2_2(key string, age string, gpa string) STORED BY 'org.apache.hcatalog.hbase.HBaseHCatStorageHandler' TBLPROPERTIES ('hbase.columns.mapping'=':key,info:age,info:gpa');
+create table pig_hbase_2_2(key string, age string, gpa string) STORED BY 'org.apache.hive.hcatalog.hbase.HBaseHCatStorageHandler' TBLPROPERTIES ('hbase.columns.mapping'=':key,info:age,info:gpa');
 \
                                 ,'pig' => q\set hcat.hbase.output.bulkMode 'false'
 a = load ':INPATH:/studenttab10k' as (name:chararray, age:int, gpa:float);
 b = group a by name;
 c = foreach b generate group as name, AVG(a.age) as age, AVG(a.gpa) as gpa;
 d = foreach c generate name as key, (chararray)age, (chararray)gpa as gpa;
-store d into 'pig_hbase_2_1' using org.apache.hcatalog.pig.HCatStorer();
-store d into 'pig_hbase_2_2' using org.apache.hcatalog.pig.HCatStorer();\,
+store d into 'pig_hbase_2_1' using org.apache.hive.hcatalog.pig.HCatStorer();
+store d into 'pig_hbase_2_2' using org.apache.hive.hcatalog.pig.HCatStorer();\,
                                 ,'result_table' => ['pig_hbase_2_1','pig_hbase_2_2']
 				,'sql'   => [ 'select name, avg(cast(age as decimal(10,5))), avg(gpa) from studenttab10k group by name;', 'select name, avg(cast(age as decimal(10,5))), avg(gpa) from studenttab10k group by name;']
                                 ,'floatpostprocess' => 1
@@ -365,7 +365,7 @@ store d into 'pig_hbase_2_2' using org.a
 create table pig_hcat_barrier_1 (name string, age int, gpa double) partitioned by (b string) CLUSTERED BY (name) INTO 1 BUCKETS STORED AS TEXTFILE;\
                                 ,'pig' => q\
 a = load ':INPATH:/studenttab10k' as (name:chararray, age:int, gpa:double);
-store a into 'pig_hcat_barrier_1' using org.apache.hcatalog.pig.HCatStorer('b=1'); \,
+store a into 'pig_hcat_barrier_1' using org.apache.hive.hcatalog.pig.HCatStorer('b=1'); \,
                                 ,'expected_err_regex' => 'not supported'
                                 },
                                 {
@@ -374,7 +374,7 @@ store a into 'pig_hcat_barrier_1' using 
 create table pig_hcat_barrier_2 (name string, age int, gpa double) partitioned by (b string) CLUSTERED BY (name) SORTED BY (name) INTO 1 BUCKETS STORED AS TEXTFILE;\
                                 ,'pig' => q\
 a = load ':INPATH:/studenttab10k' as (name:chararray, age:int, gpa:double);
-store a into 'pig_hcat_barrier_2' using org.apache.hcatalog.pig.HCatStorer('b=1'); \,
+store a into 'pig_hcat_barrier_2' using org.apache.hive.hcatalog.pig.HCatStorer('b=1'); \,
                                 ,'expected_err_regex' => 'not supported'
                                 },
                         ],
@@ -396,7 +396,7 @@ alter table pig_hcat_coop_1 partition(b=
 alter table pig_hcat_coop_1 replace columns (name string, age int);
 :
                                 ,'pig' => q\
-a = load 'pig_hcat_coop_1' using org.apache.hcatalog.pig.HCatLoader();
+a = load 'pig_hcat_coop_1' using org.apache.hive.hcatalog.pig.HCatLoader();
 store a into ':OUTPATH:';\,
 				,'sql'   => q\select name, age, '1' from studenttab10k union all select name, age, '2' from votertab10k;\
                                 ,'floatpostprocess' => 1
@@ -414,7 +414,7 @@ alter table pig_hcat_coop_2 partition(b=
 alter table pig_hcat_coop_2 replace columns (age int, name string);
 :
                                 ,'pig' => q\
-a = load 'pig_hcat_coop_2' using org.apache.hcatalog.pig.HCatLoader();
+a = load 'pig_hcat_coop_2' using org.apache.hive.hcatalog.pig.HCatLoader();
 store a into ':OUTPATH:';\,
 				,'sql'   => q\select age, name, '1' from studenttab10k union all select age, name, '2' from all100krc;\
                                 ,'floatpostprocess' => 1
@@ -428,7 +428,7 @@ store a into ':OUTPATH:';\,
                                  # test reading tuples from the complex table
                                  'num' => 1 
                                 ,'pig' => q\
-a = load 'studentcomplextab10k' using org.apache.hcatalog.pig.HCatLoader();
+a = load 'studentcomplextab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate flatten(s); 
 store b into ':OUTPATH:';\,
 				                ,'sql'   => q\select IFNULL(name, ""), IFNULL(age, ""), IFNULL(gpa, "") from studentcomplextab10k;\
@@ -438,7 +438,7 @@ store b into ':OUTPATH:';\,
                                  # test reading maps from the complex table
                                  'num' => 2 
                                 ,'pig' => q\
-a = load 'studentcomplextab10k' using org.apache.hcatalog.pig.HCatLoader();
+a = load 'studentcomplextab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate s.name as n1, m#'name' as n2;
 c = filter b by n1 != '' and n2 is not null;
 store c into ':OUTPATH:';\,
@@ -451,7 +451,7 @@ store c into ':OUTPATH:';\,
                                  # test reading arrays from the complex table
                                  'num' => 3 
                                 ,'pig' => q\
-a = load 'studentcomplextab10k' using org.apache.hcatalog.pig.HCatLoader();
+a = load 'studentcomplextab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate s.name as n1, flatten(a);
 c = filter b by n1 != '' ;
 store c into ':OUTPATH:';\,
@@ -467,11 +467,11 @@ store c into ':OUTPATH:';\,
                                 ,'hcat_prep'=>q\drop table if exists pig_complex_4;
 create table pig_complex_4 (s struct<name: string, age: int, gpa: double>) STORED AS TEXTFILE;\
                                 ,'pig' => q\
-a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
+a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate TOTUPLE(name, age, gpa) as s;
-store b into 'pig_complex_4' using org.apache.hcatalog.pig.HCatStorer();
+store b into 'pig_complex_4' using org.apache.hive.hcatalog.pig.HCatStorer();
 exec;
-c = load 'pig_complex_4' using org.apache.hcatalog.pig.HCatLoader();
+c = load 'pig_complex_4' using org.apache.hive.hcatalog.pig.HCatLoader();
 d = foreach c generate flatten(s);
 store d into ':OUTPATH:';\
 				                ,'sql'   => q\select name, age, gpa from studenttab10k;\
@@ -484,11 +484,11 @@ store d into ':OUTPATH:';\
                                 ,'hcat_prep'=>q\drop table if exists pig_complex_5;
 create table pig_complex_5 (m map<string, string>) STORED AS TEXTFILE;\
                                 ,'pig' => q\
-a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
+a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate TOMAP('name', name, 'age', (chararray)age, 'gpa', (chararray)gpa) as m;
-store b into 'pig_complex_5' using org.apache.hcatalog.pig.HCatStorer();
+store b into 'pig_complex_5' using org.apache.hive.hcatalog.pig.HCatStorer();
 exec;
-c = load 'pig_complex_5' using org.apache.hcatalog.pig.HCatLoader();
+c = load 'pig_complex_5' using org.apache.hive.hcatalog.pig.HCatLoader();
 d = foreach c generate m#'name', m#'age', m#'gpa';
 store d into ':OUTPATH:';\
 				                ,'sql'   => q\select name, age, gpa from studenttab10k;\
@@ -501,14 +501,14 @@ store d into ':OUTPATH:';\
                                 ,'hcat_prep'=>q\drop table if exists pig_complex_6;
 create table pig_complex_6 (a array<string>) STORED AS TEXTFILE;\
                                 ,'pig' => q\
-a = load 'studenttab10k' using org.apache.hcatalog.pig.HCatLoader();
+a = load 'studenttab10k' using org.apache.hive.hcatalog.pig.HCatLoader();
 b = foreach a generate name;
 c = distinct b;
 d = group c all;
 e = foreach d generate $1 as a;
-store e into 'pig_complex_6' using org.apache.hcatalog.pig.HCatStorer();
+store e into 'pig_complex_6' using org.apache.hive.hcatalog.pig.HCatStorer();
 exec;
-f = load 'pig_complex_6' using org.apache.hcatalog.pig.HCatLoader();
+f = load 'pig_complex_6' using org.apache.hive.hcatalog.pig.HCatLoader();
 g = foreach f generate flatten(a);
 store g into ':OUTPATH:';\
 				                ,'sql'   => q\select distinct name from studenttab10k;\

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/tools/generate/generate_data.pl
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/tools/generate/generate_data.pl?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/tools/generate/generate_data.pl (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/tools/generate/generate_data.pl Thu Sep  5 03:09:42 2013
@@ -319,7 +319,7 @@ stored as rcfile
 location '$location';\n";
     } elsif ($format eq "json") {
         print $hivefp "
-row format serde 'org.apache.hcatalog.data.JsonSerDe'
+row format serde 'org.apache.hive.hcatalog.data.JsonSerDe'
 stored as textfile
 location '$location'
 ;\n";
@@ -657,7 +657,7 @@ for (my $i = 0; $i < $numRows; $i++) {
             d double,
             m map<string, string>,
             bb array<struct<a: int, b: string>>)
-            row format serde 'org.apache.hcatalog.data.JsonSerDe'
+            row format serde 'org.apache.hive.hcatalog.data.JsonSerDe'
             STORED AS TEXTFILE 
             location '$hdfsTargetDir/$tableName';\n";
         open(PLAIN, ">$tableName.plain") or

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataReaderMaster.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataReaderMaster.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataReaderMaster.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataReaderMaster.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.File;
 import java.io.FileNotFoundException;
@@ -30,11 +30,11 @@ import java.util.Map;
 import java.util.Map.Entry;
 import java.util.Properties;
 
-import org.apache.hcatalog.common.HCatException;
-import org.apache.hcatalog.data.transfer.DataTransferFactory;
-import org.apache.hcatalog.data.transfer.HCatReader;
-import org.apache.hcatalog.data.transfer.ReadEntity;
-import org.apache.hcatalog.data.transfer.ReaderContext;
+import org.apache.hive.hcatalog.common.HCatException;
+import org.apache.hive.hcatalog.data.transfer.DataTransferFactory;
+import org.apache.hive.hcatalog.data.transfer.HCatReader;
+import org.apache.hive.hcatalog.data.transfer.ReadEntity;
+import org.apache.hive.hcatalog.data.transfer.ReaderContext;
 
 public class DataReaderMaster {
 

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataReaderSlave.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataReaderSlave.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataReaderSlave.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataReaderSlave.java Thu Sep  5 03:09:42 2013
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.BufferedWriter;
 import java.io.File;
@@ -28,10 +28,10 @@ import java.util.Iterator;
 import java.util.List;
 
 import org.apache.hadoop.mapreduce.InputSplit;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.transfer.DataTransferFactory;
-import org.apache.hcatalog.data.transfer.HCatReader;
-import org.apache.hcatalog.data.transfer.ReaderContext;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.transfer.DataTransferFactory;
+import org.apache.hive.hcatalog.data.transfer.HCatReader;
+import org.apache.hive.hcatalog.data.transfer.ReaderContext;
 
 public class DataReaderSlave {
 

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataWriterMaster.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataWriterMaster.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataWriterMaster.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataWriterMaster.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.File;
 import java.io.FileInputStream;
@@ -32,11 +32,11 @@ import java.util.Map;
 import java.util.Properties;
 import java.util.Map.Entry;
 
-import org.apache.hcatalog.common.HCatException;
-import org.apache.hcatalog.data.transfer.DataTransferFactory;
-import org.apache.hcatalog.data.transfer.HCatWriter;
-import org.apache.hcatalog.data.transfer.WriteEntity;
-import org.apache.hcatalog.data.transfer.WriterContext;
+import org.apache.hive.hcatalog.common.HCatException;
+import org.apache.hive.hcatalog.data.transfer.DataTransferFactory;
+import org.apache.hive.hcatalog.data.transfer.HCatWriter;
+import org.apache.hive.hcatalog.data.transfer.WriteEntity;
+import org.apache.hive.hcatalog.data.transfer.WriterContext;
 
 public class DataWriterMaster {
 

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataWriterSlave.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataWriterSlave.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataWriterSlave.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/DataWriterSlave.java Thu Sep  5 03:09:42 2013
@@ -16,7 +16,7 @@
  * specific language governing permissions and limitations
  * under the License.
  */
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.BufferedReader;
 import java.io.File;
@@ -29,11 +29,11 @@ import java.util.ArrayList;
 import java.util.Iterator;
 import java.util.List;
 
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.transfer.DataTransferFactory;
-import org.apache.hcatalog.data.transfer.HCatWriter;
-import org.apache.hcatalog.data.transfer.WriterContext;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.transfer.DataTransferFactory;
+import org.apache.hive.hcatalog.data.transfer.HCatWriter;
+import org.apache.hive.hcatalog.data.transfer.WriterContext;
 
 public class DataWriterSlave {
 

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/GroupByAge.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/GroupByAge.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/GroupByAge.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/GroupByAge.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 import java.util.Iterator;
@@ -32,14 +32,14 @@ import org.apache.hadoop.mapreduce.Reduc
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HBaseReadWrite.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HBaseReadWrite.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HBaseReadWrite.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HBaseReadWrite.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 
@@ -36,13 +36,13 @@ import org.apache.hadoop.mapreduce.lib.o
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTestDriver.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTestDriver.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTestDriver.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTestDriver.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import org.apache.hadoop.util.ProgramDriver;
 

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTypeCheck.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTypeCheck.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTypeCheck.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTypeCheck.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 import java.util.HashMap;

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTypeCheckHive.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTypeCheckHive.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTypeCheckHive.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/HCatTypeCheckHive.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 import java.util.ArrayList;
@@ -42,7 +42,7 @@ import org.apache.hadoop.hive.serde2.obj
 /**
  * A hive udf to check types of the fields read from hcat. A sample hive query which can use this is:
  *
- * create temporary function typecheck as 'org.apache.hcatalog.utils.HCatTypeCheckHive';
+ * create temporary function typecheck as 'org.apache.hive.hcatalog.utils.HCatTypeCheckHive';
  * select typecheck('map<string,string>+struct<num:int,str:string,dbl:double>+array<map<string,string>>+int', 
  * mymap, mytuple, bagofmap, rownum) from complex;
  *

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadJson.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadJson.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadJson.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadJson.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 
@@ -33,11 +33,11 @@ import org.apache.hadoop.mapreduce.lib.o
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadRC.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadRC.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadRC.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadRC.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 
@@ -33,11 +33,11 @@ import org.apache.hadoop.mapreduce.lib.o
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadText.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadText.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadText.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadText.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 
@@ -33,11 +33,11 @@ import org.apache.hadoop.mapreduce.lib.o
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadWrite.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadWrite.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadWrite.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/ReadWrite.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 
@@ -30,14 +30,14 @@ import org.apache.hadoop.mapreduce.Mappe
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SimpleRead.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SimpleRead.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SimpleRead.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SimpleRead.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 
@@ -34,10 +34,10 @@ import org.apache.hadoop.mapreduce.lib.o
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreComplex.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreComplex.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreComplex.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreComplex.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 import java.util.HashMap;
@@ -30,14 +30,14 @@ import org.apache.hadoop.io.WritableComp
 import org.apache.hadoop.mapreduce.Job;
 import org.apache.hadoop.mapreduce.Mapper;
 import org.apache.hadoop.util.GenericOptionsParser;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "complex"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreDemo.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreDemo.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreDemo.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreDemo.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 import java.util.HashMap;
@@ -29,14 +29,14 @@ import org.apache.hadoop.io.WritableComp
 import org.apache.hadoop.mapreduce.Job;
 import org.apache.hadoop.mapreduce.Mapper;
 import org.apache.hadoop.util.GenericOptionsParser;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreNumbers.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreNumbers.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreNumbers.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/StoreNumbers.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 import java.util.ArrayList;
@@ -32,15 +32,15 @@ import org.apache.hadoop.io.WritableComp
 import org.apache.hadoop.mapreduce.Job;
 import org.apache.hadoop.mapreduce.Mapper;
 import org.apache.hadoop.util.GenericOptionsParser;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatFieldSchema;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatFieldSchema;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SumNumbers.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SumNumbers.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SumNumbers.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/SumNumbers.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.DataInput;
 import java.io.DataOutput;
@@ -38,10 +38,10 @@ import org.apache.hadoop.mapreduce.Reduc
 import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
 import org.apache.hadoop.mapreduce.lib.output.TextOutputFormat;
 import org.apache.hadoop.util.GenericOptionsParser;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/TypeDataCheck.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/TypeDataCheck.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/TypeDataCheck.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/TypeDataCheck.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 import java.util.Arrays;
@@ -33,11 +33,11 @@ import org.apache.hadoop.mapreduce.lib.o
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat that checks that the columns

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/Util.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/Util.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/Util.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/Util.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 import java.util.HashMap;

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteJson.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteJson.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteJson.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteJson.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 
@@ -29,14 +29,14 @@ import org.apache.hadoop.mapreduce.Mappe
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteRC.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteRC.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteRC.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteRC.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 
@@ -29,14 +29,14 @@ import org.apache.hadoop.mapreduce.Mappe
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteText.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteText.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteText.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteText.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 
@@ -29,14 +29,14 @@ import org.apache.hadoop.mapreduce.Mappe
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat which goes against the "numbers"

Modified: hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteTextPartitioned.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteTextPartitioned.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteTextPartitioned.java (original)
+++ hive/trunk/hcatalog/src/test/e2e/hcatalog/udfs/java/org/apache/hive/hcatalog/utils/WriteTextPartitioned.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.utils;
+package org.apache.hive.hcatalog.utils;
 
 import java.io.IOException;
 import java.util.ArrayList;
@@ -32,15 +32,15 @@ import org.apache.hadoop.mapreduce.Mappe
 import org.apache.hadoop.util.GenericOptionsParser;
 import org.apache.hadoop.util.Tool;
 import org.apache.hadoop.util.ToolRunner;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.data.DefaultHCatRecord;
-import org.apache.hcatalog.data.HCatRecord;
-import org.apache.hcatalog.data.schema.HCatSchema;
-import org.apache.hcatalog.data.schema.HCatFieldSchema;
-import org.apache.hcatalog.mapreduce.HCatInputFormat;
-import org.apache.hcatalog.mapreduce.HCatOutputFormat;
-import org.apache.hcatalog.mapreduce.InputJobInfo;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.data.DefaultHCatRecord;
+import org.apache.hive.hcatalog.data.HCatRecord;
+import org.apache.hive.hcatalog.data.schema.HCatSchema;
+import org.apache.hive.hcatalog.data.schema.HCatFieldSchema;
+import org.apache.hive.hcatalog.mapreduce.HCatInputFormat;
+import org.apache.hive.hcatalog.mapreduce.HCatOutputFormat;
+import org.apache.hive.hcatalog.mapreduce.InputJobInfo;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 /**
  * This is a map reduce test for testing hcat writing to partitioned tables.
@@ -48,7 +48,7 @@ import org.apache.hcatalog.mapreduce.Out
  * other columns. This is to simulate a typical operation in a map reduce
  * program to test that hcat hands the right data to the map reduce program
  *
- * Usage: hadoop jar org.apache.hcatalog.utils.HBaseReadWrite -libjars
+ * Usage: hadoop jar org.apache.hive.hcatalog.utils.HBaseReadWrite -libjars
  * &lt;hcat_jar&gt; * &lt;serveruri&gt; &lt;input_tablename&gt; &lt;output_tablename&gt; [filter]
  * If filter is given it will be provided as the partition to write to.
  */

Modified: hive/trunk/hcatalog/storage-handlers/hbase/if/transaction.thrift
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/storage-handlers/hbase/if/transaction.thrift?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/storage-handlers/hbase/if/transaction.thrift (original)
+++ hive/trunk/hcatalog/storage-handlers/hbase/if/transaction.thrift Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-namespace java org.apache.hcatalog.hbase.snapshot.transaction.thrift
+namespace java org.apache.hive.hcatalog.hbase.snapshot.transaction.thrift
 namespace cpp Apache.HCatalog.HBase
 
 struct StoreFamilyRevision {

Modified: hive/trunk/hcatalog/storage-handlers/hbase/pom.xml
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/storage-handlers/hbase/pom.xml?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/storage-handlers/hbase/pom.xml (original)
+++ hive/trunk/hcatalog/storage-handlers/hbase/pom.xml Thu Sep  5 03:09:42 2013
@@ -22,14 +22,13 @@
          xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
 
   <parent>
-    <groupId>org.apache.hcatalog</groupId>
+    <groupId>org.apache.hive.hcatalog</groupId>
     <artifactId>hcatalog</artifactId>
     <version>0.12.0-SNAPSHOT</version>
     <relativePath>../../pom.xml</relativePath>
   </parent>
 
   <modelVersion>4.0.0</modelVersion>
-  <groupId>org.apache.hcatalog</groupId>
   <artifactId>hbase-storage-handler</artifactId>
   <packaging>jar</packaging>
   <name>hbase-storage-handler</name>
@@ -49,7 +48,7 @@
       </exclusions>
     </dependency>
     <dependency>
-      <groupId>org.apache.hcatalog</groupId>
+      <groupId>org.apache.hive.hcatalog</groupId>
       <artifactId>hcatalog-core</artifactId>
       <version>${hcatalog.version}</version>
       <scope>compile</scope>

Modified: hive/trunk/hcatalog/storage-handlers/hbase/src/gen-java/org/apache/hive/hcatalog/hbase/snapshot/transaction/thrift/StoreFamilyRevision.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/storage-handlers/hbase/src/gen-java/org/apache/hive/hcatalog/hbase/snapshot/transaction/thrift/StoreFamilyRevision.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/storage-handlers/hbase/src/gen-java/org/apache/hive/hcatalog/hbase/snapshot/transaction/thrift/StoreFamilyRevision.java (original)
+++ hive/trunk/hcatalog/storage-handlers/hbase/src/gen-java/org/apache/hive/hcatalog/hbase/snapshot/transaction/thrift/StoreFamilyRevision.java Thu Sep  5 03:09:42 2013
@@ -25,7 +25,7 @@
  *
  * DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
  */
-package org.apache.hcatalog.hbase.snapshot.transaction.thrift;
+package org.apache.hive.hcatalog.hbase.snapshot.transaction.thrift;
 
 import java.util.Map;
 import java.util.HashMap;

Modified: hive/trunk/hcatalog/storage-handlers/hbase/src/gen-java/org/apache/hive/hcatalog/hbase/snapshot/transaction/thrift/StoreFamilyRevisionList.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/storage-handlers/hbase/src/gen-java/org/apache/hive/hcatalog/hbase/snapshot/transaction/thrift/StoreFamilyRevisionList.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/storage-handlers/hbase/src/gen-java/org/apache/hive/hcatalog/hbase/snapshot/transaction/thrift/StoreFamilyRevisionList.java (original)
+++ hive/trunk/hcatalog/storage-handlers/hbase/src/gen-java/org/apache/hive/hcatalog/hbase/snapshot/transaction/thrift/StoreFamilyRevisionList.java Thu Sep  5 03:09:42 2013
@@ -25,7 +25,7 @@
  *
  * DO NOT EDIT UNLESS YOU ARE SURE THAT YOU KNOW WHAT YOU ARE DOING
  */
-package org.apache.hcatalog.hbase.snapshot.transaction.thrift;
+package org.apache.hive.hcatalog.hbase.snapshot.transaction.thrift;
 
 import java.util.ArrayList;
 import java.util.Collections;

Modified: hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseAuthorizationProvider.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseAuthorizationProvider.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseAuthorizationProvider.java (original)
+++ hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseAuthorizationProvider.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.hbase;
+package org.apache.hive.hcatalog.hbase;
 
 import java.util.List;
 

Modified: hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseBaseOutputFormat.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseBaseOutputFormat.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseBaseOutputFormat.java (original)
+++ hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseBaseOutputFormat.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.hbase;
+package org.apache.hive.hcatalog.hbase;
 
 import java.io.IOException;
 import java.util.Properties;
@@ -32,9 +32,9 @@ import org.apache.hadoop.mapred.JobConf;
 import org.apache.hadoop.mapred.OutputFormat;
 import org.apache.hadoop.mapred.RecordWriter;
 import org.apache.hadoop.util.Progressable;
-import org.apache.hcatalog.common.HCatConstants;
-import org.apache.hcatalog.common.HCatUtil;
-import org.apache.hcatalog.mapreduce.OutputJobInfo;
+import org.apache.hive.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.common.HCatUtil;
+import org.apache.hive.hcatalog.mapreduce.OutputJobInfo;
 
 public class HBaseBaseOutputFormat implements OutputFormat<WritableComparable<?>, Put>,
     HiveOutputFormat<WritableComparable<?>, Put> {

Modified: hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseBulkOutputFormat.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseBulkOutputFormat.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseBulkOutputFormat.java (original)
+++ hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseBulkOutputFormat.java Thu Sep  5 03:09:42 2013
@@ -17,7 +17,7 @@
  * under the License.
  */
 
-package org.apache.hcatalog.hbase;
+package org.apache.hive.hcatalog.hbase;
 
 import java.io.IOException;
 import java.util.List;
@@ -42,7 +42,7 @@ import org.apache.hadoop.mapred.Reporter
 import org.apache.hadoop.mapred.SequenceFileOutputFormat;
 import org.apache.hadoop.mapred.TaskAttemptContext;
 import org.apache.hadoop.util.Progressable;
-import org.apache.hcatalog.hbase.snapshot.RevisionManager;
+import org.apache.hive.hcatalog.hbase.snapshot.RevisionManager;
 
 /**
  * Class which imports data into HBase via it's "bulk load" feature. Wherein

Modified: hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseConstants.java
URL: http://svn.apache.org/viewvc/hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseConstants.java?rev=1520202&r1=1520201&r2=1520202&view=diff
==============================================================================
--- hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseConstants.java (original)
+++ hive/trunk/hcatalog/storage-handlers/hbase/src/java/org/apache/hive/hcatalog/hbase/HBaseConstants.java Thu Sep  5 03:09:42 2013
@@ -17,9 +17,9 @@
  * under the License.
  */
 
-package org.apache.hcatalog.hbase;
+package org.apache.hive.hcatalog.hbase;
 
-import org.apache.hcatalog.common.HCatConstants;
+import org.apache.hive.hcatalog.common.HCatConstants;
 
 /**
  * Constants class for constants used in HBase storage handler.