You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by we...@apache.org on 2017/05/08 20:43:30 UTC

[40/51] [partial] hive git commit: Revert "HIVE-14671 : merge master into hive-14535 (Wei Zheng)"

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java
----------------------------------------------------------------------
diff --git a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java
index 50d8878..1cf47c3 100644
--- a/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java
+++ b/itests/hcatalog-unit/src/test/java/org/apache/hive/hcatalog/listener/TestDbNotificationListener.java
@@ -31,16 +31,13 @@ import java.util.Iterator;
 import java.util.LinkedHashMap;
 import java.util.List;
 import java.util.Map;
-import java.util.Stack;
 
 import com.google.common.collect.ImmutableMap;
 import com.google.common.collect.Lists;
-import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hive.cli.CliSessionState;
 import org.apache.hadoop.hive.conf.HiveConf;
 import org.apache.hadoop.hive.metastore.HiveMetaStoreClient;
 import org.apache.hadoop.hive.metastore.IMetaStoreClient;
-import org.apache.hadoop.hive.metastore.MetaStoreEventListener;
 import org.apache.hadoop.hive.metastore.api.Database;
 import org.apache.hadoop.hive.metastore.api.FieldSchema;
 import org.apache.hadoop.hive.metastore.api.FireEventRequest;
@@ -49,7 +46,6 @@ import org.apache.hadoop.hive.metastore.api.Function;
 import org.apache.hadoop.hive.metastore.api.FunctionType;
 import org.apache.hadoop.hive.metastore.api.Index;
 import org.apache.hadoop.hive.metastore.api.InsertEventRequestData;
-import org.apache.hadoop.hive.metastore.api.MetaException;
 import org.apache.hadoop.hive.metastore.api.NotificationEvent;
 import org.apache.hadoop.hive.metastore.api.NotificationEventResponse;
 import org.apache.hadoop.hive.metastore.api.Order;
@@ -60,21 +56,6 @@ import org.apache.hadoop.hive.metastore.api.ResourceUri;
 import org.apache.hadoop.hive.metastore.api.SerDeInfo;
 import org.apache.hadoop.hive.metastore.api.StorageDescriptor;
 import org.apache.hadoop.hive.metastore.api.Table;
-import org.apache.hadoop.hive.metastore.events.AddIndexEvent;
-import org.apache.hadoop.hive.metastore.events.AddPartitionEvent;
-import org.apache.hadoop.hive.metastore.events.AlterIndexEvent;
-import org.apache.hadoop.hive.metastore.events.AlterPartitionEvent;
-import org.apache.hadoop.hive.metastore.events.AlterTableEvent;
-import org.apache.hadoop.hive.metastore.events.CreateDatabaseEvent;
-import org.apache.hadoop.hive.metastore.events.CreateFunctionEvent;
-import org.apache.hadoop.hive.metastore.events.CreateTableEvent;
-import org.apache.hadoop.hive.metastore.events.DropDatabaseEvent;
-import org.apache.hadoop.hive.metastore.events.DropFunctionEvent;
-import org.apache.hadoop.hive.metastore.events.DropIndexEvent;
-import org.apache.hadoop.hive.metastore.events.DropPartitionEvent;
-import org.apache.hadoop.hive.metastore.events.DropTableEvent;
-import org.apache.hadoop.hive.metastore.events.InsertEvent;
-import org.apache.hadoop.hive.metastore.events.ListenerEvent;
 import org.apache.hadoop.hive.metastore.messaging.AddPartitionMessage;
 import org.apache.hadoop.hive.metastore.messaging.AlterIndexMessage;
 import org.apache.hadoop.hive.metastore.messaging.AlterPartitionMessage;
@@ -94,8 +75,6 @@ import org.apache.hadoop.hive.metastore.messaging.MessageDeserializer;
 import org.apache.hadoop.hive.metastore.messaging.MessageFactory;
 import org.apache.hadoop.hive.ql.Driver;
 import org.apache.hadoop.hive.ql.session.SessionState;
-import org.apache.hive.hcatalog.data.Pair;
-import org.junit.After;
 import org.junit.Before;
 import org.junit.BeforeClass;
 import org.junit.Test;
@@ -118,105 +97,12 @@ public class TestDbNotificationListener {
   private int startTime;
   private long firstEventId;
 
-  /* This class is used to verify that HiveMetaStore calls the non-transactional listeners with the
-    * current event ID set by the DbNotificationListener class */
-  public static class MockMetaStoreEventListener extends MetaStoreEventListener {
-    private static Stack<Pair<EventType, String>> eventsIds = new Stack<>();
-
-    private static void pushEventId(EventType eventType, final ListenerEvent event) {
-      if (event.getStatus()) {
-        Map<String, String> parameters = event.getParameters();
-        if (parameters.containsKey(MetaStoreEventListenerConstants.DB_NOTIFICATION_EVENT_ID_KEY_NAME)) {
-          Pair<EventType, String> pair =
-              new Pair<>(eventType, parameters.get(MetaStoreEventListenerConstants.DB_NOTIFICATION_EVENT_ID_KEY_NAME));
-          eventsIds.push(pair);
-        }
-      }
-    }
-
-    public static void popAndVerifyLastEventId(EventType eventType, long id) {
-      if (!eventsIds.isEmpty()) {
-        Pair<EventType, String> pair = eventsIds.pop();
-
-        assertEquals("Last event type does not match.", eventType, pair.first);
-        assertEquals("Last event ID does not match.", Long.toString(id), pair.second);
-      } else {
-        assertTrue("List of events is empty.",false);
-      }
-    }
-
-    public static void clearEvents() {
-      eventsIds.clear();
-    }
-
-    public MockMetaStoreEventListener(Configuration config) {
-      super(config);
-    }
-
-    public void onCreateTable (CreateTableEvent tableEvent) throws MetaException {
-      pushEventId(EventType.CREATE_TABLE, tableEvent);
-    }
-
-    public void onDropTable (DropTableEvent tableEvent)  throws MetaException {
-      pushEventId(EventType.DROP_TABLE, tableEvent);
-    }
-
-    public void onAlterTable (AlterTableEvent tableEvent) throws MetaException {
-      pushEventId(EventType.ALTER_TABLE, tableEvent);
-    }
-
-    public void onAddPartition (AddPartitionEvent partitionEvent) throws MetaException {
-      pushEventId(EventType.ADD_PARTITION, partitionEvent);
-    }
-
-    public void onDropPartition (DropPartitionEvent partitionEvent)  throws MetaException {
-      pushEventId(EventType.DROP_PARTITION, partitionEvent);
-    }
-
-    public void onAlterPartition (AlterPartitionEvent partitionEvent)  throws MetaException {
-      pushEventId(EventType.ALTER_PARTITION, partitionEvent);
-    }
-
-    public void onCreateDatabase (CreateDatabaseEvent dbEvent) throws MetaException {
-      pushEventId(EventType.CREATE_DATABASE, dbEvent);
-    }
-
-    public void onDropDatabase (DropDatabaseEvent dbEvent) throws MetaException {
-      pushEventId(EventType.DROP_DATABASE, dbEvent);
-    }
-
-    public void onAddIndex(AddIndexEvent indexEvent) throws MetaException {
-      pushEventId(EventType.CREATE_INDEX, indexEvent);
-    }
-
-    public void onDropIndex(DropIndexEvent indexEvent) throws MetaException {
-      pushEventId(EventType.DROP_INDEX, indexEvent);
-    }
-
-    public void onAlterIndex(AlterIndexEvent indexEvent) throws MetaException {
-      pushEventId(EventType.ALTER_INDEX, indexEvent);
-    }
-
-    public void onCreateFunction (CreateFunctionEvent fnEvent) throws MetaException {
-      pushEventId(EventType.CREATE_FUNCTION, fnEvent);
-    }
-
-    public void onDropFunction (DropFunctionEvent fnEvent) throws MetaException {
-      pushEventId(EventType.DROP_FUNCTION, fnEvent);
-    }
-
-    public void onInsert(InsertEvent insertEvent) throws MetaException {
-      pushEventId(EventType.INSERT, insertEvent);
-    }
-  }
-
   @SuppressWarnings("rawtypes")
   @BeforeClass
   public static void connectToMetastore() throws Exception {
     HiveConf conf = new HiveConf();
     conf.setVar(HiveConf.ConfVars.METASTORE_TRANSACTIONAL_EVENT_LISTENERS,
         DbNotificationListener.class.getName());
-    conf.setVar(HiveConf.ConfVars.METASTORE_EVENT_LISTENERS, MockMetaStoreEventListener.class.getName());
     conf.setVar(HiveConf.ConfVars.METASTORE_EVENT_DB_LISTENER_TTL, String.valueOf(EVENTS_TTL) + "s");
     conf.setBoolVar(HiveConf.ConfVars.HIVE_SUPPORT_CONCURRENCY, false);
     conf.setBoolVar(HiveConf.ConfVars.FIRE_EVENTS_FOR_DML, true);
@@ -253,12 +139,6 @@ public class TestDbNotificationListener {
     DummyRawStoreFailEvent.setEventSucceed(true);
   }
 
-  @After
-  public void tearDown() {
-    MockMetaStoreEventListener.clearEvents();
-  }
-
-
   @Test
   public void createDatabase() throws Exception {
     String dbName = "createdb";
@@ -284,9 +164,6 @@ public class TestDbNotificationListener {
     CreateDatabaseMessage createDbMsg = md.getCreateDatabaseMessage(event.getMessage());
     assertEquals(dbName, createDbMsg.getDB());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_DATABASE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     DummyRawStoreFailEvent.setEventSucceed(false);
@@ -329,10 +206,6 @@ public class TestDbNotificationListener {
     DropDatabaseMessage dropDbMsg = md.getDropDatabaseMessage(event.getMessage());
     assertEquals(dbName, dropDbMsg.getDB());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.DROP_DATABASE, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_DATABASE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     db = new Database(dbName2, dbDescription, dbLocationUri, emptyParameters);
@@ -383,9 +256,6 @@ public class TestDbNotificationListener {
     assertEquals(tblName, createTblMsg.getTable());
     assertEquals(table, createTblMsg.getTableObj());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     table =
@@ -442,9 +312,6 @@ public class TestDbNotificationListener {
     AlterTableMessage alterTableMessage = md.getAlterTableMessage(event.getMessage());
     assertEquals(table, alterTableMessage.getTableObjAfter());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     DummyRawStoreFailEvent.setEventSucceed(false);
@@ -496,10 +363,6 @@ public class TestDbNotificationListener {
     assertEquals(defaultDbName, dropTblMsg.getDB());
     assertEquals(tblName, dropTblMsg.getTable());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.DROP_TABLE, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     table =
@@ -565,10 +428,6 @@ public class TestDbNotificationListener {
     assertTrue(ptnIter.hasNext());
     assertEquals(partition, ptnIter.next());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ADD_PARTITION, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     partition =
@@ -635,10 +494,6 @@ public class TestDbNotificationListener {
     assertEquals(tblName, alterPtnMsg.getTable());
     assertEquals(newPart, alterPtnMsg.getPtnObjAfter());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ADD_PARTITION, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     DummyRawStoreFailEvent.setEventSucceed(false);
@@ -702,11 +557,6 @@ public class TestDbNotificationListener {
     assertEquals(table.getTableName(), tableObj.getTableName());
     assertEquals(table.getOwner(), tableObj.getOwner());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.DROP_PARTITION, firstEventId + 3);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ADD_PARTITION, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     List<String> newpartCol1Vals = Arrays.asList("tomorrow");
@@ -803,13 +653,6 @@ public class TestDbNotificationListener {
     Iterator<Map<String, String>> parts = dropPtnMsg.getPartitions().iterator();
     assertTrue(parts.hasNext());
     assertEquals(part1.getValues(), Lists.newArrayList(parts.next().values()));
-
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.DROP_PARTITION, firstEventId + 5);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ADD_PARTITION, firstEventId + 4);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ADD_PARTITION, firstEventId + 3);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
   }
 
   @Test
@@ -850,9 +693,6 @@ public class TestDbNotificationListener {
     assertEquals(ResourceType.JAR, funcObj.getResourceUris().get(0).getResourceType());
     assertEquals(funcResource, funcObj.getResourceUris().get(0).getUri());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_FUNCTION, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     DummyRawStoreFailEvent.setEventSucceed(false);
@@ -902,10 +742,6 @@ public class TestDbNotificationListener {
     assertEquals(defaultDbName, dropFuncMsg.getDB());
     assertEquals(funcName, dropFuncMsg.getFunctionName());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.DROP_FUNCTION, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_FUNCTION, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     func =
@@ -971,11 +807,6 @@ public class TestDbNotificationListener {
     assertEquals(tableName, indexObj.getOrigTableName());
     assertEquals(indexTableName, indexObj.getIndexTableName());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_INDEX, firstEventId + 3);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     DummyRawStoreFailEvent.setEventSucceed(false);
@@ -1042,12 +873,6 @@ public class TestDbNotificationListener {
     assertEquals(indexTableName.toLowerCase(), dropIdxMsg.getIndexTableName());
     assertEquals(tableName.toLowerCase(), dropIdxMsg.getOrigTableName());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.DROP_INDEX, firstEventId + 4);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_INDEX, firstEventId + 3);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     index =
@@ -1122,12 +947,6 @@ public class TestDbNotificationListener {
     assertEquals(indexTableName, indexObj.getIndexTableName());
     assertTrue(indexObj.getCreateTime() < indexObj.getLastAccessTime());
 
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ALTER_INDEX, firstEventId + 4);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_INDEX, firstEventId + 3);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
-
     // When hive.metastore.transactional.event.listeners is set,
     // a failed event should not create a new notification
     DummyRawStoreFailEvent.setEventSucceed(false);
@@ -1184,10 +1003,6 @@ public class TestDbNotificationListener {
     assertEquals(tblName, event.getTableName());
     // Parse the message field
     verifyInsert(event, defaultDbName, tblName);
-
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.INSERT, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
   }
 
   @Test
@@ -1248,11 +1063,6 @@ public class TestDbNotificationListener {
     Map<String,String> partKeyValsFromNotif = insertMessage.getPartitionKeyValues();
 
     assertMapEquals(partKeyVals, partKeyValsFromNotif);
-
-    // Verify the eventID was passed to the non-transactional listener
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.INSERT, firstEventId + 3);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.ADD_PARTITION, firstEventId + 2);
-    MockMetaStoreEventListener.popAndVerifyLastEventId(EventType.CREATE_TABLE, firstEventId + 1);
   }
 
 

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/pom.xml
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/pom.xml b/itests/hive-blobstore/pom.xml
index d1c732d..b18398d 100644
--- a/itests/hive-blobstore/pom.xml
+++ b/itests/hive-blobstore/pom.xml
@@ -20,7 +20,7 @@
   <parent>
     <groupId>org.apache.hive</groupId>
     <artifactId>hive-it</artifactId>
-    <version>3.0.0-SNAPSHOT</version>
+    <version>2.2.0-SNAPSHOT</version>
     <relativePath>../pom.xml</relativePath>
   </parent>
 

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_blobstore.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_blobstore.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_blobstore.q
deleted file mode 100644
index 8fee8ed..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_blobstore.q
+++ /dev/null
@@ -1,45 +0,0 @@
--- Check we can create a partitioned table in the warehouse, 
--- export it to a blobstore, and then import its different partitions
--- using the blobstore as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE;
-
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz");
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_blobstore/export/exim_employee;
-EXPORT TABLE exim_employee
-TO '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_blobstore/export/exim_employee';
-
-DROP TABLE exim_employee;
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_blobstore/import/exim_employee;
-IMPORT TABLE exim_employee PARTITION (emp_country='us')
-FROM '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-LOCATION '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_blobstore/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-IMPORT TABLE exim_employee PARTITION (emp_country='cz')
-FROM '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-LOCATION '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_blobstore/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-IMPORT TABLE exim_employee PARTITION (emp_country='in')
-FROM '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-LOCATION '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_blobstore/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_local.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_local.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_local.q
deleted file mode 100644
index 28bc399..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_local.q
+++ /dev/null
@@ -1,44 +0,0 @@
--- Check we can create a partitioned table in the warehouse, 
--- export it to a blobstore, and then import its different partitions
--- using the local filesystem as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz");
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_local/export/exim_employee;
-EXPORT TABLE exim_employee
-TO '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_local/export/exim_employee';
-
-DROP TABLE exim_employee;
-dfs -rm -r -f ${system:build.test.dir}/import_addpartition_blobstore_to_local/import/exim_employee;
-IMPORT TABLE exim_employee PARTITION (emp_country='us')
-FROM '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_local/export/exim_employee'
-LOCATION 'file://${system:build.test.dir}/import_addpartition_blobstore_to_local/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-IMPORT TABLE exim_employee PARTITION (emp_country='cz')
-FROM '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_local/export/exim_employee'
-LOCATION 'file://${system:build.test.dir}/import_addpartition_blobstore_to_local/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-IMPORT TABLE exim_employee PARTITION (emp_country='in')
-FROM '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_local/export/exim_employee'
-LOCATION 'file://${system:build.test.dir}/import_addpartition_blobstore_to_local/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_warehouse.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_warehouse.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_warehouse.q
deleted file mode 100644
index 987dacf..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_blobstore_to_warehouse.q
+++ /dev/null
@@ -1,41 +0,0 @@
--- Check we can create a partitioned table in the warehouse, 
--- export it to a blobstore, and then import its different partitions
--- using the warehouse as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE;
-
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz");
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_warehouse/export/exim_employee;
-EXPORT TABLE exim_employee
-TO '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_warehouse/export/exim_employee';
-
-DROP TABLE exim_employee;
-IMPORT TABLE exim_employee PARTITION (emp_country='us')
-FROM '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_warehouse/export/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-IMPORT TABLE exim_employee PARTITION (emp_country='cz')
-FROM '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_warehouse/export/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-IMPORT TABLE exim_employee PARTITION (emp_country='in')
-FROM '${hiveconf:test.blobstore.path.unique}/import_addpartition_blobstore_to_warehouse/export/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_local_to_blobstore.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_local_to_blobstore.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_local_to_blobstore.q
deleted file mode 100644
index 8fde250..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_addpartition_local_to_blobstore.q
+++ /dev/null
@@ -1,44 +0,0 @@
--- Check we can create a partitioned table in the warehouse, 
--- export it to the local filesystem, and then import its 
--- different partitions using a blobstore as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE;
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz");
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${system:build.test.dir}/import_addpartition_local_to_blobstore/export/exim_employee;
-EXPORT TABLE exim_employee
-TO 'file://${system:build.test.dir}/import_addpartition_local_to_blobstore/export/exim_employee';
-
-DROP TABLE exim_employee;
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_addpartition_local_to_blobstore/import/exim_employee;
-IMPORT TABLE exim_employee PARTITION (emp_country='us')
-FROM 'file://${system:build.test.dir}/import_addpartition_local_to_blobstore/export/exim_employee'
-LOCATION '${hiveconf:test.blobstore.path.unique}/import_addpartition_local_to_blobstore/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-IMPORT TABLE exim_employee PARTITION (emp_country='cz')
-FROM 'file://${system:build.test.dir}/import_addpartition_local_to_blobstore/export/exim_employee'
-LOCATION '${hiveconf:test.blobstore.path.unique}/import_addpartition_local_to_blobstore/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-IMPORT TABLE exim_employee PARTITION (emp_country='in')
-FROM 'file://${system:build.test.dir}/import_addpartition_local_to_blobstore/export/exim_employee'
-LOCATION '${hiveconf:test.blobstore.path.unique}/import_addpartition_local_to_blobstore/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_blobstore.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_blobstore.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_blobstore.q
deleted file mode 100644
index a9f9a8f..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_blobstore.q
+++ /dev/null
@@ -1,30 +0,0 @@
--- Check we can create a partitioned table in the warehouse, 
--- export it to a blobstore, and then import the
--- whole table using the blobstore as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE;
-
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz");
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore/export/exim_employee;
-EXPORT TABLE exim_employee PARTITION (emp_country='us')
-TO '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore/export/exim_employee';
-
-DROP TABLE exim_employee;
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore/import/exim_employee;
-IMPORT FROM '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore/export/exim_employee'
-LOCATION '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_blobstore_nonpart.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_blobstore_nonpart.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_blobstore_nonpart.q
deleted file mode 100644
index 7b3f0a3..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_blobstore_nonpart.q
+++ /dev/null
@@ -1,25 +0,0 @@
--- Check we can create a non partitioned table in the warehouse, 
--- export it to a blobstore, and then import the
--- table using the blobstore as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-STORED AS TEXTFILE;
-
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee;
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore_nonpart/export/exim_employee;
-EXPORT TABLE exim_employee
-TO '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore_nonpart/export/exim_employee';
-
-DROP TABLE exim_employee;
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore_nonpart/import/exim_employee;
-IMPORT FROM '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore_nonpart/export/exim_employee'
-LOCATION '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_blobstore_nonpart/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_local.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_local.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_local.q
deleted file mode 100644
index ac3c451..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_local.q
+++ /dev/null
@@ -1,30 +0,0 @@
--- Check we can create a partitioned table in the warehouse, 
--- export it to a blobstore, and then import the
--- whole table using the local filesystem as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE;
-
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz");
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_blobstore_to_local/export/exim_employee;
-EXPORT TABLE exim_employee PARTITION (emp_country='us')
-TO '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_local/export/exim_employee';
-
-DROP TABLE exim_employee;
-dfs -rm -r -f ${system:build.test.dir}/import_blobstore_to_local/import/exim_employee;
-IMPORT FROM '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_local/export/exim_employee'
-LOCATION "file://${system:build.test.dir}/import_blobstore_to_local/import/exim_employee";
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_warehouse.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_warehouse.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_warehouse.q
deleted file mode 100644
index 9f6fc54..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_warehouse.q
+++ /dev/null
@@ -1,28 +0,0 @@
--- Check we can create a partitioned table in the warehouse, 
--- export it to a blobstore, and then import the
--- whole table using the warehouse as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE;
-
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz");
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_blobstore_to_warehouse/export/exim_employee;
-EXPORT TABLE exim_employee PARTITION (emp_country='us')
-TO '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_warehouse/export/exim_employee';
-
-DROP TABLE exim_employee;
-IMPORT FROM '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_warehouse/export/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_warehouse_nonpart.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_warehouse_nonpart.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_warehouse_nonpart.q
deleted file mode 100644
index 6f28a51..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_blobstore_to_warehouse_nonpart.q
+++ /dev/null
@@ -1,23 +0,0 @@
--- Check we can create a non partitioned table in the warehouse, 
--- export it to a blobstore, and then import the
--- table using the warehouse as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-STORED AS TEXTFILE;
-
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee;
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_blobstore_to_warehouse_nonpart/export/exim_employee;
-EXPORT TABLE exim_employee
-TO '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_warehouse_nonpart/export/exim_employee';
-
-DROP TABLE exim_employee;
-IMPORT FROM '${hiveconf:test.blobstore.path.unique}/import_blobstore_to_warehouse_nonpart/export/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/import_local_to_blobstore.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/import_local_to_blobstore.q b/itests/hive-blobstore/src/test/queries/clientpositive/import_local_to_blobstore.q
deleted file mode 100644
index 0412d3c..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/import_local_to_blobstore.q
+++ /dev/null
@@ -1,31 +0,0 @@
--- Check we can create a partitioned table in the warehouse, 
--- export it to the local filesystem, and then import the
--- whole table using a blobstore as target location
-DROP TABLE exim_employee;
-CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE;
-
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us");
-LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz");
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;
-
-dfs -rm -r -f ${system:build.test.dir}/import_local_to_blobstore/export/exim_employee;
-
-EXPORT TABLE exim_employee PARTITION (emp_country='us')
-TO 'file://${system:build.test.dir}/import_local_to_blobstore/export/exim_employee';
-
-DROP TABLE exim_employee;
-dfs -rm -r -f ${hiveconf:test.blobstore.path.unique}/import_local_to_blobstore/import/exim_employee;
-IMPORT FROM "file://${system:build.test.dir}/import_local_to_blobstore/export/exim_employee"
-LOCATION '${hiveconf:test.blobstore.path.unique}/import_local_to_blobstore/import/exim_employee';
-
-DESCRIBE EXTENDED exim_employee;
-SELECT * FROM exim_employee;

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/insert_blobstore_to_blobstore.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/insert_blobstore_to_blobstore.q b/itests/hive-blobstore/src/test/queries/clientpositive/insert_blobstore_to_blobstore.q
deleted file mode 100644
index 8219ee2..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/insert_blobstore_to_blobstore.q
+++ /dev/null
@@ -1,29 +0,0 @@
--- Test inserting into a blobstore table from another blobstore table.
-
-DROP TABLE blobstore_source;
-CREATE TABLE blobstore_source (
-    a string,
-    b string,
-    c double)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY ' '
-COLLECTION ITEMS TERMINATED BY '\t'
-LINES TERMINATED BY '\n'
-LOCATION '${hiveconf:test.blobstore.path.unique}/insert_blobstore_to_blobstore/blobstore_source';
-
-LOAD DATA LOCAL INPATH '../../data/files/3col_data.txt' INTO TABLE blobstore_source;
-
-DROP TABLE blobstore_table;
-CREATE TABLE blobstore_table LIKE blobstore_source
-LOCATION '${hiveconf:test.blobstore.path.unique}/insert_blobstore_to_blobstore/blobstore_table';
-
-INSERT OVERWRITE TABLE blobstore_table SELECT * FROM blobstore_source;
-
-SELECT COUNT(*) FROM blobstore_table;
-
--- INSERT INTO should append all records to existing ones.
-INSERT INTO TABLE blobstore_table SELECT * FROM blobstore_source;
-
-SELECT COUNT(*) FROM blobstore_table;
-
-SELECT * FROM blobstore_table;

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/insert_empty_into_blobstore.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/insert_empty_into_blobstore.q b/itests/hive-blobstore/src/test/queries/clientpositive/insert_empty_into_blobstore.q
deleted file mode 100644
index d4f0c71..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/insert_empty_into_blobstore.q
+++ /dev/null
@@ -1,53 +0,0 @@
--- Test inserting empty rows into dynamic partitioned and list bucketed blobstore tables
-
-SET hive.blobstore.optimizations.enabled=true;
-
-DROP TABLE empty;
-DROP TABLE blobstore_dynamic_partitioning;
-DROP TABLE blobstore_list_bucketing;
-
-CREATE TABLE empty (
-    id int,
-    name string,
-    dept string,
-    pt string,
-    dt string,
-    hr string);
-
-CREATE TABLE blobstore_dynamic_partitioning (
-    id int,
-    name string,
-    dept string)
-PARTITIONED BY (
-    pt string,
-    dt string,
-    hr string)
-LOCATION '${hiveconf:test.blobstore.path.unique}/insert_empty_into_blobstore/blobstore_dynamic_partitioning';
-
-INSERT INTO TABLE blobstore_dynamic_partitioning PARTITION (pt='a', dt, hr) SELECT id, name, dept, dt, hr FROM empty;
-
-SELECT COUNT(*) FROM blobstore_dynamic_partitioning;
-
-CREATE TABLE blobstore_list_bucketing (
-    id int,
-    name string,
-    dept string)
-PARTITIONED BY (
-    pt string,
-    dt string,
-    hr string)
-SKEWED BY (id) ON ('1', '2', '3') STORED AS DIRECTORIES
-LOCATION '${hiveconf:test.blobstore.path.unique}/insert_empty_into_blobstore/blobstore_list_bucketing';
-
-INSERT INTO TABLE blobstore_list_bucketing PARTITION (pt='a', dt='a', hr='a') SELECT id, name, dept FROM empty;
-
-SELECT COUNT(*) FROM blobstore_list_bucketing;
-
--- Now test empty inserts with blobstore optimizations turned off. This should give us same results.
-SET hive.blobstore.optimizations.enabled=false;
-
-INSERT INTO TABLE blobstore_dynamic_partitioning PARTITION (pt='b', dt, hr) SELECT id, name, dept, dt, hr FROM empty;
-SELECT COUNT(*) FROM blobstore_dynamic_partitioning;
-
-INSERT INTO TABLE blobstore_list_bucketing PARTITION (pt='b', dt='b', hr='b') SELECT id, name, dept FROM empty;
-SELECT COUNT(*) FROM blobstore_list_bucketing;

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/orc_buckets.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/orc_buckets.q b/itests/hive-blobstore/src/test/queries/clientpositive/orc_buckets.q
deleted file mode 100644
index 9571842..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/orc_buckets.q
+++ /dev/null
@@ -1,31 +0,0 @@
--- Test simple interaction with partitioned bucketed table with orc format in blobstore
-
-SET hive.exec.dynamic.partition=true;
-SET hive.exec.reducers.max=10;
-SET hive.exec.dynamic.partition.mode=nonstrict;
-
-DROP TABLE blobstore_source;
-CREATE TABLE blobstore_source(a STRING, b STRING, c DOUBLE)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY ' '
-COLLECTION ITEMS TERMINATED BY '\t'
-LINES TERMINATED BY '\n'
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_buckets/blobstore_source/';
-LOAD DATA LOCAL INPATH '../../data/files/3col_data.txt' INTO TABLE blobstore_source;
-
-DROP TABLE orc_buckets;
-CREATE TABLE orc_buckets (a STRING, value DOUBLE)
-PARTITIONED BY (b STRING)
-CLUSTERED BY (a) INTO 10 BUCKETS
-STORED AS ORC
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_buckets/orc_buckets';
-
-INSERT OVERWRITE TABLE orc_buckets
-PARTITION (b)
-SELECT a, c, b FROM blobstore_source;
-SELECT * FROM orc_buckets;
-
-INSERT INTO TABLE orc_buckets
-PARTITION (b)
-SELECT a, c, b FROM blobstore_source;
-SELECT * FROM orc_buckets;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/orc_format_nonpart.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/orc_format_nonpart.q b/itests/hive-blobstore/src/test/queries/clientpositive/orc_format_nonpart.q
deleted file mode 100644
index ad95459..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/orc_format_nonpart.q
+++ /dev/null
@@ -1,30 +0,0 @@
--- Test INSERT OVERWRITE and INSERT INTO on orc table in blobstore
-
-DROP TABLE blobstore_source;
-CREATE TABLE blobstore_source(a STRING, b STRING, c DOUBLE)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY ' '
-COLLECTION ITEMS TERMINATED BY '\t'
-LINES TERMINATED BY '\n'
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_format_nonpart/blobstore_source/';
-LOAD DATA LOCAL INPATH '../../data/files/3col_data.txt' INTO TABLE blobstore_source;
-
-DROP TABLE orc_table;
-CREATE EXTERNAL TABLE orc_table (a INT, b STRING, value DOUBLE) STORED AS ORC
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_format_nonpart/orc_table';
- 
-INSERT OVERWRITE TABLE orc_table
-SELECT * FROM blobstore_source;
- 
-SELECT * FROM orc_table;
-SELECT a FROM orc_table GROUP BY a;
-SELECT b FROM orc_table GROUP BY b;
-SELECT value FROM orc_table GROUP BY value;
-
-INSERT INTO TABLE orc_table
-SELECT * FROM blobstore_source;
-
-SELECT * FROM orc_table;
-SELECT a FROM orc_table GROUP BY a;
-SELECT b FROM orc_table GROUP BY b;
-SELECT value FROM orc_table GROUP BY value;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/orc_format_part.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/orc_format_part.q b/itests/hive-blobstore/src/test/queries/clientpositive/orc_format_part.q
deleted file mode 100644
index 358eccd..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/orc_format_part.q
+++ /dev/null
@@ -1,67 +0,0 @@
--- Test INSERT INTO and INSERT OVERWRITE on partitioned orc table in blobstore
-
-DROP TABLE src_events;
-CREATE TABLE src_events
-(
-  log_id      BIGINT,
-  time        BIGINT,
-  uid         BIGINT,
-  user_id     BIGINT,
-  type        INT,
-  event_data  STRING,
-  session_id  STRING,
-  full_uid    BIGINT,
-  run_date    STRING,
-  game_id     INT,
-  event_name  STRING
-)
-ROW FORMAT DELIMITED FIELDS TERMINATED BY ','
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_format_part/src_events/';
-LOAD DATA LOCAL INPATH '../../data/files/events.txt' INTO TABLE src_events;
-
-DROP TABLE orc_events;
-CREATE TABLE orc_events
-(
-  log_id      BIGINT,
-  time        BIGINT,
-  uid         BIGINT,
-  user_id     BIGINT,
-  type        INT,
-  event_data  STRING,
-  session_id  STRING,
-  full_uid    BIGINT
-)
-PARTITIONED BY (run_date STRING, game_id INT, event_name STRING)
-STORED AS ORC
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_format_part/orc_events';
-
-SET hive.exec.dynamic.partition=true;
-SET hive.exec.dynamic.partition.mode=nonstrict;
-
-INSERT OVERWRITE TABLE orc_events PARTITION (run_date, game_id, event_name)
-SELECT * FROM src_events;
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
-SELECT COUNT(*) FROM orc_events WHERE run_date=20120921;
-SELECT COUNT(*) FROM orc_events WHERE run_date=20121121;
-
-INSERT OVERWRITE TABLE orc_events PARTITION (run_date=201211, game_id, event_name)
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid,game_id,event_name FROM src_events
-WHERE SUBSTR(run_date,1,6)='201211';
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
-
-INSERT INTO TABLE orc_events PARTITION (run_date=201209, game_id=39, event_name)
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid,event_name FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209' AND game_id=39;
-SELECT COUNT(*) FROM orc_events;
-
-INSERT INTO TABLE orc_events PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209' AND game_id=39 AND event_name='hq_change';
-SELECT COUNT(*) FROM orc_events;
-
-INSERT OVERWRITE TABLE orc_events PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209' AND game_id=39 AND event_name='hq_change';
-SELECT COUNT(*) FROM orc_events;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/orc_nonstd_partitions_loc.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/orc_nonstd_partitions_loc.q b/itests/hive-blobstore/src/test/queries/clientpositive/orc_nonstd_partitions_loc.q
deleted file mode 100644
index c462538..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/orc_nonstd_partitions_loc.q
+++ /dev/null
@@ -1,100 +0,0 @@
--- Test table in orc format with non-standard partition locations in blobstore
-
-DROP TABLE src_events;
-CREATE TABLE src_events
-(
-  log_id      BIGINT,
-  time        BIGINT,
-  uid         BIGINT,
-  user_id     BIGINT,
-  type        INT,
-  event_data  STRING,
-  session_id  STRING,
-  full_uid    BIGINT,
-  run_date    STRING,
-  game_id     INT,
-  event_name  STRING
-)
-ROW FORMAT DELIMITED FIELDS TERMINATED BY ','
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_nonstd_partitions_loc/src_events/';
-LOAD DATA LOCAL INPATH '../../data/files/events.txt' INTO TABLE src_events;
-
-DROP TABLE orc_events;
-CREATE TABLE orc_events
-(
-  log_id      BIGINT,
-  time        BIGINT,
-  uid         BIGINT,
-  user_id     BIGINT,
-  type        INT,
-  event_data  STRING,
-  session_id  STRING,
-  full_uid    BIGINT
-)
-PARTITIONED BY (run_date STRING, game_id INT, event_name STRING)
-STORED AS ORC
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_nonstd_partitions_loc/orc_events/';
-
-SET hive.exec.dynamic.partition=true;
-SET hive.exec.dynamic.partition.mode=nonstrict;
-
-INSERT OVERWRITE TABLE orc_events PARTITION (run_date, game_id, event_name)
-SELECT * FROM src_events;
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
-
--- verify INSERT OVERWRITE and INSERT INTO nonstandard partition location
-ALTER TABLE orc_events ADD PARTITION (run_date=201211, game_id=39, event_name='hq_change')
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_nonstd_partitions_loc/orc_nonstd_loc/ns-part-1/';
-INSERT OVERWRITE TABLE orc_events PARTITION (run_date=201211, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201211';
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
-INSERT INTO TABLE orc_events PARTITION (run_date=201211, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201211';
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
-
-SET hive.merge.mapfiles=false;
-
--- verify INSERT OVERWRITE and INSERT INTO nonstandard partition location with hive.merge.mapfiles false
-ALTER TABLE orc_events ADD PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_nonstd_partitions_loc/orc_nonstd_loc/ns-part-2/';
-INSERT OVERWRITE TABLE orc_events PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209';
-INSERT INTO TABLE orc_events PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209';
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
-
--- verify dynamic INSERT OVERWRITE over all partitions (standard and nonstandard locations) with hive.merge.mapfiles false
-INSERT OVERWRITE TABLE orc_events PARTITION (run_date, game_id, event_name)
-SELECT * FROM src_events;
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
-
-SET hive.merge.mapfiles=true;
-
-ALTER TABLE orc_events ADD PARTITION (run_date=201207, game_id=39, event_name='hq_change')
-LOCATION '${hiveconf:test.blobstore.path.unique}/orc_nonstd_partitions_loc/orc_nonstd_loc/ns-part-3/';
-INSERT INTO TABLE orc_events PARTITION (run_date=201207, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209';
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
-
--- verify dynamic INSERT OVERWRITE over all partitions (standard and nonstandard locations) with hive.merge.mapfiles true
-INSERT OVERWRITE TABLE orc_events PARTITION (run_date, game_id, event_name)
-SELECT * FROM src_events;
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
-
-ALTER TABLE orc_events DROP PARTITION (run_date=201211, game_id=39, event_name='hq_change');
-ALTER TABLE orc_events DROP PARTITION (run_date=201209, game_id=39, event_name='hq_change');
-ALTER TABLE orc_events DROP PARTITION (run_date=201207, game_id=39, event_name='hq_change');
-SHOW PARTITIONS orc_events;
-SELECT COUNT(*) FROM orc_events;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_buckets.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_buckets.q b/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_buckets.q
deleted file mode 100644
index 606ef720..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_buckets.q
+++ /dev/null
@@ -1,31 +0,0 @@
--- Test simple interaction with partitioned bucketed table with rcfile format in blobstore
-
-SET hive.exec.dynamic.partition=true;
-SET hive.exec.reducers.max=10;
-SET hive.exec.dynamic.partition.mode=nonstrict;
-
-DROP TABLE blobstore_source;
-CREATE TABLE blobstore_source(a STRING, b STRING, c DOUBLE)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY ' '
-COLLECTION ITEMS TERMINATED BY '\t'
-LINES TERMINATED BY '\n'
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_buckets/blobstore_source/';
-LOAD DATA LOCAL INPATH '../../data/files/3col_data.txt' INTO TABLE blobstore_source;
-
-DROP TABLE rcfile_buckets;
-CREATE TABLE rcfile_buckets (a STRING, value DOUBLE)
-PARTITIONED BY (b STRING)
-CLUSTERED BY (a) INTO 10 BUCKETS
-STORED AS RCFILE
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_buckets/rcfile_buckets';
-
-INSERT OVERWRITE TABLE rcfile_buckets
-PARTITION (b)
-SELECT a, c, b FROM blobstore_source;
-SELECT * FROM rcfile_buckets;
-
-INSERT INTO TABLE rcfile_buckets
-PARTITION (b)
-SELECT a, c, b FROM blobstore_source;
-SELECT * FROM rcfile_buckets;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_nonpart.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_nonpart.q b/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_nonpart.q
deleted file mode 100644
index 9cd909e..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_nonpart.q
+++ /dev/null
@@ -1,30 +0,0 @@
--- Test INSERT OVERWRITE and INSERT INTO on rcfile table in blobstore
-
-DROP TABLE blobstore_source;
-CREATE TABLE blobstore_source(a STRING, b STRING, c DOUBLE)
-ROW FORMAT DELIMITED
-FIELDS TERMINATED BY ' '
-COLLECTION ITEMS TERMINATED BY '\t'
-LINES TERMINATED BY '\n'
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_format_nonpart/blobstore_source/';
-LOAD DATA LOCAL INPATH '../../data/files/3col_data.txt' INTO TABLE blobstore_source;
-
-DROP TABLE rcfile_table;
-CREATE TABLE rcfile_table (a INT, b STRING, value DOUBLE) STORED AS RCFILE
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_format_nonpart/rcfile_table';
- 
-INSERT OVERWRITE TABLE rcfile_table
-SELECT * FROM blobstore_source;
- 
-SELECT * FROM rcfile_table;
-SELECT a FROM rcfile_table GROUP BY a;
-SELECT b FROM rcfile_table GROUP BY b;
-SELECT VALUE FROM rcfile_table GROUP BY VALUE;
-
-INSERT INTO TABLE rcfile_table
-SELECT * FROM blobstore_source;
-
-SELECT * FROM rcfile_table;
-SELECT a FROM rcfile_table GROUP BY a;
-SELECT b FROM rcfile_table GROUP BY b;
-SELECT value FROM rcfile_table GROUP BY value;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_part.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_part.q b/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_part.q
deleted file mode 100644
index c563d3a..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_format_part.q
+++ /dev/null
@@ -1,67 +0,0 @@
--- Test INSERT INTO and INSERT OVERWRITE on partitioned rcfile table in blobstore
-
-DROP TABLE src_events;
-CREATE TABLE src_events
-(
-  log_id      BIGINT,
-  time        BIGINT,
-  uid         BIGINT,
-  user_id     BIGINT,
-  type        INT,
-  event_data  STRING,
-  session_id  STRING,
-  full_uid    BIGINT,
-  run_date    STRING,
-  game_id     INT,
-  event_name  STRING
-)
-ROW FORMAT DELIMITED FIELDS TERMINATED BY ','
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_format_part/src_events/';
-LOAD DATA LOCAL INPATH '../../data/files/events.txt' INTO TABLE src_events;
-
-DROP TABLE rcfile_events;
-CREATE TABLE rcfile_events
-(
-  log_id      BIGINT,
-  time        BIGINT,
-  uid         BIGINT,
-  user_id     BIGINT,
-  type        INT,
-  event_data  STRING,
-  session_id  STRING,
-  full_uid    BIGINT
-)
-PARTITIONED BY (run_date STRING, game_id INT, event_name STRING)
-STORED AS RCFILE
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_format_part/rcfile_events';
-
-SET hive.exec.dynamic.partition=true;
-SET hive.exec.dynamic.partition.mode=nonstrict;
-
-INSERT OVERWRITE TABLE rcfile_events PARTITION (run_date, game_id, event_name)
-SELECT * FROM src_events;
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;
-SELECT COUNT(*) FROM rcfile_events WHERE run_date=20120921;
-SELECT COUNT(*) FROM rcfile_events WHERE run_date=20121121;
-
-INSERT OVERWRITE TABLE rcfile_events PARTITION (run_date=201211, game_id, event_name)
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid,game_id,event_name FROM src_events
-WHERE SUBSTR(run_date,1,6)='201211';
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;
-
-INSERT INTO TABLE rcfile_events PARTITION (run_date=201209, game_id=39, event_name)
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid,event_name FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209' AND game_id=39;
-SELECT COUNT(*) FROM rcfile_events;
-
-INSERT INTO TABLE rcfile_events PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209' AND game_id=39 AND event_name='hq_change';
-SELECT COUNT(*) FROM rcfile_events;
-
-INSERT OVERWRITE TABLE rcfile_events PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209' AND game_id=39 AND event_name='hq_change';
-SELECT COUNT(*) FROM rcfile_events;
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_nonstd_partitions_loc.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_nonstd_partitions_loc.q b/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_nonstd_partitions_loc.q
deleted file mode 100644
index d17c281..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/rcfile_nonstd_partitions_loc.q
+++ /dev/null
@@ -1,100 +0,0 @@
--- Test table in rcfile format with non-standard partition locations in blobstore
-
-DROP TABLE src_events;
-CREATE TABLE src_events
-(
-  log_id      BIGINT,
-  time        BIGINT,
-  uid         BIGINT,
-  user_id     BIGINT,
-  type        INT,
-  event_data  STRING,
-  session_id  STRING,
-  full_uid    BIGINT,
-  run_date    STRING,
-  game_id     INT,
-  event_name  STRING
-)
-ROW FORMAT DELIMITED FIELDS TERMINATED BY ','
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_nonstd_partitions_loc/src_events/';
-LOAD DATA LOCAL INPATH '../../data/files/events.txt' INTO TABLE src_events;
-
-DROP TABLE rcfile_events;
-CREATE TABLE rcfile_events
-(
-  log_id      BIGINT,
-  time        BIGINT,
-  uid         BIGINT,
-  user_id     BIGINT,
-  type        INT,
-  event_data  STRING,
-  session_id  STRING,
-  full_uid    BIGINT
-)
-PARTITIONED BY (run_date STRING, game_id INT, event_name STRING)
-STORED AS RCFILE
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_nonstd_partitions_loc/rcfile_events/';
-
-SET hive.exec.dynamic.partition=true;
-SET hive.exec.dynamic.partition.mode=nonstrict;
-
-INSERT OVERWRITE TABLE rcfile_events PARTITION (run_date, game_id, event_name)
-SELECT * FROM src_events;
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;
-
--- verify INSERT OVERWRITE and INSERT INTO nonstandard partition location
-ALTER TABLE rcfile_events ADD PARTITION (run_date=201211, game_id=39, event_name='hq_change')
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_nonstd_partitions_loc/rcfile_nonstd_loc/ns-part-1/';
-INSERT OVERWRITE TABLE rcfile_events PARTITION (run_date=201211, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201211';
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;
-INSERT INTO TABLE rcfile_events PARTITION (run_date=201211, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201211';
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;
-
-SET hive.merge.mapfiles=false;
-
--- verify INSERT OVERWRITE and INSERT INTO nonstandard partition location with hive.merge.mapfiles false
-ALTER TABLE rcfile_events ADD PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_nonstd_partitions_loc/rcfile_nonstd_loc/ns-part-2/';
-INSERT INTO TABLE rcfile_events PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209';
-INSERT INTO TABLE rcfile_events PARTITION (run_date=201209, game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209';
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;
-
--- verify dynamic INSERT OVERWRITE over all partitions (standard and nonstandard locations) with hive.merge.mapfiles false
-INSERT OVERWRITE TABLE rcfile_events PARTITION (run_date, game_id, event_name)
-SELECT * FROM src_events;
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;
-
-SET hive.merge.mapfiles=true;
-
-ALTER TABLE rcfile_events ADD PARTITION (run_date=201207, game_id=39, event_name='hq_change')
-LOCATION '${hiveconf:test.blobstore.path.unique}/rcfile_nonstd_partitions_loc/rcfile_nonstd_loc/ns-part-3/';
-INSERT INTO TABLE rcfile_events PARTITION(run_date=201207,game_id=39, event_name='hq_change')
-SELECT log_id,time,uid,user_id,type,event_data,session_id,full_uid FROM src_events
-WHERE SUBSTR(run_date,1,6)='201209';
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;
-
--- verify dynamic INSERT OVERWRITE over all partitions (standard and nonstandard locations) with hive.merge.mapfiles true
-INSERT OVERWRITE TABLE rcfile_events PARTITION (run_date, game_id, event_name)
-SELECT * FROM src_events;
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;
-
-ALTER TABLE rcfile_events DROP PARTITION (run_date=201211,game_id=39, event_name='hq_change');
-ALTER TABLE rcfile_events DROP PARTITION (run_date=201209,game_id=39, event_name='hq_change');
-ALTER TABLE rcfile_events DROP PARTITION (run_date=201207,game_id=39, event_name='hq_change');
-SHOW PARTITIONS rcfile_events;
-SELECT COUNT(*) FROM rcfile_events;

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/zero_rows_blobstore.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/zero_rows_blobstore.q b/itests/hive-blobstore/src/test/queries/clientpositive/zero_rows_blobstore.q
deleted file mode 100644
index 1f663ef..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/zero_rows_blobstore.q
+++ /dev/null
@@ -1,19 +0,0 @@
--- Insert overwrite into blobstore when WHERE clause returns zero rows
-DROP TABLE blobstore_source;
-CREATE TABLE blobstore_source (
-    key int
-) 
-LOCATION '${hiveconf:test.blobstore.path.unique}/zero_rows_blobstore/blobstore_source/';
-LOAD DATA LOCAL INPATH '../../data/files/kv6.txt' INTO TABLE blobstore_source;
-
-DROP TABLE blobstore_target;
-CREATE TABLE blobstore_target (
-    key int
-) 
-LOCATION '${hiveconf:test.blobstore.path.unique}/zero_rows_blobstore/blobstore_target';
-
-SELECT COUNT(*) FROM blobstore_target;
-INSERT OVERWRITE TABLE blobstore_target SELECT key FROM blobstore_source;
-SELECT COUNT(*) FROM blobstore_target;
-INSERT OVERWRITE TABLE blobstore_target SELECT key FROM blobstore_source WHERE FALSE;
-SELECT COUNT(*) FROM blobstore_target;

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/queries/clientpositive/zero_rows_hdfs.q
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/queries/clientpositive/zero_rows_hdfs.q b/itests/hive-blobstore/src/test/queries/clientpositive/zero_rows_hdfs.q
deleted file mode 100644
index ef3b71d..0000000
--- a/itests/hive-blobstore/src/test/queries/clientpositive/zero_rows_hdfs.q
+++ /dev/null
@@ -1,18 +0,0 @@
--- Insert overwrite into hdfs from blobstore when WHERE clause returns zero rows
-DROP TABLE blobstore_source;
-CREATE TABLE blobstore_source (
-    key int
-)
-LOCATION '${hiveconf:test.blobstore.path.unique}/zero_rows_hdfs/blobstore_source/';
-LOAD DATA LOCAL INPATH '../../data/files/kv6.txt' INTO TABLE blobstore_source;
-
-DROP TABLE hdfs_target;
-CREATE TABLE hdfs_target (
-    key int
-);
-
-SELECT COUNT(*) FROM hdfs_target;
-INSERT OVERWRITE TABLE hdfs_target SELECT key FROM blobstore_source;
-SELECT COUNT(*) FROM hdfs_target;
-INSERT OVERWRITE TABLE hdfs_target SELECT key FROM blobstore_source WHERE FALSE;
-SELECT COUNT(*) FROM hdfs_target;

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/resources/hive-site.xml
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/resources/hive-site.xml b/itests/hive-blobstore/src/test/resources/hive-site.xml
index 038db0d..406b3b8 100644
--- a/itests/hive-blobstore/src/test/resources/hive-site.xml
+++ b/itests/hive-blobstore/src/test/resources/hive-site.xml
@@ -279,11 +279,6 @@
     <value>pblob:${system:test.tmp.dir}/bucket</value>
   </property>
 
-  <property>
-    <name>hive.exim.uri.scheme.whitelist</name>
-    <value>hdfs,pfile,file,s3,s3a,pblob</value>
-  </property>
-
   <!--
   To run these tests:
   # Create a file blobstore-conf.xml  - DO NOT ADD TO REVISION CONTROL

http://git-wip-us.apache.org/repos/asf/hive/blob/ed64a74e/itests/hive-blobstore/src/test/results/clientpositive/import_addpartition_blobstore_to_blobstore.q.out
----------------------------------------------------------------------
diff --git a/itests/hive-blobstore/src/test/results/clientpositive/import_addpartition_blobstore_to_blobstore.q.out b/itests/hive-blobstore/src/test/results/clientpositive/import_addpartition_blobstore_to_blobstore.q.out
deleted file mode 100644
index c1e57ee..0000000
--- a/itests/hive-blobstore/src/test/results/clientpositive/import_addpartition_blobstore_to_blobstore.q.out
+++ /dev/null
@@ -1,283 +0,0 @@
-PREHOOK: query: DROP TABLE exim_employee
-PREHOOK: type: DROPTABLE
-POSTHOOK: query: DROP TABLE exim_employee
-POSTHOOK: type: DROPTABLE
-PREHOOK: query: CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE
-PREHOOK: type: CREATETABLE
-PREHOOK: Output: database:default
-PREHOOK: Output: default@exim_employee
-POSTHOOK: query: CREATE TABLE exim_employee (emp_id int COMMENT "employee id")
-COMMENT "employee table"
-PARTITIONED BY (emp_country string COMMENT "two char iso code")
-STORED AS TEXTFILE
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@exim_employee
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in")
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@exim_employee
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="in")
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@exim_employee
-POSTHOOK: Output: default@exim_employee@emp_country=in
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us")
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@exim_employee
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="us")
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@exim_employee
-POSTHOOK: Output: default@exim_employee@emp_country=us
-PREHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz")
-PREHOOK: type: LOAD
-#### A masked pattern was here ####
-PREHOOK: Output: default@exim_employee
-POSTHOOK: query: LOAD DATA LOCAL INPATH "../../data/files/test.dat"
-INTO TABLE exim_employee PARTITION (emp_country="cz")
-POSTHOOK: type: LOAD
-#### A masked pattern was here ####
-POSTHOOK: Output: default@exim_employee
-POSTHOOK: Output: default@exim_employee@emp_country=cz
-PREHOOK: query: DESCRIBE EXTENDED exim_employee
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@exim_employee
-POSTHOOK: query: DESCRIBE EXTENDED exim_employee
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@exim_employee
-emp_id              	int                 	employee id         
-emp_country         	string              	two char iso code   
-	 	 
-# Partition Information	 	 
-# col_name            	data_type           	comment             
-	 	 
-emp_country         	string              	two char iso code   
-	 	 
-#### A masked pattern was here ####
-PREHOOK: query: SELECT * FROM exim_employee
-PREHOOK: type: QUERY
-PREHOOK: Input: default@exim_employee
-PREHOOK: Input: default@exim_employee@emp_country=cz
-PREHOOK: Input: default@exim_employee@emp_country=in
-PREHOOK: Input: default@exim_employee@emp_country=us
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM exim_employee
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@exim_employee
-POSTHOOK: Input: default@exim_employee@emp_country=cz
-POSTHOOK: Input: default@exim_employee@emp_country=in
-POSTHOOK: Input: default@exim_employee@emp_country=us
-#### A masked pattern was here ####
-1	cz
-2	cz
-3	cz
-4	cz
-5	cz
-6	cz
-1	in
-2	in
-3	in
-4	in
-5	in
-6	in
-1	us
-2	us
-3	us
-4	us
-5	us
-6	us
-PREHOOK: query: EXPORT TABLE exim_employee
-TO '### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-PREHOOK: type: EXPORT
-PREHOOK: Input: default@exim_employee@emp_country=cz
-PREHOOK: Input: default@exim_employee@emp_country=in
-PREHOOK: Input: default@exim_employee@emp_country=us
-PREHOOK: Output: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee
-POSTHOOK: query: EXPORT TABLE exim_employee
-TO '### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-POSTHOOK: type: EXPORT
-POSTHOOK: Input: default@exim_employee@emp_country=cz
-POSTHOOK: Input: default@exim_employee@emp_country=in
-POSTHOOK: Input: default@exim_employee@emp_country=us
-POSTHOOK: Output: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee
-PREHOOK: query: DROP TABLE exim_employee
-PREHOOK: type: DROPTABLE
-PREHOOK: Input: default@exim_employee
-PREHOOK: Output: default@exim_employee
-POSTHOOK: query: DROP TABLE exim_employee
-POSTHOOK: type: DROPTABLE
-POSTHOOK: Input: default@exim_employee
-POSTHOOK: Output: default@exim_employee
-PREHOOK: query: IMPORT TABLE exim_employee PARTITION (emp_country='us')
-FROM '### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-#### A masked pattern was here ####
-PREHOOK: type: IMPORT
-PREHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee
-PREHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/import/exim_employee
-PREHOOK: Output: database:default
-POSTHOOK: query: IMPORT TABLE exim_employee PARTITION (emp_country='us')
-FROM '### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-#### A masked pattern was here ####
-POSTHOOK: type: IMPORT
-POSTHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee
-POSTHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/import/exim_employee
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@exim_employee
-POSTHOOK: Output: default@exim_employee@emp_country=us
-PREHOOK: query: DESCRIBE EXTENDED exim_employee
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@exim_employee
-POSTHOOK: query: DESCRIBE EXTENDED exim_employee
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@exim_employee
-emp_id              	int                 	employee id         
-emp_country         	string              	two char iso code   
-	 	 
-# Partition Information	 	 
-# col_name            	data_type           	comment             
-	 	 
-emp_country         	string              	two char iso code   
-	 	 
-#### A masked pattern was here ####
-PREHOOK: query: SELECT * FROM exim_employee
-PREHOOK: type: QUERY
-PREHOOK: Input: default@exim_employee
-PREHOOK: Input: default@exim_employee@emp_country=us
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM exim_employee
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@exim_employee
-POSTHOOK: Input: default@exim_employee@emp_country=us
-#### A masked pattern was here ####
-1	us
-2	us
-3	us
-4	us
-5	us
-6	us
-PREHOOK: query: IMPORT TABLE exim_employee PARTITION (emp_country='cz')
-FROM '### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-#### A masked pattern was here ####
-PREHOOK: type: IMPORT
-PREHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee
-PREHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/import/exim_employee
-PREHOOK: Output: default@exim_employee
-POSTHOOK: query: IMPORT TABLE exim_employee PARTITION (emp_country='cz')
-FROM '### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-#### A masked pattern was here ####
-POSTHOOK: type: IMPORT
-POSTHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee
-POSTHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/import/exim_employee
-POSTHOOK: Output: default@exim_employee
-POSTHOOK: Output: default@exim_employee@emp_country=cz
-PREHOOK: query: DESCRIBE EXTENDED exim_employee
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@exim_employee
-POSTHOOK: query: DESCRIBE EXTENDED exim_employee
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@exim_employee
-emp_id              	int                 	employee id         
-emp_country         	string              	two char iso code   
-	 	 
-# Partition Information	 	 
-# col_name            	data_type           	comment             
-	 	 
-emp_country         	string              	two char iso code   
-	 	 
-#### A masked pattern was here ####
-PREHOOK: query: SELECT * FROM exim_employee
-PREHOOK: type: QUERY
-PREHOOK: Input: default@exim_employee
-PREHOOK: Input: default@exim_employee@emp_country=cz
-PREHOOK: Input: default@exim_employee@emp_country=us
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM exim_employee
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@exim_employee
-POSTHOOK: Input: default@exim_employee@emp_country=cz
-POSTHOOK: Input: default@exim_employee@emp_country=us
-#### A masked pattern was here ####
-1	cz
-2	cz
-3	cz
-4	cz
-5	cz
-6	cz
-1	us
-2	us
-3	us
-4	us
-5	us
-6	us
-PREHOOK: query: IMPORT TABLE exim_employee PARTITION (emp_country='in')
-FROM '### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-#### A masked pattern was here ####
-PREHOOK: type: IMPORT
-PREHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee
-PREHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/import/exim_employee
-PREHOOK: Output: default@exim_employee
-POSTHOOK: query: IMPORT TABLE exim_employee PARTITION (emp_country='in')
-FROM '### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee'
-#### A masked pattern was here ####
-POSTHOOK: type: IMPORT
-POSTHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/export/exim_employee
-POSTHOOK: Input: ### test.blobstore.path ###/import_addpartition_blobstore_to_blobstore/import/exim_employee
-POSTHOOK: Output: default@exim_employee
-POSTHOOK: Output: default@exim_employee@emp_country=in
-PREHOOK: query: DESCRIBE EXTENDED exim_employee
-PREHOOK: type: DESCTABLE
-PREHOOK: Input: default@exim_employee
-POSTHOOK: query: DESCRIBE EXTENDED exim_employee
-POSTHOOK: type: DESCTABLE
-POSTHOOK: Input: default@exim_employee
-emp_id              	int                 	employee id         
-emp_country         	string              	two char iso code   
-	 	 
-# Partition Information	 	 
-# col_name            	data_type           	comment             
-	 	 
-emp_country         	string              	two char iso code   
-	 	 
-#### A masked pattern was here ####
-PREHOOK: query: SELECT * FROM exim_employee
-PREHOOK: type: QUERY
-PREHOOK: Input: default@exim_employee
-PREHOOK: Input: default@exim_employee@emp_country=cz
-PREHOOK: Input: default@exim_employee@emp_country=in
-PREHOOK: Input: default@exim_employee@emp_country=us
-#### A masked pattern was here ####
-POSTHOOK: query: SELECT * FROM exim_employee
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@exim_employee
-POSTHOOK: Input: default@exim_employee@emp_country=cz
-POSTHOOK: Input: default@exim_employee@emp_country=in
-POSTHOOK: Input: default@exim_employee@emp_country=us
-#### A masked pattern was here ####
-1	cz
-2	cz
-3	cz
-4	cz
-5	cz
-6	cz
-1	in
-2	in
-3	in
-4	in
-5	in
-6	in
-1	us
-2	us
-3	us
-4	us
-5	us
-6	us