You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by br...@apache.org on 2014/08/15 17:38:13 UTC
svn commit: r1618215 [5/25] - in /hive/branches/spark: ./ bin/
common/src/java/org/apache/hadoop/hive/conf/
contrib/src/test/results/clientnegative/
contrib/src/test/results/clientpositive/
hbase-handler/src/test/results/negative/ hbase-handler/src/tes...
Modified: hive/branches/spark/pom.xml
URL: http://svn.apache.org/viewvc/hive/branches/spark/pom.xml?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/pom.xml (original)
+++ hive/branches/spark/pom.xml Fri Aug 15 15:37:46 2014
@@ -147,7 +147,7 @@
<slf4j.version>1.7.5</slf4j.version>
<ST4.version>4.0.4</ST4.version>
<super-csv.version>2.2.0</super-csv.version>
- <tez.version>0.4.0-incubating</tez.version>
+ <tez.version>0.4.1-incubating</tez.version>
<spark.version>1.1.0-SNAPSHOT</spark.version>
<scala.binary.version>2.10</scala.binary.version>
<scala.version>2.10.4</scala.version>
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/exec/TaskFactory.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/exec/TaskFactory.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/exec/TaskFactory.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/exec/TaskFactory.java Fri Aug 15 15:37:46 2014
@@ -33,6 +33,7 @@ import org.apache.hadoop.hive.ql.io.merg
import org.apache.hadoop.hive.ql.io.merge.MergeWork;
import org.apache.hadoop.hive.ql.io.rcfile.stats.PartialScanTask;
import org.apache.hadoop.hive.ql.io.rcfile.stats.PartialScanWork;
+import org.apache.hadoop.hive.ql.plan.ColumnStatsUpdateWork;
import org.apache.hadoop.hive.ql.plan.ColumnStatsWork;
import org.apache.hadoop.hive.ql.plan.ConditionalWork;
import org.apache.hadoop.hive.ql.plan.CopyWork;
@@ -94,6 +95,7 @@ public final class TaskFactory {
StatsTask.class));
taskvec.add(new TaskTuple<StatsNoJobWork>(StatsNoJobWork.class, StatsNoJobTask.class));
taskvec.add(new TaskTuple<ColumnStatsWork>(ColumnStatsWork.class, ColumnStatsTask.class));
+ taskvec.add(new TaskTuple<ColumnStatsUpdateWork>(ColumnStatsUpdateWork.class, ColumnStatsUpdateTask.class));
taskvec.add(new TaskTuple<MergeWork>(MergeWork.class,
MergeTask.class));
taskvec.add(new TaskTuple<DependencyCollectionWork>(DependencyCollectionWork.class,
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/hooks/EnforceReadOnlyTables.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/hooks/EnforceReadOnlyTables.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/hooks/EnforceReadOnlyTables.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/hooks/EnforceReadOnlyTables.java Fri Aug 15 15:37:46 2014
@@ -59,7 +59,10 @@ public class EnforceReadOnlyTables imple
public void run(SessionState sess, Set<ReadEntity> inputs,
Set<WriteEntity> outputs, UserGroupInformation ugi)
throws Exception {
- if (sess.getConf().getBoolean("hive.test.init.phase", false) == true) {
+
+ // Don't enforce during test driver setup or shutdown.
+ if (sess.getConf().getBoolean("hive.test.init.phase", false) ||
+ sess.getConf().getBoolean("hive.test.shutdown.phase", false)) {
return;
}
for (WriteEntity w: outputs) {
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java Fri Aug 15 15:37:46 2014
@@ -2716,4 +2716,19 @@ private void constructOneLBLocationMap(F
}
}
+ public void setMetaConf(String propName, String propValue) throws HiveException {
+ try {
+ getMSC().setMetaConf(propName, propValue);
+ } catch (TException te) {
+ throw new HiveException(te);
+ }
+ }
+
+ public String getMetaConf(String propName) throws HiveException {
+ try {
+ return getMSC().getMetaConf(propName);
+ } catch (TException te) {
+ throw new HiveException(te);
+ }
+ }
};
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/DDLSemanticAnalyzer.java Fri Aug 15 15:37:46 2014
@@ -25,6 +25,7 @@ import java.io.Serializable;
import java.net.URI;
import java.net.URISyntaxException;
import java.util.ArrayList;
+import java.util.Arrays;
import java.util.Collection;
import java.util.HashMap;
import java.util.HashSet;
@@ -57,6 +58,7 @@ import org.apache.hadoop.hive.metastore.
import org.apache.hadoop.hive.ql.Driver;
import org.apache.hadoop.hive.ql.ErrorMsg;
import org.apache.hadoop.hive.ql.exec.ArchiveUtils;
+import org.apache.hadoop.hive.ql.exec.ColumnStatsUpdateTask;
import org.apache.hadoop.hive.ql.exec.FetchTask;
import org.apache.hadoop.hive.ql.exec.FunctionRegistry;
import org.apache.hadoop.hive.ql.exec.Task;
@@ -91,6 +93,8 @@ import org.apache.hadoop.hive.ql.plan.Al
import org.apache.hadoop.hive.ql.plan.AlterTableDesc.AlterTableTypes;
import org.apache.hadoop.hive.ql.plan.AlterTableExchangePartition;
import org.apache.hadoop.hive.ql.plan.AlterTableSimpleDesc;
+import org.apache.hadoop.hive.ql.plan.ColumnStatsDesc;
+import org.apache.hadoop.hive.ql.plan.ColumnStatsUpdateWork;
import org.apache.hadoop.hive.ql.plan.CreateDatabaseDesc;
import org.apache.hadoop.hive.ql.plan.CreateIndexDesc;
import org.apache.hadoop.hive.ql.plan.DDLWork;
@@ -276,6 +280,8 @@ public class DDLSemanticAnalyzer extends
analyzeAlterTableClusterSort(ast, tableName, partSpec);
} else if (ast.getToken().getType() == HiveParser.TOK_COMPACT) {
analyzeAlterTableCompact(ast, tableName, partSpec);
+ } else if(ast.getToken().getType() == HiveParser.TOK_ALTERTABLE_UPDATECOLSTATS){
+ analyzeAlterTableUpdateStats(ast,tblPart);
}
break;
}
@@ -378,6 +384,9 @@ public class DDLSemanticAnalyzer extends
case HiveParser.TOK_ALTERTABLE_RENAME:
analyzeAlterTableRename(ast, false);
break;
+ case HiveParser.TOK_ALTERTABLE_UPDATECOLSTATS:
+ analyzeAlterTableUpdateStats(ast, null);
+ break;
case HiveParser.TOK_ALTERTABLE_TOUCH:
analyzeAlterTableTouch(ast);
break;
@@ -507,6 +516,57 @@ public class DDLSemanticAnalyzer extends
}
}
+ private void analyzeAlterTableUpdateStats(ASTNode ast, TablePartition tblPart)
+ throws SemanticException {
+ String tblName = null;
+ String colName = null;
+ Map<String, String> mapProp = null;
+ Map<String, String> partSpec = null;
+ String partName = null;
+ if (tblPart == null) {
+ tblName = getUnescapedName((ASTNode) ast.getChild(0));
+ colName = getUnescapedName((ASTNode) ast.getChild(1));
+ mapProp = getProps((ASTNode) (ast.getChild(2)).getChild(0));
+ } else {
+ tblName = tblPart.tableName;
+ partSpec = tblPart.partSpec;
+ try {
+ partName = Warehouse.makePartName(partSpec, false);
+ } catch (MetaException e) {
+ // TODO Auto-generated catch block
+ throw new SemanticException("partition " + partSpec.toString()
+ + " not found");
+ }
+ colName = getUnescapedName((ASTNode) ast.getChild(0));
+ mapProp = getProps((ASTNode) (ast.getChild(1)).getChild(0));
+ }
+
+ Table tbl = null;
+ try {
+ tbl = db.getTable(tblName);
+ } catch (HiveException e) {
+ throw new SemanticException("table " + tbl + " not found");
+ }
+
+ String colType = null;
+ List<FieldSchema> cols = tbl.getCols();
+ for (FieldSchema col : cols) {
+ if (colName.equalsIgnoreCase(col.getName())) {
+ colType = col.getType();
+ break;
+ }
+ }
+
+ if (colType == null)
+ throw new SemanticException("column type not found");
+
+ ColumnStatsDesc cStatsDesc = new ColumnStatsDesc(tbl.getTableName(),
+ Arrays.asList(colName), Arrays.asList(colType), partSpec == null);
+ ColumnStatsUpdateTask cStatsUpdateTask = (ColumnStatsUpdateTask) TaskFactory
+ .get(new ColumnStatsUpdateWork(cStatsDesc, partName, mapProp), conf);
+ rootTasks.add(cStatsUpdateTask);
+ }
+
private void analyzeSetShowRole(ASTNode ast) throws SemanticException {
switch (ast.getChildCount()) {
case 0:
@@ -745,6 +805,8 @@ public class DDLSemanticAnalyzer extends
if (dbProps != null) {
createDatabaseDesc.setDatabaseProperties(dbProps);
}
+ Database database = new Database(dbName, dbComment, dbLocation, dbProps);
+ outputs.add(new WriteEntity(database, WriteEntity.WriteType.DDL_NO_LOCK));
rootTasks.add(TaskFactory.get(new DDLWork(getInputs(), getOutputs(),
createDatabaseDesc), conf));
@@ -795,8 +857,12 @@ public class DDLSemanticAnalyzer extends
rootTasks.add(TaskFactory.get(new DDLWork(getInputs(), getOutputs(), dropDatabaseDesc), conf));
}
- private void analyzeSwitchDatabase(ASTNode ast) {
+ private void analyzeSwitchDatabase(ASTNode ast) throws SemanticException {
String dbName = unescapeIdentifier(ast.getChild(0).getText());
+ Database database = getDatabase(dbName, true);
+ ReadEntity dbReadEntity = new ReadEntity(database);
+ dbReadEntity.noLockNeeded();
+ inputs.add(dbReadEntity);
SwitchDatabaseDesc switchDatabaseDesc = new SwitchDatabaseDesc(dbName);
rootTasks.add(TaskFactory.get(new DDLWork(getInputs(), getOutputs(),
switchDatabaseDesc), conf));
@@ -1015,7 +1081,7 @@ public class DDLSemanticAnalyzer extends
private void analyzeCreateIndex(ASTNode ast) throws SemanticException {
String indexName = unescapeIdentifier(ast.getChild(0).getText());
String typeName = unescapeSQLString(ast.getChild(1).getText());
- String[] qualified = getQualifiedTableName((ASTNode) ast.getChild(2));
+ String[] qTabName = getQualifiedTableName((ASTNode) ast.getChild(2));
List<String> indexedCols = getColumnNames((ASTNode) ast.getChild(3));
IndexType indexType = HiveIndex.getIndexType(typeName);
@@ -1080,15 +1146,15 @@ public class DDLSemanticAnalyzer extends
}
storageFormat.fillDefaultStorageFormat();
-
if (indexTableName == null) {
- indexTableName = MetaStoreUtils.getIndexTableName(qualified[0], qualified[1], indexName);
- indexTableName = qualified[0] + "." + indexTableName; // on same database with base table
+ indexTableName = MetaStoreUtils.getIndexTableName(qTabName[0], qTabName[1], indexName);
+ indexTableName = qTabName[0] + "." + indexTableName; // on same database with base table
} else {
indexTableName = getDotName(Utilities.getDbTableName(indexTableName));
}
+ inputs.add(new ReadEntity(getTable(qTabName)));
- CreateIndexDesc crtIndexDesc = new CreateIndexDesc(getDotName(qualified), indexName,
+ CreateIndexDesc crtIndexDesc = new CreateIndexDesc(getDotName(qTabName), indexName,
indexedCols, indexTableName, deferredRebuild, storageFormat.getInputFormat(),
storageFormat.getOutputFormat(),
storageFormat.getStorageHandler(), typeName, location, idxProps, tblProps,
@@ -1116,6 +1182,8 @@ public class DDLSemanticAnalyzer extends
}
}
+ inputs.add(new ReadEntity(getTable(tableName)));
+
DropIndexDesc dropIdxDesc = new DropIndexDesc(indexName, tableName);
rootTasks.add(TaskFactory.get(new DDLWork(getInputs(), getOutputs(),
dropIdxDesc), conf));
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/HiveParser.g Fri Aug 15 15:37:46 2014
@@ -146,6 +146,7 @@ TOK_ALTERTABLE_ARCHIVE;
TOK_ALTERTABLE_UNARCHIVE;
TOK_ALTERTABLE_SERDEPROPERTIES;
TOK_ALTERTABLE_SERIALIZER;
+TOK_ALTERTABLE_UPDATECOLSTATS;
TOK_TABLE_PARTITION;
TOK_ALTERTABLE_FILEFORMAT;
TOK_ALTERTABLE_LOCATION;
@@ -938,6 +939,7 @@ alterTableStatementSuffix
: alterStatementSuffixRename
| alterStatementSuffixAddCol
| alterStatementSuffixRenameCol
+ | alterStatementSuffixUpdateStatsCol
| alterStatementSuffixDropPartitions
| alterStatementSuffixAddPartitions
| alterStatementSuffixTouch
@@ -1028,6 +1030,13 @@ alterStatementSuffixRenameCol
->^(TOK_ALTERTABLE_RENAMECOL tableName $oldName $newName colType $comment? alterStatementChangeColPosition?)
;
+alterStatementSuffixUpdateStatsCol
+@init { pushMsg("update column statistics", state); }
+@after { popMsg(state); }
+ : identifier KW_UPDATE KW_STATISTICS KW_FOR KW_COLUMN? colName=identifier KW_SET tableProperties (KW_COMMENT comment=StringLiteral)?
+ ->^(TOK_ALTERTABLE_UPDATECOLSTATS identifier $colName tableProperties $comment?)
+ ;
+
alterStatementChangeColPosition
: first=KW_FIRST|KW_AFTER afterCol=identifier
->{$first != null}? ^(TOK_ALTERTABLE_CHANGECOL_AFTER_POSITION )
@@ -1130,6 +1139,7 @@ alterTblPartitionStatementSuffix
| alterStatementSuffixMergeFiles
| alterStatementSuffixSerdeProperties
| alterStatementSuffixRenamePart
+ | alterStatementSuffixStatsPart
| alterStatementSuffixBucketNum
| alterTblPartitionStatementSuffixSkewedLocation
| alterStatementSuffixClusterbySortby
@@ -1221,6 +1231,13 @@ alterStatementSuffixRenamePart
->^(TOK_ALTERTABLE_RENAMEPART partitionSpec)
;
+alterStatementSuffixStatsPart
+@init { pushMsg("alter table stats partition statement", state); }
+@after { popMsg(state); }
+ : KW_UPDATE KW_STATISTICS KW_FOR KW_COLUMN? colName=identifier KW_SET tableProperties (KW_COMMENT comment=StringLiteral)?
+ ->^(TOK_ALTERTABLE_UPDATECOLSTATS $colName tableProperties $comment?)
+ ;
+
alterStatementSuffixMergeFiles
@init { pushMsg("", state); }
@after { popMsg(state); }
@@ -1300,6 +1317,7 @@ descStatement
| (KW_DESCRIBE|KW_DESC) (KW_DATABASE|KW_SCHEMA) KW_EXTENDED? (dbName=identifier) -> ^(TOK_DESCDATABASE $dbName KW_EXTENDED?)
;
+
analyzeStatement
@init { pushMsg("analyze statement", state); }
@after { popMsg(state); }
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java Fri Aug 15 15:37:46 2014
@@ -10163,6 +10163,7 @@ public class SemanticAnalyzer extends Ba
String dbName = qualified.length == 1 ? SessionState.get().getCurrentDatabase() : qualified[0];
Database database = getDatabase(dbName);
outputs.add(new WriteEntity(database, WriteEntity.WriteType.DDL_SHARED));
+ outputs.add(new WriteEntity(new Table(dbName, tableName), WriteEntity.WriteType.DDL_NO_LOCK));
if (isTemporary) {
if (partCols.size() > 0) {
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzerFactory.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzerFactory.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzerFactory.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzerFactory.java Fri Aug 15 15:37:46 2014
@@ -109,6 +109,7 @@ public final class SemanticAnalyzerFacto
commandType.put(HiveParser.TOK_ALTERTABLE_PARTCOLTYPE, HiveOperation.ALTERTABLE_PARTCOLTYPE);
commandType.put(HiveParser.TOK_SHOW_COMPACTIONS, HiveOperation.SHOW_COMPACTIONS);
commandType.put(HiveParser.TOK_SHOW_TRANSACTIONS, HiveOperation.SHOW_TRANSACTIONS);
+ commandType.put(HiveParser.TOK_ALTERTABLE_UPDATECOLSTATS, HiveOperation.ALTERTABLE_UPDATETABLESTATS);
}
static {
@@ -231,12 +232,14 @@ public final class SemanticAnalyzerFacto
case HiveParser.TOK_TRUNCATETABLE:
case HiveParser.TOK_EXCHANGEPARTITION:
case HiveParser.TOK_SHOW_SET_ROLE:
-
+ case HiveParser.TOK_ALTERTABLE_UPDATECOLSTATS:
return new DDLSemanticAnalyzer(conf);
case HiveParser.TOK_ALTERTABLE_PARTITION:
HiveOperation commandType = null;
Integer type = ((ASTNode) tree.getChild(1)).getToken().getType();
- if (tree.getChild(0).getChildCount() > 1) {
+ if (type == HiveParser.TOK_ALTERTABLE_UPDATECOLSTATS) {
+ commandType = HiveOperation.ALTERTABLE_UPDATEPARTSTATS;
+ } else if (tree.getChild(0).getChildCount() > 1) {
commandType = tablePartitionCommandType.get(type)[1];
} else {
commandType = tablePartitionCommandType.get(type)[0];
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/plan/HiveOperation.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/plan/HiveOperation.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/plan/HiveOperation.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/plan/HiveOperation.java Fri Aug 15 15:37:46 2014
@@ -27,7 +27,7 @@ public enum HiveOperation {
IMPORT("IMPORT", null, new Privilege[]{Privilege.ALTER_METADATA, Privilege.ALTER_DATA}),
CREATEDATABASE("CREATEDATABASE", null, new Privilege[]{Privilege.CREATE}),
DROPDATABASE("DROPDATABASE", null, new Privilege[]{Privilege.DROP}),
- SWITCHDATABASE("SWITCHDATABASE", new Privilege[]{Privilege.SELECT}, null),
+ SWITCHDATABASE("SWITCHDATABASE", null, null),
LOCKDB("LOCKDATABASE", new Privilege[]{Privilege.LOCK}, null),
UNLOCKDB("UNLOCKDATABASE", new Privilege[]{Privilege.LOCK}, null),
DROPTABLE ("DROPTABLE", null, new Privilege[]{Privilege.DROP}),
@@ -38,6 +38,8 @@ public enum HiveOperation {
ALTERTABLE_REPLACECOLS("ALTERTABLE_REPLACECOLS", new Privilege[]{Privilege.ALTER_METADATA}, null),
ALTERTABLE_RENAMECOL("ALTERTABLE_RENAMECOL", new Privilege[]{Privilege.ALTER_METADATA}, null),
ALTERTABLE_RENAMEPART("ALTERTABLE_RENAMEPART", new Privilege[]{Privilege.DROP}, new Privilege[]{Privilege.CREATE}),
+ ALTERTABLE_UPDATEPARTSTATS("ALTERTABLE_UPDATEPARTSTATS", new Privilege[]{Privilege.ALTER_METADATA}, null),
+ ALTERTABLE_UPDATETABLESTATS("ALTERTABLE_UPDATETABLESTATS", new Privilege[]{Privilege.ALTER_METADATA}, null),
ALTERTABLE_RENAME("ALTERTABLE_RENAME", new Privilege[]{Privilege.ALTER_METADATA}, null),
ALTERTABLE_DROPPARTS("ALTERTABLE_DROPPARTS", new Privilege[]{Privilege.DROP}, null),
// The location is input and table is output for alter-table add partitions
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/plan/PlanUtils.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/plan/PlanUtils.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/plan/PlanUtils.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/plan/PlanUtils.java Fri Aug 15 15:37:46 2014
@@ -157,7 +157,7 @@ public final class PlanUtils {
} catch (ClassNotFoundException e) {
// mimicking behaviour in CreateTableDesc tableDesc creation
// returning null table description for output.
- e.printStackTrace();
+ LOG.warn("Unable to find class in getDefaultTableDesc: " + e.getMessage(), e);
return null;
}
return ret;
@@ -364,8 +364,7 @@ public final class PlanUtils {
ret.setInputFileFormatClass(in_class);
ret.setOutputFileFormatClass(out_class);
} catch (ClassNotFoundException e) {
- e.printStackTrace();
- return null;
+ throw new RuntimeException("Unable to find class in getTableDesc: " + e.getMessage(), e);
}
return ret;
}
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/processors/SetProcessor.java Fri Aug 15 15:37:46 2014
@@ -32,6 +32,7 @@ import java.util.TreeMap;
import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.metastore.api.FieldSchema;
import org.apache.hadoop.hive.metastore.api.Schema;
+import org.apache.hadoop.hive.ql.metadata.Hive;
import org.apache.hadoop.hive.ql.parse.VariableSubstitution;
import org.apache.hadoop.hive.ql.session.SessionState;
@@ -104,11 +105,12 @@ public class SetProcessor implements Com
try {
return new CommandProcessorResponse(setVariable(varname, varvalue));
} catch (Exception e) {
- return new CommandProcessorResponse(1, e.getMessage(), "42000");
+ return new CommandProcessorResponse(1, e.getMessage(), "42000",
+ e instanceof IllegalArgumentException ? null : e);
}
}
- public static int setVariable(String varname, String varvalue) throws IllegalArgumentException {
+ public static int setVariable(String varname, String varvalue) throws Exception {
SessionState ss = SessionState.get();
if (varvalue.contains("\n")){
ss.err.println("Warning: Value had a \\n character in it.");
@@ -126,6 +128,10 @@ public class SetProcessor implements Com
} else if (varname.startsWith(HIVEVAR_PREFIX)) {
String propName = varname.substring(HIVEVAR_PREFIX.length());
ss.getHiveVariables().put(propName, new VariableSubstitution().substitute(ss.getConf(),varvalue));
+ } else if (varname.startsWith(METACONF_PREFIX)) {
+ String propName = varname.substring(METACONF_PREFIX.length());
+ Hive hive = Hive.get(ss.getConf());
+ hive.setMetaConf(propName, new VariableSubstitution().substitute(ss.getConf(), varvalue));
} else {
setConf(varname, varname, varvalue, true);
}
@@ -178,8 +184,7 @@ public class SetProcessor implements Com
return sortedEnvMap;
}
-
- private CommandProcessorResponse getVariable(String varname) {
+ private CommandProcessorResponse getVariable(String varname) throws Exception {
SessionState ss = SessionState.get();
if (varname.equals("silent")){
ss.out.println("silent" + "=" + ss.getIsSilent());
@@ -222,6 +227,17 @@ public class SetProcessor implements Com
ss.out.println(varname + " is undefined as a hive variable");
return new CommandProcessorResponse(1);
}
+ } else if (varname.indexOf(METACONF_PREFIX) == 0) {
+ String var = varname.substring(METACONF_PREFIX.length());
+ Hive hive = Hive.get(ss.getConf());
+ String value = hive.getMetaConf(var);
+ if (value != null) {
+ ss.out.println(METACONF_PREFIX + var + "=" + value);
+ return createProcessorSuccessResponse();
+ } else {
+ ss.out.println(varname + " is undefined as a hive meta variable");
+ return new CommandProcessorResponse(1);
+ }
} else {
dumpOption(varname);
return createProcessorSuccessResponse();
@@ -263,10 +279,12 @@ public class SetProcessor implements Com
return new CommandProcessorResponse(0);
}
return executeSetVariable(part[0],part[1]);
- } else {
+ }
+ try {
return getVariable(nwcmd);
+ } catch (Exception e) {
+ return new CommandProcessorResponse(1, e.getMessage(), "42000", e);
}
-
}
// create a Schema object containing the give column
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/HiveOperationType.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/HiveOperationType.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/HiveOperationType.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/HiveOperationType.java Fri Aug 15 15:37:46 2014
@@ -59,6 +59,8 @@ public enum HiveOperationType {
ANALYZE_TABLE,
ALTERTABLE_BUCKETNUM,
ALTERPARTITION_BUCKETNUM,
+ ALTERTABLE_UPDATETABLESTATS,
+ ALTERTABLE_UPDATEPARTSTATS,
SHOWDATABASES,
SHOWTABLES,
SHOWCOLUMNS,
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/Operation2Privilege.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/Operation2Privilege.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/Operation2Privilege.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/Operation2Privilege.java Fri Aug 15 15:37:46 2014
@@ -27,6 +27,7 @@ import java.util.Set;
import org.apache.hadoop.hive.ql.security.authorization.plugin.HiveOperationType;
import org.apache.hadoop.hive.ql.security.authorization.plugin.HivePrivilegeObject;
import org.apache.hadoop.hive.ql.security.authorization.plugin.HivePrivilegeObject.HivePrivObjectActionType;
+import org.apache.hadoop.hive.ql.security.authorization.plugin.HivePrivilegeObject.HivePrivilegeObjectType;
/**
* Mapping of operation to its required input and output privileges
@@ -43,6 +44,7 @@ public class Operation2Privilege {
// The following fields specify the criteria on objects for this priv to be required
private final IOType ioType;
private final HivePrivObjectActionType actionType;
+ private final HivePrivilegeObjectType objectType;
private PrivRequirement(SQLPrivTypeGrant[] privs, IOType ioType) {
@@ -51,11 +53,22 @@ public class Operation2Privilege {
private PrivRequirement(SQLPrivTypeGrant[] privs, IOType ioType,
HivePrivObjectActionType actionType) {
+ this(privs, ioType, actionType, null);
+ }
+
+ private PrivRequirement(SQLPrivTypeGrant[] privs, HivePrivilegeObjectType objectType) {
+ this(privs, null, null, objectType);
+ }
+
+ private PrivRequirement(SQLPrivTypeGrant[] privs, IOType ioType,
+ HivePrivObjectActionType actionType, HivePrivilegeObjectType objectType) {
this.reqPrivs = privs;
this.ioType = ioType;
this.actionType = actionType;
+ this.objectType = objectType;
}
+
/**
* Utility function that takes a input and output privilege objects
* @param inGrant
@@ -70,6 +83,15 @@ public class Operation2Privilege {
return privReqs;
}
+ /**
+ * Utility function that converts PrivRequirement array into list
+ * @param privs
+ * @return
+ */
+ static List<PrivRequirement> newPrivRequirementList(PrivRequirement... privs) {
+ return new ArrayList<PrivRequirement>(Arrays.asList(privs));
+ }
+
private SQLPrivTypeGrant[] getReqPrivs() {
return reqPrivs;
}
@@ -82,6 +104,10 @@ public class Operation2Privilege {
return actionType;
}
+ public HivePrivilegeObjectType getObjectType() {
+ return objectType;
+ }
+
}
private static Map<HiveOperationType, List<PrivRequirement>> op2Priv;
@@ -107,9 +133,9 @@ public class Operation2Privilege {
(SEL_NOGRANT_AR,
SEL_NOGRANT_AR)); //??
- op2Priv.put(HiveOperationType.CREATEDATABASE,
- PrivRequirement.newIOPrivRequirement
-(ADMIN_PRIV_AR, OWNER_INS_SEL_DEL_NOGRANT_AR));
+ op2Priv.put(HiveOperationType.CREATEDATABASE, PrivRequirement.newPrivRequirementList(
+ new PrivRequirement(OWNER_INS_SEL_DEL_NOGRANT_AR, HivePrivilegeObjectType.DFS_URI),
+ new PrivRequirement(OWNER_INS_SEL_DEL_NOGRANT_AR, HivePrivilegeObjectType.LOCAL_URI)));
op2Priv.put(HiveOperationType.DROPDATABASE, PrivRequirement.newIOPrivRequirement
(null, OWNER_PRIV_AR));
@@ -150,6 +176,10 @@ public class Operation2Privilege {
(OWNER_PRIV_AR, OWNER_PRIV_AR));
op2Priv.put(HiveOperationType.ALTERTABLE_RENAME, PrivRequirement.newIOPrivRequirement
(OWNER_PRIV_AR, OWNER_PRIV_AR));
+ op2Priv.put(HiveOperationType.ALTERTABLE_UPDATETABLESTATS, PrivRequirement.newIOPrivRequirement
+(OWNER_PRIV_AR, OWNER_PRIV_AR));
+ op2Priv.put(HiveOperationType.ALTERTABLE_UPDATEPARTSTATS, PrivRequirement.newIOPrivRequirement
+(OWNER_PRIV_AR, OWNER_PRIV_AR));
op2Priv.put(HiveOperationType.ALTERTABLE_TOUCH, PrivRequirement.newIOPrivRequirement
(OWNER_PRIV_AR, OWNER_PRIV_AR));
op2Priv.put(HiveOperationType.ALTERTABLE_ARCHIVE, PrivRequirement.newIOPrivRequirement
@@ -300,9 +330,9 @@ public class Operation2Privilege {
(null, null));
// require db ownership, if there is a file require SELECT , INSERT, and DELETE
- op2Priv.put(HiveOperationType.CREATETABLE,
- PrivRequirement.newIOPrivRequirement
-(OWNER_INS_SEL_DEL_NOGRANT_AR, OWNER_PRIV_AR));
+ op2Priv.put(HiveOperationType.CREATETABLE, PrivRequirement.newPrivRequirementList(
+ new PrivRequirement(OWNER_INS_SEL_DEL_NOGRANT_AR, IOType.INPUT),
+ new PrivRequirement(OWNER_PRIV_AR, HivePrivilegeObjectType.DATABASE)));
op2Priv.put(HiveOperationType.ALTERDATABASE, PrivRequirement.newIOPrivRequirement
(null, ADMIN_PRIV_AR));
@@ -388,20 +418,17 @@ public class Operation2Privilege {
List<PrivRequirement> opPrivs = op2Priv.get(hiveOpType);
RequiredPrivileges reqPrivs = new RequiredPrivileges();
- // Find the PrivRequirements that match on IOType and ActionType, and add
- // the privilege
- // required to reqPrivs
+ // Find the PrivRequirements that match on IOType, ActionType, and HivePrivilegeObjectType add
+ // the privilege required to reqPrivs
for (PrivRequirement opPriv : opPrivs) {
- if (opPriv.getIOType() != ioType) {
+ if (opPriv.getIOType() != null && opPriv.getIOType() != ioType) {
+ continue;
+ }
+ if (opPriv.getActionType() != null && opPriv.getActionType() != hObj.getActionType()) {
continue;
}
- if (opPriv.getActionType() != null) {
- // if action in PrivRequirement is null, it means that
- // the privileges are required irrespective of hObj's action type
- // If it is not null, action type has to match
- if (hObj.getActionType() != opPriv.getActionType()) {
- continue;
- }
+ if (opPriv.getObjectType() != null && opPriv.getObjectType() != hObj.getType()) {
+ continue;
}
reqPrivs.addAll(opPriv.getReqPrivs());
}
Modified: hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/SQLStdHiveAuthorizationValidator.java
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/SQLStdHiveAuthorizationValidator.java?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/SQLStdHiveAuthorizationValidator.java (original)
+++ hive/branches/spark/ql/src/java/org/apache/hadoop/hive/ql/security/authorization/plugin/sqlstd/SQLStdHiveAuthorizationValidator.java Fri Aug 15 15:37:46 2014
@@ -94,6 +94,11 @@ public class SQLStdHiveAuthorizationVali
RequiredPrivileges requiredPrivs = Operation2Privilege.getRequiredPrivs(hiveOpType, hiveObj,
ioType);
+ if(requiredPrivs.getRequiredPrivilegeSet().isEmpty()){
+ // no privileges required, so don't need to check this object privileges
+ continue;
+ }
+
// find available privileges
RequiredPrivileges availPrivs = new RequiredPrivileges(); //start with an empty priv set;
switch (hiveObj.getType()) {
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/repair.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/repair.q?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/repair.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/repair.q Fri Aug 15 15:37:46 2014
@@ -8,10 +8,10 @@ dfs ${system:test.dfs.mkdir} ${system:te
dfs ${system:test.dfs.mkdir} ${system:test.warehouse.dir}/repairtable/p1=b/p2=a;
dfs -touchz ${system:test.warehouse.dir}/repairtable/p1=b/p2=a/datafile;
-MSCK TABLE repairtable;
+MSCK TABLE default.repairtable;
-MSCK REPAIR TABLE repairtable;
+MSCK REPAIR TABLE default.repairtable;
MSCK TABLE repairtable;
-DROP TABLE repairtable;
+DROP TABLE default.repairtable;
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/show_partitions.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/show_partitions.q?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/show_partitions.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/show_partitions.q Fri Aug 15 15:37:46 2014
@@ -3,3 +3,30 @@ SHOW PARTITIONS default.srcpart;
SHOW PARTITIONS srcpart PARTITION(hr='11');
SHOW PARTITIONS srcpart PARTITION(ds='2008-04-08');
SHOW PARTITIONS srcpart PARTITION(ds='2008-04-08', hr='12');
+
+
+SHOW PARTITIONS default.srcpart;
+SHOW PARTITIONS default.srcpart PARTITION(hr='11');
+SHOW PARTITIONS default.srcpart PARTITION(ds='2008-04-08');
+SHOW PARTITIONS default.srcpart PARTITION(ds='2008-04-08', hr='12');
+
+CREATE DATABASE db1;
+USE db1;
+
+CREATE TABLE srcpart (key1 INT, value1 STRING) PARTITIONED BY (ds STRING, hr STRING);
+ALTER TABLE srcpart ADD PARTITION (ds='3', hr='3');
+ALTER TABLE srcpart ADD PARTITION (ds='4', hr='4');
+ALTER TABLE srcpart ADD PARTITION (ds='4', hr='5');
+
+-- from db1 to default db
+SHOW PARTITIONS default.srcpart PARTITION(hr='11');
+SHOW PARTITIONS default.srcpart PARTITION(ds='2008-04-08', hr='12');
+
+-- from db1 to db1
+SHOW PARTITIONS srcpart PARTITION(ds='4');
+SHOW PARTITIONS srcpart PARTITION(ds='3', hr='3');
+
+use default;
+-- from default to db1
+SHOW PARTITIONS db1.srcpart PARTITION(ds='4');
+SHOW PARTITIONS db1.srcpart PARTITION(ds='3', hr='3');
\ No newline at end of file
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/show_tblproperties.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/show_tblproperties.q?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/show_tblproperties.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/show_tblproperties.q Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
create table tmpfoo (a String);
show tblproperties tmpfoo("bar");
+show tblproperties default.tmpfoo("bar");
alter table tmpfoo set tblproperties ("bar" = "bar value");
alter table tmpfoo set tblproperties ("tmp" = "true");
@@ -8,4 +9,26 @@ alter table tmpfoo set tblproperties ("t
show tblproperties tmpfoo;
show tblproperties tmpfoo("bar");
-drop table tmpfoo;
+show tblproperties default.tmpfoo;
+show tblproperties default.tmpfoo("bar");
+
+CREATE DATABASE db1;
+USE db1;
+
+CREATE TABLE tmpfoo (b STRING);
+alter table tmpfoo set tblproperties ("bar" = "bar value1");
+alter table tmpfoo set tblproperties ("tmp" = "true1");
+
+-- from db1 to default db
+show tblproperties default.tmpfoo;
+show tblproperties default.tmpfoo("bar");
+
+-- from db1 to db1
+show tblproperties tmpfoo;
+show tblproperties tmpfoo("bar");
+
+use default;
+-- from default to db1
+show tblproperties db1.tmpfoo;
+show tblproperties db1.tmpfoo("bar");
+
Modified: hive/branches/spark/ql/src/test/queries/clientpositive/virtual_column.q
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/queries/clientpositive/virtual_column.q?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/queries/clientpositive/virtual_column.q (original)
+++ hive/branches/spark/ql/src/test/queries/clientpositive/virtual_column.q Fri Aug 15 15:37:46 2014
@@ -14,6 +14,3 @@ CREATE TABLE src_index_test_rc (key int,
set hive.io.rcfile.record.buffer.size = 1024;
INSERT OVERWRITE TABLE src_index_test_rc SELECT * FROM src;
select INPUT__FILE__NAME, key, BLOCK__OFFSET__INSIDE__FILE from src_index_test_rc order by key;
-
-DROP TABLE src_index_test_rc;
-DROP INDEX src_index on src_index_test_rc;
Modified: hive/branches/spark/ql/src/test/results/clientnegative/add_partition_with_whitelist.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/add_partition_with_whitelist.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/add_partition_with_whitelist.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/add_partition_with_whitelist.q.out Fri Aug 15 15:37:46 2014
@@ -3,6 +3,7 @@ PREHOOK: query: -- This pattern matches
CREATE TABLE part_whitelist_test (key STRING, value STRING) PARTITIONED BY (ds STRING)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@part_whitelist_test
POSTHOOK: query: -- This pattern matches all printable ASCII characters (disallow unicode) and disallows commas
CREATE TABLE part_whitelist_test (key STRING, value STRING) PARTITIONED BY (ds STRING)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/addpart1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/addpart1.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/addpart1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/addpart1.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@addpart1
POSTHOOK: query: create table addpart1 (a int) partitioned by (b string, c string)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_concatenate_indexed_table.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table src_rc_concatenate_test(key int, value string) stored as rcfile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@src_rc_concatenate_test
POSTHOOK: query: create table src_rc_concatenate_test(key int, value string) stored as rcfile
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
@@ -66,8 +67,10 @@ POSTHOOK: Input: default@src_rc_concaten
214 -7678496319
PREHOOK: query: create index src_rc_concatenate_test_index on table src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2")
PREHOOK: type: CREATEINDEX
+PREHOOK: Input: default@src_rc_concatenate_test
POSTHOOK: query: create index src_rc_concatenate_test_index on table src_rc_concatenate_test(key) as 'compact' WITH DEFERRED REBUILD IDXPROPERTIES ("prop1"="val1", "prop2"="val2")
POSTHOOK: type: CREATEINDEX
+POSTHOOK: Input: default@src_rc_concatenate_test
POSTHOOK: Output: default@default__src_rc_concatenate_test_src_rc_concatenate_test_index__
PREHOOK: query: show indexes on src_rc_concatenate_test
PREHOOK: type: SHOWINDEXES
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_file_format.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_file_format.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_file_format.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_file_format.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table alter_file_format_test (key int, value string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_file_format_test
POSTHOOK: query: create table alter_file_format_test (key int, value string)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_non_native.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_non_native.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_non_native.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_non_native.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: CREATE TABLE non_native1
STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@non_native1
POSTHOOK: query: CREATE TABLE non_native1(key int, value string)
STORED BY 'org.apache.hadoop.hive.ql.metadata.DefaultStorageHandler'
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_2columns.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_2columns.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_2columns.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_2columns.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- create testing table
create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_coltype
POSTHOOK: query: -- create testing table
create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_invalidcolname.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_invalidcolname.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_invalidcolname.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_invalidcolname.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- create testing table
create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_coltype
POSTHOOK: query: -- create testing table
create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_invalidtype.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_invalidtype.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_invalidtype.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_coltype_invalidtype.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- create testing table
create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_coltype
POSTHOOK: query: -- create testing table
create table alter_coltype(key string, value string) partitioned by (dt string, ts string)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_invalidspec.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_invalidspec.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_invalidspec.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_invalidspec.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- Create table
create table if not exists alter_part_invalidspec(key string, value string ) partitioned by (year string, month string) stored as textfile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_part_invalidspec
POSTHOOK: query: -- Create table
create table if not exists alter_part_invalidspec(key string, value string ) partitioned by (year string, month string) stored as textfile
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_nodrop.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_nodrop.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_nodrop.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_nodrop.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- Create table
create table if not exists alter_part_nodrop_part(key string, value string ) partitioned by (year string, month string) stored as textfile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_part_nodrop_part
POSTHOOK: query: -- Create table
create table if not exists alter_part_nodrop_part(key string, value string ) partitioned by (year string, month string) stored as textfile
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_nodrop_table.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_nodrop_table.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_nodrop_table.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_nodrop_table.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- Create table
create table if not exists alter_part_nodrop_table(key string, value string ) partitioned by (year string, month string) stored as textfile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_part_nodrop_table
POSTHOOK: query: -- Create table
create table if not exists alter_part_nodrop_table(key string, value string ) partitioned by (year string, month string) stored as textfile
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_offline.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_offline.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_offline.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_offline.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- create table
create table if not exists alter_part_offline (key string, value string ) partitioned by (year string, month string) stored as textfile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_part_offline
POSTHOOK: query: -- create table
create table if not exists alter_part_offline (key string, value string ) partitioned by (year string, month string) stored as textfile
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_with_whitelist.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_with_whitelist.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_with_whitelist.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_partition_with_whitelist.q.out Fri Aug 15 15:37:46 2014
@@ -3,6 +3,7 @@ PREHOOK: query: -- This pattern matches
CREATE TABLE part_whitelist_test (key STRING, value STRING) PARTITIONED BY (ds STRING)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@part_whitelist_test
POSTHOOK: query: -- This pattern matches all printable ASCII characters (disallow unicode) and disallows commas
CREATE TABLE part_whitelist_test (key STRING, value STRING) PARTITIONED BY (ds STRING)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_rename_partition_src
POSTHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
@@ -16,6 +17,7 @@ POSTHOOK: Output: default@alter_rename_p
PREHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_rename_partition
POSTHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure2.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure2.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_rename_partition_src
POSTHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
@@ -16,6 +17,7 @@ POSTHOOK: Output: default@alter_rename_p
PREHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_rename_partition
POSTHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure3.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_rename_partition_failure3.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_rename_partition_src
POSTHOOK: query: create table alter_rename_partition_src ( col1 string ) stored as textfile
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
@@ -16,6 +17,7 @@ POSTHOOK: Output: default@alter_rename_p
PREHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@alter_rename_partition
POSTHOOK: query: create table alter_rename_partition ( col1 string ) partitioned by (pcol1 string , pcol2 string) stored as sequencefile
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_table_add_partition.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_table_add_partition.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_table_add_partition.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_table_add_partition.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table mp (a int) partitioned by (b int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@mp
POSTHOOK: query: create table mp (a int) partitioned by (b int)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_table_wrong_regex.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_table_wrong_regex.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_table_wrong_regex.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_table_wrong_regex.q.out Fri Aug 15 15:37:46 2014
@@ -7,6 +7,7 @@ PREHOOK: query: create table aa ( test S
WITH SERDEPROPERTIES ("input.regex" = "(.*)", "output.format.string" = "$1s")
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@aa
POSTHOOK: query: create table aa ( test STRING )
ROW FORMAT SERDE 'org.apache.hadoop.hive.contrib.serde2.RegexSerDe'
WITH SERDEPROPERTIES ("input.regex" = "(.*)", "output.format.string" = "$1s")
Modified: hive/branches/spark/ql/src/test/results/clientnegative/alter_view_failure8.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/alter_view_failure8.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/alter_view_failure8.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/alter_view_failure8.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- should fail: can't u
CREATE TABLE invites (foo INT, bar STRING) PARTITIONED BY (ds STRING)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@invites
POSTHOOK: query: -- should fail: can't use ALTER VIEW on a table
CREATE TABLE invites (foo INT, bar STRING) PARTITIONED BY (ds STRING)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/altern1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/altern1.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/altern1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/altern1.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table altern1(a int, b int) partitioned by (ds string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@altern1
POSTHOOK: query: create table altern1(a int, b int) partitioned by (ds string)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive1.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive1.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE srcpart_archived LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@srcpart_archived
POSTHOOK: query: -- Tests trying to archive a partition twice.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive2.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive2.q.out Fri Aug 15 15:37:46 2014
@@ -11,6 +11,7 @@ POSTHOOK: type: DROPTABLE
PREHOOK: query: create table tstsrcpart like srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: create table tstsrcpart like srcpart
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_corrupt.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_corrupt.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_corrupt.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_corrupt.q.out Fri Aug 15 15:37:46 2014
@@ -1,7 +1,9 @@
PREHOOK: query: USE default
PREHOOK: type: SWITCHDATABASE
+PREHOOK: Input: database:default
POSTHOOK: query: USE default
POSTHOOK: type: SWITCHDATABASE
+POSTHOOK: Input: database:default
PREHOOK: query: drop table tstsrcpart
PREHOOK: type: DROPTABLE
POSTHOOK: query: drop table tstsrcpart
@@ -9,6 +11,7 @@ POSTHOOK: type: DROPTABLE
PREHOOK: query: create table tstsrcpart like srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: create table tstsrcpart like srcpart
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_insert1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_insert1.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_insert1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_insert1.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to inser
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to insert into archived partition.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_insert2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_insert2.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_insert2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_insert2.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to inser
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to insert into archived partition.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_insert3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_insert3.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_insert3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_insert3.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to creat
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to create partition inside of archived directory.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_insert4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_insert4.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_insert4.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_insert4.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to (poss
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to (possible) dynamic insert into archived partition.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_multi1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_multi1.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_multi1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_multi1.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to archive a partition twice.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_multi2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_multi2.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_multi2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_multi2.q.out Fri Aug 15 15:37:46 2014
@@ -11,6 +11,7 @@ POSTHOOK: type: DROPTABLE
PREHOOK: query: create table tstsrcpart like srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: create table tstsrcpart like srcpart
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_multi3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_multi3.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_multi3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_multi3.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to archive outer partition group containing other partition inside.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_multi4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_multi4.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_multi4.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_multi4.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to archive inner partition contained in archived partition group.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_multi5.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_multi5.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_multi5.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_multi5.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to unarc
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to unarchive outer partition group containing other partition inside.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_multi6.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_multi6.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_multi6.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_multi6.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to unarc
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to unarchive inner partition contained in archived partition group.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_multi7.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_multi7.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_multi7.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_multi7.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE tstsrcpart LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tstsrcpart
POSTHOOK: query: -- Tests trying to archive a partition group with custom locations.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec1.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec1.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE srcpart_archived LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@srcpart_archived
POSTHOOK: query: -- Tests trying to archive a partition twice.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec2.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec2.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE srcpart_archived LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@srcpart_archived
POSTHOOK: query: -- Tests trying to archive a partition twice.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec3.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec3.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE srcpart_archived LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@srcpart_archived
POSTHOOK: query: -- Tests trying to archive a partition twice.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec4.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec4.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec4.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE srcpart_archived LIKE srcpart
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@srcpart_archived
POSTHOOK: query: -- Tests trying to archive a partition twice.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec5.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec5.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec5.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/archive_partspec5.q.out Fri Aug 15 15:37:46 2014
@@ -4,6 +4,7 @@ PREHOOK: query: -- Tests trying to archi
CREATE TABLE srcpart_archived (key string, value string) partitioned by (ds string, hr int, min int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@srcpart_archived
POSTHOOK: query: -- Tests trying to archive a partition twice.
-- EXCLUDE_HADOOP_MAJOR_VERSIONS(0.17, 0.18, 0.19)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_addpartition.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_addpartition.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_addpartition.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_addpartition.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- check add partition w
create table tpart(i int, j int) partitioned by (k string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tpart
POSTHOOK: query: -- check add partition without insert privilege
create table tpart(i int, j int) partitioned by (k string)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_alter_db_owner.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_alter_db_owner.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_alter_db_owner.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_alter_db_owner.q.out Fri Aug 15 15:37:46 2014
@@ -2,8 +2,10 @@
create database dbao
PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:dbao
#### A masked pattern was here ####
create database dbao
POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:dbao
FAILED: HiveAccessControlException Permission denied: Principal [name=user1, type=USER] does not have following privileges for operation ALTERDATABASE_OWNER [[ADMIN PRIVILEGE] on Object [type=DATABASE, name=dbao]]
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_createview.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_createview.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_createview.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_createview.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- check create view wit
create table t1(i int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@t1
POSTHOOK: query: -- check create view without select privileges
create table t1(i int)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_ctas.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_ctas.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_ctas.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_ctas.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- check query without s
create table t1(i int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@t1
POSTHOOK: query: -- check query without select privilege fails
create table t1(i int)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_desc_table_nosel.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_desc_table_nosel.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_desc_table_nosel.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_desc_table_nosel.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- check if alter table
create table t1(i int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@t1
POSTHOOK: query: -- check if alter table fails as different user
create table t1(i int)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_drop_db_cascade.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_drop_db_cascade.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_drop_db_cascade.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_drop_db_cascade.q.out Fri Aug 15 15:37:46 2014
@@ -1,15 +1,19 @@
PREHOOK: query: -- ensure that drop database cascade works
create database dba1
PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:dba1
POSTHOOK: query: -- ensure that drop database cascade works
create database dba1
POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:dba1
PREHOOK: query: create table dba1.tab1(i int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:dba1
+PREHOOK: Output: dba1@dba1.tab1
POSTHOOK: query: create table dba1.tab1(i int)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:dba1
+POSTHOOK: Output: dba1@dba1.tab1
POSTHOOK: Output: dba1@tab1
PREHOOK: query: drop database dba1 cascade
PREHOOK: type: DROPDATABASE
@@ -24,15 +28,19 @@ POSTHOOK: Output: dba1@tab1
PREHOOK: query: -- check if drop database fails if the db has a table for which user does not have permission
create database dba2
PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:dba2
POSTHOOK: query: -- check if drop database fails if the db has a table for which user does not have permission
create database dba2
POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:dba2
PREHOOK: query: create table dba2.tab2(i int)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:dba2
+PREHOOK: Output: dba2@dba2.tab2
POSTHOOK: query: create table dba2.tab2(i int)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:dba2
+POSTHOOK: Output: dba2@dba2.tab2
POSTHOOK: Output: dba2@tab2
PREHOOK: query: set role ADMIN
PREHOOK: type: SHOW_ROLES
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_drop_db_empty.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_drop_db_empty.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_drop_db_empty.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_drop_db_empty.q.out Fri Aug 15 15:37:46 2014
@@ -1,9 +1,11 @@
#### A masked pattern was here ####
create database dba1
PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:dba1
#### A masked pattern was here ####
create database dba1
POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:dba1
PREHOOK: query: set role ADMIN
PREHOOK: type: SHOW_ROLES
POSTHOOK: query: set role ADMIN
@@ -38,8 +40,10 @@ public
PREHOOK: query: create database dba2
PREHOOK: type: CREATEDATABASE
+PREHOOK: Output: database:dba2
POSTHOOK: query: create database dba2
POSTHOOK: type: CREATEDATABASE
+POSTHOOK: Output: database:dba2
PREHOOK: query: show current roles
PREHOOK: type: SHOW_ROLES
POSTHOOK: query: show current roles
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_droppartition.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_droppartition.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_droppartition.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_droppartition.q.out Fri Aug 15 15:37:46 2014
@@ -2,6 +2,7 @@ PREHOOK: query: -- check drop partition
create table tpart(i int, j int) partitioned by (k string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@tpart
POSTHOOK: query: -- check drop partition without delete privilege
create table tpart(i int, j int) partitioned by (k string)
POSTHOOK: type: CREATETABLE
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_1.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_1.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_1.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_1.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table authorization_fail_1 (key int, value string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@authorization_fail_1
POSTHOOK: query: create table authorization_fail_1 (key int, value string)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_2.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_2.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_2.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_2.q.out Fri Aug 15 15:37:46 2014
@@ -1,6 +1,7 @@
PREHOOK: query: create table authorization_fail_2 (key int, value string) partitioned by (ds string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@authorization_fail_2
POSTHOOK: query: create table authorization_fail_2 (key int, value string) partitioned by (ds string)
POSTHOOK: type: CREATETABLE
POSTHOOK: Output: database:default
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_3.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_3.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_3.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_3.q.out Fri Aug 15 15:37:46 2014
@@ -3,6 +3,7 @@ PREHOOK: query: -- SORT_BEFORE_DIFF
create table authorization_fail_3 (key int, value string) partitioned by (ds string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@authorization_fail_3
POSTHOOK: query: -- SORT_BEFORE_DIFF
create table authorization_fail_3 (key int, value string) partitioned by (ds string)
Modified: hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_4.q.out
URL: http://svn.apache.org/viewvc/hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_4.q.out?rev=1618215&r1=1618214&r2=1618215&view=diff
==============================================================================
--- hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_4.q.out (original)
+++ hive/branches/spark/ql/src/test/results/clientnegative/authorization_fail_4.q.out Fri Aug 15 15:37:46 2014
@@ -3,6 +3,7 @@ PREHOOK: query: -- SORT_BEFORE_DIFF
create table authorization_fail_4 (key int, value string) partitioned by (ds string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
+PREHOOK: Output: default@authorization_fail_4
POSTHOOK: query: -- SORT_BEFORE_DIFF
create table authorization_fail_4 (key int, value string) partitioned by (ds string)