You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2018/06/03 16:25:13 UTC
hive git commit: HIVE-19633 : Remove/Migrate Minimr tests (Prasanth J
via Sergey Shelukhin)
Repository: hive
Updated Branches:
refs/heads/master fc5a5a636 -> 8cb99d680
HIVE-19633 : Remove/Migrate Minimr tests (Prasanth J via Sergey Shelukhin)
Signed-off-by: Ashutosh Chauhan <ha...@apache.org>
Project: http://git-wip-us.apache.org/repos/asf/hive/repo
Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/8cb99d68
Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/8cb99d68
Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/8cb99d68
Branch: refs/heads/master
Commit: 8cb99d680b1688ebfa617483f3010b217885ae90
Parents: fc5a5a6
Author: Prasanth Jayachandran <pr...@apache.org>
Authored: Sun Jun 3 09:23:57 2018 -0700
Committer: Ashutosh Chauhan <ha...@apache.org>
Committed: Sun Jun 3 09:23:57 2018 -0700
----------------------------------------------------------------------
.../hadoop/hive/cli/TestMinimrCliDriver.java | 2 ++
.../test/resources/testconfiguration.properties | 27 +++------------
itests/util/pom.xml | 5 +++
.../hive/ql/hooks/VerifyNumReducersHook.java | 25 +++++++++++---
.../exec/tez/monitoring/TezProgressMonitor.java | 10 +++---
.../clientpositive/root_dir_external_table.q | 3 +-
ql/src/test/queries/clientpositive/udf_using.q | 8 ++---
.../infer_bucket_sort_dyn_part.q.out | 4 +--
.../infer_bucket_sort_map_operators.q.out | 8 ++---
.../llap/bucket_num_reducers_acid.q.out | 35 ++++++++++++++++++++
.../clientpositive/parallel_orderby.q.out | 10 +++---
.../root_dir_external_table.q.out | 28 ----------------
.../results/clientpositive/scriptfile1.q.out | 4 +--
.../test/results/clientpositive/udf_using.q.out | 11 +++---
14 files changed, 98 insertions(+), 82 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/itests/qtest/src/test/java/org/apache/hadoop/hive/cli/TestMinimrCliDriver.java
----------------------------------------------------------------------
diff --git a/itests/qtest/src/test/java/org/apache/hadoop/hive/cli/TestMinimrCliDriver.java b/itests/qtest/src/test/java/org/apache/hadoop/hive/cli/TestMinimrCliDriver.java
index 417687c..f525671 100644
--- a/itests/qtest/src/test/java/org/apache/hadoop/hive/cli/TestMinimrCliDriver.java
+++ b/itests/qtest/src/test/java/org/apache/hadoop/hive/cli/TestMinimrCliDriver.java
@@ -23,6 +23,7 @@ import java.util.List;
import org.apache.hadoop.hive.cli.control.CliAdapter;
import org.apache.hadoop.hive.cli.control.CliConfigs;
import org.junit.ClassRule;
+import org.junit.Ignore;
import org.junit.Rule;
import org.junit.Test;
import org.junit.rules.TestRule;
@@ -54,6 +55,7 @@ public class TestMinimrCliDriver {
this.qfile = qfile;
}
+ @Ignore
@Test
public void testCliDriver() throws Exception {
adapter.runTest(name, qfile);
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/itests/src/test/resources/testconfiguration.properties
----------------------------------------------------------------------
diff --git a/itests/src/test/resources/testconfiguration.properties b/itests/src/test/resources/testconfiguration.properties
index 6a70a4a..fa1a4fb 100644
--- a/itests/src/test/resources/testconfiguration.properties
+++ b/itests/src/test/resources/testconfiguration.properties
@@ -1,26 +1,7 @@
# Note: the *.shared groups also run on TestCliDriver
-# NOTE: files should be listed in alphabetical order
-minimr.query.files=infer_bucket_sort_map_operators.q,\
- infer_bucket_sort_dyn_part.q,\
- infer_bucket_sort_merge.q,\
- infer_bucket_sort_reducers_power_two.q,\
- infer_bucket_sort_num_buckets.q,\
- root_dir_external_table.q,\
- parallel_orderby.q,\
- udf_using.q,\
- index_bitmap3.q,\
- index_bitmap_auto.q,\
- scriptfile1.q,\
- bucket_num_reducers_acid.q,\
- scriptfile1_win.q
-
-# These tests are disabled for minimr
-# ql_rewrite_gbtoidx.q,\
-# ql_rewrite_gbtoidx_cbo_1.q,\
-# ql_rewrite_gbtoidx_cbo_2.q,\
-# smb_mapjoin_8.q,\
-
+# DO NOT USE minimr, as MR is deprecated and MinimrCliDriver will be removed
+minimr.query.files=doesnotexist.q\
# Tests that are not enabled for CLI Driver
disabled.query.files=ql_rewrite_gbtoidx.q,\
@@ -37,7 +18,8 @@ disabled.query.files=ql_rewrite_gbtoidx.q,\
union_stats.q,\
sample2.q,\
sample4.q,\
- sample6.q
+ sample6.q,\
+ root_dir_external_table.q
# NOTE: Add tests to minitez only if it is very
@@ -433,6 +415,7 @@ minillap.query.files=acid_bucket_pruning.q,\
cttl.q
minillaplocal.query.files=\
+ bucket_num_reducers_acid.q,\
dp_counter_non_mm.q,\
dp_counter_mm.q,\
acid_no_buckets.q, \
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/itests/util/pom.xml
----------------------------------------------------------------------
diff --git a/itests/util/pom.xml b/itests/util/pom.xml
index 7b70560..9334f90 100644
--- a/itests/util/pom.xml
+++ b/itests/util/pom.xml
@@ -176,6 +176,11 @@
<artifactId>hbase-mapreduce</artifactId>
<version>${hbase.version}</version>
</dependency>
+ <dependency>
+ <groupId>org.apache.tez</groupId>
+ <artifactId>tez-api</artifactId>
+ <version>${tez.version}</version>
+ </dependency>
<!-- test inter-project -->
<dependency>
<groupId>junit</groupId>
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/VerifyNumReducersHook.java
----------------------------------------------------------------------
diff --git a/itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/VerifyNumReducersHook.java b/itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/VerifyNumReducersHook.java
index 6172ca9..a9a25f8 100644
--- a/itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/VerifyNumReducersHook.java
+++ b/itests/util/src/main/java/org/apache/hadoop/hive/ql/hooks/VerifyNumReducersHook.java
@@ -21,9 +21,13 @@ import java.util.Map;
import junit.framework.Assert;
+import org.apache.hadoop.hive.conf.HiveConf;
import org.apache.hadoop.hive.ql.MapRedStats;
+import org.apache.hadoop.hive.ql.exec.tez.monitoring.TezProgressMonitor;
import org.apache.hadoop.hive.ql.session.SessionState;
+import org.apache.tez.dag.api.client.Progress;
+
/**
*
* VerifyNumReducersHook.
@@ -42,10 +46,23 @@ public class VerifyNumReducersHook implements ExecuteWithHookContext {
Assert.assertNotNull("SessionState returned null");
int expectedReducers = hookContext.getConf().getInt(BUCKET_CONFIG, 0);
- Map<String, MapRedStats> stats = ss.getMapRedStats();
- Assert.assertEquals("Number of MapReduce jobs is incorrect", 1, stats.size());
+ if (ss.getConf().get(HiveConf.ConfVars.HIVE_EXECUTION_ENGINE.varname).equals("mr")) {
+ Map<String, MapRedStats> stats = ss.getMapRedStats();
+ Assert.assertEquals("Number of MapReduce jobs is incorrect", 1, stats.size());
- MapRedStats stat = stats.values().iterator().next();
- Assert.assertEquals("NumReducers is incorrect", expectedReducers, stat.getNumReduce());
+ MapRedStats stat = stats.values().iterator().next();
+ Assert.assertEquals("NumReducers is incorrect", expectedReducers, stat.getNumReduce());
+ } else if (ss.getConf().get(HiveConf.ConfVars.HIVE_EXECUTION_ENGINE.varname).equals("tez")) {
+ TezProgressMonitor tezProgressMonitor = (TezProgressMonitor) ss.getProgressMonitor();
+ Map<String, Progress> progressMap = tezProgressMonitor.getStatus().getVertexProgress();
+ int totalReducers = 0;
+ for (Map.Entry<String, Progress> entry : progressMap.entrySet()) {
+ // relying on the name of vertex is fragile, but this will do for now for the tests
+ if (entry.getKey().contains("Reducer")) {
+ totalReducers += entry.getValue().getTotalTaskCount();
+ }
+ }
+ Assert.assertEquals("Number of reducers is incorrect", expectedReducers, totalReducers);
+ }
}
}
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/monitoring/TezProgressMonitor.java
----------------------------------------------------------------------
diff --git a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/monitoring/TezProgressMonitor.java b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/monitoring/TezProgressMonitor.java
index b0c1659..735442d 100644
--- a/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/monitoring/TezProgressMonitor.java
+++ b/ql/src/java/org/apache/hadoop/hive/ql/exec/tez/monitoring/TezProgressMonitor.java
@@ -35,12 +35,10 @@ import java.util.HashSet;
import java.util.List;
import java.util.Map;
import java.util.Set;
-import java.util.SortedSet;
-import java.util.TreeSet;
import static org.apache.tez.dag.api.client.DAGStatus.State.KILLED;
-class TezProgressMonitor implements ProgressMonitor {
+public class TezProgressMonitor implements ProgressMonitor {
private static final int COLUMN_1_WIDTH = 16;
private final List<BaseWork> topSortedWork;
private final SessionState.LogHelper console;
@@ -211,7 +209,7 @@ class TezProgressMonitor implements ProgressMonitor {
return mode;
}
- static class VertexProgress {
+ public static class VertexProgress {
private final int totalTaskCount;
private final int succeededTaskCount;
private final int failedTaskAttemptCount;
@@ -328,4 +326,8 @@ class TezProgressMonitor implements ProgressMonitor {
return result;
}
}
+
+ public DAGStatus getStatus() {
+ return status;
+ }
}
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/test/queries/clientpositive/root_dir_external_table.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/root_dir_external_table.q b/ql/src/test/queries/clientpositive/root_dir_external_table.q
index 7763ce2..66a50e6 100644
--- a/ql/src/test/queries/clientpositive/root_dir_external_table.q
+++ b/ql/src/test/queries/clientpositive/root_dir_external_table.q
@@ -1,5 +1,6 @@
--! qt:dataset:src
set hive.mapred.mode=nonstrict;
+set hive.llap.io.enabled=false;
dfs ${system:test.dfs.mkdir} hdfs:///tmp/test_root_dir_external_table;
insert overwrite directory "hdfs:///tmp/test_root_dir_external_table" select key from src where (key < 20) order by key;
@@ -10,4 +11,4 @@ dfs -rmr hdfs:///tmp/test_root_dir_external_table;
create external table roottable_n0 (key string) row format delimited fields terminated by '\\t' stored as textfile location 'hdfs:///';
select count(*) from roottable_n0;
-dfs -rmr /000000_0;
\ No newline at end of file
+dfs -rmr /000000_0;
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/test/queries/clientpositive/udf_using.q
----------------------------------------------------------------------
diff --git a/ql/src/test/queries/clientpositive/udf_using.q b/ql/src/test/queries/clientpositive/udf_using.q
index d5e9206..732f965 100644
--- a/ql/src/test/queries/clientpositive/udf_using.q
+++ b/ql/src/test/queries/clientpositive/udf_using.q
@@ -1,9 +1,9 @@
--! qt:dataset:src
-dfs ${system:test.dfs.mkdir} hdfs:///tmp/udf_using;
+dfs ${system:test.dfs.mkdir} ${system:test.tmp.dir}/udf_using;
-dfs -copyFromLocal ../../data/files/sales.txt hdfs:///tmp/udf_using/sales.txt;
+dfs -copyFromLocal ../../data/files/sales.txt ${system:test.tmp.dir}/udf_using/sales.txt;
-create function lookup as 'org.apache.hadoop.hive.ql.udf.UDFFileLookup' using file 'hdfs:///tmp/udf_using/sales.txt';
+create function lookup as 'org.apache.hadoop.hive.ql.udf.UDFFileLookup' using file '${system:test.tmp.dir}/udf_using/sales.txt';
create table udf_using (c1 string);
insert overwrite table udf_using select 'Joe' from src limit 2;
@@ -13,4 +13,4 @@ select c1, lookup(c1) from udf_using;
drop table udf_using;
drop function lookup;
-dfs -rmr hdfs:///tmp/udf_using;
+dfs -rmr ${system:test.tmp.dir}/udf_using;
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out
index 1af3ea5..8d813a4 100644
--- a/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out
+++ b/ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out
@@ -488,7 +488,7 @@ STAGE PLANS:
Move Operator
files:
hdfs directory: true
- destination: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
Stage: Stage-0
Move Operator
@@ -529,7 +529,7 @@ STAGE PLANS:
Move Operator
files:
hdfs directory: true
- destination: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
Stage: Stage-8
Map Reduce
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out b/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out
index 63f6f5c..7b790cc 100644
--- a/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out
+++ b/ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out
@@ -133,7 +133,7 @@ STAGE PLANS:
Move Operator
files:
hdfs directory: true
- destination: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
Stage: Stage-0
Move Operator
@@ -183,7 +183,7 @@ STAGE PLANS:
Move Operator
files:
hdfs directory: true
- destination: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
PREHOOK: query: INSERT OVERWRITE TABLE test_table_out_n0 PARTITION (part = '1')
SELECT key, count(*) FROM test_table1_n14 GROUP BY key
@@ -526,7 +526,7 @@ STAGE PLANS:
Move Operator
files:
hdfs directory: true
- destination: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
Stage: Stage-0
Move Operator
@@ -576,7 +576,7 @@ STAGE PLANS:
Move Operator
files:
hdfs directory: true
- destination: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
PREHOOK: query: INSERT OVERWRITE TABLE test_table_out_n0 PARTITION (part = '1')
SELECT /*+ MAPJOIN(a) */ a.key, b.value FROM test_table1_n14 a JOIN test_table2_n13 b ON a.key = b.key
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/test/results/clientpositive/llap/bucket_num_reducers_acid.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/llap/bucket_num_reducers_acid.q.out b/ql/src/test/results/clientpositive/llap/bucket_num_reducers_acid.q.out
new file mode 100644
index 0000000..81ba5ae
--- /dev/null
+++ b/ql/src/test/results/clientpositive/llap/bucket_num_reducers_acid.q.out
@@ -0,0 +1,35 @@
+PREHOOK: query: drop table if exists bucket_nr_acid
+PREHOOK: type: DROPTABLE
+POSTHOOK: query: drop table if exists bucket_nr_acid
+POSTHOOK: type: DROPTABLE
+PREHOOK: query: create table bucket_nr_acid (a int, b int) clustered by (a) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true')
+PREHOOK: type: CREATETABLE
+PREHOOK: Output: database:default
+PREHOOK: Output: default@bucket_nr_acid
+POSTHOOK: query: create table bucket_nr_acid (a int, b int) clustered by (a) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true')
+POSTHOOK: type: CREATETABLE
+POSTHOOK: Output: database:default
+POSTHOOK: Output: default@bucket_nr_acid
+PREHOOK: query: insert into bucket_nr_acid values(1,1)
+PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
+PREHOOK: Output: default@bucket_nr_acid
+PREHOOK: query: insert into bucket_nr_acid values(0,0),(3,3)
+PREHOOK: type: QUERY
+PREHOOK: Input: _dummy_database@_dummy_table
+PREHOOK: Output: default@bucket_nr_acid
+PREHOOK: query: update bucket_nr_acid set b = -1
+PREHOOK: type: QUERY
+PREHOOK: Input: default@bucket_nr_acid
+PREHOOK: Output: default@bucket_nr_acid
+PREHOOK: query: select * from bucket_nr_acid order by a, b
+PREHOOK: type: QUERY
+PREHOOK: Input: default@bucket_nr_acid
+#### A masked pattern was here ####
+0 -1
+1 -1
+3 -1
+PREHOOK: query: drop table bucket_nr_acid
+PREHOOK: type: DROPTABLE
+PREHOOK: Input: default@bucket_nr_acid
+PREHOOK: Output: default@bucket_nr_acid
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/test/results/clientpositive/parallel_orderby.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/parallel_orderby.q.out b/ql/src/test/results/clientpositive/parallel_orderby.q.out
index 73a103d..d60a0a6 100644
--- a/ql/src/test/results/clientpositive/parallel_orderby.q.out
+++ b/ql/src/test/results/clientpositive/parallel_orderby.q.out
@@ -68,7 +68,7 @@ STAGE PLANS:
Move Operator
files:
hdfs directory: true
- destination: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
Stage: Stage-3
Create Table Operator:
@@ -134,11 +134,11 @@ Storage Desc Params:
PREHOOK: query: select * from total_ordered
PREHOOK: type: QUERY
PREHOOK: Input: default@total_ordered
-PREHOOK: Output: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
POSTHOOK: query: select * from total_ordered
POSTHOOK: type: QUERY
POSTHOOK: Input: default@total_ordered
-POSTHOOK: Output: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
128 val_128
128 val_128
150 val_150
@@ -245,11 +245,11 @@ Storage Desc Params:
PREHOOK: query: select * from total_ordered
PREHOOK: type: QUERY
PREHOOK: Input: default@total_ordered
-PREHOOK: Output: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
POSTHOOK: query: select * from total_ordered
POSTHOOK: type: QUERY
POSTHOOK: Input: default@total_ordered
-POSTHOOK: Output: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
128 val_128
128 val_128
150 val_150
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/test/results/clientpositive/root_dir_external_table.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/root_dir_external_table.q.out b/ql/src/test/results/clientpositive/root_dir_external_table.q.out
deleted file mode 100644
index 7049406..0000000
--- a/ql/src/test/results/clientpositive/root_dir_external_table.q.out
+++ /dev/null
@@ -1,28 +0,0 @@
-PREHOOK: query: insert overwrite directory "hdfs://### HDFS PATH ###" select key from src where (key < 20) order by key
-PREHOOK: type: QUERY
-PREHOOK: Input: default@src
-#### A masked pattern was here ####
-POSTHOOK: query: insert overwrite directory "hdfs://### HDFS PATH ###" select key from src where (key < 20) order by key
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@src
-#### A masked pattern was here ####
-PREHOOK: query: create external table roottable_n0 (key string) row format delimited fields terminated by '\\t' stored as textfile location 'hdfs://### HDFS PATH ###'
-PREHOOK: type: CREATETABLE
-PREHOOK: Input: hdfs://### HDFS PATH ###
-PREHOOK: Output: database:default
-PREHOOK: Output: default@roottable_n0
-POSTHOOK: query: create external table roottable_n0 (key string) row format delimited fields terminated by '\\t' stored as textfile location 'hdfs://### HDFS PATH ###'
-POSTHOOK: type: CREATETABLE
-POSTHOOK: Input: hdfs://### HDFS PATH ###
-POSTHOOK: Output: database:default
-POSTHOOK: Output: default@roottable_n0
-PREHOOK: query: select count(*) from roottable_n0
-PREHOOK: type: QUERY
-PREHOOK: Input: default@roottable_n0
-PREHOOK: Output: hdfs://### HDFS PATH ###
-POSTHOOK: query: select count(*) from roottable_n0
-POSTHOOK: type: QUERY
-POSTHOOK: Input: default@roottable_n0
-POSTHOOK: Output: hdfs://### HDFS PATH ###
-20
-#### A masked pattern was here ####
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/test/results/clientpositive/scriptfile1.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/scriptfile1.q.out b/ql/src/test/results/clientpositive/scriptfile1.q.out
index f374501..29ee2df 100644
--- a/ql/src/test/results/clientpositive/scriptfile1.q.out
+++ b/ql/src/test/results/clientpositive/scriptfile1.q.out
@@ -31,11 +31,11 @@ POSTHOOK: Lineage: dest1_n22.value SCRIPT [(src)src.FieldSchema(name:key, type:s
PREHOOK: query: SELECT dest1_n22.* FROM dest1_n22
PREHOOK: type: QUERY
PREHOOK: Input: default@dest1_n22
-PREHOOK: Output: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
POSTHOOK: query: SELECT dest1_n22.* FROM dest1_n22
POSTHOOK: type: QUERY
POSTHOOK: Input: default@dest1_n22
-POSTHOOK: Output: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
10 val_10
100 val_100
100 val_100
http://git-wip-us.apache.org/repos/asf/hive/blob/8cb99d68/ql/src/test/results/clientpositive/udf_using.q.out
----------------------------------------------------------------------
diff --git a/ql/src/test/results/clientpositive/udf_using.q.out b/ql/src/test/results/clientpositive/udf_using.q.out
index 8f92101..a226b79 100644
--- a/ql/src/test/results/clientpositive/udf_using.q.out
+++ b/ql/src/test/results/clientpositive/udf_using.q.out
@@ -1,13 +1,12 @@
-PREHOOK: query: create function lookup as 'org.apache.hadoop.hive.ql.udf.UDFFileLookup' using file 'hdfs://### HDFS PATH ###'
+#### A masked pattern was here ####
PREHOOK: type: CREATEFUNCTION
PREHOOK: Output: database:default
PREHOOK: Output: default.lookup
-PREHOOK: Output: hdfs://### HDFS PATH ###
-POSTHOOK: query: create function lookup as 'org.apache.hadoop.hive.ql.udf.UDFFileLookup' using file 'hdfs://### HDFS PATH ###'
+#### A masked pattern was here ####
POSTHOOK: type: CREATEFUNCTION
POSTHOOK: Output: database:default
POSTHOOK: Output: default.lookup
-POSTHOOK: Output: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
PREHOOK: query: create table udf_using (c1 string)
PREHOOK: type: CREATETABLE
PREHOOK: Output: database:default
@@ -28,11 +27,11 @@ POSTHOOK: Lineage: udf_using.c1 SIMPLE []
PREHOOK: query: select c1, lookup(c1) from udf_using
PREHOOK: type: QUERY
PREHOOK: Input: default@udf_using
-PREHOOK: Output: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
POSTHOOK: query: select c1, lookup(c1) from udf_using
POSTHOOK: type: QUERY
POSTHOOK: Input: default@udf_using
-POSTHOOK: Output: hdfs://### HDFS PATH ###
+#### A masked pattern was here ####
Joe 2
Joe 2
PREHOOK: query: drop table udf_using