You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kylin.apache.org by ma...@apache.org on 2015/05/26 07:37:40 UTC

[6/6] incubator-kylin git commit: KYLIN-793 modify all xmls

KYLIN-793 modify all xmls


Project: http://git-wip-us.apache.org/repos/asf/incubator-kylin/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-kylin/commit/f75658cf
Tree: http://git-wip-us.apache.org/repos/asf/incubator-kylin/tree/f75658cf
Diff: http://git-wip-us.apache.org/repos/asf/incubator-kylin/diff/f75658cf

Branch: refs/heads/new697
Commit: f75658cf86637e685f1b43d7c600535f631f529a
Parents: bebb32d
Author: honma <ho...@ebay.com>
Authored: Tue May 26 11:00:26 2015 +0800
Committer: honma <ho...@ebay.com>
Committed: Tue May 26 13:42:11 2015 +0800

----------------------------------------------------------------------
 .../sandbox/capacity-scheduler.xml              | 121 ++-
 examples/test_case_data/sandbox/core-site.xml   | 182 ++--
 .../test_case_data/sandbox/hadoop-policy.xml    |  74 +-
 .../test_case_data/sandbox/hbase-policy.xml     |  26 +-
 examples/test_case_data/sandbox/hbase-site.xml  | 232 +++--
 examples/test_case_data/sandbox/hdfs-site.xml   | 318 +++----
 examples/test_case_data/sandbox/hive-site.xml   | 953 +++++++++----------
 examples/test_case_data/sandbox/mapred-site.xml | 286 +++---
 examples/test_case_data/sandbox/yarn-site.xml   | 609 ++++++------
 9 files changed, 1323 insertions(+), 1478 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/f75658cf/examples/test_case_data/sandbox/capacity-scheduler.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox/capacity-scheduler.xml b/examples/test_case_data/sandbox/capacity-scheduler.xml
index 7b4a367..9c97e43 100644
--- a/examples/test_case_data/sandbox/capacity-scheduler.xml
+++ b/examples/test_case_data/sandbox/capacity-scheduler.xml
@@ -1,99 +1,94 @@
-<!--Tue Dec 16 19:08:08 2014-->
-    <configuration>
-    
+<!--Mon May 25 01:30:55 2015-->
+<configuration>
+
     <property>
-      <name>yarn.scheduler.capacity.default.minimum-user-limit-percent</name>
-      <value>100</value>
+        <name>yarn.scheduler.capacity.default.minimum-user-limit-percent</name>
+        <value>100</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.maximum-am-resource-percent</name>
-      <value>0.5</value>
+        <name>yarn.scheduler.capacity.maximum-am-resource-percent</name>
+        <value>0.2</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.maximum-applications</name>
-      <value>10000</value>
+        <name>yarn.scheduler.capacity.maximum-applications</name>
+        <value>10000</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.node-locality-delay</name>
-      <value>40</value>
+        <name>yarn.scheduler.capacity.node-locality-delay</name>
+        <value>40</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.resource-calculator</name>
-      <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value>
+        <name>yarn.scheduler.capacity.resource-calculator</name>
+        <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.accessible-node-labels</name>
-      <value>*</value>
+        <name>yarn.scheduler.capacity.root.accessible-node-labels</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.accessible-node-labels.default.capacity</name>
-      <value>-1</value>
+        <name>yarn.scheduler.capacity.root.accessible-node-labels.default.capacity</name>
+        <value>-1</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.accessible-node-labels.default.maximum-capacity</name>
-      <value>-1</value>
+        <name>yarn.scheduler.capacity.root.accessible-node-labels.default.maximum-capacity</name>
+        <value>-1</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.acl_administer_queue</name>
-      <value>*</value>
+        <name>yarn.scheduler.capacity.root.acl_administer_queue</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.capacity</name>
-      <value>100</value>
+        <name>yarn.scheduler.capacity.root.capacity</name>
+        <value>100</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.default-node-label-expression</name>
-      <value> </value>
+        <name>yarn.scheduler.capacity.root.default-node-label-expression</name>
+        <value> </value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.default.acl_administer_jobs</name>
-      <value>*</value>
+        <name>yarn.scheduler.capacity.root.default.acl_administer_jobs</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.default.acl_submit_applications</name>
-      <value>*</value>
+        <name>yarn.scheduler.capacity.root.default.acl_submit_applications</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.default.capacity</name>
-      <value>100</value>
+        <name>yarn.scheduler.capacity.root.default.capacity</name>
+        <value>100</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.default.maximum-am-resource-percent</name>
-      <value>0.5</value>
+        <name>yarn.scheduler.capacity.root.default.maximum-capacity</name>
+        <value>100</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.default.maximum-capacity</name>
-      <value>100</value>
+        <name>yarn.scheduler.capacity.root.default.state</name>
+        <value>RUNNING</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.default.state</name>
-      <value>RUNNING</value>
+        <name>yarn.scheduler.capacity.root.default.user-limit-factor</name>
+        <value>1</value>
     </property>
-    
+
     <property>
-      <name>yarn.scheduler.capacity.root.default.user-limit-factor</name>
-      <value>1</value>
+        <name>yarn.scheduler.capacity.root.queues</name>
+        <value>default</value>
     </property>
-    
-    <property>
-      <name>yarn.scheduler.capacity.root.queues</name>
-      <value>default</value>
-    </property>
-    
-  </configuration>
+
+</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/f75658cf/examples/test_case_data/sandbox/core-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox/core-site.xml b/examples/test_case_data/sandbox/core-site.xml
index 1b3a3ee..e2b3fbe 100644
--- a/examples/test_case_data/sandbox/core-site.xml
+++ b/examples/test_case_data/sandbox/core-site.xml
@@ -1,152 +1,106 @@
-<!--Mon Jan 19 07:28:25 2015-->
-    <configuration>
-    
+<!--Mon May 25 01:30:55 2015-->
+<configuration>
+
     <property>
-      <name>fs.defaultFS</name>
-      <value>hdfs://sandbox.hortonworks.com:8020</value>
-      <final>true</final>
+        <name>fs.defaultFS</name>
+        <value>hdfs://sandbox.hortonworks.com:8020</value>
+        <final>true</final>
     </property>
-    
+
     <property>
-      <name>fs.trash.interval</name>
-      <value>360</value>
+        <name>fs.trash.interval</name>
+        <value>360</value>
     </property>
-    
+
     <property>
-      <name>hadoop.http.authentication.simple.anonymous.allowed</name>
-      <value>true</value>
+        <name>hadoop.http.authentication.simple.anonymous.allowed</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.falcon.groups</name>
-      <value>users</value>
+        <name>hadoop.proxyuser.hcat.groups</name>
+        <value>users</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.falcon.hosts</name>
-      <value>*</value>
+        <name>hadoop.proxyuser.hcat.hosts</name>
+        <value>sandbox.hortonworks.com</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.hbase.groups</name>
-      <value>users</value>
+        <name>hadoop.proxyuser.hive.groups</name>
+        <value>users</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.hbase.hosts</name>
-      <value>*</value>
+        <name>hadoop.proxyuser.hive.hosts</name>
+        <value>sandbox.hortonworks.com</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.hcat.groups</name>
-      <value>*</value>
+        <name>hadoop.security.auth_to_local</name>
+        <value>
+            RULE:[2:$1@$0]([rn]m@.*)s/.*/yarn/
+            RULE:[2:$1@$0](jhs@.*)s/.*/mapred/
+            RULE:[2:$1@$0]([nd]n@.*)s/.*/hdfs/
+            RULE:[2:$1@$0](hm@.*)s/.*/hbase/
+            RULE:[2:$1@$0](rs@.*)s/.*/hbase/
+            DEFAULT</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.hcat.hosts</name>
-      <value>sandbox.hortonworks.com</value>
+        <name>hadoop.security.authentication</name>
+        <value>simple</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.hive.groups</name>
-      <value>users</value>
+        <name>hadoop.security.authorization</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.hive.hosts</name>
-      <value>*</value>
+        <name>io.compression.codecs</name>
+        <value>org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.SnappyCodec</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.hue.groups</name>
-      <value>*</value>
+        <name>io.file.buffer.size</name>
+        <value>131072</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.hue.hosts</name>
-      <value>*</value>
+        <name>io.serializations</name>
+        <value>org.apache.hadoop.io.serializer.WritableSerialization</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.oozie.groups</name>
-      <value>*</value>
+        <name>ipc.client.connect.max.retries</name>
+        <value>50</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.oozie.hosts</name>
-      <value>sandbox.hortonworks.com</value>
+        <name>ipc.client.connection.maxidletime</name>
+        <value>30000</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.root.groups</name>
-      <value>*</value>
+        <name>ipc.client.idlethreshold</name>
+        <value>8000</value>
     </property>
-    
+
     <property>
-      <name>hadoop.proxyuser.root.hosts</name>
-      <value>*</value>
+        <name>ipc.server.tcpnodelay</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hadoop.security.auth_to_local</name>
-      <value>
-        RULE:[2:$1@$0]([rn]m@.*)s/.*/yarn/
-        RULE:[2:$1@$0](jhs@.*)s/.*/mapred/
-        RULE:[2:$1@$0]([nd]n@.*)s/.*/hdfs/
-        RULE:[2:$1@$0](hm@.*)s/.*/hbase/
-        RULE:[2:$1@$0](rs@.*)s/.*/hbase/
-        DEFAULT
-    </value>
+        <name>mapreduce.jobtracker.webinterface.trusted</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hadoop.security.authentication</name>
-      <value>simple</value>
+        <name>proxyuser_group</name>
+        <value>users</value>
     </property>
-    
-    <property>
-      <name>hadoop.security.authorization</name>
-      <value>false</value>
-    </property>
-    
-    <property>
-      <name>io.compression.codecs</name>
-      <value>org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.SnappyCodec</value>
-    </property>
-    
-    <property>
-      <name>io.file.buffer.size</name>
-      <value>131072</value>
-    </property>
-    
-    <property>
-      <name>io.serializations</name>
-      <value>org.apache.hadoop.io.serializer.WritableSerialization</value>
-    </property>
-    
-    <property>
-      <name>ipc.client.connect.max.retries</name>
-      <value>50</value>
-    </property>
-    
-    <property>
-      <name>ipc.client.connection.maxidletime</name>
-      <value>30000</value>
-    </property>
-    
-    <property>
-      <name>ipc.client.idlethreshold</name>
-      <value>8000</value>
-    </property>
-    
-    <property>
-      <name>ipc.server.tcpnodelay</name>
-      <value>true</value>
-    </property>
-    
-    <property>
-      <name>mapreduce.jobtracker.webinterface.trusted</name>
-      <value>false</value>
-    </property>
-    
-  </configuration>
+
+</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/f75658cf/examples/test_case_data/sandbox/hadoop-policy.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox/hadoop-policy.xml b/examples/test_case_data/sandbox/hadoop-policy.xml
index 0f7e0b9..3f08465 100644
--- a/examples/test_case_data/sandbox/hadoop-policy.xml
+++ b/examples/test_case_data/sandbox/hadoop-policy.xml
@@ -1,59 +1,59 @@
-<!--Tue Dec 16 19:07:40 2014-->
-    <configuration>
-    
+<!--Mon May 25 01:30:15 2015-->
+<configuration>
+
     <property>
-      <name>security.admin.operations.protocol.acl</name>
-      <value>hadoop</value>
+        <name>security.admin.operations.protocol.acl</name>
+        <value>hadoop</value>
     </property>
-    
+
     <property>
-      <name>security.client.datanode.protocol.acl</name>
-      <value>*</value>
+        <name>security.client.datanode.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.client.protocol.acl</name>
-      <value>*</value>
+        <name>security.client.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.datanode.protocol.acl</name>
-      <value>*</value>
+        <name>security.datanode.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.inter.datanode.protocol.acl</name>
-      <value>*</value>
+        <name>security.inter.datanode.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.inter.tracker.protocol.acl</name>
-      <value>*</value>
+        <name>security.inter.tracker.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.job.client.protocol.acl</name>
-      <value>*</value>
+        <name>security.job.client.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.job.task.protocol.acl</name>
-      <value>*</value>
+        <name>security.job.task.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.namenode.protocol.acl</name>
-      <value>*</value>
+        <name>security.namenode.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.refresh.policy.protocol.acl</name>
-      <value>hadoop</value>
+        <name>security.refresh.policy.protocol.acl</name>
+        <value>hadoop</value>
     </property>
-    
+
     <property>
-      <name>security.refresh.usertogroups.mappings.protocol.acl</name>
-      <value>hadoop</value>
+        <name>security.refresh.usertogroups.mappings.protocol.acl</name>
+        <value>hadoop</value>
     </property>
-    
-  </configuration>
+
+</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/f75658cf/examples/test_case_data/sandbox/hbase-policy.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox/hbase-policy.xml b/examples/test_case_data/sandbox/hbase-policy.xml
index 36756b8..6549239 100644
--- a/examples/test_case_data/sandbox/hbase-policy.xml
+++ b/examples/test_case_data/sandbox/hbase-policy.xml
@@ -1,19 +1,19 @@
-<!--Mon Jan 19 07:29:07 2015-->
-    <configuration>
-    
+<!--Mon May 25 01:30:31 2015-->
+<configuration>
+
     <property>
-      <name>security.admin.protocol.acl</name>
-      <value>*</value>
+        <name>security.admin.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.client.protocol.acl</name>
-      <value>*</value>
+        <name>security.client.protocol.acl</name>
+        <value>*</value>
     </property>
-    
+
     <property>
-      <name>security.masterregion.protocol.acl</name>
-      <value>*</value>
+        <name>security.masterregion.protocol.acl</name>
+        <value>*</value>
     </property>
-    
-  </configuration>
+
+</configuration>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/f75658cf/examples/test_case_data/sandbox/hbase-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox/hbase-site.xml b/examples/test_case_data/sandbox/hbase-site.xml
index 9fd97c9..b63cc90 100644
--- a/examples/test_case_data/sandbox/hbase-site.xml
+++ b/examples/test_case_data/sandbox/hbase-site.xml
@@ -1,194 +1,174 @@
-<!--Mon Jan 19 07:29:07 2015-->
-    <configuration>
-    
+<!--Mon May 25 01:30:31 2015-->
+<configuration>
+
     <property>
-      <name>dfs.domain.socket.path</name>
-      <value>/var/lib/hadoop-hdfs/dn_socket</value>
+        <name>dfs.domain.socket.path</name>
+        <value>/var/lib/hadoop-hdfs/dn_socket</value>
     </property>
-    
+
     <property>
-      <name>hbase.client.keyvalue.maxsize</name>
-      <value>10485760</value>
+        <name>hbase.client.keyvalue.maxsize</name>
+        <value>10485760</value>
     </property>
-    
+
     <property>
-      <name>hbase.client.scanner.caching</name>
-      <value>100</value>
+        <name>hbase.client.scanner.caching</name>
+        <value>100</value>
     </property>
-    
+
     <property>
-      <name>hbase.cluster.distributed</name>
-      <value>true</value>
+        <name>hbase.cluster.distributed</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hbase.coprocessor.master.classes</name>
-      <value>com.xasecure.authorization.hbase.XaSecureAuthorizationCoprocessor</value>
+        <name>hbase.coprocessor.master.classes</name>
+        <value></value>
     </property>
-    
+
     <property>
-      <name>hbase.coprocessor.region.classes</name>
-      <value>com.xasecure.authorization.hbase.XaSecureAuthorizationCoprocessor</value>
+        <name>hbase.coprocessor.region.classes</name>
+        <value></value>
     </property>
-    
+
     <property>
-      <name>hbase.defaults.for.version.skip</name>
-      <value>true</value>
+        <name>hbase.defaults.for.version.skip</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hbase.hregion.majorcompaction</name>
-      <value>604800000</value>
+        <name>hbase.hregion.majorcompaction</name>
+        <value>604800000</value>
     </property>
-    
+
     <property>
-      <name>hbase.hregion.majorcompaction.jitter</name>
-      <value>0.50</value>
+        <name>hbase.hregion.majorcompaction.jitter</name>
+        <value>0.50</value>
     </property>
-    
+
     <property>
-      <name>hbase.hregion.max.filesize</name>
-      <value>10737418240</value>
+        <name>hbase.hregion.max.filesize</name>
+        <value>10737418240</value>
     </property>
-    
+
     <property>
-      <name>hbase.hregion.memstore.block.multiplier</name>
-      <value>4</value>
+        <name>hbase.hregion.memstore.block.multiplier</name>
+        <value>4</value>
     </property>
-    
+
     <property>
-      <name>hbase.hregion.memstore.flush.size</name>
-      <value>134217728</value>
+        <name>hbase.hregion.memstore.flush.size</name>
+        <value>134217728</value>
     </property>
-    
+
     <property>
-      <name>hbase.hregion.memstore.mslab.enabled</name>
-      <value>true</value>
+        <name>hbase.hregion.memstore.mslab.enabled</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hbase.hstore.blockingStoreFiles</name>
-      <value>10</value>
+        <name>hbase.hstore.blockingStoreFiles</name>
+        <value>10</value>
     </property>
-    
+
     <property>
-      <name>hbase.hstore.compactionThreshold</name>
-      <value>3</value>
+        <name>hbase.hstore.compactionThreshold</name>
+        <value>3</value>
     </property>
-    
+
     <property>
-      <name>hbase.local.dir</name>
-      <value>${hbase.tmp.dir}/local</value>
+        <name>hbase.local.dir</name>
+        <value>${hbase.tmp.dir}/local</value>
     </property>
-    
+
     <property>
-      <name>hbase.master.info.bindAddress</name>
-      <value>0.0.0.0</value>
+        <name>hbase.master.info.bindAddress</name>
+        <value>0.0.0.0</value>
     </property>
-    
+
     <property>
-      <name>hbase.master.info.port</name>
-      <value>60010</value>
+        <name>hbase.master.info.port</name>
+        <value>60010</value>
     </property>
-    
+
     <property>
-      <name>hbase.master.port</name>
-      <value>60000</value>
+        <name>hbase.master.port</name>
+        <value>60000</value>
     </property>
-    
+
     <property>
-      <name>hbase.regionserver.global.memstore.lowerLimit</name>
-      <value>0.38</value>
+        <name>hbase.regionserver.global.memstore.lowerLimit</name>
+        <value>0.38</value>
     </property>
-    
+
     <property>
-      <name>hbase.regionserver.global.memstore.upperLimit</name>
-      <value>0.4</value>
+        <name>hbase.regionserver.global.memstore.upperLimit</name>
+        <value>0.4</value>
     </property>
-    
+
     <property>
-      <name>hbase.regionserver.handler.count</name>
-      <value>60</value>
+        <name>hbase.regionserver.handler.count</name>
+        <value>60</value>
     </property>
-    
+
     <property>
-      <name>hbase.regionserver.info.port</name>
-      <value>60030</value>
+        <name>hbase.regionserver.info.port</name>
+        <value>60030</value>
     </property>
-    
+
     <property>
-      <name>hbase.rootdir</name>
-      <value>hdfs://sandbox.hortonworks.com:8020/apps/hbase/data</value>
+        <name>hbase.rootdir</name>
+        <value>hdfs://sandbox.hortonworks.com:8020/apps/hbase/data</value>
     </property>
-    
+
     <property>
-      <name>hbase.rpc.engine</name>
-      <value>org.apache.hadoop.hbase.ipc.SecureRpcEngine</value>
+        <name>hbase.security.authentication</name>
+        <value>simple</value>
     </property>
-    
+
     <property>
-      <name>hbase.rpc.protection</name>
-      <value>PRIVACY</value>
+        <name>hbase.security.authorization</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hbase.security.authentication</name>
-      <value>simple</value>
+        <name>hbase.superuser</name>
+        <value>hbase</value>
     </property>
-    
+
     <property>
-      <name>hbase.security.authorization</name>
-      <value>true</value>
+        <name>hbase.tmp.dir</name>
+        <value>/hadoop/hbase</value>
     </property>
-    
+
     <property>
-      <name>hbase.superuser</name>
-      <value>hbase</value>
+        <name>hbase.zookeeper.property.clientPort</name>
+        <value>2181</value>
     </property>
-    
+
     <property>
-      <name>hbase.tmp.dir</name>
-      <value>/hadoop/hbase</value>
+        <name>hbase.zookeeper.quorum</name>
+        <value>sandbox.hortonworks.com</value>
     </property>
-    
+
     <property>
-      <name>hbase.zookeeper.property.clientPort</name>
-      <value>2181</value>
+        <name>hbase.zookeeper.useMulti</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hbase.zookeeper.quorum</name>
-      <value>sandbox.hortonworks.com</value>
+        <name>hfile.block.cache.size</name>
+        <value>0.40</value>
     </property>
-    
+
     <property>
-      <name>hbase.zookeeper.useMulti</name>
-      <value>true</value>
+        <name>zookeeper.session.timeout</name>
+        <value>30000</value>
     </property>
-    
+
     <property>
-      <name>hbase_master_heapsize</name>
-      <value>250</value>
+        <name>zookeeper.znode.parent</name>
+        <value>/hbase-unsecure</value>
     </property>
-    
-    <property>
-      <name>hbase_regionserver_heapsize</name>
-      <value>250</value>
-    </property>
-    
-    <property>
-      <name>hfile.block.cache.size</name>
-      <value>0.40</value>
-    </property>
-    
-    <property>
-      <name>zookeeper.session.timeout</name>
-      <value>30000</value>
-    </property>
-    
-    <property>
-      <name>zookeeper.znode.parent</name>
-      <value>/hbase-unsecure</value>
-    </property>
-    
-  </configuration>
+
+</configuration>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/f75658cf/examples/test_case_data/sandbox/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox/hdfs-site.xml b/examples/test_case_data/sandbox/hdfs-site.xml
index f8347dc..c9a3a6a 100644
--- a/examples/test_case_data/sandbox/hdfs-site.xml
+++ b/examples/test_case_data/sandbox/hdfs-site.xml
@@ -1,265 +1,241 @@
-<!--Mon Jan 19 07:29:07 2015-->
-    <configuration>
-    
+<!--Mon May 25 01:30:31 2015-->
+<configuration>
+
     <property>
-      <name>dfs.block.access.token.enable</name>
-      <value>false</value>
+        <name>dfs.block.access.token.enable</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>dfs.block.size</name>
-      <value>34217472</value>
+        <name>dfs.blockreport.initialDelay</name>
+        <value>120</value>
     </property>
-    
+
     <property>
-      <name>dfs.blockreport.initialDelay</name>
-      <value>120</value>
+        <name>dfs.blocksize</name>
+        <value>134217728</value>
     </property>
-    
+
     <property>
-      <name>dfs.blocksize</name>
-      <value>134217728</value>
+        <name>dfs.client.read.shortcircuit</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>dfs.client.read.shortcircuit</name>
-      <value>true</value>
+        <name>dfs.client.read.shortcircuit.streams.cache.size</name>
+        <value>4096</value>
     </property>
-    
+
     <property>
-      <name>dfs.client.read.shortcircuit.streams.cache.size</name>
-      <value>4096</value>
+        <name>dfs.cluster.administrators</name>
+        <value> hdfs</value>
     </property>
-    
+
     <property>
-      <name>dfs.cluster.administrators</name>
-      <value> hdfs</value>
+        <name>dfs.datanode.address</name>
+        <value>0.0.0.0:50010</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.address</name>
-      <value>0.0.0.0:50010</value>
+        <name>dfs.datanode.balance.bandwidthPerSec</name>
+        <value>6250000</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.balance.bandwidthPerSec</name>
-      <value>6250000</value>
+        <name>dfs.datanode.data.dir</name>
+        <value>/hadoop/hdfs/data</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.data.dir</name>
-      <value>/hadoop/hdfs/data</value>
-      <final>true</final>
+        <name>dfs.datanode.data.dir.perm</name>
+        <value>750</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.data.dir.perm</name>
-      <value>750</value>
+        <name>dfs.datanode.du.reserved</name>
+        <value>1073741824</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.du.reserved</name>
-      <value>1073741824</value>
+        <name>dfs.datanode.failed.volumes.tolerated</name>
+        <value>0</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.failed.volumes.tolerated</name>
-      <value>0</value>
-      <final>true</final>
+        <name>dfs.datanode.http.address</name>
+        <value>0.0.0.0:50075</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.http.address</name>
-      <value>0.0.0.0:50075</value>
+        <name>dfs.datanode.https.address</name>
+        <value>0.0.0.0:50475</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.https.address</name>
-      <value>0.0.0.0:50475</value>
+        <name>dfs.datanode.ipc.address</name>
+        <value>0.0.0.0:8010</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.ipc.address</name>
-      <value>0.0.0.0:8010</value>
+        <name>dfs.datanode.max.transfer.threads</name>
+        <value>16384</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.max.transfer.threads</name>
-      <value>1024</value>
+        <name>dfs.domain.socket.path</name>
+        <value>/var/lib/hadoop-hdfs/dn_socket</value>
     </property>
-    
+
     <property>
-      <name>dfs.datanode.max.xcievers</name>
-      <value>1024</value>
+        <name>dfs.heartbeat.interval</name>
+        <value>3</value>
     </property>
-    
+
     <property>
-      <name>dfs.domain.socket.path</name>
-      <value>/var/lib/hadoop-hdfs/dn_socket</value>
+        <name>dfs.hosts.exclude</name>
+        <value>/etc/hadoop/conf/dfs.exclude</value>
     </property>
-    
+
     <property>
-      <name>dfs.heartbeat.interval</name>
-      <value>3</value>
+        <name>dfs.http.policy</name>
+        <value>HTTP_ONLY</value>
     </property>
-    
+
     <property>
-      <name>dfs.hosts.exclude</name>
-      <value>/etc/hadoop/conf/dfs.exclude</value>
+        <name>dfs.https.port</name>
+        <value>50470</value>
     </property>
-    
+
     <property>
-      <name>dfs.http.policy</name>
-      <value>HTTP_ONLY</value>
+        <name>dfs.journalnode.edits.dir</name>
+        <value>/hadoop/hdfs/journalnode</value>
     </property>
-    
+
     <property>
-      <name>dfs.https.port</name>
-      <value>50470</value>
+        <name>dfs.journalnode.http-address</name>
+        <value>0.0.0.0:8480</value>
     </property>
-    
+
     <property>
-      <name>dfs.journalnode.edits.dir</name>
-      <value>/hadoop/hdfs/journalnode</value>
+        <name>dfs.namenode.accesstime.precision</name>
+        <value>0</value>
     </property>
-    
+
     <property>
-      <name>dfs.journalnode.http-address</name>
-      <value>0.0.0.0:8480</value>
+        <name>dfs.namenode.avoid.read.stale.datanode</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.accesstime.precision</name>
-      <value>3600000</value>
+        <name>dfs.namenode.avoid.write.stale.datanode</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.avoid.read.stale.datanode</name>
-      <value>true</value>
+        <name>dfs.namenode.checkpoint.dir</name>
+        <value>/hadoop/hdfs/namesecondary</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.avoid.write.stale.datanode</name>
-      <value>true</value>
+        <name>dfs.namenode.checkpoint.edits.dir</name>
+        <value>${dfs.namenode.checkpoint.dir}</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.checkpoint.dir</name>
-      <value>/hadoop/hdfs/namesecondary</value>
+        <name>dfs.namenode.checkpoint.period</name>
+        <value>21600</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.checkpoint.edits.dir</name>
-      <value>${dfs.namenode.checkpoint.dir}</value>
+        <name>dfs.namenode.checkpoint.txns</name>
+        <value>1000000</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.checkpoint.period</name>
-      <value>21600</value>
+        <name>dfs.namenode.handler.count</name>
+        <value>40</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.checkpoint.txns</name>
-      <value>1000000</value>
+        <name>dfs.namenode.http-address</name>
+        <value>sandbox.hortonworks.com:50070</value>
+        <final>true</final>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.handler.count</name>
-      <value>100</value>
+        <name>dfs.namenode.https-address</name>
+        <value>sandbox.hortonworks.com:50470</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.http-address</name>
-      <value>sandbox.hortonworks.com:50070</value>
-      <final>true</final>
+        <name>dfs.namenode.name.dir</name>
+        <value>/hadoop/hdfs/namenode</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.https-address</name>
-      <value>sandbox.hortonworks.com:50470</value>
+        <name>dfs.namenode.name.dir.restore</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.name.dir</name>
-      <value>/hadoop/hdfs/namenode</value>
-      <final>true</final>
+        <name>dfs.namenode.safemode.threshold-pct</name>
+        <value>1.0f</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.name.dir.restore</name>
-      <value>true</value>
+        <name>dfs.namenode.secondary.http-address</name>
+        <value>sandbox.hortonworks.com:50090</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.safemode.threshold-pct</name>
-      <value>1.0f</value>
+        <name>dfs.namenode.stale.datanode.interval</name>
+        <value>30000</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.secondary.http-address</name>
-      <value>sandbox.hortonworks.com:50090</value>
+        <name>dfs.namenode.startup.delay.block.deletion.sec</name>
+        <value>3600</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.stale.datanode.interval</name>
-      <value>30000</value>
+        <name>dfs.namenode.write.stale.datanode.ratio</name>
+        <value>1.0f</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.startup.delay.block.deletion.sec</name>
-      <value>3600</value>
+        <name>dfs.permissions.enabled</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>dfs.namenode.write.stale.datanode.ratio</name>
-      <value>1.0f</value>
+        <name>dfs.permissions.superusergroup</name>
+        <value>hdfs</value>
     </property>
-    
+
     <property>
-      <name>dfs.nfs.exports.allowed.hosts</name>
-      <value>* rw</value>
+        <name>dfs.replication</name>
+        <value>3</value>
     </property>
-    
+
     <property>
-      <name>dfs.nfs3.dump.dir</name>
-      <value>/tmp/.hdfs-nfs</value>
+        <name>dfs.replication.max</name>
+        <value>50</value>
     </property>
-    
+
     <property>
-      <name>dfs.permissions.enabled</name>
-      <value>true</value>
+        <name>dfs.support.append</name>
+        <value>true</value>
+        <final>true</final>
     </property>
-    
+
     <property>
-      <name>dfs.permissions.superusergroup</name>
-      <value>hdfs</value>
+        <name>dfs.webhdfs.enabled</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>dfs.replication</name>
-      <value>1</value>
+        <name>fs.permissions.umask-mode</name>
+        <value>022</value>
     </property>
-    
-    <property>
-      <name>dfs.replication.max</name>
-      <value>50</value>
-    </property>
-    
-    <property>
-      <name>dfs.support.append</name>
-      <value>true</value>
-      <final>true</final>
-    </property>
-    
-    <property>
-      <name>dfs.webhdfs.enabled</name>
-      <value>true</value>
-      <final>true</final>
-    </property>
-    
-    <property>
-      <name>fs.permissions.umask-mode</name>
-      <value>022</value>
-    </property>
-    
-  </configuration>
+
+</configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/f75658cf/examples/test_case_data/sandbox/hive-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox/hive-site.xml b/examples/test_case_data/sandbox/hive-site.xml
index 40afbf6..7efaf4f 100644
--- a/examples/test_case_data/sandbox/hive-site.xml
+++ b/examples/test_case_data/sandbox/hive-site.xml
@@ -1,794 +1,779 @@
-<!--Tue Dec 16 19:33:41 2014-->
-    <configuration>
-    
+<!--Mon May 25 01:23:17 2015-->
+<configuration>
+
     <property>
-      <name>ambari.hive.db.schema.name</name>
-      <value>hive</value>
+        <name>ambari.hive.db.schema.name</name>
+        <value>hive</value>
     </property>
-    
+
     <property>
-      <name>datanucleus.cache.level2.type</name>
-      <value>none</value>
+        <name>datanucleus.cache.level2.type</name>
+        <value>none</value>
     </property>
-    
+
     <property>
-      <name>hive.auto.convert.join</name>
-      <value>true</value>
+        <name>hive.auto.convert.join</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.auto.convert.join.noconditionaltask</name>
-      <value>true</value>
+        <name>hive.auto.convert.join.noconditionaltask</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.auto.convert.join.noconditionaltask.size</name>
-      <value>1000000000</value>
+        <name>hive.auto.convert.join.noconditionaltask.size</name>
+        <value>1251999744</value>
     </property>
-    
+
     <property>
-      <name>hive.auto.convert.sortmerge.join</name>
-      <value>true</value>
+        <name>hive.auto.convert.sortmerge.join</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.auto.convert.sortmerge.join.noconditionaltask</name>
-      <value>true</value>
+        <name>hive.auto.convert.sortmerge.join.noconditionaltask</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.auto.convert.sortmerge.join.to.mapjoin</name>
-      <value>false</value>
+        <name>hive.auto.convert.sortmerge.join.to.mapjoin</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.cbo.enable</name>
-      <value>true</value>
+        <name>hive.cbo.enable</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.cli.print.header</name>
-      <value>false</value>
+        <name>hive.cli.print.header</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.cluster.delegation.token.store.class</name>
-      <value>org.apache.hadoop.hive.thrift.ZooKeeperTokenStore</value>
+        <name>hive.cluster.delegation.token.store.class</name>
+        <value>org.apache.hadoop.hive.thrift.ZooKeeperTokenStore</value>
     </property>
-    
+
     <property>
-      <name>hive.cluster.delegation.token.store.zookeeper.connectString</name>
-      <value>sandbox.hortonworks.com:2181</value>
+        <name>hive.cluster.delegation.token.store.zookeeper.connectString</name>
+        <value>sandbox.hortonworks.com:2181</value>
     </property>
-    
+
     <property>
-      <name>hive.cluster.delegation.token.store.zookeeper.znode</name>
-      <value>/hive/cluster/delegation</value>
+        <name>hive.cluster.delegation.token.store.zookeeper.znode</name>
+        <value>/hive/cluster/delegation</value>
     </property>
-    
+
     <property>
-      <name>hive.compactor.abortedtxn.threshold</name>
-      <value>1000</value>
+        <name>hive.compactor.abortedtxn.threshold</name>
+        <value>1000</value>
     </property>
-    
+
     <property>
-      <name>hive.compactor.check.interval</name>
-      <value>300s</value>
+        <name>hive.compactor.check.interval</name>
+        <value>300L</value>
     </property>
-    
+
     <property>
-      <name>hive.compactor.delta.num.threshold</name>
-      <value>10</value>
+        <name>hive.compactor.delta.num.threshold</name>
+        <value>10</value>
     </property>
-    
+
     <property>
-      <name>hive.compactor.delta.pct.threshold</name>
-      <value>0.1f</value>
+        <name>hive.compactor.delta.pct.threshold</name>
+        <value>0.1f</value>
     </property>
-    
+
     <property>
-      <name>hive.compactor.initiator.on</name>
-      <value>true</value>
+        <name>hive.compactor.initiator.on</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.compactor.worker.threads</name>
-      <value>0</value>
+        <name>hive.compactor.worker.threads</name>
+        <value>0</value>
     </property>
-    
+
     <property>
-      <name>hive.compactor.worker.timeout</name>
-      <value>86400s</value>
+        <name>hive.compactor.worker.timeout</name>
+        <value>86400L</value>
     </property>
-    
+
     <property>
-      <name>hive.compute.query.using.stats</name>
-      <value>true</value>
+        <name>hive.compute.query.using.stats</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.conf.restricted.list</name>
-      <value>hive.security.authenticator.manager,hive.security.authorization.manager,hive.users.in.admin.role</value>
+        <name>hive.conf.restricted.list</name>
+        <value>hive.security.authenticator.manager,hive.security.authorization.manager,hive.users.in.admin.role</value>
     </property>
-    
+
     <property>
-      <name>hive.convert.join.bucket.mapjoin.tez</name>
-      <value>false</value>
+        <name>hive.convert.join.bucket.mapjoin.tez</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.enforce.bucketing</name>
-      <value>true</value>
+        <name>hive.enforce.bucketing</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.enforce.sorting</name>
-      <value>true</value>
+        <name>hive.enforce.sorting</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.enforce.sortmergebucketmapjoin</name>
-      <value>true</value>
+        <name>hive.enforce.sortmergebucketmapjoin</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.compress.intermediate</name>
-      <value>false</value>
+        <name>hive.exec.compress.intermediate</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.compress.output</name>
-      <value>false</value>
+        <name>hive.exec.compress.output</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.dynamic.partition</name>
-      <value>true</value>
+        <name>hive.exec.dynamic.partition</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.dynamic.partition.mode</name>
-      <value>nonstrict</value>
+        <name>hive.exec.dynamic.partition.mode</name>
+        <value>nonstrict</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.failure.hooks</name>
-      <value>org.apache.hadoop.hive.ql.hooks.ATSHook</value>
+        <name>hive.exec.failure.hooks</name>
+        <value>org.apache.hadoop.hive.ql.hooks.ATSHook</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.max.created.files</name>
-      <value>100000</value>
+        <name>hive.exec.max.created.files</name>
+        <value>100000</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.max.dynamic.partitions</name>
-      <value>5000</value>
+        <name>hive.exec.max.dynamic.partitions</name>
+        <value>5000</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.max.dynamic.partitions.pernode</name>
-      <value>2000</value>
+        <name>hive.exec.max.dynamic.partitions.pernode</name>
+        <value>2000</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.orc.compression.strategy</name>
-      <value>SPEED</value>
+        <name>hive.exec.orc.compression.strategy</name>
+        <value>SPEED</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.orc.default.compress</name>
-      <value>ZLIB</value>
+        <name>hive.exec.orc.default.compress</name>
+        <value>ZLIB</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.orc.default.stripe.size</name>
-      <value>67108864</value>
+        <name>hive.exec.orc.default.stripe.size</name>
+        <value>67108864</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.parallel</name>
-      <value>false</value>
+        <name>hive.exec.parallel</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.parallel.thread.number</name>
-      <value>8</value>
+        <name>hive.exec.parallel.thread.number</name>
+        <value>8</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.post.hooks</name>
-      <value>org.apache.hadoop.hive.ql.hooks.ATSHook</value>
+        <name>hive.exec.post.hooks</name>
+        <value>org.apache.hadoop.hive.ql.hooks.ATSHook</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.pre.hooks</name>
-      <value>org.apache.hadoop.hive.ql.hooks.ATSHook</value>
+        <name>hive.exec.pre.hooks</name>
+        <value>org.apache.hadoop.hive.ql.hooks.ATSHook</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.reducers.bytes.per.reducer</name>
-      <value>67108864</value>
+        <name>hive.exec.reducers.bytes.per.reducer</name>
+        <value>67108864</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.reducers.max</name>
-      <value>1009</value>
+        <name>hive.exec.reducers.max</name>
+        <value>1009</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.scratchdir</name>
-      <value>/tmp/hive</value>
+        <name>hive.exec.scratchdir</name>
+        <value>/tmp/hive</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.submit.local.task.via.child</name>
-      <value>true</value>
+        <name>hive.exec.submit.local.task.via.child</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.exec.submitviachild</name>
-      <value>false</value>
+        <name>hive.exec.submitviachild</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.execution.engine</name>
-      <value>mr</value>
+        <name>hive.execution.engine</name>
+        <value>mr</value>
     </property>
-    
+
     <property>
-      <name>hive.fetch.task.aggr</name>
-      <value>false</value>
+        <name>hive.fetch.task.aggr</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.fetch.task.conversion</name>
-      <value>more</value>
+        <name>hive.fetch.task.conversion</name>
+        <value>more</value>
     </property>
-    
+
     <property>
-      <name>hive.fetch.task.conversion.threshold</name>
-      <value>1073741824</value>
+        <name>hive.fetch.task.conversion.threshold</name>
+        <value>1073741824</value>
     </property>
-    
+
     <property>
-      <name>hive.heapsize</name>
-      <value>250</value>
+        <name>hive.heapsize</name>
+        <value>1024</value>
     </property>
-    
+
     <property>
-      <name>hive.limit.optimize.enable</name>
-      <value>true</value>
+        <name>hive.limit.optimize.enable</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.limit.pushdown.memory.usage</name>
-      <value>0.04</value>
+        <name>hive.limit.pushdown.memory.usage</name>
+        <value>0.04</value>
     </property>
-    
+
     <property>
-      <name>hive.map.aggr</name>
-      <value>true</value>
+        <name>hive.map.aggr</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.map.aggr.hash.force.flush.memory.threshold</name>
-      <value>0.9</value>
+        <name>hive.map.aggr.hash.force.flush.memory.threshold</name>
+        <value>0.9</value>
     </property>
-    
+
     <property>
-      <name>hive.map.aggr.hash.min.reduction</name>
-      <value>0.5</value>
+        <name>hive.map.aggr.hash.min.reduction</name>
+        <value>0.5</value>
     </property>
-    
+
     <property>
-      <name>hive.map.aggr.hash.percentmemory</name>
-      <value>0.5</value>
+        <name>hive.map.aggr.hash.percentmemory</name>
+        <value>0.5</value>
     </property>
-    
+
     <property>
-      <name>hive.mapjoin.bucket.cache.size</name>
-      <value>10000</value>
+        <name>hive.mapjoin.bucket.cache.size</name>
+        <value>10000</value>
     </property>
-    
+
     <property>
-      <name>hive.mapjoin.optimized.hashtable</name>
-      <value>true</value>
+        <name>hive.mapjoin.optimized.hashtable</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.mapred.reduce.tasks.speculative.execution</name>
-      <value>false</value>
+        <name>hive.mapred.reduce.tasks.speculative.execution</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.merge.mapfiles</name>
-      <value>true</value>
+        <name>hive.merge.mapfiles</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.merge.mapredfiles</name>
-      <value>false</value>
+        <name>hive.merge.mapredfiles</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.merge.orcfile.stripe.level</name>
-      <value>true</value>
+        <name>hive.merge.orcfile.stripe.level</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.merge.rcfile.block.level</name>
-      <value>true</value>
+        <name>hive.merge.rcfile.block.level</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.merge.size.per.task</name>
-      <value>256000000</value>
+        <name>hive.merge.size.per.task</name>
+        <value>256000000</value>
     </property>
-    
+
     <property>
-      <name>hive.merge.smallfiles.avgsize</name>
-      <value>16000000</value>
+        <name>hive.merge.smallfiles.avgsize</name>
+        <value>16000000</value>
     </property>
-    
+
     <property>
-      <name>hive.merge.tezfiles</name>
-      <value>false</value>
+        <name>hive.merge.tezfiles</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.authorization.storage.checks</name>
-      <value>false</value>
+        <name>hive.metastore.authorization.storage.checks</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.cache.pinobjtypes</name>
-      <value>Table,Database,Type,FieldSchema,Order</value>
+        <name>hive.metastore.cache.pinobjtypes</name>
+        <value>Table,Database,Type,FieldSchema,Order</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.client.connect.retry.delay</name>
-      <value>5s</value>
+        <name>hive.metastore.client.connect.retry.delay</name>
+        <value>5s</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.client.socket.timeout</name>
-      <value>1800s</value>
+        <name>hive.metastore.client.socket.timeout</name>
+        <value>1800s</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.connect.retries</name>
-      <value>24</value>
+        <name>hive.metastore.connect.retries</name>
+        <value>24</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.execute.setugi</name>
-      <value>true</value>
+        <name>hive.metastore.execute.setugi</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.failure.retries</name>
-      <value>24</value>
+        <name>hive.metastore.failure.retries</name>
+        <value>24</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.kerberos.keytab.file</name>
-      <value>/etc/security/keytabs/hive.service.keytab</value>
+        <name>hive.metastore.kerberos.keytab.file</name>
+        <value>/etc/security/keytabs/hive.service.keytab</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.kerberos.principal</name>
-      <value>hive/_HOST@EXAMPLE.COM</value>
+        <name>hive.metastore.kerberos.principal</name>
+        <value>hive/_HOST@EXAMPLE.COM</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.pre.event.listeners</name>
-      <value>org.apache.hadoop.hive.ql.security.authorization.AuthorizationPreEventListener</value>
+        <name>hive.metastore.pre.event.listeners</name>
+        <value>org.apache.hadoop.hive.ql.security.authorization.AuthorizationPreEventListener</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.sasl.enabled</name>
-      <value>false</value>
+        <name>hive.metastore.sasl.enabled</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.server.max.threads</name>
-      <value>100000</value>
+        <name>hive.metastore.server.max.threads</name>
+        <value>100000</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.uris</name>
-      <value>thrift://sandbox.hortonworks.com:9083</value>
+        <name>hive.metastore.uris</name>
+        <value>thrift://sandbox.hortonworks.com:9083</value>
     </property>
-    
+
     <property>
-      <name>hive.metastore.warehouse.dir</name>
-      <value>/apps/hive/warehouse</value>
+        <name>hive.metastore.warehouse.dir</name>
+        <value>/apps/hive/warehouse</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.bucketmapjoin</name>
-      <value>true</value>
+        <name>hive.optimize.bucketmapjoin</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.bucketmapjoin.sortedmerge</name>
-      <value>false</value>
+        <name>hive.optimize.bucketmapjoin.sortedmerge</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.constant.propagation</name>
-      <value>true</value>
+        <name>hive.optimize.constant.propagation</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.index.filter</name>
-      <value>true</value>
+        <name>hive.optimize.index.filter</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.mapjoin.mapreduce</name>
-      <value>true</value>
+        <name>hive.optimize.mapjoin.mapreduce</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.metadataonly</name>
-      <value>true</value>
+        <name>hive.optimize.metadataonly</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.null.scan</name>
-      <value>true</value>
+        <name>hive.optimize.null.scan</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.reducededuplication</name>
-      <value>true</value>
+        <name>hive.optimize.reducededuplication</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.reducededuplication.min.reducer</name>
-      <value>4</value>
+        <name>hive.optimize.reducededuplication.min.reducer</name>
+        <value>4</value>
     </property>
-    
+
     <property>
-      <name>hive.optimize.sort.dynamic.partition</name>
-      <value>false</value>
+        <name>hive.optimize.sort.dynamic.partition</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.orc.compute.splits.num.threads</name>
-      <value>10</value>
+        <name>hive.orc.compute.splits.num.threads</name>
+        <value>10</value>
     </property>
-    
+
     <property>
-      <name>hive.orc.splits.include.file.footer</name>
-      <value>false</value>
+        <name>hive.orc.splits.include.file.footer</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.prewarm.enabled</name>
-      <value>false</value>
+        <name>hive.prewarm.enabled</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.prewarm.numcontainers</name>
-      <value>10</value>
+        <name>hive.prewarm.numcontainers</name>
+        <value>10</value>
     </property>
-    
+
     <property>
-      <name>hive.security.authenticator.manager</name>
-      <value>org.apache.hadoop.hive.ql.security.ProxyUserAuthenticator</value>
+        <name>hive.security.authenticator.manager</name>
+        <value>org.apache.hadoop.hive.ql.security.ProxyUserAuthenticator</value>
     </property>
-    
+
     <property>
-      <name>hive.security.authorization.enabled</name>
-      <value>false</value>
+        <name>hive.security.authorization.enabled</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.security.authorization.manager</name>
-      <value>org.apache.hadoop.hive.ql.security.authorization.plugin.sqlstd.SQLStdConfOnlyAuthorizerFactory</value>
+        <name>hive.security.authorization.manager</name>
+        <value>org.apache.hadoop.hive.ql.security.authorization.plugin.sqlstd.SQLStdConfOnlyAuthorizerFactory</value>
     </property>
-    
+
     <property>
-      <name>hive.security.metastore.authenticator.manager</name>
-      <value>org.apache.hadoop.hive.ql.security.HadoopDefaultMetastoreAuthenticator</value>
+        <name>hive.security.metastore.authenticator.manager</name>
+        <value>org.apache.hadoop.hive.ql.security.HadoopDefaultMetastoreAuthenticator</value>
     </property>
-    
+
     <property>
-      <name>hive.security.metastore.authorization.auth.reads</name>
-      <value>true</value>
+        <name>hive.security.metastore.authorization.auth.reads</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.security.metastore.authorization.manager</name>
-      <value>org.apache.hadoop.hive.ql.security.authorization.StorageBasedAuthorizationProvider,org.apache.hadoop.hive.ql.security.authorization.MetaStoreAuthzAPIAuthorizerEmbedOnly</value>
+        <name>hive.security.metastore.authorization.manager</name>
+        <value>org.apache.hadoop.hive.ql.security.authorization.StorageBasedAuthorizationProvider,org.apache.hadoop.hive.ql.security.authorization.MetaStoreAuthzAPIAuthorizerEmbedOnly</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.allow.user.substitution</name>
-      <value>true</value>
+        <name>hive.server2.allow.user.substitution</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.authentication</name>
-      <value>NONE</value>
+        <name>hive.server2.authentication</name>
+        <value>NONE</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.authentication.spnego.keytab</name>
-      <value>HTTP/_HOST@EXAMPLE.COM</value>
+        <name>hive.server2.authentication.spnego.keytab</name>
+        <value>HTTP/_HOST@EXAMPLE.COM</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.authentication.spnego.principal</name>
-      <value>/etc/security/keytabs/spnego.service.keytab</value>
+        <name>hive.server2.authentication.spnego.principal</name>
+        <value>/etc/security/keytabs/spnego.service.keytab</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.enable.doAs</name>
-      <value>false</value>
+        <name>hive.server2.enable.doAs</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.enable.impersonation</name>
-      <value>true</value>
+        <name>hive.server2.enable.impersonation</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.logging.operation.enabled</name>
-      <value>true</value>
+        <name>hive.server2.logging.operation.enabled</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.logging.operation.log.location</name>
-      <value>${system:java.io.tmpdir}/${system:user.name}/operation_logs</value>
+        <name>hive.server2.logging.operation.log.location</name>
+        <value>${system:java.io.tmpdir}/${system:user.name}/operation_logs</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.support.dynamic.service.discovery</name>
-      <value>true</value>
+        <name>hive.server2.support.dynamic.service.discovery</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.table.type.mapping</name>
-      <value>CLASSIC</value>
+        <name>hive.server2.table.type.mapping</name>
+        <value>CLASSIC</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.tez.default.queues</name>
-      <value>default</value>
+        <name>hive.server2.tez.default.queues</name>
+        <value>default</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.tez.initialize.default.sessions</name>
-      <value>false</value>
+        <name>hive.server2.tez.initialize.default.sessions</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.tez.sessions.per.default.queue</name>
-      <value>1</value>
+        <name>hive.server2.tez.sessions.per.default.queue</name>
+        <value>1</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.thrift.http.path</name>
-      <value>cliservice</value>
+        <name>hive.server2.thrift.http.path</name>
+        <value>cliservice</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.thrift.http.port</name>
-      <value>10001</value>
+        <name>hive.server2.thrift.http.port</name>
+        <value>10001</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.thrift.max.worker.threads</name>
-      <value>500</value>
+        <name>hive.server2.thrift.max.worker.threads</name>
+        <value>500</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.thrift.port</name>
-      <value>10000</value>
+        <name>hive.server2.thrift.port</name>
+        <value>10000</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.thrift.sasl.qop</name>
-      <value>auth</value>
+        <name>hive.server2.thrift.sasl.qop</name>
+        <value>auth</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.transport.mode</name>
-      <value>binary</value>
+        <name>hive.server2.transport.mode</name>
+        <value>binary</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.use.SSL</name>
-      <value>false</value>
+        <name>hive.server2.use.SSL</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.server2.zookeeper.namespace</name>
-      <value>hiveserver2</value>
+        <name>hive.server2.zookeeper.namespace</name>
+        <value>hiveserver2</value>
     </property>
-    
+
     <property>
-      <name>hive.smbjoin.cache.rows</name>
-      <value>10000</value>
+        <name>hive.smbjoin.cache.rows</name>
+        <value>10000</value>
     </property>
-    
+
     <property>
-      <name>hive.stats.autogather</name>
-      <value>true</value>
+        <name>hive.stats.autogather</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.stats.dbclass</name>
-      <value>fs</value>
+        <name>hive.stats.dbclass</name>
+        <value>fs</value>
     </property>
-    
+
     <property>
-      <name>hive.stats.fetch.column.stats</name>
-      <value>false</value>
+        <name>hive.stats.fetch.column.stats</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.stats.fetch.partition.stats</name>
-      <value>true</value>
+        <name>hive.stats.fetch.partition.stats</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.support.concurrency</name>
-      <value>true</value>
+        <name>hive.support.concurrency</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.auto.reducer.parallelism</name>
-      <value>false</value>
+        <name>hive.tez.auto.reducer.parallelism</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.container.size</name>
-      <value>250</value>
+        <name>hive.tez.container.size</name>
+        <value>3584</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.cpu.vcores</name>
-      <value>-1</value>
+        <name>hive.tez.cpu.vcores</name>
+        <value>-1</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.dynamic.partition.pruning</name>
-      <value>true</value>
+        <name>hive.tez.dynamic.partition.pruning</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.dynamic.partition.pruning.max.data.size</name>
-      <value>104857600</value>
+        <name>hive.tez.dynamic.partition.pruning.max.data.size</name>
+        <value>104857600</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.dynamic.partition.pruning.max.event.size</name>
-      <value>1048576</value>
+        <name>hive.tez.dynamic.partition.pruning.max.event.size</name>
+        <value>1048576</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.input.format</name>
-      <value>org.apache.hadoop.hive.ql.io.HiveInputFormat</value>
+        <name>hive.tez.input.format</name>
+        <value>org.apache.hadoop.hive.ql.io.HiveInputFormat</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.java.opts</name>
-      <value>-server -Xmx200m -Djava.net.preferIPv4Stack=true</value>
+        <name>hive.tez.java.opts</name>
+        <value>-server -Xmx2868m -Djava.net.preferIPv4Stack=true -XX:NewRatio=8 -XX:+UseNUMA -XX:+UseParallelGC -XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.log.level</name>
-      <value>INFO</value>
+        <name>hive.tez.log.level</name>
+        <value>INFO</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.max.partition.factor</name>
-      <value>2.0</value>
+        <name>hive.tez.max.partition.factor</name>
+        <value>2.0</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.min.partition.factor</name>
-      <value>0.25</value>
+        <name>hive.tez.min.partition.factor</name>
+        <value>0.25</value>
     </property>
-    
+
     <property>
-      <name>hive.tez.smb.number.waves</name>
-      <value>0.5</value>
+        <name>hive.tez.smb.number.waves</name>
+        <value>0.5</value>
     </property>
-    
+
     <property>
-      <name>hive.txn.manager</name>
-      <value>org.apache.hadoop.hive.ql.lockmgr.DbTxnManager</value>
+        <name>hive.txn.manager</name>
+        <value>org.apache.hadoop.hive.ql.lockmgr.DummyTxnManager</value>
     </property>
-    
+
     <property>
-      <name>hive.txn.max.open.batch</name>
-      <value>1000</value>
+        <name>hive.txn.max.open.batch</name>
+        <value>1000</value>
     </property>
-    
+
     <property>
-      <name>hive.txn.timeout</name>
-      <value>300</value>
+        <name>hive.txn.timeout</name>
+        <value>300</value>
     </property>
-    
+
     <property>
-      <name>hive.user.install.directory</name>
-      <value>/user/</value>
+        <name>hive.user.install.directory</name>
+        <value>/user/</value>
     </property>
-    
+
     <property>
-      <name>hive.users.in.admin.role</name>
-      <value>hue,hive</value>
+        <name>hive.vectorized.execution.enabled</name>
+        <value>true</value>
     </property>
-    
+
     <property>
-      <name>hive.vectorized.execution.enabled</name>
-      <value>true</value>
+        <name>hive.vectorized.execution.reduce.enabled</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>hive.vectorized.execution.reduce.enabled</name>
-      <value>false</value>
+        <name>hive.vectorized.groupby.checkinterval</name>
+        <value>4096</value>
     </property>
-    
+
     <property>
-      <name>hive.vectorized.groupby.checkinterval</name>
-      <value>4096</value>
+        <name>hive.vectorized.groupby.flush.percent</name>
+        <value>0.1</value>
     </property>
-    
+
     <property>
-      <name>hive.vectorized.groupby.flush.percent</name>
-      <value>0.1</value>
+        <name>hive.vectorized.groupby.maxentries</name>
+        <value>100000</value>
     </property>
-    
+
     <property>
-      <name>hive.vectorized.groupby.maxentries</name>
-      <value>100000</value>
+        <name>hive.zookeeper.client.port</name>
+        <value>2181</value>
     </property>
-    
+
     <property>
-      <name>hive.zookeeper.client.port</name>
-      <value>2181</value>
+        <name>hive.zookeeper.namespace</name>
+        <value>hive_zookeeper_namespace</value>
     </property>
-    
+
     <property>
-      <name>hive.zookeeper.namespace</name>
-      <value>hive_zookeeper_namespace</value>
+        <name>hive.zookeeper.quorum</name>
+        <value>sandbox.hortonworks.com:2181</value>
     </property>
-    
+
     <property>
-      <name>hive.zookeeper.quorum</name>
-      <value>sandbox.hortonworks.com:2181</value>
+        <name>javax.jdo.option.ConnectionDriverName</name>
+        <value>com.mysql.jdbc.Driver</value>
     </property>
-    
+
     <property>
-      <name>hive_metastore_user_passwd</name>
-      <value>hive</value>
+        <name>javax.jdo.option.ConnectionURL</name>
+        <value>jdbc:mysql://sandbox.hortonworks.com/hive?createDatabaseIfNotExist=true</value>
     </property>
-    
+
     <property>
-      <name>javax.jdo.option.ConnectionDriverName</name>
-      <value>com.mysql.jdbc.Driver</value>
+        <name>javax.jdo.option.ConnectionUserName</name>
+        <value>hive</value>
     </property>
-    
-    <property>
-      <name>javax.jdo.option.ConnectionPassword</name>
-      <value>hive</value>
-    </property>
-    
-    <property>
-      <name>javax.jdo.option.ConnectionURL</name>
-      <value>jdbc:mysql://sandbox.hortonworks.com/hive?createDatabaseIfNotExist=true</value>
-    </property>
-    
-    <property>
-      <name>javax.jdo.option.ConnectionUserName</name>
-      <value>hive</value>
-    </property>
-    
-  </configuration>
+
+</configuration>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/f75658cf/examples/test_case_data/sandbox/mapred-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox/mapred-site.xml b/examples/test_case_data/sandbox/mapred-site.xml
index ee3837c..f6a1c2e 100644
--- a/examples/test_case_data/sandbox/mapred-site.xml
+++ b/examples/test_case_data/sandbox/mapred-site.xml
@@ -1,239 +1,219 @@
-<!--Tue Dec 16 19:08:08 2014-->
-    <configuration>
-    
+<!--Mon May 25 01:30:55 2015-->
+<configuration>
+
     <property>
-      <name>io.sort.mb</name>
-      <value>64</value>
+        <name>mapreduce.admin.map.child.java.opts</name>
+        <value>-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=${hdp.version}</value>
     </property>
-    
+
     <property>
-      <name>mapred.child.java.opts</name>
-      <value>-Xmx200m</value>
+        <name>mapreduce.admin.reduce.child.java.opts</name>
+        <value>-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=${hdp.version}</value>
     </property>
-    
+
     <property>
-      <name>mapred.job.map.memory.mb</name>
-      <value>250</value>
+        <name>mapreduce.admin.user.env</name>
+        <value>LD_LIBRARY_PATH=/usr/hdp/${hdp.version}/hadoop/lib/native:/usr/hdp/${hdp.version}/hadoop/lib/native/Linux-amd64-64</value>
     </property>
-    
+
     <property>
-      <name>mapred.job.reduce.memory.mb</name>
-      <value>250</value>
+        <name>mapreduce.am.max-attempts</name>
+        <value>2</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.admin.map.child.java.opts</name>
-      <value>-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=${hdp.version}</value>
+        <name>mapreduce.application.classpath</name>
+        <value>$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:/usr/hdp/${hdp.version}/hadoop/lib/hadoop-lzo-0.6.0.${hdp.version}.jar:/etc/hadoop/conf/secure</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.admin.reduce.child.java.opts</name>
-      <value>-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=${hdp.version}</value>
+        <name>mapreduce.application.framework.path</name>
+        <value>/hdp/apps/${hdp.version}/mapreduce/mapreduce.tar.gz#mr-framework</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.admin.user.env</name>
-      <value>LD_LIBRARY_PATH=/usr/hdp/${hdp.version}/hadoop/lib/native:/usr/hdp/${hdp.version}/hadoop/lib/native/Linux-amd64-64</value>
+        <name>mapreduce.cluster.administrators</name>
+        <value> hadoop</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.am.max-attempts</name>
-      <value>2</value>
+        <name>mapreduce.framework.name</name>
+        <value>yarn</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.application.classpath</name>
-      <value>/tmp/kylin/*,$HADOOP_CONF_DIR,/usr/hdp/current/hive-client/conf/,/usr/hdp/${hdp.version}/hive/lib/hive-metastore.jar,/usr/hdp/${hdp.version}/hive/lib/hive-exec.jar,/usr/hdp/${hdp.version}/hive-hcatalog/share/hcatalog/*,$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*,$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*,$PWD/mr-framework/hadoop/share/hadoop/common/*,$PWD/mr-framework/hadoop/share/hadoop/common/lib/*,$PWD/mr-framework/hadoop/share/hadoop/yarn/*,$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*,$PWD/mr-framework/hadoop/share/hadoop/hdfs/*,$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*,/usr/hdp/${hdp.version}/hadoop/lib/hadoop-lzo-0.6.0.${hdp.version}.jar,/etc/hadoop/conf/secure</value>
+        <name>mapreduce.job.emit-timeline-data</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.application.framework.path</name>
-      <value>/hdp/apps/${hdp.version}/mapreduce/mapreduce.tar.gz#mr-framework</value>
+        <name>mapreduce.job.reduce.slowstart.completedmaps</name>
+        <value>0.05</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.cluster.administrators</name>
-      <value> hadoop</value>
+        <name>mapreduce.jobhistory.address</name>
+        <value>sandbox.hortonworks.com:10020</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.framework.name</name>
-      <value>yarn</value>
+        <name>mapreduce.jobhistory.bind-host</name>
+        <value>0.0.0.0</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.job.emit-timeline-data</name>
-      <value>false</value>
+        <name>mapreduce.jobhistory.done-dir</name>
+        <value>/mr-history/done</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.job.reduce.slowstart.completedmaps</name>
-      <value>0.05</value>
+        <name>mapreduce.jobhistory.intermediate-done-dir</name>
+        <value>/mr-history/tmp</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.jobhistory.address</name>
-      <value>sandbox.hortonworks.com:10020</value>
+        <name>mapreduce.jobhistory.webapp.address</name>
+        <value>sandbox.hortonworks.com:19888</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.jobhistory.bind-host</name>
-      <value>0.0.0.0</value>
+        <name>mapreduce.map.java.opts</name>
+        <value>-Xmx2867m</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.jobhistory.done-dir</name>
-      <value>/mr-history/done</value>
+        <name>mapreduce.map.log.level</name>
+        <value>INFO</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.jobhistory.intermediate-done-dir</name>
-      <value>/mr-history/tmp</value>
+        <name>mapreduce.map.memory.mb</name>
+        <value>3584</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.jobhistory.webapp.address</name>
-      <value>sandbox.hortonworks.com:19888</value>
+        <name>mapreduce.map.output.compress</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.map.java.opts</name>
-      <value>-Xmx200m</value>
+        <name>mapreduce.map.sort.spill.percent</name>
+        <value>0.7</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.map.log.level</name>
-      <value>INFO</value>
+        <name>mapreduce.map.speculative</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.map.memory.mb</name>
-      <value>250</value>
+        <name>mapreduce.output.fileoutputformat.compress</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.map.output.compress</name>
-      <value>false</value>
+        <name>mapreduce.output.fileoutputformat.compress.type</name>
+        <value>BLOCK</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.map.sort.spill.percent</name>
-      <value>0.7</value>
+        <name>mapreduce.reduce.input.buffer.percent</name>
+        <value>0.0</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.map.speculative</name>
-      <value>false</value>
+        <name>mapreduce.reduce.java.opts</name>
+        <value>-Xmx2867m</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.output.fileoutputformat.compress</name>
-      <value>false</value>
+        <name>mapreduce.reduce.log.level</name>
+        <value>INFO</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.output.fileoutputformat.compress.type</name>
-      <value>BLOCK</value>
+        <name>mapreduce.reduce.memory.mb</name>
+        <value>3584</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.input.buffer.percent</name>
-      <value>0.0</value>
+        <name>mapreduce.reduce.shuffle.fetch.retry.enabled</name>
+        <value>1</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.java.opts</name>
-      <value>-Xmx200m</value>
+        <name>mapreduce.reduce.shuffle.fetch.retry.interval-ms</name>
+        <value>1000</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.log.level</name>
-      <value>INFO</value>
+        <name>mapreduce.reduce.shuffle.fetch.retry.timeout-ms</name>
+        <value>30000</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.memory.mb</name>
-      <value>250</value>
+        <name>mapreduce.reduce.shuffle.input.buffer.percent</name>
+        <value>0.7</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.shuffle.fetch.retry.enabled</name>
-      <value>1</value>
+        <name>mapreduce.reduce.shuffle.merge.percent</name>
+        <value>0.66</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.shuffle.fetch.retry.interval-ms</name>
-      <value>1000</value>
+        <name>mapreduce.reduce.shuffle.parallelcopies</name>
+        <value>30</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.shuffle.fetch.retry.timeout-ms</name>
-      <value>30000</value>
+        <name>mapreduce.reduce.speculative</name>
+        <value>false</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.shuffle.input.buffer.percent</name>
-      <value>0.7</value>
+        <name>mapreduce.shuffle.port</name>
+        <value>13562</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.shuffle.merge.percent</name>
-      <value>0.66</value>
+        <name>mapreduce.task.io.sort.factor</name>
+        <value>100</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.shuffle.parallelcopies</name>
-      <value>30</value>
+        <name>mapreduce.task.io.sort.mb</name>
+        <value>1024</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.reduce.speculative</name>
-      <value>false</value>
+        <name>mapreduce.task.timeout</name>
+        <value>300000</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.shuffle.port</name>
-      <value>13562</value>
+        <name>yarn.app.mapreduce.am.admin-command-opts</name>
+        <value>-Dhdp.version=${hdp.version}</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.task.io.sort.factor</name>
-      <value>100</value>
+        <name>yarn.app.mapreduce.am.command-opts</name>
+        <value>-Xmx2867m -Dhdp.version=${hdp.version}</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.task.io.sort.mb</name>
-      <value>64</value>
+        <name>yarn.app.mapreduce.am.log.level</name>
+        <value>INFO</value>
     </property>
-    
+
     <property>
-      <name>mapreduce.task.timeout</name>
-      <value>300000</value>
+        <name>yarn.app.mapreduce.am.resource.mb</name>
+        <value>3584</value>
     </property>
-    
+
     <property>
-      <name>yarn.app.mapreduce.am.admin-command-opts</name>
-      <value>-Dhdp.version=${hdp.version}</value>
+        <name>yarn.app.mapreduce.am.staging-dir</name>
+        <value>/user</value>
     </property>
-    
-    <property>
-      <name>yarn.app.mapreduce.am.command-opts</name>
-      <value>-Xmx200m</value>
-    </property>
-    
-    <property>
-      <name>yarn.app.mapreduce.am.log.level</name>
-      <value>INFO</value>
-    </property>
-    
-    <property>
-      <name>yarn.app.mapreduce.am.resource.mb</name>
-      <value>250</value>
-    </property>
-    
-    <property>
-      <name>yarn.app.mapreduce.am.staging-dir</name>
-      <value>/user</value>
-    </property>
-    
-  </configuration>
+
+</configuration>