You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@kylin.apache.org by ma...@apache.org on 2015/05/26 05:12:03 UTC

[3/4] incubator-kylin git commit: KYLIN-697 backup all xmls for hdp 2.2.0

KYLIN-697 backup all xmls for hdp 2.2.0


Project: http://git-wip-us.apache.org/repos/asf/incubator-kylin/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-kylin/commit/b647f155
Tree: http://git-wip-us.apache.org/repos/asf/incubator-kylin/tree/b647f155
Diff: http://git-wip-us.apache.org/repos/asf/incubator-kylin/diff/b647f155

Branch: refs/heads/KYLIN-783
Commit: b647f155642a5dae778d6b466a1e6c8d79022935
Parents: 4fb9c8b
Author: honma <ho...@ebay.com>
Authored: Tue May 26 11:03:13 2015 +0800
Committer: honma <ho...@ebay.com>
Committed: Tue May 26 11:03:13 2015 +0800

----------------------------------------------------------------------
 .../test_case_data/sandbox-hdp220/.gitignore    |   2 +
 .../sandbox-hdp220/capacity-scheduler.xml       |  99 +++
 .../test_case_data/sandbox-hdp220/core-site.xml | 152 ++++
 .../sandbox-hdp220/hadoop-policy.xml            |  59 ++
 .../sandbox-hdp220/hbase-policy.xml             |  19 +
 .../sandbox-hdp220/hbase-site.xml               | 194 +++++
 .../test_case_data/sandbox-hdp220/hdfs-site.xml | 265 +++++++
 .../test_case_data/sandbox-hdp220/hive-site.xml | 794 +++++++++++++++++++
 .../sandbox-hdp220/kylin.properties             |  98 +++
 .../sandbox-hdp220/kylin_hive_conf.xml          |  54 ++
 .../sandbox-hdp220/kylin_job_conf.xml           |  61 ++
 .../sandbox-hdp220/mapred-site.xml              | 239 ++++++
 .../test_case_data/sandbox-hdp220/yarn-site.xml | 509 ++++++++++++
 13 files changed, 2545 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/.gitignore
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/.gitignore b/examples/test_case_data/sandbox-hdp220/.gitignore
new file mode 100644
index 0000000..5c501e0
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/.gitignore
@@ -0,0 +1,2 @@
+kylin.properties.override
+

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/capacity-scheduler.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/capacity-scheduler.xml b/examples/test_case_data/sandbox-hdp220/capacity-scheduler.xml
new file mode 100644
index 0000000..7b4a367
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/capacity-scheduler.xml
@@ -0,0 +1,99 @@
+<!--Tue Dec 16 19:08:08 2014-->
+    <configuration>
+    
+    <property>
+      <name>yarn.scheduler.capacity.default.minimum-user-limit-percent</name>
+      <value>100</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.maximum-am-resource-percent</name>
+      <value>0.5</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.maximum-applications</name>
+      <value>10000</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.node-locality-delay</name>
+      <value>40</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.resource-calculator</name>
+      <value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.accessible-node-labels</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.accessible-node-labels.default.capacity</name>
+      <value>-1</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.accessible-node-labels.default.maximum-capacity</name>
+      <value>-1</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.acl_administer_queue</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.capacity</name>
+      <value>100</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.default-node-label-expression</name>
+      <value> </value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.default.acl_administer_jobs</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.default.acl_submit_applications</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.default.capacity</name>
+      <value>100</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.default.maximum-am-resource-percent</name>
+      <value>0.5</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.default.maximum-capacity</name>
+      <value>100</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.default.state</name>
+      <value>RUNNING</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.default.user-limit-factor</name>
+      <value>1</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.capacity.root.queues</name>
+      <value>default</value>
+    </property>
+    
+  </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/core-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/core-site.xml b/examples/test_case_data/sandbox-hdp220/core-site.xml
new file mode 100644
index 0000000..1b3a3ee
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/core-site.xml
@@ -0,0 +1,152 @@
+<!--Mon Jan 19 07:28:25 2015-->
+    <configuration>
+    
+    <property>
+      <name>fs.defaultFS</name>
+      <value>hdfs://sandbox.hortonworks.com:8020</value>
+      <final>true</final>
+    </property>
+    
+    <property>
+      <name>fs.trash.interval</name>
+      <value>360</value>
+    </property>
+    
+    <property>
+      <name>hadoop.http.authentication.simple.anonymous.allowed</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.falcon.groups</name>
+      <value>users</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.falcon.hosts</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.hbase.groups</name>
+      <value>users</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.hbase.hosts</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.hcat.groups</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.hcat.hosts</name>
+      <value>sandbox.hortonworks.com</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.hive.groups</name>
+      <value>users</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.hive.hosts</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.hue.groups</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.hue.hosts</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.oozie.groups</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.oozie.hosts</name>
+      <value>sandbox.hortonworks.com</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.root.groups</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>hadoop.proxyuser.root.hosts</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>hadoop.security.auth_to_local</name>
+      <value>
+        RULE:[2:$1@$0]([rn]m@.*)s/.*/yarn/
+        RULE:[2:$1@$0](jhs@.*)s/.*/mapred/
+        RULE:[2:$1@$0]([nd]n@.*)s/.*/hdfs/
+        RULE:[2:$1@$0](hm@.*)s/.*/hbase/
+        RULE:[2:$1@$0](rs@.*)s/.*/hbase/
+        DEFAULT
+    </value>
+    </property>
+    
+    <property>
+      <name>hadoop.security.authentication</name>
+      <value>simple</value>
+    </property>
+    
+    <property>
+      <name>hadoop.security.authorization</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>io.compression.codecs</name>
+      <value>org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.SnappyCodec</value>
+    </property>
+    
+    <property>
+      <name>io.file.buffer.size</name>
+      <value>131072</value>
+    </property>
+    
+    <property>
+      <name>io.serializations</name>
+      <value>org.apache.hadoop.io.serializer.WritableSerialization</value>
+    </property>
+    
+    <property>
+      <name>ipc.client.connect.max.retries</name>
+      <value>50</value>
+    </property>
+    
+    <property>
+      <name>ipc.client.connection.maxidletime</name>
+      <value>30000</value>
+    </property>
+    
+    <property>
+      <name>ipc.client.idlethreshold</name>
+      <value>8000</value>
+    </property>
+    
+    <property>
+      <name>ipc.server.tcpnodelay</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.jobtracker.webinterface.trusted</name>
+      <value>false</value>
+    </property>
+    
+  </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/hadoop-policy.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/hadoop-policy.xml b/examples/test_case_data/sandbox-hdp220/hadoop-policy.xml
new file mode 100644
index 0000000..0f7e0b9
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/hadoop-policy.xml
@@ -0,0 +1,59 @@
+<!--Tue Dec 16 19:07:40 2014-->
+    <configuration>
+    
+    <property>
+      <name>security.admin.operations.protocol.acl</name>
+      <value>hadoop</value>
+    </property>
+    
+    <property>
+      <name>security.client.datanode.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.client.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.datanode.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.inter.datanode.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.inter.tracker.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.job.client.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.job.task.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.namenode.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.refresh.policy.protocol.acl</name>
+      <value>hadoop</value>
+    </property>
+    
+    <property>
+      <name>security.refresh.usertogroups.mappings.protocol.acl</name>
+      <value>hadoop</value>
+    </property>
+    
+  </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/hbase-policy.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/hbase-policy.xml b/examples/test_case_data/sandbox-hdp220/hbase-policy.xml
new file mode 100644
index 0000000..36756b8
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/hbase-policy.xml
@@ -0,0 +1,19 @@
+<!--Mon Jan 19 07:29:07 2015-->
+    <configuration>
+    
+    <property>
+      <name>security.admin.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.client.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>security.masterregion.protocol.acl</name>
+      <value>*</value>
+    </property>
+    
+  </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/hbase-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/hbase-site.xml b/examples/test_case_data/sandbox-hdp220/hbase-site.xml
new file mode 100644
index 0000000..9fd97c9
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/hbase-site.xml
@@ -0,0 +1,194 @@
+<!--Mon Jan 19 07:29:07 2015-->
+    <configuration>
+    
+    <property>
+      <name>dfs.domain.socket.path</name>
+      <value>/var/lib/hadoop-hdfs/dn_socket</value>
+    </property>
+    
+    <property>
+      <name>hbase.client.keyvalue.maxsize</name>
+      <value>10485760</value>
+    </property>
+    
+    <property>
+      <name>hbase.client.scanner.caching</name>
+      <value>100</value>
+    </property>
+    
+    <property>
+      <name>hbase.cluster.distributed</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hbase.coprocessor.master.classes</name>
+      <value>com.xasecure.authorization.hbase.XaSecureAuthorizationCoprocessor</value>
+    </property>
+    
+    <property>
+      <name>hbase.coprocessor.region.classes</name>
+      <value>com.xasecure.authorization.hbase.XaSecureAuthorizationCoprocessor</value>
+    </property>
+    
+    <property>
+      <name>hbase.defaults.for.version.skip</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hbase.hregion.majorcompaction</name>
+      <value>604800000</value>
+    </property>
+    
+    <property>
+      <name>hbase.hregion.majorcompaction.jitter</name>
+      <value>0.50</value>
+    </property>
+    
+    <property>
+      <name>hbase.hregion.max.filesize</name>
+      <value>10737418240</value>
+    </property>
+    
+    <property>
+      <name>hbase.hregion.memstore.block.multiplier</name>
+      <value>4</value>
+    </property>
+    
+    <property>
+      <name>hbase.hregion.memstore.flush.size</name>
+      <value>134217728</value>
+    </property>
+    
+    <property>
+      <name>hbase.hregion.memstore.mslab.enabled</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hbase.hstore.blockingStoreFiles</name>
+      <value>10</value>
+    </property>
+    
+    <property>
+      <name>hbase.hstore.compactionThreshold</name>
+      <value>3</value>
+    </property>
+    
+    <property>
+      <name>hbase.local.dir</name>
+      <value>${hbase.tmp.dir}/local</value>
+    </property>
+    
+    <property>
+      <name>hbase.master.info.bindAddress</name>
+      <value>0.0.0.0</value>
+    </property>
+    
+    <property>
+      <name>hbase.master.info.port</name>
+      <value>60010</value>
+    </property>
+    
+    <property>
+      <name>hbase.master.port</name>
+      <value>60000</value>
+    </property>
+    
+    <property>
+      <name>hbase.regionserver.global.memstore.lowerLimit</name>
+      <value>0.38</value>
+    </property>
+    
+    <property>
+      <name>hbase.regionserver.global.memstore.upperLimit</name>
+      <value>0.4</value>
+    </property>
+    
+    <property>
+      <name>hbase.regionserver.handler.count</name>
+      <value>60</value>
+    </property>
+    
+    <property>
+      <name>hbase.regionserver.info.port</name>
+      <value>60030</value>
+    </property>
+    
+    <property>
+      <name>hbase.rootdir</name>
+      <value>hdfs://sandbox.hortonworks.com:8020/apps/hbase/data</value>
+    </property>
+    
+    <property>
+      <name>hbase.rpc.engine</name>
+      <value>org.apache.hadoop.hbase.ipc.SecureRpcEngine</value>
+    </property>
+    
+    <property>
+      <name>hbase.rpc.protection</name>
+      <value>PRIVACY</value>
+    </property>
+    
+    <property>
+      <name>hbase.security.authentication</name>
+      <value>simple</value>
+    </property>
+    
+    <property>
+      <name>hbase.security.authorization</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hbase.superuser</name>
+      <value>hbase</value>
+    </property>
+    
+    <property>
+      <name>hbase.tmp.dir</name>
+      <value>/hadoop/hbase</value>
+    </property>
+    
+    <property>
+      <name>hbase.zookeeper.property.clientPort</name>
+      <value>2181</value>
+    </property>
+    
+    <property>
+      <name>hbase.zookeeper.quorum</name>
+      <value>sandbox.hortonworks.com</value>
+    </property>
+    
+    <property>
+      <name>hbase.zookeeper.useMulti</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hbase_master_heapsize</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>hbase_regionserver_heapsize</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>hfile.block.cache.size</name>
+      <value>0.40</value>
+    </property>
+    
+    <property>
+      <name>zookeeper.session.timeout</name>
+      <value>30000</value>
+    </property>
+    
+    <property>
+      <name>zookeeper.znode.parent</name>
+      <value>/hbase-unsecure</value>
+    </property>
+    
+  </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/hdfs-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/hdfs-site.xml b/examples/test_case_data/sandbox-hdp220/hdfs-site.xml
new file mode 100644
index 0000000..f8347dc
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/hdfs-site.xml
@@ -0,0 +1,265 @@
+<!--Mon Jan 19 07:29:07 2015-->
+    <configuration>
+    
+    <property>
+      <name>dfs.block.access.token.enable</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>dfs.block.size</name>
+      <value>34217472</value>
+    </property>
+    
+    <property>
+      <name>dfs.blockreport.initialDelay</name>
+      <value>120</value>
+    </property>
+    
+    <property>
+      <name>dfs.blocksize</name>
+      <value>134217728</value>
+    </property>
+    
+    <property>
+      <name>dfs.client.read.shortcircuit</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>dfs.client.read.shortcircuit.streams.cache.size</name>
+      <value>4096</value>
+    </property>
+    
+    <property>
+      <name>dfs.cluster.administrators</name>
+      <value> hdfs</value>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.address</name>
+      <value>0.0.0.0:50010</value>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.balance.bandwidthPerSec</name>
+      <value>6250000</value>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.data.dir</name>
+      <value>/hadoop/hdfs/data</value>
+      <final>true</final>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.data.dir.perm</name>
+      <value>750</value>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.du.reserved</name>
+      <value>1073741824</value>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.failed.volumes.tolerated</name>
+      <value>0</value>
+      <final>true</final>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.http.address</name>
+      <value>0.0.0.0:50075</value>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.https.address</name>
+      <value>0.0.0.0:50475</value>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.ipc.address</name>
+      <value>0.0.0.0:8010</value>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.max.transfer.threads</name>
+      <value>1024</value>
+    </property>
+    
+    <property>
+      <name>dfs.datanode.max.xcievers</name>
+      <value>1024</value>
+    </property>
+    
+    <property>
+      <name>dfs.domain.socket.path</name>
+      <value>/var/lib/hadoop-hdfs/dn_socket</value>
+    </property>
+    
+    <property>
+      <name>dfs.heartbeat.interval</name>
+      <value>3</value>
+    </property>
+    
+    <property>
+      <name>dfs.hosts.exclude</name>
+      <value>/etc/hadoop/conf/dfs.exclude</value>
+    </property>
+    
+    <property>
+      <name>dfs.http.policy</name>
+      <value>HTTP_ONLY</value>
+    </property>
+    
+    <property>
+      <name>dfs.https.port</name>
+      <value>50470</value>
+    </property>
+    
+    <property>
+      <name>dfs.journalnode.edits.dir</name>
+      <value>/hadoop/hdfs/journalnode</value>
+    </property>
+    
+    <property>
+      <name>dfs.journalnode.http-address</name>
+      <value>0.0.0.0:8480</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.accesstime.precision</name>
+      <value>3600000</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.avoid.read.stale.datanode</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.avoid.write.stale.datanode</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.checkpoint.dir</name>
+      <value>/hadoop/hdfs/namesecondary</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.checkpoint.edits.dir</name>
+      <value>${dfs.namenode.checkpoint.dir}</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.checkpoint.period</name>
+      <value>21600</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.checkpoint.txns</name>
+      <value>1000000</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.handler.count</name>
+      <value>100</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.http-address</name>
+      <value>sandbox.hortonworks.com:50070</value>
+      <final>true</final>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.https-address</name>
+      <value>sandbox.hortonworks.com:50470</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.name.dir</name>
+      <value>/hadoop/hdfs/namenode</value>
+      <final>true</final>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.name.dir.restore</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.safemode.threshold-pct</name>
+      <value>1.0f</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.secondary.http-address</name>
+      <value>sandbox.hortonworks.com:50090</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.stale.datanode.interval</name>
+      <value>30000</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.startup.delay.block.deletion.sec</name>
+      <value>3600</value>
+    </property>
+    
+    <property>
+      <name>dfs.namenode.write.stale.datanode.ratio</name>
+      <value>1.0f</value>
+    </property>
+    
+    <property>
+      <name>dfs.nfs.exports.allowed.hosts</name>
+      <value>* rw</value>
+    </property>
+    
+    <property>
+      <name>dfs.nfs3.dump.dir</name>
+      <value>/tmp/.hdfs-nfs</value>
+    </property>
+    
+    <property>
+      <name>dfs.permissions.enabled</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>dfs.permissions.superusergroup</name>
+      <value>hdfs</value>
+    </property>
+    
+    <property>
+      <name>dfs.replication</name>
+      <value>1</value>
+    </property>
+    
+    <property>
+      <name>dfs.replication.max</name>
+      <value>50</value>
+    </property>
+    
+    <property>
+      <name>dfs.support.append</name>
+      <value>true</value>
+      <final>true</final>
+    </property>
+    
+    <property>
+      <name>dfs.webhdfs.enabled</name>
+      <value>true</value>
+      <final>true</final>
+    </property>
+    
+    <property>
+      <name>fs.permissions.umask-mode</name>
+      <value>022</value>
+    </property>
+    
+  </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/hive-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/hive-site.xml b/examples/test_case_data/sandbox-hdp220/hive-site.xml
new file mode 100644
index 0000000..40afbf6
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/hive-site.xml
@@ -0,0 +1,794 @@
+<!--Tue Dec 16 19:33:41 2014-->
+    <configuration>
+    
+    <property>
+      <name>ambari.hive.db.schema.name</name>
+      <value>hive</value>
+    </property>
+    
+    <property>
+      <name>datanucleus.cache.level2.type</name>
+      <value>none</value>
+    </property>
+    
+    <property>
+      <name>hive.auto.convert.join</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.auto.convert.join.noconditionaltask</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.auto.convert.join.noconditionaltask.size</name>
+      <value>1000000000</value>
+    </property>
+    
+    <property>
+      <name>hive.auto.convert.sortmerge.join</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.auto.convert.sortmerge.join.noconditionaltask</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.auto.convert.sortmerge.join.to.mapjoin</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.cbo.enable</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.cli.print.header</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.cluster.delegation.token.store.class</name>
+      <value>org.apache.hadoop.hive.thrift.ZooKeeperTokenStore</value>
+    </property>
+    
+    <property>
+      <name>hive.cluster.delegation.token.store.zookeeper.connectString</name>
+      <value>sandbox.hortonworks.com:2181</value>
+    </property>
+    
+    <property>
+      <name>hive.cluster.delegation.token.store.zookeeper.znode</name>
+      <value>/hive/cluster/delegation</value>
+    </property>
+    
+    <property>
+      <name>hive.compactor.abortedtxn.threshold</name>
+      <value>1000</value>
+    </property>
+    
+    <property>
+      <name>hive.compactor.check.interval</name>
+      <value>300s</value>
+    </property>
+    
+    <property>
+      <name>hive.compactor.delta.num.threshold</name>
+      <value>10</value>
+    </property>
+    
+    <property>
+      <name>hive.compactor.delta.pct.threshold</name>
+      <value>0.1f</value>
+    </property>
+    
+    <property>
+      <name>hive.compactor.initiator.on</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.compactor.worker.threads</name>
+      <value>0</value>
+    </property>
+    
+    <property>
+      <name>hive.compactor.worker.timeout</name>
+      <value>86400s</value>
+    </property>
+    
+    <property>
+      <name>hive.compute.query.using.stats</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.conf.restricted.list</name>
+      <value>hive.security.authenticator.manager,hive.security.authorization.manager,hive.users.in.admin.role</value>
+    </property>
+    
+    <property>
+      <name>hive.convert.join.bucket.mapjoin.tez</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.enforce.bucketing</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.enforce.sorting</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.enforce.sortmergebucketmapjoin</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.compress.intermediate</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.compress.output</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.dynamic.partition</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.dynamic.partition.mode</name>
+      <value>nonstrict</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.failure.hooks</name>
+      <value>org.apache.hadoop.hive.ql.hooks.ATSHook</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.max.created.files</name>
+      <value>100000</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.max.dynamic.partitions</name>
+      <value>5000</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.max.dynamic.partitions.pernode</name>
+      <value>2000</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.orc.compression.strategy</name>
+      <value>SPEED</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.orc.default.compress</name>
+      <value>ZLIB</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.orc.default.stripe.size</name>
+      <value>67108864</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.parallel</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.parallel.thread.number</name>
+      <value>8</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.post.hooks</name>
+      <value>org.apache.hadoop.hive.ql.hooks.ATSHook</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.pre.hooks</name>
+      <value>org.apache.hadoop.hive.ql.hooks.ATSHook</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.reducers.bytes.per.reducer</name>
+      <value>67108864</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.reducers.max</name>
+      <value>1009</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.scratchdir</name>
+      <value>/tmp/hive</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.submit.local.task.via.child</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.exec.submitviachild</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.execution.engine</name>
+      <value>mr</value>
+    </property>
+    
+    <property>
+      <name>hive.fetch.task.aggr</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.fetch.task.conversion</name>
+      <value>more</value>
+    </property>
+    
+    <property>
+      <name>hive.fetch.task.conversion.threshold</name>
+      <value>1073741824</value>
+    </property>
+    
+    <property>
+      <name>hive.heapsize</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>hive.limit.optimize.enable</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.limit.pushdown.memory.usage</name>
+      <value>0.04</value>
+    </property>
+    
+    <property>
+      <name>hive.map.aggr</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.map.aggr.hash.force.flush.memory.threshold</name>
+      <value>0.9</value>
+    </property>
+    
+    <property>
+      <name>hive.map.aggr.hash.min.reduction</name>
+      <value>0.5</value>
+    </property>
+    
+    <property>
+      <name>hive.map.aggr.hash.percentmemory</name>
+      <value>0.5</value>
+    </property>
+    
+    <property>
+      <name>hive.mapjoin.bucket.cache.size</name>
+      <value>10000</value>
+    </property>
+    
+    <property>
+      <name>hive.mapjoin.optimized.hashtable</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.mapred.reduce.tasks.speculative.execution</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.merge.mapfiles</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.merge.mapredfiles</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.merge.orcfile.stripe.level</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.merge.rcfile.block.level</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.merge.size.per.task</name>
+      <value>256000000</value>
+    </property>
+    
+    <property>
+      <name>hive.merge.smallfiles.avgsize</name>
+      <value>16000000</value>
+    </property>
+    
+    <property>
+      <name>hive.merge.tezfiles</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.authorization.storage.checks</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.cache.pinobjtypes</name>
+      <value>Table,Database,Type,FieldSchema,Order</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.client.connect.retry.delay</name>
+      <value>5s</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.client.socket.timeout</name>
+      <value>1800s</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.connect.retries</name>
+      <value>24</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.execute.setugi</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.failure.retries</name>
+      <value>24</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.kerberos.keytab.file</name>
+      <value>/etc/security/keytabs/hive.service.keytab</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.kerberos.principal</name>
+      <value>hive/_HOST@EXAMPLE.COM</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.pre.event.listeners</name>
+      <value>org.apache.hadoop.hive.ql.security.authorization.AuthorizationPreEventListener</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.sasl.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.server.max.threads</name>
+      <value>100000</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.uris</name>
+      <value>thrift://sandbox.hortonworks.com:9083</value>
+    </property>
+    
+    <property>
+      <name>hive.metastore.warehouse.dir</name>
+      <value>/apps/hive/warehouse</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.bucketmapjoin</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.bucketmapjoin.sortedmerge</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.constant.propagation</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.index.filter</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.mapjoin.mapreduce</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.metadataonly</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.null.scan</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.reducededuplication</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.reducededuplication.min.reducer</name>
+      <value>4</value>
+    </property>
+    
+    <property>
+      <name>hive.optimize.sort.dynamic.partition</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.orc.compute.splits.num.threads</name>
+      <value>10</value>
+    </property>
+    
+    <property>
+      <name>hive.orc.splits.include.file.footer</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.prewarm.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.prewarm.numcontainers</name>
+      <value>10</value>
+    </property>
+    
+    <property>
+      <name>hive.security.authenticator.manager</name>
+      <value>org.apache.hadoop.hive.ql.security.ProxyUserAuthenticator</value>
+    </property>
+    
+    <property>
+      <name>hive.security.authorization.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.security.authorization.manager</name>
+      <value>org.apache.hadoop.hive.ql.security.authorization.plugin.sqlstd.SQLStdConfOnlyAuthorizerFactory</value>
+    </property>
+    
+    <property>
+      <name>hive.security.metastore.authenticator.manager</name>
+      <value>org.apache.hadoop.hive.ql.security.HadoopDefaultMetastoreAuthenticator</value>
+    </property>
+    
+    <property>
+      <name>hive.security.metastore.authorization.auth.reads</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.security.metastore.authorization.manager</name>
+      <value>org.apache.hadoop.hive.ql.security.authorization.StorageBasedAuthorizationProvider,org.apache.hadoop.hive.ql.security.authorization.MetaStoreAuthzAPIAuthorizerEmbedOnly</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.allow.user.substitution</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.authentication</name>
+      <value>NONE</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.authentication.spnego.keytab</name>
+      <value>HTTP/_HOST@EXAMPLE.COM</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.authentication.spnego.principal</name>
+      <value>/etc/security/keytabs/spnego.service.keytab</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.enable.doAs</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.enable.impersonation</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.logging.operation.enabled</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.logging.operation.log.location</name>
+      <value>${system:java.io.tmpdir}/${system:user.name}/operation_logs</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.support.dynamic.service.discovery</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.table.type.mapping</name>
+      <value>CLASSIC</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.tez.default.queues</name>
+      <value>default</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.tez.initialize.default.sessions</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.tez.sessions.per.default.queue</name>
+      <value>1</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.thrift.http.path</name>
+      <value>cliservice</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.thrift.http.port</name>
+      <value>10001</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.thrift.max.worker.threads</name>
+      <value>500</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.thrift.port</name>
+      <value>10000</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.thrift.sasl.qop</name>
+      <value>auth</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.transport.mode</name>
+      <value>binary</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.use.SSL</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.server2.zookeeper.namespace</name>
+      <value>hiveserver2</value>
+    </property>
+    
+    <property>
+      <name>hive.smbjoin.cache.rows</name>
+      <value>10000</value>
+    </property>
+    
+    <property>
+      <name>hive.stats.autogather</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.stats.dbclass</name>
+      <value>fs</value>
+    </property>
+    
+    <property>
+      <name>hive.stats.fetch.column.stats</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.stats.fetch.partition.stats</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.support.concurrency</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.auto.reducer.parallelism</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.container.size</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.cpu.vcores</name>
+      <value>-1</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.dynamic.partition.pruning</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.dynamic.partition.pruning.max.data.size</name>
+      <value>104857600</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.dynamic.partition.pruning.max.event.size</name>
+      <value>1048576</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.input.format</name>
+      <value>org.apache.hadoop.hive.ql.io.HiveInputFormat</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.java.opts</name>
+      <value>-server -Xmx200m -Djava.net.preferIPv4Stack=true</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.log.level</name>
+      <value>INFO</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.max.partition.factor</name>
+      <value>2.0</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.min.partition.factor</name>
+      <value>0.25</value>
+    </property>
+    
+    <property>
+      <name>hive.tez.smb.number.waves</name>
+      <value>0.5</value>
+    </property>
+    
+    <property>
+      <name>hive.txn.manager</name>
+      <value>org.apache.hadoop.hive.ql.lockmgr.DbTxnManager</value>
+    </property>
+    
+    <property>
+      <name>hive.txn.max.open.batch</name>
+      <value>1000</value>
+    </property>
+    
+    <property>
+      <name>hive.txn.timeout</name>
+      <value>300</value>
+    </property>
+    
+    <property>
+      <name>hive.user.install.directory</name>
+      <value>/user/</value>
+    </property>
+    
+    <property>
+      <name>hive.users.in.admin.role</name>
+      <value>hue,hive</value>
+    </property>
+    
+    <property>
+      <name>hive.vectorized.execution.enabled</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>hive.vectorized.execution.reduce.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hive.vectorized.groupby.checkinterval</name>
+      <value>4096</value>
+    </property>
+    
+    <property>
+      <name>hive.vectorized.groupby.flush.percent</name>
+      <value>0.1</value>
+    </property>
+    
+    <property>
+      <name>hive.vectorized.groupby.maxentries</name>
+      <value>100000</value>
+    </property>
+    
+    <property>
+      <name>hive.zookeeper.client.port</name>
+      <value>2181</value>
+    </property>
+    
+    <property>
+      <name>hive.zookeeper.namespace</name>
+      <value>hive_zookeeper_namespace</value>
+    </property>
+    
+    <property>
+      <name>hive.zookeeper.quorum</name>
+      <value>sandbox.hortonworks.com:2181</value>
+    </property>
+    
+    <property>
+      <name>hive_metastore_user_passwd</name>
+      <value>hive</value>
+    </property>
+    
+    <property>
+      <name>javax.jdo.option.ConnectionDriverName</name>
+      <value>com.mysql.jdbc.Driver</value>
+    </property>
+    
+    <property>
+      <name>javax.jdo.option.ConnectionPassword</name>
+      <value>hive</value>
+    </property>
+    
+    <property>
+      <name>javax.jdo.option.ConnectionURL</name>
+      <value>jdbc:mysql://sandbox.hortonworks.com/hive?createDatabaseIfNotExist=true</value>
+    </property>
+    
+    <property>
+      <name>javax.jdo.option.ConnectionUserName</name>
+      <value>hive</value>
+    </property>
+    
+  </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/kylin.properties
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/kylin.properties b/examples/test_case_data/sandbox-hdp220/kylin.properties
new file mode 100644
index 0000000..f211137
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/kylin.properties
@@ -0,0 +1,98 @@
+## Config for Kylin Engine ##
+
+# List of web servers in use, this enables one web server instance to sync up with other servers.
+kylin.rest.servers=localhost:7070
+
+# The metadata store in hbase
+kylin.metadata.url=kylin_metadata@hbase
+
+# The storage for final cube file in hbase
+kylin.storage.url=hbase
+
+# Temp folder in hdfs
+kylin.hdfs.working.dir=/tmp
+
+kylin.job.mapreduce.default.reduce.input.mb=500
+
+# If true, job engine will not assume that hadoop CLI reside on the same server as it self
+# you will have to specify kylin.job.remote.cli.hostname, kylin.job.remote.cli.username and kylin.job.remote.cli.password
+kylin.job.run.as.remote.cmd=false
+
+# Only necessary when kylin.job.run.as.remote.cmd=true
+kylin.job.remote.cli.hostname=
+
+# Only necessary when kylin.job.run.as.remote.cmd=true
+kylin.job.remote.cli.username=
+
+# Only necessary when kylin.job.run.as.remote.cmd=true
+kylin.job.remote.cli.password=
+
+# Used by test cases to prepare synthetic data for sample cube
+kylin.job.remote.cli.working.dir=/tmp/kylin
+
+# Max count of concurrent jobs running
+kylin.job.concurrent.max.limit=10
+
+# Whether calculate cube in mem in each mapper;
+kylin.job.cubing.inMem=true
+
+#the percentage of the sampling, default 25%
+kylin.job.cubing.inMem.sampling.percent=25
+
+# The cut size for hbase region, in GB.
+# E.g, for cube whose capacity be marked as "SMALL", split region per 5GB by default
+kylin.job.hbase.region.cut.small=5
+kylin.job.hbase.region.cut.medium=10
+kylin.job.hbase.region.cut.large=50
+
+# Time interval to check hadoop job status
+kylin.job.yarn.app.rest.check.interval.seconds=10
+
+## Config for Restful APP ##
+# database connection settings:
+ldap.server=
+ldap.username=
+ldap.password=
+ldap.user.searchBase=
+ldap.user.searchPattern=
+ldap.user.groupSearchBase=
+ldap.service.searchBase=OU=
+ldap.service.searchPattern=
+ldap.service.groupSearchBase=
+acl.adminRole=
+acl.defaultRole=
+ganglia.group=
+ganglia.port=8664
+
+## Config for mail service
+
+# If true, will send email notification;
+mail.enabled=false
+mail.host=
+mail.username=
+mail.password=
+mail.sender=
+
+###########################config info for web#######################
+
+#help info ,format{name|displayName|link} ,optional
+kylin.web.help.length=4
+kylin.web.help.0=start|Getting Started|
+kylin.web.help.1=odbc|ODBC Driver|
+kylin.web.help.2=tableau|Tableau Guide|
+kylin.web.help.3=onboard|Cube Design Tutorial|
+#hadoop url link ,optional
+kylin.web.hadoop=
+#job diagnostic url link ,optional
+kylin.web.diagnostic=
+#contact mail on web page ,optional
+kylin.web.contact_mail=
+
+###########################config info for front#######################
+
+#env DEV|QA|PROD
+deploy.env=DEV
+
+###########################config info for sandbox#######################
+kylin.sandbox=true
+

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/kylin_hive_conf.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/kylin_hive_conf.xml b/examples/test_case_data/sandbox-hdp220/kylin_hive_conf.xml
new file mode 100644
index 0000000..d990175
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/kylin_hive_conf.xml
@@ -0,0 +1,54 @@
+<?xml version="1.0"?>
+<configuration>
+    <property>
+        <name>dfs.replication</name>
+        <value>2</value>
+        <description>Block replication</description>
+    </property>
+
+    <property>
+        <name>hive.exec.compress.output</name>
+        <value>true</value>
+        <description>enable compress</description>
+    </property>
+
+    <property>
+        <name>hive.auto.convert.join.noconditionaltask</name>
+        <value>true</value>
+        <description>enable map-side join</description>
+    </property>
+
+    <property>
+        <name>hive.auto.convert.join.noconditionaltask.size</name>
+        <value>300000000</value>
+        <description>enable map-side join</description>
+    </property>
+
+    <!--
+    <property>
+        <name>mapreduce.map.output.compress.codec</name>
+        <value>com.hadoop.compression.lzo.LzoCodec</value>
+        <description></description>
+    </property>
+    <property>
+        <name>mapreduce.output.fileoutputformat.compress.codec</name>
+        <value>com.hadoop.compression.lzo.LzoCodec</value>
+        <description></description>
+    </property>
+    -->
+    <property>
+        <name>hive.merge.mapfiles</name>
+        <value>true</value>
+        <description>Enable hive file merge on mapper only job</description>
+    </property>
+    <property>
+        <name>hive.merge.mapredfiles</name>
+        <value>true</value>
+        <description>Enable hive file merge on map-reduce job</description>
+    </property>
+    <property>
+        <name>hive.merge.size.per.task</name>
+        <value>64000000</value>
+        <description>Size for the merged file</description>
+    </property>
+</configuration>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/kylin_job_conf.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/kylin_job_conf.xml b/examples/test_case_data/sandbox-hdp220/kylin_job_conf.xml
new file mode 100644
index 0000000..f811ed4
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/kylin_job_conf.xml
@@ -0,0 +1,61 @@
+<?xml version="1.0"?>
+<configuration>
+
+    <property>
+        <name>mapreduce.job.split.metainfo.maxsize</name>
+        <value>-1</value>
+        <description>The maximum permissible size of the split metainfo file.
+            The JobTracker won't attempt to read split metainfo files bigger than
+            the configured value. No limits if set to -1.
+        </description>
+    </property>
+
+    <!-- uncomment the following 5 properties to enable lzo compressing
+
+	<property>
+		<name>mapred.compress.map.output</name>
+		<value>true</value>
+		<description>Compress map outputs</description>
+	</property>
+
+	<property>
+		<name>mapred.map.output.compression.codec</name>
+		<value>com.hadoop.compression.lzo.LzoCodec</value>
+		<description>The compression codec to use for map outputs
+		</description>
+	</property>
+
+	<property>
+		<name>mapred.output.compress</name>
+		<value>true</value>
+		<description>Compress the output of a MapReduce job</description>
+	</property>
+
+	<property>
+		<name>mapred.output.compression.codec</name>
+		<value>com.hadoop.compression.lzo.LzoCodec</value>
+		<description>The compression codec to use for job outputs
+		</description>
+	</property>
+
+	<property>
+		<name>mapred.output.compression.type</name>
+		<value>BLOCK</value>
+		<description>The compression type to use for job outputs</description>
+	</property>
+
+	!-->
+
+    <property>
+        <name>mapreduce.job.max.split.locations</name>
+        <value>2000</value>
+        <description>No description</description>
+    </property>
+
+    <property>
+        <name>dfs.replication</name>
+        <value>2</value>
+        <description>Block replication</description>
+    </property>
+
+</configuration>
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/mapred-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/mapred-site.xml b/examples/test_case_data/sandbox-hdp220/mapred-site.xml
new file mode 100644
index 0000000..ee3837c
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/mapred-site.xml
@@ -0,0 +1,239 @@
+<!--Tue Dec 16 19:08:08 2014-->
+    <configuration>
+    
+    <property>
+      <name>io.sort.mb</name>
+      <value>64</value>
+    </property>
+    
+    <property>
+      <name>mapred.child.java.opts</name>
+      <value>-Xmx200m</value>
+    </property>
+    
+    <property>
+      <name>mapred.job.map.memory.mb</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>mapred.job.reduce.memory.mb</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.admin.map.child.java.opts</name>
+      <value>-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=${hdp.version}</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.admin.reduce.child.java.opts</name>
+      <value>-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=${hdp.version}</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.admin.user.env</name>
+      <value>LD_LIBRARY_PATH=/usr/hdp/${hdp.version}/hadoop/lib/native:/usr/hdp/${hdp.version}/hadoop/lib/native/Linux-amd64-64</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.am.max-attempts</name>
+      <value>2</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.application.classpath</name>
+      <value>/tmp/kylin/*,$HADOOP_CONF_DIR,/usr/hdp/current/hive-client/conf/,/usr/hdp/${hdp.version}/hive/lib/hive-metastore.jar,/usr/hdp/${hdp.version}/hive/lib/hive-exec.jar,/usr/hdp/${hdp.version}/hive-hcatalog/share/hcatalog/*,$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*,$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*,$PWD/mr-framework/hadoop/share/hadoop/common/*,$PWD/mr-framework/hadoop/share/hadoop/common/lib/*,$PWD/mr-framework/hadoop/share/hadoop/yarn/*,$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*,$PWD/mr-framework/hadoop/share/hadoop/hdfs/*,$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*,/usr/hdp/${hdp.version}/hadoop/lib/hadoop-lzo-0.6.0.${hdp.version}.jar,/etc/hadoop/conf/secure</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.application.framework.path</name>
+      <value>/hdp/apps/${hdp.version}/mapreduce/mapreduce.tar.gz#mr-framework</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.cluster.administrators</name>
+      <value> hadoop</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.framework.name</name>
+      <value>yarn</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.job.emit-timeline-data</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.job.reduce.slowstart.completedmaps</name>
+      <value>0.05</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.jobhistory.address</name>
+      <value>sandbox.hortonworks.com:10020</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.jobhistory.bind-host</name>
+      <value>0.0.0.0</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.jobhistory.done-dir</name>
+      <value>/mr-history/done</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.jobhistory.intermediate-done-dir</name>
+      <value>/mr-history/tmp</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.jobhistory.webapp.address</name>
+      <value>sandbox.hortonworks.com:19888</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.map.java.opts</name>
+      <value>-Xmx200m</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.map.log.level</name>
+      <value>INFO</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.map.memory.mb</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.map.output.compress</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.map.sort.spill.percent</name>
+      <value>0.7</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.map.speculative</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.output.fileoutputformat.compress</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.output.fileoutputformat.compress.type</name>
+      <value>BLOCK</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.input.buffer.percent</name>
+      <value>0.0</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.java.opts</name>
+      <value>-Xmx200m</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.log.level</name>
+      <value>INFO</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.memory.mb</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.shuffle.fetch.retry.enabled</name>
+      <value>1</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.shuffle.fetch.retry.interval-ms</name>
+      <value>1000</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.shuffle.fetch.retry.timeout-ms</name>
+      <value>30000</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.shuffle.input.buffer.percent</name>
+      <value>0.7</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.shuffle.merge.percent</name>
+      <value>0.66</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.shuffle.parallelcopies</name>
+      <value>30</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.reduce.speculative</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.shuffle.port</name>
+      <value>13562</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.task.io.sort.factor</name>
+      <value>100</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.task.io.sort.mb</name>
+      <value>64</value>
+    </property>
+    
+    <property>
+      <name>mapreduce.task.timeout</name>
+      <value>300000</value>
+    </property>
+    
+    <property>
+      <name>yarn.app.mapreduce.am.admin-command-opts</name>
+      <value>-Dhdp.version=${hdp.version}</value>
+    </property>
+    
+    <property>
+      <name>yarn.app.mapreduce.am.command-opts</name>
+      <value>-Xmx200m</value>
+    </property>
+    
+    <property>
+      <name>yarn.app.mapreduce.am.log.level</name>
+      <value>INFO</value>
+    </property>
+    
+    <property>
+      <name>yarn.app.mapreduce.am.resource.mb</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>yarn.app.mapreduce.am.staging-dir</name>
+      <value>/user</value>
+    </property>
+    
+  </configuration>

http://git-wip-us.apache.org/repos/asf/incubator-kylin/blob/b647f155/examples/test_case_data/sandbox-hdp220/yarn-site.xml
----------------------------------------------------------------------
diff --git a/examples/test_case_data/sandbox-hdp220/yarn-site.xml b/examples/test_case_data/sandbox-hdp220/yarn-site.xml
new file mode 100644
index 0000000..4e0ed9c
--- /dev/null
+++ b/examples/test_case_data/sandbox-hdp220/yarn-site.xml
@@ -0,0 +1,509 @@
+<!--Tue Dec 16 19:08:08 2014-->
+    <configuration>
+    
+    <property>
+      <name>hadoop.registry.rm.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>hadoop.registry.zk.quorum</name>
+      <value>sandbox.hortonworks.com:2181</value>
+    </property>
+    
+    <property>
+      <name>yarn.acl.enable</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.admin.acl</name>
+      <value></value>
+    </property>
+    
+    <property>
+      <name>yarn.application.classpath</name>
+      <value>$HADOOP_CONF_DIR,/usr/hdp/current/hadoop-client/*,/usr/hdp/current/hadoop-client/lib/*,/usr/hdp/current/hadoop-hdfs-client/*,/usr/hdp/current/hadoop-hdfs-client/lib/*,/usr/hdp/current/hadoop-yarn-client/*,/usr/hdp/current/hadoop-yarn-client/lib/*</value>
+    </property>
+    
+    <property>
+      <name>yarn.client.nodemanager-connect.max-wait-ms</name>
+      <value>900000</value>
+    </property>
+    
+    <property>
+      <name>yarn.client.nodemanager-connect.retry-interval-ms</name>
+      <value>10000</value>
+    </property>
+    
+    <property>
+      <name>yarn.log-aggregation-enable</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>yarn.log-aggregation.retain-seconds</name>
+      <value>2592000</value>
+    </property>
+    
+    <property>
+      <name>yarn.log.server.url</name>
+      <value>http://sandbox.hortonworks.com:19888/jobhistory/logs</value>
+    </property>
+    
+    <property>
+      <name>yarn.node-labels.fs-store.retry-policy-spec</name>
+      <value>2000, 500</value>
+    </property>
+    
+    <property>
+      <name>yarn.node-labels.fs-store.root-dir</name>
+      <value>/system/yarn/node-labels</value>
+    </property>
+    
+    <property>
+      <name>yarn.node-labels.manager-class</name>
+      <value>org.apache.hadoop.yarn.server.resourcemanager.nodelabels.MemoryRMNodeLabelsManager</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.address</name>
+      <value>0.0.0.0:45454</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.admin-env</name>
+      <value>MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.aux-services</name>
+      <value>mapreduce_shuffle</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.aux-services.mapreduce_shuffle.class</name>
+      <value>org.apache.hadoop.mapred.ShuffleHandler</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.bind-host</name>
+      <value>0.0.0.0</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.container-executor.class</name>
+      <value>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.container-monitor.interval-ms</name>
+      <value>3000</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.delete.debug-delay-sec</name>
+      <value>0</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.disk-health-checker.max-disk-utilization-per-disk-percentage</name>
+      <value>90</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb</name>
+      <value>1000</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.disk-health-checker.min-healthy-disks</name>
+      <value>0.25</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.health-checker.interval-ms</name>
+      <value>135000</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.health-checker.script.timeout-ms</name>
+      <value>60000</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.linux-container-executor.cgroups.hierarchy</name>
+      <value>hadoop-yarn</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.linux-container-executor.cgroups.mount</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.linux-container-executor.cgroups.strict-resource-usage</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.linux-container-executor.group</name>
+      <value>hadoop</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.linux-container-executor.resources-handler.class</name>
+      <value>org.apache.hadoop.yarn.server.nodemanager.util.DefaultLCEResourcesHandler</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.local-dirs</name>
+      <value>/hadoop/yarn/local</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.log-aggregation.compression-type</name>
+      <value>gz</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.log-aggregation.debug-enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.log-aggregation.num-log-files-per-app</name>
+      <value>30</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.log-aggregation.roll-monitoring-interval-seconds</name>
+      <value>-1</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.log-dirs</name>
+      <value>/hadoop/yarn/log</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.log.retain-second</name>
+      <value>604800</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.pmem-check-enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.recovery.dir</name>
+      <value>/var/log/hadoop-yarn/nodemanager/recovery-state</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.recovery.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.remote-app-log-dir</name>
+      <value>/app-logs</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.remote-app-log-dir-suffix</name>
+      <value>logs</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.resource.cpu-vcores</name>
+      <value>8</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.resource.memory-mb</name>
+      <value>2250</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.resource.percentage-physical-cpu-limit</name>
+      <value>100</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.vmem-check-enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.nodemanager.vmem-pmem-ratio</name>
+      <value>10</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.address</name>
+      <value>sandbox.hortonworks.com:8050</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.admin.address</name>
+      <value>sandbox.hortonworks.com:8141</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.am.max-attempts</name>
+      <value>2</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.bind-host</name>
+      <value>0.0.0.0</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.connect.max-wait.ms</name>
+      <value>900000</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.connect.retry-interval.ms</name>
+      <value>30000</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.fs.state-store.retry-policy-spec</name>
+      <value>2000, 500</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.fs.state-store.uri</name>
+      <value> </value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.ha.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.hostname</name>
+      <value>sandbox.hortonworks.com</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.nodes.exclude-path</name>
+      <value>/etc/hadoop/conf/yarn.exclude</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.recovery.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.resource-tracker.address</name>
+      <value>sandbox.hortonworks.com:8025</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.scheduler.address</name>
+      <value>sandbox.hortonworks.com:8030</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.scheduler.class</name>
+      <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.state-store.max-completed-applications</name>
+      <value>${yarn.resourcemanager.max-completed-applications}</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.store.class</name>
+      <value>org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.system-metrics-publisher.dispatcher.pool-size</name>
+      <value>10</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.system-metrics-publisher.enabled</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.webapp.address</name>
+      <value>sandbox.hortonworks.com:8088</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.webapp.delegation-token-auth-filter.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.webapp.proxyuser.hcat.groups</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.webapp.proxyuser.hcat.hosts</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.webapp.proxyuser.oozie.groups</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.webapp.proxyuser.oozie.hosts</name>
+      <value>*</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.work-preserving-recovery.enabled</name>
+      <value>false</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.work-preserving-recovery.scheduling-wait-ms</name>
+      <value>10000</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.zk-acl</name>
+      <value>world:anyone:rwcda </value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.zk-address</name>
+      <value>localhost:2181</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.zk-num-retries</name>
+      <value>1000</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.zk-retry-interval-ms</name>
+      <value>1000</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.zk-state-store.parent-path</name>
+      <value>/rmstore</value>
+    </property>
+    
+    <property>
+      <name>yarn.resourcemanager.zk-timeout-ms</name>
+      <value>10000</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.maximum-allocation-mb</name>
+      <value>2250</value>
+    </property>
+    
+    <property>
+      <name>yarn.scheduler.minimum-allocation-mb</name>
+      <value>250</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.address</name>
+      <value>sandbox.hortonworks.com:10200</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.bind-host</name>
+      <value>0.0.0.0</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.client.max-retries</name>
+      <value>30</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.client.retry-interval-ms</name>
+      <value>1000</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.enabled</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.generic-application-history.store-class</name>
+      <value>org.apache.hadoop.yarn.server.applicationhistoryservice.NullApplicationHistoryStore</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.http-authentication.simple.anonymous.allowed</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.http-authentication.type</name>
+      <value>simple</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.leveldb-timeline-store.path</name>
+      <value>/hadoop/yarn/timeline</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.leveldb-timeline-store.read-cache-size</name>
+      <value>104857600</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.leveldb-timeline-store.start-time-read-cache-size</name>
+      <value>10000</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.leveldb-timeline-store.start-time-write-cache-size</name>
+      <value>10000</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.leveldb-timeline-store.ttl-interval-ms</name>
+      <value>300000</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.store-class</name>
+      <value>org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.ttl-enable</name>
+      <value>true</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.ttl-ms</name>
+      <value>2678400000</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.webapp.address</name>
+      <value>sandbox.hortonworks.com:8188</value>
+    </property>
+    
+    <property>
+      <name>yarn.timeline-service.webapp.https.address</name>
+      <value>sandbox.hortonworks.com:8190</value>
+    </property>
+    
+  </configuration>