You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@griffin.apache.org by gu...@apache.org on 2019/10/28 13:57:37 UTC
[griffin-site] branch asf-site updated: Updated asf-site site from
master (4491dcba19e9671b04216f041febf89ccda9ca1b)
This is an automated email from the ASF dual-hosted git repository.
guoyp pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/griffin-site.git
The following commit(s) were added to refs/heads/asf-site by this push:
new cbaf0e7 Updated asf-site site from master (4491dcba19e9671b04216f041febf89ccda9ca1b)
cbaf0e7 is described below
commit cbaf0e7d6f1b5102cb1e920fd785876ce619c6ee
Author: William Guo <gu...@apache.org>
AuthorDate: Mon Oct 28 21:57:33 2019 +0800
Updated asf-site site from master (4491dcba19e9671b04216f041febf89ccda9ca1b)
---
docs/quickstart-cn.html | 28 ++++++++++++++--------------
1 file changed, 14 insertions(+), 14 deletions(-)
diff --git a/docs/quickstart-cn.html b/docs/quickstart-cn.html
index 2c5f43e..3ef0757 100644
--- a/docs/quickstart-cn.html
+++ b/docs/quickstart-cn.html
@@ -267,14 +267,14 @@ livy.repl.enable-hive-context = true
<div class="highlighter-rouge"><div class="highlight"><pre class="highlight"><code># Apache Griffin应用名称
spring.application.name=griffin_service
# MySQL数据库配置信息
-spring.datasource.url=jdbc:mysql://10.104.20.126:3306/griffin_quartz?useSSL=false
-spring.datasource.username=xnuser
-spring.datasource.password=Xn20!@n0oLk
+spring.datasource.url=jdbc:mysql://10.xxx.xx.xxx:3306/griffin_quartz?useSSL=false
+spring.datasource.username=xxxxx
+spring.datasource.password=xxxxx
spring.jpa.generate-ddl=true
spring.datasource.driver-class-name=com.mysql.jdbc.Driver
spring.jpa.show-sql=true
# Hive metastore配置信息
-hive.metastore.uris=thrift://namenodetest01.bi:9083
+hive.metastore.uris=thrift://namenode.test01.xxx:9083
hive.metastore.dbname=default
hive.hmshandler.retry.attempts=15
hive.hmshandler.retry.interval=2000ms
@@ -387,7 +387,7 @@ org.quartz.jobStore.clusterCheckinInterval=20000
{
"type": "HDFS",
"config": {
- "path": "hdfs://namenodetest01.bi.10101111.com:9001/griffin/persist",
+ "path": "hdfs://namenodetest01.xx.xxxx.com:9001/griffin/persist",
"max.persist.lines": 10000,
"max.lines.per.file": 10000
}
@@ -396,7 +396,7 @@ org.quartz.jobStore.clusterCheckinInterval=20000
"type": "ELASTICSEARCH",
"config": {
"method": "post",
- "api": "http://10.104.110.119:9200/griffin/accuracy",
+ "api": "http://10.xxx.xxx.xxx:9200/griffin/accuracy",
"connection.timeout": "1m",
"retry": 10
}
@@ -497,11 +497,11 @@ hive <span class="nt">-f</span> create-table.hql
<span class="c">#current hour</span>
<span class="nb">sudo</span> ./gen_demo_data.sh
-<span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span><span class="nb">date</span> +%Y%m%d%H<span class="sb">`</span>
+<span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span>date +%Y%m%d%H<span class="sb">`</span>
<span class="nv">dt</span><span class="o">=</span><span class="k">${</span><span class="nv">cur_date</span>:0:8<span class="k">}</span>
<span class="nv">hour</span><span class="o">=</span><span class="k">${</span><span class="nv">cur_date</span>:8:2<span class="k">}</span>
<span class="nv">partition_date</span><span class="o">=</span><span class="s2">"dt='</span><span class="nv">$dt</span><span class="s2">',hour='</span><span class="nv">$hour</span><span class="s2">'"</span>
-<span class="nb">sed </span>s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">></span> insert-data.hql
+sed s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">></span> insert-data.hql
hive <span class="nt">-f</span> insert-data.hql
<span class="nv">src_done_path</span><span class="o">=</span>/griffin/data/batch/demo_src/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
<span class="nv">tgt_done_path</span><span class="o">=</span>/griffin/data/batch/demo_tgt/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
@@ -513,11 +513,11 @@ hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="
<span class="c">#last hour</span>
<span class="nb">sudo</span> ./gen_demo_data.sh
-<span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span><span class="nb">date</span> <span class="nt">-d</span> <span class="s1">'1 hour ago'</span> +%Y%m%d%H<span class="sb">`</span>
+<span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span>date <span class="nt">-d</span> <span class="s1">'1 hour ago'</span> +%Y%m%d%H<span class="sb">`</span>
<span class="nv">dt</span><span class="o">=</span><span class="k">${</span><span class="nv">cur_date</span>:0:8<span class="k">}</span>
<span class="nv">hour</span><span class="o">=</span><span class="k">${</span><span class="nv">cur_date</span>:8:2<span class="k">}</span>
<span class="nv">partition_date</span><span class="o">=</span><span class="s2">"dt='</span><span class="nv">$dt</span><span class="s2">',hour='</span><span class="nv">$hour</span><span class="s2">'"</span>
-<span class="nb">sed </span>s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">></span> insert-data.hql
+sed s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">></span> insert-data.hql
hive <span class="nt">-f</span> insert-data.hql
<span class="nv">src_done_path</span><span class="o">=</span>/griffin/data/batch/demo_src/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
<span class="nv">tgt_done_path</span><span class="o">=</span>/griffin/data/batch/demo_tgt/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
@@ -532,12 +532,12 @@ hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="
<span class="k">while </span><span class="nb">true
</span><span class="k">do
</span><span class="nb">sudo</span> ./gen_demo_data.sh
- <span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span><span class="nb">date</span> +%Y%m%d%H<span class="sb">`</span>
- <span class="nv">next_date</span><span class="o">=</span><span class="sb">`</span><span class="nb">date</span> <span class="nt">-d</span> <span class="s2">"+1hour"</span> <span class="s1">'+%Y%m%d%H'</span><span class="sb">`</span>
+ <span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span>date +%Y%m%d%H<span class="sb">`</span>
+ <span class="nv">next_date</span><span class="o">=</span><span class="sb">`</span>date <span class="nt">-d</span> <span class="s2">"+1hour"</span> <span class="s1">'+%Y%m%d%H'</span><span class="sb">`</span>
<span class="nv">dt</span><span class="o">=</span><span class="k">${</span><span class="nv">next_date</span>:0:8<span class="k">}</span>
<span class="nv">hour</span><span class="o">=</span><span class="k">${</span><span class="nv">next_date</span>:8:2<span class="k">}</span>
<span class="nv">partition_date</span><span class="o">=</span><span class="s2">"dt='</span><span class="nv">$dt</span><span class="s2">',hour='</span><span class="nv">$hour</span><span class="s2">'"</span>
- <span class="nb">sed </span>s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">></span> insert-data.hql
+ sed s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">></span> insert-data.hql
hive <span class="nt">-f</span> insert-data.hql
<span class="nv">src_done_path</span><span class="o">=</span>/griffin/data/batch/demo_src/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
<span class="nv">tgt_done_path</span><span class="o">=</span>/griffin/data/batch/demo_tgt/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
@@ -546,7 +546,7 @@ hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="
hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="nv">src_done_path</span><span class="k">}</span>
hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="nv">tgt_done_path</span><span class="k">}</span>
<span class="nb">echo</span> <span class="s2">"insert data [</span><span class="nv">$partition_date</span><span class="s2">] done"</span>
- <span class="nb">sleep </span>3600
+ sleep 3600
<span class="k">done
</span><span class="nb">set</span> <span class="nt">-e</span>
</code></pre></div></div>