You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@griffin.apache.org by gu...@apache.org on 2019/10/28 13:57:37 UTC

[griffin-site] branch asf-site updated: Updated asf-site site from master (4491dcba19e9671b04216f041febf89ccda9ca1b)

This is an automated email from the ASF dual-hosted git repository.

guoyp pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/griffin-site.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new cbaf0e7  Updated asf-site site from master (4491dcba19e9671b04216f041febf89ccda9ca1b)
cbaf0e7 is described below

commit cbaf0e7d6f1b5102cb1e920fd785876ce619c6ee
Author: William Guo <gu...@apache.org>
AuthorDate: Mon Oct 28 21:57:33 2019 +0800

    Updated asf-site site from master (4491dcba19e9671b04216f041febf89ccda9ca1b)
---
 docs/quickstart-cn.html | 28 ++++++++++++++--------------
 1 file changed, 14 insertions(+), 14 deletions(-)

diff --git a/docs/quickstart-cn.html b/docs/quickstart-cn.html
index 2c5f43e..3ef0757 100644
--- a/docs/quickstart-cn.html
+++ b/docs/quickstart-cn.html
@@ -267,14 +267,14 @@ livy.repl.enable-hive-context = true
 <div class="highlighter-rouge"><div class="highlight"><pre class="highlight"><code># Apache Griffin应用名称
 spring.application.name=griffin_service
 # MySQL数据库配置信息
-spring.datasource.url=jdbc:mysql://10.104.20.126:3306/griffin_quartz?useSSL=false
-spring.datasource.username=xnuser
-spring.datasource.password=Xn20!@n0oLk
+spring.datasource.url=jdbc:mysql://10.xxx.xx.xxx:3306/griffin_quartz?useSSL=false
+spring.datasource.username=xxxxx
+spring.datasource.password=xxxxx
 spring.jpa.generate-ddl=true
 spring.datasource.driver-class-name=com.mysql.jdbc.Driver
 spring.jpa.show-sql=true
 # Hive metastore配置信息
-hive.metastore.uris=thrift://namenodetest01.bi:9083
+hive.metastore.uris=thrift://namenode.test01.xxx:9083
 hive.metastore.dbname=default
 hive.hmshandler.retry.attempts=15
 hive.hmshandler.retry.interval=2000ms
@@ -387,7 +387,7 @@ org.quartz.jobStore.clusterCheckinInterval=20000
     {
       "type": "HDFS",
       "config": {
-        "path": "hdfs://namenodetest01.bi.10101111.com:9001/griffin/persist",
+        "path": "hdfs://namenodetest01.xx.xxxx.com:9001/griffin/persist",
         "max.persist.lines": 10000,
         "max.lines.per.file": 10000
       }
@@ -396,7 +396,7 @@ org.quartz.jobStore.clusterCheckinInterval=20000
       "type": "ELASTICSEARCH",
       "config": {
         "method": "post",
-        "api": "http://10.104.110.119:9200/griffin/accuracy",
+        "api": "http://10.xxx.xxx.xxx:9200/griffin/accuracy",
         "connection.timeout": "1m",
         "retry": 10
       }
@@ -497,11 +497,11 @@ hive <span class="nt">-f</span> create-table.hql
 
 <span class="c">#current hour</span>
 <span class="nb">sudo</span> ./gen_demo_data.sh
-<span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span><span class="nb">date</span> +%Y%m%d%H<span class="sb">`</span>
+<span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span>date +%Y%m%d%H<span class="sb">`</span>
 <span class="nv">dt</span><span class="o">=</span><span class="k">${</span><span class="nv">cur_date</span>:0:8<span class="k">}</span>
 <span class="nv">hour</span><span class="o">=</span><span class="k">${</span><span class="nv">cur_date</span>:8:2<span class="k">}</span>
 <span class="nv">partition_date</span><span class="o">=</span><span class="s2">"dt='</span><span class="nv">$dt</span><span class="s2">',hour='</span><span class="nv">$hour</span><span class="s2">'"</span>
-<span class="nb">sed </span>s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">&gt;</span> insert-data.hql
+sed s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">&gt;</span> insert-data.hql
 hive <span class="nt">-f</span> insert-data.hql
 <span class="nv">src_done_path</span><span class="o">=</span>/griffin/data/batch/demo_src/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
 <span class="nv">tgt_done_path</span><span class="o">=</span>/griffin/data/batch/demo_tgt/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
@@ -513,11 +513,11 @@ hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="
 
 <span class="c">#last hour</span>
 <span class="nb">sudo</span> ./gen_demo_data.sh
-<span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span><span class="nb">date</span> <span class="nt">-d</span> <span class="s1">'1 hour ago'</span> +%Y%m%d%H<span class="sb">`</span>
+<span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span>date <span class="nt">-d</span> <span class="s1">'1 hour ago'</span> +%Y%m%d%H<span class="sb">`</span>
 <span class="nv">dt</span><span class="o">=</span><span class="k">${</span><span class="nv">cur_date</span>:0:8<span class="k">}</span>
 <span class="nv">hour</span><span class="o">=</span><span class="k">${</span><span class="nv">cur_date</span>:8:2<span class="k">}</span>
 <span class="nv">partition_date</span><span class="o">=</span><span class="s2">"dt='</span><span class="nv">$dt</span><span class="s2">',hour='</span><span class="nv">$hour</span><span class="s2">'"</span>
-<span class="nb">sed </span>s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">&gt;</span> insert-data.hql
+sed s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">&gt;</span> insert-data.hql
 hive <span class="nt">-f</span> insert-data.hql
 <span class="nv">src_done_path</span><span class="o">=</span>/griffin/data/batch/demo_src/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
 <span class="nv">tgt_done_path</span><span class="o">=</span>/griffin/data/batch/demo_tgt/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
@@ -532,12 +532,12 @@ hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="
 <span class="k">while </span><span class="nb">true
 </span><span class="k">do
   </span><span class="nb">sudo</span> ./gen_demo_data.sh
-  <span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span><span class="nb">date</span> +%Y%m%d%H<span class="sb">`</span>
-  <span class="nv">next_date</span><span class="o">=</span><span class="sb">`</span><span class="nb">date</span> <span class="nt">-d</span> <span class="s2">"+1hour"</span> <span class="s1">'+%Y%m%d%H'</span><span class="sb">`</span>
+  <span class="nv">cur_date</span><span class="o">=</span><span class="sb">`</span>date +%Y%m%d%H<span class="sb">`</span>
+  <span class="nv">next_date</span><span class="o">=</span><span class="sb">`</span>date <span class="nt">-d</span> <span class="s2">"+1hour"</span> <span class="s1">'+%Y%m%d%H'</span><span class="sb">`</span>
   <span class="nv">dt</span><span class="o">=</span><span class="k">${</span><span class="nv">next_date</span>:0:8<span class="k">}</span>
   <span class="nv">hour</span><span class="o">=</span><span class="k">${</span><span class="nv">next_date</span>:8:2<span class="k">}</span>
   <span class="nv">partition_date</span><span class="o">=</span><span class="s2">"dt='</span><span class="nv">$dt</span><span class="s2">',hour='</span><span class="nv">$hour</span><span class="s2">'"</span>
-  <span class="nb">sed </span>s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">&gt;</span> insert-data.hql
+  sed s/PARTITION_DATE/<span class="nv">$partition_date</span>/ ./insert-data.hql.template <span class="o">&gt;</span> insert-data.hql
   hive <span class="nt">-f</span> insert-data.hql
   <span class="nv">src_done_path</span><span class="o">=</span>/griffin/data/batch/demo_src/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
   <span class="nv">tgt_done_path</span><span class="o">=</span>/griffin/data/batch/demo_tgt/dt<span class="o">=</span><span class="k">${</span><span class="nv">dt</span><span class="k">}</span>/hour<span class="o">=</span><span class="k">${</span><span class="nv">hour</span><span class="k">}</span>/_DONE
@@ -546,7 +546,7 @@ hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="
   hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="nv">src_done_path</span><span class="k">}</span>
   hadoop fs <span class="nt">-touchz</span> <span class="k">${</span><span class="nv">tgt_done_path</span><span class="k">}</span>
   <span class="nb">echo</span> <span class="s2">"insert data [</span><span class="nv">$partition_date</span><span class="s2">] done"</span>
-  <span class="nb">sleep </span>3600
+  sleep 3600
 <span class="k">done
 </span><span class="nb">set</span> <span class="nt">-e</span>
 </code></pre></div></div>