You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@zeppelin.apache.org by ah...@apache.org on 2016/11/14 22:43:47 UTC
svn commit: r1769702 [2/2] - in /zeppelin/site/docs/0.7.0-SNAPSHOT: ./
assets/themes/zeppelin/img/docs-img/ development/ displaysystem/ install/
interpreter/ manual/ quickstart/ rest-api/ security/ storage/
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/manual/publish.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/manual/publish.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/manual/publish.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/manual/publish.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/pleasecontribute.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/pleasecontribute.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/pleasecontribute.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/pleasecontribute.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/explorezeppelinui.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/explorezeppelinui.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/explorezeppelinui.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/explorezeppelinui.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/tutorial.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/tutorial.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/tutorial.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/quickstart/tutorial.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-configuration.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-configuration.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-configuration.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-configuration.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-credential.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-credential.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-credential.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-credential.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-interpreter.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-interpreter.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-interpreter.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-interpreter.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-notebook.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-notebook.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-notebook.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/rest-api/rest-notebook.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/rss.xml
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/rss.xml?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/rss.xml (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/rss.xml Mon Nov 14 22:43:45 2016
@@ -5,8 +5,8 @@
<description>Apache Zeppelin - The Apache Software Foundation</description>
<link>http://zeppelin.apache.org</link>
<link>http://zeppelin.apache.org</link>
- <lastBuildDate>2016-11-14T07:08:53+01:00</lastBuildDate>
- <pubDate>2016-11-14T07:08:53+01:00</pubDate>
+ <lastBuildDate>2016-11-14T23:31:22+01:00</lastBuildDate>
+ <pubDate>2016-11-14T23:31:22+01:00</pubDate>
<ttl>1800</ttl>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/screenshots.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/screenshots.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/screenshots.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/screenshots.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/search.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/search.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/search.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/search.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/search_data.json
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/search_data.json?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/search_data.json (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/search_data.json Mon Nov 14 22:43:45 2016
@@ -281,7 +281,7 @@
"/interpreter/jdbc.html": {
"title": "Generic JDBC Interpreter for Apache Zeppelin",
- "content" : "<!--Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.-->Generic JDBC Interpreter for Apache ZeppelinOverviewJDBC interpreter lets you create a JDBC connection to any data sources seamlessly. By now, it has been tested with: Postgresql - JDBC Driver Mysql - JDBC Driver MariaDB - JDBC Driver Redshift - JDBC Driver Apache Hive - JDBC Driver Apache Phoenix itself is a JDBC driver
Apache Drill - JDBC Driver Apache Tajo - JDBC Driver If you are using other databases not in the above list, please feel free to share your use case. It would be helpful to improve the functionality of JDBC interpreter.Create a new JDBC InterpreterFirst, click + Create button at the top-right corner in the interpreter setting page.Fill Interpreter name field with whatever you want to use as the alias(e.g. mysql, mysql2, hive, redshift, and etc..). Please note that this alias will be used as %interpreter_name to call the interpreter in the paragraph. Then select jdbc as an Interpreter group. The default driver of JDBC interpreter is set as PostgreSQL. It means Zeppelin includes PostgreSQL driver jar in itself.So you don&#39;t need to add any dependencies(e.g. the artifact name or path for PostgreSQL driver jar) for PostgreSQL connection.The JDBC interpreter properties are defined by default like below. Name Default Value Descrip
tion common.max_count 1000 The maximun number of SQL result to display default.driver org.postgresql.Driver JDBC Driver Name default.password The JDBC user password default.url jdbc:postgresql://localhost:5432/ The URL for JDBC default.user gpadmin The JDBC user name If you want to connect other databases such as Mysql, Redshift and Hive, you need to edit the property values. The below example is for Mysql connection.The last step is Dependency Setting. Since Zeppelin only includes PostgreSQL driver jar by default, you need to add each driver&#39;s maven coordinates or JDBC driver&#39;s jar file path for the other databases.That&#39;s it. You can find more JDBC connection setting examples(Mysql, Apache Hive, Apache Phoenix, and Apache Tajo) in this section.More propertiesThere are more JDBC interpreter properties you can specify like below. Property Name Description common.max_result Max
number of SQL result to display to prevent the browser overload. This is common properties for all connections zeppelin.jdbc.auth.type Types of authentications' methods supported are SIMPLE, and KERBEROS zeppelin.jdbc.principal The principal name to load from the keytab zeppelin.jdbc.keytab.location The path to the keytab file You can also add more properties by using this method.For example, if a connection needs a schema parameter, it would have to add the property as follows: name value default.schema schema_name Binding JDBC interpter to notebookTo bind the interpreters created in the interpreter setting page, click the gear icon at the top-right corner.Select(blue) or deselect(white) the interpreter buttons depending on your use cases. If you need to use more than one interpreter in the notebook, activate several buttons.Don&#39;t forget to click Save button, or you will face Interpreter *** is not found error.How to u
seRun the paragraph with JDBC interpreterTo test whether your databases and Zeppelin are successfully connected or not, type %jdbc_interpreter_name(e.g. %mysql) at the top of the paragraph and run show databases.%jdbc_interpreter_nameshow databasesIf the paragraph is FINISHED without any errors, a new paragraph will be automatically added after the previous one with %jdbc_interpreter_name.So you don&#39;t need to type this prefix in every paragraphs&#39; header.Apply Zeppelin Dynamic FormsYou can leverage Zeppelin Dynamic Form inside your queries. You can use both the text input and select form parametrization features.%jdbc_interpreter_nameSELECT name, country, performerFROM demo.performersWHERE name=&#39;{{performer=Sheryl Crow|Doof|Fanfarlo|Los Paranoia}}&#39;ExamplesHere are some examples you can refer to. Including the below connectors, you can connect every databases as long as it can be configured with it&#39;s JDBC driver.MysqlProperties Name Valu
e default.driver com.mysql.jdbc.Driver default.url jdbc:mysql://localhost:3306/ default.user mysql_user default.password mysql_password Dependencies Artifact Excludes mysql:mysql-connector-java:5.1.38 Apache HiveProperties Name Value default.driver org.apache.hive.jdbc.HiveDriver default.url jdbc:hive2://localhost:10000 default.user hive_user default.password hive_password Dependencies Artifact Excludes org.apache.hive:hive-jdbc:0.14.0 org.apache.hadoop:hadoop-common:2.6.0 Apache PhoenixPhoenix supports thick and thin connection types:Thick client is faster, but must connect directly to ZooKeeper and HBase RegionServers.Thin client has fewer dependencies and connects through a Phoenix Query Server instance.Use the appropriate default.driver, default.url, and the dependency artifact for your connection type.Thick client connectionProperties
Name Value default.driver org.apache.phoenix.jdbc.PhoenixDriver default.url jdbc:phoenix:localhost:2181:/hbase-unsecure default.user phoenix_user default.password phoenix_password Dependencies Artifact Excludes org.apache.phoenix:phoenix-core:4.4.0-HBase-1.0 Thin client connectionProperties Name Value default.driver org.apache.phoenix.queryserver.client.Driver default.url jdbc:phoenix:thin:url=http://localhost:8765;serialization=PROTOBUF default.user phoenix_user default.password phoenix_password DependenciesBefore Adding one of the below dependencies, check the Phoenix version first. Artifact Excludes Description org.apache.phoenix:phoenix-server-client:4.7.0-HBase-1.1 For Phoenix 4.7 org.apache.phoenix:phoenix-queryserver-client:4.8.0-HBase-1.2 For Phoenix 4.8+ Apache TajoProperties Name Value default.driver org
.apache.tajo.jdbc.TajoDriver default.url jdbc:tajo://localhost:26002/default Dependencies Artifact Excludes org.apache.tajo:tajo-jdbc:0.11.0 Bug reportingIf you find a bug using JDBC interpreter, please create a JIRA ticket.",
+ "content" : "<!--Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.-->Generic JDBC Interpreter for Apache ZeppelinOverviewJDBC interpreter lets you create a JDBC connection to any data sources seamlessly. By now, it has been tested with: Postgresql - JDBC Driver Mysql - JDBC Driver MariaDB - JDBC Driver Redshift - JDBC Driver Apache Hive - JDBC Driver Apache Phoenix itself is a JDBC driver
Apache Drill - JDBC Driver Apache Tajo - JDBC Driver If you are using other databases not in the above list, please feel free to share your use case. It would be helpful to improve the functionality of JDBC interpreter.Create a new JDBC InterpreterFirst, click + Create button at the top-right corner in the interpreter setting page.Fill Interpreter name field with whatever you want to use as the alias(e.g. mysql, mysql2, hive, redshift, and etc..). Please note that this alias will be used as %interpreter_name to call the interpreter in the paragraph. Then select jdbc as an Interpreter group. The default driver of JDBC interpreter is set as PostgreSQL. It means Zeppelin includes PostgreSQL driver jar in itself.So you don&#39;t need to add any dependencies(e.g. the artifact name or path for PostgreSQL driver jar) for PostgreSQL connection.The JDBC interpreter properties are defined by default like below. Name Default Value Descrip
tion common.max_count 1000 The maximun number of SQL result to display default.driver org.postgresql.Driver JDBC Driver Name default.password The JDBC user password default.url jdbc:postgresql://localhost:5432/ The URL for JDBC default.user gpadmin The JDBC user name If you want to connect other databases such as Mysql, Redshift and Hive, you need to edit the property values. The below example is for Mysql connection.The last step is Dependency Setting. Since Zeppelin only includes PostgreSQL driver jar by default, you need to add each driver&#39;s maven coordinates or JDBC driver&#39;s jar file path for the other databases.That&#39;s it. You can find more JDBC connection setting examples(Mysql, MariaDB, Redshift, Apache Hive, Apache Phoenix, and Apache Tajo) in this section.More propertiesThere are more JDBC interpreter properties you can specify like below. Property Name Description commo
n.max_result Max number of SQL result to display to prevent the browser overload. This is common properties for all connections zeppelin.jdbc.auth.type Types of authentications' methods supported are SIMPLE, and KERBEROS zeppelin.jdbc.principal The principal name to load from the keytab zeppelin.jdbc.keytab.location The path to the keytab file You can also add more properties by using this method.For example, if a connection needs a schema parameter, it would have to add the property as follows: name value default.schema schema_name Binding JDBC interpter to notebookTo bind the interpreters created in the interpreter setting page, click the gear icon at the top-right corner.Select(blue) or deselect(white) the interpreter buttons depending on your use cases. If you need to use more than one interpreter in the notebook, activate several buttons.Don&#39;t forget to click Save button, or you will face Interpreter *** is not f
ound error.How to useRun the paragraph with JDBC interpreterTo test whether your databases and Zeppelin are successfully connected or not, type %jdbc_interpreter_name(e.g. %mysql) at the top of the paragraph and run show databases.%jdbc_interpreter_nameshow databasesIf the paragraph is FINISHED without any errors, a new paragraph will be automatically added after the previous one with %jdbc_interpreter_name.So you don&#39;t need to type this prefix in every paragraphs&#39; header.Apply Zeppelin Dynamic FormsYou can leverage Zeppelin Dynamic Form inside your queries. You can use both the text input and select form parametrization features.%jdbc_interpreter_nameSELECT name, country, performerFROM demo.performersWHERE name=&#39;{{performer=Sheryl Crow|Doof|Fanfarlo|Los Paranoia}}&#39;ExamplesHere are some examples you can refer to. Including the below connectors, you can connect every databases as long as it can be configured with it&#39;s JDBC driver.PostgresProper
ties Name Value default.driver org.postgresql.Driver default.url jdbc:postgresql://localhost:5432/ default.user mysql_user default.password mysql_password Postgres JDBC Driver DocsDependencies Artifact Excludes org.postgresql:postgresql:9.4.1211 Maven Repository: org.postgresql:postgresqlMysqlProperties Name Value default.driver com.mysql.jdbc.Driver default.url jdbc:mysql://localhost:3306/ default.user mysql_user default.password mysql_password Mysql JDBC Driver DocsDependencies Artifact Excludes mysql:mysql-connector-java:5.1.38 Maven Repository: mysql:mysql-connector-javaMariaDBProperties Name Value default.driver org.mariadb.jdbc.Driver default.url jdbc:mariadb://localhost:3306 default.user mariadb_user default.password mariadb_password MariaDB JDBC Driver DocsDependencies Artifact Exclu
des org.mariadb.jdbc:mariadb-java-client:1.5.4 Maven Repository: org.mariadb.jdbc:mariadb-java-clientRedshiftProperties Name Value default.driver com.amazon.redshift.jdbc42.Driver default.url jdbc:redshift://your-redshift-instance-address.redshift.amazonaws.com:5439/your-database default.user redshift_user default.password redshift_password AWS Redshift JDBC Driver DocsDependencies Artifact Excludes com.amazonaws:aws-java-sdk-redshift:1.11.51 Maven Repository: com.amazonaws:aws-java-sdk-redshiftApache HiveProperties Name Value default.driver org.apache.hive.jdbc.HiveDriver default.url jdbc:hive2://localhost:10000 default.user hive_user default.password hive_password Apache Hive 1 JDBC Driver DocsApache Hive 2 JDBC Driver DocsDependencies Artifact Excludes org.apache.hive:hive-jdbc:0.14.0 org.apache.hadoop:hadoop-common:2.6.0
Maven Repository : org.apache.hive:hive-jdbcApache PhoenixPhoenix supports thick and thin connection types:Thick client is faster, but must connect directly to ZooKeeper and HBase RegionServers.Thin client has fewer dependencies and connects through a Phoenix Query Server instance.Use the appropriate default.driver, default.url, and the dependency artifact for your connection type.Thick client connectionProperties Name Value default.driver org.apache.phoenix.jdbc.PhoenixDriver default.url jdbc:phoenix:localhost:2181:/hbase-unsecure default.user phoenix_user default.password phoenix_password Dependencies Artifact Excludes org.apache.phoenix:phoenix-core:4.4.0-HBase-1.0 Maven Repository: org.apache.phoenix:phoenix-coreThin client connectionProperties Name Value default.driver org.apache.phoenix.queryserver.client.Driver default.url jdbc:phoenix:thin:url=http://localhost:8765;serialization
=PROTOBUF default.user phoenix_user default.password phoenix_password DependenciesBefore Adding one of the below dependencies, check the Phoenix version first. Artifact Excludes Description org.apache.phoenix:phoenix-server-client:4.7.0-HBase-1.1 For Phoenix 4.7 org.apache.phoenix:phoenix-queryserver-client:4.8.0-HBase-1.2 For Phoenix 4.8+ Maven Repository: org.apache.phoenix:phoenix-queryserver-clientApache TajoProperties Name Value default.driver org.apache.tajo.jdbc.TajoDriver default.url jdbc:tajo://localhost:26002/default Apache Tajo JDBC Driver DocsDependencies Artifact Excludes org.apache.tajo:tajo-jdbc:0.11.0 Maven Repository: org.apache.tajo:tajo-jdbcBug reportingIf you find a bug using JDBC interpreter, please create a JIRA ticket.",
"url": " /interpreter/jdbc.html",
"group": "interpreter",
"excerpt": "Generic JDBC Interpreter lets you create a JDBC connection to any data source. You can use Postgres, MySql, MariaDB, Redshift, Apache Hive, Apache Phoenix, Apache Drill and Apache Tajo using JDBC interpreter."
@@ -398,6 +398,17 @@
}
,
+
+
+ "/interpreter/scio.html": {
+ "title": "Scio Interpreter for Apache Zeppelin",
+ "content" : "<!--Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.-->Scio Interpreter for Apache ZeppelinOverviewScio is a Scala DSL for Google Cloud Dataflow and Apache Beam inspired by Spark and Scalding. See the current wiki and API documentation for more information.Configuration Name Default Value Description zeppelin.scio.argz --runner=InProcessPipelineRunner Scio interpreter wide arguments. Documentation: https://github.com/spotify/scio/wiki#options and https://cloud.g
oogle.com/dataflow/pipelines/specifying-exec-params zeppelin.scio.maxResult 1000 Max number of SCollection results to display Enabling the Scio InterpreterIn a notebook, to enable the Scio interpreter, click the Gear icon and select beam (beam.scio).Using the Scio InterpreterIn a paragraph, use $beam.scio to select the Scio interpreter. You can use it much the same way as vanilla Scala REPL and Scio REPL. State (like variables, imports, execution etc) is shared among all Scio paragraphs. There is a special variable argz which holds arguments from Scio interpreter settings. The easiest way to proceed is to create a Scio context via standard ContextAndArgs.$beam.scioval (sc, args) = ContextAndArgs(argz)Use sc context the way you would in regular pipeline/REPL.Example:$beam.scioval (sc, args) = ContextAndArgs(argz)sc.parallelize(Seq(&quot;foo&quot;, &quot;foo&quot;, &quot;bar&quot;)).countByValue.closeAndDisplay()If you close Scio context, go ahead
an create a new one using ContextAndArgs. Please refer to Scio wiki for more complex examples. You can close Scio context much the same way as in Scio REPL, and use Zeppelin display helpers to synchronously close and display results - read more below.ProgressThere can be only one paragraph running at a time. There is no notion of overall progress, thus progress bar will be 0.SCollection display helpersScio interpreter comes with display helpers to ease working with Zeppelin notebooks. Simply use closeAndDisplay() on SCollection to close context and display the results. The number of results is limited by zeppelin.scio.maxResult (by default 1000).Supported SCollection types:Scio&#39;s typed BigQueryScala&#39;s Products (case classes, tuples)Google BigQuery&#39;s TableRowApache AvroAll Scala&#39;s AnyValHelper methodsThere are different helper methods for different objects. You can easily display results from SCollection, Future[Tap] and Tap.SCollection helperSCollect
ion has closeAndDisplay Zeppelin helper method for types listed above. Use it to synchronously close Scio context, once available pull and display results.Future[Tap] helperFuture[Tap] has waitAndDisplay Zeppelin helper method for types listed above. Use it synchronously wait for results, once available pull and display results.Tap helperTap has display Zeppelin helper method for types listed above. Use it to pull and display results.ExamplesBigQuery example:$beam.scio@BigQueryType.fromQuery(&quot;&quot;&quot;|SELECT departure_airport,count(case when departure_delay&gt;0 then 1 else 0 end) as no_of_delays |FROM [bigquery-samples:airline_ontime_data.flights] |group by departure_airport |order by 2 desc |limit 10&quot;&quot;&quot;.stripMargin) class Flightsval (sc, args) = ContextAndArgs(argz)sc.bigQuerySelect(Flights.query).closeAndDisplay(Flights.schema
)BigQuery typed example:$beam.scio@BigQueryType.fromQuery(&quot;&quot;&quot;|SELECT departure_airport,count(case when departure_delay&gt;0 then 1 else 0 end) as no_of_delays |FROM [bigquery-samples:airline_ontime_data.flights] |group by departure_airport |order by 2 desc |limit 10&quot;&quot;&quot;.stripMargin) class Flightsval (sc, args) = ContextAndArgs(argz)sc.typedBigQuery[Flights]().flatMap(_.no_of_delays).mean.closeAndDisplay()Avro example:$beam.scioimport com.spotify.data.ExampleAvroval (sc, args) = ContextAndArgs(argz)sc.avroFile[ExampleAvro](&quot;gs://&lt;bucket&gt;/tmp/my.avro&quot;).take(10).closeAndDisplay()Avro example with a view schema:$beam.scioimport com.spotify.data.ExampleAvroimport org.apache.avro.Schemaval (sc, args) = ContextAndArgs(argz)val view = Schema.parse(&quot;&quot;&quot;{&quot;type&quot
;:&quot;record&quot;,&quot;name&quot;:&quot;ExampleAvro&quot;,&quot;namespace&quot;:&quot;com.spotify.data&quot;,&quot;fields&quot;:[{&quot;name&quot;:&quot;track&quot;,&quot;type&quot;:&quot;string&quot;}, {&quot;name&quot;:&quot;artist&quot;, &quot;type&quot;:&quot;string&quot;}]}&quot;&quot;&quot;)sc.avroFile[EndSongCleaned](&quot;gs://&lt;bucket&gt;/tmp/my.avro&quot;).take(10).closeAndDisplay(view)Google credentialsScio Interpreter will try to infer your Google Cloud credentials from its environment, it will take into the account: * argz interpreter settings (doc) * environment variable (GOOGLE_APPLICATION_CREDENTIALS) * gcloud configurationBigQuery macro credentialsCurrently BigQuery project for macro expansion is inferred using Google Dataflow&#39;s DefaultProjectFactory().create()",
+ "url": " /interpreter/scio.html",
+ "group": "interpreter",
+ "excerpt": "Scio is a Scala DSL for Apache Beam/Google Dataflow model."
+ }
+ ,
+
"/interpreter/shell.html": {
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/security/authentication.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/security/authentication.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/security/authentication.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/security/authentication.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/security/datasource_authorization.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/security/datasource_authorization.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/security/datasource_authorization.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/security/datasource_authorization.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/security/notebook_authorization.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/security/notebook_authorization.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/security/notebook_authorization.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/security/notebook_authorization.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/security/shiroauthentication.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/security/shiroauthentication.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/security/shiroauthentication.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/security/shiroauthentication.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/sitemap.txt
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/sitemap.txt?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/sitemap.txt (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/sitemap.txt Mon Nov 14 22:43:45 2016
@@ -38,6 +38,7 @@ http://zeppelin.apache.org/interpreter/p
http://zeppelin.apache.org/interpreter/python.html
http://zeppelin.apache.org/interpreter/r.html
http://zeppelin.apache.org/interpreter/scalding.html
+http://zeppelin.apache.org/interpreter/scio.html
http://zeppelin.apache.org/interpreter/shell.html
http://zeppelin.apache.org/interpreter/spark.html
http://zeppelin.apache.org/manual/dependencymanagement.html
Modified: zeppelin/site/docs/0.7.0-SNAPSHOT/storage/storage.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.7.0-SNAPSHOT/storage/storage.html?rev=1769702&r1=1769701&r2=1769702&view=diff
==============================================================================
--- zeppelin/site/docs/0.7.0-SNAPSHOT/storage/storage.html (original)
+++ zeppelin/site/docs/0.7.0-SNAPSHOT/storage/storage.html Mon Nov 14 22:43:45 2016
@@ -116,6 +116,7 @@
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/postgresql.html">Postgresql, HAWQ</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/r.html">R</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scalding.html">Scalding</a></li>
+ <li><a href="/docs/0.7.0-SNAPSHOT/interpreter/scio.html">Scio</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/shell.html">Shell</a></li>
<li><a href="/docs/0.7.0-SNAPSHOT/interpreter/spark.html">Spark</a></li>
</ul>