You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@zeppelin.apache.org by ah...@apache.org on 2017/04/04 06:06:18 UTC
svn commit: r1790064 [2/2] - in /zeppelin/site/docs/0.8.0-SNAPSHOT: ./
assets/themes/zeppelin/img/screenshots/ development/ displaysystem/
install/ interpreter/ manual/ quickstart/ rest-api/ security/ storage/
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/manual/userimpersonation.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/manual/userimpersonation.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/manual/userimpersonation.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/manual/userimpersonation.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/pleasecontribute.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/pleasecontribute.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/pleasecontribute.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/pleasecontribute.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/explorezeppelinui.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/explorezeppelinui.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/explorezeppelinui.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/explorezeppelinui.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/install_with_flink_and_spark_cluster.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/tutorial.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/tutorial.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/tutorial.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/quickstart/tutorial.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-configuration.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-configuration.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-configuration.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-configuration.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-credential.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-credential.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-credential.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-credential.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-helium.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-helium.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-helium.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-helium.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-interpreter.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-interpreter.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-interpreter.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-interpreter.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-notebook.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-notebook.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-notebook.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-notebook.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-notebookRepo.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-notebookRepo.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-notebookRepo.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/rest-api/rest-notebookRepo.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/rss.xml
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/rss.xml?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/rss.xml (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/rss.xml Tue Apr 4 06:06:16 2017
@@ -5,8 +5,8 @@
<description>Apache Zeppelin - The Apache Software Foundation</description>
<link>http://zeppelin.apache.org</link>
<link>http://zeppelin.apache.org</link>
- <lastBuildDate>2017-04-01T18:59:50+09:00</lastBuildDate>
- <pubDate>2017-04-01T18:59:50+09:00</pubDate>
+ <lastBuildDate>2017-04-04T15:02:14+09:00</lastBuildDate>
+ <pubDate>2017-04-04T15:02:14+09:00</pubDate>
<ttl>1800</ttl>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/screenshots.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/screenshots.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/screenshots.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/screenshots.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/search.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/search.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/search.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/search.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/search_data.json
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/search_data.json?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/search_data.json (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/search_data.json Tue Apr 4 06:06:16 2017
@@ -279,6 +279,17 @@
+ "/interpreter/groovy.html": {
+ "title": "Apache Groovy Interpreter for Apache Zeppelin",
+ "content" : "<!--Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.-->Groovy Interpreter for Apache ZeppelinSamples%groovy//get a parameter defined as z.angularBind(&#39;ngSearchParam&#39;, value, &#39;paragraph_id&#39;)//g is a context object for groovy to avoid mix with z objectdef param = g.angular(&#39;ngSearchParam&#39;)//send request https://www.googleapis.com/customsearch/v1?q=ngSearchParam_valuedef r = HTTP.get( //assume you defined the groovy interpreter property //
`search_baseurl`=&#39;https://www.googleapis.com/customsearch/v1&#39; //in groovy object o.getProperty(&#39;A&#39;) == o.&#39;A&#39; == o.A == o[&#39;A&#39;] url : g.search_baseurl, query: [ q: param ], headers: [ &#39;Accept&#39;:&#39;application/json&#39;, //&#39;Authorization:&#39; : g.getProperty(&#39;search_auth&#39;), ] )//check response codeif( r.response.code==200 ) { g.html().with{ //g.html() renders %angular to output and returns groovy.xml.MarkupBuilder h2(&quot;the response ${r.response.code}&quot;) span( r.response.body ) h2(&quot;headers&quot;) pre( r.response.headers.join(&#39;n&#39;) ) }} else { //just to show that it&#39;s possible to use println with multiline groovy string to render output println(&quot;&quot;&quot;%angular &lt;script&gt; alert (&quot;code=${r.response.code} n msg=${r.response.message}&quot;)
&lt;/script&gt; &quot;&quot;&quot;)}%groovy//renders a table with headers a, b, c and two rowsg.table( [ [&#39;a&#39;,&#39;b&#39;,&#39;c&#39;], [&#39;a1&#39;,&#39;b1&#39;,&#39;c1&#39;], [&#39;a2&#39;,&#39;b2&#39;,&#39;c2&#39;], ])the g objectg.angular(String name)Returns angular object by name. Look up notebook scope first and then global scope.g.angularBind(String name, Object value)Assign a new value into angular object namejava.util.Properties g.getProperties()returns all properties defined for this interpreterString g.getProperty(&#39;PROPERTY_NAME&#39;) groovy g.PROPERTY_NAMEg.&#39;PROPERTY_NAME&#39;g[&#39;PROPERTY_NAME&#39;]g.getProperties().getProperty(&#39;PROPERTY_NAME&#39;)All above the accessor to named property defined in groovy interpreter.In this case with name PROPERTY_NAMEgroovy.xml.MarkupBuilder g.html()Starts or continues renderin
g of %angular to output and returns groovy.xml.MarkupBuilderMarkupBuilder is usefull to generate html (xml)void g.table(obj)starts or continues rendering table rows.obj: List(rows) of List(columns) where first line is a header ",
+ "url": " /interpreter/groovy.html",
+ "group": "interpreter",
+ "excerpt": "Apache Groovy is a powerful, optionally typed and dynamic language, with static-typing and static compilation capabilities, for the Java platform aimed at improving developer productivity thanks to a concise, familiar and easy to learn syntax."
+ }
+ ,
+
+
+
"/interpreter/hbase.html": {
"title": "HBase Shell Interpreter for Apache Zeppelin",
"content" : "<!--Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.-->HBase Shell Interpreter for Apache ZeppelinOverviewHBase Shell is a JRuby IRB client for Apache HBase. This interpreter provides all capabilities of Apache HBase shell within Apache Zeppelin. The interpreter assumes that Apache HBase client software has been installed and it can connect to the Apache HBase cluster from the machine on where Apache Zeppelin is installed.To get start with HBase, please see HBase Quickstart.HBase relea
se supportedBy default, Zeppelin is built against HBase 1.0.x releases. To work with HBase 1.1.x releases, use the following build command:# HBase 1.1.4mvn clean package -DskipTests -Phadoop-2.6 -Dhadoop.version=2.6.0 -P build-distr -Dhbase.hbase.version=1.1.4 -Dhbase.hadoop.version=2.6.0To work with HBase 1.2.0+, use the following build command:# HBase 1.2.0mvn clean package -DskipTests -Phadoop-2.6 -Dhadoop.version=2.6.0 -P build-distr -Dhbase.hbase.version=1.2.0 -Dhbase.hadoop.version=2.6.0Configuration Property Default Description hbase.home /usr/lib/hbase Installation directory of HBase, defaults to HBASE_HOME in environment hbase.ruby.sources lib/ruby Path to Ruby scripts relative to 'hbase.home' zeppelin.hbase.test.mode false Disable checks for unit and manual tests If you want to connect to HBase running on a cluster, you&#39;ll need to follow the next step.Export HBASE_HOMEIn conf/zeppelin-env.sh, export HBASE_H
OME environment variable with your HBase installation path. This ensures hbase-site.xml can be loaded.for exampleexport HBASE_HOME=/usr/lib/hbaseor, when running with CDHexport HBASE_HOME=&quot;/opt/cloudera/parcels/CDH/lib/hbase&quot;You can optionally export HBASE_CONF_DIR instead of HBASE_HOME should you have custom HBase configurations.Enabling the HBase Shell InterpreterIn a notebook, to enable the HBase Shell interpreter, click the Gear icon and select HBase Shell.Using the HBase Shell InterpreterIn a paragraph, use %hbase to select the HBase Shell interpreter and then input all commands. To get the list of available commands, use help.%hbasehelpFor example, to create a table%hbasecreate &#39;test&#39;, &#39;cf&#39;And then to put data into that table%hbaseput &#39;test&#39;, &#39;row1&#39;, &#39;cf:a&#39;, &#39;value1&#39;For more information on all commands available, refer to HBase shell commands.",
@@ -358,7 +369,7 @@
"/interpreter/livy.html": {
"title": "Livy Interpreter for Apache Zeppelin",
- "content" : "<!--Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.-->Livy Interpreter for Apache ZeppelinOverviewLivy is an open source REST interface for interacting with Spark from anywhere. It supports executing snippets of code or programs in a Spark context that runs locally or in YARN.Interactive Scala, Python and R shellsBatch submissions in Scala, Java, PythonMulti users can share the same server (impersonation support)Can be used for submitting jobs from anywhere with RESTDoes not require a
ny code change to your programsRequirementsAdditional requirements for the Livy interpreter are:Spark 1.3 or above.Livy server.ConfigurationWe added some common configurations for spark, and you can set any configuration you want.You can find all Spark configurations in here.And instead of starting property with spark. it should be replaced with livy.spark..Example: spark.driver.memory to livy.spark.driver.memory Property Default Description zeppelin.livy.url http://localhost:8998 URL where livy server is running zeppelin.livy.spark.maxResult 1000 Max number of Spark SQL result to display. zeppelin.livy.session.create_timeout 120 Timeout in seconds for session creation zeppelin.livy.displayAppInfo false Whether to display app info zeppelin.livy.pull_status.interval.millis 1000 The interval for checking paragraph execution status livy.spark.driver.cores Driver cores. ex) 1, 2. livy.spar
k.driver.memory Driver memory. ex) 512m, 32g. livy.spark.executor.instances Executor instances. ex) 1, 4. livy.spark.executor.cores Num cores per executor. ex) 1, 4. livy.spark.executor.memory Executor memory per worker instance. ex) 512m, 32g. livy.spark.dynamicAllocation.enabled Use dynamic resource allocation. ex) True, False. livy.spark.dynamicAllocation.cachedExecutorIdleTimeout Remove an executor which has cached data blocks. livy.spark.dynamicAllocation.minExecutors Lower bound for the number of executors. livy.spark.dynamicAllocation.initialExecutors Initial number of executors to run. livy.spark.dynamicAllocation.maxExecutors Upper bound for the number of executors. livy.spark.jars.packages Adding extra libraries to livy interpreter zeppelin.livy.ssl.trustStore client trustStore file. Used when livy ssl is enabled
zeppelin.livy.ssl.trustStorePassword password for trustStore file. Used when livy ssl is enabled We remove livy.spark.master in zeppelin-0.7. Because we sugguest user to use livy 0.3 in zeppelin-0.7. And livy 0.3 don&#39;t allow to specify livy.spark.master, it enfornce yarn-cluster mode.Adding External librariesYou can load dynamic library to livy interpreter by set livy.spark.jars.packages property to comma-separated list of maven coordinates of jars to include on the driver and executor classpaths. The format for the coordinates should be groupId:artifactId:version. Example Property Example Description livy.spark.jars.packages io.spray:spray-json_2.10:1.3.1 Adding extra libraries to livy interpreter How to useBasically, you can usespark%livy.sparksc.versionpyspark%livy.pysparkprint &quot;1&quot;sparkR%livy.sparkrhello &lt;- function( name ) { sprintf( &quot;Hello, %s&quot;, name );}hello(&quot;liv
y&quot;)ImpersonationWhen Zeppelin server is running with authentication enabled, then this interpreter utilizes Livy’s user impersonation feature i.e. sends extra parameter for creating and running a session (&quot;proxyUser&quot;: &quot;${loggedInUser}&quot;). This is particularly useful when multi users are sharing a Notebook server.Apply Zeppelin Dynamic FormsYou can leverage Zeppelin Dynamic Form. You can use both the text input and select form parameterization features.%livy.pysparkprint &quot;${group_by=product_id,product_id|product_name|customer_id|store_id}&quot;FAQLivy debugging: If you see any of these in error consoleConnect to livyhost:8998 [livyhost/127.0.0.1, livyhost/0:0:0:0:0:0:0:1] failed: Connection refusedLooks like the livy server is not up yet or the config is wrongException: Session not found, Livy server would have restarted, or lost session.The session would have timed out, you may need to restart the interpreter.Blacklisted
configuration values in session config: spark.masterEdit conf/spark-blacklist.conf file in livy server and comment out #spark.master line.If you choose to work on livy in apps/spark/java directory in https://github.com/cloudera/hue,copy spark-user-configurable-options.template to spark-user-configurable-options.conf file in livy server and comment out #spark.master. ",
+ "content" : "<!--Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.-->Livy Interpreter for Apache ZeppelinOverviewLivy is an open source REST interface for interacting with Spark from anywhere. It supports executing snippets of code or programs in a Spark context that runs locally or in YARN.Interactive Scala, Python and R shellsBatch submissions in Scala, Java, PythonMulti users can share the same server (impersonation support)Can be used for submitting jobs from anywhere with RESTDoes not require a
ny code change to your programsRequirementsAdditional requirements for the Livy interpreter are:Spark 1.3 or above.Livy server.ConfigurationWe added some common configurations for spark, and you can set any configuration you want.You can find all Spark configurations in here.And instead of starting property with spark. it should be replaced with livy.spark..Example: spark.driver.memory to livy.spark.driver.memory Property Default Description zeppelin.livy.url http://localhost:8998 URL where livy server is running zeppelin.livy.spark.sql.maxResult 1000 Max number of Spark SQL result to display. zeppelin.livy.spark.sql.field.truncate true Whether to truncate field values longer than 20 characters or not zeppelin.livy.session.create_timeout 120 Timeout in seconds for session creation zeppelin.livy.displayAppInfo false Whether to display app info zeppelin.livy.pull_status.interval.millis 1000 The in
terval for checking paragraph execution status livy.spark.driver.cores Driver cores. ex) 1, 2. livy.spark.driver.memory Driver memory. ex) 512m, 32g. livy.spark.executor.instances Executor instances. ex) 1, 4. livy.spark.executor.cores Num cores per executor. ex) 1, 4. livy.spark.executor.memory Executor memory per worker instance. ex) 512m, 32g. livy.spark.dynamicAllocation.enabled Use dynamic resource allocation. ex) True, False. livy.spark.dynamicAllocation.cachedExecutorIdleTimeout Remove an executor which has cached data blocks. livy.spark.dynamicAllocation.minExecutors Lower bound for the number of executors. livy.spark.dynamicAllocation.initialExecutors Initial number of executors to run. livy.spark.dynamicAllocation.maxExecutors Upper bound for the number of executors. livy.spark.jars.packages Adding extra lib
raries to livy interpreter zeppelin.livy.ssl.trustStore client trustStore file. Used when livy ssl is enabled zeppelin.livy.ssl.trustStorePassword password for trustStore file. Used when livy ssl is enabled We remove livy.spark.master in zeppelin-0.7. Because we sugguest user to use livy 0.3 in zeppelin-0.7. And livy 0.3 don&#39;t allow to specify livy.spark.master, it enfornce yarn-cluster mode.Adding External librariesYou can load dynamic library to livy interpreter by set livy.spark.jars.packages property to comma-separated list of maven coordinates of jars to include on the driver and executor classpaths. The format for the coordinates should be groupId:artifactId:version. Example Property Example Description livy.spark.jars.packages io.spray:spray-json_2.10:1.3.1 Adding extra libraries to livy interpreter How to useBasically, you can usespark%livy.sparksc.versionpyspark%livy.pysparkprint &quot;1&q
uot;sparkR%livy.sparkrhello &lt;- function( name ) { sprintf( &quot;Hello, %s&quot;, name );}hello(&quot;livy&quot;)ImpersonationWhen Zeppelin server is running with authentication enabled, then this interpreter utilizes Livy’s user impersonation feature i.e. sends extra parameter for creating and running a session (&quot;proxyUser&quot;: &quot;${loggedInUser}&quot;). This is particularly useful when multi users are sharing a Notebook server.Apply Zeppelin Dynamic FormsYou can leverage Zeppelin Dynamic Form. You can use both the text input and select form parameterization features.%livy.pysparkprint &quot;${group_by=product_id,product_id|product_name|customer_id|store_id}&quot;FAQLivy debugging: If you see any of these in error consoleConnect to livyhost:8998 [livyhost/127.0.0.1, livyhost/0:0:0:0:0:0:0:1] failed: Connection refusedLooks like the livy server is not up yet or the config is wrongException: Session not found, Livy serv
er would have restarted, or lost session.The session would have timed out, you may need to restart the interpreter.Blacklisted configuration values in session config: spark.masterEdit conf/spark-blacklist.conf file in livy server and comment out #spark.master line.If you choose to work on livy in apps/spark/java directory in https://github.com/cloudera/hue,copy spark-user-configurable-options.template to spark-user-configurable-options.conf file in livy server and comment out #spark.master. ",
"url": " /interpreter/livy.html",
"group": "interpreter",
"excerpt": "Livy is an open source REST interface for interacting with Spark from anywhere. It supports executing snippets of code or programs in a Spark context that runs locally or in YARN."
@@ -468,7 +479,7 @@
"/interpreter/spark.html": {
"title": "Apache Spark Interpreter for Apache Zeppelin",
- "content" : "<!--Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.-->Spark Interpreter for Apache ZeppelinOverviewApache Spark is a fast and general-purpose cluster computing system.It provides high-level APIs in Java, Scala, Python and R, and an optimized engine that supports general execution graphs.Apache Spark is supported in Zeppelin with Spark interpreter group which consists of below five interpreters. Name Class Description %spark SparkInterpreter Creates a SparkConte
xt and provides a Scala environment %spark.pyspark PySparkInterpreter Provides a Python environment %spark.r SparkRInterpreter Provides an R environment with SparkR support %spark.sql SparkSQLInterpreter Provides a SQL environment %spark.dep DepInterpreter Dependency loader ConfigurationThe Spark interpreter can be configured with properties provided by Zeppelin.You can also set other Spark properties which are not listed in the table. For a list of additional properties, refer to Spark Available Properties. Property Default Description args Spark commandline args master local[*] Spark master uri. ex) spark://masterhost:7077 spark.app.name Zeppelin The name of spark application. spark.cores.max Total number of cores to use. Empty value uses all available core. spark.executor.memory 1g Executor memory per worker instance. ex) 512m, 32g zeppelin.dep
.additionalRemoteRepository spark-packages, http://dl.bintray.com/spark-packages/maven, false; A list of id,remote-repository-URL,is-snapshot; for each remote repository. zeppelin.dep.localrepo local-repo Local repository for dependency loader PYSPARKPYTHON python Python binary executable to use for PySpark in both driver and workers (default is python). Property spark.pyspark.python take precedence if it is set PYSPARKDRIVERPYTHON python Python binary executable to use for PySpark in driver only (default is PYSPARKPYTHON). Property spark.pyspark.driver.python take precedence if it is set zeppelin.spark.concurrentSQL false Execute multiple SQL concurrently if set true. zeppelin.spark.maxResult 1000 Max number of Spark SQL result to display. zeppelin.spark.printREPLOutput true Print REPL output zeppelin.spark.useHiveContext true Use HiveContext instead of SQLConte
xt if it is true. zeppelin.spark.importImplicit true Import implicits, UDF collection, and sql if set true. Without any configuration, Spark interpreter works out of box in local mode. But if you want to connect to your Spark cluster, you&#39;ll need to follow below two simple steps.1. Export SPARK_HOMEIn conf/zeppelin-env.sh, export SPARK_HOME environment variable with your Spark installation path.For example,export SPARK_HOME=/usr/lib/sparkYou can optionally set more environment variables# set hadoop conf direxport HADOOP_CONF_DIR=/usr/lib/hadoop# set options to pass spark-submit commandexport SPARK_SUBMIT_OPTIONS=&quot;--packages com.databricks:spark-csv_2.10:1.2.0&quot;# extra classpath. e.g. set classpath for hive-site.xmlexport ZEPPELIN_INTP_CLASSPATH_OVERRIDES=/etc/hive/confFor Windows, ensure you have winutils.exe in %HADOOP_HOME%bin. Please see Problems running Hadoop on Windows for the details.2. Set master in Interpreter menuAfter start Zeppelin,
go to Interpreter menu and edit master property in your Spark interpreter setting. The value may vary depending on your Spark cluster deployment type.For example,local[*] in local modespark://master:7077 in standalone clusteryarn-client in Yarn client modemesos://host:5050 in Mesos clusterThat&#39;s it. Zeppelin will work with any version of Spark and any deployment type without rebuilding Zeppelin in this way. For the further information about Spark &amp; Zeppelin version compatibility, please refer to &quot;Available Interpreters&quot; section in Zeppelin download page.Note that without exporting SPARK_HOME, it&#39;s running in local mode with included version of Spark. The included version may vary depending on the build profile.SparkContext, SQLContext, SparkSession, ZeppelinContextSparkContext, SQLContext and ZeppelinContext are automatically created and exposed as variable names sc, sqlContext and z, respectively, in Scala, Python and R environments.Starin
g from 0.6.1 SparkSession is available as variable spark when you are using Spark 2.x.Note that Scala/Python/R environment shares the same SparkContext, SQLContext and ZeppelinContext instance. Dependency ManagementThere are two ways to load external libraries in Spark interpreter. First is using interpreter setting menu and second is loading Spark properties.1. Setting Dependencies via Interpreter SettingPlease see Dependency Management for the details.2. Loading Spark PropertiesOnce SPARK_HOME is set in conf/zeppelin-env.sh, Zeppelin uses spark-submit as spark interpreter runner. spark-submit supports two ways to load configurations. The first is command line options such as --master and Zeppelin can pass these options to spark-submit by exporting SPARK_SUBMIT_OPTIONS in conf/zeppelin-env.sh. Second is reading configuration options from SPARK_HOME/conf/spark-defaults.conf. Spark properties that user can set to distribute libraries are: spark-defaults.conf SPARK_SUBMIT_OPTI
ONS Description spark.jars --jars Comma-separated list of local jars to include on the driver and executor classpaths. spark.jars.packages --packages Comma-separated list of maven coordinates of jars to include on the driver and executor classpaths. Will search the local maven repo, then maven central and any additional remote repositories given by --repositories. The format for the coordinates should be groupId:artifactId:version. spark.files --files Comma-separated list of files to be placed in the working directory of each executor. Here are few examples:SPARK_SUBMIT_OPTIONS in conf/zeppelin-env.shexport SPARK_SUBMIT_OPTIONS=&quot;--packages com.databricks:spark-csv_2.10:1.2.0 --jars /path/mylib1.jar,/path/mylib2.jar --files /path/mylib1.py,/path/mylib2.zip,/path/mylib3.egg&quot;SPARK_HOME/conf/spark-defaults.confspark.jars /path/mylib1.jar,/path/mylib2.jarspark.jars.packages com.databricks:spark-csv_2.10:1.2.0spark.file
s /path/mylib1.py,/path/mylib2.egg,/path/mylib3.zip3. Dynamic Dependency Loading via %spark.dep interpreterNote: %spark.dep interpreter loads libraries to %spark and %spark.pyspark but not to %spark.sql interpreter. So we recommend you to use the first option instead.When your code requires external library, instead of doing download/copy/restart Zeppelin, you can easily do following jobs using %spark.dep interpreter.Load libraries recursively from maven repositoryLoad libraries from local filesystemAdd additional maven repositoryAutomatically add libraries to SparkCluster (You can turn off)Dep interpreter leverages Scala environment. So you can write any Scala code here.Note that %spark.dep interpreter should be used before %spark, %spark.pyspark, %spark.sql.Here&#39;s usages.%spark.depz.reset() // clean up previously added artifact and repository// add maven repositoryz.addRepo(&quot;RepoName&quot;).url(&quot;RepoURL&quot;)// add maven snapshot repositor
yz.addRepo(&quot;RepoName&quot;).url(&quot;RepoURL&quot;).snapshot()// add credentials for private maven repositoryz.addRepo(&quot;RepoName&quot;).url(&quot;RepoURL&quot;).username(&quot;username&quot;).password(&quot;password&quot;)// add artifact from filesystemz.load(&quot;/path/to.jar&quot;)// add artifact from maven repository, with no dependencyz.load(&quot;groupId:artifactId:version&quot;).excludeAll()// add artifact recursivelyz.load(&quot;groupId:artifactId:version&quot;)// add artifact recursively except comma separated GroupID:ArtifactId listz.load(&quot;groupId:artifactId:version&quot;).exclude(&quot;groupId:artifactId,groupId:artifactId, ...&quot;)// exclude with patternz.load(&quot;groupId:artifactId:version&quot;).exclude(*)z.load(&quot;groupId:artifactId:version&quot;).exclude(&quot;groupId:artifactId:*&quot;)z.load(&quot;groupId:artifactId:version&
quot;).exclude(&quot;groupId:*&quot;)// local() skips adding artifact to spark clusters (skipping sc.addJar())z.load(&quot;groupId:artifactId:version&quot;).local()ZeppelinContextZeppelin automatically injects ZeppelinContext as variable z in your Scala/Python environment. ZeppelinContext provides some additional functions and utilities.Object ExchangeZeppelinContext extends map and it&#39;s shared between Scala and Python environment.So you can put some objects from Scala and read it from Python, vice versa. // Put object from scala%sparkval myObject = ...z.put(&quot;objName&quot;, myObject)// Exchanging data framesmyScalaDataFrame = ...z.put(&quot;myScalaDataFrame&quot;, myScalaDataFrame)val myPythonDataFrame = z.get(&quot;myPythonDataFrame&quot;).asInstanceOf[DataFrame] # Get object from python%spark.pysparkmyObject = z.get(&quot;objName&quot;)# Exchanging data framesmyPythonDataFrame = ...z.put(&quot;myPythonDataFrame&a
mp;quot;, postsDf._jdf)myScalaDataFrame = DataFrame(z.get(&quot;myScalaDataFrame&quot;), sqlContext) Form CreationZeppelinContext provides functions for creating forms.In Scala and Python environments, you can create forms programmatically. %spark/* Create text input form */z.input(&quot;formName&quot;)/* Create text input form with default value */z.input(&quot;formName&quot;, &quot;defaultValue&quot;)/* Create select form */z.select(&quot;formName&quot;, Seq((&quot;option1&quot;, &quot;option1DisplayName&quot;), (&quot;option2&quot;, &quot;option2DisplayName&quot;)))/* Create select form with default value*/z.select(&quot;formName&quot;, &quot;option1&quot;, Seq((&quot;option1&quot;, &quot;option1DisplayName&quot;), (&quot;option2&quot;, &quot;option2DisplayName&quot;))) %spark.pyspark# Create text i
nput formz.input(&quot;formName&quot;)# Create text input form with default valuez.input(&quot;formName&quot;, &quot;defaultValue&quot;)# Create select formz.select(&quot;formName&quot;, [(&quot;option1&quot;, &quot;option1DisplayName&quot;), (&quot;option2&quot;, &quot;option2DisplayName&quot;)])# Create select form with default valuez.select(&quot;formName&quot;, [(&quot;option1&quot;, &quot;option1DisplayName&quot;), (&quot;option2&quot;, &quot;option2DisplayName&quot;)], &quot;option1&quot;) In sql environment, you can create form in simple template.%spark.sqlselect * from ${table=defaultTableName} where text like &#39;%${search}%&#39;To learn more about dynamic form, checkout Dynamic Form.Matplotlib Integration (pyspark)Both the python and pyspark interpreters have built-in support for inline visualization using matpl
otlib, a popular plotting library for python. More details can be found in the python interpreter documentation, since matplotlib support is identical. More advanced interactive plotting can be done with pyspark through utilizing Zeppelin&#39;s built-in Angular Display System, as shown below:Interpreter setting optionYou can choose one of shared, scoped and isolated options wheh you configure Spark interpreter. Spark interpreter creates separated Scala compiler per each notebook but share a single SparkContext in scoped mode (experimental). It creates separated SparkContext per each notebook in isolated mode.Setting up Zeppelin with KerberosLogical setup with Zeppelin, Kerberos Key Distribution Center (KDC), and Spark on YARN:Configuration SetupOn the server that Zeppelin is installed, install Kerberos client modules and configuration, krb5.conf.This is to make the server communicate with KDC.Set SPARK_HOME in [ZEPPELIN_HOME]/conf/zeppelin-env.sh to use spark-submit(Additionally
, you might have to set export HADOOP_CONF_DIR=/etc/hadoop/conf)Add the two properties below to Spark configuration ([SPARK_HOME]/conf/spark-defaults.conf):spark.yarn.principalspark.yarn.keytabNOTE: If you do not have permission to access for the above spark-defaults.conf file, optionally, you can add the above lines to the Spark Interpreter setting through the Interpreter tab in the Zeppelin UI.That&#39;s it. Play with Zeppelin!",
+ "content" : "<!--Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.-->Spark Interpreter for Apache ZeppelinOverviewApache Spark is a fast and general-purpose cluster computing system.It provides high-level APIs in Java, Scala, Python and R, and an optimized engine that supports general execution graphs.Apache Spark is supported in Zeppelin with Spark interpreter group which consists of below five interpreters. Name Class Description %spark SparkInterpreter Creates a SparkConte
xt and provides a Scala environment %spark.pyspark PySparkInterpreter Provides a Python environment %spark.r SparkRInterpreter Provides an R environment with SparkR support %spark.sql SparkSQLInterpreter Provides a SQL environment %spark.dep DepInterpreter Dependency loader ConfigurationThe Spark interpreter can be configured with properties provided by Zeppelin.You can also set other Spark properties which are not listed in the table. For a list of additional properties, refer to Spark Available Properties. Property Default Description args Spark commandline args master local[*] Spark master uri. ex) spark://masterhost:7077 spark.app.name Zeppelin The name of spark application. spark.cores.max Total number of cores to use. Empty value uses all available core. spark.executor.memory 1g Executor memory per worker instance. ex) 512m, 32g zeppelin.dep
.additionalRemoteRepository spark-packages, http://dl.bintray.com/spark-packages/maven, false; A list of id,remote-repository-URL,is-snapshot; for each remote repository. zeppelin.dep.localrepo local-repo Local repository for dependency loader PYSPARKPYTHON python Python binary executable to use for PySpark in both driver and workers (default is python). Property spark.pyspark.python take precedence if it is set PYSPARKDRIVERPYTHON python Python binary executable to use for PySpark in driver only (default is PYSPARKPYTHON). Property spark.pyspark.driver.python take precedence if it is set zeppelin.spark.concurrentSQL false Execute multiple SQL concurrently if set true. zeppelin.spark.maxResult 1000 Max number of Spark SQL result to display. zeppelin.spark.printREPLOutput true Print REPL output zeppelin.spark.useHiveContext true Use HiveContext instead of SQLConte
xt if it is true. zeppelin.spark.importImplicit true Import implicits, UDF collection, and sql if set true. zeppelin.spark.enableSupportedVersionCheck true Do not change - developer only setting, not for production use Without any configuration, Spark interpreter works out of box in local mode. But if you want to connect to your Spark cluster, you&#39;ll need to follow below two simple steps.1. Export SPARK_HOMEIn conf/zeppelin-env.sh, export SPARK_HOME environment variable with your Spark installation path.For example,export SPARK_HOME=/usr/lib/sparkYou can optionally set more environment variables# set hadoop conf direxport HADOOP_CONF_DIR=/usr/lib/hadoop# set options to pass spark-submit commandexport SPARK_SUBMIT_OPTIONS=&quot;--packages com.databricks:spark-csv_2.10:1.2.0&quot;# extra classpath. e.g. set classpath for hive-site.xmlexport ZEPPELIN_INTP_CLASSPATH_OVERRIDES=/etc/hive/confFor Windows, ensure you have winutils.exe in %HADOOP_HO
ME%bin. Please see Problems running Hadoop on Windows for the details.2. Set master in Interpreter menuAfter start Zeppelin, go to Interpreter menu and edit master property in your Spark interpreter setting. The value may vary depending on your Spark cluster deployment type.For example,local[*] in local modespark://master:7077 in standalone clusteryarn-client in Yarn client modemesos://host:5050 in Mesos clusterThat&#39;s it. Zeppelin will work with any version of Spark and any deployment type without rebuilding Zeppelin in this way. For the further information about Spark &amp; Zeppelin version compatibility, please refer to &quot;Available Interpreters&quot; section in Zeppelin download page.Note that without exporting SPARK_HOME, it&#39;s running in local mode with included version of Spark. The included version may vary depending on the build profile.SparkContext, SQLContext, SparkSession, ZeppelinContextSparkContext, SQLContext and ZeppelinContext are automa
tically created and exposed as variable names sc, sqlContext and z, respectively, in Scala, Python and R environments.Staring from 0.6.1 SparkSession is available as variable spark when you are using Spark 2.x.Note that Scala/Python/R environment shares the same SparkContext, SQLContext and ZeppelinContext instance. Dependency ManagementThere are two ways to load external libraries in Spark interpreter. First is using interpreter setting menu and second is loading Spark properties.1. Setting Dependencies via Interpreter SettingPlease see Dependency Management for the details.2. Loading Spark PropertiesOnce SPARK_HOME is set in conf/zeppelin-env.sh, Zeppelin uses spark-submit as spark interpreter runner. spark-submit supports two ways to load configurations. The first is command line options such as --master and Zeppelin can pass these options to spark-submit by exporting SPARK_SUBMIT_OPTIONS in conf/zeppelin-env.sh. Second is reading configuration options from SPARK_HOME/conf/spark-
defaults.conf. Spark properties that user can set to distribute libraries are: spark-defaults.conf SPARK_SUBMIT_OPTIONS Description spark.jars --jars Comma-separated list of local jars to include on the driver and executor classpaths. spark.jars.packages --packages Comma-separated list of maven coordinates of jars to include on the driver and executor classpaths. Will search the local maven repo, then maven central and any additional remote repositories given by --repositories. The format for the coordinates should be groupId:artifactId:version. spark.files --files Comma-separated list of files to be placed in the working directory of each executor. Here are few examples:SPARK_SUBMIT_OPTIONS in conf/zeppelin-env.shexport SPARK_SUBMIT_OPTIONS=&quot;--packages com.databricks:spark-csv_2.10:1.2.0 --jars /path/mylib1.jar,/path/mylib2.jar --files /path/mylib1.py,/path/mylib2.zip,/path/mylib3.egg&quot;SPARK_HOME/conf/spark-default
s.confspark.jars /path/mylib1.jar,/path/mylib2.jarspark.jars.packages com.databricks:spark-csv_2.10:1.2.0spark.files /path/mylib1.py,/path/mylib2.egg,/path/mylib3.zip3. Dynamic Dependency Loading via %spark.dep interpreterNote: %spark.dep interpreter loads libraries to %spark and %spark.pyspark but not to %spark.sql interpreter. So we recommend you to use the first option instead.When your code requires external library, instead of doing download/copy/restart Zeppelin, you can easily do following jobs using %spark.dep interpreter.Load libraries recursively from maven repositoryLoad libraries from local filesystemAdd additional maven repositoryAutomatically add libraries to SparkCluster (You can turn off)Dep interpreter leverages Scala environment. So you can write any Scala code here.Note that %spark.dep interpreter should be used before %spark, %spark.pyspark, %spark.sql.Here&#39;s usages.%spark.depz.reset() // clean up previously added artifact and repository//
add maven repositoryz.addRepo(&quot;RepoName&quot;).url(&quot;RepoURL&quot;)// add maven snapshot repositoryz.addRepo(&quot;RepoName&quot;).url(&quot;RepoURL&quot;).snapshot()// add credentials for private maven repositoryz.addRepo(&quot;RepoName&quot;).url(&quot;RepoURL&quot;).username(&quot;username&quot;).password(&quot;password&quot;)// add artifact from filesystemz.load(&quot;/path/to.jar&quot;)// add artifact from maven repository, with no dependencyz.load(&quot;groupId:artifactId:version&quot;).excludeAll()// add artifact recursivelyz.load(&quot;groupId:artifactId:version&quot;)// add artifact recursively except comma separated GroupID:ArtifactId listz.load(&quot;groupId:artifactId:version&quot;).exclude(&quot;groupId:artifactId,groupId:artifactId, ...&quot;)// exclude with patternz.load(&quot;groupId:artifactId:version&quot;).exclude(*)z.load(&quot;groupId:art
ifactId:version&quot;).exclude(&quot;groupId:artifactId:*&quot;)z.load(&quot;groupId:artifactId:version&quot;).exclude(&quot;groupId:*&quot;)// local() skips adding artifact to spark clusters (skipping sc.addJar())z.load(&quot;groupId:artifactId:version&quot;).local()ZeppelinContextZeppelin automatically injects ZeppelinContext as variable z in your Scala/Python environment. ZeppelinContext provides some additional functions and utilities.Object ExchangeZeppelinContext extends map and it&#39;s shared between Scala and Python environment.So you can put some objects from Scala and read it from Python, vice versa. // Put object from scala%sparkval myObject = ...z.put(&quot;objName&quot;, myObject)// Exchanging data framesmyScalaDataFrame = ...z.put(&quot;myScalaDataFrame&quot;, myScalaDataFrame)val myPythonDataFrame = z.get(&quot;myPythonDataFrame&quot;).asInstanceOf[DataFrame] # Get object from python%spark.pysparkmyO
bject = z.get(&quot;objName&quot;)# Exchanging data framesmyPythonDataFrame = ...z.put(&quot;myPythonDataFrame&quot;, postsDf._jdf)myScalaDataFrame = DataFrame(z.get(&quot;myScalaDataFrame&quot;), sqlContext) Form CreationZeppelinContext provides functions for creating forms.In Scala and Python environments, you can create forms programmatically. %spark/* Create text input form */z.input(&quot;formName&quot;)/* Create text input form with default value */z.input(&quot;formName&quot;, &quot;defaultValue&quot;)/* Create select form */z.select(&quot;formName&quot;, Seq((&quot;option1&quot;, &quot;option1DisplayName&quot;), (&quot;option2&quot;, &quot;option2DisplayName&quot;)))/* Create select form with default value*/z.select(&quot;formName&quot;, &quot;option1&quot;, Seq((&quot;option1&quot;, &quot;option1DisplayName&quot;),
(&quot;option2&quot;, &quot;option2DisplayName&quot;))) %spark.pyspark# Create text input formz.input(&quot;formName&quot;)# Create text input form with default valuez.input(&quot;formName&quot;, &quot;defaultValue&quot;)# Create select formz.select(&quot;formName&quot;, [(&quot;option1&quot;, &quot;option1DisplayName&quot;), (&quot;option2&quot;, &quot;option2DisplayName&quot;)])# Create select form with default valuez.select(&quot;formName&quot;, [(&quot;option1&quot;, &quot;option1DisplayName&quot;), (&quot;option2&quot;, &quot;option2DisplayName&quot;)], &quot;option1&quot;) In sql environment, you can create form in simple template.%spark.sqlselect * from ${table=defaultTableName} where text like &#39;%${search}%&#39;To learn more about dynamic form, checkout Dynamic Form.Matplot
lib Integration (pyspark)Both the python and pyspark interpreters have built-in support for inline visualization using matplotlib, a popular plotting library for python. More details can be found in the python interpreter documentation, since matplotlib support is identical. More advanced interactive plotting can be done with pyspark through utilizing Zeppelin&#39;s built-in Angular Display System, as shown below:Interpreter setting optionYou can choose one of shared, scoped and isolated options wheh you configure Spark interpreter. Spark interpreter creates separated Scala compiler per each notebook but share a single SparkContext in scoped mode (experimental). It creates separated SparkContext per each notebook in isolated mode.Setting up Zeppelin with KerberosLogical setup with Zeppelin, Kerberos Key Distribution Center (KDC), and Spark on YARN:Configuration SetupOn the server that Zeppelin is installed, install Kerberos client modules and configuration, krb5.conf.This is to
make the server communicate with KDC.Set SPARK_HOME in [ZEPPELIN_HOME]/conf/zeppelin-env.sh to use spark-submit(Additionally, you might have to set export HADOOP_CONF_DIR=/etc/hadoop/conf)Add the two properties below to Spark configuration ([SPARK_HOME]/conf/spark-defaults.conf):spark.yarn.principalspark.yarn.keytabNOTE: If you do not have permission to access for the above spark-defaults.conf file, optionally, you can add the above lines to the Spark Interpreter setting through the Interpreter tab in the Zeppelin UI.That&#39;s it. Play with Zeppelin!",
"url": " /interpreter/spark.html",
"group": "interpreter",
"excerpt": "Apache Spark is a fast and general-purpose cluster computing system. It provides high-level APIs in Java, Scala, Python and R, and an optimized engine that supports general execution engine."
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/security/authentication.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/security/authentication.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/security/authentication.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/security/authentication.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/security/datasource_authorization.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/security/datasource_authorization.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/security/datasource_authorization.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/security/datasource_authorization.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/security/notebook_authorization.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/security/notebook_authorization.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/security/notebook_authorization.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/security/notebook_authorization.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/security/shiroauthentication.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/security/shiroauthentication.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/security/shiroauthentication.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/security/shiroauthentication.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/sitemap.txt
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/sitemap.txt?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/sitemap.txt (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/sitemap.txt Tue Apr 4 06:06:16 2017
@@ -27,6 +27,7 @@ http://zeppelin.apache.org/interpreter/c
http://zeppelin.apache.org/interpreter/elasticsearch.html
http://zeppelin.apache.org/interpreter/flink.html
http://zeppelin.apache.org/interpreter/geode.html
+http://zeppelin.apache.org/interpreter/groovy.html
http://zeppelin.apache.org/interpreter/hbase.html
http://zeppelin.apache.org/interpreter/hdfs.html
http://zeppelin.apache.org/interpreter/hive.html
Modified: zeppelin/site/docs/0.8.0-SNAPSHOT/storage/storage.html
URL: http://svn.apache.org/viewvc/zeppelin/site/docs/0.8.0-SNAPSHOT/storage/storage.html?rev=1790064&r1=1790063&r2=1790064&view=diff
==============================================================================
--- zeppelin/site/docs/0.8.0-SNAPSHOT/storage/storage.html (original)
+++ zeppelin/site/docs/0.8.0-SNAPSHOT/storage/storage.html Tue Apr 4 06:06:16 2017
@@ -108,6 +108,7 @@
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/elasticsearch.html">Elasticsearch</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/flink.html">Flink</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/geode.html">Geode</a></li>
+ <li><a href="/docs/0.8.0-SNAPSHOT/interpreter/groovy.html">Groovy</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hbase.html">HBase</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hdfs.html">HDFS</a></li>
<li><a href="/docs/0.8.0-SNAPSHOT/interpreter/hive.html">Hive</a></li>