You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@bahir.apache.org by lr...@apache.org on 2020/03/14 04:28:28 UTC

[bahir-website] branch asf-site updated: Publishing from 7e902acd3d8358dd288d4c0bc90e34b8adb00e24

This is an automated email from the ASF dual-hosted git repository.

lresende pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/bahir-website.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new cb1bec3  Publishing from 7e902acd3d8358dd288d4c0bc90e34b8adb00e24
cb1bec3 is described below

commit cb1bec381f06130f30ca51959d3d8706c1e9b000
Author: Luciano Resende <lr...@apache.org>
AuthorDate: Fri Mar 13 21:28:12 2020 -0700

    Publishing from 7e902acd3d8358dd288d4c0bc90e34b8adb00e24
---
 content/community-members/index.html               |  20 +-
 content/community/index.html                       |  20 +-
 content/contributing-extensions/index.html         |  20 +-
 content/contributing/index.html                    |  32 +-
 content/docs/flink/1.0/documentation/index.html    |  20 +-
 .../flink/1.0/flink-streaming-activemq/index.html  |  25 +-
 .../docs/flink/1.0/flink-streaming-akka/index.html |  36 ++-
 .../flink/1.0/flink-streaming-flume/index.html     |  28 +-
 .../flink/1.0/flink-streaming-netty/index.html     |  70 +++--
 .../flink/1.0/flink-streaming-redis/index.html     |  41 ++-
 .../docs/flink/current/documentation/index.html    |  20 +-
 .../current/flink-streaming-activemq/index.html    |  25 +-
 .../flink/current/flink-streaming-akka/index.html  |  36 ++-
 .../flink/current/flink-streaming-flume/index.html |  28 +-
 .../current/flink-streaming-influxdb/index.html    |  26 +-
 .../flink/current/flink-streaming-kudu/index.html  | 125 ++++----
 .../flink/current/flink-streaming-netty/index.html |  70 +++--
 .../flink/current/flink-streaming-redis/index.html |  41 ++-
 content/docs/flink/overview/index.html             |  20 +-
 content/docs/spark/2.0.0/documentation/index.html  |  20 +-
 .../2.0.0/spark-sql-streaming-mqtt/index.html      |  53 ++--
 .../spark/2.0.0/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.0.0/spark-streaming-mqtt/index.html    |  45 ++-
 .../spark/2.0.0/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.0.0/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.0.1/documentation/index.html  |  20 +-
 .../2.0.1/spark-sql-streaming-mqtt/index.html      |  75 +++--
 .../spark/2.0.1/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.0.1/spark-streaming-mqtt/index.html    |  72 +++--
 .../spark/2.0.1/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.0.1/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.0.2/documentation/index.html  |  20 +-
 .../2.0.2/spark-sql-streaming-mqtt/index.html      |  75 +++--
 .../spark/2.0.2/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.0.2/spark-streaming-mqtt/index.html    |  72 +++--
 .../spark/2.0.2/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.0.2/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.1.0/documentation/index.html  |  20 +-
 .../2.1.0/spark-sql-streaming-mqtt/index.html      |  75 +++--
 .../spark/2.1.0/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.1.0/spark-streaming-mqtt/index.html    |  72 +++--
 .../spark/2.1.0/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.1.0/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.1.1/documentation/index.html  |  20 +-
 .../docs/spark/2.1.1/spark-sql-cloudant/index.html | 302 +++++++++---------
 .../2.1.1/spark-sql-streaming-akka/index.html      |  57 ++--
 .../2.1.1/spark-sql-streaming-mqtt/index.html      |  75 +++--
 .../spark/2.1.1/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.1.1/spark-streaming-mqtt/index.html    |  74 +++--
 .../spark/2.1.1/spark-streaming-pubsub/index.html  |  49 +--
 .../spark/2.1.1/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.1.1/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.1.2/documentation/index.html  |  20 +-
 .../docs/spark/2.1.2/spark-sql-cloudant/index.html | 338 +++++++++++----------
 .../2.1.2/spark-sql-streaming-akka/index.html      |  57 ++--
 .../2.1.2/spark-sql-streaming-mqtt/index.html      |  75 +++--
 .../spark/2.1.2/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.1.2/spark-streaming-mqtt/index.html    |  84 ++---
 .../spark/2.1.2/spark-streaming-pubsub/index.html  |  71 +++--
 .../spark/2.1.2/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.1.2/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.1.3/documentation/index.html  |  20 +-
 .../docs/spark/2.1.3/spark-sql-cloudant/index.html | 338 +++++++++++----------
 .../2.1.3/spark-sql-streaming-akka/index.html      |  57 ++--
 .../2.1.3/spark-sql-streaming-mqtt/index.html      |  75 +++--
 .../spark/2.1.3/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.1.3/spark-streaming-mqtt/index.html    |  84 ++---
 .../spark/2.1.3/spark-streaming-pubsub/index.html  |  71 +++--
 .../spark/2.1.3/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.1.3/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.2.0/documentation/index.html  |  20 +-
 .../docs/spark/2.2.0/spark-sql-cloudant/index.html | 302 +++++++++---------
 .../2.2.0/spark-sql-streaming-akka/index.html      |  57 ++--
 .../2.2.0/spark-sql-streaming-mqtt/index.html      |  75 +++--
 .../spark/2.2.0/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.2.0/spark-streaming-mqtt/index.html    |  74 +++--
 .../spark/2.2.0/spark-streaming-pubsub/index.html  |  49 +--
 .../spark/2.2.0/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.2.0/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.2.1/documentation/index.html  |  20 +-
 .../docs/spark/2.2.1/spark-sql-cloudant/index.html | 338 +++++++++++----------
 .../2.2.1/spark-sql-streaming-akka/index.html      |  57 ++--
 .../2.2.1/spark-sql-streaming-mqtt/index.html      |  75 +++--
 .../spark/2.2.1/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.2.1/spark-streaming-mqtt/index.html    |  84 ++---
 .../spark/2.2.1/spark-streaming-pubsub/index.html  |  71 +++--
 .../spark/2.2.1/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.2.1/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.2.2/documentation/index.html  |  20 +-
 .../docs/spark/2.2.2/spark-sql-cloudant/index.html | 338 +++++++++++----------
 .../2.2.2/spark-sql-streaming-akka/index.html      |  57 ++--
 .../2.2.2/spark-sql-streaming-mqtt/index.html      |  75 +++--
 .../spark/2.2.2/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.2.2/spark-streaming-mqtt/index.html    |  84 ++---
 .../spark/2.2.2/spark-streaming-pubsub/index.html  |  71 +++--
 .../spark/2.2.2/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.2.2/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.3.0/documentation/index.html  |  20 +-
 .../docs/spark/2.3.0/spark-sql-cloudant/index.html | 338 +++++++++++----------
 .../2.3.0/spark-sql-streaming-akka/index.html      |  57 ++--
 .../2.3.0/spark-sql-streaming-mqtt/index.html      |  97 +++---
 .../spark/2.3.0/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.3.0/spark-streaming-mqtt/index.html    |  84 ++---
 .../spark/2.3.0/spark-streaming-pubnub/index.html  |  51 ++--
 .../spark/2.3.0/spark-streaming-pubsub/index.html  |  71 +++--
 .../spark/2.3.0/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.3.0/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.3.1/documentation/index.html  |  20 +-
 .../docs/spark/2.3.1/spark-sql-cloudant/index.html | 338 +++++++++++----------
 .../2.3.1/spark-sql-streaming-akka/index.html      |  57 ++--
 .../2.3.1/spark-sql-streaming-mqtt/index.html      |  97 +++---
 .../spark/2.3.1/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.3.1/spark-streaming-mqtt/index.html    |  84 ++---
 .../spark/2.3.1/spark-streaming-pubnub/index.html  |  51 ++--
 .../spark/2.3.1/spark-streaming-pubsub/index.html  |  71 +++--
 .../spark/2.3.1/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.3.1/spark-streaming-zeromq/index.html  |  41 ++-
 content/docs/spark/2.3.2/documentation/index.html  |  20 +-
 .../docs/spark/2.3.2/spark-sql-cloudant/index.html | 338 +++++++++++----------
 .../2.3.2/spark-sql-streaming-akka/index.html      |  57 ++--
 .../2.3.2/spark-sql-streaming-mqtt/index.html      |  97 +++---
 .../spark/2.3.2/spark-streaming-akka/index.html    |  47 +--
 .../spark/2.3.2/spark-streaming-mqtt/index.html    |  84 ++---
 .../spark/2.3.2/spark-streaming-pubnub/index.html  |  51 ++--
 .../spark/2.3.2/spark-streaming-pubsub/index.html  |  71 +++--
 .../spark/2.3.2/spark-streaming-twitter/index.html |  45 ++-
 .../spark/2.3.2/spark-streaming-zeromq/index.html  |  41 ++-
 .../docs/spark/current/documentation/index.html    |  20 +-
 .../spark/current/spark-sql-cloudant/index.html    | 338 +++++++++++----------
 .../current/spark-sql-streaming-akka/index.html    |  57 ++--
 .../current/spark-sql-streaming-mqtt/index.html    |  97 +++---
 .../spark/current/spark-streaming-akka/index.html  |  47 +--
 .../spark/current/spark-streaming-mqtt/index.html  |  84 ++---
 .../current/spark-streaming-pubnub/index.html      |  51 ++--
 .../current/spark-streaming-pubsub/index.html      |  71 +++--
 .../current/spark-streaming-twitter/index.html     |  45 ++-
 .../current/spark-streaming-zeromq/index.html      |  41 ++-
 content/docs/spark/overview/index.html             |  20 +-
 content/downloads/flink/index.html                 |  32 +-
 content/downloads/spark/index.html                 |  42 ++-
 content/feed.xml                                   |  74 +----
 content/history/index.html                         |  24 +-
 content/index.html                                 |  27 +-
 content/news/2015/09/25/release-0.1.0/index.html   |   7 -
 content/news/2015/11/09/new-committers/index.html  |  11 -
 content/news/2015/11/10/release-0.2.0/index.html   |  10 -
 content/privacy-policy/index.html                  |  23 +-
 .../releases/spark/2.0.0/release-notes/index.html  |  35 ++-
 .../releases/spark/2.0.1/release-notes/index.html  |  29 +-
 .../releases/spark/2.0.2/release-notes/index.html  |  29 +-
 .../releases/spark/2.1.0/release-notes/index.html  |  23 +-
 .../releases/spark/2.3.3/release-notes/index.html  |  23 +-
 .../{2.3.3 => 2.3.4}/release-notes/index.html      |  26 +-
 153 files changed, 6103 insertions(+), 4122 deletions(-)

diff --git a/content/community-members/index.html b/content/community-members/index.html
index c28821e..58657c5 100644
--- a/content/community-members/index.html
+++ b/content/community-members/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -554,7 +572,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/community/index.html b/content/community/index.html
index c708a3a..2042a15 100644
--- a/content/community/index.html
+++ b/content/community/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -257,7 +275,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/contributing-extensions/index.html b/content/contributing-extensions/index.html
index 7875ea3..77eab9f 100644
--- a/content/contributing-extensions/index.html
+++ b/content/contributing-extensions/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -248,7 +266,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/contributing/index.html b/content/contributing/index.html
index 961a4a4..24e0df2 100644
--- a/content/contributing/index.html
+++ b/content/contributing/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -279,29 +297,26 @@
 
 <p>Make sure you do not have any uncommitted changes and rebase master with latest changes from upstream:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>git fetch upstream
+<pre><code>git fetch upstream
 git checkout master
 git rebase upstream/master
 </code></pre>
-</div>
 
 <p>Now you should rebase your branch with master, to receive the upstream changes</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>git checkout branch
+<pre><code>git checkout branch
 git rebase master
 </code></pre>
-</div>
 
 <p>In both cases, you can have conflicts:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>error: could not apply fa39187... something to add to patch A
+<pre><code>error: could not apply fa39187... something to add to patch A
 
 When you have resolved this problem, run "git rebase --continue".
 If you prefer to skip this patch, run "git rebase --skip" instead.
 To check out the original branch and stop rebasing, run "git rebase --abort".
 Could not apply fa39187f3c3dfd2ab5faa38ac01cf3de7ce2e841... Change fake file
 </code></pre>
-</div>
 
 <p>Here, Git is telling you which commit is causing the conflict (fa39187). You’re given three choices:</p>
 
@@ -339,12 +354,11 @@ Could not apply fa39187f3c3dfd2ab5faa38ac01cf3de7ce2e841... Change fake file
 
 <p>Below is an example of a good commit message</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>[BAHIR-130] Performance enhancements for decision tree
+<pre><code>[BAHIR-130] Performance enhancements for decision tree
 
 Generate Matrix with random values through local memory
 if there is sufficient memory.
 </code></pre>
-</div>
 
 <h3 id="code-review-criteria">Code Review Criteria</h3>
 <p>Before considering how to contribute code, it’s useful to understand how code is reviewed, and why changes may be rejected. Simply put, changes that have many or large positives, and few negative effects or risks, are much more likely to be merged, and merged quickly. Risky and less valuable changes are very unlikely to be merged, and may be rejected outright rather than receive iterations of review.</p>
@@ -391,7 +405,7 @@ if there is sufficient memory.
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/1.0/documentation/index.html b/content/docs/flink/1.0/documentation/index.html
index a1c3163..3657911 100644
--- a/content/docs/flink/1.0/documentation/index.html
+++ b/content/docs/flink/1.0/documentation/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -218,7 +236,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/1.0/flink-streaming-activemq/index.html b/content/docs/flink/1.0/flink-streaming-activemq/index.html
index af12cb8..64decc1 100644
--- a/content/docs/flink/1.0/flink-streaming-activemq/index.html
+++ b/content/docs/flink/1.0/flink-streaming-activemq/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -200,20 +218,19 @@
 <p>This connector provides a source and sink to <a href="http://activemq.apache.org/">Apache ActiveMQ</a>™
 To use this connector, add the following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-activemq_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with ActiveMQ 5.14.0.</p>
 
 <p>Note that the streaming connectors are not part of the binary distribution of Flink. You need to link them into your job jar for cluster execution.
 See how to link with them for cluster execution <a href="https://ci.apache.org/projects/flink/flink-docs-release-1.2/dev/linking.html">here</a>.</p>
 
-<p>The source class is called <code class="highlighter-rouge">AMQSource</code>, and the sink is <code class="highlighter-rouge">AMQSink</code>.</p>
+<p>The source class is called <code>AMQSource</code>, and the sink is <code>AMQSink</code>.</p>
 
   </div>
 </div>
@@ -222,7 +239,7 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/1.0/flink-streaming-akka/index.html b/content/docs/flink/1.0/flink-streaming-akka/index.html
index 582ad6b..0ae47a7 100644
--- a/content/docs/flink/1.0/flink-streaming-akka/index.html
+++ b/content/docs/flink/1.0/flink-streaming-akka/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -200,13 +218,12 @@
 <p>This connector provides a sink to <a href="http://akka.io/">Akka</a> source actors in an ActorSystem.
 To use this connector, add the following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-akka_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with Akka 2.0+.</p>
 
@@ -215,19 +232,18 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 
 <h2 id="configuration">Configuration</h2>
 
-<p>The configurations for the Receiver Actor System in Flink Akka connector can be created using the standard typesafe <code class="highlighter-rouge">Config (com.typesafe.config.Config)</code> object.</p>
+<p>The configurations for the Receiver Actor System in Flink Akka connector can be created using the standard typesafe <code>Config (com.typesafe.config.Config)</code> object.</p>
 
-<p>To enable acknowledgements, the custom configuration <code class="highlighter-rouge">akka.remote.auto-ack</code> can be used.</p>
+<p>To enable acknowledgements, the custom configuration <code>akka.remote.auto-ack</code> can be used.</p>
 
 <p>The user can set any of the default configurations allowed by Akka as well as custom configurations allowed by the connector.</p>
 
 <p>A sample configuration can be defined as follows:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>String configFile = getClass().getClassLoader()
+<pre><code>String configFile = getClass().getClassLoader()
       .getResource("feeder_actor.conf").getFile();
 Config config = ConfigFactory.parseFile(new File(configFile));    
 </code></pre>
-</div>
 
 <h2 id="message-types">Message Types</h2>
 
@@ -235,13 +251,13 @@ Config config = ConfigFactory.parseFile(new File(configFile));
 
 <ul>
   <li>
-    <p>message containing <code class="highlighter-rouge">Iterable&lt;Object&gt;</code> data</p>
+    <p>message containing <code>Iterable&lt;Object&gt;</code> data</p>
   </li>
   <li>
-    <p>message containing generic <code class="highlighter-rouge">Object</code> data</p>
+    <p>message containing generic <code>Object</code> data</p>
   </li>
   <li>
-    <p>message containing generic <code class="highlighter-rouge">Object</code> data and a <code class="highlighter-rouge">Timestamp</code> value passed as <code class="highlighter-rouge">Tuple2&lt;Object, Long&gt;</code>.</p>
+    <p>message containing generic <code>Object</code> data and a <code>Timestamp</code> value passed as <code>Tuple2&lt;Object, Long&gt;</code>.</p>
   </li>
 </ul>
 
@@ -252,7 +268,7 @@ Config config = ConfigFactory.parseFile(new File(configFile));
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/1.0/flink-streaming-flume/index.html b/content/docs/flink/1.0/flink-streaming-flume/index.html
index 13a3f01..1404374 100644
--- a/content/docs/flink/1.0/flink-streaming-flume/index.html
+++ b/content/docs/flink/1.0/flink-streaming-flume/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -200,24 +218,22 @@
 <p>This connector provides a sink that can send data to <a href="https://flume.apache.org/">Apache Flume</a>™. To use this connector, add the
 following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-flume_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with Flume 1.5.0.</p>
 
 <p>Note that the streaming connectors are not part of the binary distribution of Flink. You need to link them into your job jar for cluster execution.
 See how to link with them for cluster execution <a href="https://ci.apache.org/projects/flink/flink-docs-release-1.2/dev/linking.html">here</a>.</p>
 
-<p>To create a <code class="highlighter-rouge">FlumeSink</code> instantiate the following constructor:</p>
+<p>To create a <code>FlumeSink</code> instantiate the following constructor:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>FlumeSink(String host, int port, SerializationSchema&lt;IN&gt; schema)
+<pre><code>FlumeSink(String host, int port, SerializationSchema&lt;IN&gt; schema)
 </code></pre>
-</div>
 
   </div>
 </div>
@@ -226,7 +242,7 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/1.0/flink-streaming-netty/index.html b/content/docs/flink/1.0/flink-streaming-netty/index.html
index 3d849eb..dd1bad6 100644
--- a/content/docs/flink/1.0/flink-streaming-netty/index.html
+++ b/content/docs/flink/1.0/flink-streaming-netty/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -204,7 +222,8 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 
 <h2 id="data-flow">Data Flow</h2>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>+-------------+      (2)    +------------------------+
+<p><code>
++-------------+      (2)    +------------------------+
 | user system |    &lt;-----   | Third Register Service |           
 +-------------+             +------------------------+
        |                                ^
@@ -214,52 +233,47 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 +--------------------+                  |
 | Flink Netty Source |  ----------------+
 +--------------------+         (1)
-</code></pre>
-</div>
+</code></p>
 
 <p>There are three components:</p>
 
 <ul>
   <li>User System - where the data stream is coming from</li>
-  <li>Third Register Service - receive <code class="highlighter-rouge">Flink Netty Source</code>’s register request (ip and port)</li>
-  <li>Flink Netty Source - Netty Server for receiving pushed streaming data from <code class="highlighter-rouge">User System</code></li>
+  <li>Third Register Service - receive <code>Flink Netty Source</code>’s register request (ip and port)</li>
+  <li>Flink Netty Source - Netty Server for receiving pushed streaming data from <code>User System</code></li>
 </ul>
 
 <h2 id="maven-dependency">Maven Dependency</h2>
 <p>To use this connector, add the following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
-  &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
-  &lt;artifactId&gt;flink-connector-netty_2.11&lt;/artifactId&gt;
-  &lt;version&gt;1.0&lt;/version&gt;
-&lt;/dependency&gt;
-</code></pre>
-</div>
+<p>```</p>
+<dependency>
+  <groupid>org.apache.bahir</groupid>
+  <artifactid>flink-connector-netty_2.11</artifactid>
+  <version>1.0</version>
+</dependency>
+<p>```</p>
 
 <h2 id="usage">Usage</h2>
 
 <p><em>Tcp Source:</em></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val env = StreamExecutionEnvironment.getExecutionEnvironment
+<p><code>
+val env = StreamExecutionEnvironment.getExecutionEnvironment
 env.addSource(new TcpReceiverSource("msg", 7070, Some("http://localhost:9090/cb")))
-</code></pre>
-</div>
-<blockquote>
-  <p>paramKey:  the http query param key
-tryPort:   try to use this point, if this point is used then try a new port
-callbackUrl:   register connector’s ip and port to a <code class="highlighter-rouge">Third Register Service</code></p>
-</blockquote>
+</code>
+&gt;paramKey:  the http query param key
+&gt;tryPort:   try to use this point, if this point is used then try a new port
+&gt;callbackUrl:   register connector’s ip and port to a <code>Third Register Service</code></p>
 
 <p><em>Http Source:</em></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val env = StreamExecutionEnvironment.getExecutionEnvironment
+<p><code>
+val env = StreamExecutionEnvironment.getExecutionEnvironment
 env.addSource(new TcpReceiverSource(7070, Some("http://localhost:9090/cb")))
-</code></pre>
-</div>
-<blockquote>
-  <p>tryPort:   try to use this port, if this point is used then try a new port
-callbackUrl:   register connector’s ip and port to a <code class="highlighter-rouge">Third Register Service</code></p>
-</blockquote>
+</code>
+&gt;tryPort:   try to use this port, if this point is used then try a new port
+&gt;callbackUrl:   register connector’s ip and port to a <code>Third Register Service</code></p>
 
 <h2 id="full-example">Full Example</h2>
 
@@ -277,7 +291,7 @@ callbackUrl:   register connector’s ip and port to a <code class="highlighter-
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/1.0/flink-streaming-redis/index.html b/content/docs/flink/1.0/flink-streaming-redis/index.html
index 7acd66f..6e3d121 100644
--- a/content/docs/flink/1.0/flink-streaming-redis/index.html
+++ b/content/docs/flink/1.0/flink-streaming-redis/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,13 +219,12 @@
 to <a href="http://redis.io/topics/pubsub">Redis PubSub</a>. To use this connector, add the
 following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-redis_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with Redis 2.8.5.</p>
 
@@ -233,7 +250,7 @@ The sink can use three different methods for communicating with different type o
 
 <p><strong>Java:</strong></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>public static class RedisExampleMapper implements RedisMapper&lt;Tuple2&lt;String, String&gt;&gt;{
+<pre><code>public static class RedisExampleMapper implements RedisMapper&lt;Tuple2&lt;String, String&gt;&gt;{
 
     @Override
     public RedisCommandDescription getCommandDescription() {
@@ -255,11 +272,10 @@ FlinkJedisPoolConfig conf = new FlinkJedisPoolConfig.Builder().setHost("127.0.0.
 DataStream&lt;String&gt; stream = ...;
 stream.addSink(new RedisSink&lt;Tuple2&lt;String, String&gt;&gt;(conf, new RedisExampleMapper());
 </code></pre>
-</div>
 
 <p><strong>Scala:</strong></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class RedisExampleMapper extends RedisMapper[(String, String)]{
+<pre><code>class RedisExampleMapper extends RedisMapper[(String, String)]{
   override def getCommandDescription: RedisCommandDescription = {
     new RedisCommandDescription(RedisCommand.HSET, "HASH_NAME")
   }
@@ -271,45 +287,40 @@ stream.addSink(new RedisSink&lt;Tuple2&lt;String, String&gt;&gt;(conf, new Redis
 val conf = new FlinkJedisPoolConfig.Builder().setHost("127.0.0.1").build()
 stream.addSink(new RedisSink[(String, String)](conf, new RedisExampleMapper))
 </code></pre>
-</div>
 
 <p>This example code does the same, but for Redis Cluster:</p>
 
 <p><strong>Java:</strong></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>FlinkJedisPoolConfig conf = new FlinkJedisPoolConfig.Builder()
+<pre><code>FlinkJedisPoolConfig conf = new FlinkJedisPoolConfig.Builder()
     .setNodes(new HashSet&lt;InetSocketAddress&gt;(Arrays.asList(new InetSocketAddress(5601)))).build();
 
 DataStream&lt;String&gt; stream = ...;
 stream.addSink(new RedisSink&lt;Tuple2&lt;String, String&gt;&gt;(conf, new RedisExampleMapper());
 </code></pre>
-</div>
 
 <p><strong>Scala:</strong></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val conf = new FlinkJedisPoolConfig.Builder().setNodes(...).build()
+<pre><code>val conf = new FlinkJedisPoolConfig.Builder().setNodes(...).build()
 stream.addSink(new RedisSink[(String, String)](conf, new RedisExampleMapper))
 </code></pre>
-</div>
 
 <p>This example shows when the Redis environment is with Sentinels:</p>
 
 <p>Java:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>FlinkJedisSentinelConfig conf = new FlinkJedisSentinelConfig.Builder()
+<pre><code>FlinkJedisSentinelConfig conf = new FlinkJedisSentinelConfig.Builder()
     .setMasterName("master").setSentinels(...).build();
 
 DataStream&lt;String&gt; stream = ...;
 stream.addSink(new RedisSink&lt;Tuple2&lt;String, String&gt;&gt;(conf, new RedisExampleMapper());
 </code></pre>
-</div>
 
 <p>Scala:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val conf = new FlinkJedisSentinelConfig.Builder().setMasterName("master").setSentinels(...).build()
+<pre><code>val conf = new FlinkJedisSentinelConfig.Builder().setMasterName("master").setSentinels(...).build()
 stream.addSink(new RedisSink[(String, String)](conf, new RedisExampleMapper))
 </code></pre>
-</div>
 
 <p>This section gives a description of all the available data types and what Redis command used for that.</p>
 
@@ -358,7 +369,7 @@ stream.addSink(new RedisSink[(String, String)](conf, new RedisExampleMapper))
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/current/documentation/index.html b/content/docs/flink/current/documentation/index.html
index 3dcbbf1..347525f 100644
--- a/content/docs/flink/current/documentation/index.html
+++ b/content/docs/flink/current/documentation/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -222,7 +240,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/current/flink-streaming-activemq/index.html b/content/docs/flink/current/flink-streaming-activemq/index.html
index debbefe..5442d1b 100644
--- a/content/docs/flink/current/flink-streaming-activemq/index.html
+++ b/content/docs/flink/current/flink-streaming-activemq/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -200,20 +218,19 @@
 <p>This connector provides a source and sink to <a href="http://activemq.apache.org/">Apache ActiveMQ</a>™
 To use this connector, add the following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-activemq_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.1-SNAPSHOT&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with ActiveMQ 5.14.0.</p>
 
 <p>Note that the streaming connectors are not part of the binary distribution of Flink. You need to link them into your job jar for cluster execution.
 See how to link with them for cluster execution <a href="https://ci.apache.org/projects/flink/flink-docs-release-1.2/dev/linking.html">here</a>.</p>
 
-<p>The source class is called <code class="highlighter-rouge">AMQSource</code>, and the sink is <code class="highlighter-rouge">AMQSink</code>.</p>
+<p>The source class is called <code>AMQSource</code>, and the sink is <code>AMQSink</code>.</p>
 
   </div>
 </div>
@@ -222,7 +239,7 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/current/flink-streaming-akka/index.html b/content/docs/flink/current/flink-streaming-akka/index.html
index adae8e4..42f4c7b 100644
--- a/content/docs/flink/current/flink-streaming-akka/index.html
+++ b/content/docs/flink/current/flink-streaming-akka/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -200,13 +218,12 @@
 <p>This connector provides a sink to <a href="http://akka.io/">Akka</a> source actors in an ActorSystem.
 To use this connector, add the following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-akka_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.1-SNAPSHOT&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with Akka 2.0+.</p>
 
@@ -215,19 +232,18 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 
 <h2 id="configuration">Configuration</h2>
 
-<p>The configurations for the Receiver Actor System in Flink Akka connector can be created using the standard typesafe <code class="highlighter-rouge">Config (com.typesafe.config.Config)</code> object.</p>
+<p>The configurations for the Receiver Actor System in Flink Akka connector can be created using the standard typesafe <code>Config (com.typesafe.config.Config)</code> object.</p>
 
-<p>To enable acknowledgements, the custom configuration <code class="highlighter-rouge">akka.remote.auto-ack</code> can be used.</p>
+<p>To enable acknowledgements, the custom configuration <code>akka.remote.auto-ack</code> can be used.</p>
 
 <p>The user can set any of the default configurations allowed by Akka as well as custom configurations allowed by the connector.</p>
 
 <p>A sample configuration can be defined as follows:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>String configFile = getClass().getClassLoader()
+<pre><code>String configFile = getClass().getClassLoader()
       .getResource("feeder_actor.conf").getFile();
 Config config = ConfigFactory.parseFile(new File(configFile));    
 </code></pre>
-</div>
 
 <h2 id="message-types">Message Types</h2>
 
@@ -235,13 +251,13 @@ Config config = ConfigFactory.parseFile(new File(configFile));
 
 <ul>
   <li>
-    <p>message containing <code class="highlighter-rouge">Iterable&lt;Object&gt;</code> data</p>
+    <p>message containing <code>Iterable&lt;Object&gt;</code> data</p>
   </li>
   <li>
-    <p>message containing generic <code class="highlighter-rouge">Object</code> data</p>
+    <p>message containing generic <code>Object</code> data</p>
   </li>
   <li>
-    <p>message containing generic <code class="highlighter-rouge">Object</code> data and a <code class="highlighter-rouge">Timestamp</code> value passed as <code class="highlighter-rouge">Tuple2&lt;Object, Long&gt;</code>.</p>
+    <p>message containing generic <code>Object</code> data and a <code>Timestamp</code> value passed as <code>Tuple2&lt;Object, Long&gt;</code>.</p>
   </li>
 </ul>
 
@@ -252,7 +268,7 @@ Config config = ConfigFactory.parseFile(new File(configFile));
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/current/flink-streaming-flume/index.html b/content/docs/flink/current/flink-streaming-flume/index.html
index 01ec5f3..3a1aac4 100644
--- a/content/docs/flink/current/flink-streaming-flume/index.html
+++ b/content/docs/flink/current/flink-streaming-flume/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -200,24 +218,22 @@
 <p>This connector provides a sink that can send data to <a href="https://flume.apache.org/">Apache Flume</a>™. To use this connector, add the
 following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-flume_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.1-SNAPSHOT&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with Flume 1.8.0.</p>
 
 <p>Note that the streaming connectors are not part of the binary distribution of Flink. You need to link them into your job jar for cluster execution.
 See how to link with them for cluster execution <a href="https://ci.apache.org/projects/flink/flink-docs-release-1.2/dev/linking.html">here</a>.</p>
 
-<p>To create a <code class="highlighter-rouge">FlumeSink</code> instantiate the following constructor:</p>
+<p>To create a <code>FlumeSink</code> instantiate the following constructor:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>FlumeSink(String host, int port, SerializationSchema&lt;IN&gt; schema)
+<pre><code>FlumeSink(String host, int port, SerializationSchema&lt;IN&gt; schema)
 </code></pre>
-</div>
 
 
   </div>
@@ -227,7 +243,7 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/current/flink-streaming-influxdb/index.html b/content/docs/flink/current/flink-streaming-influxdb/index.html
index 05fe052..1f5d9ca 100644
--- a/content/docs/flink/current/flink-streaming-influxdb/index.html
+++ b/content/docs/flink/current/flink-streaming-influxdb/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -200,13 +218,12 @@
 <p>This connector provides a sink that can send data to <a href="https://www.influxdata.com/">InfluxDB</a>. To use this connector, add the
 following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-influxdb_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.1-SNAPSHOT&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with InfluxDB 1.3.x <br />
 <em>Requirements</em>: Java 1.8+</p>
@@ -221,11 +238,10 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 
 <h3 id="java-api">JAVA API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>DataStream&lt;InfluxDBPoint&gt; dataStream = ...
+<pre><code>DataStream&lt;InfluxDBPoint&gt; dataStream = ...
 InfluxDBConfig influxDBConfig = InfluxDBConfig.builder(String host, String username, String password, String dbName)
 dataStream.addSink(new InfluxDBSink(influxDBConfig));
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir-flink/tree/master/flink-connector-influxdb/examples">InfluxDB Examples</a></p>
 
@@ -237,7 +253,7 @@ dataStream.addSink(new InfluxDBSink(influxDBConfig));
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/current/flink-streaming-kudu/index.html b/content/docs/flink/current/flink-streaming-kudu/index.html
index 921dea3..b951ac2 100644
--- a/content/docs/flink/current/flink-streaming-kudu/index.html
+++ b/content/docs/flink/current/flink-streaming-kudu/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -197,16 +215,15 @@
 
 <h1 id="flink-kudu-connector">Flink Kudu Connector</h1>
 
-<p>This connector provides a source (<code class="highlighter-rouge">KuduInputFormat</code>) and a sink/output (<code class="highlighter-rouge">KuduSink</code> and <code class="highlighter-rouge">KuduOutputFormat</code>, respectively) that can read and write to <a href="https://kudu.apache.org/">Kudu</a>. To use this connector, add the
+<p>This connector provides a source (<code>KuduInputFormat</code>) and a sink/output (<code>KuduSink</code> and <code>KuduOutputFormat</code>, respectively) that can read and write to <a href="https://kudu.apache.org/">Kudu</a>. To use this connector, add the
 following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-kudu_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.1-SNAPSHOT&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with Apache Kudu <em>1.7.1</em> (last stable version).</p>
 
@@ -220,81 +237,81 @@ Optionally, you can use the docker images provided in dockers folder.</p>
 
 <h2 id="kuduinputformat">KuduInputFormat</h2>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
+<p>```
+ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();</p>
 
-env.setParallelism(PARALLELISM);
+<p>env.setParallelism(PARALLELISM);</p>
 
-// create a table info object
+<p>// create a table info object
 KuduTableInfo tableInfo = KuduTableInfo.Builder
-        .create("books")
-        .addColumn(KuduColumnInfo.Builder.create("id", Type.INT32).key(true).hashKey(true).build())
-        .addColumn(KuduColumnInfo.Builder.create("title", Type.STRING).build())
-        .addColumn(KuduColumnInfo.Builder.create("author", Type.STRING).build())
-        .addColumn(KuduColumnInfo.Builder.create("price", Type.DOUBLE).build())
-        .addColumn(KuduColumnInfo.Builder.create("quantity", Type.INT32).build())
-        .build();
-    
-// Pass the tableInfo to the KuduInputFormat and provide kuduMaster ips
-env.createInput(new KuduInputFormat&lt;&gt;("172.25.0.6", tableInfo))
-        .count();
-        
-env.execute();
-</code></pre>
-</div>
+        .create(“books”)
+        .addColumn(KuduColumnInfo.Builder.create(“id”, Type.INT32).key(true).hashKey(true).build())
+        .addColumn(KuduColumnInfo.Builder.create(“title”, Type.STRING).build())
+        .addColumn(KuduColumnInfo.Builder.create(“author”, Type.STRING).build())
+        .addColumn(KuduColumnInfo.Builder.create(“price”, Type.DOUBLE).build())
+        .addColumn(KuduColumnInfo.Builder.create(“quantity”, Type.INT32).build())
+        .build();</p>
+
+<p>// Pass the tableInfo to the KuduInputFormat and provide kuduMaster ips
+env.createInput(new KuduInputFormat&lt;&gt;(“172.25.0.6”, tableInfo))
+        .count();</p>
+
+<p>env.execute();
+```</p>
 
 <h2 id="kuduoutputformat">KuduOutputFormat</h2>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
+<p>```
+ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();</p>
 
-env.setParallelism(PARALLELISM);
+<p>env.setParallelism(PARALLELISM);</p>
 
-// create a table info object
+<p>// create a table info object
 KuduTableInfo tableInfo = KuduTableInfo.Builder
-        .create("books")
+        .create(“books”)
         .createIfNotExist(true)
         .replicas(1)
-        .addColumn(KuduColumnInfo.Builder.create("id", Type.INT32).key(true).hashKey(true).build())
-        .addColumn(KuduColumnInfo.Builder.create("title", Type.STRING).build())
-        .addColumn(KuduColumnInfo.Builder.create("author", Type.STRING).build())
-        .addColumn(KuduColumnInfo.Builder.create("price", Type.DOUBLE).build())
-        .addColumn(KuduColumnInfo.Builder.create("quantity", Type.INT32).build())
-        .build();
+        .addColumn(KuduColumnInfo.Builder.create(“id”, Type.INT32).key(true).hashKey(true).build())
+        .addColumn(KuduColumnInfo.Builder.create(“title”, Type.STRING).build())
+        .addColumn(KuduColumnInfo.Builder.create(“author”, Type.STRING).build())
+        .addColumn(KuduColumnInfo.Builder.create(“price”, Type.DOUBLE).build())
+        .addColumn(KuduColumnInfo.Builder.create(“quantity”, Type.INT32).build())
+        .build();</p>
 
-...
+<p>…</p>
 
-env.fromCollection(books)
-        .output(new KuduOutputFormat&lt;&gt;("172.25.0.6", tableInfo));
+<p>env.fromCollection(books)
+        .output(new KuduOutputFormat&lt;&gt;(“172.25.0.6”, tableInfo));</p>
 
-env.execute();
-</code></pre>
-</div>
+<p>env.execute();
+```</p>
 
 <h2 id="kudusink">KuduSink</h2>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
+<p>```
+StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();</p>
 
-env.setParallelism(PARALLELISM);
+<p>env.setParallelism(PARALLELISM);</p>
 
-// create a table info object
+<p>// create a table info object
 KuduTableInfo tableInfo = KuduTableInfo.Builder
-        .create("books")
+        .create(“books”)
         .createIfNotExist(true)
         .replicas(1)
-        .addColumn(KuduColumnInfo.Builder.create("id", Type.INT32).key(true).hashKey(true).build())
-        .addColumn(KuduColumnInfo.Builder.create("title", Type.STRING).build())
-        .addColumn(KuduColumnInfo.Builder.create("author", Type.STRING).build())
-        .addColumn(KuduColumnInfo.Builder.create("price", Type.DOUBLE).build())
-        .addColumn(KuduColumnInfo.Builder.create("quantity", Type.INT32).build())
-        .build();
+        .addColumn(KuduColumnInfo.Builder.create(“id”, Type.INT32).key(true).hashKey(true).build())
+        .addColumn(KuduColumnInfo.Builder.create(“title”, Type.STRING).build())
+        .addColumn(KuduColumnInfo.Builder.create(“author”, Type.STRING).build())
+        .addColumn(KuduColumnInfo.Builder.create(“price”, Type.DOUBLE).build())
+        .addColumn(KuduColumnInfo.Builder.create(“quantity”, Type.INT32).build())
+        .build();</p>
 
-...
+<p>…</p>
 
-env.fromCollection(books)
-    .addSink(new KuduSink&lt;&gt;("172.25.0.6", tableInfo));
+<p>env.fromCollection(books)
+    .addSink(new KuduSink&lt;&gt;(“172.25.0.6”, tableInfo));</p>
 
-env.execute();
-</code></pre>
-</div>
+<p>env.execute();
+```</p>
 
   </div>
 </div>
@@ -303,7 +320,7 @@ env.execute();
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/current/flink-streaming-netty/index.html b/content/docs/flink/current/flink-streaming-netty/index.html
index 845f8e8..11e280f 100644
--- a/content/docs/flink/current/flink-streaming-netty/index.html
+++ b/content/docs/flink/current/flink-streaming-netty/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -204,7 +222,8 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 
 <h2 id="data-flow">Data Flow</h2>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>+-------------+      (2)    +------------------------+
+<p><code>
++-------------+      (2)    +------------------------+
 | user system |    &lt;-----   | Third Register Service |           
 +-------------+             +------------------------+
        |                                ^
@@ -214,52 +233,47 @@ See how to link with them for cluster execution <a href="https://ci.apache.org/p
 +--------------------+                  |
 | Flink Netty Source |  ----------------+
 +--------------------+         (1)
-</code></pre>
-</div>
+</code></p>
 
 <p>There are three components:</p>
 
 <ul>
   <li>User System - where the data stream is coming from</li>
-  <li>Third Register Service - receive <code class="highlighter-rouge">Flink Netty Source</code>’s register request (ip and port)</li>
-  <li>Flink Netty Source - Netty Server for receiving pushed streaming data from <code class="highlighter-rouge">User System</code></li>
+  <li>Third Register Service - receive <code>Flink Netty Source</code>’s register request (ip and port)</li>
+  <li>Flink Netty Source - Netty Server for receiving pushed streaming data from <code>User System</code></li>
 </ul>
 
 <h2 id="maven-dependency">Maven Dependency</h2>
 <p>To use this connector, add the following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
-  &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
-  &lt;artifactId&gt;flink-connector-netty_2.11&lt;/artifactId&gt;
-  &lt;version&gt;1.1-SNAPSHOT&lt;/version&gt;
-&lt;/dependency&gt;
-</code></pre>
-</div>
+<p>```</p>
+<dependency>
+  <groupid>org.apache.bahir</groupid>
+  <artifactid>flink-connector-netty_2.11</artifactid>
+  <version>1.1-SNAPSHOT</version>
+</dependency>
+<p>```</p>
 
 <h2 id="usage">Usage</h2>
 
 <p><em>Tcp Source:</em></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val env = StreamExecutionEnvironment.getExecutionEnvironment
+<p><code>
+val env = StreamExecutionEnvironment.getExecutionEnvironment
 env.addSource(new TcpReceiverSource("msg", 7070, Some("http://localhost:9090/cb")))
-</code></pre>
-</div>
-<blockquote>
-  <p>paramKey:  the http query param key
-tryPort:   try to use this point, if this point is used then try a new port
-callbackUrl:   register connector’s ip and port to a <code class="highlighter-rouge">Third Register Service</code></p>
-</blockquote>
+</code>
+&gt;paramKey:  the http query param key
+&gt;tryPort:   try to use this point, if this point is used then try a new port
+&gt;callbackUrl:   register connector’s ip and port to a <code>Third Register Service</code></p>
 
 <p><em>Http Source:</em></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val env = StreamExecutionEnvironment.getExecutionEnvironment
+<p><code>
+val env = StreamExecutionEnvironment.getExecutionEnvironment
 env.addSource(new TcpReceiverSource(7070, Some("http://localhost:9090/cb")))
-</code></pre>
-</div>
-<blockquote>
-  <p>tryPort:   try to use this port, if this point is used then try a new port
-callbackUrl:   register connector’s ip and port to a <code class="highlighter-rouge">Third Register Service</code></p>
-</blockquote>
+</code>
+&gt;tryPort:   try to use this port, if this point is used then try a new port
+&gt;callbackUrl:   register connector’s ip and port to a <code>Third Register Service</code></p>
 
 <h2 id="full-example">Full Example</h2>
 
@@ -277,7 +291,7 @@ callbackUrl:   register connector’s ip and port to a <code class="highlighter-
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/current/flink-streaming-redis/index.html b/content/docs/flink/current/flink-streaming-redis/index.html
index 28f2900..df76a63 100644
--- a/content/docs/flink/current/flink-streaming-redis/index.html
+++ b/content/docs/flink/current/flink-streaming-redis/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,13 +219,12 @@
 to <a href="http://redis.io/topics/pubsub">Redis PubSub</a>. To use this connector, add the
 following dependency to your project:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
   &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
   &lt;artifactId&gt;flink-connector-redis_2.11&lt;/artifactId&gt;
   &lt;version&gt;1.1-SNAPSHOT&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
 <p><em>Version Compatibility</em>: This module is compatible with Redis 2.8.5.</p>
 
@@ -233,7 +250,7 @@ The sink can use three different methods for communicating with different type o
 
 <p><strong>Java:</strong></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>public static class RedisExampleMapper implements RedisMapper&lt;Tuple2&lt;String, String&gt;&gt;{
+<pre><code>public static class RedisExampleMapper implements RedisMapper&lt;Tuple2&lt;String, String&gt;&gt;{
 
     @Override
     public RedisCommandDescription getCommandDescription() {
@@ -255,11 +272,10 @@ FlinkJedisPoolConfig conf = new FlinkJedisPoolConfig.Builder().setHost("127.0.0.
 DataStream&lt;String&gt; stream = ...;
 stream.addSink(new RedisSink&lt;Tuple2&lt;String, String&gt;&gt;(conf, new RedisExampleMapper());
 </code></pre>
-</div>
 
 <p><strong>Scala:</strong></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class RedisExampleMapper extends RedisMapper[(String, String)]{
+<pre><code>class RedisExampleMapper extends RedisMapper[(String, String)]{
   override def getCommandDescription: RedisCommandDescription = {
     new RedisCommandDescription(RedisCommand.HSET, "HASH_NAME")
   }
@@ -271,45 +287,40 @@ stream.addSink(new RedisSink&lt;Tuple2&lt;String, String&gt;&gt;(conf, new Redis
 val conf = new FlinkJedisPoolConfig.Builder().setHost("127.0.0.1").build()
 stream.addSink(new RedisSink[(String, String)](conf, new RedisExampleMapper))
 </code></pre>
-</div>
 
 <p>This example code does the same, but for Redis Cluster:</p>
 
 <p><strong>Java:</strong></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>FlinkJedisPoolConfig conf = new FlinkJedisPoolConfig.Builder()
+<pre><code>FlinkJedisPoolConfig conf = new FlinkJedisPoolConfig.Builder()
     .setNodes(new HashSet&lt;InetSocketAddress&gt;(Arrays.asList(new InetSocketAddress(5601)))).build();
 
 DataStream&lt;String&gt; stream = ...;
 stream.addSink(new RedisSink&lt;Tuple2&lt;String, String&gt;&gt;(conf, new RedisExampleMapper());
 </code></pre>
-</div>
 
 <p><strong>Scala:</strong></p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val conf = new FlinkJedisPoolConfig.Builder().setNodes(...).build()
+<pre><code>val conf = new FlinkJedisPoolConfig.Builder().setNodes(...).build()
 stream.addSink(new RedisSink[(String, String)](conf, new RedisExampleMapper))
 </code></pre>
-</div>
 
 <p>This example shows when the Redis environment is with Sentinels:</p>
 
 <p>Java:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>FlinkJedisSentinelConfig conf = new FlinkJedisSentinelConfig.Builder()
+<pre><code>FlinkJedisSentinelConfig conf = new FlinkJedisSentinelConfig.Builder()
     .setMasterName("master").setSentinels(...).build();
 
 DataStream&lt;String&gt; stream = ...;
 stream.addSink(new RedisSink&lt;Tuple2&lt;String, String&gt;&gt;(conf, new RedisExampleMapper());
 </code></pre>
-</div>
 
 <p>Scala:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val conf = new FlinkJedisSentinelConfig.Builder().setMasterName("master").setSentinels(...).build()
+<pre><code>val conf = new FlinkJedisSentinelConfig.Builder().setMasterName("master").setSentinels(...).build()
 stream.addSink(new RedisSink[(String, String)](conf, new RedisExampleMapper))
 </code></pre>
-</div>
 
 <p>This section gives a description of all the available data types and what Redis command used for that.</p>
 
@@ -358,7 +369,7 @@ stream.addSink(new RedisSink[(String, String)](conf, new RedisExampleMapper))
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/flink/overview/index.html b/content/docs/flink/overview/index.html
index b4a8ff3..ab83496 100644
--- a/content/docs/flink/overview/index.html
+++ b/content/docs/flink/overview/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -209,7 +227,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.0/documentation/index.html b/content/docs/spark/2.0.0/documentation/index.html
index e300330..bfd7399 100644
--- a/content/docs/spark/2.0.0/documentation/index.html
+++ b/content/docs/spark/2.0.0/documentation/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -222,7 +240,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.0/spark-sql-streaming-mqtt/index.html b/content/docs/spark/2.0.0/spark-sql-streaming-mqtt/index.html
index 48e41ca..81ed176 100644
--- a/content/docs/spark/2.0.0/spark-sql-streaming-mqtt/index.html
+++ b/content/docs/spark/2.0.0/spark-sql-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.0.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.0.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-sql-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.0.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.0.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is compiled for Scala 2.11 only, and intends to support Spark 2.0 onwards.</p>
 
@@ -231,31 +246,29 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <p>A SQL Stream can be created with data streams received through MQTT Server using,</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h2 id="enable-recovering-from-failures">Enable recovering from failures.</h2>
 
-<p>Setting values for option <code class="highlighter-rouge">localStorage</code> and <code class="highlighter-rouge">clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
+<p>Setting values for option <code>localStorage</code> and <code>clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .option("localStorage", "/path/to/localdir")
     .option("clientId", "some-client-id")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h3 id="scala-api">Scala API</h3>
 
 <p>An example, for scala API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
 val lines = spark.readStream
   .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
   .option("topic", topic)
@@ -275,15 +288,14 @@ val query = wordCounts.writeStream
 
 query.awaitTermination()
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">MQTTStreamWordCount.scala</code> for full example.</p>
+<p>Please see <code>MQTTStreamWordCount.scala</code> for full example.</p>
 
 <h3 id="java-api">Java API</h3>
 
 <p>An example, for Java API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
 Dataset&lt;String&gt; lines = spark
         .readStream()
         .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
@@ -309,9 +321,8 @@ StreamingQuery query = wordCounts.writeStream()
 
 query.awaitTermination();
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">JavaMQTTStreamWordCount.java</code> for full example.</p>
+<p>Please see <code>JavaMQTTStreamWordCount.java</code> for full example.</p>
 
 
   </div>
@@ -321,7 +332,7 @@ query.awaitTermination();
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.0/spark-streaming-akka/index.html b/content/docs/spark/2.0.0/spark-streaming-akka/index.html
index 0f4c28a..60419ce 100644
--- a/content/docs/spark/2.0.0/spark-streaming-akka/index.html
+++ b/content/docs/spark/2.0.0/spark-streaming-akka/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,42 +219,39 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.0.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.0.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-akka_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.0.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.0.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p>DStreams can be created with data streams received through Akka actors by using <code class="highlighter-rouge">AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
+<p>DStreams can be created with data streams received through Akka actors by using <code>AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends ActorReceiver {
+<pre><code>class CustomActor extends ActorReceiver {
   def receive = {
     case data: String =&gt; store(data)
   }
@@ -246,14 +261,13 @@ this actor can be configured to handle failures, etc.</p>
 val ssc: StreamingContext = ...
 val lines = AkkaUtils.createStream[String](ssc, Props[CustomActor](), "CustomReceiver")
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends JavaActorReceiver {
+<pre><code>class CustomActor extends JavaActorReceiver {
     @Override
     public void onReceive(Object msg) throws Exception {
         store((String) msg);
@@ -264,7 +278,6 @@ this actor can be configured to handle failures, etc.</p>
 JavaStreamingContext jssc = ...;
 JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Props.create(CustomActor.class), "CustomReceiver");
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-akka/examples">Akka Examples</a></p>
 
@@ -275,7 +288,7 @@ JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Pro
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.0/spark-streaming-mqtt/index.html b/content/docs/spark/2.0.0/spark-streaming-mqtt/index.html
index b895ee1..c90a519 100644
--- a/content/docs/spark/2.0.0/spark-streaming-mqtt/index.html
+++ b/content/docs/spark/2.0.0/spark-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.0.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.0.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.0.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.0.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -231,21 +246,19 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
+<pre><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
+<pre><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-mqtt/examples">MQTT Examples</a></p>
 
@@ -256,7 +269,7 @@ this actor can be configured to handle failures, etc.</p>
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.0/spark-streaming-twitter/index.html b/content/docs/spark/2.0.0/spark-streaming-twitter/index.html
index fcbf7e6..cd79e59 100644
--- a/content/docs/spark/2.0.0/spark-streaming-twitter/index.html
+++ b/content/docs/spark/2.0.0/spark-streaming-twitter/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,52 +219,47 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.0.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.0.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-twitter_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.0.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.0.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p><code class="highlighter-rouge">TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
-can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code class="highlighter-rouge">TwitterUtils</code> class and create a DStream with <code class="highlighter-rouge">TwitterUtils.createStream</code> as shown below.</p>
+<p><code>TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
+can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code>TwitterUtils</code> class and create a DStream with <code>TwitterUtils.createStream</code> as shown below.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter._
+<pre><code>import org.apache.spark.streaming.twitter._
 
 TwitterUtils.createStream(ssc, None)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter.*;
+<pre><code>import org.apache.spark.streaming.twitter.*;
 
 TwitterUtils.createStream(jssc);
 </code></pre>
-</div>
 
 <p>You can also either get the public stream, or get the filtered stream based on keywords. 
 See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-twitter/examples">Twitter Examples</a></p>
@@ -258,7 +271,7 @@ See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.0/spark-streaming-zeromq/index.html b/content/docs/spark/2.0.0/spark-streaming-zeromq/index.html
index d40de94..400934e 100644
--- a/content/docs/spark/2.0.0/spark-streaming-zeromq/index.html
+++ b/content/docs/spark/2.0.0/spark-streaming-zeromq/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.0.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.0.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-zeromq_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.0.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.0.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -231,15 +246,13 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = ZeroMQUtils.createStream(ssc, ...)
+<pre><code>val lines = ZeroMQUtils.createStream(ssc, ...)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
+<pre><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-zeromq/examples">ZeroMQ Examples</a></p>
 
@@ -250,7 +263,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.1/documentation/index.html b/content/docs/spark/2.0.1/documentation/index.html
index 779bf78..2a6d7de 100644
--- a/content/docs/spark/2.0.1/documentation/index.html
+++ b/content/docs/spark/2.0.1/documentation/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -222,7 +240,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.1/spark-sql-streaming-mqtt/index.html b/content/docs/spark/2.0.1/spark-sql-streaming-mqtt/index.html
index 2243e71..f67f5c5 100644
--- a/content/docs/spark/2.0.1/spark-sql-streaming-mqtt/index.html
+++ b/content/docs/spark/2.0.1/spark-sql-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.0.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.0.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-sql-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.0.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.0.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is compiled for Scala 2.11 only, and intends to support Spark 2.0 onwards.</p>
 
@@ -231,49 +246,47 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <p>A SQL Stream can be created with data streams received through MQTT Server using,</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h2 id="enable-recovering-from-failures">Enable recovering from failures.</h2>
 
-<p>Setting values for option <code class="highlighter-rouge">localStorage</code> and <code class="highlighter-rouge">clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
+<p>Setting values for option <code>localStorage</code> and <code>clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .option("localStorage", "/path/to/localdir")
     .option("clientId", "some-client-id")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h2 id="configuration-options">Configuration options.</h2>
 
 <p>This source uses <a href="https://eclipse.org/paho/clients/java/">Eclipse Paho Java Client</a>. Client API documentation is located <a href="http://www.eclipse.org/paho/files/javadoc/index.html">here</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">brokerUrl</code> A url MqttClient connects to. Set this or <code class="highlighter-rouge">path</code> as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
-  <li><code class="highlighter-rouge">persistence</code> By default it is used for storing incoming messages on disk. If <code class="highlighter-rouge">memory</code> is provided as value for this option, then recovery on restart is not supported.</li>
-  <li><code class="highlighter-rouge">topic</code> Topic MqttClient subscribes to.</li>
-  <li><code class="highlighter-rouge">clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
-  <li><code class="highlighter-rouge">QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
-  <li><code class="highlighter-rouge">username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
-  <li><code class="highlighter-rouge">password</code> Sets the password to use for the connection.</li>
-  <li><code class="highlighter-rouge">cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
-  <li><code class="highlighter-rouge">connectionTimeout</code> Sets the connection timeout, a value of 0 is interpretted as wait until client connects. See <code class="highlighter-rouge">MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
-  <li><code class="highlighter-rouge">keepAlive</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setKeepAliveInterval</code>.</li>
-  <li><code class="highlighter-rouge">mqttVersion</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setMqttVersion</code>.</li>
+  <li><code>brokerUrl</code> A url MqttClient connects to. Set this or <code>path</code> as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
+  <li><code>persistence</code> By default it is used for storing incoming messages on disk. If <code>memory</code> is provided as value for this option, then recovery on restart is not supported.</li>
+  <li><code>topic</code> Topic MqttClient subscribes to.</li>
+  <li><code>clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
+  <li><code>QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
+  <li><code>username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
+  <li><code>password</code> Sets the password to use for the connection.</li>
+  <li><code>cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
+  <li><code>connectionTimeout</code> Sets the connection timeout, a value of 0 is interpretted as wait until client connects. See <code>MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
+  <li><code>keepAlive</code> Same as <code>MqttConnectOptions.setKeepAliveInterval</code>.</li>
+  <li><code>mqttVersion</code> Same as <code>MqttConnectOptions.setMqttVersion</code>.</li>
 </ul>
 
 <h3 id="scala-api">Scala API</h3>
 
 <p>An example, for scala API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
 val lines = spark.readStream
   .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
   .option("topic", topic)
@@ -293,15 +306,14 @@ val query = wordCounts.writeStream
 
 query.awaitTermination()
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">MQTTStreamWordCount.scala</code> for full example.</p>
+<p>Please see <code>MQTTStreamWordCount.scala</code> for full example.</p>
 
 <h3 id="java-api">Java API</h3>
 
 <p>An example, for Java API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
 Dataset&lt;String&gt; lines = spark
         .readStream()
         .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
@@ -327,9 +339,8 @@ StreamingQuery query = wordCounts.writeStream()
 
 query.awaitTermination();
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">JavaMQTTStreamWordCount.java</code> for full example.</p>
+<p>Please see <code>JavaMQTTStreamWordCount.java</code> for full example.</p>
 
 
   </div>
@@ -339,7 +350,7 @@ query.awaitTermination();
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.1/spark-streaming-akka/index.html b/content/docs/spark/2.0.1/spark-streaming-akka/index.html
index 78e0a00..19fdfe0 100644
--- a/content/docs/spark/2.0.1/spark-streaming-akka/index.html
+++ b/content/docs/spark/2.0.1/spark-streaming-akka/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,42 +219,39 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.0.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.0.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-akka_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.0.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.0.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p>DStreams can be created with data streams received through Akka actors by using <code class="highlighter-rouge">AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
+<p>DStreams can be created with data streams received through Akka actors by using <code>AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends ActorReceiver {
+<pre><code>class CustomActor extends ActorReceiver {
   def receive = {
     case data: String =&gt; store(data)
   }
@@ -246,14 +261,13 @@ this actor can be configured to handle failures, etc.</p>
 val ssc: StreamingContext = ...
 val lines = AkkaUtils.createStream[String](ssc, Props[CustomActor](), "CustomReceiver")
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends JavaActorReceiver {
+<pre><code>class CustomActor extends JavaActorReceiver {
     @Override
     public void onReceive(Object msg) throws Exception {
         store((String) msg);
@@ -264,7 +278,6 @@ this actor can be configured to handle failures, etc.</p>
 JavaStreamingContext jssc = ...;
 JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Props.create(CustomActor.class), "CustomReceiver");
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-akka/examples">Akka Examples</a></p>
 
@@ -275,7 +288,7 @@ JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Pro
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.1/spark-streaming-mqtt/index.html b/content/docs/spark/2.0.1/spark-streaming-mqtt/index.html
index edf483f..912a0cd 100644
--- a/content/docs/spark/2.0.1/spark-streaming-mqtt/index.html
+++ b/content/docs/spark/2.0.1/spark-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.0.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.0.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.0.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.0.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -232,43 +247,42 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 <p>This source uses the <a href="https://eclipse.org/paho/clients/java/">Eclipse Paho Java Client</a>. Client API documentation is located <a href="http://www.eclipse.org/paho/files/javadoc/index.html">here</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">brokerUrl</code> A url MqttClient connects to. Set this as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
-  <li><code class="highlighter-rouge">storageLevel</code> By default it is used for storing incoming messages on disk.</li>
-  <li><code class="highlighter-rouge">topic</code> Topic MqttClient subscribes to.</li>
-  <li><code class="highlighter-rouge">clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
-  <li><code class="highlighter-rouge">QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
-  <li><code class="highlighter-rouge">username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
-  <li><code class="highlighter-rouge">password</code> Sets the password to use for the connection.</li>
-  <li><code class="highlighter-rouge">cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
-  <li><code class="highlighter-rouge">connectionTimeout</code> Sets the connection timeout, a value of 0 is interpreted as wait until client connects. See <code class="highlighter-rouge">MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
-  <li><code class="highlighter-rouge">keepAlive</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setKeepAliveInterval</code>.</li>
-  <li><code class="highlighter-rouge">mqttVersion</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setMqttVersion</code>.</li>
+  <li><code>brokerUrl</code> A url MqttClient connects to. Set this as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
+  <li><code>storageLevel</code> By default it is used for storing incoming messages on disk.</li>
+  <li><code>topic</code> Topic MqttClient subscribes to.</li>
+  <li><code>clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
+  <li><code>QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
+  <li><code>username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
+  <li><code>password</code> Sets the password to use for the connection.</li>
+  <li><code>cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
+  <li><code>connectionTimeout</code> Sets the connection timeout, a value of 0 is interpreted as wait until client connects. See <code>MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
+  <li><code>keepAlive</code> Same as <code>MqttConnectOptions.setKeepAliveInterval</code>.</li>
+  <li><code>mqttVersion</code> Same as <code>MqttConnectOptions.setMqttVersion</code>.</li>
 </ul>
 
 <h2 id="examples">Examples</h2>
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
+<pre><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
 </code></pre>
-</div>
 
 <p>Additional mqtt connection options can be provided:</p>
 
-<pre><code class="language-Scala">val lines = MQTTUtils.createStream(ssc, brokerUrl, topic, storageLevel, clientId, username, password, cleanSession, qos, connectionTimeout, keepAliveInterval, mqttVersion)
-</code></pre>
+<p><code>Scala
+val lines = MQTTUtils.createStream(ssc, brokerUrl, topic, storageLevel, clientId, username, password, cleanSession, qos, connectionTimeout, keepAliveInterval, mqttVersion)
+</code></p>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
+<pre><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-mqtt/examples">MQTT Examples</a></p>
 
@@ -279,7 +293,7 @@ this actor can be configured to handle failures, etc.</p>
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.1/spark-streaming-twitter/index.html b/content/docs/spark/2.0.1/spark-streaming-twitter/index.html
index cb87d48..aebcb17 100644
--- a/content/docs/spark/2.0.1/spark-streaming-twitter/index.html
+++ b/content/docs/spark/2.0.1/spark-streaming-twitter/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,52 +219,47 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.0.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.0.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-twitter_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.0.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.0.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p><code class="highlighter-rouge">TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
-can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code class="highlighter-rouge">TwitterUtils</code> class and create a DStream with <code class="highlighter-rouge">TwitterUtils.createStream</code> as shown below.</p>
+<p><code>TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
+can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code>TwitterUtils</code> class and create a DStream with <code>TwitterUtils.createStream</code> as shown below.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter._
+<pre><code>import org.apache.spark.streaming.twitter._
 
 TwitterUtils.createStream(ssc, None)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter.*;
+<pre><code>import org.apache.spark.streaming.twitter.*;
 
 TwitterUtils.createStream(jssc);
 </code></pre>
-</div>
 
 <p>You can also either get the public stream, or get the filtered stream based on keywords. 
 See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-twitter/examples">Twitter Examples</a></p>
@@ -258,7 +271,7 @@ See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.1/spark-streaming-zeromq/index.html b/content/docs/spark/2.0.1/spark-streaming-zeromq/index.html
index 0d17f6c..9c123ff 100644
--- a/content/docs/spark/2.0.1/spark-streaming-zeromq/index.html
+++ b/content/docs/spark/2.0.1/spark-streaming-zeromq/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.0.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.0.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-zeromq_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.0.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.0.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -231,15 +246,13 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = ZeroMQUtils.createStream(ssc, ...)
+<pre><code>val lines = ZeroMQUtils.createStream(ssc, ...)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
+<pre><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-zeromq/examples">ZeroMQ Examples</a></p>
 
@@ -250,7 +263,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.2/documentation/index.html b/content/docs/spark/2.0.2/documentation/index.html
index a4a02a1..3cdf493 100644
--- a/content/docs/spark/2.0.2/documentation/index.html
+++ b/content/docs/spark/2.0.2/documentation/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -222,7 +240,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.2/spark-sql-streaming-mqtt/index.html b/content/docs/spark/2.0.2/spark-sql-streaming-mqtt/index.html
index 26aee30..7679139 100644
--- a/content/docs/spark/2.0.2/spark-sql-streaming-mqtt/index.html
+++ b/content/docs/spark/2.0.2/spark-sql-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.0.2"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.0.2"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-sql-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.2&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.0.2
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.0.2
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is compiled for Scala 2.11 only, and intends to support Spark 2.0 onwards.</p>
 
@@ -231,49 +246,47 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <p>A SQL Stream can be created with data streams received through MQTT Server using,</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h2 id="enable-recovering-from-failures">Enable recovering from failures.</h2>
 
-<p>Setting values for option <code class="highlighter-rouge">localStorage</code> and <code class="highlighter-rouge">clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
+<p>Setting values for option <code>localStorage</code> and <code>clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .option("localStorage", "/path/to/localdir")
     .option("clientId", "some-client-id")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h2 id="configuration-options">Configuration options.</h2>
 
 <p>This source uses <a href="https://eclipse.org/paho/clients/java/">Eclipse Paho Java Client</a>. Client API documentation is located <a href="http://www.eclipse.org/paho/files/javadoc/index.html">here</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">brokerUrl</code> A url MqttClient connects to. Set this or <code class="highlighter-rouge">path</code> as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
-  <li><code class="highlighter-rouge">persistence</code> By default it is used for storing incoming messages on disk. If <code class="highlighter-rouge">memory</code> is provided as value for this option, then recovery on restart is not supported.</li>
-  <li><code class="highlighter-rouge">topic</code> Topic MqttClient subscribes to.</li>
-  <li><code class="highlighter-rouge">clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
-  <li><code class="highlighter-rouge">QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
-  <li><code class="highlighter-rouge">username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
-  <li><code class="highlighter-rouge">password</code> Sets the password to use for the connection.</li>
-  <li><code class="highlighter-rouge">cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
-  <li><code class="highlighter-rouge">connectionTimeout</code> Sets the connection timeout, a value of 0 is interpretted as wait until client connects. See <code class="highlighter-rouge">MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
-  <li><code class="highlighter-rouge">keepAlive</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setKeepAliveInterval</code>.</li>
-  <li><code class="highlighter-rouge">mqttVersion</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setMqttVersion</code>.</li>
+  <li><code>brokerUrl</code> A url MqttClient connects to. Set this or <code>path</code> as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
+  <li><code>persistence</code> By default it is used for storing incoming messages on disk. If <code>memory</code> is provided as value for this option, then recovery on restart is not supported.</li>
+  <li><code>topic</code> Topic MqttClient subscribes to.</li>
+  <li><code>clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
+  <li><code>QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
+  <li><code>username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
+  <li><code>password</code> Sets the password to use for the connection.</li>
+  <li><code>cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
+  <li><code>connectionTimeout</code> Sets the connection timeout, a value of 0 is interpretted as wait until client connects. See <code>MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
+  <li><code>keepAlive</code> Same as <code>MqttConnectOptions.setKeepAliveInterval</code>.</li>
+  <li><code>mqttVersion</code> Same as <code>MqttConnectOptions.setMqttVersion</code>.</li>
 </ul>
 
 <h3 id="scala-api">Scala API</h3>
 
 <p>An example, for scala API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
 val lines = spark.readStream
   .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
   .option("topic", topic)
@@ -293,15 +306,14 @@ val query = wordCounts.writeStream
 
 query.awaitTermination()
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">MQTTStreamWordCount.scala</code> for full example.</p>
+<p>Please see <code>MQTTStreamWordCount.scala</code> for full example.</p>
 
 <h3 id="java-api">Java API</h3>
 
 <p>An example, for Java API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
 Dataset&lt;String&gt; lines = spark
         .readStream()
         .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
@@ -327,9 +339,8 @@ StreamingQuery query = wordCounts.writeStream()
 
 query.awaitTermination();
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">JavaMQTTStreamWordCount.java</code> for full example.</p>
+<p>Please see <code>JavaMQTTStreamWordCount.java</code> for full example.</p>
 
 
   </div>
@@ -339,7 +350,7 @@ query.awaitTermination();
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.2/spark-streaming-akka/index.html b/content/docs/spark/2.0.2/spark-streaming-akka/index.html
index 89a14fd..ecfb9e7 100644
--- a/content/docs/spark/2.0.2/spark-streaming-akka/index.html
+++ b/content/docs/spark/2.0.2/spark-streaming-akka/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,42 +219,39 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.0.2"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.0.2"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-akka_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.2&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.0.2
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.0.2
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p>DStreams can be created with data streams received through Akka actors by using <code class="highlighter-rouge">AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
+<p>DStreams can be created with data streams received through Akka actors by using <code>AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends ActorReceiver {
+<pre><code>class CustomActor extends ActorReceiver {
   def receive = {
     case data: String =&gt; store(data)
   }
@@ -246,14 +261,13 @@ this actor can be configured to handle failures, etc.</p>
 val ssc: StreamingContext = ...
 val lines = AkkaUtils.createStream[String](ssc, Props[CustomActor](), "CustomReceiver")
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends JavaActorReceiver {
+<pre><code>class CustomActor extends JavaActorReceiver {
     @Override
     public void onReceive(Object msg) throws Exception {
         store((String) msg);
@@ -264,7 +278,6 @@ this actor can be configured to handle failures, etc.</p>
 JavaStreamingContext jssc = ...;
 JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Props.create(CustomActor.class), "CustomReceiver");
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-akka/examples">Akka Examples</a></p>
 
@@ -275,7 +288,7 @@ JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Pro
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.2/spark-streaming-mqtt/index.html b/content/docs/spark/2.0.2/spark-streaming-mqtt/index.html
index b3d5a9f..ad0a27c 100644
--- a/content/docs/spark/2.0.2/spark-streaming-mqtt/index.html
+++ b/content/docs/spark/2.0.2/spark-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.0.2"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.0.2"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.2&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.0.2
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.0.2
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -232,43 +247,42 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 <p>This source uses the <a href="https://eclipse.org/paho/clients/java/">Eclipse Paho Java Client</a>. Client API documentation is located <a href="http://www.eclipse.org/paho/files/javadoc/index.html">here</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">brokerUrl</code> A url MqttClient connects to. Set this as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
-  <li><code class="highlighter-rouge">storageLevel</code> By default it is used for storing incoming messages on disk.</li>
-  <li><code class="highlighter-rouge">topic</code> Topic MqttClient subscribes to.</li>
-  <li><code class="highlighter-rouge">clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
-  <li><code class="highlighter-rouge">QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
-  <li><code class="highlighter-rouge">username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
-  <li><code class="highlighter-rouge">password</code> Sets the password to use for the connection.</li>
-  <li><code class="highlighter-rouge">cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
-  <li><code class="highlighter-rouge">connectionTimeout</code> Sets the connection timeout, a value of 0 is interpreted as wait until client connects. See <code class="highlighter-rouge">MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
-  <li><code class="highlighter-rouge">keepAlive</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setKeepAliveInterval</code>.</li>
-  <li><code class="highlighter-rouge">mqttVersion</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setMqttVersion</code>.</li>
+  <li><code>brokerUrl</code> A url MqttClient connects to. Set this as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
+  <li><code>storageLevel</code> By default it is used for storing incoming messages on disk.</li>
+  <li><code>topic</code> Topic MqttClient subscribes to.</li>
+  <li><code>clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
+  <li><code>QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
+  <li><code>username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
+  <li><code>password</code> Sets the password to use for the connection.</li>
+  <li><code>cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
+  <li><code>connectionTimeout</code> Sets the connection timeout, a value of 0 is interpreted as wait until client connects. See <code>MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
+  <li><code>keepAlive</code> Same as <code>MqttConnectOptions.setKeepAliveInterval</code>.</li>
+  <li><code>mqttVersion</code> Same as <code>MqttConnectOptions.setMqttVersion</code>.</li>
 </ul>
 
 <h2 id="examples">Examples</h2>
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
+<pre><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
 </code></pre>
-</div>
 
 <p>Additional mqtt connection options can be provided:</p>
 
-<pre><code class="language-Scala">val lines = MQTTUtils.createStream(ssc, brokerUrl, topic, storageLevel, clientId, username, password, cleanSession, qos, connectionTimeout, keepAliveInterval, mqttVersion)
-</code></pre>
+<p><code>Scala
+val lines = MQTTUtils.createStream(ssc, brokerUrl, topic, storageLevel, clientId, username, password, cleanSession, qos, connectionTimeout, keepAliveInterval, mqttVersion)
+</code></p>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
+<pre><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-mqtt/examples">MQTT Examples</a></p>
 
@@ -279,7 +293,7 @@ this actor can be configured to handle failures, etc.</p>
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.2/spark-streaming-twitter/index.html b/content/docs/spark/2.0.2/spark-streaming-twitter/index.html
index b220e2b..d4d3df0 100644
--- a/content/docs/spark/2.0.2/spark-streaming-twitter/index.html
+++ b/content/docs/spark/2.0.2/spark-streaming-twitter/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,52 +219,47 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.0.2"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.0.2"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-twitter_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.2&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.0.2
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.0.2
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p><code class="highlighter-rouge">TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
-can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code class="highlighter-rouge">TwitterUtils</code> class and create a DStream with <code class="highlighter-rouge">TwitterUtils.createStream</code> as shown below.</p>
+<p><code>TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
+can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code>TwitterUtils</code> class and create a DStream with <code>TwitterUtils.createStream</code> as shown below.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter._
+<pre><code>import org.apache.spark.streaming.twitter._
 
 TwitterUtils.createStream(ssc, None)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter.*;
+<pre><code>import org.apache.spark.streaming.twitter.*;
 
 TwitterUtils.createStream(jssc);
 </code></pre>
-</div>
 
 <p>You can also either get the public stream, or get the filtered stream based on keywords. 
 See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-twitter/examples">Twitter Examples</a></p>
@@ -258,7 +271,7 @@ See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.0.2/spark-streaming-zeromq/index.html b/content/docs/spark/2.0.2/spark-streaming-zeromq/index.html
index e0884db..b16dcfd 100644
--- a/content/docs/spark/2.0.2/spark-streaming-zeromq/index.html
+++ b/content/docs/spark/2.0.2/spark-streaming-zeromq/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.0.2"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.0.2"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-zeromq_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.0.2&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.0.2
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.0.2
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -231,15 +246,13 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = ZeroMQUtils.createStream(ssc, ...)
+<pre><code>val lines = ZeroMQUtils.createStream(ssc, ...)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
+<pre><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-zeromq/examples">ZeroMQ Examples</a></p>
 
@@ -250,7 +263,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.0/documentation/index.html b/content/docs/spark/2.1.0/documentation/index.html
index 24e7ba9..ce6a7ef 100644
--- a/content/docs/spark/2.1.0/documentation/index.html
+++ b/content/docs/spark/2.1.0/documentation/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -222,7 +240,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.0/spark-sql-streaming-mqtt/index.html b/content/docs/spark/2.1.0/spark-sql-streaming-mqtt/index.html
index bf7270b..59adcd0 100644
--- a/content/docs/spark/2.1.0/spark-sql-streaming-mqtt/index.html
+++ b/content/docs/spark/2.1.0/spark-sql-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.1.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.1.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-sql-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.1.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.1.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is compiled for Scala 2.11 only, and intends to support Spark 2.0 onwards.</p>
 
@@ -231,49 +246,47 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <p>A SQL Stream can be created with data streams received through MQTT Server using,</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h2 id="enable-recovering-from-failures">Enable recovering from failures.</h2>
 
-<p>Setting values for option <code class="highlighter-rouge">localStorage</code> and <code class="highlighter-rouge">clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
+<p>Setting values for option <code>localStorage</code> and <code>clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .option("localStorage", "/path/to/localdir")
     .option("clientId", "some-client-id")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h2 id="configuration-options">Configuration options.</h2>
 
 <p>This source uses <a href="https://eclipse.org/paho/clients/java/">Eclipse Paho Java Client</a>. Client API documentation is located <a href="http://www.eclipse.org/paho/files/javadoc/index.html">here</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">brokerUrl</code> A url MqttClient connects to. Set this or <code class="highlighter-rouge">path</code> as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
-  <li><code class="highlighter-rouge">persistence</code> By default it is used for storing incoming messages on disk. If <code class="highlighter-rouge">memory</code> is provided as value for this option, then recovery on restart is not supported.</li>
-  <li><code class="highlighter-rouge">topic</code> Topic MqttClient subscribes to.</li>
-  <li><code class="highlighter-rouge">clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
-  <li><code class="highlighter-rouge">QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
-  <li><code class="highlighter-rouge">username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
-  <li><code class="highlighter-rouge">password</code> Sets the password to use for the connection.</li>
-  <li><code class="highlighter-rouge">cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
-  <li><code class="highlighter-rouge">connectionTimeout</code> Sets the connection timeout, a value of 0 is interpretted as wait until client connects. See <code class="highlighter-rouge">MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
-  <li><code class="highlighter-rouge">keepAlive</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setKeepAliveInterval</code>.</li>
-  <li><code class="highlighter-rouge">mqttVersion</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setMqttVersion</code>.</li>
+  <li><code>brokerUrl</code> A url MqttClient connects to. Set this or <code>path</code> as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
+  <li><code>persistence</code> By default it is used for storing incoming messages on disk. If <code>memory</code> is provided as value for this option, then recovery on restart is not supported.</li>
+  <li><code>topic</code> Topic MqttClient subscribes to.</li>
+  <li><code>clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
+  <li><code>QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
+  <li><code>username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
+  <li><code>password</code> Sets the password to use for the connection.</li>
+  <li><code>cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
+  <li><code>connectionTimeout</code> Sets the connection timeout, a value of 0 is interpretted as wait until client connects. See <code>MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
+  <li><code>keepAlive</code> Same as <code>MqttConnectOptions.setKeepAliveInterval</code>.</li>
+  <li><code>mqttVersion</code> Same as <code>MqttConnectOptions.setMqttVersion</code>.</li>
 </ul>
 
 <h3 id="scala-api">Scala API</h3>
 
 <p>An example, for scala API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
 val lines = spark.readStream
   .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
   .option("topic", topic)
@@ -293,15 +306,14 @@ val query = wordCounts.writeStream
 
 query.awaitTermination()
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">MQTTStreamWordCount.scala</code> for full example.</p>
+<p>Please see <code>MQTTStreamWordCount.scala</code> for full example.</p>
 
 <h3 id="java-api">Java API</h3>
 
 <p>An example, for Java API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
 Dataset&lt;String&gt; lines = spark
         .readStream()
         .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
@@ -327,9 +339,8 @@ StreamingQuery query = wordCounts.writeStream()
 
 query.awaitTermination();
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">JavaMQTTStreamWordCount.java</code> for full example.</p>
+<p>Please see <code>JavaMQTTStreamWordCount.java</code> for full example.</p>
 
 
   </div>
@@ -339,7 +350,7 @@ query.awaitTermination();
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.0/spark-streaming-akka/index.html b/content/docs/spark/2.1.0/spark-streaming-akka/index.html
index b27a7d3..dd96ad7 100644
--- a/content/docs/spark/2.1.0/spark-streaming-akka/index.html
+++ b/content/docs/spark/2.1.0/spark-streaming-akka/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,42 +219,39 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.1.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.1.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-akka_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.1.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.1.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p>DStreams can be created with data streams received through Akka actors by using <code class="highlighter-rouge">AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
+<p>DStreams can be created with data streams received through Akka actors by using <code>AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends ActorReceiver {
+<pre><code>class CustomActor extends ActorReceiver {
   def receive = {
     case data: String =&gt; store(data)
   }
@@ -246,14 +261,13 @@ this actor can be configured to handle failures, etc.</p>
 val ssc: StreamingContext = ...
 val lines = AkkaUtils.createStream[String](ssc, Props[CustomActor](), "CustomReceiver")
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends JavaActorReceiver {
+<pre><code>class CustomActor extends JavaActorReceiver {
     @Override
     public void onReceive(Object msg) throws Exception {
         store((String) msg);
@@ -264,7 +278,6 @@ this actor can be configured to handle failures, etc.</p>
 JavaStreamingContext jssc = ...;
 JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Props.create(CustomActor.class), "CustomReceiver");
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-akka/examples">Akka Examples</a></p>
 
@@ -275,7 +288,7 @@ JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Pro
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.0/spark-streaming-mqtt/index.html b/content/docs/spark/2.1.0/spark-streaming-mqtt/index.html
index afb28fc..2c9597b 100644
--- a/content/docs/spark/2.1.0/spark-streaming-mqtt/index.html
+++ b/content/docs/spark/2.1.0/spark-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.1.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.1.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.1.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.1.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -232,43 +247,42 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 <p>This source uses the <a href="https://eclipse.org/paho/clients/java/">Eclipse Paho Java Client</a>. Client API documentation is located <a href="http://www.eclipse.org/paho/files/javadoc/index.html">here</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">brokerUrl</code> A url MqttClient connects to. Set this as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
-  <li><code class="highlighter-rouge">storageLevel</code> By default it is used for storing incoming messages on disk.</li>
-  <li><code class="highlighter-rouge">topic</code> Topic MqttClient subscribes to.</li>
-  <li><code class="highlighter-rouge">clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
-  <li><code class="highlighter-rouge">QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
-  <li><code class="highlighter-rouge">username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
-  <li><code class="highlighter-rouge">password</code> Sets the password to use for the connection.</li>
-  <li><code class="highlighter-rouge">cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
-  <li><code class="highlighter-rouge">connectionTimeout</code> Sets the connection timeout, a value of 0 is interpreted as wait until client connects. See <code class="highlighter-rouge">MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
-  <li><code class="highlighter-rouge">keepAlive</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setKeepAliveInterval</code>.</li>
-  <li><code class="highlighter-rouge">mqttVersion</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setMqttVersion</code>.</li>
+  <li><code>brokerUrl</code> A url MqttClient connects to. Set this as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
+  <li><code>storageLevel</code> By default it is used for storing incoming messages on disk.</li>
+  <li><code>topic</code> Topic MqttClient subscribes to.</li>
+  <li><code>clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
+  <li><code>QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
+  <li><code>username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
+  <li><code>password</code> Sets the password to use for the connection.</li>
+  <li><code>cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
+  <li><code>connectionTimeout</code> Sets the connection timeout, a value of 0 is interpreted as wait until client connects. See <code>MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
+  <li><code>keepAlive</code> Same as <code>MqttConnectOptions.setKeepAliveInterval</code>.</li>
+  <li><code>mqttVersion</code> Same as <code>MqttConnectOptions.setMqttVersion</code>.</li>
 </ul>
 
 <h2 id="examples">Examples</h2>
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
+<pre><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
 </code></pre>
-</div>
 
 <p>Additional mqtt connection options can be provided:</p>
 
-<pre><code class="language-Scala">val lines = MQTTUtils.createStream(ssc, brokerUrl, topic, storageLevel, clientId, username, password, cleanSession, qos, connectionTimeout, keepAliveInterval, mqttVersion)
-</code></pre>
+<p><code>Scala
+val lines = MQTTUtils.createStream(ssc, brokerUrl, topic, storageLevel, clientId, username, password, cleanSession, qos, connectionTimeout, keepAliveInterval, mqttVersion)
+</code></p>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
+<pre><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-mqtt/examples">MQTT Examples</a></p>
 
@@ -279,7 +293,7 @@ this actor can be configured to handle failures, etc.</p>
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.0/spark-streaming-twitter/index.html b/content/docs/spark/2.1.0/spark-streaming-twitter/index.html
index d4ca4ea..3a9ee28 100644
--- a/content/docs/spark/2.1.0/spark-streaming-twitter/index.html
+++ b/content/docs/spark/2.1.0/spark-streaming-twitter/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,52 +219,47 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.1.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.1.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-twitter_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.1.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.1.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p><code class="highlighter-rouge">TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
-can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code class="highlighter-rouge">TwitterUtils</code> class and create a DStream with <code class="highlighter-rouge">TwitterUtils.createStream</code> as shown below.</p>
+<p><code>TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
+can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code>TwitterUtils</code> class and create a DStream with <code>TwitterUtils.createStream</code> as shown below.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter._
+<pre><code>import org.apache.spark.streaming.twitter._
 
 TwitterUtils.createStream(ssc, None)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter.*;
+<pre><code>import org.apache.spark.streaming.twitter.*;
 
 TwitterUtils.createStream(jssc);
 </code></pre>
-</div>
 
 <p>You can also either get the public stream, or get the filtered stream based on keywords. 
 See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-twitter/examples">Twitter Examples</a></p>
@@ -258,7 +271,7 @@ See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.0/spark-streaming-zeromq/index.html b/content/docs/spark/2.1.0/spark-streaming-zeromq/index.html
index 57ce9fe..2ef450a 100644
--- a/content/docs/spark/2.1.0/spark-streaming-zeromq/index.html
+++ b/content/docs/spark/2.1.0/spark-streaming-zeromq/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.1.0"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.1.0"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-zeromq_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.0&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.1.0
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.1.0
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -231,15 +246,13 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = ZeroMQUtils.createStream(ssc, ...)
+<pre><code>val lines = ZeroMQUtils.createStream(ssc, ...)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
+<pre><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-zeromq/examples">ZeroMQ Examples</a></p>
 
@@ -250,7 +263,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.1/documentation/index.html b/content/docs/spark/2.1.1/documentation/index.html
index df48613..20448fb 100644
--- a/content/docs/spark/2.1.1/documentation/index.html
+++ b/content/docs/spark/2.1.1/documentation/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -232,7 +250,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.1/spark-sql-cloudant/index.html b/content/docs/spark/2.1.1/spark-sql-cloudant/index.html
index d51f702..cab5160 100644
--- a/content/docs/spark/2.1.1/spark-sql-cloudant/index.html
+++ b/content/docs/spark/2.1.1/spark-sql-cloudant/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -209,40 +227,35 @@ clusters, desktop PCs, and mobile devices.</p>
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-cloudant" % "2.1.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-cloudant" % "2.1.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-sql-cloudant_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.</p>
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>Submit a job in Python:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>spark-submit  --master local[4] --jars &lt;path to cloudant-spark.jar&gt;  &lt;path to python script&gt;
+<pre><code>spark-submit  --master local[4] --jars &lt;path to cloudant-spark.jar&gt;  &lt;path to python script&gt;
 </code></pre>
-</div>
 
 <p>Submit a job in Scala:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>spark-submit --class "&lt;your class&gt;" --master local[4] --jars &lt;path to cloudant-spark.jar&gt; &lt;path to your app jar&gt;
+<pre><code>spark-submit --class "&lt;your class&gt;" --master local[4] --jars &lt;path to cloudant-spark.jar&gt; &lt;path to your app jar&gt;
 </code></pre>
-</div>
 
 <p>This library is compiled for Scala 2.11 only, and intends to support Spark 2.0 onwards.</p>
 
@@ -391,12 +404,12 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
   </tbody>
 </table>
 
-<p>For fast loading, views are loaded without include_docs. Thus, a derived schema will always be: <code class="highlighter-rouge"><span class="p">{</span><span class="err">id,</span><span class="w"> </span><span class="err">key,</span><span class="w"> </span><span class="err">value</span><span class="p">}</span></code>, where <code class="highlighter-rouge">value </code>can be a compount field. An example of loading data from a view:</p>
+<p>For fast loading, views are loaded without include_docs. Thus, a derived schema will always be: <code>{id, key, value}</code>, where <code>value </code>can be a compount field. An example of loading data from a view:</p>
 
-<div class="language-python highlighter-rouge"><pre class="highlight"><code><span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="p">(</span><span class="s">" CREATE TEMPORARY TABLE flightTable1 USING org.apache.bahir.cloudant OPTIONS ( database 'n_flight', view '_design/view/_view/AA0')"</span><span class="p">)</span>
+<p>```python
+spark.sql(“ CREATE TEMPORARY TABLE flightTable1 USING org.apache.bahir.cloudant OPTIONS ( database ‘n_flight’, view ‘_design/view/_view/AA0’)”)</p>
 
-</code></pre>
-</div>
+<p>```</p>
 
 <h3 id="configuration-on-cloudant-receiver-for-spark-streaming">Configuration on Cloudant Receiver for Spark Streaming</h3>
 
@@ -439,7 +452,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <h3 id="configuration-in-spark-submit-using---conf-option">Configuration in spark-submit using –conf option</h3>
 
-<p>The above stated configuration keys can also be set using <code class="highlighter-rouge">spark-submit --conf</code> option. When passing configuration in spark-submit, make sure adding “spark.” as prefix to the keys.</p>
+<p>The above stated configuration keys can also be set using <code>spark-submit --conf</code> option. When passing configuration in spark-submit, make sure adding “spark.” as prefix to the keys.</p>
 
 <h2 id="examples">Examples</h2>
 
@@ -447,62 +460,61 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <h4 id="using-sql-in-python">Using SQL In Python</h4>
 
-<div class="language-python highlighter-rouge"><pre class="highlight"><code><span class="n">spark</span> <span class="o">=</span> <span class="n">SparkSession</span>\
-    <span class="o">.</span><span class="n">builder</span>\
-    <span class="o">.</span><span class="n">appName</span><span class="p">(</span><span class="s">"Cloudant Spark SQL Example in Python using temp tables"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.host"</span><span class="p">,</span><span class="s">"ACCOUNT.cloudant.com"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.username"</span><span class="p">,</span> <span class="s">"USERNAME"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.password"</span><span class="p">,</span><span class="s">"PASSWORD"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">getOrCreate</span><span class="p">()</span>
-
-
-<span class="c"># Loading temp table from Cloudant db</span>
-<span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="p">(</span><span class="s">" CREATE TEMPORARY TABLE airportTable USING org.apache.bahir.cloudant OPTIONS ( database 'n_airportcodemapping')"</span><span class="p">)</span>
-<span class="n">airportData</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="p">(</span><span class="s">"SELECT _id, airportName FROM airportTable WHERE _id &gt;= 'CAA' AND _id &lt;= 'GAA' ORDER BY _id"</span><span class="p">)</span>
-<span class="n">airportData</span><span class="o">.</span><span class="n">printSchema</span><span class="p">()</span>
-<span class="k">print</span> <span class="s">'Total # of rows in airportData: '</span> <span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">airportData</span><span class="o">.</span><span class="n">count</span><span class="p">())</span>
-<span class="k">for</span> <span class="n">code</span> <span class="ow">in</span> <span class="n">airportData</span><span class="o">.</span><span class="n">collect</span><span class="p">():</span>
-    <span class="k">print</span> <span class="n">code</span><span class="o">.</span><span class="n">_id</span>
-</code></pre>
-</div>
+<p>```python
+spark = SparkSession\
+    .builder\
+    .appName(“Cloudant Spark SQL Example in Python using temp tables”)\
+    .config(“cloudant.host”,”ACCOUNT.cloudant.com”)\
+    .config(“cloudant.username”, “USERNAME”)\
+    .config(“cloudant.password”,”PASSWORD”)\
+    .getOrCreate()</p>
+
+<h1 id="loading-temp-table-from-cloudant-db">Loading temp table from Cloudant db</h1>
+<p>spark.sql(“ CREATE TEMPORARY TABLE airportTable USING org.apache.bahir.cloudant OPTIONS ( database ‘n_airportcodemapping’)”)
+airportData = spark.sql(“SELECT _id, airportName FROM airportTable WHERE _id &gt;= ‘CAA’ AND _id &lt;= ‘GAA’ ORDER BY _id”)
+airportData.printSchema()
+print ‘Total # of rows in airportData: ‘ + str(airportData.count())
+for code in airportData.collect():
+    print code._id
+```</p>
 
 <p>See <a href="examples/python/CloudantApp.py">CloudantApp.py</a> for examples.</p>
 
 <p>Submit job example:
-<code class="highlighter-rouge">
+<code>
 spark-submit  --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.1 --conf spark.cloudant.host=ACCOUNT.cloudant.com --conf spark.cloudant.username=USERNAME --conf spark.cloudant.password=PASSWORD sql-cloudant/examples/python/CloudantApp.py
 </code></p>
 
 <h4 id="using-dataframe-in-python">Using DataFrame In Python</h4>
 
-<div class="language-python highlighter-rouge"><pre class="highlight"><code><span class="n">spark</span> <span class="o">=</span> <span class="n">SparkSession</span>\
-    <span class="o">.</span><span class="n">builder</span>\
-    <span class="o">.</span><span class="n">appName</span><span class="p">(</span><span class="s">"Cloudant Spark SQL Example in Python using dataframes"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.host"</span><span class="p">,</span><span class="s">"ACCOUNT.cloudant.com"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.username"</span><span class="p">,</span> <span class="s">"USERNAME"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.password"</span><span class="p">,</span><span class="s">"PASSWORD"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"jsonstore.rdd.partitions"</span><span class="p">,</span> <span class="mi">8</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">getOrCreate</span><span class="p">()</span>
-
-<span class="c"># ***1. Loading dataframe from Cloudant db</span>
-<span class="n">df</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">read</span><span class="o">.</span><span class="n">load</span><span class="p">(</span><span class="s">"n_airportcodemapping"</span><span class="p">,</span> <span class="s">"org.apache.bahir.cloudant"</span><span class="p">)</span>
-<span class="n">df</span><span class="o">.</span><span class="n">cache</span><span class="p">()</span>
-<span class="n">df</span><span class="o">.</span><span class="n">printSchema</span><span class="p">()</span>
-<span class="n">df</span><span class="o">.</span><span class="nb">filter</span><span class="p">(</span><span class="n">df</span><span class="o">.</span><span class="n">airportName</span> <span class="o">&gt;=</span> <span class="s">'Moscow'</span><span class="p">)</span><span class="o">.</span><span class="n">select</span><span class="p">(</span><span class="s">"_id"</span><span class="p">,</span><span class="s">'airportName'</span><span class="p">)</span><span class="o">.</span><span cl [...]
-<span class="n">df</span><span class="o">.</span><span class="nb">filter</span><span class="p">(</span><span class="n">df</span><span class="o">.</span><span class="n">_id</span> <span class="o">&gt;=</span> <span class="s">'CAA'</span><span class="p">)</span><span class="o">.</span><span class="n">select</span><span class="p">(</span><span class="s">"_id"</span><span class="p">,</span><span class="s">'airportName'</span><span class="p">)</span><span class="o">.</span><span class="n">sho [...]
-</code></pre>
-</div>
+<p>```python
+spark = SparkSession\
+    .builder\
+    .appName(“Cloudant Spark SQL Example in Python using dataframes”)\
+    .config(“cloudant.host”,”ACCOUNT.cloudant.com”)\
+    .config(“cloudant.username”, “USERNAME”)\
+    .config(“cloudant.password”,”PASSWORD”)\
+    .config(“jsonstore.rdd.partitions”, 8)\
+    .getOrCreate()</p>
+
+<h1 id="loading-dataframe-from-cloudant-db">***1. Loading dataframe from Cloudant db</h1>
+<p>df = spark.read.load(“n_airportcodemapping”, “org.apache.bahir.cloudant”)
+df.cache()
+df.printSchema()
+df.filter(df.airportName &gt;= ‘Moscow’).select(“_id”,’airportName’).show()
+df.filter(df._id &gt;= ‘CAA’).select(“_id”,’airportName’).show()	  <br />
+```</p>
 
 <p>See <a href="examples/python/CloudantDF.py">CloudantDF.py</a> for examples.</p>
 
 <p>In case of doing multiple operations on a dataframe (select, filter etc.),
 you should persist a dataframe. Otherwise, every operation on a dataframe will load the same data from Cloudant again.
-Persisting will also speed up computation. This statement will persist an RDD in memory: <code class="highlighter-rouge">df.cache()</code>.  Alternatively for large dbs to persist in memory &amp; disk, use:</p>
+Persisting will also speed up computation. This statement will persist an RDD in memory: <code>df.cache()</code>.  Alternatively for large dbs to persist in memory &amp; disk, use:</p>
 
-<div class="language-python highlighter-rouge"><pre class="highlight"><code><span class="kn">from</span> <span class="nn">pyspark</span> <span class="kn">import</span> <span class="n">StorageLevel</span>
-<span class="n">df</span><span class="o">.</span><span class="n">persist</span><span class="p">(</span><span class="n">storageLevel</span> <span class="o">=</span> <span class="n">StorageLevel</span><span class="p">(</span><span class="bp">True</span><span class="p">,</span> <span class="bp">True</span><span class="p">,</span> <span class="bp">False</span><span class="p">,</span> <span class="bp">True</span><span class="p">,</span> <span class="mi">1</span><span class="p">))</span>
-</code></pre>
-</div>
+<p><code>python
+from pyspark import StorageLevel
+df.persist(storageLevel = StorageLevel(True, True, False, True, 1))
+</code></p>
 
 <p><a href="examples/python/CloudantDFOption.py">Sample code</a> on using DataFrame option to define cloudant configuration</p>
 
@@ -510,65 +522,65 @@ Persisting will also speed up computation. This statement will persist an RDD in
 
 <h4 id="using-sql-in-scala">Using SQL In Scala</h4>
 
-<div class="language-scala highlighter-rouge"><pre class="highlight"><code><span class="k">val</span> <span class="n">spark</span> <span class="k">=</span> <span class="nc">SparkSession</span>
-      <span class="o">.</span><span class="n">builder</span><span class="o">()</span>
-      <span class="o">.</span><span class="n">appName</span><span class="o">(</span><span class="s">"Cloudant Spark SQL Example"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.host"</span><span class="o">,</span><span class="s">"ACCOUNT.cloudant.com"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.username"</span><span class="o">,</span> <span class="s">"USERNAME"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.password"</span><span class="o">,</span><span class="s">"PASSWORD"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">getOrCreate</span><span class="o">()</span>
-
-<span class="c1">// For implicit conversions of Dataframe to RDDs
-</span><span class="k">import</span> <span class="nn">spark.implicits._</span>
-
-<span class="c1">// create a temp table from Cloudant db and query it using sql syntax
-</span><span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="o">(</span>
-    <span class="n">s</span><span class="s">"""
+<p>```scala
+val spark = SparkSession
+      .builder()
+      .appName(“Cloudant Spark SQL Example”)
+      .config(“cloudant.host”,”ACCOUNT.cloudant.com”)
+      .config(“cloudant.username”, “USERNAME”)
+      .config(“cloudant.password”,”PASSWORD”)
+      .getOrCreate()</p>
+
+<p>// For implicit conversions of Dataframe to RDDs
+import spark.implicits._</p>
+
+<p>// create a temp table from Cloudant db and query it using sql syntax
+spark.sql(
+    s”””
     |CREATE TEMPORARY TABLE airportTable
     |USING org.apache.bahir.cloudant
-    |OPTIONS ( database 'n_airportcodemapping')
-    """</span><span class="o">.</span><span class="n">stripMargin</span><span class="o">)</span>
-<span class="c1">// create a dataframe
-</span><span class="k">val</span> <span class="n">airportData</span> <span class="k">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="o">(</span><span class="s">"SELECT _id, airportName FROM airportTable WHERE _id &gt;= 'CAA' AND _id &lt;= 'GAA' ORDER BY _id"</span><span class="o">)</span>
-<span class="n">airportData</span><span class="o">.</span><span class="n">printSchema</span><span class="o">()</span>
-<span class="n">println</span><span class="o">(</span><span class="n">s</span><span class="s">"Total # of rows in airportData: "</span> <span class="o">+</span> <span class="n">airportData</span><span class="o">.</span><span class="n">count</span><span class="o">())</span>
-<span class="c1">// convert dataframe to array of Rows, and process each row
-</span><span class="n">airportData</span><span class="o">.</span><span class="n">map</span><span class="o">(</span><span class="n">t</span> <span class="k">=&gt;</span> <span class="s">"code: "</span> <span class="o">+</span> <span class="n">t</span><span class="o">(</span><span class="mi">0</span><span class="o">)</span> <span class="o">+</span> <span class="s">",name:"</span> <span class="o">+</span> <span class="n">t</span><span class="o">(</span><span class="mi">1</span><span class=" [...]
-</code></pre>
-</div>
-<p>See <a href="examples/scala/src/main/scala/mytest/spark/CloudantApp.scala">CloudantApp.scala</a> for examples.</p>
+    |OPTIONS ( database ‘n_airportcodemapping’)
+    “”“.stripMargin)
+// create a dataframe
+val airportData = spark.sql(“SELECT _id, airportName FROM airportTable WHERE _id &gt;= ‘CAA’ AND _id &lt;= ‘GAA’ ORDER BY _id”)
+airportData.printSchema()
+println(s”Total # of rows in airportData: “ + airportData.count())
+// convert dataframe to array of Rows, and process each row
+airportData.map(t =&gt; “code: “ + t(0) + “,name:” + t(1)).collect().foreach(println)
+```
+See <a href="examples/scala/src/main/scala/mytest/spark/CloudantApp.scala">CloudantApp.scala</a> for examples.</p>
 
 <p>Submit job example:
-<code class="highlighter-rouge">
+<code>
 spark-submit --class org.apache.spark.examples.sql.cloudant.CloudantApp --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.1 --conf spark.cloudant.host=ACCOUNT.cloudant.com --conf spark.cloudant.username=USERNAME --conf spark.cloudant.password=PASSWORD  /path/to/spark-sql-cloudant_2.11-2.1.1-tests.jar
 </code></p>
 
 <h3 id="using-dataframe-in-scala">Using DataFrame In Scala</h3>
 
-<div class="language-scala highlighter-rouge"><pre class="highlight"><code><span class="k">val</span> <span class="n">spark</span> <span class="k">=</span> <span class="nc">SparkSession</span>
-      <span class="o">.</span><span class="n">builder</span><span class="o">()</span>
-      <span class="o">.</span><span class="n">appName</span><span class="o">(</span><span class="s">"Cloudant Spark SQL Example with Dataframe"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.host"</span><span class="o">,</span><span class="s">"ACCOUNT.cloudant.com"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.username"</span><span class="o">,</span> <span class="s">"USERNAME"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.password"</span><span class="o">,</span><span class="s">"PASSWORD"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"createDBOnSave"</span><span class="o">,</span><span class="s">"true"</span><span class="o">)</span> <span class="c1">// to create a db on save
-</span>      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"jsonstore.rdd.partitions"</span><span class="o">,</span> <span class="s">"20"</span><span class="o">)</span> <span class="c1">// using 20 partitions
-</span>      <span class="o">.</span><span class="n">getOrCreate</span><span class="o">()</span>
-
-<span class="c1">// 1. Loading data from Cloudant db
-</span><span class="k">val</span> <span class="n">df</span> <span class="k">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">read</span><span class="o">.</span><span class="n">format</span><span class="o">(</span><span class="s">"org.apache.bahir.cloudant"</span><span class="o">).</span><span class="n">load</span><span class="o">(</span><span class="s">"n_flight"</span><span class="o">)</span>
-<span class="c1">// Caching df in memory to speed computations
+<p>```scala
+val spark = SparkSession
+      .builder()
+      .appName(“Cloudant Spark SQL Example with Dataframe”)
+      .config(“cloudant.host”,”ACCOUNT.cloudant.com”)
+      .config(“cloudant.username”, “USERNAME”)
+      .config(“cloudant.password”,”PASSWORD”)
+      .config(“createDBOnSave”,”true”) // to create a db on save
+      .config(“jsonstore.rdd.partitions”, “20”) // using 20 partitions
+      .getOrCreate()</p>
+
+<p>// 1. Loading data from Cloudant db
+val df = spark.read.format(“org.apache.bahir.cloudant”).load(“n_flight”)
+// Caching df in memory to speed computations
 // and not to retrieve data from cloudant again
-</span><span class="n">df</span><span class="o">.</span><span class="n">cache</span><span class="o">()</span>
-<span class="n">df</span><span class="o">.</span><span class="n">printSchema</span><span class="o">()</span>
-
-<span class="c1">// 2. Saving dataframe to Cloudant db
-</span><span class="k">val</span> <span class="n">df2</span> <span class="k">=</span> <span class="n">df</span><span class="o">.</span><span class="n">filter</span><span class="o">(</span><span class="n">df</span><span class="o">(</span><span class="s">"flightSegmentId"</span><span class="o">)</span> <span class="o">===</span> <span class="s">"AA106"</span><span class="o">)</span>
-    <span class="o">.</span><span class="n">select</span><span class="o">(</span><span class="s">"flightSegmentId"</span><span class="o">,</span><span class="s">"economyClassBaseCost"</span><span class="o">)</span>
-<span class="n">df2</span><span class="o">.</span><span class="n">show</span><span class="o">()</span>
-<span class="n">df2</span><span class="o">.</span><span class="n">write</span><span class="o">.</span><span class="n">format</span><span class="o">(</span><span class="s">"org.apache.bahir.cloudant"</span><span class="o">).</span><span class="n">save</span><span class="o">(</span><span class="s">"n_flight2"</span><span class="o">)</span>
-</code></pre>
-</div>
+df.cache()
+df.printSchema()</p>
+
+<p>// 2. Saving dataframe to Cloudant db
+val df2 = df.filter(df(“flightSegmentId”) === “AA106”)
+    .select(“flightSegmentId”,”economyClassBaseCost”)
+df2.show()
+df2.write.format(“org.apache.bahir.cloudant”).save(“n_flight2”)
+```</p>
 
 <p>See <a href="examples/scala/src/main/scala/mytest/spark/CloudantDF.scala">CloudantDF.scala</a> for examples.</p>
 
@@ -576,49 +588,49 @@ spark-submit --class org.apache.spark.examples.sql.cloudant.CloudantApp --packag
 
 <h3 id="using-streams-in-scala">Using Streams In Scala</h3>
 
-<div class="language-scala highlighter-rouge"><pre class="highlight"><code><span class="k">val</span> <span class="n">ssc</span> <span class="k">=</span> <span class="k">new</span> <span class="nc">StreamingContext</span><span class="o">(</span><span class="n">sparkConf</span><span class="o">,</span> <span class="nc">Seconds</span><span class="o">(</span><span class="mi">10</span><span class="o">))</span>
-<span class="k">val</span> <span class="n">changes</span> <span class="k">=</span> <span class="n">ssc</span><span class="o">.</span><span class="n">receiverStream</span><span class="o">(</span><span class="k">new</span> <span class="nc">CloudantReceiver</span><span class="o">(</span><span class="nc">Map</span><span class="o">(</span>
-  <span class="s">"cloudant.host"</span> <span class="o">-&gt;</span> <span class="s">"ACCOUNT.cloudant.com"</span><span class="o">,</span>
-  <span class="s">"cloudant.username"</span> <span class="o">-&gt;</span> <span class="s">"USERNAME"</span><span class="o">,</span>
-  <span class="s">"cloudant.password"</span> <span class="o">-&gt;</span> <span class="s">"PASSWORD"</span><span class="o">,</span>
-  <span class="s">"database"</span> <span class="o">-&gt;</span> <span class="s">"n_airportcodemapping"</span><span class="o">)))</span>
-
-<span class="n">changes</span><span class="o">.</span><span class="n">foreachRDD</span><span class="o">((</span><span class="n">rdd</span><span class="k">:</span> <span class="kt">RDD</span><span class="o">[</span><span class="kt">String</span><span class="o">],</span> <span class="n">time</span><span class="k">:</span> <span class="kt">Time</span><span class="o">)</span> <span class="k">=&gt;</span> <span class="o">{</span>
-  <span class="c1">// Get the singleton instance of SparkSession
-</span>  <span class="k">val</span> <span class="n">spark</span> <span class="k">=</span> <span class="nc">SparkSessionSingleton</span><span class="o">.</span><span class="n">getInstance</span><span class="o">(</span><span class="n">rdd</span><span class="o">.</span><span class="n">sparkContext</span><span class="o">.</span><span class="n">getConf</span><span class="o">)</span>
-
-  <span class="n">println</span><span class="o">(</span><span class="n">s</span><span class="s">"========= $time ========="</span><span class="o">)</span>
-  <span class="c1">// Convert RDD[String] to DataFrame
-</span>  <span class="k">val</span> <span class="n">changesDataFrame</span> <span class="k">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">read</span><span class="o">.</span><span class="n">json</span><span class="o">(</span><span class="n">rdd</span><span class="o">)</span>
-  <span class="k">if</span> <span class="o">(!</span><span class="n">changesDataFrame</span><span class="o">.</span><span class="n">schema</span><span class="o">.</span><span class="n">isEmpty</span><span class="o">)</span> <span class="o">{</span>
-    <span class="n">changesDataFrame</span><span class="o">.</span><span class="n">printSchema</span><span class="o">()</span>
-    <span class="n">changesDataFrame</span><span class="o">.</span><span class="n">select</span><span class="o">(</span><span class="s">"*"</span><span class="o">).</span><span class="n">show</span><span class="o">()</span>
-    <span class="o">....</span>
-  <span class="o">}</span>
-<span class="o">})</span>
-<span class="n">ssc</span><span class="o">.</span><span class="n">start</span><span class="o">()</span>
-<span class="c1">// run streaming for 120 secs
-</span><span class="nc">Thread</span><span class="o">.</span><span class="n">sleep</span><span class="o">(</span><span class="mi">120000L</span><span class="o">)</span>
-<span class="n">ssc</span><span class="o">.</span><span class="n">stop</span><span class="o">(</span><span class="kc">true</span><span class="o">)</span>
-
-</code></pre>
-</div>
+<p>```scala
+val ssc = new StreamingContext(sparkConf, Seconds(10))
+val changes = ssc.receiverStream(new CloudantReceiver(Map(
+  “cloudant.host” -&gt; “ACCOUNT.cloudant.com”,
+  “cloudant.username” -&gt; “USERNAME”,
+  “cloudant.password” -&gt; “PASSWORD”,
+  “database” -&gt; “n_airportcodemapping”)))</p>
+
+<p>changes.foreachRDD((rdd: RDD[String], time: Time) =&gt; {
+  // Get the singleton instance of SparkSession
+  val spark = SparkSessionSingleton.getInstance(rdd.sparkContext.getConf)</p>
+
+<p>println(s”========= $time =========”)
+  // Convert RDD[String] to DataFrame
+  val changesDataFrame = spark.read.json(rdd)
+  if (!changesDataFrame.schema.isEmpty) {
+    changesDataFrame.printSchema()
+    changesDataFrame.select(“*”).show()
+    ….
+  }
+})
+ssc.start()
+// run streaming for 120 secs
+Thread.sleep(120000L)
+ssc.stop(true)</p>
+
+<p>```</p>
 
 <p>See <a href="examples/scala/src/main/scala/mytest/spark/CloudantStreaming.scala">CloudantStreaming.scala</a> for examples.</p>
 
 <p>By default, Spark Streaming will load all documents from a database. If you want to limit the loading to
-specific documents, use <code class="highlighter-rouge">selector</code> option of <code class="highlighter-rouge">CloudantReceiver</code> and specify your conditions
+specific documents, use <code>selector</code> option of <code>CloudantReceiver</code> and specify your conditions
 (See <a href="examples/scala/src/main/scala/mytest/spark/CloudantStreamingSelector.scala">CloudantStreamingSelector.scala</a>
 example for more details):</p>
 
-<div class="language-scala highlighter-rouge"><pre class="highlight"><code><span class="k">val</span> <span class="n">changes</span> <span class="k">=</span> <span class="n">ssc</span><span class="o">.</span><span class="n">receiverStream</span><span class="o">(</span><span class="k">new</span> <span class="nc">CloudantReceiver</span><span class="o">(</span><span class="nc">Map</span><span class="o">(</span>
-  <span class="s">"cloudant.host"</span> <span class="o">-&gt;</span> <span class="s">"ACCOUNT.cloudant.com"</span><span class="o">,</span>
-  <span class="s">"cloudant.username"</span> <span class="o">-&gt;</span> <span class="s">"USERNAME"</span><span class="o">,</span>
-  <span class="s">"cloudant.password"</span> <span class="o">-&gt;</span> <span class="s">"PASSWORD"</span><span class="o">,</span>
-  <span class="s">"database"</span> <span class="o">-&gt;</span> <span class="s">"sales"</span><span class="o">,</span>
-  <span class="s">"selector"</span> <span class="o">-&gt;</span> <span class="s">"{\"month\":\"May\", \"rep\":\"John\"}"</span><span class="o">)))</span>
-</code></pre>
-</div>
+<p><code>scala
+val changes = ssc.receiverStream(new CloudantReceiver(Map(
+  "cloudant.host" -&gt; "ACCOUNT.cloudant.com",
+  "cloudant.username" -&gt; "USERNAME",
+  "cloudant.password" -&gt; "PASSWORD",
+  "database" -&gt; "sales",
+  "selector" -&gt; "{\"month\":\"May\", \"rep\":\"John\"}")))
+</code></p>
 
   </div>
 </div>
@@ -627,7 +639,7 @@ example for more details):</p>
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.1/spark-sql-streaming-akka/index.html b/content/docs/spark/2.1.1/spark-sql-streaming-akka/index.html
index 93fe3e4..d5b2817 100644
--- a/content/docs/spark/2.1.1/spark-sql-streaming-akka/index.html
+++ b/content/docs/spark/2.1.1/spark-sql-streaming-akka/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-akka" % "2.1.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-akka" % "2.1.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-sql-streaming-akka_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-akka_2.11:2.1.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-akka_2.11:2.1.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is compiled for Scala 2.11 only, and intends to support Spark 2.0 onwards.</p>
 
@@ -231,39 +246,37 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <p>A SQL Stream can be created with data streams received from Akka Feeder actor using,</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>    sqlContext.readStream
+<pre><code>    sqlContext.readStream
             .format("org.apache.bahir.sql.streaming.akka.AkkaStreamSourceProvider")
             .option("urlOfPublisher", "feederActorUri")
             .load()
 </code></pre>
-</div>
 
 <h2 id="enable-recovering-from-failures">Enable recovering from failures.</h2>
 
-<p>Setting values for option <code class="highlighter-rouge">persistenceDirPath</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
+<p>Setting values for option <code>persistenceDirPath</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>    sqlContext.readStream
+<pre><code>    sqlContext.readStream
             .format("org.apache.bahir.sql.streaming.akka.AkkaStreamSourceProvider")
             .option("urlOfPublisher", "feederActorUri")
             .option("persistenceDirPath", "/path/to/localdir")
             .load()
 </code></pre>
-</div>
 
 <h2 id="configuration-options">Configuration options.</h2>
 
 <p>This source uses <a href="http://doc.akka.io/api/akka/2.4/akka/actor/Actor.html">Akka Actor api</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">urlOfPublisher</code> The url of Publisher or Feeder actor that the Receiver actor connects to. Set this as the tcp url of the Publisher or Feeder actor.</li>
-  <li><code class="highlighter-rouge">persistenceDirPath</code> By default it is used for storing incoming messages on disk.</li>
+  <li><code>urlOfPublisher</code> The url of Publisher or Feeder actor that the Receiver actor connects to. Set this as the tcp url of the Publisher or Feeder actor.</li>
+  <li><code>persistenceDirPath</code> By default it is used for storing incoming messages on disk.</li>
 </ul>
 
 <h3 id="scala-api">Scala API</h3>
 
 <p>An example, for scala API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>    // Create DataFrame representing the stream of input lines from connection
+<pre><code>    // Create DataFrame representing the stream of input lines from connection
     // to publisher or feeder actor
     val lines = spark.readStream
                 .format("org.apache.bahir.sql.streaming.akka.AkkaStreamSourceProvider")
@@ -284,15 +297,14 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
     query.awaitTermination()
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">AkkaStreamWordCount.scala</code> for full example.</p>
+<p>Please see <code>AkkaStreamWordCount.scala</code> for full example.</p>
 
 <h3 id="java-api">Java API</h3>
 
 <p>An example, for Java API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>    // Create DataFrame representing the stream of input lines from connection
+<pre><code>    // Create DataFrame representing the stream of input lines from connection
     // to publisher or feeder actor
     Dataset&lt;String&gt; lines = spark
                             .readStream()
@@ -319,9 +331,8 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
     query.awaitTermination();   
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">JavaAkkaStreamWordCount.java</code> for full example.</p>
+<p>Please see <code>JavaAkkaStreamWordCount.java</code> for full example.</p>
 
   </div>
 </div>
@@ -330,7 +341,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.1/spark-sql-streaming-mqtt/index.html b/content/docs/spark/2.1.1/spark-sql-streaming-mqtt/index.html
index f72991f..642df95 100644
--- a/content/docs/spark/2.1.1/spark-sql-streaming-mqtt/index.html
+++ b/content/docs/spark/2.1.1/spark-sql-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.1.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-mqtt" % "2.1.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-sql-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.1.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-mqtt_2.11:2.1.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is compiled for Scala 2.11 only, and intends to support Spark 2.0 onwards.</p>
 
@@ -231,49 +246,47 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <p>A SQL Stream can be created with data streams received through MQTT Server using,</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h2 id="enable-recovering-from-failures">Enable recovering from failures.</h2>
 
-<p>Setting values for option <code class="highlighter-rouge">localStorage</code> and <code class="highlighter-rouge">clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
+<p>Setting values for option <code>localStorage</code> and <code>clientId</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>sqlContext.readStream
+<pre><code>sqlContext.readStream
     .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
     .option("topic", "mytopic")
     .option("localStorage", "/path/to/localdir")
     .option("clientId", "some-client-id")
     .load("tcp://localhost:1883")
 </code></pre>
-</div>
 
 <h2 id="configuration-options">Configuration options.</h2>
 
 <p>This source uses <a href="https://eclipse.org/paho/clients/java/">Eclipse Paho Java Client</a>. Client API documentation is located <a href="http://www.eclipse.org/paho/files/javadoc/index.html">here</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">brokerUrl</code> A url MqttClient connects to. Set this or <code class="highlighter-rouge">path</code> as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
-  <li><code class="highlighter-rouge">persistence</code> By default it is used for storing incoming messages on disk. If <code class="highlighter-rouge">memory</code> is provided as value for this option, then recovery on restart is not supported.</li>
-  <li><code class="highlighter-rouge">topic</code> Topic MqttClient subscribes to.</li>
-  <li><code class="highlighter-rouge">clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
-  <li><code class="highlighter-rouge">QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
-  <li><code class="highlighter-rouge">username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
-  <li><code class="highlighter-rouge">password</code> Sets the password to use for the connection.</li>
-  <li><code class="highlighter-rouge">cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
-  <li><code class="highlighter-rouge">connectionTimeout</code> Sets the connection timeout, a value of 0 is interpretted as wait until client connects. See <code class="highlighter-rouge">MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
-  <li><code class="highlighter-rouge">keepAlive</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setKeepAliveInterval</code>.</li>
-  <li><code class="highlighter-rouge">mqttVersion</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setMqttVersion</code>.</li>
+  <li><code>brokerUrl</code> A url MqttClient connects to. Set this or <code>path</code> as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
+  <li><code>persistence</code> By default it is used for storing incoming messages on disk. If <code>memory</code> is provided as value for this option, then recovery on restart is not supported.</li>
+  <li><code>topic</code> Topic MqttClient subscribes to.</li>
+  <li><code>clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
+  <li><code>QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
+  <li><code>username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
+  <li><code>password</code> Sets the password to use for the connection.</li>
+  <li><code>cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
+  <li><code>connectionTimeout</code> Sets the connection timeout, a value of 0 is interpretted as wait until client connects. See <code>MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
+  <li><code>keepAlive</code> Same as <code>MqttConnectOptions.setKeepAliveInterval</code>.</li>
+  <li><code>mqttVersion</code> Same as <code>MqttConnectOptions.setMqttVersion</code>.</li>
 </ul>
 
 <h3 id="scala-api">Scala API</h3>
 
 <p>An example, for scala API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server
 val lines = spark.readStream
   .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
   .option("topic", topic)
@@ -293,15 +306,14 @@ val query = wordCounts.writeStream
 
 query.awaitTermination()
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">MQTTStreamWordCount.scala</code> for full example.</p>
+<p>Please see <code>MQTTStreamWordCount.scala</code> for full example.</p>
 
 <h3 id="java-api">Java API</h3>
 
 <p>An example, for Java API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
+<pre><code>// Create DataFrame representing the stream of input lines from connection to mqtt server.
 Dataset&lt;String&gt; lines = spark
         .readStream()
         .format("org.apache.bahir.sql.streaming.mqtt.MQTTStreamSourceProvider")
@@ -327,9 +339,8 @@ StreamingQuery query = wordCounts.writeStream()
 
 query.awaitTermination();
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">JavaMQTTStreamWordCount.java</code> for full example.</p>
+<p>Please see <code>JavaMQTTStreamWordCount.java</code> for full example.</p>
 
   </div>
 </div>
@@ -338,7 +349,7 @@ query.awaitTermination();
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.1/spark-streaming-akka/index.html b/content/docs/spark/2.1.1/spark-streaming-akka/index.html
index 459785b..eba5792 100644
--- a/content/docs/spark/2.1.1/spark-streaming-akka/index.html
+++ b/content/docs/spark/2.1.1/spark-streaming-akka/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,42 +219,39 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.1.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-akka" % "2.1.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-akka_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.1.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-akka_2.11:2.1.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p>DStreams can be created with data streams received through Akka actors by using <code class="highlighter-rouge">AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
+<p>DStreams can be created with data streams received through Akka actors by using <code>AkkaUtils.createStream(ssc, actorProps, actor-name)</code>.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends ActorReceiver {
+<pre><code>class CustomActor extends ActorReceiver {
   def receive = {
     case data: String =&gt; store(data)
   }
@@ -246,14 +261,13 @@ this actor can be configured to handle failures, etc.</p>
 val ssc: StreamingContext = ...
 val lines = AkkaUtils.createStream[String](ssc, Props[CustomActor](), "CustomReceiver")
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>class CustomActor extends JavaActorReceiver {
+<pre><code>class CustomActor extends JavaActorReceiver {
     @Override
     public void onReceive(Object msg) throws Exception {
         store((String) msg);
@@ -264,7 +278,6 @@ this actor can be configured to handle failures, etc.</p>
 JavaStreamingContext jssc = ...;
 JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Props.create(CustomActor.class), "CustomReceiver");
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-akka/examples">Akka Examples</a></p>
 
@@ -275,7 +288,7 @@ JavaDStream&lt;String&gt; lines = AkkaUtils.&lt;String&gt;createStream(jssc, Pro
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.1/spark-streaming-mqtt/index.html b/content/docs/spark/2.1.1/spark-streaming-mqtt/index.html
index 3127d82..c70bee3 100644
--- a/content/docs/spark/2.1.1/spark-streaming-mqtt/index.html
+++ b/content/docs/spark/2.1.1/spark-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.1.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-mqtt" % "2.1.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-mqtt_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.1.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-mqtt_2.11:2.1.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -232,47 +247,46 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 <p>This source uses the <a href="https://eclipse.org/paho/clients/java/">Eclipse Paho Java Client</a>. Client API documentation is located <a href="http://www.eclipse.org/paho/files/javadoc/index.html">here</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">brokerUrl</code> A url MqttClient connects to. Set this as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
-  <li><code class="highlighter-rouge">storageLevel</code> By default it is used for storing incoming messages on disk.</li>
-  <li><code class="highlighter-rouge">topic</code> Topic MqttClient subscribes to.</li>
-  <li><code class="highlighter-rouge">topics</code> List of topics MqttClient subscribes to.</li>
-  <li><code class="highlighter-rouge">clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
-  <li><code class="highlighter-rouge">QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
-  <li><code class="highlighter-rouge">username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
-  <li><code class="highlighter-rouge">password</code> Sets the password to use for the connection.</li>
-  <li><code class="highlighter-rouge">cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
-  <li><code class="highlighter-rouge">connectionTimeout</code> Sets the connection timeout, a value of 0 is interpreted as wait until client connects. See <code class="highlighter-rouge">MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
-  <li><code class="highlighter-rouge">keepAlive</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setKeepAliveInterval</code>.</li>
-  <li><code class="highlighter-rouge">mqttVersion</code> Same as <code class="highlighter-rouge">MqttConnectOptions.setMqttVersion</code>.</li>
+  <li><code>brokerUrl</code> A url MqttClient connects to. Set this as the url of the Mqtt Server. e.g. tcp://localhost:1883.</li>
+  <li><code>storageLevel</code> By default it is used for storing incoming messages on disk.</li>
+  <li><code>topic</code> Topic MqttClient subscribes to.</li>
+  <li><code>topics</code> List of topics MqttClient subscribes to.</li>
+  <li><code>clientId</code> clientId, this client is assoicated with. Provide the same value to recover a stopped client.</li>
+  <li><code>QoS</code> The maximum quality of service to subscribe each topic at. Messages published at a lower quality of service will be received at the published QoS. Messages published at a higher quality of service will be received using the QoS specified on the subscribe.</li>
+  <li><code>username</code> Sets the user name to use for the connection to Mqtt Server. Do not set it, if server does not need this. Setting it empty will lead to errors.</li>
+  <li><code>password</code> Sets the password to use for the connection.</li>
+  <li><code>cleanSession</code> Setting it true starts a clean session, removes all checkpointed messages by a previous run of this source. This is set to false by default.</li>
+  <li><code>connectionTimeout</code> Sets the connection timeout, a value of 0 is interpreted as wait until client connects. See <code>MqttConnectOptions.setConnectionTimeout</code> for more information.</li>
+  <li><code>keepAlive</code> Same as <code>MqttConnectOptions.setKeepAliveInterval</code>.</li>
+  <li><code>mqttVersion</code> Same as <code>MqttConnectOptions.setMqttVersion</code>.</li>
 </ul>
 
 <h2 id="examples">Examples</h2>
 
 <h3 id="scala-api">Scala API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">ActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>ActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
+<pre><code>val lines = MQTTUtils.createStream(ssc, brokerUrl, topic)
 val lines = MQTTUtils.createPairedStream(ssc, brokerUrl, topic)
 </code></pre>
-</div>
 
 <p>Additional mqtt connection options can be provided:</p>
 
-<pre><code class="language-Scala">val lines = MQTTUtils.createStream(ssc, brokerUrl, topic, storageLevel, clientId, username, password, cleanSession, qos, connectionTimeout, keepAliveInterval, mqttVersion)
+<p><code>Scala
+val lines = MQTTUtils.createStream(ssc, brokerUrl, topic, storageLevel, clientId, username, password, cleanSession, qos, connectionTimeout, keepAliveInterval, mqttVersion)
 val lines = MQTTUtils.createPairedStream(ssc, brokerUrl, topics, storageLevel, clientId, username, password, cleanSession, qos, connectionTimeout, keepAliveInterval, mqttVersion)
-</code></pre>
+</code></p>
 
 <h3 id="java-api">Java API</h3>
 
-<p>You need to extend <code class="highlighter-rouge">JavaActorReceiver</code> so as to store received data into Spark using <code class="highlighter-rouge">store(...)</code> methods. The supervisor strategy of
+<p>You need to extend <code>JavaActorReceiver</code> so as to store received data into Spark using <code>store(...)</code> methods. The supervisor strategy of
 this actor can be configured to handle failures, etc.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
+<pre><code>JavaDStream&lt;String&gt; lines = MQTTUtils.createStream(jssc, brokerUrl, topic);
 JavaReceiverInputDStream&lt;Tuple2&lt;String, String&gt;&gt; lines = MQTTUtils.createPairedStream(jssc, brokerUrl, topics);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-mqtt/examples">MQTT Examples</a></p>
 
@@ -283,7 +297,7 @@ JavaReceiverInputDStream&lt;Tuple2&lt;String, String&gt;&gt; lines = MQTTUtils.c
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.1/spark-streaming-pubsub/index.html b/content/docs/spark/2.1.1/spark-streaming-pubsub/index.html
index 8c5cc22..38617d5 100644
--- a/content/docs/spark/2.1.1/spark-streaming-pubsub/index.html
+++ b/content/docs/spark/2.1.1/spark-streaming-pubsub/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,53 +219,48 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-pubsub" % "2.1.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-pubsub" % "2.1.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-pubsub_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-pubsub_2.11:2.1.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-pubsub_2.11:2.1.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <h2 id="examples">Examples</h2>
 
 <p>First you need to create credential by SparkGCPCredentials, it support four type of credentials
 * application default
-    <code class="highlighter-rouge">SparkGCPCredentials.builder.build()</code>
+    <code>SparkGCPCredentials.builder.build()</code>
 * json type service account
-    <code class="highlighter-rouge">SparkGCPCredentials.builder.jsonServiceAccount(PATH_TO_JSON_KEY).build()</code>
+    <code>SparkGCPCredentials.builder.jsonServiceAccount(PATH_TO_JSON_KEY).build()</code>
 * p12 type service account
-    <code class="highlighter-rouge">SparkGCPCredentials.builder.p12ServiceAccount(PATH_TO_P12_KEY, EMAIL_ACCOUNT).build()</code>
+    <code>SparkGCPCredentials.builder.p12ServiceAccount(PATH_TO_P12_KEY, EMAIL_ACCOUNT).build()</code>
 * metadata service account(running on dataproc)
-    <code class="highlighter-rouge">SparkGCPCredentials.builder.metadataServiceAccount().build()</code></p>
+    <code>SparkGCPCredentials.builder.metadataServiceAccount().build()</code></p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = PubsubUtils.createStream(ssc, projectId, subscriptionName, credential, ..)
+<pre><code>val lines = PubsubUtils.createStream(ssc, projectId, subscriptionName, credential, ..)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;SparkPubsubMessage&gt; lines = PubsubUtils.createStream(jssc, projectId, subscriptionName, credential...)
+<pre><code>JavaDStream&lt;SparkPubsubMessage&gt; lines = PubsubUtils.createStream(jssc, projectId, subscriptionName, credential...)
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="streaming-pubsub/examples">Google Cloud Pubsub Examples</a></p>
 
@@ -258,7 +271,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.1/spark-streaming-twitter/index.html b/content/docs/spark/2.1.1/spark-streaming-twitter/index.html
index 164b578..57e2849 100644
--- a/content/docs/spark/2.1.1/spark-streaming-twitter/index.html
+++ b/content/docs/spark/2.1.1/spark-streaming-twitter/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,52 +219,47 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.1.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-twitter" % "2.1.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-twitter_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.1.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-twitter_2.11:2.1.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
 <h2 id="examples">Examples</h2>
 
-<p><code class="highlighter-rouge">TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
-can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code class="highlighter-rouge">TwitterUtils</code> class and create a DStream with <code class="highlighter-rouge">TwitterUtils.createStream</code> as shown below.</p>
+<p><code>TwitterUtils</code> uses Twitter4j to get the public stream of tweets using <a href="https://dev.twitter.com/docs/streaming-apis">Twitter’s Streaming API</a>. Authentication information
+can be provided by any of the <a href="http://twitter4j.org/en/configuration.html">methods</a> supported by Twitter4J library. You can import the <code>TwitterUtils</code> class and create a DStream with <code>TwitterUtils.createStream</code> as shown below.</p>
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter._
+<pre><code>import org.apache.spark.streaming.twitter._
 
 TwitterUtils.createStream(ssc, None)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>import org.apache.spark.streaming.twitter.*;
+<pre><code>import org.apache.spark.streaming.twitter.*;
 
 TwitterUtils.createStream(jssc);
 </code></pre>
-</div>
 
 <p>You can also either get the public stream, or get the filtered stream based on keywords.
 See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-twitter/examples">Twitter Examples</a></p>
@@ -258,7 +271,7 @@ See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.1/spark-streaming-zeromq/index.html b/content/docs/spark/2.1.1/spark-streaming-zeromq/index.html
index f639572..493cf2f 100644
--- a/content/docs/spark/2.1.1/spark-streaming-zeromq/index.html
+++ b/content/docs/spark/2.1.1/spark-streaming-zeromq/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.1.1"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-streaming-zeromq" % "2.1.1"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-streaming-zeromq_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.1&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.1.1
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-streaming-zeromq_2.11:2.1.1
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is cross-published for Scala 2.10 and Scala 2.11, so users should replace the proper Scala version (2.10 or 2.11) in the commands listed above.</p>
 
@@ -231,15 +246,13 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <h3 id="scala-api">Scala API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>val lines = ZeroMQUtils.createStream(ssc, ...)
+<pre><code>val lines = ZeroMQUtils.createStream(ssc, ...)
 </code></pre>
-</div>
 
 <h3 id="java-api">Java API</h3>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
+<pre><code>JavaDStream&lt;String&gt; lines = ZeroMQUtils.createStream(jssc, ...);
 </code></pre>
-</div>
 
 <p>See end-to-end examples at <a href="https://github.com/apache/bahir/tree/master/streaming-zeromq/examples">ZeroMQ Examples</a></p>
 
@@ -250,7 +263,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.2/documentation/index.html b/content/docs/spark/2.1.2/documentation/index.html
index d535561..a949286 100644
--- a/content/docs/spark/2.1.2/documentation/index.html
+++ b/content/docs/spark/2.1.2/documentation/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -234,7 +252,7 @@
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.2/spark-sql-cloudant/index.html b/content/docs/spark/2.1.2/spark-sql-cloudant/index.html
index 7ad52c9..c5bf0fc 100644
--- a/content/docs/spark/2.1.2/spark-sql-cloudant/index.html
+++ b/content/docs/spark/2.1.2/spark-sql-cloudant/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -209,40 +227,35 @@ clusters, desktop PCs, and mobile devices.</p>
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-cloudant" % "2.1.2"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-cloudant" % "2.1.2"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-sql-cloudant_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.2&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.</p>
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.2
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.2
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>Submit a job in Python:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>spark-submit  --master local[4] --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.2  &lt;path to python script&gt;
+<pre><code>spark-submit  --master local[4] --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.2  &lt;path to python script&gt;
 </code></pre>
-</div>
 
 <p>Submit a job in Scala:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>spark-submit --class "&lt;your class&gt;" --master local[4] --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.2 &lt;path to spark-sql-cloudant jar&gt;
+<pre><code>spark-submit --class "&lt;your class&gt;" --master local[4] --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.2 &lt;path to spark-sql-cloudant jar&gt;
 </code></pre>
-</div>
 
 <p>This library is compiled for Scala 2.11 only, and intends to support Spark 2.0 onwards.</p>
 
@@ -275,12 +288,12 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
     <tr>
       <td>cloudant.batchInterval</td>
       <td style="text-align: center">8</td>
-      <td>number of seconds to set for streaming all documents from <code class="highlighter-rouge">_changes</code> endpoint into Spark dataframe.  See <a href="https://spark.apache.org/docs/latest/streaming-programming-guide.html#setting-the-right-batch-interval">Setting the right batch interval</a> for tuning this value.</td>
+      <td>number of seconds to set for streaming all documents from <code>_changes</code> endpoint into Spark dataframe.  See <a href="https://spark.apache.org/docs/latest/streaming-programming-guide.html#setting-the-right-batch-interval">Setting the right batch interval</a> for tuning this value.</td>
     </tr>
     <tr>
       <td>cloudant.endpoint</td>
-      <td style="text-align: center"><code class="highlighter-rouge">_all_docs</code></td>
-      <td>endpoint for RelationProvider when loading data from Cloudant to DataFrames or SQL temporary tables. Select between the Cloudant <code class="highlighter-rouge">_all_docs</code> or <code class="highlighter-rouge">_changes</code> API endpoint.  See <strong>Note</strong> below for differences between endpoints.</td>
+      <td style="text-align: center"><code>_all_docs</code></td>
+      <td>endpoint for RelationProvider when loading data from Cloudant to DataFrames or SQL temporary tables. Select between the Cloudant <code>_all_docs</code> or <code>_changes</code> API endpoint.  See <strong>Note</strong> below for differences between endpoints.</td>
     </tr>
     <tr>
       <td>cloudant.protocol</td>
@@ -305,32 +318,32 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
     <tr>
       <td>cloudant.numberOfRetries</td>
       <td style="text-align: center">3</td>
-      <td>number of times to replay a request that received a 429 <code class="highlighter-rouge">Too Many Requests</code> response</td>
+      <td>number of times to replay a request that received a 429 <code>Too Many Requests</code> response</td>
     </tr>
     <tr>
       <td>cloudant.useQuery</td>
       <td style="text-align: center">false</td>
-      <td>by default, <code class="highlighter-rouge">_all_docs</code> endpoint is used if configuration ‘view’ and ‘index’ (see below) are not set. When useQuery is enabled, <code class="highlighter-rouge">_find</code> endpoint will be used in place of <code class="highlighter-rouge">_all_docs</code> when query condition is not on primary key field (_id), so that query predicates may be driven into datastore.</td>
+      <td>by default, <code>_all_docs</code> endpoint is used if configuration ‘view’ and ‘index’ (see below) are not set. When useQuery is enabled, <code>_find</code> endpoint will be used in place of <code>_all_docs</code> when query condition is not on primary key field (_id), so that query predicates may be driven into datastore.</td>
     </tr>
     <tr>
       <td>cloudant.queryLimit</td>
       <td style="text-align: center">25</td>
-      <td>the maximum number of results returned when querying the <code class="highlighter-rouge">_find</code> endpoint.</td>
+      <td>the maximum number of results returned when querying the <code>_find</code> endpoint.</td>
     </tr>
     <tr>
       <td>cloudant.storageLevel</td>
       <td style="text-align: center">MEMORY_ONLY</td>
-      <td>the storage level for persisting Spark RDDs during load when <code class="highlighter-rouge">cloudant.endpoint</code> is set to <code class="highlighter-rouge">_changes</code>.  See <a href="https://spark.apache.org/docs/latest/programming-guide.html#rdd-persistence">RDD Persistence section</a> in Spark’s Progamming Guide for all available storage level options.</td>
+      <td>the storage level for persisting Spark RDDs during load when <code>cloudant.endpoint</code> is set to <code>_changes</code>.  See <a href="https://spark.apache.org/docs/latest/programming-guide.html#rdd-persistence">RDD Persistence section</a> in Spark’s Progamming Guide for all available storage level options.</td>
     </tr>
     <tr>
       <td>cloudant.timeout</td>
       <td style="text-align: center">60000</td>
-      <td>stop the response after waiting the defined number of milliseconds for data.  Only supported with <code class="highlighter-rouge">changes</code> endpoint.</td>
+      <td>stop the response after waiting the defined number of milliseconds for data.  Only supported with <code>changes</code> endpoint.</td>
     </tr>
     <tr>
       <td>jsonstore.rdd.partitions</td>
       <td style="text-align: center">10</td>
-      <td>the number of partitions intent used to drive JsonStoreRDD loading query result in parallel. The actual number is calculated based on total rows returned and satisfying maxInPartition and minInPartition. Only supported with <code class="highlighter-rouge">_all_docs</code> endpoint.</td>
+      <td>the number of partitions intent used to drive JsonStoreRDD loading query result in parallel. The actual number is calculated based on total rows returned and satisfying maxInPartition and minInPartition. Only supported with <code>_all_docs</code> endpoint.</td>
     </tr>
     <tr>
       <td>jsonstore.rdd.maxInPartition</td>
@@ -355,7 +368,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
     <tr>
       <td>schemaSampleSize</td>
       <td style="text-align: center">-1</td>
-      <td>the sample size for RDD schema discovery. 1 means we are using only the first document for schema discovery; -1 means all documents; 0 will be treated as 1; any number N means min(N, total) docs. Only supported with <code class="highlighter-rouge">_all_docs</code> endpoint.</td>
+      <td>the sample size for RDD schema discovery. 1 means we are using only the first document for schema discovery; -1 means all documents; 0 will be treated as 1; any number N means min(N, total) docs. Only supported with <code>_all_docs</code> endpoint.</td>
     </tr>
     <tr>
       <td>createDBOnSave</td>
@@ -365,27 +378,27 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
   </tbody>
 </table>
 
-<p>The <code class="highlighter-rouge">cloudant.endpoint</code> option sets ` _changes<code class="highlighter-rouge"> or </code>_all_docs` API endpoint to be called while loading Cloudant data into Spark DataFrames or SQL Tables.</p>
+<p>The <code>cloudant.endpoint</code> option sets ` _changes<code> or </code>_all_docs` API endpoint to be called while loading Cloudant data into Spark DataFrames or SQL Tables.</p>
 
-<p><strong>Note:</strong> When using <code class="highlighter-rouge">_changes</code> API, please consider:
+<p><strong>Note:</strong> When using <code>_changes</code> API, please consider:
 1. Results are partially ordered and may not be be presented in order in
 which documents were updated.
 2. In case of shards’ unavailability, you may see duplicate results (changes that have been seen already)
-3. Can use <code class="highlighter-rouge">selector</code> option to filter Cloudant docs during load
+3. Can use <code>selector</code> option to filter Cloudant docs during load
 4. Supports a real snapshot of the database and represents it in a single point of time.
 5. Only supports a single partition.</p>
 
-<p>When using <code class="highlighter-rouge">_all_docs</code> API:
+<p>When using <code>_all_docs</code> API:
 1. Supports parallel reads (using offset and range) and partitioning.
 2. Using partitions may not represent the true snapshot of a database.  Some docs
    may be added or deleted in the database between loading data into different
    Spark partitions.</p>
 
-<p>If loading Cloudant docs from a database greater than 100 MB, set <code class="highlighter-rouge">cloudant.endpoint</code> to <code class="highlighter-rouge">_changes</code> and <code class="highlighter-rouge">spark.streaming.unpersist</code> to <code class="highlighter-rouge">false</code>.
-This will enable RDD persistence during load against <code class="highlighter-rouge">_changes</code> endpoint and allow the persisted RDDs to be accessible after streaming completes.</p>
+<p>If loading Cloudant docs from a database greater than 100 MB, set <code>cloudant.endpoint</code> to <code>_changes</code> and <code>spark.streaming.unpersist</code> to <code>false</code>.
+This will enable RDD persistence during load against <code>_changes</code> endpoint and allow the persisted RDDs to be accessible after streaming completes.</p>
 
 <p>See <a href="src/test/scala/org/apache/bahir/cloudant/CloudantChangesDFSuite.scala">CloudantChangesDFSuite</a>
-for examples of loading data into a Spark DataFrame with <code class="highlighter-rouge">_changes</code> API.</p>
+for examples of loading data into a Spark DataFrame with <code>_changes</code> API.</p>
 
 <h3 id="configuration-on-spark-sql-temporary-table-or-dataframe">Configuration on Spark SQL Temporary Table or DataFrame</h3>
 
@@ -433,7 +446,7 @@ for examples of loading data into a Spark DataFrame with <code class="highlighte
     <tr>
       <td>selector</td>
       <td style="text-align: center">all documents</td>
-      <td>a selector written in Cloudant Query syntax, specifying conditions for selecting documents when the <code class="highlighter-rouge">cloudant.endpoint</code> option is set to <code class="highlighter-rouge">_changes</code>. Only documents satisfying the selector’s conditions will be retrieved from Cloudant and loaded into Spark.</td>
+      <td>a selector written in Cloudant Query syntax, specifying conditions for selecting documents when the <code>cloudant.endpoint</code> option is set to <code>_changes</code>. Only documents satisfying the selector’s conditions will be retrieved from Cloudant and loaded into Spark.</td>
     </tr>
     <tr>
       <td>view</td>
@@ -443,12 +456,12 @@ for examples of loading data into a Spark DataFrame with <code class="highlighte
   </tbody>
 </table>
 
-<p>For fast loading, views are loaded without include_docs. Thus, a derived schema will always be: <code class="highlighter-rouge"><span class="p">{</span><span class="err">id,</span><span class="w"> </span><span class="err">key,</span><span class="w"> </span><span class="err">value</span><span class="p">}</span></code>, where <code class="highlighter-rouge">value </code>can be a compount field. An example of loading data from a view:</p>
+<p>For fast loading, views are loaded without include_docs. Thus, a derived schema will always be: <code>{id, key, value}</code>, where <code>value </code>can be a compount field. An example of loading data from a view:</p>
 
-<div class="language-python highlighter-rouge"><pre class="highlight"><code><span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="p">(</span><span class="s">" CREATE TEMPORARY TABLE flightTable1 USING org.apache.bahir.cloudant OPTIONS ( database 'n_flight', view '_design/view/_view/AA0')"</span><span class="p">)</span>
+<p>```python
+spark.sql(“ CREATE TEMPORARY TABLE flightTable1 USING org.apache.bahir.cloudant OPTIONS ( database ‘n_flight’, view ‘_design/view/_view/AA0’)”)</p>
 
-</code></pre>
-</div>
+<p>```</p>
 
 <h3 id="configuration-on-cloudant-receiver-for-spark-streaming">Configuration on Cloudant Receiver for Spark Streaming</h3>
 
@@ -491,7 +504,7 @@ for examples of loading data into a Spark DataFrame with <code class="highlighte
 
 <h3 id="configuration-in-spark-submit-using---conf-option">Configuration in spark-submit using –conf option</h3>
 
-<p>The above stated configuration keys can also be set using <code class="highlighter-rouge">spark-submit --conf</code> option. When passing configuration in spark-submit, make sure adding “spark.” as prefix to the keys.</p>
+<p>The above stated configuration keys can also be set using <code>spark-submit --conf</code> option. When passing configuration in spark-submit, make sure adding “spark.” as prefix to the keys.</p>
 
 <h2 id="examples">Examples</h2>
 
@@ -499,62 +512,61 @@ for examples of loading data into a Spark DataFrame with <code class="highlighte
 
 <h4 id="using-sql-in-python">Using SQL In Python</h4>
 
-<div class="language-python highlighter-rouge"><pre class="highlight"><code><span class="n">spark</span> <span class="o">=</span> <span class="n">SparkSession</span>\
-    <span class="o">.</span><span class="n">builder</span>\
-    <span class="o">.</span><span class="n">appName</span><span class="p">(</span><span class="s">"Cloudant Spark SQL Example in Python using temp tables"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.host"</span><span class="p">,</span><span class="s">"ACCOUNT.cloudant.com"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.username"</span><span class="p">,</span> <span class="s">"USERNAME"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.password"</span><span class="p">,</span><span class="s">"PASSWORD"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">getOrCreate</span><span class="p">()</span>
-
-
-<span class="c"># Loading temp table from Cloudant db</span>
-<span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="p">(</span><span class="s">" CREATE TEMPORARY TABLE airportTable USING org.apache.bahir.cloudant OPTIONS ( database 'n_airportcodemapping')"</span><span class="p">)</span>
-<span class="n">airportData</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="p">(</span><span class="s">"SELECT _id, airportName FROM airportTable WHERE _id &gt;= 'CAA' AND _id &lt;= 'GAA' ORDER BY _id"</span><span class="p">)</span>
-<span class="n">airportData</span><span class="o">.</span><span class="n">printSchema</span><span class="p">()</span>
-<span class="k">print</span> <span class="s">'Total # of rows in airportData: '</span> <span class="o">+</span> <span class="nb">str</span><span class="p">(</span><span class="n">airportData</span><span class="o">.</span><span class="n">count</span><span class="p">())</span>
-<span class="k">for</span> <span class="n">code</span> <span class="ow">in</span> <span class="n">airportData</span><span class="o">.</span><span class="n">collect</span><span class="p">():</span>
-    <span class="k">print</span> <span class="n">code</span><span class="o">.</span><span class="n">_id</span>
-</code></pre>
-</div>
+<p>```python
+spark = SparkSession\
+    .builder\
+    .appName(“Cloudant Spark SQL Example in Python using temp tables”)\
+    .config(“cloudant.host”,”ACCOUNT.cloudant.com”)\
+    .config(“cloudant.username”, “USERNAME”)\
+    .config(“cloudant.password”,”PASSWORD”)\
+    .getOrCreate()</p>
+
+<h1 id="loading-temp-table-from-cloudant-db">Loading temp table from Cloudant db</h1>
+<p>spark.sql(“ CREATE TEMPORARY TABLE airportTable USING org.apache.bahir.cloudant OPTIONS ( database ‘n_airportcodemapping’)”)
+airportData = spark.sql(“SELECT _id, airportName FROM airportTable WHERE _id &gt;= ‘CAA’ AND _id &lt;= ‘GAA’ ORDER BY _id”)
+airportData.printSchema()
+print ‘Total # of rows in airportData: ‘ + str(airportData.count())
+for code in airportData.collect():
+    print code._id
+```</p>
 
 <p>See <a href="examples/python/CloudantApp.py">CloudantApp.py</a> for examples.</p>
 
 <p>Submit job example:
-<code class="highlighter-rouge">
+<code>
 spark-submit  --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.2 --conf spark.cloudant.host=ACCOUNT.cloudant.com --conf spark.cloudant.username=USERNAME --conf spark.cloudant.password=PASSWORD sql-cloudant/examples/python/CloudantApp.py
 </code></p>
 
 <h4 id="using-dataframe-in-python">Using DataFrame In Python</h4>
 
-<div class="language-python highlighter-rouge"><pre class="highlight"><code><span class="n">spark</span> <span class="o">=</span> <span class="n">SparkSession</span>\
-    <span class="o">.</span><span class="n">builder</span>\
-    <span class="o">.</span><span class="n">appName</span><span class="p">(</span><span class="s">"Cloudant Spark SQL Example in Python using dataframes"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.host"</span><span class="p">,</span><span class="s">"ACCOUNT.cloudant.com"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.username"</span><span class="p">,</span> <span class="s">"USERNAME"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"cloudant.password"</span><span class="p">,</span><span class="s">"PASSWORD"</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">config</span><span class="p">(</span><span class="s">"jsonstore.rdd.partitions"</span><span class="p">,</span> <span class="mi">8</span><span class="p">)</span>\
-    <span class="o">.</span><span class="n">getOrCreate</span><span class="p">()</span>
-
-<span class="c"># ***1. Loading dataframe from Cloudant db</span>
-<span class="n">df</span> <span class="o">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">read</span><span class="o">.</span><span class="n">load</span><span class="p">(</span><span class="s">"n_airportcodemapping"</span><span class="p">,</span> <span class="s">"org.apache.bahir.cloudant"</span><span class="p">)</span>
-<span class="n">df</span><span class="o">.</span><span class="n">cache</span><span class="p">()</span>
-<span class="n">df</span><span class="o">.</span><span class="n">printSchema</span><span class="p">()</span>
-<span class="n">df</span><span class="o">.</span><span class="nb">filter</span><span class="p">(</span><span class="n">df</span><span class="o">.</span><span class="n">airportName</span> <span class="o">&gt;=</span> <span class="s">'Moscow'</span><span class="p">)</span><span class="o">.</span><span class="n">select</span><span class="p">(</span><span class="s">"_id"</span><span class="p">,</span><span class="s">'airportName'</span><span class="p">)</span><span class="o">.</span><span cl [...]
-<span class="n">df</span><span class="o">.</span><span class="nb">filter</span><span class="p">(</span><span class="n">df</span><span class="o">.</span><span class="n">_id</span> <span class="o">&gt;=</span> <span class="s">'CAA'</span><span class="p">)</span><span class="o">.</span><span class="n">select</span><span class="p">(</span><span class="s">"_id"</span><span class="p">,</span><span class="s">'airportName'</span><span class="p">)</span><span class="o">.</span><span class="n">sho [...]
-</code></pre>
-</div>
+<p>```python
+spark = SparkSession\
+    .builder\
+    .appName(“Cloudant Spark SQL Example in Python using dataframes”)\
+    .config(“cloudant.host”,”ACCOUNT.cloudant.com”)\
+    .config(“cloudant.username”, “USERNAME”)\
+    .config(“cloudant.password”,”PASSWORD”)\
+    .config(“jsonstore.rdd.partitions”, 8)\
+    .getOrCreate()</p>
+
+<h1 id="loading-dataframe-from-cloudant-db">***1. Loading dataframe from Cloudant db</h1>
+<p>df = spark.read.load(“n_airportcodemapping”, “org.apache.bahir.cloudant”)
+df.cache()
+df.printSchema()
+df.filter(df.airportName &gt;= ‘Moscow’).select(“_id”,’airportName’).show()
+df.filter(df._id &gt;= ‘CAA’).select(“_id”,’airportName’).show()	  <br />
+```</p>
 
 <p>See <a href="examples/python/CloudantDF.py">CloudantDF.py</a> for examples.</p>
 
 <p>In case of doing multiple operations on a dataframe (select, filter etc.),
 you should persist a dataframe. Otherwise, every operation on a dataframe will load the same data from Cloudant again.
-Persisting will also speed up computation. This statement will persist an RDD in memory: <code class="highlighter-rouge">df.cache()</code>.  Alternatively for large dbs to persist in memory &amp; disk, use:</p>
+Persisting will also speed up computation. This statement will persist an RDD in memory: <code>df.cache()</code>.  Alternatively for large dbs to persist in memory &amp; disk, use:</p>
 
-<div class="language-python highlighter-rouge"><pre class="highlight"><code><span class="kn">from</span> <span class="nn">pyspark</span> <span class="kn">import</span> <span class="n">StorageLevel</span>
-<span class="n">df</span><span class="o">.</span><span class="n">persist</span><span class="p">(</span><span class="n">storageLevel</span> <span class="o">=</span> <span class="n">StorageLevel</span><span class="p">(</span><span class="bp">True</span><span class="p">,</span> <span class="bp">True</span><span class="p">,</span> <span class="bp">False</span><span class="p">,</span> <span class="bp">True</span><span class="p">,</span> <span class="mi">1</span><span class="p">))</span>
-</code></pre>
-</div>
+<p><code>python
+from pyspark import StorageLevel
+df.persist(storageLevel = StorageLevel(True, True, False, True, 1))
+</code></p>
 
 <p><a href="examples/python/CloudantDFOption.py">Sample code</a> on using DataFrame option to define cloudant configuration</p>
 
@@ -562,65 +574,65 @@ Persisting will also speed up computation. This statement will persist an RDD in
 
 <h4 id="using-sql-in-scala">Using SQL In Scala</h4>
 
-<div class="language-scala highlighter-rouge"><pre class="highlight"><code><span class="k">val</span> <span class="n">spark</span> <span class="k">=</span> <span class="nc">SparkSession</span>
-      <span class="o">.</span><span class="n">builder</span><span class="o">()</span>
-      <span class="o">.</span><span class="n">appName</span><span class="o">(</span><span class="s">"Cloudant Spark SQL Example"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.host"</span><span class="o">,</span><span class="s">"ACCOUNT.cloudant.com"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.username"</span><span class="o">,</span> <span class="s">"USERNAME"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.password"</span><span class="o">,</span><span class="s">"PASSWORD"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">getOrCreate</span><span class="o">()</span>
-
-<span class="c1">// For implicit conversions of Dataframe to RDDs
-</span><span class="k">import</span> <span class="nn">spark.implicits._</span>
-
-<span class="c1">// create a temp table from Cloudant db and query it using sql syntax
-</span><span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="o">(</span>
-    <span class="n">s</span><span class="s">"""
+<p>```scala
+val spark = SparkSession
+      .builder()
+      .appName(“Cloudant Spark SQL Example”)
+      .config(“cloudant.host”,”ACCOUNT.cloudant.com”)
+      .config(“cloudant.username”, “USERNAME”)
+      .config(“cloudant.password”,”PASSWORD”)
+      .getOrCreate()</p>
+
+<p>// For implicit conversions of Dataframe to RDDs
+import spark.implicits._</p>
+
+<p>// create a temp table from Cloudant db and query it using sql syntax
+spark.sql(
+    s”””
     |CREATE TEMPORARY TABLE airportTable
     |USING org.apache.bahir.cloudant
-    |OPTIONS ( database 'n_airportcodemapping')
-    """</span><span class="o">.</span><span class="n">stripMargin</span><span class="o">)</span>
-<span class="c1">// create a dataframe
-</span><span class="k">val</span> <span class="n">airportData</span> <span class="k">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">sql</span><span class="o">(</span><span class="s">"SELECT _id, airportName FROM airportTable WHERE _id &gt;= 'CAA' AND _id &lt;= 'GAA' ORDER BY _id"</span><span class="o">)</span>
-<span class="n">airportData</span><span class="o">.</span><span class="n">printSchema</span><span class="o">()</span>
-<span class="n">println</span><span class="o">(</span><span class="n">s</span><span class="s">"Total # of rows in airportData: "</span> <span class="o">+</span> <span class="n">airportData</span><span class="o">.</span><span class="n">count</span><span class="o">())</span>
-<span class="c1">// convert dataframe to array of Rows, and process each row
-</span><span class="n">airportData</span><span class="o">.</span><span class="n">map</span><span class="o">(</span><span class="n">t</span> <span class="k">=&gt;</span> <span class="s">"code: "</span> <span class="o">+</span> <span class="n">t</span><span class="o">(</span><span class="mi">0</span><span class="o">)</span> <span class="o">+</span> <span class="s">",name:"</span> <span class="o">+</span> <span class="n">t</span><span class="o">(</span><span class="mi">1</span><span class=" [...]
-</code></pre>
-</div>
-<p>See <a href="examples/scala/src/main/scala/mytest/spark/CloudantApp.scala">CloudantApp.scala</a> for examples.</p>
+    |OPTIONS ( database ‘n_airportcodemapping’)
+    “”“.stripMargin)
+// create a dataframe
+val airportData = spark.sql(“SELECT _id, airportName FROM airportTable WHERE _id &gt;= ‘CAA’ AND _id &lt;= ‘GAA’ ORDER BY _id”)
+airportData.printSchema()
+println(s”Total # of rows in airportData: “ + airportData.count())
+// convert dataframe to array of Rows, and process each row
+airportData.map(t =&gt; “code: “ + t(0) + “,name:” + t(1)).collect().foreach(println)
+```
+See <a href="examples/scala/src/main/scala/mytest/spark/CloudantApp.scala">CloudantApp.scala</a> for examples.</p>
 
 <p>Submit job example:
-<code class="highlighter-rouge">
+<code>
 spark-submit --class org.apache.spark.examples.sql.cloudant.CloudantApp --packages org.apache.bahir:spark-sql-cloudant_2.11:2.1.2 --conf spark.cloudant.host=ACCOUNT.cloudant.com --conf spark.cloudant.username=USERNAME --conf spark.cloudant.password=PASSWORD  /path/to/spark-sql-cloudant_2.11-2.1.2-tests.jar
 </code></p>
 
 <h3 id="using-dataframe-in-scala">Using DataFrame In Scala</h3>
 
-<div class="language-scala highlighter-rouge"><pre class="highlight"><code><span class="k">val</span> <span class="n">spark</span> <span class="k">=</span> <span class="nc">SparkSession</span>
-      <span class="o">.</span><span class="n">builder</span><span class="o">()</span>
-      <span class="o">.</span><span class="n">appName</span><span class="o">(</span><span class="s">"Cloudant Spark SQL Example with Dataframe"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.host"</span><span class="o">,</span><span class="s">"ACCOUNT.cloudant.com"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.username"</span><span class="o">,</span> <span class="s">"USERNAME"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"cloudant.password"</span><span class="o">,</span><span class="s">"PASSWORD"</span><span class="o">)</span>
-      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"createDBOnSave"</span><span class="o">,</span><span class="s">"true"</span><span class="o">)</span> <span class="c1">// to create a db on save
-</span>      <span class="o">.</span><span class="n">config</span><span class="o">(</span><span class="s">"jsonstore.rdd.partitions"</span><span class="o">,</span> <span class="s">"20"</span><span class="o">)</span> <span class="c1">// using 20 partitions
-</span>      <span class="o">.</span><span class="n">getOrCreate</span><span class="o">()</span>
-
-<span class="c1">// 1. Loading data from Cloudant db
-</span><span class="k">val</span> <span class="n">df</span> <span class="k">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">read</span><span class="o">.</span><span class="n">format</span><span class="o">(</span><span class="s">"org.apache.bahir.cloudant"</span><span class="o">).</span><span class="n">load</span><span class="o">(</span><span class="s">"n_flight"</span><span class="o">)</span>
-<span class="c1">// Caching df in memory to speed computations
+<p>```scala
+val spark = SparkSession
+      .builder()
+      .appName(“Cloudant Spark SQL Example with Dataframe”)
+      .config(“cloudant.host”,”ACCOUNT.cloudant.com”)
+      .config(“cloudant.username”, “USERNAME”)
+      .config(“cloudant.password”,”PASSWORD”)
+      .config(“createDBOnSave”,”true”) // to create a db on save
+      .config(“jsonstore.rdd.partitions”, “20”) // using 20 partitions
+      .getOrCreate()</p>
+
+<p>// 1. Loading data from Cloudant db
+val df = spark.read.format(“org.apache.bahir.cloudant”).load(“n_flight”)
+// Caching df in memory to speed computations
 // and not to retrieve data from cloudant again
-</span><span class="n">df</span><span class="o">.</span><span class="n">cache</span><span class="o">()</span>
-<span class="n">df</span><span class="o">.</span><span class="n">printSchema</span><span class="o">()</span>
-
-<span class="c1">// 2. Saving dataframe to Cloudant db
-</span><span class="k">val</span> <span class="n">df2</span> <span class="k">=</span> <span class="n">df</span><span class="o">.</span><span class="n">filter</span><span class="o">(</span><span class="n">df</span><span class="o">(</span><span class="s">"flightSegmentId"</span><span class="o">)</span> <span class="o">===</span> <span class="s">"AA106"</span><span class="o">)</span>
-    <span class="o">.</span><span class="n">select</span><span class="o">(</span><span class="s">"flightSegmentId"</span><span class="o">,</span><span class="s">"economyClassBaseCost"</span><span class="o">)</span>
-<span class="n">df2</span><span class="o">.</span><span class="n">show</span><span class="o">()</span>
-<span class="n">df2</span><span class="o">.</span><span class="n">write</span><span class="o">.</span><span class="n">format</span><span class="o">(</span><span class="s">"org.apache.bahir.cloudant"</span><span class="o">).</span><span class="n">save</span><span class="o">(</span><span class="s">"n_flight2"</span><span class="o">)</span>
-</code></pre>
-</div>
+df.cache()
+df.printSchema()</p>
+
+<p>// 2. Saving dataframe to Cloudant db
+val df2 = df.filter(df(“flightSegmentId”) === “AA106”)
+    .select(“flightSegmentId”,”economyClassBaseCost”)
+df2.show()
+df2.write.format(“org.apache.bahir.cloudant”).save(“n_flight2”)
+```</p>
 
 <p>See <a href="examples/scala/src/main/scala/mytest/spark/CloudantDF.scala">CloudantDF.scala</a> for examples.</p>
 
@@ -628,49 +640,49 @@ spark-submit --class org.apache.spark.examples.sql.cloudant.CloudantApp --packag
 
 <h3 id="using-streams-in-scala">Using Streams In Scala</h3>
 
-<div class="language-scala highlighter-rouge"><pre class="highlight"><code><span class="k">val</span> <span class="n">ssc</span> <span class="k">=</span> <span class="k">new</span> <span class="nc">StreamingContext</span><span class="o">(</span><span class="n">sparkConf</span><span class="o">,</span> <span class="nc">Seconds</span><span class="o">(</span><span class="mi">10</span><span class="o">))</span>
-<span class="k">val</span> <span class="n">changes</span> <span class="k">=</span> <span class="n">ssc</span><span class="o">.</span><span class="n">receiverStream</span><span class="o">(</span><span class="k">new</span> <span class="nc">CloudantReceiver</span><span class="o">(</span><span class="nc">Map</span><span class="o">(</span>
-  <span class="s">"cloudant.host"</span> <span class="o">-&gt;</span> <span class="s">"ACCOUNT.cloudant.com"</span><span class="o">,</span>
-  <span class="s">"cloudant.username"</span> <span class="o">-&gt;</span> <span class="s">"USERNAME"</span><span class="o">,</span>
-  <span class="s">"cloudant.password"</span> <span class="o">-&gt;</span> <span class="s">"PASSWORD"</span><span class="o">,</span>
-  <span class="s">"database"</span> <span class="o">-&gt;</span> <span class="s">"n_airportcodemapping"</span><span class="o">)))</span>
-
-<span class="n">changes</span><span class="o">.</span><span class="n">foreachRDD</span><span class="o">((</span><span class="n">rdd</span><span class="k">:</span> <span class="kt">RDD</span><span class="o">[</span><span class="kt">String</span><span class="o">],</span> <span class="n">time</span><span class="k">:</span> <span class="kt">Time</span><span class="o">)</span> <span class="k">=&gt;</span> <span class="o">{</span>
-  <span class="c1">// Get the singleton instance of SparkSession
-</span>  <span class="k">val</span> <span class="n">spark</span> <span class="k">=</span> <span class="nc">SparkSessionSingleton</span><span class="o">.</span><span class="n">getInstance</span><span class="o">(</span><span class="n">rdd</span><span class="o">.</span><span class="n">sparkContext</span><span class="o">.</span><span class="n">getConf</span><span class="o">)</span>
-
-  <span class="n">println</span><span class="o">(</span><span class="n">s</span><span class="s">"========= $time ========="</span><span class="o">)</span>
-  <span class="c1">// Convert RDD[String] to DataFrame
-</span>  <span class="k">val</span> <span class="n">changesDataFrame</span> <span class="k">=</span> <span class="n">spark</span><span class="o">.</span><span class="n">read</span><span class="o">.</span><span class="n">json</span><span class="o">(</span><span class="n">rdd</span><span class="o">)</span>
-  <span class="k">if</span> <span class="o">(!</span><span class="n">changesDataFrame</span><span class="o">.</span><span class="n">schema</span><span class="o">.</span><span class="n">isEmpty</span><span class="o">)</span> <span class="o">{</span>
-    <span class="n">changesDataFrame</span><span class="o">.</span><span class="n">printSchema</span><span class="o">()</span>
-    <span class="n">changesDataFrame</span><span class="o">.</span><span class="n">select</span><span class="o">(</span><span class="s">"*"</span><span class="o">).</span><span class="n">show</span><span class="o">()</span>
-    <span class="o">....</span>
-  <span class="o">}</span>
-<span class="o">})</span>
-<span class="n">ssc</span><span class="o">.</span><span class="n">start</span><span class="o">()</span>
-<span class="c1">// run streaming for 120 secs
-</span><span class="nc">Thread</span><span class="o">.</span><span class="n">sleep</span><span class="o">(</span><span class="mi">120000L</span><span class="o">)</span>
-<span class="n">ssc</span><span class="o">.</span><span class="n">stop</span><span class="o">(</span><span class="kc">true</span><span class="o">)</span>
-
-</code></pre>
-</div>
+<p>```scala
+val ssc = new StreamingContext(sparkConf, Seconds(10))
+val changes = ssc.receiverStream(new CloudantReceiver(Map(
+  “cloudant.host” -&gt; “ACCOUNT.cloudant.com”,
+  “cloudant.username” -&gt; “USERNAME”,
+  “cloudant.password” -&gt; “PASSWORD”,
+  “database” -&gt; “n_airportcodemapping”)))</p>
+
+<p>changes.foreachRDD((rdd: RDD[String], time: Time) =&gt; {
+  // Get the singleton instance of SparkSession
+  val spark = SparkSessionSingleton.getInstance(rdd.sparkContext.getConf)</p>
+
+<p>println(s”========= $time =========”)
+  // Convert RDD[String] to DataFrame
+  val changesDataFrame = spark.read.json(rdd)
+  if (!changesDataFrame.schema.isEmpty) {
+    changesDataFrame.printSchema()
+    changesDataFrame.select(“*”).show()
+    ….
+  }
+})
+ssc.start()
+// run streaming for 120 secs
+Thread.sleep(120000L)
+ssc.stop(true)</p>
+
+<p>```</p>
 
 <p>See <a href="examples/scala/src/main/scala/mytest/spark/CloudantStreaming.scala">CloudantStreaming.scala</a> for examples.</p>
 
 <p>By default, Spark Streaming will load all documents from a database. If you want to limit the loading to
-specific documents, use <code class="highlighter-rouge">selector</code> option of <code class="highlighter-rouge">CloudantReceiver</code> and specify your conditions
+specific documents, use <code>selector</code> option of <code>CloudantReceiver</code> and specify your conditions
 (See <a href="examples/scala/src/main/scala/mytest/spark/CloudantStreamingSelector.scala">CloudantStreamingSelector.scala</a>
 example for more details):</p>
 
-<div class="language-scala highlighter-rouge"><pre class="highlight"><code><span class="k">val</span> <span class="n">changes</span> <span class="k">=</span> <span class="n">ssc</span><span class="o">.</span><span class="n">receiverStream</span><span class="o">(</span><span class="k">new</span> <span class="nc">CloudantReceiver</span><span class="o">(</span><span class="nc">Map</span><span class="o">(</span>
-  <span class="s">"cloudant.host"</span> <span class="o">-&gt;</span> <span class="s">"ACCOUNT.cloudant.com"</span><span class="o">,</span>
-  <span class="s">"cloudant.username"</span> <span class="o">-&gt;</span> <span class="s">"USERNAME"</span><span class="o">,</span>
-  <span class="s">"cloudant.password"</span> <span class="o">-&gt;</span> <span class="s">"PASSWORD"</span><span class="o">,</span>
-  <span class="s">"database"</span> <span class="o">-&gt;</span> <span class="s">"sales"</span><span class="o">,</span>
-  <span class="s">"selector"</span> <span class="o">-&gt;</span> <span class="s">"{\"month\":\"May\", \"rep\":\"John\"}"</span><span class="o">)))</span>
-</code></pre>
-</div>
+<p><code>scala
+val changes = ssc.receiverStream(new CloudantReceiver(Map(
+  "cloudant.host" -&gt; "ACCOUNT.cloudant.com",
+  "cloudant.username" -&gt; "USERNAME",
+  "cloudant.password" -&gt; "PASSWORD",
+  "database" -&gt; "sales",
+  "selector" -&gt; "{\"month\":\"May\", \"rep\":\"John\"}")))
+</code></p>
 
   </div>
 </div>
@@ -679,7 +691,7 @@ example for more details):</p>
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.2/spark-sql-streaming-akka/index.html b/content/docs/spark/2.1.2/spark-sql-streaming-akka/index.html
index 7023959..7139e4b 100644
--- a/content/docs/spark/2.1.2/spark-sql-streaming-akka/index.html
+++ b/content/docs/spark/2.1.2/spark-sql-streaming-akka/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="http://www.apache.org/foundation/how-it-works.html" target="_blank">Apache Software Foundation</a></li>
               
               
+              
               <li><a href="http://www.apache.org/licenses/" target="_blank">Apache License</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/sponsorship" target="_blank">Sponsorship</a></li>
               
               
+              
               <li><a href="http://www.apache.org/foundation/thanks.html" target="_blank">Thanks</a></li>
               
               
+              
               <li><a href="/privacy-policy" target="_self">Privacy Policy</a></li>
               
             </ul>
@@ -201,29 +219,26 @@
 
 <p>Using SBT:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-akka" % "2.1.2"
+<pre><code>libraryDependencies += "org.apache.bahir" %% "spark-sql-streaming-akka" % "2.1.2"
 </code></pre>
-</div>
 
 <p>Using Maven:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>&lt;dependency&gt;
+<pre><code>&lt;dependency&gt;
     &lt;groupId&gt;org.apache.bahir&lt;/groupId&gt;
     &lt;artifactId&gt;spark-sql-streaming-akka_2.11&lt;/artifactId&gt;
     &lt;version&gt;2.1.2&lt;/version&gt;
 &lt;/dependency&gt;
 </code></pre>
-</div>
 
-<p>This library can also be added to Spark jobs launched through <code class="highlighter-rouge">spark-shell</code> or <code class="highlighter-rouge">spark-submit</code> by using the <code class="highlighter-rouge">--packages</code> command line option.
+<p>This library can also be added to Spark jobs launched through <code>spark-shell</code> or <code>spark-submit</code> by using the <code>--packages</code> command line option.
 For example, to include it when starting the spark shell:</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-akka_2.11:2.1.2
+<pre><code>$ bin/spark-shell --packages org.apache.bahir:spark-sql-streaming-akka_2.11:2.1.2
 </code></pre>
-</div>
 
-<p>Unlike using <code class="highlighter-rouge">--jars</code>, using <code class="highlighter-rouge">--packages</code> ensures that this library and its dependencies will be added to the classpath.
-The <code class="highlighter-rouge">--packages</code> argument can also be used with <code class="highlighter-rouge">bin/spark-submit</code>.</p>
+<p>Unlike using <code>--jars</code>, using <code>--packages</code> ensures that this library and its dependencies will be added to the classpath.
+The <code>--packages</code> argument can also be used with <code>bin/spark-submit</code>.</p>
 
 <p>This library is compiled for Scala 2.11 only, and intends to support Spark 2.0 onwards.</p>
 
@@ -231,39 +246,37 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
 <p>A SQL Stream can be created with data streams received from Akka Feeder actor using,</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>    sqlContext.readStream
+<pre><code>    sqlContext.readStream
             .format("org.apache.bahir.sql.streaming.akka.AkkaStreamSourceProvider")
             .option("urlOfPublisher", "feederActorUri")
             .load()
 </code></pre>
-</div>
 
 <h2 id="enable-recovering-from-failures">Enable recovering from failures.</h2>
 
-<p>Setting values for option <code class="highlighter-rouge">persistenceDirPath</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
+<p>Setting values for option <code>persistenceDirPath</code> helps in recovering in case of a restart, by restoring the state where it left off before the shutdown.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>    sqlContext.readStream
+<pre><code>    sqlContext.readStream
             .format("org.apache.bahir.sql.streaming.akka.AkkaStreamSourceProvider")
             .option("urlOfPublisher", "feederActorUri")
             .option("persistenceDirPath", "/path/to/localdir")
             .load() 
 </code></pre>
-</div>
 
 <h2 id="configuration-options">Configuration options.</h2>
 
 <p>This source uses <a href="http://doc.akka.io/api/akka/2.4/akka/actor/Actor.html">Akka Actor api</a>.</p>
 
 <ul>
-  <li><code class="highlighter-rouge">urlOfPublisher</code> The url of Publisher or Feeder actor that the Receiver actor connects to. Set this as the tcp url of the Publisher or Feeder actor.</li>
-  <li><code class="highlighter-rouge">persistenceDirPath</code> By default it is used for storing incoming messages on disk.</li>
+  <li><code>urlOfPublisher</code> The url of Publisher or Feeder actor that the Receiver actor connects to. Set this as the tcp url of the Publisher or Feeder actor.</li>
+  <li><code>persistenceDirPath</code> By default it is used for storing incoming messages on disk.</li>
 </ul>
 
 <h3 id="scala-api">Scala API</h3>
 
 <p>An example, for scala API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>    // Create DataFrame representing the stream of input lines from connection
+<pre><code>    // Create DataFrame representing the stream of input lines from connection
     // to publisher or feeder actor
     val lines = spark.readStream
                 .format("org.apache.bahir.sql.streaming.akka.AkkaStreamSourceProvider")
@@ -284,15 +297,14 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
     query.awaitTermination()
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">AkkaStreamWordCount.scala</code> for full example.</p>
+<p>Please see <code>AkkaStreamWordCount.scala</code> for full example.</p>
 
 <h3 id="java-api">Java API</h3>
 
 <p>An example, for Java API to count words from incoming message stream.</p>
 
-<div class="highlighter-rouge"><pre class="highlight"><code>    // Create DataFrame representing the stream of input lines from connection
+<pre><code>    // Create DataFrame representing the stream of input lines from connection
     // to publisher or feeder actor
     Dataset&lt;String&gt; lines = spark
                             .readStream()
@@ -319,9 +331,8 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
     query.awaitTermination();   
 </code></pre>
-</div>
 
-<p>Please see <code class="highlighter-rouge">JavaAkkaStreamWordCount.java</code> for full example.</p>
+<p>Please see <code>JavaAkkaStreamWordCount.java</code> for full example.</p>
 
   </div>
 </div>
@@ -330,7 +341,7 @@ The <code class="highlighter-rouge">--packages</code> argument can also be used
 
       <hr>
 
-      <!-- <p>&copy; 2019 </p>-->
+      <!-- <p>&copy; 2020 </p>-->
       <footer class="site-footer">
     <div class="wrapper">
         <div class="footer-col-wrapper">
diff --git a/content/docs/spark/2.1.2/spark-sql-streaming-mqtt/index.html b/content/docs/spark/2.1.2/spark-sql-streaming-mqtt/index.html
index 190b0de..d71815a 100644
--- a/content/docs/spark/2.1.2/spark-sql-streaming-mqtt/index.html
+++ b/content/docs/spark/2.1.2/spark-sql-streaming-mqtt/index.html
@@ -65,9 +65,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/downloads/spark" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/downloads/flink" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -83,21 +85,27 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/community" target="_self">Get Involved</a></li>
               
               
+              
               <li><a href="/contributing" target="_self">Contributing</a></li>
               
               
+              
               <li><a href="/contributing-extensions" target="_self">Contributing Extensions</a></li>
               
               
+              
               <li><a href="https://issues.apache.org/jira/browse/BAHIR" target="_blank">Issue Tracker</a></li>
               
               
+              
               <li><a href="/community#source-code" target="_self">Source Code</a></li>
               
               
+              
               <li><a href="/community-members" target="_self">Project Committers</a></li>
               
             </ul>
@@ -113,9 +121,11 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="/docs/spark/overview" target="_self">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="/docs/flink/overview" target="_self">Bahir Flink Extensions</a></li>
               
             </ul>
@@ -131,12 +141,15 @@
             <ul class="dropdown-menu dropdown-left">
               
               
+              
               <li><a href="https://github.com/apache/bahir" target="_blank">Bahir Spark Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-flink" target="_blank">Bahir Flink Extensions</a></li>
               
               
+              
               <li><a href="https://github.com/apache/bahir-website" target="_blank">Bahir Website</a></li>
               
             </ul>
@@ -152,18 +165,23 @@
             <ul class="dropdown-menu dropdown-left">
               
               
... 20397 lines suppressed ...