You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@bigtop.apache.org by of...@apache.org on 2015/01/29 21:53:06 UTC

[1/3] bigtop git commit: BIGTOP-1636. Missing patches and files for hue

Repository: bigtop
Updated Branches:
  refs/heads/master 1357959c0 -> 59fa03c0f


BIGTOP-1636. Missing patches and files for hue


Project: http://git-wip-us.apache.org/repos/asf/bigtop/repo
Commit: http://git-wip-us.apache.org/repos/asf/bigtop/commit/e392f6a1
Tree: http://git-wip-us.apache.org/repos/asf/bigtop/tree/e392f6a1
Diff: http://git-wip-us.apache.org/repos/asf/bigtop/diff/e392f6a1

Branch: refs/heads/master
Commit: e392f6a1652652dfe1d4149f145b8b94e1d6fb85
Parents: 1357959
Author: Olaf Flebbe <of...@oflebbe.de>
Authored: Thu Jan 29 21:15:53 2015 +0100
Committer: Olaf Flebbe <of...@oflebbe.de>
Committed: Thu Jan 29 21:50:27 2015 +0100

----------------------------------------------------------------------
 .../src/common/hue/patch0-getlog.diff           | 264 +++++++++++++++++++
 .../src/common/hue/patch1-bigtop-conf.diff      |  24 ++
 bigtop-packages/src/deb/hue/hue-impala.install  |   1 +
 .../src/deb/hue/hue-security.install            |   1 +
 4 files changed, 290 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/bigtop/blob/e392f6a1/bigtop-packages/src/common/hue/patch0-getlog.diff
----------------------------------------------------------------------
diff --git a/bigtop-packages/src/common/hue/patch0-getlog.diff b/bigtop-packages/src/common/hue/patch0-getlog.diff
new file mode 100644
index 0000000..957c6ab
--- /dev/null
+++ b/bigtop-packages/src/common/hue/patch0-getlog.diff
@@ -0,0 +1,264 @@
+commit 6a0246710f7deeb0fd2e1f2b3b209ad119c30b72
+Author: Peter Slawski <pe...@amazon.com>
+Date:   Fri Oct 10 21:11:28 2014 -0700
+
+    HUE-2387 [beeswax] Support getting logs using FetchResults() api
+
+diff --git a/apps/beeswax/gen-py/TCLIService/ttypes.py b/apps/beeswax/gen-py/TCLIService/ttypes.py
+index b273bf0..d1e1a5e 100644
+--- a/apps/beeswax/gen-py/TCLIService/ttypes.py
++++ b/apps/beeswax/gen-py/TCLIService/ttypes.py
+@@ -5124,6 +5124,7 @@ class TFetchResultsReq(object):
+    - operationHandle
+    - orientation
+    - maxRows
++   - fetchType
+   """
+ 
+   thrift_spec = (
+@@ -5131,12 +5132,14 @@ class TFetchResultsReq(object):
+     (1, TType.STRUCT, 'operationHandle', (TOperationHandle, TOperationHandle.thrift_spec), None, ), # 1
+     (2, TType.I32, 'orientation', None,     0, ), # 2
+     (3, TType.I64, 'maxRows', None, None, ), # 3
++    (4, TType.I16, 'fetchType', None, 0, ), # 4
+   )
+ 
+-  def __init__(self, operationHandle=None, orientation=thrift_spec[2][4], maxRows=None,):
++  def __init__(self, operationHandle=None, orientation=thrift_spec[2][4], maxRows=None, fetchType=thrift_spec[4][4],):
+     self.operationHandle = operationHandle
+     self.orientation = orientation
+     self.maxRows = maxRows
++    self.fetchType = fetchType
+ 
+   def read(self, iprot):
+     if iprot.__class__ == TBinaryProtocol.TBinaryProtocolAccelerated and isinstance(iprot.trans, TTransport.CReadableTransport) and self.thrift_spec is not None and fastbinary is not None:
+@@ -5163,6 +5166,11 @@ class TFetchResultsReq(object):
+           self.maxRows = iprot.readI64();
+         else:
+           iprot.skip(ftype)
++      elif fid == 4:
++        if ftype == TType.I16:
++          self.fetchType = iprot.readI16();
++        else:
++          iprot.skip(ftype)
+       else:
+         iprot.skip(ftype)
+       iprot.readFieldEnd()
+@@ -5185,6 +5193,10 @@ class TFetchResultsReq(object):
+       oprot.writeFieldBegin('maxRows', TType.I64, 3)
+       oprot.writeI64(self.maxRows)
+       oprot.writeFieldEnd()
++    if self.fetchType is not None:
++      oprot.writeFieldBegin('fetchType', TType.I16, 4)
++      oprot.writeI16(self.fetchType)
++      oprot.writeFieldEnd()
+     oprot.writeFieldStop()
+     oprot.writeStructEnd()
+ 
+diff --git a/apps/beeswax/src/beeswax/api.py b/apps/beeswax/src/beeswax/api.py
+index 91da7dd..f56b44a 100644
+--- a/apps/beeswax/src/beeswax/api.py
++++ b/apps/beeswax/src/beeswax/api.py
+@@ -172,7 +172,8 @@ def watch_query_refresh_json(request, id):
+     handle, state = _get_query_handle_and_state(query_history)
+ 
+   try:
+-    log = db.get_log(handle)
++    start_over = request.POST.get('log-start-over') == 'true'
++    log = db.get_log(handle, start_over=start_over)
+   except Exception, ex:
+     log = str(ex)
+ 
+diff --git a/apps/beeswax/src/beeswax/conf.py b/apps/beeswax/src/beeswax/conf.py
+index 1667481..e7ffc8e 100644
+--- a/apps/beeswax/src/beeswax/conf.py
++++ b/apps/beeswax/src/beeswax/conf.py
+@@ -60,6 +60,14 @@ SERVER_CONN_TIMEOUT = Config(
+   type=int,
+   help=_t('Timeout in seconds for Thrift calls.'))
+ 
++USE_GET_LOG_API = Config(
++  key='use_get_log_api',
++  default=True,
++  type=coerce_bool,
++  help=_t('Choose whether Hue uses the GetLog() thrift call to retrieve Hive logs.'
++          'If false, Hue will use the FetchResults() thrift call instead.')
++)
++
+ BROWSE_PARTITIONED_TABLE_LIMIT = Config(
+   key='browse_partitioned_table_limit',
+   default=250,
+diff --git a/apps/beeswax/src/beeswax/server/dbms.py b/apps/beeswax/src/beeswax/server/dbms.py
+index baaba68..86b012b 100644
+--- a/apps/beeswax/src/beeswax/server/dbms.py
++++ b/apps/beeswax/src/beeswax/server/dbms.py
+@@ -338,8 +338,8 @@ class HiveServer2Dbms(object):
+     return self.client.use(query)
+ 
+ 
+-  def get_log(self, query_handle):
+-    return self.client.get_log(query_handle)
++  def get_log(self, query_handle, start_over=True):
++    return self.client.get_log(query_handle, start_over)
+ 
+ 
+   def get_state(self, handle):
+diff --git a/apps/beeswax/src/beeswax/server/hive_server2_lib.py b/apps/beeswax/src/beeswax/server/hive_server2_lib.py
+index d7c222c..f7f2562 100644
+--- a/apps/beeswax/src/beeswax/server/hive_server2_lib.py
++++ b/apps/beeswax/src/beeswax/server/hive_server2_lib.py
+@@ -14,6 +14,7 @@
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ # See the License for the specific language governing permissions and
+ # limitations under the License.
++from itertools import imap
+ 
+ import logging
+ import re
+@@ -590,6 +591,15 @@ class HiveServerClient:
+     return res, schema
+ 
+ 
++  def fetch_log(self, operation_handle, orientation=TFetchOrientation.FETCH_NEXT, max_rows=1000):
++    req = TFetchResultsReq(operationHandle=operation_handle, orientation=orientation,
++                           maxRows=max_rows, fetchType=1)
++    res = self.call(self._client.FetchResults, req)
++
++    lines = imap(lambda r: r.colVals[0].stringVal.value, res.results.rows)
++    return '\n'.join(lines)
++
++
+   def get_operation_status(self, operation_handle):
+     req = TGetOperationStatusReq(operationHandle=operation_handle)
+     return self.call(self._client.GetOperationStatus, req)
+@@ -759,9 +769,18 @@ class HiveServerClientCompatible(object):
+     return 'Does not exist in HS2'
+ 
+ 
+-  def get_log(self, handle):
++  def get_log(self, handle, start_over=True):
+     operationHandle = handle.get_rpc_handle()
+-    return self._client.get_log(operationHandle)
++
++    if beeswax_conf.USE_GET_LOG_API.get() or self.query_server['server_name'] == 'impala':
++      return self._client.get_log(operationHandle)
++    else:
++      if start_over:
++        orientation = TFetchOrientation.FETCH_FIRST
++      else:
++        orientation = TFetchOrientation.FETCH_NEXT
++
++      return self._client.fetch_log(operationHandle, orientation=orientation, max_rows=-1)
+ 
+ 
+   def get_databases(self):
+diff --git a/apps/beeswax/src/beeswax/templates/execute.mako b/apps/beeswax/src/beeswax/templates/execute.mako
+index 6b1f086..27909b0 100644
+--- a/apps/beeswax/src/beeswax/templates/execute.mako
++++ b/apps/beeswax/src/beeswax/templates/execute.mako
+@@ -2637,6 +2637,11 @@ function setupCodeMirrorSubscription() {
+ 
+ // Knockout
+ viewModel = new BeeswaxViewModel("${app_name}");
++
++% if not beeswax_conf.USE_GET_LOG_API.get() and app_name != 'impala':
++  viewModel.shouldAppendLogs = true;
++% endif
++
+ % if query_history:
+   loadQueryHistory(${query_history.id});
+ % elif design.id:
+diff --git a/apps/beeswax/static/js/beeswax.vm.js b/apps/beeswax/static/js/beeswax.vm.js
+index fdd2f78..6e6c65c 100644
+--- a/apps/beeswax/static/js/beeswax.vm.js
++++ b/apps/beeswax/static/js/beeswax.vm.js
+@@ -560,7 +560,8 @@ function BeeswaxViewModel(server) {
+   self.watchQuery = function() {
+     var data = {
+       'query-query': self.design.query.value(),
+-      'query-database': self.database()
++      'query-database': self.database(),
++      'log-start-over': self.design.watch.logs().length == 0
+     };
+     $.extend(data, self.getSettingsFormData());
+     $.extend(data, self.getFileResourcesFormData());
+@@ -589,6 +590,17 @@ function BeeswaxViewModel(server) {
+     $.ajax(request);
+   };
+ 
++  self.shouldAppendLogs = false;
++
++  self.applyLogs = function(log) {
++    var lines = log.split("\n")
++
++    if (self.shouldAppendLogs) {
++      lines = self.design.watch.logs().concat(lines);
++    }
++    self.design.watch.logs(lines);
++  };
++
+   self.watchQueryLoop = function(fn) {
+     var TIMEOUT = 100;
+     var timer = null;
+@@ -607,7 +619,7 @@ function BeeswaxViewModel(server) {
+           clearTimeout(timer);
+           self.design.isRunning(false);
+           if (data.log) {
+-            self.design.watch.logs(data.log.split("\n"));
++            self.applyLogs(data.log)
+             // scroll logs
+             self.design.watch.jobUrls(data.jobUrls);
+           }
+@@ -623,7 +635,7 @@ function BeeswaxViewModel(server) {
+         } else {
+           self.design.statement(data.statement); // In case new no result statement executed
+           if (data.log) {
+-            self.design.watch.logs(data.log.split("\n"));
++            self.applyLogs(data.log)
+             // scroll logs
+             self.design.watch.jobUrls(data.jobUrls);
+           }
+diff --git a/apps/beeswax/thrift/TCLIService.thrift b/apps/beeswax/thrift/TCLIService.thrift
+index 53ea3cf..63dad7e 100644
+--- a/apps/beeswax/thrift/TCLIService.thrift
++++ b/apps/beeswax/thrift/TCLIService.thrift
+@@ -1002,6 +1002,9 @@ struct TFetchResultsReq {
+   // Max number of rows that should be returned in
+   // the rowset.
+   3: required i64 maxRows
++
++  // The type of a fetch results request. 0 represents Query output. 1 represents Log
++  4: optional i16 fetchType = 0
+ }
+ 
+ struct TFetchResultsResp {
+diff --git a/desktop/conf.dist/hue.ini b/desktop/conf.dist/hue.ini
+index dc06a85..98c6bbb 100644
+--- a/desktop/conf.dist/hue.ini
++++ b/desktop/conf.dist/hue.ini
+@@ -724,6 +724,10 @@
+   # Timeout in seconds for thrift calls to Hive service
+   ## server_conn_timeout=120
+ 
++  # Choose whether Hue uses the GetLog() thrift call to retrieve Hive logs.
++  # If false, Hue will use the FetchResults() thrift call instead.
++  ## use_get_log_api=true
++
+   # Set a LIMIT clause when browsing a partitioned table.
+   # A positive value will be set as the LIMIT. If 0 or negative, do not set any limit.
+   ## browse_partitioned_table_limit=250
+diff --git a/desktop/conf/pseudo-distributed.ini.tmpl b/desktop/conf/pseudo-distributed.ini.tmpl
+index 784c235..52f719b 100644
+--- a/desktop/conf/pseudo-distributed.ini.tmpl
++++ b/desktop/conf/pseudo-distributed.ini.tmpl
+@@ -731,6 +731,10 @@
+   # Timeout in seconds for thrift calls to Hive service
+   ## server_conn_timeout=120
+ 
++  # Choose whether Hue uses the GetLog() thrift call to retrieve Hive logs.
++  # If false, Hue will use the FetchResults() thrift call instead.
++  ## use_get_log_api=true
++
+   # Set a LIMIT clause when browsing a partitioned table.
+   # A positive value will be set as the LIMIT. If 0 or negative, do not set any limit.
+   ## browse_partitioned_table_limit=250

http://git-wip-us.apache.org/repos/asf/bigtop/blob/e392f6a1/bigtop-packages/src/common/hue/patch1-bigtop-conf.diff
----------------------------------------------------------------------
diff --git a/bigtop-packages/src/common/hue/patch1-bigtop-conf.diff b/bigtop-packages/src/common/hue/patch1-bigtop-conf.diff
new file mode 100644
index 0000000..36e1532
--- /dev/null
+++ b/bigtop-packages/src/common/hue/patch1-bigtop-conf.diff
@@ -0,0 +1,24 @@
+diff --git a/desktop/conf.dist/hue.ini~ b/desktop/conf.dist/hue.ini
+index 9b30901..a2b9bea 100644
+--- a/desktop/conf.dist/hue.ini~
++++ b/desktop/conf.dist/hue.ini
+@@ -95,7 +95,8 @@
+ 
+   # Comma separated list of apps to not load at server startup.
+   # e.g.: pig,zookeeper
+-  ## app_blacklist=
++  # Bigtop does not bundle impala
++  app_blacklist=impala
+ 
+   # The directory where to store the auditing logs. Auditing is disable if the value is empty.
+   # e.g. /var/log/hue/audit.log
+@@ -722,7 +723,8 @@
+ 
+   # Choose whether Hue uses the GetLog() thrift call to retrieve Hive logs.
+   # If false, Hue will use the FetchResults() thrift call instead.
+-  ## use_get_log_api=true
++  # Bigtop Hive only supports FetchResults() 
++  use_get_log_api=false
+ 
+   # Set a LIMIT clause when browsing a partitioned table.
+   # A positive value will be set as the LIMIT. If 0 or negative, do not set any limit.

http://git-wip-us.apache.org/repos/asf/bigtop/blob/e392f6a1/bigtop-packages/src/deb/hue/hue-impala.install
----------------------------------------------------------------------
diff --git a/bigtop-packages/src/deb/hue/hue-impala.install b/bigtop-packages/src/deb/hue/hue-impala.install
new file mode 100644
index 0000000..1ff1a4e
--- /dev/null
+++ b/bigtop-packages/src/deb/hue/hue-impala.install
@@ -0,0 +1 @@
+/usr/lib/hue/apps/impala

http://git-wip-us.apache.org/repos/asf/bigtop/blob/e392f6a1/bigtop-packages/src/deb/hue/hue-security.install
----------------------------------------------------------------------
diff --git a/bigtop-packages/src/deb/hue/hue-security.install b/bigtop-packages/src/deb/hue/hue-security.install
new file mode 100644
index 0000000..86816f6
--- /dev/null
+++ b/bigtop-packages/src/deb/hue/hue-security.install
@@ -0,0 +1 @@
+/usr/lib/hue/apps/security


[2/3] bigtop git commit: BIGTOP-1620. Hadoop deb Packages for architecture "all" containing binaries

Posted by of...@apache.org.
BIGTOP-1620. Hadoop deb Packages for architecture "all" containing binaries


Project: http://git-wip-us.apache.org/repos/asf/bigtop/repo
Commit: http://git-wip-us.apache.org/repos/asf/bigtop/commit/34b92a14
Tree: http://git-wip-us.apache.org/repos/asf/bigtop/tree/34b92a14
Diff: http://git-wip-us.apache.org/repos/asf/bigtop/diff/34b92a14

Branch: refs/heads/master
Commit: 34b92a145f3f3f6b96169198ad7b82f25cbd86f1
Parents: e392f6a
Author: Olaf Flebbe <of...@oflebbe.de>
Authored: Thu Jan 29 20:57:30 2015 +0100
Committer: Olaf Flebbe <of...@oflebbe.de>
Committed: Thu Jan 29 21:50:55 2015 +0100

----------------------------------------------------------------------
 bigtop-packages/src/deb/hadoop/control | 78 ++++++++++++++---------------
 1 file changed, 39 insertions(+), 39 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/bigtop/blob/34b92a14/bigtop-packages/src/deb/hadoop/control
----------------------------------------------------------------------
diff --git a/bigtop-packages/src/deb/hadoop/control b/bigtop-packages/src/deb/hadoop/control
index 01e1767..5234bb5 100644
--- a/bigtop-packages/src/deb/hadoop/control
+++ b/bigtop-packages/src/deb/hadoop/control
@@ -22,7 +22,7 @@ Standards-Version: 3.8.0
 Homepage: http://hadoop.apache.org/core/
 
 Package: hadoop
-Architecture: all
+Architecture: any
 Depends: ${shlibs:Depends}, ${misc:Depends}, adduser, bigtop-utils (>= 0.7), zookeeper (>= 3.4.0), psmisc, netcat-openbsd
 Description: Hadoop is a software platform for processing vast amounts of data
  Hadoop is a software platform that lets one easily write and
@@ -46,8 +46,8 @@ Description: Hadoop is a software platform for processing vast amounts of data
  located.
 
 Package: hadoop-hdfs
-Architecture: all
-Depends: ${shlibs:Depends}, ${misc:Depends}, adduser, bigtop-utils (>= 0.7), hadoop (= ${source:Version}), bigtop-jsvc
+Architecture: any
+Depends: ${shlibs:Depends}, ${misc:Depends}, adduser, bigtop-utils (>= 0.7), hadoop (= ${binary:Version}), bigtop-jsvc
 Description: The Hadoop Distributed File System
  Hadoop Distributed File System (HDFS) is the primary storage system used by 
  Hadoop applications. HDFS creates multiple replicas of data blocks and distributes 
@@ -55,8 +55,8 @@ Description: The Hadoop Distributed File System
  computations.
 
 Package: hadoop-yarn
-Architecture: all
-Depends: ${shlibs:Depends}, ${misc:Depends}, adduser, bigtop-utils (>= 0.7), hadoop (= ${source:Version})
+Architecture: any
+Depends: ${shlibs:Depends}, ${misc:Depends}, adduser, bigtop-utils (>= 0.7), hadoop (= ${binary:Version})
 Description: The Hadoop NextGen MapReduce (YARN)
  YARN (Hadoop NextGen MapReduce) is a general purpose data-computation framework.
  The fundamental idea of YARN is to split up the two major functionalities of the 
@@ -73,15 +73,15 @@ Description: The Hadoop NextGen MapReduce (YARN)
  execute and monitor the tasks. 
 
 Package: hadoop-mapreduce
-Architecture: all
-Depends: ${shlibs:Depends}, ${misc:Depends}, adduser, bigtop-utils (>= 0.7), hadoop-yarn (= ${source:Version})
+Architecture: any
+Depends: ${shlibs:Depends}, ${misc:Depends}, adduser, bigtop-utils (>= 0.7), hadoop-yarn (= ${binary:Version})
 Description: The Hadoop MapReduce (MRv2)
  Hadoop MapReduce is a programming model and software framework for writing applications 
  that rapidly process vast amounts of data in parallel on large clusters of compute nodes.
 
 Package: hadoop-hdfs-fuse
-Architecture: i386 amd64
-Depends: ${shlibs:Depends}, hadoop-hdfs (= ${source:Version}), hadoop-client (= ${source:Version}), libfuse2, fuse-utils|fuse, bigtop-utils (>= 0.7)
+Architecture: any
+Depends: ${shlibs:Depends}, hadoop-hdfs (= ${binary:Version}), hadoop-client (= ${binary:Version}), libfuse2, fuse-utils|fuse, bigtop-utils (>= 0.7)
 Enhances: hadoop
 Description: Mountable HDFS
  These projects (enumerated below) allow HDFS to be mounted (on most flavors of Unix) as a standard file system using
@@ -93,60 +93,60 @@ Description: Hadoop Documentation
  Documentation for Hadoop
 
 Package: hadoop-conf-pseudo
-Architecture: all
-Depends: hadoop (= ${source:Version}), hadoop-hdfs-namenode (= ${source:Version}),
-         hadoop-hdfs-datanode (= ${source:Version}), hadoop-hdfs-secondarynamenode (= ${source:Version}),
-         hadoop-yarn-resourcemanager (= ${source:Version}), hadoop-yarn-nodemanager (= ${source:Version}),
-         hadoop-mapreduce-historyserver (= ${source:Version})
+Architecture: any
+Depends: hadoop (= ${binary:Version}), hadoop-hdfs-namenode (= ${binary:Version}),
+         hadoop-hdfs-datanode (= ${binary:Version}), hadoop-hdfs-secondarynamenode (= ${binary:Version}),
+         hadoop-yarn-resourcemanager (= ${binary:Version}), hadoop-yarn-nodemanager (= ${binary:Version}),
+         hadoop-mapreduce-historyserver (= ${binary:Version})
 Description: Pseudo-distributed Hadoop configuration
  Contains configuration files for a "pseudo-distributed" Hadoop deployment.
  In this mode, each of the hadoop components runs as a separate Java process,
  but all on the same machine.
 
 Package: hadoop-mapreduce-historyserver
-Architecture: all
-Depends: hadoop-mapreduce (= ${source:Version}), hadoop-hdfs (= ${source:Version})
+Architecture: any
+Depends: hadoop-mapreduce (= ${binary:Version}), hadoop-hdfs (= ${binary:Version})
 Description: MapReduce History Server
  The History server keeps records of the different activities being performed on a Apache Hadoop cluster
 
 Package: hadoop-yarn-nodemanager
-Architecture: all
-Depends: hadoop-yarn (= ${source:Version})
+Architecture: any
+Depends: hadoop-yarn (= ${binary:Version})
 Description: YARN Node Manager
  The NodeManager is the per-machine framework agent who is responsible for
  containers, monitoring their resource usage (cpu, memory, disk, network) and
  reporting the same to the ResourceManager/Scheduler.
 
 Package: hadoop-yarn-resourcemanager
-Architecture: all
-Depends: hadoop-yarn (= ${source:Version})
+Architecture: any
+Depends: hadoop-yarn (= ${binary:Version})
 Description: YARN Resource Manager
  The resource manager manages the global assignment of compute resources to applications
 
 Package: hadoop-yarn-proxyserver
-Architecture: all
-Depends: hadoop-yarn (= ${source:Version})
+Architecture: any
+Depends: hadoop-yarn (= ${binary:Version})
 Description: YARN Web Proxy
  The web proxy server sits in front of the YARN application master web UI.
 
 Package: hadoop-hdfs-namenode
-Architecture: all
-Depends: hadoop-hdfs (= ${source:Version})
+Architecture: any
+Depends: hadoop-hdfs (= ${binary:Version})
 Description: The Hadoop namenode manages the block locations of HDFS files
  The Hadoop Distributed Filesystem (HDFS) requires one unique server, the
  namenode, which manages the block locations of files on the filesystem.
 
 Package: hadoop-hdfs-secondarynamenode
-Architecture: all
-Depends: hadoop-hdfs (= ${source:Version})
+Architecture: any
+Depends: hadoop-hdfs (= ${binary:Version})
 Description: Hadoop Secondary namenode
  The Secondary Name Node periodically compacts the Name Node EditLog
  into a checkpoint.  This compaction ensures that Name Node restarts
  do not incur unnecessary downtime.
 
 Package: hadoop-hdfs-zkfc
-Architecture: all
-Depends: hadoop-hdfs (= ${source:Version})
+Architecture: any
+Depends: hadoop-hdfs (= ${binary:Version})
 Description: Hadoop HDFS failover controller
  The Hadoop HDFS failover controller is a ZooKeeper client which also
  monitors and manages the state of the NameNode. Each of the machines
@@ -156,16 +156,16 @@ Description: Hadoop HDFS failover controller
 
 Package: hadoop-hdfs-journalnode
 Provides: hadoop-hdfs-journalnode
-Architecture: all
-Depends: hadoop-hdfs (= ${source:Version})
+Architecture: any
+Depends: hadoop-hdfs (= ${binary:Version})
 Description: Hadoop HDFS JournalNode 
  The HDFS JournalNode is responsible for persisting NameNode edit logs. 
  In a typical deployment the JournalNode daemon runs on at least three 
  separate machines in the cluster.
 
 Package: hadoop-hdfs-datanode
-Architecture: all
-Depends: hadoop-hdfs (= ${source:Version})
+Architecture: any
+Depends: hadoop-hdfs (= ${binary:Version})
 Description: Hadoop Data Node
  The Data Nodes in the Hadoop Cluster are responsible for serving up
  blocks of data over the network to Hadoop Distributed Filesystem
@@ -173,27 +173,27 @@ Description: Hadoop Data Node
 
 Package: libhdfs0
 Architecture: any
-Depends: hadoop (= ${source:Version}), ${shlibs:Depends}
+Depends: hadoop (= ${binary:Version}), ${shlibs:Depends}
 Description: Hadoop Filesystem Library
  Hadoop Filesystem Library
 
 Package: libhdfs0-dev
 Architecture: any
 Section: libdevel
-Depends: hadoop (= ${source:Version}), libhdfs0 (= ${binary:Version})
+Depends: hadoop (= ${binary:Version}), libhdfs0 (= ${binary:Version})
 Description: Development support for libhdfs0
  Includes examples and header files for accessing HDFS from C
 
 Package: hadoop-httpfs
-Architecture: all
-Depends: hadoop-hdfs (= ${source:Version}), bigtop-tomcat
+Architecture: any
+Depends: hadoop-hdfs (= ${binary:Version}), bigtop-tomcat
 Description: HTTPFS for Hadoop
   The server providing HTTP REST API support for the complete FileSystem/FileContext
   interface in HDFS.
 
 Package: hadoop-client
-Architecture: all
-Depends: hadoop (= ${source:Version}), hadoop-hdfs (= ${source:Version}),
-         hadoop-yarn (= ${source:Version}), hadoop-mapreduce (= ${source:Version})
+Architecture: any
+Depends: hadoop (= ${binary:Version}), hadoop-hdfs (= ${binary:Version}),
+         hadoop-yarn (= ${binary:Version}), hadoop-mapreduce (= ${binary:Version})
 Description: Hadoop client side dependencies
  Installation of this package will provide you with all the dependencies for Hadoop clients.


[3/3] bigtop git commit: BIGTOP-1135. Hue has google analytics enabled by default

Posted by of...@apache.org.
BIGTOP-1135. Hue has google analytics enabled by default


Project: http://git-wip-us.apache.org/repos/asf/bigtop/repo
Commit: http://git-wip-us.apache.org/repos/asf/bigtop/commit/59fa03c0
Tree: http://git-wip-us.apache.org/repos/asf/bigtop/tree/59fa03c0
Diff: http://git-wip-us.apache.org/repos/asf/bigtop/diff/59fa03c0

Branch: refs/heads/master
Commit: 59fa03c0f18c06a0afd3ba570a5a42e74552f2e0
Parents: 34b92a1
Author: Olaf Flebbe <of...@oflebbe.de>
Authored: Thu Jan 29 21:23:48 2015 +0100
Committer: Olaf Flebbe <of...@oflebbe.de>
Committed: Thu Jan 29 21:51:08 2015 +0100

----------------------------------------------------------------------
 .../common/hue/patch2-disable-google-analytics.diff    | 13 +++++++++++++
 1 file changed, 13 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/bigtop/blob/59fa03c0/bigtop-packages/src/common/hue/patch2-disable-google-analytics.diff
----------------------------------------------------------------------
diff --git a/bigtop-packages/src/common/hue/patch2-disable-google-analytics.diff b/bigtop-packages/src/common/hue/patch2-disable-google-analytics.diff
new file mode 100644
index 0000000..83f7b37
--- /dev/null
+++ b/bigtop-packages/src/common/hue/patch2-disable-google-analytics.diff
@@ -0,0 +1,13 @@
+--- hue-3.7.1/desktop/conf.dist/hue.ini~	2015-01-29 21:23:06.000000000 +0100
++++ hue-3.7.1/desktop/conf.dist/hue.ini	2015-01-29 21:25:47.497503619 +0100
+@@ -79,8 +79,8 @@
+ 
+   # Help improve Hue with anonymous usage analytics.
+   # Use Google Analytics to see how many times an application or specific section of an application is used, nothing more.
+-  ## collect_usage=true
+-  
++  collect_usage=false
++
+   # Support for HTTPS termination at the load-balancer level with SECURE_PROXY_SSL_HEADER.
+   ## secure_proxy_ssl_header=false
+