You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@madlib.apache.org by ok...@apache.org on 2020/03/26 20:53:45 UTC

[madlib] branch master updated: DL: Clean up user docs

This is an automated email from the ASF dual-hosted git repository.

okislal pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/madlib.git


The following commit(s) were added to refs/heads/master by this push:
     new 1a9b557  DL: Clean up user docs
1a9b557 is described below

commit 1a9b557ab3c221a2c3956a201cfe2ef060977f39
Author: Orhan Kislal <ok...@apache.org>
AuthorDate: Thu Mar 26 16:53:33 2020 -0400

    DL: Clean up user docs
---
 doc/mainpage.dox.in                                          |  1 -
 src/ports/postgres/modules/deep_learning/madlib_keras.sql_in | 12 ++++++------
 .../deep_learning/madlib_keras_fit_multiple_model.sql_in     | 12 ++++++------
 3 files changed, 12 insertions(+), 13 deletions(-)

diff --git a/doc/mainpage.dox.in b/doc/mainpage.dox.in
index 82be4a5..189b948 100644
--- a/doc/mainpage.dox.in
+++ b/doc/mainpage.dox.in
@@ -296,7 +296,6 @@ Interface and implementation are subject to change.
         @brief Train multiple deep learning models at the same time for model architecture search and hyperparameter selection.
         @details Train multiple deep learning models at the same time for model architecture search and hyperparameter selection.
         @{
-            @defgroup grp_automl AutoML
             @defgroup grp_keras_run_model_selection Run Model Selection
             @defgroup grp_keras_setup_model_selection Setup Model Selection
         @}
diff --git a/src/ports/postgres/modules/deep_learning/madlib_keras.sql_in b/src/ports/postgres/modules/deep_learning/madlib_keras.sql_in
index 75fa56a..b0c77a3 100644
--- a/src/ports/postgres/modules/deep_learning/madlib_keras.sql_in
+++ b/src/ports/postgres/modules/deep_learning/madlib_keras.sql_in
@@ -87,15 +87,15 @@ Note that the following MADlib functions are targeting a specific Keras
 version (2.2.4) with a specific TensorFlow kernel version (1.14).
 Using a newer or older version may or may not work as intended.
 
-@note CUDA GPU memory cannot be released until the process holding it is terminated. 
-When a MADlib deep learning function is called with GPUs, Greenplum internally 
-creates a process (called a slice) which calls TensorFlow to do the computation. 
+@note CUDA GPU memory cannot be released until the process holding it is terminated.
+When a MADlib deep learning function is called with GPUs, Greenplum internally
+creates a process (called a slice) which calls TensorFlow to do the computation.
 This process holds the GPU memory until one of the following two things happen:
-query finishes and user logs out of the Postgres client/session; or, 
-query finishes and user waits for the timeout set by `gp_vmem_idle_resource_timeout`.  
+query finishes and user logs out of the Postgres client/session; or,
+query finishes and user waits for the timeout set by gp_vmem_idle_resource_timeout.
 The default value for this timeout is 18 sec [8].  So the recommendation is:
 log out/reconnect to the session after every GPU query; or
-wait for `gp_vmem_idle_resource_timeout` before you run another GPU query (you can 
+wait for gp_vmem_idle_resource_timeout before you run another GPU query (you can
 also set it to a lower value).
 
 @anchor keras_fit
diff --git a/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in b/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
index b929724..4d1eb09 100644
--- a/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
+++ b/src/ports/postgres/modules/deep_learning/madlib_keras_fit_multiple_model.sql_in
@@ -94,15 +94,15 @@ release the disk space once the fit multiple query has completed execution.
 This is not the case for GPDB 6+ where disk space is released during the
 fit multiple query.
 
-@note CUDA GPU memory cannot be released until the process holding it is terminated. 
-When a MADlib deep learning function is called with GPUs, Greenplum internally 
-creates a process (called a slice) which calls TensorFlow to do the computation. 
+@note CUDA GPU memory cannot be released until the process holding it is terminated.
+When a MADlib deep learning function is called with GPUs, Greenplum internally
+creates a process (called a slice) which calls TensorFlow to do the computation.
 This process holds the GPU memory until one of the following two things happen:
-query finishes and user logs out of the Postgres client/session; or, 
-query finishes and user waits for the timeout set by `gp_vmem_idle_resource_timeout`.  
+query finishes and user logs out of the Postgres client/session; or,
+query finishes and user waits for the timeout set by gp_vmem_idle_resource_timeout.
 The default value for this timeout is 18 sec [8].  So the recommendation is:
 log out/reconnect to the session after every GPU query; or
-wait for `gp_vmem_idle_resource_timeout` before you run another GPU query (you can 
+wait for gp_vmem_idle_resource_timeout before you run another GPU query (you can
 also set it to a lower value).
 
 @anchor keras_fit