You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@datalab.apache.org by lf...@apache.org on 2021/04/30 06:58:19 UTC
[incubator-datalab] branch DATALAB-2091 updated: [DATALAB-2091]:
fixed jupyter kernel configuration after emr creation
This is an automated email from the ASF dual-hosted git repository.
lfrolov pushed a commit to branch DATALAB-2091
in repository https://gitbox.apache.org/repos/asf/incubator-datalab.git
The following commit(s) were added to refs/heads/DATALAB-2091 by this push:
new 5c70c93 [DATALAB-2091]: fixed jupyter kernel configuration after emr creation
5c70c93 is described below
commit 5c70c9310b63ebad9e9b74f9fdfa2d66b26e3b1b
Author: leonidfrolov <fr...@gmail.com>
AuthorDate: Fri Apr 30 09:57:59 2021 +0300
[DATALAB-2091]: fixed jupyter kernel configuration after emr creation
---
.../jupyter_install_dataengine-service_kernels.py | 20 ++++++++++----------
1 file changed, 10 insertions(+), 10 deletions(-)
diff --git a/infrastructure-provisioning/src/general/scripts/aws/jupyter_install_dataengine-service_kernels.py b/infrastructure-provisioning/src/general/scripts/aws/jupyter_install_dataengine-service_kernels.py
index b098f4f..91a93e8 100644
--- a/infrastructure-provisioning/src/general/scripts/aws/jupyter_install_dataengine-service_kernels.py
+++ b/infrastructure-provisioning/src/general/scripts/aws/jupyter_install_dataengine-service_kernels.py
@@ -106,31 +106,31 @@ if __name__ == "__main__":
global conn
conn = datalab.fab.init_datalab_connection(args.notebook_ip, args.os_user, args.keyfile)
configure_notebook(args)
- spark_version = get_spark_version(args.cluster_name)
- hadoop_version = get_hadoop_version(args.cluster_name)
+ args.spark_version = get_spark_version(args.cluster_name)
+ args.hadoop_version = get_hadoop_version(args.cluster_name)
r_enabled = os.environ['notebook_r_enabled']
- numpy_version = os.environ['notebook_numpy_version']
+ #numpy_version = os.environ['notebook_numpy_version']
s3_client = boto3.client('s3', config=botoConfig(signature_version='s3v4'), region_name=args.region)
s3_client.download_file(args.bucket, args.project_name + '/' + args.cluster_name + '/scala_version',
'/tmp/scala_version')
s3_client.download_file(args.bucket, args.project_name + '/' + args.cluster_name + '/python_version',
'/tmp/python_version')
with open('/tmp/scala_version') as f:
- scala_version = str(f.read()).rstrip()
- print(scala_version)
+ args.scala_version = str(f.read()).rstrip()
+ print(args.scala_version)
with open('/tmp/python_version') as f:
- python_version = str(f.read()).rstrip()
- print(python_version)
+ args.python_version = str(f.read()).rstrip()
+ print(args.python_version)
if r_enabled == 'true':
s3_client.download_file(args.bucket, args.project_name + '/' + args.cluster_name + '/r_version', '/tmp/r_version')
with open('/tmp/r_version') as g:
- r_version = str(g.read()).rstrip()
- print(r_version)
+ args.r_version = str(g.read()).rstrip()
+ print(args.r_version)
else:
r_version = 'false'
cluster_id = get_emr_id_by_name(args.cluster_name)
master_instances = get_emr_instances_list(cluster_id, 'MASTER')
- master_ip = master_instances[0].get('PrivateIpAddress')
+ args.master_ip = master_instances[0].get('PrivateIpAddress')
#conn.sudo("/usr/bin/python3 /usr/local/bin/jupyter_dataengine-service_create_configs.py --bucket " + args.bucket
# + " --cluster_name " + args.cluster_name + " --emr_version " + args.emr_version + " --spark_version "
# + spark_version + " --scala_version " + scala_version + " --r_version " + r_version + " --hadoop_version "
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@datalab.apache.org
For additional commands, e-mail: commits-help@datalab.apache.org