You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@ambari.apache.org by "Andrew Onischuk (JIRA)" <ji...@apache.org> on 2015/02/13 14:16:11 UTC
[jira] [Resolved] (AMBARI-9618) Add checks in Ambaripreupload.py so
that there is minimum chances of tarball uploads in parallel.
[ https://issues.apache.org/jira/browse/AMBARI-9618?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Andrew Onischuk resolved AMBARI-9618.
-------------------------------------
Resolution: Fixed
> Add checks in Ambaripreupload.py so that there is minimum chances of tarball uploads in parallel.
> -------------------------------------------------------------------------------------------------
>
> Key: AMBARI-9618
> URL: https://issues.apache.org/jira/browse/AMBARI-9618
> Project: Ambari
> Issue Type: Bug
> Reporter: Andrew Onischuk
> Assignee: Andrew Onischuk
> Fix For: 2.0.0
>
>
> Add checks in Ambaripreupload.py so that there is minimum chances of tarball
> uploads in parallel.
> We should add the following checks in:
>
>
>
>
> if dir exists wasb:///hdp/apps/{{ hdp_stack_version }}/mapreduce/:
> copy_tarballs_to_hdfs("/usr/hdp/current/hadoop-client/mapreduce.tar.gz", "wasb:///hdp/apps/{{ hdp_stack_version }}/mapreduce/", 'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, params.user_group)
>
> copy_tarballs_to_hdfs("/usr/hdp/current/sqoop-client/sqoop.tar.gz", "wasb:///hdp/apps/{{ hdp_stack_version }}/sqoop/", 'hadoop-mapreduce-historyserver', params.mapred_user, params.hdfs_user, params.user_group)
>
>
>
> We should do this with all copy tarballs to hdfs in the file - meaning checks
> if the directories exist and then only copy the tarballs.
> The reason for doing this is so that when Ambari is doing the install/upload
> tarballs there is less chances of the ambaripreupload.py script creating a
> race condition on tarball upload.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)