You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@ambari.apache.org by al...@apache.org on 2016/06/29 14:37:06 UTC
[07/11] ambari git commit: Revert "AMBARI-17079. Moved Hue to Ambari
migrator to standard view architecture (Pradarttana Panda via dipayanb)"
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HiveInstanceDetailsUtility.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HiveInstanceDetailsUtility.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HiveInstanceDetailsUtility.java
deleted file mode 100644
index bcbe4de..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HiveInstanceDetailsUtility.java
+++ /dev/null
@@ -1,106 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.ambari.view.huetoambarimigration.migration.configuration;
-
-import org.apache.ambari.view.ViewContext;
-import org.apache.ambari.view.huetoambarimigration.datasource.DataSourceAmbariDatabase;
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.models.InstanceModel;
-
-import java.beans.PropertyVetoException;
-import java.io.IOException;
-import java.sql.*;
-import java.util.ArrayList;
-import java.util.List;
-
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.ambariqueryset.hive.instancedetail.*;
-
-
-public class HiveInstanceDetailsUtility {
-
- public List<InstanceModel> getInstancedetails(ViewContext view) throws PropertyVetoException, SQLException, IOException {
-
- List<InstanceModel> instancelist = new ArrayList<>();
- Connection conn = null;
- conn = DataSourceAmbariDatabase.getInstance(view.getProperties().get("ambaridrivername"), view.getProperties().get("ambarijdbcurl"), view.getProperties().get("ambaridbusername"), view.getProperties().get("ambaridbpassword")).getConnection();
- conn.setAutoCommit(false);
- PreparedStatement prSt;
-
- QuerySetAmbariDB ambaridatabase = null;
-
- if (view.getProperties().get("ambaridrivername").contains("mysql")) {
- ambaridatabase = new MysqlQuerySetAmbariDB();
- } else if (view.getProperties().get("ambaridrivername").contains("postgresql")) {
- ambaridatabase = new PostgressQuerySetAmbariDB();
- } else if (view.getProperties().get("ambaridrivername").contains("oracle")) {
- ambaridatabase = new OracleQuerySetAmbariDB();
- }
-
- ResultSet rs1 = null;
- prSt = ambaridatabase.getHiveInstanceDeatil(conn);
- rs1 = prSt.executeQuery();
- int i = 0;
-
- while (rs1.next()) {
- InstanceModel I = new InstanceModel();
- I.setInstanceName(rs1.getString(1));
- I.setId(i);
- instancelist.add(I);
- i++;
- }
- return instancelist;
-
- }
-
- public List<InstanceModel> getAllInstancedetails(ViewContext view) throws PropertyVetoException, SQLException, IOException {
-
- List<InstanceModel> instancelist = new ArrayList<>();
- Connection conn = null;
- Statement stmt = null;
- conn = DataSourceAmbariDatabase.getInstance(view.getProperties().get("ambaridrivername"), view.getProperties().get("ambarijdbcurl"), view.getProperties().get("ambaridbusername"), view.getProperties().get("ambaridbpassword")).getConnection();
- conn.setAutoCommit(false);
- PreparedStatement prSt;
-
- QuerySetAmbariDB ambaridatabase = null;
-
- if (view.getProperties().get("ambaridrivername").contains("mysql")) {
- ambaridatabase = new MysqlQuerySetAmbariDB();
- } else if (view.getProperties().get("ambaridrivername").contains("postgresql")) {
- ambaridatabase = new PostgressQuerySetAmbariDB();
- } else if (view.getProperties().get("ambaridrivername").contains("oracle")) {
- ambaridatabase = new OracleQuerySetAmbariDB();
- }
-
- ResultSet rs1 = null;
- int i = 0;
- prSt = ambaridatabase.getAllInstanceDeatil(conn);
- rs1 = prSt.executeQuery();
-
- while (rs1.next()) {
- InstanceModel I = new InstanceModel();
- I.setInstanceName(rs1.getString(1));
- I.setId(i);
- instancelist.add(I);
- i++;
- }
- rs1.close();
- return instancelist;
-
- }
-
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueDatabaseCheck.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueDatabaseCheck.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueDatabaseCheck.java
deleted file mode 100644
index a5ec758..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueDatabaseCheck.java
+++ /dev/null
@@ -1,60 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.ambari.view.huetoambarimigration.migration.configuration;
-
-
-import com.google.inject.Inject;
-import org.apache.ambari.view.ViewContext;
-import org.apache.ambari.view.huetoambarimigration.utils.ServiceFormattedException;
-import org.json.simple.JSONObject;
-
-import javax.ws.rs.GET;
-import javax.ws.rs.Path;
-import javax.ws.rs.Produces;
-import javax.ws.rs.WebApplicationException;
-import javax.ws.rs.core.MediaType;
-import javax.ws.rs.core.Response;
-import java.io.IOException;
-
-/**
- * Service class to check hue database check
- */
-@Path("/huedatabases")
-
-public class HueDatabaseCheck {
-
- @Inject
- ViewContext view;
-
- @GET
- @Produces(MediaType.APPLICATION_JSON)
- public Response hueDatabase() throws IOException {
- JSONObject response = new JSONObject();
- try {
- response.put("huedatabase", ConfigurationCheckImplementation.checkHueDatabaseConnection(view.getProperties().get("huedrivername"), view.getProperties().get("huejdbcurl"), view.getProperties().get("huedbusername"), view.getProperties().get("huedbpassword")));
- return Response.ok(response).build();
- } catch (WebApplicationException ex) {
- throw ex;
- } catch (Exception ex) {
- throw new ServiceFormattedException(ex.getMessage(), ex);
- }
-
- }
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueHttpUrlCheck.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueHttpUrlCheck.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueHttpUrlCheck.java
deleted file mode 100644
index dd5f409..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueHttpUrlCheck.java
+++ /dev/null
@@ -1,60 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.ambari.view.huetoambarimigration.migration.configuration;
-
-
-import com.google.inject.Inject;
-import org.apache.ambari.view.ViewContext;
-import org.apache.ambari.view.huetoambarimigration.utils.ServiceFormattedException;
-import org.json.simple.JSONObject;
-
-import javax.ws.rs.GET;
-import javax.ws.rs.Path;
-import javax.ws.rs.Produces;
-import javax.ws.rs.WebApplicationException;
-import javax.ws.rs.core.MediaType;
-import javax.ws.rs.core.Response;
-
-/**
- * Service class to check hue http url
- */
-@Path("/huehttpurls")
-
-public class HueHttpUrlCheck {
-
- @Inject
- ViewContext view;
-
- @GET
- @Produces(MediaType.APPLICATION_JSON)
- public Response hueHtttpUrl() {
-
- JSONObject response = new JSONObject();
- try {
- response.put("huehttpurl", ConfigurationCheckImplementation.checkConfigurationForHue(view.getProperties().get("Hue_URL")));
- return Response.ok(response).build();
-
- } catch (WebApplicationException ex) {
- throw ex;
- } catch (Exception ex) {
- throw new ServiceFormattedException(ex.getMessage(), ex);
- }
- }
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueWebHdfsCheck.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueWebHdfsCheck.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueWebHdfsCheck.java
deleted file mode 100644
index 7b5e7a0..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/HueWebHdfsCheck.java
+++ /dev/null
@@ -1,60 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.ambari.view.huetoambarimigration.migration.configuration;
-
-
-import com.google.inject.Inject;
-import org.apache.ambari.view.ViewContext;
-import org.apache.ambari.view.huetoambarimigration.utils.ServiceFormattedException;
-import org.json.simple.JSONObject;
-
-import javax.ws.rs.GET;
-import javax.ws.rs.Path;
-import javax.ws.rs.Produces;
-import javax.ws.rs.WebApplicationException;
-import javax.ws.rs.core.MediaType;
-import javax.ws.rs.core.Response;
-import java.io.IOException;
-import java.net.URISyntaxException;
-
-/**
- * Service class to check hueWebHdfs
- */
-
-@Path("/huewebhdfsurls")
-
-public class HueWebHdfsCheck {
-
- @Inject
- ViewContext view;
-
- @GET
- @Produces(MediaType.APPLICATION_JSON)
- public Response hueWebHdfs() throws IOException, URISyntaxException {
-
- JSONObject response = new JSONObject();
- try {
- response.put("huewebhdfsurl", ConfigurationCheckImplementation.checkNamenodeURIConnectionforHue(view.getProperties().get("namenode_URI_Hue")));
- return Response.ok(response).build();
- } catch (WebApplicationException ex) {
- throw ex;
- } catch (Exception ex) {
- throw new ServiceFormattedException(ex.getMessage(), ex);
- }
- }
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/PigInstanceDetailsAmbari.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/PigInstanceDetailsAmbari.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/PigInstanceDetailsAmbari.java
deleted file mode 100644
index cc14385..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/PigInstanceDetailsAmbari.java
+++ /dev/null
@@ -1,61 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.ambari.view.huetoambarimigration.migration.configuration;
-
-
-import com.google.inject.Inject;
-import org.apache.ambari.view.ViewContext;
-import org.json.simple.JSONObject;
-
-import javax.ws.rs.GET;
-import javax.ws.rs.Path;
-import javax.ws.rs.Produces;
-import javax.ws.rs.core.MediaType;
-import javax.ws.rs.core.Response;
-import java.beans.PropertyVetoException;
-import java.io.IOException;
-import java.sql.SQLException;
-
-/**
- * Service class to fetch Piginstance detail
- */
-
-@Path("/piginstancedetails")
-
-public class PigInstanceDetailsAmbari {
-
- @Inject
- ViewContext view;
-
- @GET
- @Produces(MediaType.APPLICATION_JSON)
- public Response instancelist() throws IOException, PropertyVetoException, SQLException {
-
-
- PigInstanceDetailsUtility instance=new PigInstanceDetailsUtility();
-
- JSONObject response = new JSONObject();
- response.put("piginstancedetails",instance.getInstancedetails(view));
- return Response.ok(response).build();
-
-
- }
-
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/PigInstanceDetailsUtility.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/PigInstanceDetailsUtility.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/PigInstanceDetailsUtility.java
deleted file mode 100644
index d3cfb97..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/PigInstanceDetailsUtility.java
+++ /dev/null
@@ -1,79 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.ambari.view.huetoambarimigration.migration.configuration;
-
-import org.apache.ambari.view.ViewContext;
-import org.apache.ambari.view.huetoambarimigration.datasource.DataSourceAmbariDatabase;
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.models.InstanceModel;
-
-import java.beans.PropertyVetoException;
-import java.io.IOException;
-import java.sql.*;
-import java.util.ArrayList;
-import java.util.List;
-
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.ambariqueryset.pig.instancedetail.*;
-
-/**
- * Utility class to fetch Pig Instance details
- */
-
-public class PigInstanceDetailsUtility {
-
- public List<InstanceModel> getInstancedetails(ViewContext view) throws PropertyVetoException, SQLException, IOException {
-
- List<InstanceModel> instancelist = new ArrayList<>();
- Connection conn = null;
- Statement stmt = null;
- PreparedStatement prSt;
- conn = DataSourceAmbariDatabase.getInstance(view.getProperties().get("ambaridrivername"), view.getProperties().get("ambarijdbcurl"), view.getProperties().get("ambaridbusername"), view.getProperties().get("ambaridbpassword")).getConnection();
- conn.setAutoCommit(false);
- stmt = conn.createStatement();
- int i = 0;
-
- QuerySetAmbariDB ambaridatabase = null;
-
- if (view.getProperties().get("ambaridrivername").contains("mysql")) {
- ambaridatabase = new MysqlQuerySetAmbariDB();
- } else if (view.getProperties().get("ambaridrivername").contains("postgresql")) {
- ambaridatabase = new PostgressQuerySetAmbariDB();
- } else if (view.getProperties().get("ambaridrivername").contains("oracle")) {
- ambaridatabase = new OracleQuerySetAmbariDB();
- }
-
- ResultSet rs1 = null;
-
- prSt = ambaridatabase.getAllPigInstance(conn);
-
- rs1 = prSt.executeQuery();
-
- while (rs1.next()) {
- InstanceModel I = new InstanceModel();
- I.setInstanceName(rs1.getString(1));
- I.setId(i);
- instancelist.add(I);
- i++;
- }
- rs1.close();
- stmt.close();
- return instancelist;
-
- }
-
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/UserDetailHue.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/UserDetailHue.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/UserDetailHue.java
deleted file mode 100644
index d993f3a..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/UserDetailHue.java
+++ /dev/null
@@ -1,58 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.ambari.view.huetoambarimigration.migration.configuration;
-
-
-import com.google.inject.Inject;
-import org.apache.ambari.view.ViewContext;
-import org.json.simple.JSONObject;
-
-import javax.ws.rs.GET;
-import javax.ws.rs.Path;
-import javax.ws.rs.Produces;
-import javax.ws.rs.core.MediaType;
-import javax.ws.rs.core.Response;
-import java.beans.PropertyVetoException;
-import java.io.IOException;
-import java.sql.SQLException;
-
-/**
- * Service class to fetch user detail
- */
-
-@Path("/usersdetails")
-
-public class UserDetailHue {
-
- @Inject
- ViewContext view;
-
- @GET
- @Produces(MediaType.APPLICATION_JSON)
- public Response userList() throws IOException, PropertyVetoException, SQLException {
-
- UserDetailsUtility user=new UserDetailsUtility();
-
- JSONObject response = new JSONObject();
- response.put("usersdetails",user.getUserDetails(view));
- return Response.ok(response).build();
- }
-
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/UserDetailsUtility.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/UserDetailsUtility.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/UserDetailsUtility.java
deleted file mode 100644
index 09f175b..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/configuration/UserDetailsUtility.java
+++ /dev/null
@@ -1,80 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.ambari.view.huetoambarimigration.migration.configuration;
-
-import org.apache.ambari.view.ViewContext;
-import org.apache.ambari.view.huetoambarimigration.datasource.DataSourceHueDatabase;
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.models.UserModel;
-
-import java.beans.PropertyVetoException;
-import java.io.IOException;
-import java.sql.*;
-import java.util.ArrayList;
-import java.util.List;
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.huequeryset.userdetails.*;
-
-
-public class UserDetailsUtility {
-
-
- public List<UserModel> getUserDetails(ViewContext view) throws PropertyVetoException, SQLException, IOException {
-
- List<UserModel> userlist=new ArrayList<>();
- Connection conn = null;
- Statement stmt = null;
- conn = DataSourceHueDatabase.getInstance(view.getProperties().get("huedrivername"),view.getProperties().get("huejdbcurl"),view.getProperties().get("huedbusername"),view.getProperties().get("huedbpassword")).getConnection();
- conn.setAutoCommit(false);
- stmt = conn.createStatement();
- UserModel all=new UserModel();
- all.setId(-1);
- all.setUsername("all");
- PreparedStatement prSt;
- userlist.add(all);
- ResultSet rs1=null;
-
- QuerySet huedatabase = null;
-
- if (view.getProperties().get("huedrivername").contains("mysql")) {
- huedatabase = new MysqlQuerySet();
- } else if (view.getProperties().get("huedrivername").contains("postgresql")) {
- huedatabase = new PostgressQuerySet();
- } else if (view.getProperties().get("huedrivername").contains("sqlite")) {
- huedatabase = new SqliteQuerySet();
- } else if (view.getProperties().get("huedrivername").contains("oracle")) {
- huedatabase = new OracleQuerySet();
- }
-
- prSt = huedatabase.getUserDetails(conn);
-
- rs1 = prSt.executeQuery();
-
- while (rs1.next()) {
- UserModel I=new UserModel();
- I.setUsername(rs1.getString(2));
- I.setId(rs1.getInt(1));
- System.out.println(rs1.getString(2));
- userlist.add(I);
- }
- rs1.close();
- stmt.close();
- return userlist;
-
- }
-
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryMigrationUtility.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryMigrationUtility.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryMigrationUtility.java
deleted file mode 100755
index bdcf293..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryMigrationUtility.java
+++ /dev/null
@@ -1,255 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-
-package org.apache.ambari.view.huetoambarimigration.migration.hive.historyquery;
-
-import org.apache.ambari.view.ViewContext;
-import org.apache.ambari.view.huetoambarimigration.datasource.DataSourceAmbariDatabase;
-import org.apache.ambari.view.huetoambarimigration.datasource.DataSourceHueDatabase;
-import org.apache.ambari.view.huetoambarimigration.migration.InitiateJobMigration;
-import org.apache.ambari.view.huetoambarimigration.migration.configuration.ConfigurationCheckImplementation;
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.models.HiveModel;
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.models.MigrationModel;
-import org.apache.ambari.view.huetoambarimigration.persistence.utils.ItemNotFound;
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.ambariqueryset.hive.historyqueryset.MysqlQuerySetAmbariDB;
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.ambariqueryset.hive.historyqueryset.OracleQuerySetAmbariDB;
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.ambariqueryset.hive.historyqueryset.PostgressQuerySetAmbariDB;
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.ambariqueryset.hive.historyqueryset.QuerySetAmbariDB;
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.huequeryset.hive.historyqueryset.*;
-import org.apache.ambari.view.huetoambarimigration.resources.PersonalCRUDResourceManager;
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.MigrationResourceManager;
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.models.MigrationResponse;
-import org.apache.log4j.Logger;
-
-import java.beans.PropertyVetoException;
-import java.io.IOException;
-import java.net.URISyntaxException;
-import java.sql.Connection;
-import java.sql.SQLException;
-import java.text.ParseException;
-import java.util.ArrayList;
-
-
-public class HiveHistoryMigrationUtility {
-
-
- protected MigrationResourceManager resourceManager = null;
-
- public synchronized PersonalCRUDResourceManager<MigrationResponse> getResourceManager(ViewContext view) {
- if (resourceManager == null) {
- resourceManager = new MigrationResourceManager(view);
- }
- return resourceManager;
- }
-
-
- public void hiveHistoryQueryMigration(String username, String instance, String startDate, String endDate, ViewContext view, MigrationResponse migrationresult, String jobid) throws IOException, ItemNotFound {
-
- InitiateJobMigration migrationservice = new InitiateJobMigration();
-
- long startTime = System.currentTimeMillis();
-
- final Logger logger = Logger.getLogger(HiveHistoryMigrationUtility.class);
- Connection connectionHuedb = null;
- Connection connectionAmbaridb = null;
-
- logger.info("--------------------------------------");
- logger.info("hive History query Migration started");
- logger.info("--------------------------------------");
- logger.info("start date: " + startDate);
- logger.info("enddate date: " + endDate);
- logger.info("instance is: " + username);
- logger.info("hue username is : " + instance);
-
- MigrationModel model = new MigrationModel();
-
- int maxCountOfAmbariDb, i = 0;
- String time = null;
- Long epochTime = null;
- String dirNameforHiveHistroy;
- ArrayList<HiveModel> dbpojoHiveHistoryQuery = new ArrayList<HiveModel>();
-
- HiveHistoryQueryMigrationImplementation hiveHistoryQueryImpl = new HiveHistoryQueryMigrationImplementation();// creating objects of HiveHistroy implementation
-
- QuerySet huedatabase = null;
-
- /*instanciang queryset
- * according to driver name
- */
-
- if (view.getProperties().get("huedrivername").contains("mysql")) {
- huedatabase = new MysqlQuerySet();
- } else if (view.getProperties().get("huedrivername").contains("postgresql")) {
- huedatabase = new PostgressQuerySet();
- } else if (view.getProperties().get("huedrivername").contains("sqlite")) {
- huedatabase = new SqliteQuerySet();
- } else if (view.getProperties().get("huedrivername").contains("oracle")) {
- huedatabase = new OracleQuerySet();
- }
-
-
- QuerySetAmbariDB ambaridatabase = null;
-
-
- if (view.getProperties().get("ambaridrivername").contains("mysql")) {
- ambaridatabase = new MysqlQuerySetAmbariDB();
- } else if (view.getProperties().get("ambaridrivername").contains("postgresql")) {
- ambaridatabase = new PostgressQuerySetAmbariDB();
- } else if (view.getProperties().get("ambaridrivername").contains("oracle")) {
- ambaridatabase = new OracleQuerySetAmbariDB();
- }
-
-
- try {
-
- connectionHuedb = DataSourceHueDatabase.getInstance(view.getProperties().get("huedrivername"), view.getProperties().get("huejdbcurl"), view.getProperties().get("huedbusername"), view.getProperties().get("huedbpassword")).getConnection();
-
- dbpojoHiveHistoryQuery = hiveHistoryQueryImpl.fetchFromHue(username, startDate, endDate, connectionHuedb, huedatabase);
-
- for (int j = 0; j < dbpojoHiveHistoryQuery.size(); j++) {
- logger.info("the query fetched from hue" + dbpojoHiveHistoryQuery.get(j).getQuery());
-
- }
-
- /* if No migration query selected from Hue Database according to our search criteria */
-
- if (dbpojoHiveHistoryQuery.size() == 0) {
- migrationresult.setIsNoQuerySelected("yes");
- migrationresult.setProgressPercentage(0);
- migrationresult.setNumberOfQueryTransfered(0);
- migrationresult.setTotalNoQuery(dbpojoHiveHistoryQuery.size());
- getResourceManager(view).update(migrationresult, jobid);
- logger.info("No queries has been selected acccording to your criteria");
-
- } else {
- /* If hive queries are selected based on our search criteria */
-
- connectionAmbaridb = DataSourceAmbariDatabase.getInstance(view.getProperties().get("ambaridrivername"), view.getProperties().get("ambarijdbcurl"), view.getProperties().get("ambaridbusername"), view.getProperties().get("ambaridbpassword")).getConnection();// connecting to ambari db
- connectionAmbaridb.setAutoCommit(false);
-
- // for each queries fetched from Hue database//
-
- for (i = 0; i < dbpojoHiveHistoryQuery.size(); i++) {
-
- float calc = ((float) (i + 1)) / dbpojoHiveHistoryQuery.size() * 100;
- int progressPercentage = Math.round(calc);
- migrationresult.setIsNoQuerySelected("no");
- migrationresult.setProgressPercentage(progressPercentage);
- migrationresult.setNumberOfQueryTransfered(i + 1);
- migrationresult.setTotalNoQuery(dbpojoHiveHistoryQuery.size());
- getResourceManager(view).update(migrationresult, jobid);
-
- logger.info("_____________________");
- logger.info("Loop No." + (i + 1));
- logger.info("_____________________");
- logger.info("Hue query that has been fetched" + dbpojoHiveHistoryQuery.get(i).getQuery());
- int id = 0;
-
- id = hiveHistoryQueryImpl.fetchInstanceTablename(connectionAmbaridb, instance, ambaridatabase); // feching table name according to the given instance name
-
- logger.info("Table name has been fetched from intance name");
-
- hiveHistoryQueryImpl.writetoFileQueryhql(dbpojoHiveHistoryQuery.get(i).getQuery(), ConfigurationCheckImplementation.getHomeDir());// writing to .hql file to a temp file on local disk
-
- logger.info(".hql file created in Temp directory");
-
- hiveHistoryQueryImpl.writetoFileLogs(ConfigurationCheckImplementation.getHomeDir());// writing to logs file to a temp file on local disk
-
- logger.info("Log file created in Temp directory");
-
- maxCountOfAmbariDb = (hiveHistoryQueryImpl.fetchMaximumIdfromAmbaridb(connectionAmbaridb, id, ambaridatabase) + 1);// fetching the maximum count for ambari db to insert
-
- time = hiveHistoryQueryImpl.getTime();// getting the system current time.
-
- epochTime = hiveHistoryQueryImpl.getEpochTime();// getting system time as epoch format
-
- dirNameforHiveHistroy = "/user/admin/migration/jobs/migration-job-" + maxCountOfAmbariDb + "-" + time + "/";// creating the directory name
-
- logger.info("Directory name where .hql will be saved: " + dirNameforHiveHistroy);
-
- hiveHistoryQueryImpl.insertRowinAmbaridb(dirNameforHiveHistroy, maxCountOfAmbariDb, epochTime, connectionAmbaridb, id, instance, i, ambaridatabase);// inserting in ambari database
-
- if (view.getProperties().get("KerberoseEnabled").equals("y")) {
-
- logger.info("kerberose enabled");
- hiveHistoryQueryImpl.createDirKerberorisedSecured(dirNameforHiveHistroy, view.getProperties().get("namenode_URI_Ambari"));// creating directory in kerborized secured hdfs
- logger.info("Directory created in hdfs");
- hiveHistoryQueryImpl.putFileinHdfsKerborizedSecured(ConfigurationCheckImplementation.getHomeDir() + "query.hql", dirNameforHiveHistroy, view.getProperties().get("namenode_URI_Ambari"));// copying the .hql file to kerborized hdfs
- hiveHistoryQueryImpl.putFileinHdfsKerborizedSecured(ConfigurationCheckImplementation.getHomeDir() + "logs", dirNameforHiveHistroy, view.getProperties().get("namenode_URI_Ambari"));// copying the log file to kerborized hdfs
- } else {
-
- logger.info("kerberose not enabled");
- hiveHistoryQueryImpl.createDir(dirNameforHiveHistroy, view.getProperties().get("namenode_URI_Ambari"));// creating directory in hdfs
- logger.info("Directory created in hdfs");
- hiveHistoryQueryImpl.putFileinHdfs(ConfigurationCheckImplementation.getHomeDir() + "query.hql", dirNameforHiveHistroy, view.getProperties().get("namenode_URI_Ambari"));// copying the .hql file to hdfs
- hiveHistoryQueryImpl.putFileinHdfs(ConfigurationCheckImplementation.getHomeDir() + "logs", dirNameforHiveHistroy, view.getProperties().get("namenode_URI_Ambari"));// copying the log file to hdfs
- }
-
- }
- connectionAmbaridb.commit();
-
-
- }
- } catch (SQLException e) {
- logger.error("Sql exception in ambari database: ", e);
- try {
- connectionAmbaridb.rollback();
- model.setIfSuccess(false);
- logger.error("Sql statement are Rolledback");
- } catch (SQLException e1) {
- logger.error("Sql rollback exception in ambari database",
- e1);
- }
- } catch (ClassNotFoundException e) {
- logger.error("Class not found :- ", e);
- } catch (ParseException e) {
- logger.error("Parse Exception : ", e);
- } catch (URISyntaxException e) {
- logger.error("URI Syntax Exception: ", e);
- } catch (PropertyVetoException e) {
- logger.error("PropertyVetoException: ", e);
- } catch (ItemNotFound itemNotFound) {
- itemNotFound.printStackTrace();
- } finally {
- if (connectionAmbaridb != null) try {
- connectionAmbaridb.close();
- } catch (SQLException e) {
- logger.error("Exception in closing the connection :", e);
- }
- }
- //deleteing the temprary files that are created while execution
- hiveHistoryQueryImpl.deleteFileQueryhql(ConfigurationCheckImplementation.getHomeDir());
- hiveHistoryQueryImpl.deleteFileQueryLogs(ConfigurationCheckImplementation.getHomeDir());
-
- //session.setAttribute(ProgressBarStatus.TASK_PROGRESS_VARIABLE, 0);
- logger.info("------------------------------");
- logger.info("hive History query Migration Ends");
- logger.info("------------------------------");
-
- long stopTime = System.currentTimeMillis();
- long elapsedTime = stopTime - startTime;
-
- migrationresult.setJobtype("hivehistoryquerymigration");
- migrationresult.setTotalTimeTaken(String.valueOf(elapsedTime));
- getResourceManager(view).update(migrationresult, jobid);
-
-
- }
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryQueryMigrationImplementation.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryQueryMigrationImplementation.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryQueryMigrationImplementation.java
deleted file mode 100644
index a0182f6..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryQueryMigrationImplementation.java
+++ /dev/null
@@ -1,551 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.ambari.view.huetoambarimigration.migration.hive.historyquery;
-
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.ambariqueryset.hive.historyqueryset.QuerySetAmbariDB;
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.huequeryset.hive.historyqueryset.QuerySet;
-import org.apache.ambari.view.huetoambarimigration.migration.configuration.ConfigurationCheckImplementation;
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.models.HiveModel;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FSDataOutputStream;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.security.UserGroupInformation;
-import org.apache.log4j.Logger;
-import org.jdom.Attribute;
-import org.jdom.Document;
-import org.jdom.Element;
-import org.jdom.JDOMException;
-import org.jdom.input.SAXBuilder;
-import org.jdom.output.Format;
-import org.jdom.output.XMLOutputter;
-
-import java.io.*;
-import java.net.URISyntaxException;
-import java.security.PrivilegedExceptionAction;
-import java.sql.*;
-import java.text.ParseException;
-import java.text.SimpleDateFormat;
-import java.util.ArrayList;
-import java.util.Calendar;
-import java.util.Date;
-import java.util.GregorianCalendar;
-
-public class HiveHistoryQueryMigrationImplementation {
-
- static final Logger logger = Logger.getLogger(HiveHistoryQueryMigrationImplementation.class);
-
- public void wrtitetoalternatesqlfile(String dirname, String content, String instance, int i) throws IOException {
-
- Date dNow = new Date();
- SimpleDateFormat ft = new SimpleDateFormat("YYYY-MM-dd hh:mm:ss");
- String currentDate = ft.format(dNow);
-
- XMLOutputter xmlOutput = new XMLOutputter();
- xmlOutput.setFormat(Format.getPrettyFormat());
-
- File xmlfile = new File(ConfigurationCheckImplementation.getHomeDir() + "RevertChangesService.xml");
-
- if (xmlfile.exists()) {
- String iteration = Integer.toString(i + 1);
- SAXBuilder builder = new SAXBuilder();
- Document doc;
- try {
- doc = (Document) builder.build(xmlfile);
- Element rootNode = doc.getRootElement();
- Element record = new Element("RevertRecord");
- record.setAttribute(new Attribute("id", iteration));
- record.addContent(new Element("datetime").setText(currentDate.toString()));
- record.addContent(new Element("dirname").setText(dirname));
- record.addContent(new Element("instance").setText(instance));
- record.addContent(new Element("query").setText(content));
- rootNode.addContent(record);
- xmlOutput.output(doc, new FileWriter(ConfigurationCheckImplementation.getHomeDir() + "RevertChangesService.xml"));
-
- } catch (JDOMException e) {
- logger.error("JDOMException", e);
-
- }
-
- } else {
-
- try {
- String iteration = Integer.toString(i + 1);
- Element revertrecord = new Element("RevertChangePage");
- Document doc = new Document(revertrecord);
- doc.setRootElement(revertrecord);
-
- Element record = new Element("RevertRecord");
- record.setAttribute(new Attribute("id", iteration));
- record.addContent(new Element("datetime").setText(currentDate.toString()));
- record.addContent(new Element("dirname").setText(dirname));
- record.addContent(new Element("instance").setText(instance));
- record.addContent(new Element("query").setText(content));
- doc.getRootElement().addContent(record);
- xmlOutput.output(doc, new FileWriter(ConfigurationCheckImplementation.getHomeDir() + "RevertChangesService.xml"));
- } catch (IOException io) {
- logger.error("JDOMException", io);
- }
-
- }
-
- }
-
- public int fetchMaximumIdfromAmbaridb(Connection c, int id, QuerySetAmbariDB ambaridatabase) throws SQLException {
-
- String ds_id = null;
- ResultSet rs = null;
- PreparedStatement prSt = null;
-
- prSt = ambaridatabase.getMaxDsIdFromTableId(c, id);
-
- rs = prSt.executeQuery();
-
- while (rs.next()) {
- ds_id = rs.getString("max");
- }
-
- int num;
- if (ds_id == null) {
- num = 1;
- } else {
- num = Integer.parseInt(ds_id);
- }
- return num;
- }
-
-
- public void insertRowinAmbaridb(String dirname, int maxcount, long epochtime, Connection c, int id, String instance, int i, QuerySetAmbariDB ambaridatabase) throws SQLException, IOException {
-
- String maxcount1 = Integer.toString(maxcount);
- String epochtime1 = Long.toString(epochtime);
- PreparedStatement prSt = null;
- String revsql = null;
-
- prSt = ambaridatabase.insertToHiveHistory(c, id, maxcount1, epochtime, dirname);
-
- logger.info("The actual insert statement is " + prSt);
-
- prSt.executeUpdate();
-
- revsql = ambaridatabase.RevertSql(id, maxcount1);
-
- logger.info("adding revert sql hive history");
-
- wrtitetoalternatesqlfile(dirname, revsql, instance, i);
-
-
- }
-
- public int fetchInstanceTablename(Connection c, String instance, QuerySetAmbariDB ambaridatabase) throws SQLException {
-
- String ds_id = new String();
- int id = 0;
- Statement stmt = null;
- PreparedStatement prSt = null;
-
-
- ResultSet rs = null;
-
-
- prSt = ambaridatabase.getTableIdFromInstanceName(c, instance);
-
- logger.info("sql statement to fetch is from ambari instance:= = " + prSt);
-
- rs = prSt.executeQuery();
-
- while (rs.next()) {
- id = rs.getInt("id");
- }
- return id;
- }
-
- public long getEpochTime() throws ParseException {
- long seconds = System.currentTimeMillis() / 1000l;
- return seconds;
-
- }
-
- public String getTime() throws ParseException {
- int day, month, year;
- int second, minute, hour;
- int milisecond;
- GregorianCalendar date = new GregorianCalendar();
-
- day = date.get(Calendar.DAY_OF_MONTH);
- month = date.get(Calendar.MONTH);
- year = date.get(Calendar.YEAR);
-
- second = date.get(Calendar.SECOND);
- minute = date.get(Calendar.MINUTE);
- hour = date.get(Calendar.HOUR);
- milisecond = date.get(Calendar.MILLISECOND);
-
- String s = year + "-" + (month + 1) + "-" + day + "_" + hour + "-"
- + minute;
- String s1 = year + "-" + (month + 1) + "-" + day + "_" + hour + "-"
- + minute + "-" + second + "-" + milisecond;
- SimpleDateFormat df = new SimpleDateFormat("yyyy-MM-dd_HH-mm-ss-SSS");
- Date date1 = df.parse(s1);
- long epoch = date1.getTime();
- return s;
-
- }
-
- public ArrayList<HiveModel> fetchFromHue(String username, String startdate, String endtime, Connection connection, QuerySet huedatabase) throws ClassNotFoundException, SQLException {
- int id = 0;
- int i = 0;
- ArrayList<HiveModel> hiveArrayList = new ArrayList<HiveModel>();
-
-
- try {
- connection.setAutoCommit(false);
- PreparedStatement prSt = null;
- Statement statement = connection.createStatement();
- String query;
- ResultSet rs;
-
- ResultSet rs1 = null;
- if (username.equals("all")) {
- } else {
-
-
- prSt = huedatabase.getUseridfromUserName(connection, username);
-
- rs = prSt.executeQuery();
-
- while (rs.next()) {
- id = rs.getInt("id");
- }
- }
-
- if (startdate.equals("") && endtime.equals("")) {
- if (username.equals("all")) {
- prSt = huedatabase.getQueriesNoStartDateNoEndDateAllUser(connection);
- } else {
- prSt = huedatabase.getQueriesNoStartDateNoEndDate(connection, id);
-
- }
-
- } else if ((startdate.equals("")) && !(endtime.equals(""))) {
- if (username.equals("all")) {
- prSt = huedatabase.getQueriesNoStartDateYesEndDateAllUser(connection, endtime);
- } else {
- prSt = huedatabase.getQueriesNoStartDateYesEndDate(connection, id, endtime);
-
- }
- } else if (!(startdate.equals("")) && (endtime.equals(""))) {
- if (username.equals("all")) {
- prSt = huedatabase.getQueriesYesStartDateNoEndDateAllUser(connection, startdate);
- } else {
- prSt = huedatabase.getQueriesYesStartDateNoEndDate(connection, id, startdate);
-
- }
-
- } else if (!(startdate.equals("")) && !(endtime.equals(""))) {
- if (username.equals("all")) {
- prSt = huedatabase.getQueriesYesStartDateYesEndDateAllUser(connection, startdate, endtime);
- } else {
- prSt = huedatabase.getQueriesYesStartDateYesEndDate(connection, id, startdate, endtime);
- }
- }
-
- rs1 = prSt.executeQuery();
-
- while (rs1.next()) {
- HiveModel hivepojo = new HiveModel();
- query = rs1.getString("query");
- hivepojo.setQuery(query);
- hiveArrayList.add(hivepojo);
- i++;
- }
-
- connection.commit();
-
- } catch (SQLException e) {
- connection.rollback();
-
- } finally {
- try {
- if (connection != null)
- connection.close();
- } catch (SQLException e) {
- logger.error("Sql exception error: " + e);
- }
- }
- return hiveArrayList;
-
- }
-
- public void writetoFileQueryhql(String content, String homedir) {
- try {
- File file = new File(homedir + "query.hql");
- // if file doesnt exists, then create it
- if (!file.exists()) {
- file.createNewFile();
- }
- FileWriter fw = new FileWriter(file.getAbsoluteFile());
- BufferedWriter bw = new BufferedWriter(fw);
- bw.write(content);
- bw.close();
- } catch (IOException e) {
- logger.error("IOException", e);
- }
-
- }
-
- public void deleteFileQueryhql(String homedir) {
- try {
- File file = new File(homedir + "query.hql");
-
- if (file.delete()) {
- logger.info("temporary hql file deleted");
- } else {
- logger.info("temporary hql file delete failed");
- }
-
- } catch (Exception e) {
-
- logger.error("File Exception ", e);
-
- }
-
- }
-
- public void deleteFileQueryLogs(String homedir) {
- try {
- File file = new File(homedir + "logs");
-
- if (file.delete()) {
- logger.info("temporary logs file deleted");
- } else {
- logger.info("temporary logs file delete failed");
- }
-
- } catch (Exception e) {
-
- logger.error("File Exception ", e);
-
- }
-
- }
-
- public void writetoFileLogs(String homedir) {
- try {
- String content = "";
- File file = new File(homedir + "logs");
- // if file doesnt exists, then create it
- if (!file.exists()) {
- file.createNewFile();
- }
- FileWriter fw = new FileWriter(file.getAbsoluteFile());
- BufferedWriter bw = new BufferedWriter(fw);
- bw.write(content);
- bw.close();
- } catch (IOException e) {
- logger.error("IOException", e);
- }
-
- }
-
- public void createDir(final String dir, final String namenodeuri) throws IOException,
- URISyntaxException {
-
- try {
- final Configuration conf = new Configuration();
-
- conf.set("fs.hdfs.impl",
- org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()
- );
- conf.set("fs.file.impl",
- org.apache.hadoop.fs.LocalFileSystem.class.getName()
- );
- conf.set("fs.defaultFS", namenodeuri);
- conf.set("hadoop.job.ugi", "hdfs");
- UserGroupInformation.setConfiguration(conf);
-
- UserGroupInformation ugi = UserGroupInformation.createRemoteUser("hdfs");
-
- ugi.doAs(new PrivilegedExceptionAction<Boolean>() {
-
- public Boolean run() throws Exception {
-
- FileSystem fs = FileSystem.get(conf);
- Path src = new Path(dir);
- Boolean b = fs.mkdirs(src);
- return b;
- }
- });
- } catch (Exception e) {
- logger.error("Exception in Webhdfs", e);
- }
- }
-
- public void createDirKerberorisedSecured(final String dir, final String namenodeuri) throws IOException,
- URISyntaxException {
-
- try {
- final Configuration conf = new Configuration();
-
- conf.set("fs.hdfs.impl",
- org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()
- );
- conf.set("fs.file.impl",
- org.apache.hadoop.fs.LocalFileSystem.class.getName()
- );
- conf.set("fs.defaultFS", namenodeuri);
- conf.set("hadoop.job.ugi", "hdfs");
- conf.set("hadoop.security.authentication", "Kerberos");
- UserGroupInformation.setConfiguration(conf);
- UserGroupInformation ugi = UserGroupInformation.createRemoteUser("hdfs");
- ugi.doAs(new PrivilegedExceptionAction<Boolean>() {
-
- public Boolean run() throws Exception {
- FileSystem fs = FileSystem.get(conf);
- Path src = new Path(dir);
- Boolean b = fs.mkdirs(src);
- return b;
- }
- });
- } catch (Exception e) {
- logger.error("Exception in Webhdfs", e);
- }
- }
-
-
- public void putFileinHdfs(final String source, final String dest, final String namenodeuri)
- throws IOException {
-
- try {
- UserGroupInformation ugi = UserGroupInformation.createRemoteUser("hdfs");
-
- ugi.doAs(new PrivilegedExceptionAction<Void>() {
-
- public Void run() throws Exception {
-
- Configuration conf = new Configuration();
- conf.set("fs.defaultFS", namenodeuri);
- conf.set("hadoop.job.ugi", "hdfs");
- conf.set("fs.hdfs.impl",
- org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()
- );
- conf.set("fs.file.impl",
- org.apache.hadoop.fs.LocalFileSystem.class.getName()
- );
- FileSystem fileSystem = FileSystem.get(conf);
-
- String filename = source.substring(
- source.lastIndexOf('/') + 1, source.length());
- String dest1;
- if (dest.charAt(dest.length() - 1) != '/') {
- dest1 = dest + "/" + filename;
- } else {
- dest1 = dest + filename;
- }
-
- Path path = new Path(dest1);
- if (fileSystem.exists(path)) {
-
- }
- // Path pathsource = new Path(source);
- FSDataOutputStream out = fileSystem.create(path);
-
- InputStream in = new BufferedInputStream(
- new FileInputStream(new File(source)));
-
- byte[] b = new byte[1024];
- int numBytes = 0;
- while ((numBytes = in.read(b)) > 0) {
- out.write(b, 0, numBytes);
- }
- in.close();
- out.close();
- fileSystem.close();
- return null;
- }
- });
- } catch (Exception e) {
- logger.error("Webhdfs exception", e);
- }
-
- }
-
- public void putFileinHdfsKerborizedSecured(final String source, final String dest, final String namenodeuri)
- throws IOException {
-
- try {
-
- final Configuration conf = new Configuration();
-
- conf.set("fs.hdfs.impl",
- org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()
- );
- conf.set("fs.file.impl",
- org.apache.hadoop.fs.LocalFileSystem.class.getName()
- );
- conf.set("fs.defaultFS", namenodeuri);
- conf.set("hadoop.job.ugi", "hdfs");
- conf.set("hadoop.security.authentication", "Kerberos");
- UserGroupInformation.setConfiguration(conf);
- UserGroupInformation ugi = UserGroupInformation.createRemoteUser("hdfs");
-
- ugi.doAs(new PrivilegedExceptionAction<Void>() {
-
- public Void run() throws Exception {
-
- FileSystem fileSystem = FileSystem.get(conf);
-
- String filename = source.substring(
- source.lastIndexOf('/') + 1, source.length());
- String dest1;
- if (dest.charAt(dest.length() - 1) != '/') {
- dest1 = dest + "/" + filename;
- } else {
- dest1 = dest + filename;
- }
-
- Path path = new Path(dest1);
- if (fileSystem.exists(path)) {
-
- }
-
- FSDataOutputStream out = fileSystem.create(path);
-
- InputStream in = new BufferedInputStream(
- new FileInputStream(new File(source)));
-
- byte[] b = new byte[1024];
- int numBytes = 0;
- while ((numBytes = in.read(b)) > 0) {
- out.write(b, 0, numBytes);
- }
- in.close();
- out.close();
- fileSystem.close();
- return null;
- }
- });
- } catch (Exception e) {
- logger.error("Webhdfs exception", e);
-
- }
-
- }
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryStartJob.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryStartJob.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryStartJob.java
deleted file mode 100644
index 5228bf6..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/historyquery/HiveHistoryStartJob.java
+++ /dev/null
@@ -1,74 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-package org.apache.ambari.view.huetoambarimigration.migration.hive.historyquery;
-
-import org.apache.ambari.view.ViewContext;
-import org.apache.ambari.view.huetoambarimigration.persistence.utils.ItemNotFound;
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.models.MigrationResponse;
-import org.json.simple.JSONObject;
-
-import java.io.IOException;
-
-
-public class HiveHistoryStartJob extends Thread{
-
- String username;
- String instance;
- String startdate;
- String enddate;
- String jobid;
- ViewContext view;
-
- public HiveHistoryStartJob(String username, String instance, String startdate, String enddate, String jobid, ViewContext view) {
- this.username = username;
- this.instance=instance;
- this.startdate=startdate;
- this.enddate=enddate;
- this.jobid=jobid;
- this.view=view;
- }
-
- @Override
- public void run() {
-
- MigrationResponse migrationresult=new MigrationResponse();
-
- migrationresult.setId(jobid);
- migrationresult.setIntanceName(instance);
- migrationresult.setUserNameofhue(username);
- migrationresult.setProgressPercentage(0);
-
- JSONObject response = new JSONObject();
-
- /**
- * creating a separete thread
- */
-
- HiveHistoryMigrationUtility hivehistoryquery=new HiveHistoryMigrationUtility();
- try {
- hivehistoryquery.hiveHistoryQueryMigration(username,instance,startdate,enddate,view,migrationresult,jobid);
- }
- catch (IOException e) {
- e.printStackTrace();
- } catch (ItemNotFound itemNotFound) {
- itemNotFound.printStackTrace();
- }
-
- }
-
-}
http://git-wip-us.apache.org/repos/asf/ambari/blob/283256c8/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/savedquery/HiveSavedQueryMigrationImplementation.java
----------------------------------------------------------------------
diff --git a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/savedquery/HiveSavedQueryMigrationImplementation.java b/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/savedquery/HiveSavedQueryMigrationImplementation.java
deleted file mode 100644
index 7bd48b2..0000000
--- a/contrib/views/hueambarimigration/src/main/java/org/apache/ambari/view/huetoambarimigration/migration/hive/savedquery/HiveSavedQueryMigrationImplementation.java
+++ /dev/null
@@ -1,673 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- * <p/>
- * http://www.apache.org/licenses/LICENSE-2.0
- * <p/>
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.ambari.view.huetoambarimigration.migration.hive.savedquery;
-
-import org.apache.ambari.view.huetoambarimigration.resources.scripts.models.HiveModel;
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.ambariqueryset.hive.savedqueryset.QuerySetAmbariDB;
-import org.apache.ambari.view.huetoambarimigration.datasource.queryset.huequeryset.hive.savedqueryset.QuerySet;
-import org.apache.ambari.view.huetoambarimigration.migration.configuration.ConfigurationCheckImplementation;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.FSDataOutputStream;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.security.UserGroupInformation;
-import org.apache.log4j.Logger;
-import org.jdom.Attribute;
-import org.jdom.Document;
-import org.jdom.Element;
-import org.jdom.JDOMException;
-import org.jdom.input.SAXBuilder;
-import org.jdom.output.Format;
-import org.jdom.output.XMLOutputter;
-import org.json.JSONObject;
-
-import java.io.*;
-import java.net.URISyntaxException;
-import java.nio.charset.Charset;
-import java.security.PrivilegedExceptionAction;
-import java.sql.*;
-import java.text.ParseException;
-import java.text.SimpleDateFormat;
-import java.util.ArrayList;
-import java.util.Calendar;
-import java.util.Date;
-import java.util.GregorianCalendar;
-
-public class HiveSavedQueryMigrationImplementation {
-
- static final Logger logger = Logger.getLogger(HiveSavedQueryMigrationImplementation.class);
-
- private static String readAll(Reader rd) throws IOException {
- StringBuilder sb = new StringBuilder();
- int cp;
- while ((cp = rd.read()) != -1) {
- sb.append((char) cp);
- }
- return sb.toString();
- }
-
- public void wrtitetoalternatesqlfile(String dirname, String content,
- String instance, int i) throws IOException {
-
- Date dNow = new Date();
- SimpleDateFormat ft = new SimpleDateFormat("YYYY-MM-dd hh:mm:ss");
- String currentDate = ft.format(dNow);
-
- XMLOutputter xmlOutput = new XMLOutputter();
-
- xmlOutput.setFormat(Format.getPrettyFormat());
-
- File xmlfile = new File(ConfigurationCheckImplementation.getHomeDir() + "RevertChangesService.xml");
-
- if (xmlfile.exists()) {
- String iteration = Integer.toString(i + 1);
- SAXBuilder builder = new SAXBuilder();
- Document doc;
- try {
- doc = (Document) builder.build(xmlfile);
-
- Element rootNode = doc.getRootElement();
-
- Element record = new Element("RevertRecord");
- record.setAttribute(new Attribute("id", iteration));
- record.addContent(new Element("datetime").setText(currentDate.toString()));
- record.addContent(new Element("dirname").setText(dirname));
- record.addContent(new Element("instance").setText(instance));
- record.addContent(new Element("query").setText(content));
-
- rootNode.addContent(record);
- xmlOutput.output(doc, new FileWriter(ConfigurationCheckImplementation.getHomeDir() + "RevertChangesService.xml"));
-
- } catch (JDOMException e) {
- // TODO Auto-generated catch block
- logger.error("JDOMException: ", e);
- }
-
- } else {
-
- try {
- String iteration = Integer.toString(i + 1);
- Element revertrecord = new Element("RevertChangePage");
- Document doc = new Document(revertrecord);
- doc.setRootElement(revertrecord);
-
- Element record = new Element("RevertRecord");
- record.setAttribute(new Attribute("id", iteration));
- record.addContent(new Element("datetime").setText(currentDate
- .toString()));
- record.addContent(new Element("dirname").setText(dirname));
- record.addContent(new Element("instance").setText(instance));
- record.addContent(new Element("query").setText(content));
-
- doc.getRootElement().addContent(record);
-
- xmlOutput.output(doc, new FileWriter(ConfigurationCheckImplementation.getHomeDir() + "RevertChangesService.xml"));
-
- } catch (IOException io) {
-
- }
-
- }
-
- }
-
- public int fetchMaxidforSavedQueryHive(Connection c, int id, QuerySetAmbariDB ambaridatabase) throws SQLException {
-
- String ds_id = null;
- ResultSet rs = null;
- PreparedStatement prSt = null;
-
- prSt = ambaridatabase.getMaxDsIdFromTableIdSavedquery(c, id);
-
- rs = prSt.executeQuery();
-
- while (rs.next()) {
- ds_id = rs.getString("max");
- }
-
- int num;
- if (ds_id == null) {
- num = 1;
- } else {
- num = Integer.parseInt(ds_id);
- }
- return num;
- }
-
- public int fetchInstancetablenameForSavedqueryHive(Connection c, String instance, QuerySetAmbariDB ambaridatabase) throws SQLException {
-
- String ds_id = new String();
- int id = 0;
- Statement stmt = null;
- PreparedStatement prSt = null;
-
-
- ResultSet rs = null;
-
-
- prSt = ambaridatabase.getTableIdFromInstanceNameSavedquery(c, instance);
-
- logger.info("sql statement to fetch is from ambari instance:= = " + prSt);
-
- rs = prSt.executeQuery();
-
- while (rs.next()) {
- id = rs.getInt("id");
- }
- return id;
- }
-
- public int fetchInstanceTablenameHiveHistory(Connection c, String instance, QuerySetAmbariDB ambaridatabase) throws SQLException {
- String ds_id = new String();
- int id = 0;
- Statement stmt = null;
- PreparedStatement prSt = null;
-
-
- ResultSet rs = null;
-
-
- prSt = ambaridatabase.getTableIdFromInstanceNameHistoryquery(c, instance);
-
- logger.info("sql statement to fetch is from ambari instance:= = " + prSt);
-
- rs = prSt.executeQuery();
-
- while (rs.next()) {
- id = rs.getInt("id");
- }
- return id;
-
- }
-
- public int fetchMaxdsidFromHiveHistory(Connection c, int id, QuerySetAmbariDB ambaridatabase)
- throws SQLException {
-
- String ds_id = null;
- ResultSet rs = null;
- PreparedStatement prSt = null;
-
- prSt = ambaridatabase.getMaxDsIdFromTableIdHistoryquery(c, id);
-
- rs = prSt.executeQuery();
-
- while (rs.next()) {
- ds_id = rs.getString("max");
- }
-
- int num;
- if (ds_id == null) {
- num = 1;
- } else {
- num = Integer.parseInt(ds_id);
- }
- return num;
- }
-
-
- /**/
- public void insertRowHiveHistory(String dirname, int maxcount, long epochtime, Connection c, int id, String instance, int i, QuerySetAmbariDB ambaridatabase)
- throws SQLException, IOException {
-
- String maxcount1 = Integer.toString(maxcount);
- String epochtime1 = Long.toString(epochtime);
- PreparedStatement prSt = null;
- String revsql = null;
-
- prSt = ambaridatabase.insertToHiveHistory(c, id, maxcount1, epochtime, dirname);
-
- System.out.println("the actual query is " + prSt);
-
- logger.info("The actual insert statement is " + prSt);
-
- prSt.executeUpdate();
-
- revsql = ambaridatabase.revertSqlHistoryQuery(id, maxcount1);
-
- logger.info("adding revert sqlsavedquery in hivehistory ");
-
- wrtitetoalternatesqlfile(dirname, revsql, instance, i);
- }
-
- public void insertRowinSavedQuery(int maxcount, String database, String dirname, String query, String name, Connection c, int id, String instance, int i, QuerySetAmbariDB ambaridatabase) throws SQLException, IOException {
-
- String maxcount1 = Integer.toString(maxcount);
- String revsql = null;
-
- PreparedStatement prSt = null;
-
- prSt = ambaridatabase.insertToHiveSavedQuery(c, id, maxcount1, database, dirname, query, name);
-
- System.out.println("the actual query is " + prSt);
-
- logger.info("The actual insert statement is " + prSt);
-
- prSt.executeUpdate();
-
- revsql = ambaridatabase.revertSqlSavedQuery(id, maxcount1);
-
- logger.info("adding revert sqlsavedquery ");
-
- wrtitetoalternatesqlfile(dirname, revsql, instance, i);
-
- }
-
- public long getEpochTime() throws ParseException {
-
- long seconds = System.currentTimeMillis() / 1000l;
- return seconds;
-
- }
-
- public String getTime() throws ParseException {
- int day, month, year;
- int second, minute, hour;
- int milisecond;
- GregorianCalendar date = new GregorianCalendar();
-
- day = date.get(Calendar.DAY_OF_MONTH);
- month = date.get(Calendar.MONTH);
- year = date.get(Calendar.YEAR);
-
- second = date.get(Calendar.SECOND);
- minute = date.get(Calendar.MINUTE);
- hour = date.get(Calendar.HOUR);
- milisecond = date.get(Calendar.MILLISECOND);
-
- String s = year + "-" + (month + 1) + "-" + day + "_" + hour + "-"
- + minute;
- String s1 = year + "-" + (month + 1) + "-" + day + "_" + hour + "-"
- + minute + "-" + second + "-" + milisecond;
- SimpleDateFormat df = new SimpleDateFormat("yyyy-MM-dd_HH-mm-ss-SSS");
- Date date1 = df.parse(s1);
- long epoch = date1.getTime();
-
- return s;
-
- }
-
- public ArrayList<HiveModel> fetchFromHuedb(String username, String startdate, String endtime, Connection connection, QuerySet huedatabase)
- throws ClassNotFoundException, IOException {
- int id = 0;
- int i = 0;
- String[] query = new String[100];
- ArrayList<HiveModel> hiveArrayList = new ArrayList<HiveModel>();
- ResultSet rs1 = null;
-
- try {
- Statement statement = connection.createStatement();
- connection.setAutoCommit(false);
- PreparedStatement prSt = null;
- ResultSet rs;
- if (username.equals("all")) {
- } else {
-
- prSt = huedatabase.getUseridfromUserName(connection, username);
-
- rs = prSt.executeQuery();
-
- while (rs.next()) {
- id = rs.getInt("id");
- }
- }
-
- if (startdate.equals("") && endtime.equals("")) {
- if (username.equals("all")) {
- prSt = huedatabase.getQueriesNoStartDateNoEndDateAllUser(connection);
- } else {
- prSt = huedatabase.getQueriesNoStartDateNoEndDate(connection, id);
-
- }
-
- } else if ((startdate.equals("")) && !(endtime.equals(""))) {
- if (username.equals("all")) {
- prSt = huedatabase.getQueriesNoStartDateYesEndDateAllUser(connection, endtime);
- } else {
- prSt = huedatabase.getQueriesNoStartDateYesEndDate(connection, id, endtime);
-
- }
- } else if (!(startdate.equals("")) && (endtime.equals(""))) {
- if (username.equals("all")) {
- prSt = huedatabase.getQueriesYesStartDateNoEndDateAllUser(connection, startdate);
- } else {
- prSt = huedatabase.getQueriesYesStartDateNoEndDate(connection, id, startdate);
-
- }
-
- } else if (!(startdate.equals("")) && !(endtime.equals(""))) {
- if (username.equals("all")) {
- prSt = huedatabase.getQueriesYesStartDateYesEndDateAllUser(connection, startdate, endtime);
- } else {
- prSt = huedatabase.getQueriesYesStartDateYesEndDate(connection, id, startdate, endtime);
- }
-
- }
-
- rs1 = prSt.executeQuery();
-
-
- while (rs1.next()) {
- HiveModel hivepojo = new HiveModel();
- String name = rs1.getString("name");
- String temp = rs1.getString("data");
- InputStream is = new ByteArrayInputStream(temp.getBytes());
- BufferedReader rd = new BufferedReader(new InputStreamReader(
- is, Charset.forName("UTF-8")));
- String jsonText = readAll(rd);
- JSONObject json = new JSONObject(jsonText);
- String resources = json.get("query").toString();
- json = new JSONObject(resources);
-
- String resarr = (json.get("query")).toString();
-
- json = new JSONObject(resources);
- String database = (json.get("database")).toString();
- hivepojo.setQuery(resarr);
- hivepojo.setDatabase(database);
- hivepojo.setOwner(name);
- hiveArrayList.add(hivepojo);
- i++;
- }
-
- } catch (SQLException e2) {
- e2.printStackTrace();
- } finally
-
- {
- try {
- if (connection != null)
- connection.close();
- } catch (SQLException e) {
- logger.error("sql connection exception", e);
- }
- }
-
- return hiveArrayList;
-
- }
-
-
- public void writetoFilequeryHql(String content, String homedir) {
- try {
- File file = new File(homedir + "query.hql");
- if (!file.exists()) {
- file.createNewFile();
- }
-
- FileWriter fw = new FileWriter(file.getAbsoluteFile());
- BufferedWriter bw = new BufferedWriter(fw);
- bw.write(content);
- bw.close();
-
- } catch (IOException e) {
- logger.error("IOException: ", e);
- }
-
- }
-
- public void deleteFileQueryhql(String homedir) {
- try {
- File file = new File(homedir + "query.hql");
-
- if (file.delete()) {
- logger.info("temporary hql file deleted");
- } else {
- logger.info("temporary hql file delete failed");
- }
-
- } catch (Exception e) {
-
- logger.error("File Exception ", e);
-
- }
-
- }
-
- public void deleteFileQueryLogs(String homedir) {
- try {
- File file = new File(homedir + "logs");
-
- if (file.delete()) {
- logger.info("temporary logs file deleted");
- } else {
- logger.info("temporary logs file delete failed");
- }
-
- } catch (Exception e) {
-
- logger.error("File Exception ", e);
-
- }
-
- }
-
-
- public void writetoFileLogs(String homedir) {
- try {
-
- String content = "";
- File file = new File(homedir + "logs");
-
- // if file doesnt exists, then create it
- if (!file.exists()) {
- file.createNewFile();
- }
-
- FileWriter fw = new FileWriter(file.getAbsoluteFile());
- BufferedWriter bw = new BufferedWriter(fw);
- bw.write(content);
- bw.close();
-
- } catch (IOException e) {
- logger.error("IOException: ", e);
- }
-
- }
-
- public void createDirHive(final String dir, final String namenodeuri)
- throws IOException, URISyntaxException {
-
- try {
- final Configuration conf = new Configuration();
-
- conf.set("fs.hdfs.impl",
- org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()
- );
- conf.set("fs.file.impl",
- org.apache.hadoop.fs.LocalFileSystem.class.getName()
- );
- conf.set("fs.defaultFS", namenodeuri);
- conf.set("hadoop.job.ugi", "hdfs");
- conf.set("hadoop.security.authentication", "Kerberos");
-
- UserGroupInformation.setConfiguration(conf);
- UserGroupInformation ugi = UserGroupInformation.createRemoteUser("hdfs");
-
- ugi.doAs(new PrivilegedExceptionAction<Void>() {
-
- public Void run() throws Exception {
-
- FileSystem fs = FileSystem.get(conf);
- Path src = new Path(dir);
- fs.mkdirs(src);
- return null;
- }
- });
- } catch (Exception e) {
- logger.error("Webhdfs: ", e);
- }
- }
-
- public void createDirHiveSecured(final String dir, final String namenodeuri)
- throws IOException, URISyntaxException {
-
- try {
- final Configuration conf = new Configuration();
-
- conf.set("fs.hdfs.impl",
- org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()
- );
- conf.set("fs.file.impl",
- org.apache.hadoop.fs.LocalFileSystem.class.getName()
- );
- conf.set("fs.defaultFS", namenodeuri);
- conf.set("hadoop.job.ugi", "hdfs");
- conf.set("hadoop.security.authentication", "Kerberos");
-
- UserGroupInformation.setConfiguration(conf);
- UserGroupInformation ugi = UserGroupInformation.createRemoteUser("hdfs");
-
- ugi.doAs(new PrivilegedExceptionAction<Void>() {
-
- public Void run() throws Exception {
-
- FileSystem fs = FileSystem.get(conf);
- Path src = new Path(dir);
- fs.mkdirs(src);
- return null;
- }
- });
- } catch (Exception e) {
- logger.error("Webhdfs: ", e);
- }
- }
-
- public void putFileinHdfs(final String source, final String dest,
- final String namenodeuri) throws IOException {
-
- try {
- final Configuration conf = new Configuration();
-
- conf.set("fs.hdfs.impl",
- org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()
- );
- conf.set("fs.file.impl",
- org.apache.hadoop.fs.LocalFileSystem.class.getName()
- );
- conf.set("fs.defaultFS", namenodeuri);
- conf.set("hadoop.job.ugi", "hdfs");
- conf.set("hadoop.security.authentication", "Kerberos");
-
- UserGroupInformation.setConfiguration(conf);
- UserGroupInformation ugi = UserGroupInformation.createRemoteUser("hdfs");
- ugi.doAs(new PrivilegedExceptionAction<Void>() {
-
- public Void run() throws Exception {
-
- FileSystem fileSystem = FileSystem.get(conf);
- String filename = source.substring(
- source.lastIndexOf('/') + 1, source.length());
- String dest1;
- if (dest.charAt(dest.length() - 1) != '/') {
- dest1 = dest + "/" + filename;
- } else {
- dest1 = dest + filename;
- }
-
- Path path = new Path(dest1);
- if (fileSystem.exists(path)) {
-
- }
- FSDataOutputStream out = fileSystem.create(path);
-
- InputStream in = new BufferedInputStream(
- new FileInputStream(new File(source)));
-
- byte[] b = new byte[1024];
- int numBytes = 0;
- while ((numBytes = in.read(b)) > 0) {
- out.write(b, 0, numBytes);
- }
- in.close();
- out.close();
- fileSystem.close();
- return null;
- }
- });
- } catch (Exception e) {
- logger.error("Webhdfs exception", e);
- }
-
- }
-
-
- public void putFileinHdfsSecured(final String source, final String dest,
- final String namenodeuri) throws IOException {
-
- try {
- final Configuration conf = new Configuration();
-
- conf.set("fs.hdfs.impl",
- org.apache.hadoop.hdfs.DistributedFileSystem.class.getName()
- );
- conf.set("fs.file.impl",
- org.apache.hadoop.fs.LocalFileSystem.class.getName()
- );
- conf.set("fs.defaultFS", namenodeuri);
- conf.set("hadoop.job.ugi", "hdfs");
- conf.set("hadoop.security.authentication", "Kerberos");
-
- UserGroupInformation.setConfiguration(conf);
- UserGroupInformation ugi = UserGroupInformation.createRemoteUser("hdfs");
- ugi.doAs(new PrivilegedExceptionAction<Void>() {
-
- public Void run() throws Exception {
-
-
- FileSystem fileSystem = FileSystem.get(conf);
-
- String filename = source.substring(
- source.lastIndexOf('/') + 1, source.length());
- String dest1;
- if (dest.charAt(dest.length() - 1) != '/') {
- dest1 = dest + "/" + filename;
- } else {
- dest1 = dest + filename;
- }
-
- Path path = new Path(dest1);
- if (fileSystem.exists(path)) {
-
- }
- // Path pathsource = new Path(source);
- FSDataOutputStream out = fileSystem.create(path);
-
- InputStream in = new BufferedInputStream(
- new FileInputStream(new File(source)));
-
- byte[] b = new byte[1024];
- int numBytes = 0;
- while ((numBytes = in.read(b)) > 0) {
- out.write(b, 0, numBytes);
- }
- in.close();
- out.close();
- fileSystem.close();
-
-
- return null;
- }
- });
- } catch (Exception e) {
- logger.error("Webhdfs exception", e);
- }
-
- }
-
-}