You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@druid.apache.org by ka...@apache.org on 2023/03/16 05:34:50 UTC

[druid] branch master updated: Update container creation in AzureTestUtil.java (#13911)

This is an automated email from the ASF dual-hosted git repository.

karan pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/druid.git


The following commit(s) were added to refs/heads/master by this push:
     new c7d864d3bc Update container creation in AzureTestUtil.java (#13911)
c7d864d3bc is described below

commit c7d864d3bccf82601c757f44c859eaeae34cfc1d
Author: abhagraw <99...@users.noreply.github.com>
AuthorDate: Thu Mar 16 11:04:43 2023 +0530

    Update container creation in AzureTestUtil.java (#13911)
    
    *
    1. Handling deletion/creation of container created during the previously run test in AzureTestUtil.java.
    2. Adding/updating log messages and comments in Azure and GCS deep storage tests.
---
 .github/workflows/reusable-revised-its.yml          |  2 +-
 .github/workflows/revised-its.yml                   |  2 +-
 .../AbstractAzureInputSourceParallelIndexTest.java  |  1 +
 .../indexer/ITGcsToGcsParallelIndexTest.java        | 14 ++++++--------
 .../testsEx/msq/ITAzureSQLBasedIngestionTest.java   |  4 ++--
 .../testsEx/msq/ITGcsSQLBasedIngestionTest.java     |  9 +++++++++
 .../apache/druid/testsEx/utils/AzureTestUtil.java   | 21 ++++++++++++++++++++-
 7 files changed, 40 insertions(+), 13 deletions(-)

diff --git a/.github/workflows/reusable-revised-its.yml b/.github/workflows/reusable-revised-its.yml
index 91ff249b2c..5a220d5755 100644
--- a/.github/workflows/reusable-revised-its.yml
+++ b/.github/workflows/reusable-revised-its.yml
@@ -48,7 +48,7 @@ env:
   SEGMENT_DOWNLOAD_TIMEOUT_MINS: 5
 
 jobs:
-  test: # Github job that runs a given revised/new IT against retrieved cached druid docker image
+  test: # GitHub job that runs a given revised/new IT against retrieved cached druid docker image
     name: ${{ inputs.it }} integration test (Compile=jdk${{ inputs.build_jdk }}, Run=jdk${{ inputs.runtime_jdk }}, Indexer=${{ inputs.use_indexer }})
     runs-on: ubuntu-22.04
     steps:
diff --git a/.github/workflows/revised-its.yml b/.github/workflows/revised-its.yml
index e4cf409ee6..83cad23f42 100644
--- a/.github/workflows/revised-its.yml
+++ b/.github/workflows/revised-its.yml
@@ -13,7 +13,7 @@
 # See the License for the specific language governing permissions and
 # limitations under the License.
 
-# Github workflow that runs revised/new ITs
+# GitHub workflow that runs revised/new ITs
 on:
   workflow_call:
   workflow_dispatch:
diff --git a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/indexer/AbstractAzureInputSourceParallelIndexTest.java b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/indexer/AbstractAzureInputSourceParallelIndexTest.java
index 3098956a8c..81852df11c 100644
--- a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/indexer/AbstractAzureInputSourceParallelIndexTest.java
+++ b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/indexer/AbstractAzureInputSourceParallelIndexTest.java
@@ -42,6 +42,7 @@ public class AbstractAzureInputSourceParallelIndexTest extends AbstractCloudInpu
   public static void uploadDataFilesToAzure()
   {
     try {
+      LOG.info("Uploading files to Azure");
       azure = new AzureTestUtil();
       // Creating a container with name set in AZURE_CONTAINER env variable.
       azure.createStorageContainer();
diff --git a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/indexer/ITGcsToGcsParallelIndexTest.java b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/indexer/ITGcsToGcsParallelIndexTest.java
index 0eaef4e64c..ca202797bd 100644
--- a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/indexer/ITGcsToGcsParallelIndexTest.java
+++ b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/indexer/ITGcsToGcsParallelIndexTest.java
@@ -31,15 +31,13 @@ import java.util.List;
 
 /**
  * IMPORTANT:
- * To run this test, you must:
- * 1) Set the bucket and path for your data. This can be done by setting -Ddruid.test.config.cloudBucket and
- *    -Ddruid.test.config.cloudPath or setting "cloud_bucket" and "cloud_path" in the config file.
- * 2) Copy wikipedia_index_data1.json, wikipedia_index_data2.json, and wikipedia_index_data3.json
- *    located in integration-tests/src/test/resources/data/batch_index/json to your GCS at the location set in step 1.
- * 3) Provide -Doverride.config.path=<PATH_TO_FILE> with gcs configs set. See
- *    integration-tests/docker/environment-configs/override-examples/gcs for env vars to provide.
- * 4) Provide -Dresource.file.dir.path=<PATH_TO_FOLDER> with folder that contains GOOGLE_APPLICATION_CREDENTIALS file
+ * To run this test, you must set the following env variables in the build environment -
+ * GOOGLE_PREFIX - path inside the bucket where the test data files will be uploaded
+ * GOOGLE_BUCKET - Google cloud bucket name
+ * GOOGLE_APPLICATION_CREDENTIALS - path to the json file containing google cloud credentials
+ * <a href="https://druid.apache.org/docs/latest/development/extensions-core/google.html">Google Cloud Storage setup in druid</a>
  */
+
 @RunWith(DruidTestRunner.class)
 @Category(GcsDeepStorage.class)
 public class ITGcsToGcsParallelIndexTest extends AbstractGcsInputSourceParallelIndexTest
diff --git a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/msq/ITAzureSQLBasedIngestionTest.java b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/msq/ITAzureSQLBasedIngestionTest.java
index c54e891696..54f207704a 100644
--- a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/msq/ITAzureSQLBasedIngestionTest.java
+++ b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/msq/ITAzureSQLBasedIngestionTest.java
@@ -22,7 +22,7 @@ package org.apache.druid.testsEx.msq;
 import junitparams.Parameters;
 import junitparams.naming.TestCaseName;
 import org.apache.druid.java.util.common.Pair;
-import org.apache.druid.testsEx.categories.S3DeepStorage;
+import org.apache.druid.testsEx.categories.AzureDeepStorage;
 import org.apache.druid.testsEx.config.DruidTestRunner;
 import org.apache.druid.testsEx.indexer.AbstractAzureInputSourceParallelIndexTest;
 import org.junit.Test;
@@ -41,7 +41,7 @@ import java.util.List;
  */
 
 @RunWith(DruidTestRunner.class)
-@Category(S3DeepStorage.class)
+@Category(AzureDeepStorage.class)
 public class ITAzureSQLBasedIngestionTest extends AbstractAzureInputSourceParallelIndexTest
 {
   private static final String CLOUD_INGEST_SQL = "/multi-stage-query/wikipedia_cloud_index_msq.sql";
diff --git a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/msq/ITGcsSQLBasedIngestionTest.java b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/msq/ITGcsSQLBasedIngestionTest.java
index a6031c6e1f..125b30ef81 100644
--- a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/msq/ITGcsSQLBasedIngestionTest.java
+++ b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/msq/ITGcsSQLBasedIngestionTest.java
@@ -31,6 +31,15 @@ import org.junit.runner.RunWith;
 
 import java.util.List;
 
+/**
+ * IMPORTANT:
+ * To run this test, you must set the following env variables in the build environment -
+ * GOOGLE_PREFIX - path inside the bucket where the test data files will be uploaded
+ * GOOGLE_BUCKET - Google cloud bucket name
+ * GOOGLE_APPLICATION_CREDENTIALS - path to the json file containing google cloud credentials
+ * <a href="https://druid.apache.org/docs/latest/development/extensions-core/google.html">Google Cloud Storage setup in druid</a>
+ */
+
 @RunWith(DruidTestRunner.class)
 @Category(GcsDeepStorage.class)
 public class ITGcsSQLBasedIngestionTest extends AbstractGcsInputSourceParallelIndexTest
diff --git a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/utils/AzureTestUtil.java b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/utils/AzureTestUtil.java
index 4b31206a1d..90be0cbd46 100644
--- a/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/utils/AzureTestUtil.java
+++ b/integration-tests-ex/cases/src/test/java/org/apache/druid/testsEx/utils/AzureTestUtil.java
@@ -25,6 +25,7 @@ import com.microsoft.azure.storage.blob.CloudBlobClient;
 import com.microsoft.azure.storage.blob.CloudBlobContainer;
 import com.microsoft.azure.storage.blob.CloudBlockBlob;
 import org.apache.druid.java.util.common.logger.Logger;
+import org.apache.druid.testing.utils.ITRetryUtil;
 
 import java.io.File;
 import java.io.IOException;
@@ -83,9 +84,26 @@ public class AzureTestUtil
 
   public void createStorageContainer() throws URISyntaxException, StorageException
   {
+    LOG.info("Creating azure container " + AZURE_CONTAINER);
     CloudBlobContainer container = azureStorageClient.getContainerReference(AZURE_CONTAINER);
     // Create the container if it does not exist.
-    container.createIfNotExists();
+
+    // From the azure documentation -
+    // When a container is deleted, a container with the same name can't be created for at least 30 seconds.
+    // The container might not be available for more than 30 seconds if the service is still processing the request.
+    // While the container is being deleted, attempts to create a container of the same name fail with status
+    // code 409 (Conflict). The service indicates that the container is being deleted.
+    // All other operations, including operations on any blobs under the container,
+    // fail with status code 404 (Not Found) while the container is being deleted.
+    ITRetryUtil.retryUntil(
+        () -> container.createIfNotExists(),
+        true,
+        10000,
+        13,
+        "Create Azure container : " + AZURE_CONTAINER + " "
+    );
+
+    LOG.info("Azure container " + AZURE_CONTAINER + " created");
   }
 
   public void deleteStorageContainer() throws URISyntaxException, StorageException
@@ -109,6 +127,7 @@ public class AzureTestUtil
     // Create or overwrite the "myimage.jpg" blob with contents from a local file.
     File source = new File(filePath);
     CloudBlockBlob blob = container.getBlockBlobReference(DRUID_CLOUD_PATH + '/' + source.getName());
+    LOG.info("Uploading file " + DRUID_CLOUD_PATH + '/' + source.getName() + " in azure container " + AZURE_CONTAINER);
     blob.upload(Files.newInputStream(source.toPath()), source.length());
   }
 }


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@druid.apache.org
For additional commands, e-mail: commits-help@druid.apache.org