You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by xy...@apache.org on 2019/08/20 23:47:32 UTC
[hadoop] 01/02: Revert "HDDS-1965. Compile error due to leftover
ScmBlockLocationTestIngClient file (#1293)"
This is an automated email from the ASF dual-hosted git repository.
xyao pushed a commit to branch trunk
in repository https://gitbox.apache.org/repos/asf/hadoop.git
commit 7653ebdbb21e0f489cb113e6b878e40aa54c3b3a
Author: Xiaoyu Yao <xy...@apache.org>
AuthorDate: Tue Aug 20 16:14:12 2019 -0700
Revert "HDDS-1965. Compile error due to leftover ScmBlockLocationTestIngClient file (#1293)"
This reverts commit 83e452eceac63559c2f5146510ae3e89e310ac1e.
---
.../ozone/om/ScmBlockLocationTestIngClient.java | 195 +++++++++++++++++++++
1 file changed, 195 insertions(+)
diff --git a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ScmBlockLocationTestIngClient.java b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ScmBlockLocationTestIngClient.java
new file mode 100644
index 0000000..982e87e
--- /dev/null
+++ b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/ScmBlockLocationTestIngClient.java
@@ -0,0 +1,195 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ *
+ */
+
+package org.apache.hadoop.ozone.om;
+
+import org.apache.commons.lang3.StringUtils;
+import org.apache.hadoop.hdds.client.BlockID;
+import org.apache.hadoop.hdds.client.ContainerBlockID;
+import org.apache.hadoop.hdds.protocol.DatanodeDetails;
+import org.apache.hadoop.hdds.protocol.proto.HddsProtos;
+import org.apache.hadoop.hdds.scm.ScmInfo;
+import org.apache.hadoop.hdds.scm.TestUtils;
+import org.apache.hadoop.hdds.scm.container.common.helpers.AllocatedBlock;
+import org.apache.hadoop.hdds.scm.container.common.helpers.DeleteBlockResult;
+import org.apache.hadoop.hdds.scm.container.common.helpers.ExcludeList;
+import org.apache.hadoop.hdds.scm.pipeline.Pipeline;
+import org.apache.hadoop.hdds.scm.pipeline.PipelineID;
+import org.apache.hadoop.hdds.scm.protocol.ScmBlockLocationProtocol;
+import org.apache.hadoop.ozone.common.BlockGroup;
+import org.apache.hadoop.ozone.common.DeleteBlockGroupResult;
+import org.apache.hadoop.util.Time;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.Collections;
+import java.util.List;
+import java.util.UUID;
+
+import static org.apache.hadoop.hdds.protocol.proto
+ .ScmBlockLocationProtocolProtos.DeleteScmBlockResult.Result;
+import static org.apache.hadoop.hdds.protocol.proto
+ .ScmBlockLocationProtocolProtos.DeleteScmBlockResult.Result.success;
+import static org.apache.hadoop.hdds.protocol.proto
+ .ScmBlockLocationProtocolProtos.DeleteScmBlockResult.Result.unknownFailure;
+
+/**
+ * This is a testing client that allows us to intercept calls from OzoneManager
+ * to SCM.
+ * <p>
+ * TODO: OzoneManager#getScmBlockClient -- so that we can load this class up via
+ * config setting into OzoneManager. Right now, we just pass this to
+ * KeyDeletingService only.
+ * <p>
+ * TODO: Move this class to a generic test utils so we can use this class in
+ * other Ozone Manager tests.
+ */
+public class ScmBlockLocationTestingClient implements ScmBlockLocationProtocol {
+ private static final Logger LOG =
+ LoggerFactory.getLogger(ScmBlockLocationTestingClient.class);
+ private final String clusterID;
+ private final String scmId;
+
+ // 0 means no calls will fail, +1 means all calls will fail, +2 means every
+ // second call will fail, +3 means every third and so on.
+ private final int failCallsFrequency;
+ private int currentCall = 0;
+
+ /**
+ * If ClusterID or SCMID is blank a per instance ID is generated.
+ *
+ * @param clusterID - String or blank.
+ * @param scmId - String or Blank.
+ * @param failCallsFrequency - Set to 0 for no failures, 1 for always to fail,
+ * a positive number for that frequency of failure.
+ */
+ public ScmBlockLocationTestingClient(String clusterID, String scmId,
+ int failCallsFrequency) {
+ this.clusterID = StringUtils.isNotBlank(clusterID) ? clusterID :
+ UUID.randomUUID().toString();
+ this.scmId = StringUtils.isNotBlank(scmId) ? scmId :
+ UUID.randomUUID().toString();
+ this.failCallsFrequency = Math.abs(failCallsFrequency);
+ switch (this.failCallsFrequency) {
+ case 0:
+ LOG.debug("Set to no failure mode, all delete block calls will " +
+ "succeed.");
+ break;
+ case 1:
+ LOG.debug("Set to all failure mode. All delete block calls to SCM" +
+ " will fail.");
+ break;
+ default:
+ LOG.debug("Set to Mix mode, every {} -th call will fail",
+ this.failCallsFrequency);
+ }
+
+ }
+
+ /**
+ * Returns Fake blocks to the BlockManager so we get blocks in the Database.
+ * @param size - size of the block.
+ * @param type Replication Type
+ * @param factor - Replication factor
+ * @param owner - String owner.
+ * @param excludeList list of dns/pipelines to exclude
+ * @return
+ * @throws IOException
+ */
+ @Override
+ public List<AllocatedBlock> allocateBlock(long size, int num,
+ HddsProtos.ReplicationType type, HddsProtos.ReplicationFactor factor,
+ String owner, ExcludeList excludeList) throws IOException {
+ DatanodeDetails datanodeDetails = TestUtils.randomDatanodeDetails();
+ Pipeline pipeline = createPipeline(datanodeDetails);
+ long containerID = Time.monotonicNow();
+ long localID = Time.monotonicNow();
+ AllocatedBlock.Builder abb =
+ new AllocatedBlock.Builder()
+ .setContainerBlockID(new ContainerBlockID(containerID, localID))
+ .setPipeline(pipeline);
+ return Collections.singletonList(abb.build());
+ }
+
+ private Pipeline createPipeline(DatanodeDetails datanode) {
+ List<DatanodeDetails> dns = new ArrayList<>();
+ dns.add(datanode);
+ Pipeline pipeline = Pipeline.newBuilder()
+ .setState(Pipeline.PipelineState.OPEN)
+ .setId(PipelineID.randomId())
+ .setType(HddsProtos.ReplicationType.STAND_ALONE)
+ .setFactor(HddsProtos.ReplicationFactor.ONE)
+ .setNodes(dns)
+ .build();
+ return pipeline;
+ }
+
+ @Override
+ public List<DeleteBlockGroupResult> deleteKeyBlocks(
+ List<BlockGroup> keyBlocksInfoList) throws IOException {
+ List<DeleteBlockGroupResult> results = new ArrayList<>();
+ List<DeleteBlockResult> blockResultList = new ArrayList<>();
+ Result result;
+ for (BlockGroup keyBlocks : keyBlocksInfoList) {
+ for (BlockID blockKey : keyBlocks.getBlockIDList()) {
+ currentCall++;
+ switch (this.failCallsFrequency) {
+ case 0:
+ result = success;
+ break;
+ case 1:
+ result = unknownFailure;
+ break;
+ default:
+ if (currentCall % this.failCallsFrequency == 0) {
+ result = unknownFailure;
+ } else {
+ result = success;
+ }
+ }
+ blockResultList.add(new DeleteBlockResult(blockKey, result));
+ }
+ results.add(new DeleteBlockGroupResult(keyBlocks.getGroupID(),
+ blockResultList));
+ }
+ return results;
+ }
+
+ @Override
+ public ScmInfo getScmInfo() throws IOException {
+ ScmInfo.Builder builder =
+ new ScmInfo.Builder()
+ .setClusterId(clusterID)
+ .setScmId(scmId);
+ return builder.build();
+ }
+
+ @Override
+ public List<DatanodeDetails> sortDatanodes(List<String> nodes,
+ String clientMachine) throws IOException {
+ return null;
+ }
+
+ @Override
+ public void close() throws IOException {
+
+ }
+}
---------------------------------------------------------------------
To unsubscribe, e-mail: common-commits-unsubscribe@hadoop.apache.org
For additional commands, e-mail: common-commits-help@hadoop.apache.org