You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@nifi.apache.org by mt...@apache.org on 2018/07/02 10:21:26 UTC

nifi git commit: NIFI-5147 Add CalculateAttributeHash processor

Repository: nifi
Updated Branches:
  refs/heads/master 4ea6e6a40 -> 7fb931889


NIFI-5147 Add CalculateAttributeHash processor

This closes #2802

Signed-off-by: Mike Thomsen <mi...@gmail.com>


Project: http://git-wip-us.apache.org/repos/asf/nifi/repo
Commit: http://git-wip-us.apache.org/repos/asf/nifi/commit/7fb93188
Tree: http://git-wip-us.apache.org/repos/asf/nifi/tree/7fb93188
Diff: http://git-wip-us.apache.org/repos/asf/nifi/diff/7fb93188

Branch: refs/heads/master
Commit: 7fb931889f37ecec7d14665e58c3332423916b64
Parents: 4ea6e6a
Author: Otto Fowler <ot...@gmail.com>
Authored: Mon Jun 18 11:00:38 2018 -0400
Committer: Mike Thomsen <mi...@gmail.com>
Committed: Mon Jul 2 06:20:21 2018 -0400

----------------------------------------------------------------------
 .../standard/CalculateAttributeHash.java        | 250 +++++++++++++++++++
 .../org.apache.nifi.processor.Processor         |   1 +
 .../standard/TestCalculateAttributeHash.java    | 178 +++++++++++++
 3 files changed, 429 insertions(+)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/nifi/blob/7fb93188/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/main/java/org/apache/nifi/processors/standard/CalculateAttributeHash.java
----------------------------------------------------------------------
diff --git a/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/main/java/org/apache/nifi/processors/standard/CalculateAttributeHash.java b/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/main/java/org/apache/nifi/processors/standard/CalculateAttributeHash.java
new file mode 100644
index 0000000..aa91cda
--- /dev/null
+++ b/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/main/java/org/apache/nifi/processors/standard/CalculateAttributeHash.java
@@ -0,0 +1,250 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.nifi.processors.standard;
+
+import org.apache.commons.codec.binary.Hex;
+import org.apache.commons.codec.digest.DigestUtils;
+import org.apache.commons.lang3.StringUtils;
+import org.apache.nifi.annotation.behavior.DynamicProperty;
+import org.apache.nifi.annotation.behavior.EventDriven;
+import org.apache.nifi.annotation.behavior.InputRequirement;
+import org.apache.nifi.annotation.behavior.SideEffectFree;
+import org.apache.nifi.annotation.behavior.SupportsBatching;
+import org.apache.nifi.annotation.behavior.WritesAttribute;
+import org.apache.nifi.annotation.documentation.CapabilityDescription;
+import org.apache.nifi.annotation.documentation.Tags;
+import org.apache.nifi.components.AllowableValue;
+import org.apache.nifi.components.PropertyDescriptor;
+import org.apache.nifi.flowfile.FlowFile;
+import org.apache.nifi.logging.ComponentLog;
+import org.apache.nifi.processor.AbstractProcessor;
+import org.apache.nifi.processor.ProcessContext;
+import org.apache.nifi.processor.ProcessSession;
+import org.apache.nifi.processor.Relationship;
+import org.apache.nifi.processor.util.StandardValidators;
+
+import java.nio.charset.Charset;
+import java.util.ArrayList;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.SortedMap;
+import java.util.TreeMap;
+import java.util.concurrent.atomic.AtomicReference;
+
+@EventDriven
+@SideEffectFree
+@SupportsBatching
+@Tags({"attributes", "hash"})
+@InputRequirement(InputRequirement.Requirement.INPUT_REQUIRED)
+@CapabilityDescription("Calculates a hash value for the value specified attributes and write it to an output attribute.  Please refer to https://csrc.nist.gov/Projects/Hash-Functions/NIST-Policy-on-Hash-Functions help decide on which algorithm to use.")
+@WritesAttribute(attribute = "<Specified Attribute Name per Dynamic Property>", description = "This Processor adds an attribute whose value is the result of "
+        + "Hashing of the found attribute. The name of this attribute is specified by the value of the dynamic property.")
+@DynamicProperty(name = "A flowfile attribute key for attribute inspection", value = "Attribute Name",
+        description = "The property name defines the attribute to look for and hash in the incoming flowfile."
+                + "The property value defines the name to give the generated attribute."
+                + "Attribute names must be unique.")
+public class CalculateAttributeHash extends AbstractProcessor {
+    public enum PartialAttributePolicy {
+        ALLOW,
+        PROHIBIT
+    }
+
+    static final AllowableValue MD2_VALUE = new AllowableValue("MD2", "MD2 Hashing Algorithm", "MD2 Hashing Algorithm");
+    static final AllowableValue MD5_VALUE = new AllowableValue("MD5", "MD5 Hashing Algorithm", "MD5 Hashing Algorithm");
+    static final AllowableValue SHA1_VALUE = new AllowableValue("SHA-1", "SHA-1 Hashing Algorithm", "SHA-1 Hashing Algorithm");
+    static final AllowableValue SHA256_VALUE = new AllowableValue("SHA-256", "SHA-256 Hashing Algorithm", "SHA-256 Hashing Algorithm");
+    static final AllowableValue SHA384_VALUE = new AllowableValue("SHA-384", "SHA-384 Hashing Algorithm", "SHA-384 Hashing Algorithm");
+    static final AllowableValue SHA512_VALUE = new AllowableValue("SHA-512", "SHA-512 Hashing Algorithm", "SHA-512 Hashing Algorithm");
+
+    static final AllowableValue ALLOW_PARITAL_ATTRIBUTES_VALUE = new AllowableValue(PartialAttributePolicy.ALLOW.name(),
+            "Allow missing attributes",
+            "Do not route to failure if there are attributes configured for hashing that are not present in the flow file");
+
+    static final AllowableValue FAIL_PARTIAL_ATTRIBUTES_VALUE = new AllowableValue(PartialAttributePolicy.PROHIBIT.name(),
+            "Fail if missing attributes",
+            "Route to failure if there are attributes configured for hashing that are not present in the flow file");
+
+    public static final PropertyDescriptor CHARACTER_SET = new PropertyDescriptor.Builder()
+            .name("Character Set")
+            .description("The Character Set in which the file is encoded")
+            .required(true)
+            .addValidator(StandardValidators.CHARACTER_SET_VALIDATOR)
+            .defaultValue("UTF-8")
+            .build();
+
+    public static final PropertyDescriptor FAIL_WHEN_EMPTY = new PropertyDescriptor.Builder()
+            .name("fail_when_empty")
+            .displayName("Fail when no attributes")
+            .description("Route to failure when none of the attributes that are configured for hashing are found.  " +
+            "If set to false, then flow files that do not contain any of the attributes that are configured for hashing will just pass through to success.")
+            .allowableValues("true","false")
+            .required(true)
+            .addValidator(StandardValidators.BOOLEAN_VALIDATOR)
+            .defaultValue("true")
+            .build();
+
+    public static final PropertyDescriptor HASH_ALGORITHM = new PropertyDescriptor.Builder()
+            .name("hash_algorithm")
+            .displayName("Hash Algorithm")
+            .description("The Hash Algorithm to use.  Note that not all of the algorithms available are recommended for use.  " +
+            "There are many things to consider when picking an algorithm, it is recommended that use most secure algorithm when possible.")
+            .required(true)
+            .allowableValues(MD2_VALUE, MD5_VALUE, SHA1_VALUE, SHA256_VALUE, SHA384_VALUE, SHA512_VALUE)
+            .defaultValue(SHA256_VALUE.getValue())
+            .addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
+            .build();
+
+    public static final PropertyDescriptor PARTIAL_ATTR_ROUTE_POLICY = new PropertyDescriptor.Builder()
+            .name("missing_attr_policy")
+            .displayName("Missing attribute policy")
+            .description("Policy for how the processor handles attributes that are configured for hashing but are not found in the flowfile.")
+            .required(true)
+            .allowableValues(ALLOW_PARITAL_ATTRIBUTES_VALUE, FAIL_PARTIAL_ATTRIBUTES_VALUE)
+            .addValidator(StandardValidators.NON_BLANK_VALIDATOR)
+            .defaultValue(ALLOW_PARITAL_ATTRIBUTES_VALUE.getValue())
+            .build();
+
+    public static final Relationship REL_SUCCESS = new Relationship.Builder()
+            .name("success")
+            .description("Used for FlowFiles that have a hash value added")
+            .build();
+    public static final Relationship REL_FAILURE = new Relationship.Builder()
+            .name("failure")
+            .description("Used for FlowFiles that are missing required attributes")
+            .build();
+
+    private final static Set<Relationship> relationships;
+    private final static List<PropertyDescriptor> properties;
+
+    private final AtomicReference<Map<String, String>> attributeToGenerateNameMapRef = new AtomicReference<>(Collections.emptyMap());
+
+    static {
+        final Set<Relationship> _relationships = new HashSet<>();
+        _relationships.add(REL_FAILURE);
+        _relationships.add(REL_SUCCESS);
+        relationships = Collections.unmodifiableSet(_relationships);
+
+        final List<PropertyDescriptor> _properties = new ArrayList<>();
+        _properties.add(CHARACTER_SET);
+        _properties.add(FAIL_WHEN_EMPTY);
+        _properties.add(HASH_ALGORITHM);
+        _properties.add(PARTIAL_ATTR_ROUTE_POLICY);
+        properties = Collections.unmodifiableList(_properties);
+    }
+
+    @Override
+    public Set<Relationship> getRelationships() {
+        return relationships;
+    }
+
+    @Override
+    protected List<PropertyDescriptor> getSupportedPropertyDescriptors() {
+        return properties;
+    }
+
+    @Override
+    protected PropertyDescriptor getSupportedDynamicPropertyDescriptor(final String propertyDescriptorName) {
+        return new PropertyDescriptor.Builder()
+                .name(propertyDescriptorName).addValidator(StandardValidators.NON_BLANK_VALIDATOR).build();
+    }
+
+    @Override
+    public void onPropertyModified(final PropertyDescriptor descriptor, final String oldValue, final String newValue) {
+        if (descriptor.isRequired()) {
+            return;
+        }
+
+        final Map<String, String> attributeToGeneratedNameMap = new HashMap<>(attributeToGenerateNameMapRef.get());
+        if (newValue == null) {
+            attributeToGeneratedNameMap.remove(descriptor.getName());
+        } else {
+            attributeToGeneratedNameMap.put(descriptor.getName(), newValue);
+        }
+
+        attributeToGenerateNameMapRef.set(Collections.unmodifiableMap(attributeToGeneratedNameMap));
+    }
+
+    @Override
+    public void onTrigger(final ProcessContext context, final ProcessSession session) {
+        FlowFile flowFile = session.get();
+        if (flowFile == null) {
+            return;
+        }
+        final Charset charset = Charset.forName(context.getProperty(CHARACTER_SET).getValue());
+        final Map<String, String> attributeToGeneratedNameMap = attributeToGenerateNameMapRef.get();
+        final ComponentLog logger = getLogger();
+
+        final SortedMap<String, String> attributes = getRelevantAttributes(flowFile, attributeToGeneratedNameMap);
+        if(attributes.isEmpty()) {
+            if( context.getProperty(FAIL_WHEN_EMPTY).asBoolean()) {
+                logger.info("routing {} to 'failure' because of missing all attributes: {}", new Object[]{flowFile, getMissingKeysString(null,attributeToGeneratedNameMap.keySet())});
+                session.transfer(flowFile, REL_FAILURE);
+                return;
+            }
+        }
+        if (attributes.size() != attributeToGeneratedNameMap.size()) {
+            if (PartialAttributePolicy.valueOf(context.getProperty(PARTIAL_ATTR_ROUTE_POLICY).getValue()) == PartialAttributePolicy.PROHIBIT) {
+                logger.info("routing {} to 'failure' because of missing attributes: {}", new Object[]{flowFile, getMissingKeysString(attributes.keySet(),attributeToGeneratedNameMap.keySet())});
+                session.transfer(flowFile, REL_FAILURE);
+                return;
+            }
+        }
+        // Generate a hash with the configured algorithm for each attribute value
+        // and create a new attribute with the configured name
+        for (final Map.Entry<String, String> entry : attributes.entrySet()) {
+            String value = hashValue(context.getProperty(HASH_ALGORITHM).getValue(), entry.getValue(), charset);
+            session.putAttribute(flowFile, attributeToGeneratedNameMap.get(entry.getKey()), value);
+        }
+        session.getProvenanceReporter().modifyAttributes(flowFile);
+        session.transfer(flowFile, REL_SUCCESS);
+    }
+
+    private static SortedMap<String, String> getRelevantAttributes(final FlowFile flowFile, final Map<String, String> attributeToGeneratedNameMap) {
+        final SortedMap<String, String> attributeMap = new TreeMap<>();
+        for (final Map.Entry<String, String> entry : attributeToGeneratedNameMap.entrySet()) {
+            final String attributeName = entry.getKey();
+            final String attributeValue = flowFile.getAttribute(attributeName);
+            if (attributeValue != null) {
+                attributeMap.put(attributeName, attributeValue);
+            }
+        }
+        return attributeMap;
+    }
+
+    private static String hashValue(String algorithm, String value, Charset charset) {
+        if (StringUtils.isBlank(value)) {
+            return value;
+        }
+        return Hex.encodeHexString(DigestUtils.getDigest(algorithm).digest(value.getBytes(charset)));
+    }
+
+    private static String getMissingKeysString(Set<String> foundKeys, Set<String> wantedKeys) {
+        final StringBuilder missingKeys = new StringBuilder();
+        for (final String wantedKey : wantedKeys) {
+            if (foundKeys == null || !foundKeys.contains(wantedKey)) {
+                missingKeys.append(wantedKey).append(" ");
+            }
+        }
+        return missingKeys.toString();
+    }
+}
+

http://git-wip-us.apache.org/repos/asf/nifi/blob/7fb93188/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/main/resources/META-INF/services/org.apache.nifi.processor.Processor
----------------------------------------------------------------------
diff --git a/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/main/resources/META-INF/services/org.apache.nifi.processor.Processor b/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/main/resources/META-INF/services/org.apache.nifi.processor.Processor
index 8fc361d..0c0bf89 100644
--- a/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/main/resources/META-INF/services/org.apache.nifi.processor.Processor
+++ b/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/main/resources/META-INF/services/org.apache.nifi.processor.Processor
@@ -15,6 +15,7 @@
 org.apache.nifi.processors.standard.AttributesToJSON
 org.apache.nifi.processors.standard.AttributesToCSV
 org.apache.nifi.processors.standard.Base64EncodeContent
+org.apache.nifi.processors.standard.CalculateAttributeHash
 org.apache.nifi.processors.standard.CalculateRecordStats
 org.apache.nifi.processors.standard.CompressContent
 org.apache.nifi.processors.standard.ControlRate

http://git-wip-us.apache.org/repos/asf/nifi/blob/7fb93188/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/test/java/org/apache/nifi/processors/standard/TestCalculateAttributeHash.java
----------------------------------------------------------------------
diff --git a/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/test/java/org/apache/nifi/processors/standard/TestCalculateAttributeHash.java b/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/test/java/org/apache/nifi/processors/standard/TestCalculateAttributeHash.java
new file mode 100644
index 0000000..9e21a5c
--- /dev/null
+++ b/nifi-nar-bundles/nifi-standard-bundle/nifi-standard-processors/src/test/java/org/apache/nifi/processors/standard/TestCalculateAttributeHash.java
@@ -0,0 +1,178 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.nifi.processors.standard;
+
+import org.apache.commons.codec.binary.Hex;
+import org.apache.commons.codec.digest.DigestUtils;
+import org.apache.nifi.util.MockFlowFile;
+import org.apache.nifi.util.TestRunner;
+import org.apache.nifi.util.TestRunners;
+import org.junit.Assert;
+import org.junit.Test;
+
+import java.nio.charset.Charset;
+import java.nio.charset.StandardCharsets;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
+public class TestCalculateAttributeHash {
+
+    private static final Charset UTF8 = StandardCharsets.UTF_8;
+    @Test
+    public void testMD2() throws Exception {
+        testAllAlgorithm("MD2");
+        testParitalAlgorithm("MD2");
+        testMissingAlgorithm("MD2");
+    }
+
+    @Test
+    public void testMD5() throws Exception {
+        testAllAlgorithm("MD5");
+        testParitalAlgorithm("MD5");
+        testMissingAlgorithm("MD5");
+    }
+
+    @Test
+    public void testSHA1() throws Exception {
+        testAllAlgorithm("SHA-1");
+        testParitalAlgorithm("SHA-1");
+        testMissingAlgorithm("SHA-1");
+    }
+
+    @Test
+    public void testSHA256() throws Exception {
+        testAllAlgorithm("SHA-256");
+        testParitalAlgorithm("SHA-256");
+        testMissingAlgorithm("SHA-256");
+    }
+
+    @Test
+    public void testSHA384() throws Exception {
+        testAllAlgorithm("SHA-384");
+        testParitalAlgorithm("SHA-384");
+        testMissingAlgorithm("SHA-384");
+    }
+
+    @Test
+    public void testSHA512() throws Exception {
+        testAllAlgorithm("SHA-512");
+        testParitalAlgorithm("SHA-512");
+        testMissingAlgorithm("SHA-512");
+    }
+
+    public void testAllAlgorithm(String algorithm) {
+        final TestRunner runner = TestRunners.newTestRunner(new CalculateAttributeHash());
+        runner.setProperty(CalculateAttributeHash.HASH_ALGORITHM.getName(), algorithm);
+        runner.setProperty("name", String.format("%s_%s", "name", algorithm));
+        runner.setProperty("value", String.format("%s_%s", "value", algorithm));
+
+        final Map<String, String> attributeMap = new HashMap<>();
+        attributeMap.put("name", "abcdefg");
+        attributeMap.put("value", "hijklmnop");
+        runner.enqueue(new byte[0], attributeMap);
+
+        runner.run(1);
+
+        runner.assertTransferCount(HashAttribute.REL_FAILURE, 0);
+        runner.assertTransferCount(HashAttribute.REL_SUCCESS, 1);
+
+        final List<MockFlowFile> success = runner.getFlowFilesForRelationship(HashAttribute.REL_SUCCESS);
+
+        for (final MockFlowFile flowFile : success) {
+            Assert.assertEquals(Hex.encodeHexString(DigestUtils.getDigest(algorithm).digest("abcdefg".getBytes(UTF8))),
+                    flowFile.getAttribute(String.format("%s_%s", "name", algorithm)));
+            Assert.assertEquals(Hex.encodeHexString(DigestUtils.getDigest(algorithm).digest("hijklmnop".getBytes(UTF8))),
+                    flowFile.getAttribute(String.format("%s_%s", "value", algorithm)));
+        }
+    }
+
+    public void testParitalAlgorithm(String algorithm) {
+        final TestRunner runner = TestRunners.newTestRunner(new CalculateAttributeHash());
+        runner.setProperty(CalculateAttributeHash.HASH_ALGORITHM.getName(), algorithm);
+        runner.setProperty("name", String.format("%s_%s", "name", algorithm));
+        runner.setProperty("value", String.format("%s_%s", "value", algorithm));
+
+        // test default ALLOW
+
+        final Map<String, String> paritalAttributeMap = new HashMap<>();
+        paritalAttributeMap.put("name", "abcdefg");
+        runner.enqueue(new byte[0], paritalAttributeMap);
+
+        runner.run(1);
+
+        runner.assertTransferCount(HashAttribute.REL_FAILURE, 0);
+        runner.assertTransferCount(HashAttribute.REL_SUCCESS, 1);
+
+        final List<MockFlowFile> success = runner.getFlowFilesForRelationship(HashAttribute.REL_SUCCESS);
+
+        for (final MockFlowFile flowFile : success) {
+            Assert.assertEquals(Hex.encodeHexString(DigestUtils.getDigest(algorithm).digest("abcdefg".getBytes(UTF8))),
+                    flowFile.getAttribute(String.format("%s_%s", "name", algorithm)));
+                   }
+
+        runner.clearTransferState();
+
+        // test PROHIBIT
+        runner.setProperty(CalculateAttributeHash.PARTIAL_ATTR_ROUTE_POLICY, CalculateAttributeHash.PartialAttributePolicy.PROHIBIT.name());
+        runner.enqueue(new byte[0], paritalAttributeMap);
+
+        runner.run(1);
+
+        runner.assertTransferCount(HashAttribute.REL_FAILURE, 1);
+        runner.assertTransferCount(HashAttribute.REL_SUCCESS, 0);
+    }
+
+    public void testMissingAlgorithm(String algorithm) {
+        final TestRunner runner = TestRunners.newTestRunner(new CalculateAttributeHash());
+        runner.setProperty(CalculateAttributeHash.HASH_ALGORITHM.getName(), algorithm);
+        runner.setProperty("name", String.format("%s_%s", "name", algorithm));
+        runner.setProperty("value", String.format("%s_%s", "value", algorithm));
+
+        // test default of true
+        final Map<String, String> attributeMap = new HashMap<>();
+        attributeMap.put("not_name", "abcdefg");
+        attributeMap.put("not_value", "hijklmnop");
+        runner.enqueue(new byte[0], attributeMap);
+
+        runner.run(1);
+
+        runner.assertTransferCount(HashAttribute.REL_FAILURE, 1);
+        runner.assertTransferCount(HashAttribute.REL_SUCCESS, 0);
+
+        runner.clearTransferState();
+
+        runner.setProperty(CalculateAttributeHash.FAIL_WHEN_EMPTY,"false");
+
+        runner.enqueue(new byte[0], attributeMap);
+
+        runner.run(1);
+
+        runner.assertTransferCount(HashAttribute.REL_FAILURE, 0);
+        runner.assertTransferCount(HashAttribute.REL_SUCCESS, 1);
+
+        final List<MockFlowFile> success = runner.getFlowFilesForRelationship(HashAttribute.REL_SUCCESS);
+        final Map<String, Integer> correlationCount = new HashMap<>();
+
+        for (final MockFlowFile flowFile : success) {
+            flowFile.assertAttributeNotExists(String.format("%s_%s", "name", algorithm));
+            flowFile.assertAttributeNotExists(String.format("%s_%s", "value", algorithm));
+            Assert.assertEquals("abcdefg",flowFile.getAttribute("not_name"));
+            Assert.assertEquals("hijklmnop",flowFile.getAttribute("not_value"));
+        }
+    }
+}