mirror of https://github.com/apache/nifi.git
Revert "NIFI-5147 Add CalculateAttributeHash processor"
This reverts commit 7fb931889f
.
This commit is contained in:
parent
b549fbb67b
commit
790102d8b1
|
@ -1,250 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.nifi.processors.standard;
|
||||
|
||||
import org.apache.commons.codec.binary.Hex;
|
||||
import org.apache.commons.codec.digest.DigestUtils;
|
||||
import org.apache.commons.lang3.StringUtils;
|
||||
import org.apache.nifi.annotation.behavior.DynamicProperty;
|
||||
import org.apache.nifi.annotation.behavior.EventDriven;
|
||||
import org.apache.nifi.annotation.behavior.InputRequirement;
|
||||
import org.apache.nifi.annotation.behavior.SideEffectFree;
|
||||
import org.apache.nifi.annotation.behavior.SupportsBatching;
|
||||
import org.apache.nifi.annotation.behavior.WritesAttribute;
|
||||
import org.apache.nifi.annotation.documentation.CapabilityDescription;
|
||||
import org.apache.nifi.annotation.documentation.Tags;
|
||||
import org.apache.nifi.components.AllowableValue;
|
||||
import org.apache.nifi.components.PropertyDescriptor;
|
||||
import org.apache.nifi.flowfile.FlowFile;
|
||||
import org.apache.nifi.logging.ComponentLog;
|
||||
import org.apache.nifi.processor.AbstractProcessor;
|
||||
import org.apache.nifi.processor.ProcessContext;
|
||||
import org.apache.nifi.processor.ProcessSession;
|
||||
import org.apache.nifi.processor.Relationship;
|
||||
import org.apache.nifi.processor.util.StandardValidators;
|
||||
|
||||
import java.nio.charset.Charset;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
import java.util.SortedMap;
|
||||
import java.util.TreeMap;
|
||||
import java.util.concurrent.atomic.AtomicReference;
|
||||
|
||||
@EventDriven
|
||||
@SideEffectFree
|
||||
@SupportsBatching
|
||||
@Tags({"attributes", "hash"})
|
||||
@InputRequirement(InputRequirement.Requirement.INPUT_REQUIRED)
|
||||
@CapabilityDescription("Calculates a hash value for the value specified attributes and write it to an output attribute. Please refer to https://csrc.nist.gov/Projects/Hash-Functions/NIST-Policy-on-Hash-Functions help decide on which algorithm to use.")
|
||||
@WritesAttribute(attribute = "<Specified Attribute Name per Dynamic Property>", description = "This Processor adds an attribute whose value is the result of "
|
||||
+ "Hashing of the found attribute. The name of this attribute is specified by the value of the dynamic property.")
|
||||
@DynamicProperty(name = "A flowfile attribute key for attribute inspection", value = "Attribute Name",
|
||||
description = "The property name defines the attribute to look for and hash in the incoming flowfile."
|
||||
+ "The property value defines the name to give the generated attribute."
|
||||
+ "Attribute names must be unique.")
|
||||
public class CalculateAttributeHash extends AbstractProcessor {
|
||||
public enum PartialAttributePolicy {
|
||||
ALLOW,
|
||||
PROHIBIT
|
||||
}
|
||||
|
||||
static final AllowableValue MD2_VALUE = new AllowableValue("MD2", "MD2 Hashing Algorithm", "MD2 Hashing Algorithm");
|
||||
static final AllowableValue MD5_VALUE = new AllowableValue("MD5", "MD5 Hashing Algorithm", "MD5 Hashing Algorithm");
|
||||
static final AllowableValue SHA1_VALUE = new AllowableValue("SHA-1", "SHA-1 Hashing Algorithm", "SHA-1 Hashing Algorithm");
|
||||
static final AllowableValue SHA256_VALUE = new AllowableValue("SHA-256", "SHA-256 Hashing Algorithm", "SHA-256 Hashing Algorithm");
|
||||
static final AllowableValue SHA384_VALUE = new AllowableValue("SHA-384", "SHA-384 Hashing Algorithm", "SHA-384 Hashing Algorithm");
|
||||
static final AllowableValue SHA512_VALUE = new AllowableValue("SHA-512", "SHA-512 Hashing Algorithm", "SHA-512 Hashing Algorithm");
|
||||
|
||||
static final AllowableValue ALLOW_PARITAL_ATTRIBUTES_VALUE = new AllowableValue(PartialAttributePolicy.ALLOW.name(),
|
||||
"Allow missing attributes",
|
||||
"Do not route to failure if there are attributes configured for hashing that are not present in the flow file");
|
||||
|
||||
static final AllowableValue FAIL_PARTIAL_ATTRIBUTES_VALUE = new AllowableValue(PartialAttributePolicy.PROHIBIT.name(),
|
||||
"Fail if missing attributes",
|
||||
"Route to failure if there are attributes configured for hashing that are not present in the flow file");
|
||||
|
||||
public static final PropertyDescriptor CHARACTER_SET = new PropertyDescriptor.Builder()
|
||||
.name("Character Set")
|
||||
.description("The Character Set in which the file is encoded")
|
||||
.required(true)
|
||||
.addValidator(StandardValidators.CHARACTER_SET_VALIDATOR)
|
||||
.defaultValue("UTF-8")
|
||||
.build();
|
||||
|
||||
public static final PropertyDescriptor FAIL_WHEN_EMPTY = new PropertyDescriptor.Builder()
|
||||
.name("fail_when_empty")
|
||||
.displayName("Fail when no attributes")
|
||||
.description("Route to failure when none of the attributes that are configured for hashing are found. " +
|
||||
"If set to false, then flow files that do not contain any of the attributes that are configured for hashing will just pass through to success.")
|
||||
.allowableValues("true","false")
|
||||
.required(true)
|
||||
.addValidator(StandardValidators.BOOLEAN_VALIDATOR)
|
||||
.defaultValue("true")
|
||||
.build();
|
||||
|
||||
public static final PropertyDescriptor HASH_ALGORITHM = new PropertyDescriptor.Builder()
|
||||
.name("hash_algorithm")
|
||||
.displayName("Hash Algorithm")
|
||||
.description("The Hash Algorithm to use. Note that not all of the algorithms available are recommended for use. " +
|
||||
"There are many things to consider when picking an algorithm, it is recommended that use most secure algorithm when possible.")
|
||||
.required(true)
|
||||
.allowableValues(MD2_VALUE, MD5_VALUE, SHA1_VALUE, SHA256_VALUE, SHA384_VALUE, SHA512_VALUE)
|
||||
.defaultValue(SHA256_VALUE.getValue())
|
||||
.addValidator(StandardValidators.NON_EMPTY_VALIDATOR)
|
||||
.build();
|
||||
|
||||
public static final PropertyDescriptor PARTIAL_ATTR_ROUTE_POLICY = new PropertyDescriptor.Builder()
|
||||
.name("missing_attr_policy")
|
||||
.displayName("Missing attribute policy")
|
||||
.description("Policy for how the processor handles attributes that are configured for hashing but are not found in the flowfile.")
|
||||
.required(true)
|
||||
.allowableValues(ALLOW_PARITAL_ATTRIBUTES_VALUE, FAIL_PARTIAL_ATTRIBUTES_VALUE)
|
||||
.addValidator(StandardValidators.NON_BLANK_VALIDATOR)
|
||||
.defaultValue(ALLOW_PARITAL_ATTRIBUTES_VALUE.getValue())
|
||||
.build();
|
||||
|
||||
public static final Relationship REL_SUCCESS = new Relationship.Builder()
|
||||
.name("success")
|
||||
.description("Used for FlowFiles that have a hash value added")
|
||||
.build();
|
||||
public static final Relationship REL_FAILURE = new Relationship.Builder()
|
||||
.name("failure")
|
||||
.description("Used for FlowFiles that are missing required attributes")
|
||||
.build();
|
||||
|
||||
private final static Set<Relationship> relationships;
|
||||
private final static List<PropertyDescriptor> properties;
|
||||
|
||||
private final AtomicReference<Map<String, String>> attributeToGenerateNameMapRef = new AtomicReference<>(Collections.emptyMap());
|
||||
|
||||
static {
|
||||
final Set<Relationship> _relationships = new HashSet<>();
|
||||
_relationships.add(REL_FAILURE);
|
||||
_relationships.add(REL_SUCCESS);
|
||||
relationships = Collections.unmodifiableSet(_relationships);
|
||||
|
||||
final List<PropertyDescriptor> _properties = new ArrayList<>();
|
||||
_properties.add(CHARACTER_SET);
|
||||
_properties.add(FAIL_WHEN_EMPTY);
|
||||
_properties.add(HASH_ALGORITHM);
|
||||
_properties.add(PARTIAL_ATTR_ROUTE_POLICY);
|
||||
properties = Collections.unmodifiableList(_properties);
|
||||
}
|
||||
|
||||
@Override
|
||||
public Set<Relationship> getRelationships() {
|
||||
return relationships;
|
||||
}
|
||||
|
||||
@Override
|
||||
protected List<PropertyDescriptor> getSupportedPropertyDescriptors() {
|
||||
return properties;
|
||||
}
|
||||
|
||||
@Override
|
||||
protected PropertyDescriptor getSupportedDynamicPropertyDescriptor(final String propertyDescriptorName) {
|
||||
return new PropertyDescriptor.Builder()
|
||||
.name(propertyDescriptorName).addValidator(StandardValidators.NON_BLANK_VALIDATOR).build();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void onPropertyModified(final PropertyDescriptor descriptor, final String oldValue, final String newValue) {
|
||||
if (descriptor.isRequired()) {
|
||||
return;
|
||||
}
|
||||
|
||||
final Map<String, String> attributeToGeneratedNameMap = new HashMap<>(attributeToGenerateNameMapRef.get());
|
||||
if (newValue == null) {
|
||||
attributeToGeneratedNameMap.remove(descriptor.getName());
|
||||
} else {
|
||||
attributeToGeneratedNameMap.put(descriptor.getName(), newValue);
|
||||
}
|
||||
|
||||
attributeToGenerateNameMapRef.set(Collections.unmodifiableMap(attributeToGeneratedNameMap));
|
||||
}
|
||||
|
||||
@Override
|
||||
public void onTrigger(final ProcessContext context, final ProcessSession session) {
|
||||
FlowFile flowFile = session.get();
|
||||
if (flowFile == null) {
|
||||
return;
|
||||
}
|
||||
final Charset charset = Charset.forName(context.getProperty(CHARACTER_SET).getValue());
|
||||
final Map<String, String> attributeToGeneratedNameMap = attributeToGenerateNameMapRef.get();
|
||||
final ComponentLog logger = getLogger();
|
||||
|
||||
final SortedMap<String, String> attributes = getRelevantAttributes(flowFile, attributeToGeneratedNameMap);
|
||||
if(attributes.isEmpty()) {
|
||||
if( context.getProperty(FAIL_WHEN_EMPTY).asBoolean()) {
|
||||
logger.info("routing {} to 'failure' because of missing all attributes: {}", new Object[]{flowFile, getMissingKeysString(null,attributeToGeneratedNameMap.keySet())});
|
||||
session.transfer(flowFile, REL_FAILURE);
|
||||
return;
|
||||
}
|
||||
}
|
||||
if (attributes.size() != attributeToGeneratedNameMap.size()) {
|
||||
if (PartialAttributePolicy.valueOf(context.getProperty(PARTIAL_ATTR_ROUTE_POLICY).getValue()) == PartialAttributePolicy.PROHIBIT) {
|
||||
logger.info("routing {} to 'failure' because of missing attributes: {}", new Object[]{flowFile, getMissingKeysString(attributes.keySet(),attributeToGeneratedNameMap.keySet())});
|
||||
session.transfer(flowFile, REL_FAILURE);
|
||||
return;
|
||||
}
|
||||
}
|
||||
// Generate a hash with the configured algorithm for each attribute value
|
||||
// and create a new attribute with the configured name
|
||||
for (final Map.Entry<String, String> entry : attributes.entrySet()) {
|
||||
String value = hashValue(context.getProperty(HASH_ALGORITHM).getValue(), entry.getValue(), charset);
|
||||
session.putAttribute(flowFile, attributeToGeneratedNameMap.get(entry.getKey()), value);
|
||||
}
|
||||
session.getProvenanceReporter().modifyAttributes(flowFile);
|
||||
session.transfer(flowFile, REL_SUCCESS);
|
||||
}
|
||||
|
||||
private static SortedMap<String, String> getRelevantAttributes(final FlowFile flowFile, final Map<String, String> attributeToGeneratedNameMap) {
|
||||
final SortedMap<String, String> attributeMap = new TreeMap<>();
|
||||
for (final Map.Entry<String, String> entry : attributeToGeneratedNameMap.entrySet()) {
|
||||
final String attributeName = entry.getKey();
|
||||
final String attributeValue = flowFile.getAttribute(attributeName);
|
||||
if (attributeValue != null) {
|
||||
attributeMap.put(attributeName, attributeValue);
|
||||
}
|
||||
}
|
||||
return attributeMap;
|
||||
}
|
||||
|
||||
private static String hashValue(String algorithm, String value, Charset charset) {
|
||||
if (StringUtils.isBlank(value)) {
|
||||
return value;
|
||||
}
|
||||
return Hex.encodeHexString(DigestUtils.getDigest(algorithm).digest(value.getBytes(charset)));
|
||||
}
|
||||
|
||||
private static String getMissingKeysString(Set<String> foundKeys, Set<String> wantedKeys) {
|
||||
final StringBuilder missingKeys = new StringBuilder();
|
||||
for (final String wantedKey : wantedKeys) {
|
||||
if (foundKeys == null || !foundKeys.contains(wantedKey)) {
|
||||
missingKeys.append(wantedKey).append(" ");
|
||||
}
|
||||
}
|
||||
return missingKeys.toString();
|
||||
}
|
||||
}
|
||||
|
|
@ -15,7 +15,6 @@
|
|||
org.apache.nifi.processors.standard.AttributesToJSON
|
||||
org.apache.nifi.processors.standard.AttributesToCSV
|
||||
org.apache.nifi.processors.standard.Base64EncodeContent
|
||||
org.apache.nifi.processors.standard.CalculateAttributeHash
|
||||
org.apache.nifi.processors.standard.CalculateRecordStats
|
||||
org.apache.nifi.processors.standard.CompressContent
|
||||
org.apache.nifi.processors.standard.ControlRate
|
||||
|
|
|
@ -1,178 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.nifi.processors.standard;
|
||||
|
||||
import org.apache.commons.codec.binary.Hex;
|
||||
import org.apache.commons.codec.digest.DigestUtils;
|
||||
import org.apache.nifi.util.MockFlowFile;
|
||||
import org.apache.nifi.util.TestRunner;
|
||||
import org.apache.nifi.util.TestRunners;
|
||||
import org.junit.Assert;
|
||||
import org.junit.Test;
|
||||
|
||||
import java.nio.charset.Charset;
|
||||
import java.nio.charset.StandardCharsets;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
||||
public class TestCalculateAttributeHash {
|
||||
|
||||
private static final Charset UTF8 = StandardCharsets.UTF_8;
|
||||
@Test
|
||||
public void testMD2() throws Exception {
|
||||
testAllAlgorithm("MD2");
|
||||
testParitalAlgorithm("MD2");
|
||||
testMissingAlgorithm("MD2");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testMD5() throws Exception {
|
||||
testAllAlgorithm("MD5");
|
||||
testParitalAlgorithm("MD5");
|
||||
testMissingAlgorithm("MD5");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testSHA1() throws Exception {
|
||||
testAllAlgorithm("SHA-1");
|
||||
testParitalAlgorithm("SHA-1");
|
||||
testMissingAlgorithm("SHA-1");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testSHA256() throws Exception {
|
||||
testAllAlgorithm("SHA-256");
|
||||
testParitalAlgorithm("SHA-256");
|
||||
testMissingAlgorithm("SHA-256");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testSHA384() throws Exception {
|
||||
testAllAlgorithm("SHA-384");
|
||||
testParitalAlgorithm("SHA-384");
|
||||
testMissingAlgorithm("SHA-384");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testSHA512() throws Exception {
|
||||
testAllAlgorithm("SHA-512");
|
||||
testParitalAlgorithm("SHA-512");
|
||||
testMissingAlgorithm("SHA-512");
|
||||
}
|
||||
|
||||
public void testAllAlgorithm(String algorithm) {
|
||||
final TestRunner runner = TestRunners.newTestRunner(new CalculateAttributeHash());
|
||||
runner.setProperty(CalculateAttributeHash.HASH_ALGORITHM.getName(), algorithm);
|
||||
runner.setProperty("name", String.format("%s_%s", "name", algorithm));
|
||||
runner.setProperty("value", String.format("%s_%s", "value", algorithm));
|
||||
|
||||
final Map<String, String> attributeMap = new HashMap<>();
|
||||
attributeMap.put("name", "abcdefg");
|
||||
attributeMap.put("value", "hijklmnop");
|
||||
runner.enqueue(new byte[0], attributeMap);
|
||||
|
||||
runner.run(1);
|
||||
|
||||
runner.assertTransferCount(HashAttribute.REL_FAILURE, 0);
|
||||
runner.assertTransferCount(HashAttribute.REL_SUCCESS, 1);
|
||||
|
||||
final List<MockFlowFile> success = runner.getFlowFilesForRelationship(HashAttribute.REL_SUCCESS);
|
||||
|
||||
for (final MockFlowFile flowFile : success) {
|
||||
Assert.assertEquals(Hex.encodeHexString(DigestUtils.getDigest(algorithm).digest("abcdefg".getBytes(UTF8))),
|
||||
flowFile.getAttribute(String.format("%s_%s", "name", algorithm)));
|
||||
Assert.assertEquals(Hex.encodeHexString(DigestUtils.getDigest(algorithm).digest("hijklmnop".getBytes(UTF8))),
|
||||
flowFile.getAttribute(String.format("%s_%s", "value", algorithm)));
|
||||
}
|
||||
}
|
||||
|
||||
public void testParitalAlgorithm(String algorithm) {
|
||||
final TestRunner runner = TestRunners.newTestRunner(new CalculateAttributeHash());
|
||||
runner.setProperty(CalculateAttributeHash.HASH_ALGORITHM.getName(), algorithm);
|
||||
runner.setProperty("name", String.format("%s_%s", "name", algorithm));
|
||||
runner.setProperty("value", String.format("%s_%s", "value", algorithm));
|
||||
|
||||
// test default ALLOW
|
||||
|
||||
final Map<String, String> paritalAttributeMap = new HashMap<>();
|
||||
paritalAttributeMap.put("name", "abcdefg");
|
||||
runner.enqueue(new byte[0], paritalAttributeMap);
|
||||
|
||||
runner.run(1);
|
||||
|
||||
runner.assertTransferCount(HashAttribute.REL_FAILURE, 0);
|
||||
runner.assertTransferCount(HashAttribute.REL_SUCCESS, 1);
|
||||
|
||||
final List<MockFlowFile> success = runner.getFlowFilesForRelationship(HashAttribute.REL_SUCCESS);
|
||||
|
||||
for (final MockFlowFile flowFile : success) {
|
||||
Assert.assertEquals(Hex.encodeHexString(DigestUtils.getDigest(algorithm).digest("abcdefg".getBytes(UTF8))),
|
||||
flowFile.getAttribute(String.format("%s_%s", "name", algorithm)));
|
||||
}
|
||||
|
||||
runner.clearTransferState();
|
||||
|
||||
// test PROHIBIT
|
||||
runner.setProperty(CalculateAttributeHash.PARTIAL_ATTR_ROUTE_POLICY, CalculateAttributeHash.PartialAttributePolicy.PROHIBIT.name());
|
||||
runner.enqueue(new byte[0], paritalAttributeMap);
|
||||
|
||||
runner.run(1);
|
||||
|
||||
runner.assertTransferCount(HashAttribute.REL_FAILURE, 1);
|
||||
runner.assertTransferCount(HashAttribute.REL_SUCCESS, 0);
|
||||
}
|
||||
|
||||
public void testMissingAlgorithm(String algorithm) {
|
||||
final TestRunner runner = TestRunners.newTestRunner(new CalculateAttributeHash());
|
||||
runner.setProperty(CalculateAttributeHash.HASH_ALGORITHM.getName(), algorithm);
|
||||
runner.setProperty("name", String.format("%s_%s", "name", algorithm));
|
||||
runner.setProperty("value", String.format("%s_%s", "value", algorithm));
|
||||
|
||||
// test default of true
|
||||
final Map<String, String> attributeMap = new HashMap<>();
|
||||
attributeMap.put("not_name", "abcdefg");
|
||||
attributeMap.put("not_value", "hijklmnop");
|
||||
runner.enqueue(new byte[0], attributeMap);
|
||||
|
||||
runner.run(1);
|
||||
|
||||
runner.assertTransferCount(HashAttribute.REL_FAILURE, 1);
|
||||
runner.assertTransferCount(HashAttribute.REL_SUCCESS, 0);
|
||||
|
||||
runner.clearTransferState();
|
||||
|
||||
runner.setProperty(CalculateAttributeHash.FAIL_WHEN_EMPTY,"false");
|
||||
|
||||
runner.enqueue(new byte[0], attributeMap);
|
||||
|
||||
runner.run(1);
|
||||
|
||||
runner.assertTransferCount(HashAttribute.REL_FAILURE, 0);
|
||||
runner.assertTransferCount(HashAttribute.REL_SUCCESS, 1);
|
||||
|
||||
final List<MockFlowFile> success = runner.getFlowFilesForRelationship(HashAttribute.REL_SUCCESS);
|
||||
final Map<String, Integer> correlationCount = new HashMap<>();
|
||||
|
||||
for (final MockFlowFile flowFile : success) {
|
||||
flowFile.assertAttributeNotExists(String.format("%s_%s", "name", algorithm));
|
||||
flowFile.assertAttributeNotExists(String.format("%s_%s", "value", algorithm));
|
||||
Assert.assertEquals("abcdefg",flowFile.getAttribute("not_name"));
|
||||
Assert.assertEquals("hijklmnop",flowFile.getAttribute("not_value"));
|
||||
}
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue