HDDS-1482. Use strongly typed codec implementations for the S3Table

Closes #789
This commit is contained in:
Bharat Viswanadham 2019-05-02 13:56:30 +02:00 committed by Márton Elek
parent e2f0f72677
commit 4605db369e
No known key found for this signature in database
GPG Key ID: D51EA8F00EE79B28
7 changed files with 191 additions and 35 deletions

View File

@ -25,6 +25,7 @@ import org.apache.hadoop.ozone.om.helpers.OmBucketInfo;
import org.apache.hadoop.ozone.om.helpers.OmKeyInfo; import org.apache.hadoop.ozone.om.helpers.OmKeyInfo;
import org.apache.hadoop.ozone.om.helpers.OmMultipartKeyInfo; import org.apache.hadoop.ozone.om.helpers.OmMultipartKeyInfo;
import org.apache.hadoop.ozone.om.helpers.OmVolumeArgs; import org.apache.hadoop.ozone.om.helpers.OmVolumeArgs;
import org.apache.hadoop.ozone.om.helpers.S3SecretValue;
import org.apache.hadoop.ozone.protocol.proto.OzoneManagerProtocolProtos.VolumeList; import org.apache.hadoop.ozone.protocol.proto.OzoneManagerProtocolProtos.VolumeList;
import org.apache.hadoop.ozone.security.OzoneTokenIdentifier; import org.apache.hadoop.ozone.security.OzoneTokenIdentifier;
import org.apache.hadoop.utils.db.DBStore; import org.apache.hadoop.utils.db.DBStore;
@ -259,7 +260,7 @@ public interface OMMetadataManager {
* @return Table. * @return Table.
*/ */
Table<byte[], byte[]> getS3Table(); Table<String, String> getS3Table();
/** /**
* Returns the DB key name of a multipart upload key in OM metadata store. * Returns the DB key name of a multipart upload key in OM metadata store.
@ -285,7 +286,7 @@ public interface OMMetadataManager {
* Gets the S3 Secrets table. * Gets the S3 Secrets table.
* @return Table * @return Table
*/ */
Table<byte[], byte[]> getS3SecretTable(); Table<String, S3SecretValue> getS3SecretTable();
/** /**
* Returns number of rows in a table. This should not be used for very * Returns number of rows in a table. This should not be used for very

View File

@ -23,7 +23,6 @@ import org.apache.commons.codec.digest.DigestUtils;
import org.apache.hadoop.hdds.conf.OzoneConfiguration; import org.apache.hadoop.hdds.conf.OzoneConfiguration;
import org.apache.hadoop.ozone.OmUtils; import org.apache.hadoop.ozone.OmUtils;
import org.apache.hadoop.ozone.om.helpers.S3SecretValue; import org.apache.hadoop.ozone.om.helpers.S3SecretValue;
import org.apache.hadoop.ozone.protocol.proto.OzoneManagerProtocolProtos;
import org.apache.hadoop.ozone.security.OzoneSecurityException; import org.apache.hadoop.ozone.security.OzoneSecurityException;
import org.apache.logging.log4j.util.Strings; import org.apache.logging.log4j.util.Strings;
import org.slf4j.Logger; import org.slf4j.Logger;
@ -31,7 +30,6 @@ import org.slf4j.LoggerFactory;
import java.io.IOException; import java.io.IOException;
import static java.nio.charset.StandardCharsets.UTF_8;
import static org.apache.hadoop.ozone.security.OzoneSecurityException.ResultCodes.S3_SECRET_NOT_FOUND; import static org.apache.hadoop.ozone.security.OzoneSecurityException.ResultCodes.S3_SECRET_NOT_FOUND;
/** /**
@ -61,20 +59,17 @@ public class S3SecretManagerImpl implements S3SecretManager {
public S3SecretValue getS3Secret(String kerberosID) throws IOException { public S3SecretValue getS3Secret(String kerberosID) throws IOException {
Preconditions.checkArgument(Strings.isNotBlank(kerberosID), Preconditions.checkArgument(Strings.isNotBlank(kerberosID),
"kerberosID cannot be null or empty."); "kerberosID cannot be null or empty.");
byte[] awsAccessKey = kerberosID.getBytes(UTF_8);
S3SecretValue result = null; S3SecretValue result = null;
omMetadataManager.getLock().acquireS3SecretLock(kerberosID); omMetadataManager.getLock().acquireS3SecretLock(kerberosID);
try { try {
byte[] s3Secret = S3SecretValue s3Secret =
omMetadataManager.getS3SecretTable().get(awsAccessKey); omMetadataManager.getS3SecretTable().get(kerberosID);
if(s3Secret == null) { if(s3Secret == null) {
byte[] secret = OmUtils.getSHADigest(); byte[] secret = OmUtils.getSHADigest();
result = new S3SecretValue(kerberosID, DigestUtils.sha256Hex(secret)); result = new S3SecretValue(kerberosID, DigestUtils.sha256Hex(secret));
omMetadataManager.getS3SecretTable() omMetadataManager.getS3SecretTable().put(kerberosID, result);
.put(awsAccessKey, result.getProtobuf().toByteArray());
} else { } else {
result = S3SecretValue.fromProtobuf( return s3Secret;
OzoneManagerProtocolProtos.S3Secret.parseFrom(s3Secret));
} }
} finally { } finally {
omMetadataManager.getLock().releaseS3SecretLock(kerberosID); omMetadataManager.getLock().releaseS3SecretLock(kerberosID);
@ -90,11 +85,10 @@ public class S3SecretManagerImpl implements S3SecretManager {
"awsAccessKeyId cannot be null or empty."); "awsAccessKeyId cannot be null or empty.");
LOG.trace("Get secret for awsAccessKey:{}", kerberosID); LOG.trace("Get secret for awsAccessKey:{}", kerberosID);
byte[] s3Secret; S3SecretValue s3Secret;
omMetadataManager.getLock().acquireS3SecretLock(kerberosID); omMetadataManager.getLock().acquireS3SecretLock(kerberosID);
try { try {
s3Secret = omMetadataManager.getS3SecretTable() s3Secret = omMetadataManager.getS3SecretTable().get(kerberosID);
.get(kerberosID.getBytes(UTF_8));
if (s3Secret == null) { if (s3Secret == null) {
throw new OzoneSecurityException("S3 secret not found for " + throw new OzoneSecurityException("S3 secret not found for " +
"awsAccessKeyId " + kerberosID, S3_SECRET_NOT_FOUND); "awsAccessKeyId " + kerberosID, S3_SECRET_NOT_FOUND);
@ -103,8 +97,7 @@ public class S3SecretManagerImpl implements S3SecretManager {
omMetadataManager.getLock().releaseS3SecretLock(kerberosID); omMetadataManager.getLock().releaseS3SecretLock(kerberosID);
} }
return OzoneManagerProtocolProtos.S3Secret.parseFrom(s3Secret) return s3Secret.getAwsSecret();
.getAwsSecret();
} }
public OMMetadataManager getOmMetadataManager() { public OMMetadataManager getOmMetadataManager() {

View File

@ -0,0 +1,57 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
* <p>
* http://www.apache.org/licenses/LICENSE-2.0
* <p>
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.ozone.om.codec;
import java.io.IOException;
import com.google.common.base.Preconditions;
import com.google.protobuf.InvalidProtocolBufferException;
import org.apache.hadoop.ozone.om.helpers.S3SecretValue;
import org.apache.hadoop.ozone.protocol.proto.OzoneManagerProtocolProtos;
import org.apache.hadoop.utils.db.Codec;
/**
* Codec to encode S3SecretValue as byte array.
*/
public class S3SecretValueCodec implements Codec<S3SecretValue> {
@Override
public byte[] toPersistedFormat(S3SecretValue object) throws IOException {
Preconditions
.checkNotNull(object, "Null object can't be converted to byte array.");
return object.getProtobuf().toByteArray();
}
@Override
public S3SecretValue fromPersistedFormat(byte[] rawData) throws IOException {
Preconditions
.checkNotNull(rawData,
"Null byte array can't converted to real object.");
try {
return S3SecretValue.fromProtobuf(
OzoneManagerProtocolProtos.S3Secret.parseFrom(rawData));
} catch (InvalidProtocolBufferException e) {
throw new IllegalArgumentException(
"Can't encode the the raw data from the byte array", e);
}
}
}

View File

@ -19,6 +19,8 @@ package org.apache.hadoop.ozone.om.helpers;
import org.apache.hadoop.ozone.protocol.proto.OzoneManagerProtocolProtos; import org.apache.hadoop.ozone.protocol.proto.OzoneManagerProtocolProtos;
import java.util.Objects;
/** /**
* S3Secret to be saved in database. * S3Secret to be saved in database.
*/ */
@ -67,4 +69,22 @@ public class S3SecretValue {
public String toString() { public String toString() {
return "awsAccessKey=" + kerberosID + "\nawsSecret=" + awsSecret; return "awsAccessKey=" + kerberosID + "\nawsSecret=" + awsSecret;
} }
@Override
public boolean equals(Object o) {
if (this == o) {
return true;
}
if (o == null || getClass() != o.getClass()) {
return false;
}
S3SecretValue that = (S3SecretValue) o;
return kerberosID.equals(that.kerberosID) &&
awsSecret.equals(that.awsSecret);
}
@Override
public int hashCode() {
return Objects.hash(kerberosID, awsSecret);
}
} }

View File

@ -0,0 +1,88 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
* <p>
* http://www.apache.org/licenses/LICENSE-2.0
* <p>
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.hadoop.ozone.om.codec;
import java.nio.charset.StandardCharsets;
import java.util.UUID;
import org.junit.Assert;
import org.junit.Before;
import org.junit.Rule;
import org.junit.Test;
import org.junit.rules.ExpectedException;
import org.apache.hadoop.ozone.om.helpers.S3SecretValue;
import org.apache.hadoop.test.GenericTestUtils;
import static org.junit.Assert.fail;
/**
* This class test S3SecretValueCodec.
*/
public class TestS3SecretValueCodec {
@Rule
public ExpectedException thrown = ExpectedException.none();
private S3SecretValueCodec codec;
@Before
public void initialize() {
codec = new S3SecretValueCodec();
}
@Test
public void testCodecWithCorrectData() throws Exception {
S3SecretValue s3SecretValue =
new S3SecretValue(UUID.randomUUID().toString(),
UUID.randomUUID().toString());
byte[] data = codec.toPersistedFormat(s3SecretValue);
Assert.assertNotNull(data);
S3SecretValue docdedS3Secret = codec.fromPersistedFormat(data);
Assert.assertEquals(s3SecretValue, docdedS3Secret);
}
@Test
public void testCodecWithIncorrectValues() throws Exception {
try {
codec.fromPersistedFormat("random".getBytes(StandardCharsets.UTF_8));
fail("testCodecWithIncorrectValues failed");
} catch (IllegalArgumentException ex) {
GenericTestUtils.assertExceptionContains("Can't encode the the raw " +
"data from the byte array", ex);
}
}
@Test
public void testCodecWithNullDataFromTable() throws Exception {
thrown.expect(NullPointerException.class);
codec.fromPersistedFormat(null);
}
@Test
public void testCodecWithNullDataFromUser() throws Exception {
thrown.expect(NullPointerException.class);
codec.toPersistedFormat(null);
}
}

View File

@ -32,6 +32,7 @@ import org.apache.hadoop.ozone.om.codec.OmBucketInfoCodec;
import org.apache.hadoop.ozone.om.codec.OmKeyInfoCodec; import org.apache.hadoop.ozone.om.codec.OmKeyInfoCodec;
import org.apache.hadoop.ozone.om.codec.OmMultipartKeyInfoCodec; import org.apache.hadoop.ozone.om.codec.OmMultipartKeyInfoCodec;
import org.apache.hadoop.ozone.om.codec.OmVolumeArgsCodec; import org.apache.hadoop.ozone.om.codec.OmVolumeArgsCodec;
import org.apache.hadoop.ozone.om.codec.S3SecretValueCodec;
import org.apache.hadoop.ozone.om.codec.TokenIdentifierCodec; import org.apache.hadoop.ozone.om.codec.TokenIdentifierCodec;
import org.apache.hadoop.ozone.om.codec.VolumeListCodec; import org.apache.hadoop.ozone.om.codec.VolumeListCodec;
import org.apache.hadoop.ozone.om.exceptions.OMException; import org.apache.hadoop.ozone.om.exceptions.OMException;
@ -41,6 +42,7 @@ import org.apache.hadoop.ozone.om.helpers.OmKeyInfo;
import org.apache.hadoop.ozone.om.helpers.OmKeyLocationInfoGroup; import org.apache.hadoop.ozone.om.helpers.OmKeyLocationInfoGroup;
import org.apache.hadoop.ozone.om.helpers.OmMultipartKeyInfo; import org.apache.hadoop.ozone.om.helpers.OmMultipartKeyInfo;
import org.apache.hadoop.ozone.om.helpers.OmVolumeArgs; import org.apache.hadoop.ozone.om.helpers.OmVolumeArgs;
import org.apache.hadoop.ozone.om.helpers.S3SecretValue;
import org.apache.hadoop.ozone.protocol.proto.OzoneManagerProtocolProtos.VolumeList; import org.apache.hadoop.ozone.protocol.proto.OzoneManagerProtocolProtos.VolumeList;
import org.apache.hadoop.ozone.security.OzoneTokenIdentifier; import org.apache.hadoop.ozone.security.OzoneTokenIdentifier;
import org.apache.hadoop.utils.db.DBStore; import org.apache.hadoop.utils.db.DBStore;
@ -176,7 +178,7 @@ public class OmMetadataManagerImpl implements OMMetadataManager {
} }
@Override @Override
public Table<byte[], byte[]> getS3Table() { public Table<String, String> getS3Table() {
return s3Table; return s3Table;
} }
@ -232,7 +234,8 @@ public class OmMetadataManagerImpl implements OMMetadataManager {
.addCodec(OmBucketInfo.class, new OmBucketInfoCodec()) .addCodec(OmBucketInfo.class, new OmBucketInfoCodec())
.addCodec(OmVolumeArgs.class, new OmVolumeArgsCodec()) .addCodec(OmVolumeArgs.class, new OmVolumeArgsCodec())
.addCodec(VolumeList.class, new VolumeListCodec()) .addCodec(VolumeList.class, new VolumeListCodec())
.addCodec(OmMultipartKeyInfo.class, new OmMultipartKeyInfoCodec()); .addCodec(OmMultipartKeyInfo.class, new OmMultipartKeyInfoCodec())
.addCodec(S3SecretValue.class, new S3SecretValueCodec());
} }
/** /**
@ -265,7 +268,7 @@ public class OmMetadataManagerImpl implements OMMetadataManager {
this.store.getTable(OPEN_KEY_TABLE, String.class, OmKeyInfo.class); this.store.getTable(OPEN_KEY_TABLE, String.class, OmKeyInfo.class);
checkTableStatus(openKeyTable, OPEN_KEY_TABLE); checkTableStatus(openKeyTable, OPEN_KEY_TABLE);
s3Table = this.store.getTable(S3_TABLE); s3Table = this.store.getTable(S3_TABLE, String.class, String.class);
checkTableStatus(s3Table, S3_TABLE); checkTableStatus(s3Table, S3_TABLE);
multipartInfoTable = this.store.getTable(MULTIPARTINFO_TABLE, multipartInfoTable = this.store.getTable(MULTIPARTINFO_TABLE,
@ -276,7 +279,8 @@ public class OmMetadataManagerImpl implements OMMetadataManager {
OzoneTokenIdentifier.class, Long.class); OzoneTokenIdentifier.class, Long.class);
checkTableStatus(dTokenTable, DELEGATION_TOKEN_TABLE); checkTableStatus(dTokenTable, DELEGATION_TOKEN_TABLE);
s3SecretTable = this.store.getTable(S3_SECRET_TABLE); s3SecretTable = this.store.getTable(S3_SECRET_TABLE, String.class,
S3SecretValue.class);
checkTableStatus(s3SecretTable, S3_SECRET_TABLE); checkTableStatus(s3SecretTable, S3_SECRET_TABLE);
} }
@ -701,7 +705,7 @@ public class OmMetadataManagerImpl implements OMMetadataManager {
} }
@Override @Override
public Table<byte[], byte[]> getS3SecretTable() { public Table<String, S3SecretValue> getS3SecretTable() {
return s3SecretTable; return s3SecretTable;
} }

View File

@ -32,7 +32,6 @@ import org.slf4j.Logger;
import org.slf4j.LoggerFactory; import org.slf4j.LoggerFactory;
import java.io.IOException; import java.io.IOException;
import java.nio.charset.StandardCharsets;
import java.util.Objects; import java.util.Objects;
import static org.apache.hadoop.ozone.OzoneConsts.OM_S3_VOLUME_PREFIX; import static org.apache.hadoop.ozone.OzoneConsts.OM_S3_VOLUME_PREFIX;
@ -104,9 +103,8 @@ public class S3BucketManagerImpl implements S3BucketManager {
omMetadataManager.getLock().acquireS3Lock(bucketName); omMetadataManager.getLock().acquireS3Lock(bucketName);
try { try {
byte[] bucket = String bucket =
omMetadataManager.getS3Table().get( omMetadataManager.getS3Table().get(bucketName);
bucketName.getBytes(StandardCharsets.UTF_8));
if (bucket != null) { if (bucket != null) {
LOG.debug("Bucket already exists. {}", bucketName); LOG.debug("Bucket already exists. {}", bucketName);
@ -119,9 +117,7 @@ public class S3BucketManagerImpl implements S3BucketManager {
String finalName = String.format("%s/%s", ozoneVolumeName, String finalName = String.format("%s/%s", ozoneVolumeName,
ozoneBucketName); ozoneBucketName);
omMetadataManager.getS3Table().put( omMetadataManager.getS3Table().put(bucketName, finalName);
bucketName.getBytes(StandardCharsets.UTF_8),
finalName.getBytes(StandardCharsets.UTF_8));
} finally { } finally {
omMetadataManager.getLock().releaseS3Lock(bucketName); omMetadataManager.getLock().releaseS3Lock(bucketName);
} }
@ -134,8 +130,7 @@ public class S3BucketManagerImpl implements S3BucketManager {
omMetadataManager.getLock().acquireS3Lock(bucketName); omMetadataManager.getLock().acquireS3Lock(bucketName);
try { try {
byte[] bucket = bucketName.getBytes(StandardCharsets.UTF_8); String map = omMetadataManager.getS3Table().get(bucketName);
byte[] map = omMetadataManager.getS3Table().get(bucket);
if (map == null) { if (map == null) {
throw new OMException("No such S3 bucket. " + bucketName, throw new OMException("No such S3 bucket. " + bucketName,
@ -149,7 +144,7 @@ public class S3BucketManagerImpl implements S3BucketManager {
} else { } else {
bucketManager.deleteBucket(getOzoneVolumeName(bucketName), bucketName); bucketManager.deleteBucket(getOzoneVolumeName(bucketName), bucketName);
} }
omMetadataManager.getS3Table().delete(bucket); omMetadataManager.getS3Table().delete(bucketName);
} catch(IOException ex) { } catch(IOException ex) {
throw ex; throw ex;
} finally { } finally {
@ -224,11 +219,9 @@ public class S3BucketManagerImpl implements S3BucketManager {
"Length of the S3 Bucket is not correct."); "Length of the S3 Bucket is not correct.");
omMetadataManager.getLock().acquireS3Lock(s3BucketName); omMetadataManager.getLock().acquireS3Lock(s3BucketName);
try { try {
byte[] mapping = String mapping = omMetadataManager.getS3Table().get(s3BucketName);
omMetadataManager.getS3Table().get(
s3BucketName.getBytes(StandardCharsets.UTF_8));
if (mapping != null) { if (mapping != null) {
return new String(mapping, StandardCharsets.UTF_8); return mapping;
} }
throw new OMException("No such S3 bucket.", throw new OMException("No such S3 bucket.",
OMException.ResultCodes.S3_BUCKET_NOT_FOUND); OMException.ResultCodes.S3_BUCKET_NOT_FOUND);