Cleanup Bulk Delete Exception Logging (#41693) (#42606)

* Cleanup Bulk Delete Exception Logging

* Follow up to #41368
* Collect all failed blob deletes and add them to the exception message
* Remove logging of blob name list from caller exception logging
This commit is contained in:
Armin Braun 2019-05-28 11:00:28 +02:00 committed by GitHub
parent 44bf784fe1
commit 116b050cc6
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
3 changed files with 30 additions and 15 deletions

View File

@ -329,6 +329,7 @@ class GoogleCloudStorageBlobStore implements BlobStore {
if (e != null) { if (e != null) {
throw new IOException("Exception when deleting blobs [" + failedBlobs + "]", e); throw new IOException("Exception when deleting blobs [" + failedBlobs + "]", e);
} }
assert failedBlobs.isEmpty();
} }
private static String buildKey(String keyPath, String s) { private static String buildKey(String keyPath, String s) {

View File

@ -25,6 +25,7 @@ import com.amazonaws.services.s3.model.AmazonS3Exception;
import com.amazonaws.services.s3.model.CompleteMultipartUploadRequest; import com.amazonaws.services.s3.model.CompleteMultipartUploadRequest;
import com.amazonaws.services.s3.model.DeleteObjectsRequest; import com.amazonaws.services.s3.model.DeleteObjectsRequest;
import com.amazonaws.services.s3.model.InitiateMultipartUploadRequest; import com.amazonaws.services.s3.model.InitiateMultipartUploadRequest;
import com.amazonaws.services.s3.model.MultiObjectDeleteException;
import com.amazonaws.services.s3.model.ObjectListing; import com.amazonaws.services.s3.model.ObjectListing;
import com.amazonaws.services.s3.model.ObjectMetadata; import com.amazonaws.services.s3.model.ObjectMetadata;
import com.amazonaws.services.s3.model.PartETag; import com.amazonaws.services.s3.model.PartETag;
@ -34,6 +35,7 @@ import com.amazonaws.services.s3.model.S3ObjectSummary;
import com.amazonaws.services.s3.model.UploadPartRequest; import com.amazonaws.services.s3.model.UploadPartRequest;
import com.amazonaws.services.s3.model.UploadPartResult; import com.amazonaws.services.s3.model.UploadPartResult;
import org.apache.lucene.util.SetOnce; import org.apache.lucene.util.SetOnce;
import org.elasticsearch.ExceptionsHelper;
import org.elasticsearch.common.Nullable; import org.elasticsearch.common.Nullable;
import org.elasticsearch.common.Strings; import org.elasticsearch.common.Strings;
import org.elasticsearch.common.blobstore.BlobMetaData; import org.elasticsearch.common.blobstore.BlobMetaData;
@ -50,6 +52,8 @@ import java.nio.file.NoSuchFileException;
import java.util.ArrayList; import java.util.ArrayList;
import java.util.List; import java.util.List;
import java.util.Map; import java.util.Map;
import java.util.Set;
import java.util.stream.Collectors;
import static org.elasticsearch.repositories.s3.S3Repository.MAX_FILE_SIZE; import static org.elasticsearch.repositories.s3.S3Repository.MAX_FILE_SIZE;
import static org.elasticsearch.repositories.s3.S3Repository.MAX_FILE_SIZE_USING_MULTIPART; import static org.elasticsearch.repositories.s3.S3Repository.MAX_FILE_SIZE_USING_MULTIPART;
@ -127,12 +131,13 @@ class S3BlobContainer extends AbstractBlobContainer {
if (blobNames.isEmpty()) { if (blobNames.isEmpty()) {
return; return;
} }
final Set<String> outstanding = blobNames.stream().map(this::buildKey).collect(Collectors.toSet());
try (AmazonS3Reference clientReference = blobStore.clientReference()) { try (AmazonS3Reference clientReference = blobStore.clientReference()) {
// S3 API only allows 1k blobs per delete so we split up the given blobs into requests of max. 1k deletes // S3 API only allows 1k blobs per delete so we split up the given blobs into requests of max. 1k deletes
final List<DeleteObjectsRequest> deleteRequests = new ArrayList<>(); final List<DeleteObjectsRequest> deleteRequests = new ArrayList<>();
final List<String> partition = new ArrayList<>(); final List<String> partition = new ArrayList<>();
for (String blob : blobNames) { for (String key : outstanding) {
partition.add(buildKey(blob)); partition.add(key);
if (partition.size() == MAX_BULK_DELETES ) { if (partition.size() == MAX_BULK_DELETES ) {
deleteRequests.add(bulkDelete(blobStore.bucket(), partition)); deleteRequests.add(bulkDelete(blobStore.bucket(), partition));
partition.clear(); partition.clear();
@ -144,23 +149,32 @@ class S3BlobContainer extends AbstractBlobContainer {
SocketAccess.doPrivilegedVoid(() -> { SocketAccess.doPrivilegedVoid(() -> {
AmazonClientException aex = null; AmazonClientException aex = null;
for (DeleteObjectsRequest deleteRequest : deleteRequests) { for (DeleteObjectsRequest deleteRequest : deleteRequests) {
List<String> keysInRequest =
deleteRequest.getKeys().stream().map(DeleteObjectsRequest.KeyVersion::getKey).collect(Collectors.toList());
try { try {
clientReference.client().deleteObjects(deleteRequest); clientReference.client().deleteObjects(deleteRequest);
outstanding.removeAll(keysInRequest);
} catch (MultiObjectDeleteException e) {
// We are sending quiet mode requests so we can't use the deleted keys entry on the exception and instead
// first remove all keys that were sent in the request and then add back those that ran into an exception.
outstanding.removeAll(keysInRequest);
outstanding.addAll(
e.getErrors().stream().map(MultiObjectDeleteException.DeleteError::getKey).collect(Collectors.toSet()));
aex = ExceptionsHelper.useOrSuppress(aex, e);
} catch (AmazonClientException e) { } catch (AmazonClientException e) {
if (aex == null) { // The AWS client threw any unexpected exception and did not execute the request at all so we do not
aex = e; // remove any keys from the outstanding deletes set.
} else { aex = ExceptionsHelper.useOrSuppress(aex, e);
aex.addSuppressed(e);
}
} }
} }
if (aex != null) { if (aex != null) {
throw aex; throw aex;
} }
}); });
} catch (final AmazonClientException e) { } catch (Exception e) {
throw new IOException("Exception when deleting blobs [" + blobNames + "]", e); throw new IOException("Failed to delete blobs [" + outstanding + "]", e);
} }
assert outstanding.isEmpty();
} }
private static DeleteObjectsRequest bulkDelete(String bucket, List<String> blobs) { private static DeleteObjectsRequest bulkDelete(String bucket, List<String> blobs) {

View File

@ -998,8 +998,8 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
try { try {
blobContainer.deleteBlobsIgnoringIfNotExists(blobNames); blobContainer.deleteBlobsIgnoringIfNotExists(blobNames);
} catch (IOException e) { } catch (IOException e) {
logger.warn(() -> new ParameterizedMessage("[{}][{}] failed to delete index blobs {} during finalization", logger.warn(() -> new ParameterizedMessage("[{}][{}] failed to delete index blobs during finalization",
snapshotId, shardId, blobNames), e); snapshotId, shardId), e);
throw e; throw e;
} }
@ -1014,8 +1014,8 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
try { try {
blobContainer.deleteBlobsIgnoringIfNotExists(indexBlobs); blobContainer.deleteBlobsIgnoringIfNotExists(indexBlobs);
} catch (IOException e) { } catch (IOException e) {
logger.warn(() -> new ParameterizedMessage("[{}][{}] failed to delete index blobs {} during finalization", logger.warn(() -> new ParameterizedMessage("[{}][{}] failed to delete index blobs during finalization",
snapshotId, shardId, indexBlobs), e); snapshotId, shardId), e);
throw e; throw e;
} }
@ -1027,8 +1027,8 @@ public abstract class BlobStoreRepository extends AbstractLifecycleComponent imp
try { try {
blobContainer.deleteBlobsIgnoringIfNotExists(orphanedBlobs); blobContainer.deleteBlobsIgnoringIfNotExists(orphanedBlobs);
} catch (IOException e) { } catch (IOException e) {
logger.warn(() -> new ParameterizedMessage("[{}][{}] failed to delete data blobs {} during finalization", logger.warn(() -> new ParameterizedMessage("[{}][{}] failed to delete data blobs during finalization",
snapshotId, shardId, orphanedBlobs), e); snapshotId, shardId), e);
} }
} catch (IOException e) { } catch (IOException e) {
String message = String message =