separate segment mover and segment archiver

This commit is contained in:
Xavier Léauté 2013-12-13 16:40:57 -08:00
parent 123bddd615
commit 6b90372002
5 changed files with 82 additions and 18 deletions

View File

@ -41,7 +41,7 @@
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
<metamx.java-util.version>0.25.1</metamx.java-util.version>
<apache.curator.version>2.1.0-incubating</apache.curator.version>
<druid.api.version>0.1.6</druid.api.version>
<druid.api.version>0.1.7</druid.api.version>
</properties>
<modules>

View File

@ -0,0 +1,59 @@
/*
* Druid - a distributed column store.
* Copyright (C) 2012, 2013 Metamarkets Group Inc.
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version 2
* of the License, or (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
*/
package io.druid.storage.s3;
import com.google.common.collect.ImmutableMap;
import com.google.inject.Inject;
import com.metamx.common.MapUtils;
import io.druid.segment.loading.DataSegmentArchiver;
import io.druid.segment.loading.SegmentLoadingException;
import io.druid.timeline.DataSegment;
import org.jets3t.service.impl.rest.httpclient.RestS3Service;
public class S3DataSegmentArchiver extends S3DataSegmentMover implements DataSegmentArchiver
{
private final S3DataSegmentArchiverConfig config;
@Inject
public S3DataSegmentArchiver(
RestS3Service s3Client,
S3DataSegmentArchiverConfig config
)
{
super(s3Client);
this.config = config;
}
@Override
public DataSegment archive(DataSegment segment) throws SegmentLoadingException
{
String targetS3Bucket = config.getArchiveBucket();
String targetS3Path = MapUtils.getString(segment.getLoadSpec(), "key");
return move(
segment,
ImmutableMap.<String, Object>of(
"bucket", targetS3Bucket,
"key", targetS3Path
)
);
}
}

View File

@ -21,7 +21,7 @@ package io.druid.storage.s3;
import com.fasterxml.jackson.annotation.JsonProperty;
public class S3DataSegmentMoverConfig
public class S3DataSegmentArchiverConfig
{
@JsonProperty
public String archiveBucket = "";

View File

@ -37,20 +37,17 @@ public class S3DataSegmentMover implements DataSegmentMover
private static final Logger log = new Logger(S3DataSegmentMover.class);
private final RestS3Service s3Client;
private final S3DataSegmentMoverConfig config;
@Inject
public S3DataSegmentMover(
RestS3Service s3Client,
S3DataSegmentMoverConfig config
RestS3Service s3Client
)
{
this.s3Client = s3Client;
this.config = config;
}
@Override
public DataSegment move(DataSegment segment) throws SegmentLoadingException
public DataSegment move(DataSegment segment, Map<String, Object> targetLoadSpec) throws SegmentLoadingException
{
try {
Map<String, Object> loadSpec = segment.getLoadSpec();
@ -58,10 +55,15 @@ public class S3DataSegmentMover implements DataSegmentMover
String s3Path = MapUtils.getString(loadSpec, "key");
String s3DescriptorPath = S3Utils.descriptorPathForSegmentPath(s3Path);
final String s3ArchiveBucket = config.getArchiveBucket();
final String targetS3Bucket = MapUtils.getString(targetLoadSpec, "bucket");
final String targetS3Path = MapUtils.getString(targetLoadSpec, "key");
String targetS3DescriptorPath = S3Utils.descriptorPathForSegmentPath(targetS3Path);
if (s3ArchiveBucket.isEmpty()) {
throw new SegmentLoadingException("S3 archive bucket not specified");
if (targetS3Bucket.isEmpty()) {
throw new SegmentLoadingException("Target S3 bucket is not specified");
}
if (targetS3Path.isEmpty()) {
throw new SegmentLoadingException("Target S3 path is not specified");
}
if (s3Client.isObjectInBucket(s3Bucket, s3Path)) {
@ -69,26 +71,28 @@ public class S3DataSegmentMover implements DataSegmentMover
"Moving index file[s3://%s/%s] to [s3://%s/%s]",
s3Bucket,
s3Path,
s3ArchiveBucket,
s3Path
targetS3Bucket,
targetS3Path
);
s3Client.moveObject(s3Bucket, s3Path, s3ArchiveBucket, new S3Object(s3Path), false);
s3Client.moveObject(s3Bucket, s3Path, targetS3Bucket, new S3Object(targetS3Path), false);
}
if (s3Client.isObjectInBucket(s3Bucket, s3DescriptorPath)) {
log.info(
"Moving descriptor file[s3://%s/%s] to [s3://%s/%s]",
s3Bucket,
s3DescriptorPath,
s3ArchiveBucket,
s3DescriptorPath
targetS3Bucket,
targetS3DescriptorPath
);
s3Client.moveObject(s3Bucket, s3DescriptorPath, s3ArchiveBucket, new S3Object(s3DescriptorPath), false);
s3Client.moveObject(s3Bucket, s3DescriptorPath, targetS3Bucket, new S3Object(targetS3DescriptorPath), false);
}
return segment.withLoadSpec(
ImmutableMap.<String, Object>builder()
.putAll(loadSpec)
.put("bucket", s3ArchiveBucket).build()
.put("bucket", targetS3Bucket)
.put("key", targetS3Path)
.build()
);
}
catch (ServiceException e) {

View File

@ -52,9 +52,10 @@ public class S3StorageDruidModule implements DruidModule
Binders.dataSegmentPullerBinder(binder).addBinding("s3_zip").to(S3DataSegmentPuller.class).in(LazySingleton.class);
Binders.dataSegmentKillerBinder(binder).addBinding("s3_zip").to(S3DataSegmentKiller.class).in(LazySingleton.class);
Binders.dataSegmentMoverBinder(binder).addBinding("s3_zip").to(S3DataSegmentMover.class).in(LazySingleton.class);
Binders.dataSegmentArchiverBinder(binder).addBinding("s3_zip").to(S3DataSegmentArchiver.class).in(LazySingleton.class);
Binders.dataSegmentPusherBinder(binder).addBinding("s3").to(S3DataSegmentPusher.class).in(LazySingleton.class);
JsonConfigProvider.bind(binder, "druid.storage", S3DataSegmentPusherConfig.class);
JsonConfigProvider.bind(binder, "druid.storage", S3DataSegmentMoverConfig.class);
JsonConfigProvider.bind(binder, "druid.storage", S3DataSegmentArchiverConfig.class);
Binders.taskLogsBinder(binder).addBinding("s3").to(S3TaskLogs.class);
JsonConfigProvider.bind(binder, "druid.indexer.logs", S3TaskLogsConfig.class);