HDFS-7404. Remove o.a.h.hdfs.server.datanode.web.resources. Contributed by Li Lu.
This commit is contained in:
parent
541172f2ee
commit
ad32886ae6
|
@ -112,6 +112,9 @@ Release 2.7.0 - UNRELEASED
|
||||||
|
|
||||||
HDFS-7279. Use netty to implement DatanodeWebHdfsMethods. (wheat9)
|
HDFS-7279. Use netty to implement DatanodeWebHdfsMethods. (wheat9)
|
||||||
|
|
||||||
|
HDFS-7404. Remove o.a.h.hdfs.server.datanode.web.resources.
|
||||||
|
(Li Lu via wheat9)
|
||||||
|
|
||||||
OPTIMIZATIONS
|
OPTIMIZATIONS
|
||||||
|
|
||||||
BUG FIXES
|
BUG FIXES
|
||||||
|
|
|
@ -1,477 +0,0 @@
|
||||||
/**
|
|
||||||
* Licensed to the Apache Software Foundation (ASF) under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. The ASF licenses this file
|
|
||||||
* to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
package org.apache.hadoop.hdfs.server.datanode.web.resources;
|
|
||||||
|
|
||||||
import java.io.IOException;
|
|
||||||
import java.io.InputStream;
|
|
||||||
import java.net.URI;
|
|
||||||
import java.net.URISyntaxException;
|
|
||||||
import java.security.PrivilegedExceptionAction;
|
|
||||||
import java.util.EnumSet;
|
|
||||||
|
|
||||||
import javax.servlet.ServletContext;
|
|
||||||
import javax.servlet.http.HttpServletRequest;
|
|
||||||
import javax.servlet.http.HttpServletResponse;
|
|
||||||
import javax.ws.rs.Consumes;
|
|
||||||
import javax.ws.rs.DefaultValue;
|
|
||||||
import javax.ws.rs.GET;
|
|
||||||
import javax.ws.rs.POST;
|
|
||||||
import javax.ws.rs.PUT;
|
|
||||||
import javax.ws.rs.Path;
|
|
||||||
import javax.ws.rs.PathParam;
|
|
||||||
import javax.ws.rs.Produces;
|
|
||||||
import javax.ws.rs.QueryParam;
|
|
||||||
import javax.ws.rs.core.Context;
|
|
||||||
import javax.ws.rs.core.MediaType;
|
|
||||||
import javax.ws.rs.core.Response;
|
|
||||||
|
|
||||||
import com.google.common.annotations.VisibleForTesting;
|
|
||||||
import org.apache.commons.logging.Log;
|
|
||||||
import org.apache.commons.logging.LogFactory;
|
|
||||||
import org.apache.hadoop.conf.Configuration;
|
|
||||||
import org.apache.hadoop.fs.CreateFlag;
|
|
||||||
import org.apache.hadoop.fs.FSDataOutputStream;
|
|
||||||
import org.apache.hadoop.fs.MD5MD5CRC32FileChecksum;
|
|
||||||
import org.apache.hadoop.fs.permission.FsPermission;
|
|
||||||
import org.apache.hadoop.hdfs.DFSClient;
|
|
||||||
import org.apache.hadoop.hdfs.HAUtil;
|
|
||||||
import org.apache.hadoop.hdfs.client.HdfsDataInputStream;
|
|
||||||
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
|
|
||||||
import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier;
|
|
||||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
|
||||||
import org.apache.hadoop.hdfs.web.JsonUtil;
|
|
||||||
import org.apache.hadoop.hdfs.web.ParamFilter;
|
|
||||||
import org.apache.hadoop.hdfs.web.SWebHdfsFileSystem;
|
|
||||||
import org.apache.hadoop.hdfs.web.WebHdfsFileSystem;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.BlockSizeParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.BufferSizeParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.DelegationParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.GetOpParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.HttpOpParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.LengthParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.NamenodeAddressParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.OffsetParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.OverwriteParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.Param;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.PermissionParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.PostOpParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.PutOpParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.ReplicationParam;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.UriFsPathParam;
|
|
||||||
import org.apache.hadoop.io.IOUtils;
|
|
||||||
import org.apache.hadoop.security.SecurityUtil;
|
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
|
||||||
import org.apache.hadoop.security.token.Token;
|
|
||||||
|
|
||||||
import com.sun.jersey.spi.container.ResourceFilters;
|
|
||||||
|
|
||||||
/** Web-hdfs DataNode implementation. */
|
|
||||||
@Path("")
|
|
||||||
@ResourceFilters(ParamFilter.class)
|
|
||||||
public class DatanodeWebHdfsMethods {
|
|
||||||
public static final Log LOG = LogFactory.getLog(DatanodeWebHdfsMethods.class);
|
|
||||||
|
|
||||||
private static final UriFsPathParam ROOT = new UriFsPathParam("");
|
|
||||||
|
|
||||||
private @Context ServletContext context;
|
|
||||||
private @Context HttpServletRequest request;
|
|
||||||
private @Context HttpServletResponse response;
|
|
||||||
|
|
||||||
private void init(final UserGroupInformation ugi,
|
|
||||||
final DelegationParam delegation, final String nnId,
|
|
||||||
final UriFsPathParam path, final HttpOpParam<?> op,
|
|
||||||
final Param<?, ?>... parameters) throws IOException {
|
|
||||||
if (LOG.isTraceEnabled()) {
|
|
||||||
LOG.trace("HTTP " + op.getValue().getType() + ": " + op + ", " + path
|
|
||||||
+ ", ugi=" + ugi + Param.toSortedString(", ", parameters));
|
|
||||||
}
|
|
||||||
if (nnId == null) {
|
|
||||||
throw new IllegalArgumentException(NamenodeAddressParam.NAME
|
|
||||||
+ " is not specified.");
|
|
||||||
}
|
|
||||||
|
|
||||||
//clear content type
|
|
||||||
response.setContentType(null);
|
|
||||||
|
|
||||||
if (UserGroupInformation.isSecurityEnabled()) {
|
|
||||||
//add a token for RPC.
|
|
||||||
final Token<DelegationTokenIdentifier> token = deserializeToken
|
|
||||||
(delegation.getValue(), nnId);
|
|
||||||
ugi.addToken(token);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
@VisibleForTesting
|
|
||||||
Token<DelegationTokenIdentifier> deserializeToken
|
|
||||||
(String delegation,String nnId) throws IOException {
|
|
||||||
final DataNode datanode = (DataNode) context.getAttribute("datanode");
|
|
||||||
final Configuration conf = datanode.getConf();
|
|
||||||
final Token<DelegationTokenIdentifier> token = new
|
|
||||||
Token<DelegationTokenIdentifier>();
|
|
||||||
token.decodeFromUrlString(delegation);
|
|
||||||
URI nnUri = URI.create(HdfsConstants.HDFS_URI_SCHEME +
|
|
||||||
"://" + nnId);
|
|
||||||
boolean isLogical = HAUtil.isLogicalUri(conf, nnUri);
|
|
||||||
if (isLogical) {
|
|
||||||
token.setService(HAUtil.buildTokenServiceForLogicalUri(nnUri,
|
|
||||||
HdfsConstants.HDFS_URI_SCHEME));
|
|
||||||
} else {
|
|
||||||
token.setService(SecurityUtil.buildTokenService(nnUri));
|
|
||||||
}
|
|
||||||
token.setKind(DelegationTokenIdentifier.HDFS_DELEGATION_KIND);
|
|
||||||
return token;
|
|
||||||
}
|
|
||||||
|
|
||||||
/** Handle HTTP PUT request for the root. */
|
|
||||||
@PUT
|
|
||||||
@Path("/")
|
|
||||||
@Consumes({"*/*"})
|
|
||||||
@Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
|
|
||||||
public Response putRoot(
|
|
||||||
final InputStream in,
|
|
||||||
@Context final UserGroupInformation ugi,
|
|
||||||
@QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
|
|
||||||
final DelegationParam delegation,
|
|
||||||
@QueryParam(NamenodeAddressParam.NAME)
|
|
||||||
@DefaultValue(NamenodeAddressParam.DEFAULT)
|
|
||||||
final NamenodeAddressParam namenode,
|
|
||||||
@QueryParam(PutOpParam.NAME) @DefaultValue(PutOpParam.DEFAULT)
|
|
||||||
final PutOpParam op,
|
|
||||||
@QueryParam(PermissionParam.NAME) @DefaultValue(PermissionParam.DEFAULT)
|
|
||||||
final PermissionParam permission,
|
|
||||||
@QueryParam(OverwriteParam.NAME) @DefaultValue(OverwriteParam.DEFAULT)
|
|
||||||
final OverwriteParam overwrite,
|
|
||||||
@QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
|
|
||||||
final BufferSizeParam bufferSize,
|
|
||||||
@QueryParam(ReplicationParam.NAME) @DefaultValue(ReplicationParam.DEFAULT)
|
|
||||||
final ReplicationParam replication,
|
|
||||||
@QueryParam(BlockSizeParam.NAME) @DefaultValue(BlockSizeParam.DEFAULT)
|
|
||||||
final BlockSizeParam blockSize
|
|
||||||
) throws IOException, InterruptedException {
|
|
||||||
return put(in, ugi, delegation, namenode, ROOT, op, permission,
|
|
||||||
overwrite, bufferSize, replication, blockSize);
|
|
||||||
}
|
|
||||||
|
|
||||||
/** Handle HTTP PUT request. */
|
|
||||||
@PUT
|
|
||||||
@Path("{" + UriFsPathParam.NAME + ":.*}")
|
|
||||||
@Consumes({"*/*"})
|
|
||||||
@Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
|
|
||||||
public Response put(
|
|
||||||
final InputStream in,
|
|
||||||
@Context final UserGroupInformation ugi,
|
|
||||||
@QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
|
|
||||||
final DelegationParam delegation,
|
|
||||||
@QueryParam(NamenodeAddressParam.NAME)
|
|
||||||
@DefaultValue(NamenodeAddressParam.DEFAULT)
|
|
||||||
final NamenodeAddressParam namenode,
|
|
||||||
@PathParam(UriFsPathParam.NAME) final UriFsPathParam path,
|
|
||||||
@QueryParam(PutOpParam.NAME) @DefaultValue(PutOpParam.DEFAULT)
|
|
||||||
final PutOpParam op,
|
|
||||||
@QueryParam(PermissionParam.NAME) @DefaultValue(PermissionParam.DEFAULT)
|
|
||||||
final PermissionParam permission,
|
|
||||||
@QueryParam(OverwriteParam.NAME) @DefaultValue(OverwriteParam.DEFAULT)
|
|
||||||
final OverwriteParam overwrite,
|
|
||||||
@QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
|
|
||||||
final BufferSizeParam bufferSize,
|
|
||||||
@QueryParam(ReplicationParam.NAME) @DefaultValue(ReplicationParam.DEFAULT)
|
|
||||||
final ReplicationParam replication,
|
|
||||||
@QueryParam(BlockSizeParam.NAME) @DefaultValue(BlockSizeParam.DEFAULT)
|
|
||||||
final BlockSizeParam blockSize
|
|
||||||
) throws IOException, InterruptedException {
|
|
||||||
|
|
||||||
final String nnId = namenode.getValue();
|
|
||||||
init(ugi, delegation, nnId, path, op, permission,
|
|
||||||
overwrite, bufferSize, replication, blockSize);
|
|
||||||
|
|
||||||
return ugi.doAs(new PrivilegedExceptionAction<Response>() {
|
|
||||||
@Override
|
|
||||||
public Response run() throws IOException, URISyntaxException {
|
|
||||||
return put(in, nnId, path.getAbsolutePath(), op,
|
|
||||||
permission, overwrite, bufferSize, replication, blockSize);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
private Response put(
|
|
||||||
final InputStream in,
|
|
||||||
final String nnId,
|
|
||||||
final String fullpath,
|
|
||||||
final PutOpParam op,
|
|
||||||
final PermissionParam permission,
|
|
||||||
final OverwriteParam overwrite,
|
|
||||||
final BufferSizeParam bufferSize,
|
|
||||||
final ReplicationParam replication,
|
|
||||||
final BlockSizeParam blockSize
|
|
||||||
) throws IOException, URISyntaxException {
|
|
||||||
final DataNode datanode = (DataNode)context.getAttribute("datanode");
|
|
||||||
|
|
||||||
switch(op.getValue()) {
|
|
||||||
case CREATE:
|
|
||||||
{
|
|
||||||
final Configuration conf = new Configuration(datanode.getConf());
|
|
||||||
conf.set(FsPermission.UMASK_LABEL, "000");
|
|
||||||
|
|
||||||
final int b = bufferSize.getValue(conf);
|
|
||||||
DFSClient dfsclient = newDfsClient(nnId, conf);
|
|
||||||
FSDataOutputStream out = null;
|
|
||||||
try {
|
|
||||||
out = dfsclient.createWrappedOutputStream(dfsclient.create(
|
|
||||||
fullpath, permission.getFsPermission(),
|
|
||||||
overwrite.getValue() ?
|
|
||||||
EnumSet.of(CreateFlag.CREATE, CreateFlag.OVERWRITE) :
|
|
||||||
EnumSet.of(CreateFlag.CREATE),
|
|
||||||
replication.getValue(conf), blockSize.getValue(conf), null,
|
|
||||||
b, null), null);
|
|
||||||
IOUtils.copyBytes(in, out, b);
|
|
||||||
out.close();
|
|
||||||
out = null;
|
|
||||||
dfsclient.close();
|
|
||||||
dfsclient = null;
|
|
||||||
} finally {
|
|
||||||
IOUtils.cleanup(LOG, out);
|
|
||||||
IOUtils.cleanup(LOG, dfsclient);
|
|
||||||
}
|
|
||||||
final String scheme = "http".equals(request.getScheme()) ?
|
|
||||||
WebHdfsFileSystem.SCHEME : SWebHdfsFileSystem.SCHEME;
|
|
||||||
final URI uri = new URI(scheme, nnId, fullpath, null, null);
|
|
||||||
return Response.created(uri).type(MediaType.APPLICATION_OCTET_STREAM).build();
|
|
||||||
}
|
|
||||||
default:
|
|
||||||
throw new UnsupportedOperationException(op + " is not supported");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/** Handle HTTP POST request for the root for the root. */
|
|
||||||
@POST
|
|
||||||
@Path("/")
|
|
||||||
@Consumes({"*/*"})
|
|
||||||
@Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
|
|
||||||
public Response postRoot(
|
|
||||||
final InputStream in,
|
|
||||||
@Context final UserGroupInformation ugi,
|
|
||||||
@QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
|
|
||||||
final DelegationParam delegation,
|
|
||||||
@QueryParam(NamenodeAddressParam.NAME)
|
|
||||||
@DefaultValue(NamenodeAddressParam.DEFAULT)
|
|
||||||
final NamenodeAddressParam namenode,
|
|
||||||
@QueryParam(PostOpParam.NAME) @DefaultValue(PostOpParam.DEFAULT)
|
|
||||||
final PostOpParam op,
|
|
||||||
@QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
|
|
||||||
final BufferSizeParam bufferSize
|
|
||||||
) throws IOException, InterruptedException {
|
|
||||||
return post(in, ugi, delegation, namenode, ROOT, op, bufferSize);
|
|
||||||
}
|
|
||||||
|
|
||||||
/** Handle HTTP POST request. */
|
|
||||||
@POST
|
|
||||||
@Path("{" + UriFsPathParam.NAME + ":.*}")
|
|
||||||
@Consumes({"*/*"})
|
|
||||||
@Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
|
|
||||||
public Response post(
|
|
||||||
final InputStream in,
|
|
||||||
@Context final UserGroupInformation ugi,
|
|
||||||
@QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
|
|
||||||
final DelegationParam delegation,
|
|
||||||
@QueryParam(NamenodeAddressParam.NAME)
|
|
||||||
@DefaultValue(NamenodeAddressParam.DEFAULT)
|
|
||||||
final NamenodeAddressParam namenode,
|
|
||||||
@PathParam(UriFsPathParam.NAME) final UriFsPathParam path,
|
|
||||||
@QueryParam(PostOpParam.NAME) @DefaultValue(PostOpParam.DEFAULT)
|
|
||||||
final PostOpParam op,
|
|
||||||
@QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
|
|
||||||
final BufferSizeParam bufferSize
|
|
||||||
) throws IOException, InterruptedException {
|
|
||||||
|
|
||||||
final String nnId = namenode.getValue();
|
|
||||||
init(ugi, delegation, nnId, path, op, bufferSize);
|
|
||||||
|
|
||||||
return ugi.doAs(new PrivilegedExceptionAction<Response>() {
|
|
||||||
@Override
|
|
||||||
public Response run() throws IOException {
|
|
||||||
return post(in, nnId, path.getAbsolutePath(), op,
|
|
||||||
bufferSize);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
private Response post(
|
|
||||||
final InputStream in,
|
|
||||||
final String nnId,
|
|
||||||
final String fullpath,
|
|
||||||
final PostOpParam op,
|
|
||||||
final BufferSizeParam bufferSize
|
|
||||||
) throws IOException {
|
|
||||||
final DataNode datanode = (DataNode)context.getAttribute("datanode");
|
|
||||||
|
|
||||||
switch(op.getValue()) {
|
|
||||||
case APPEND:
|
|
||||||
{
|
|
||||||
final Configuration conf = new Configuration(datanode.getConf());
|
|
||||||
final int b = bufferSize.getValue(conf);
|
|
||||||
DFSClient dfsclient = newDfsClient(nnId, conf);
|
|
||||||
FSDataOutputStream out = null;
|
|
||||||
try {
|
|
||||||
out = dfsclient.append(fullpath, b, null, null);
|
|
||||||
IOUtils.copyBytes(in, out, b);
|
|
||||||
out.close();
|
|
||||||
out = null;
|
|
||||||
dfsclient.close();
|
|
||||||
dfsclient = null;
|
|
||||||
} finally {
|
|
||||||
IOUtils.cleanup(LOG, out);
|
|
||||||
IOUtils.cleanup(LOG, dfsclient);
|
|
||||||
}
|
|
||||||
return Response.ok().type(MediaType.APPLICATION_OCTET_STREAM).build();
|
|
||||||
}
|
|
||||||
default:
|
|
||||||
throw new UnsupportedOperationException(op + " is not supported");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/** Handle HTTP GET request for the root. */
|
|
||||||
@GET
|
|
||||||
@Path("/")
|
|
||||||
@Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
|
|
||||||
public Response getRoot(
|
|
||||||
@Context final UserGroupInformation ugi,
|
|
||||||
@QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
|
|
||||||
final DelegationParam delegation,
|
|
||||||
@QueryParam(NamenodeAddressParam.NAME)
|
|
||||||
@DefaultValue(NamenodeAddressParam.DEFAULT)
|
|
||||||
final NamenodeAddressParam namenode,
|
|
||||||
@QueryParam(GetOpParam.NAME) @DefaultValue(GetOpParam.DEFAULT)
|
|
||||||
final GetOpParam op,
|
|
||||||
@QueryParam(OffsetParam.NAME) @DefaultValue(OffsetParam.DEFAULT)
|
|
||||||
final OffsetParam offset,
|
|
||||||
@QueryParam(LengthParam.NAME) @DefaultValue(LengthParam.DEFAULT)
|
|
||||||
final LengthParam length,
|
|
||||||
@QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
|
|
||||||
final BufferSizeParam bufferSize
|
|
||||||
) throws IOException, InterruptedException {
|
|
||||||
return get(ugi, delegation, namenode, ROOT, op, offset, length,
|
|
||||||
bufferSize);
|
|
||||||
}
|
|
||||||
|
|
||||||
/** Handle HTTP GET request. */
|
|
||||||
@GET
|
|
||||||
@Path("{" + UriFsPathParam.NAME + ":.*}")
|
|
||||||
@Produces({MediaType.APPLICATION_OCTET_STREAM, MediaType.APPLICATION_JSON})
|
|
||||||
public Response get(
|
|
||||||
@Context final UserGroupInformation ugi,
|
|
||||||
@QueryParam(DelegationParam.NAME) @DefaultValue(DelegationParam.DEFAULT)
|
|
||||||
final DelegationParam delegation,
|
|
||||||
@QueryParam(NamenodeAddressParam.NAME)
|
|
||||||
@DefaultValue(NamenodeAddressParam.DEFAULT)
|
|
||||||
final NamenodeAddressParam namenode,
|
|
||||||
@PathParam(UriFsPathParam.NAME) final UriFsPathParam path,
|
|
||||||
@QueryParam(GetOpParam.NAME) @DefaultValue(GetOpParam.DEFAULT)
|
|
||||||
final GetOpParam op,
|
|
||||||
@QueryParam(OffsetParam.NAME) @DefaultValue(OffsetParam.DEFAULT)
|
|
||||||
final OffsetParam offset,
|
|
||||||
@QueryParam(LengthParam.NAME) @DefaultValue(LengthParam.DEFAULT)
|
|
||||||
final LengthParam length,
|
|
||||||
@QueryParam(BufferSizeParam.NAME) @DefaultValue(BufferSizeParam.DEFAULT)
|
|
||||||
final BufferSizeParam bufferSize
|
|
||||||
) throws IOException, InterruptedException {
|
|
||||||
|
|
||||||
final String nnId = namenode.getValue();
|
|
||||||
init(ugi, delegation, nnId, path, op, offset, length, bufferSize);
|
|
||||||
|
|
||||||
return ugi.doAs(new PrivilegedExceptionAction<Response>() {
|
|
||||||
@Override
|
|
||||||
public Response run() throws IOException {
|
|
||||||
return get(nnId, path.getAbsolutePath(), op, offset,
|
|
||||||
length, bufferSize);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
private Response get(
|
|
||||||
final String nnId,
|
|
||||||
final String fullpath,
|
|
||||||
final GetOpParam op,
|
|
||||||
final OffsetParam offset,
|
|
||||||
final LengthParam length,
|
|
||||||
final BufferSizeParam bufferSize
|
|
||||||
) throws IOException {
|
|
||||||
final DataNode datanode = (DataNode)context.getAttribute("datanode");
|
|
||||||
final Configuration conf = new Configuration(datanode.getConf());
|
|
||||||
|
|
||||||
switch(op.getValue()) {
|
|
||||||
case OPEN:
|
|
||||||
{
|
|
||||||
final int b = bufferSize.getValue(conf);
|
|
||||||
final DFSClient dfsclient = newDfsClient(nnId, conf);
|
|
||||||
HdfsDataInputStream in = null;
|
|
||||||
try {
|
|
||||||
in = dfsclient.createWrappedInputStream(
|
|
||||||
dfsclient.open(fullpath, b, true));
|
|
||||||
in.seek(offset.getValue());
|
|
||||||
} catch(IOException ioe) {
|
|
||||||
IOUtils.cleanup(LOG, in);
|
|
||||||
IOUtils.cleanup(LOG, dfsclient);
|
|
||||||
throw ioe;
|
|
||||||
}
|
|
||||||
|
|
||||||
final long n = length.getValue() != null ?
|
|
||||||
Math.min(length.getValue(), in.getVisibleLength() - offset.getValue()) :
|
|
||||||
in.getVisibleLength() - offset.getValue();
|
|
||||||
|
|
||||||
// jetty 6 reserves 12 bytes in the out buffer for chunked responses
|
|
||||||
// (file length > 2GB) which causes extremely poor performance when
|
|
||||||
// 12 bytes of the output spill into another buffer which results
|
|
||||||
// in a big and little write
|
|
||||||
int outBufferSize = response.getBufferSize();
|
|
||||||
if (n > Integer.MAX_VALUE) {
|
|
||||||
outBufferSize -= 12;
|
|
||||||
}
|
|
||||||
/**
|
|
||||||
* Allow the Web UI to perform an AJAX request to get the data.
|
|
||||||
*/
|
|
||||||
return Response.ok(new OpenEntity(in, n, outBufferSize, dfsclient))
|
|
||||||
.type(MediaType.APPLICATION_OCTET_STREAM)
|
|
||||||
.header("Access-Control-Allow-Methods", "GET")
|
|
||||||
.header("Access-Control-Allow-Origin", "*")
|
|
||||||
.build();
|
|
||||||
}
|
|
||||||
case GETFILECHECKSUM:
|
|
||||||
{
|
|
||||||
MD5MD5CRC32FileChecksum checksum = null;
|
|
||||||
DFSClient dfsclient = newDfsClient(nnId, conf);
|
|
||||||
try {
|
|
||||||
checksum = dfsclient.getFileChecksum(fullpath, Long.MAX_VALUE);
|
|
||||||
dfsclient.close();
|
|
||||||
dfsclient = null;
|
|
||||||
} finally {
|
|
||||||
IOUtils.cleanup(LOG, dfsclient);
|
|
||||||
}
|
|
||||||
final String js = JsonUtil.toJsonString(checksum);
|
|
||||||
return Response.ok(js).type(MediaType.APPLICATION_JSON).build();
|
|
||||||
}
|
|
||||||
default:
|
|
||||||
throw new UnsupportedOperationException(op + " is not supported");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
private static DFSClient newDfsClient(String nnId,
|
|
||||||
Configuration conf) throws IOException {
|
|
||||||
URI uri = URI.create(HdfsConstants.HDFS_URI_SCHEME + "://" + nnId);
|
|
||||||
return new DFSClient(uri, conf);
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -1,92 +0,0 @@
|
||||||
/**
|
|
||||||
* Licensed to the Apache Software Foundation (ASF) under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. The ASF licenses this file
|
|
||||||
* to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing, software
|
|
||||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
* See the License for the specific language governing permissions and
|
|
||||||
* limitations under the License.
|
|
||||||
*/
|
|
||||||
package org.apache.hadoop.hdfs.server.datanode.web.resources;
|
|
||||||
|
|
||||||
import org.apache.hadoop.hdfs.DFSClient;
|
|
||||||
import org.apache.hadoop.hdfs.client.HdfsDataInputStream;
|
|
||||||
import org.apache.hadoop.io.IOUtils;
|
|
||||||
|
|
||||||
import javax.ws.rs.core.MediaType;
|
|
||||||
import javax.ws.rs.core.MultivaluedMap;
|
|
||||||
import javax.ws.rs.ext.MessageBodyWriter;
|
|
||||||
import javax.ws.rs.ext.Provider;
|
|
||||||
import java.io.IOException;
|
|
||||||
import java.io.OutputStream;
|
|
||||||
import java.lang.annotation.Annotation;
|
|
||||||
import java.lang.reflect.Type;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A response entity for a HdfsDataInputStream.
|
|
||||||
*/
|
|
||||||
public class OpenEntity {
|
|
||||||
private final HdfsDataInputStream in;
|
|
||||||
private final long length;
|
|
||||||
private final int outBufferSize;
|
|
||||||
private final DFSClient dfsclient;
|
|
||||||
|
|
||||||
OpenEntity(final HdfsDataInputStream in, final long length,
|
|
||||||
final int outBufferSize, final DFSClient dfsclient) {
|
|
||||||
this.in = in;
|
|
||||||
this.length = length;
|
|
||||||
this.outBufferSize = outBufferSize;
|
|
||||||
this.dfsclient = dfsclient;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A {@link MessageBodyWriter} for {@link OpenEntity}.
|
|
||||||
*/
|
|
||||||
@Provider
|
|
||||||
public static class Writer implements MessageBodyWriter<OpenEntity> {
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public boolean isWriteable(Class<?> clazz, Type genericType,
|
|
||||||
Annotation[] annotations, MediaType mediaType) {
|
|
||||||
return clazz == OpenEntity.class
|
|
||||||
&& MediaType.APPLICATION_OCTET_STREAM_TYPE.isCompatible(mediaType);
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public long getSize(OpenEntity e, Class<?> type, Type genericType,
|
|
||||||
Annotation[] annotations, MediaType mediaType) {
|
|
||||||
return e.length;
|
|
||||||
}
|
|
||||||
|
|
||||||
@Override
|
|
||||||
public void writeTo(OpenEntity e, Class<?> type, Type genericType,
|
|
||||||
Annotation[] annotations, MediaType mediaType,
|
|
||||||
MultivaluedMap<String, Object> httpHeaders, OutputStream out
|
|
||||||
) throws IOException {
|
|
||||||
try {
|
|
||||||
byte[] buf = new byte[e.outBufferSize];
|
|
||||||
long remaining = e.length;
|
|
||||||
while (remaining > 0) {
|
|
||||||
int read = e.in.read(buf, 0, (int)Math.min(buf.length, remaining));
|
|
||||||
if (read == -1) { // EOF
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
out.write(buf, 0, read);
|
|
||||||
out.flush();
|
|
||||||
remaining -= read;
|
|
||||||
}
|
|
||||||
} finally {
|
|
||||||
IOUtils.cleanup(DatanodeWebHdfsMethods.LOG, e.in);
|
|
||||||
IOUtils.cleanup(DatanodeWebHdfsMethods.LOG, e.dfsclient);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -15,40 +15,41 @@
|
||||||
* See the License for the specific language governing permissions and
|
* See the License for the specific language governing permissions and
|
||||||
* limitations under the License.
|
* limitations under the License.
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hdfs.server.datanode.web.resources;
|
package org.apache.hadoop.hdfs.server.datanode.web.webhdfs;
|
||||||
|
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.hdfs.DFSTestUtil;
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
import org.apache.hadoop.hdfs.HAUtil;
|
import org.apache.hadoop.hdfs.HAUtil;
|
||||||
import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier;
|
import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.DataNode;
|
import org.apache.hadoop.hdfs.web.resources.DelegationParam;
|
||||||
|
import org.apache.hadoop.hdfs.web.resources.NamenodeAddressParam;
|
||||||
import org.apache.hadoop.security.token.Token;
|
import org.apache.hadoop.security.token.Token;
|
||||||
import org.junit.Assert;
|
import org.junit.Assert;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.mockito.internal.util.reflection.Whitebox;
|
|
||||||
|
import io.netty.handler.codec.http.QueryStringDecoder;
|
||||||
|
|
||||||
import javax.servlet.ServletContext;
|
import javax.servlet.ServletContext;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
|
|
||||||
import static org.mockito.Mockito.doReturn;
|
import static org.mockito.Mockito.doReturn;
|
||||||
import static org.mockito.Mockito.mock;
|
import static org.mockito.Mockito.mock;
|
||||||
|
|
||||||
public class TestDatanodeWebHdfsMethods {
|
public class TestParameterParser {
|
||||||
private static final String LOGICAL_NAME = "minidfs";
|
private static final String LOGICAL_NAME = "minidfs";
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testDeserializeHAToken() throws IOException {
|
public void testDeserializeHAToken() throws IOException {
|
||||||
Configuration conf = DFSTestUtil.newHAConfiguration(LOGICAL_NAME);
|
Configuration conf = DFSTestUtil.newHAConfiguration(LOGICAL_NAME);
|
||||||
DataNode dn = mock(DataNode.class);
|
|
||||||
doReturn(conf).when(dn).getConf();
|
|
||||||
ServletContext context = mock(ServletContext.class);
|
|
||||||
doReturn(dn).when(context).getAttribute("datanode");
|
|
||||||
final Token<DelegationTokenIdentifier> token = new
|
final Token<DelegationTokenIdentifier> token = new
|
||||||
Token<DelegationTokenIdentifier>();
|
Token<DelegationTokenIdentifier>();
|
||||||
DatanodeWebHdfsMethods method = new DatanodeWebHdfsMethods();
|
QueryStringDecoder decoder = new QueryStringDecoder(
|
||||||
Whitebox.setInternalState(method, "context", context);
|
WebHdfsHandler.WEBHDFS_PREFIX + "/?"
|
||||||
final Token<DelegationTokenIdentifier> tok2 = method.deserializeToken
|
+ NamenodeAddressParam.NAME + "=" + LOGICAL_NAME + "&"
|
||||||
(token.encodeToUrlString(), LOGICAL_NAME);
|
+ DelegationParam.NAME + "=" + token.encodeToUrlString());
|
||||||
|
ParameterParser testParser = new ParameterParser(decoder, conf);
|
||||||
|
final Token<DelegationTokenIdentifier> tok2 = testParser.delegationToken();
|
||||||
Assert.assertTrue(HAUtil.isTokenForLogicalUri(tok2));
|
Assert.assertTrue(HAUtil.isTokenForLogicalUri(tok2));
|
||||||
}
|
}
|
||||||
}
|
}
|
|
@ -31,7 +31,6 @@ import org.apache.hadoop.fs.permission.FsPermission;
|
||||||
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
import org.apache.hadoop.hdfs.DFSConfigKeys;
|
||||||
import org.apache.hadoop.hdfs.DFSTestUtil;
|
import org.apache.hadoop.hdfs.DFSTestUtil;
|
||||||
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
import org.apache.hadoop.hdfs.MiniDFSCluster;
|
||||||
import org.apache.hadoop.hdfs.server.datanode.web.resources.DatanodeWebHdfsMethods;
|
|
||||||
import org.apache.hadoop.hdfs.web.resources.ExceptionHandler;
|
import org.apache.hadoop.hdfs.web.resources.ExceptionHandler;
|
||||||
import org.apache.hadoop.security.AccessControlException;
|
import org.apache.hadoop.security.AccessControlException;
|
||||||
import org.apache.hadoop.security.UserGroupInformation;
|
import org.apache.hadoop.security.UserGroupInformation;
|
||||||
|
@ -44,7 +43,6 @@ import org.junit.Test;
|
||||||
public class TestFSMainOperationsWebHdfs extends FSMainOperationsBaseTest {
|
public class TestFSMainOperationsWebHdfs extends FSMainOperationsBaseTest {
|
||||||
{
|
{
|
||||||
((Log4JLogger)ExceptionHandler.LOG).getLogger().setLevel(Level.ALL);
|
((Log4JLogger)ExceptionHandler.LOG).getLogger().setLevel(Level.ALL);
|
||||||
((Log4JLogger)DatanodeWebHdfsMethods.LOG).getLogger().setLevel(Level.ALL);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private static MiniDFSCluster cluster = null;
|
private static MiniDFSCluster cluster = null;
|
||||||
|
|
Loading…
Reference in New Issue