HDFS-15902. Improve the log for HTTPFS server operation. Contributed by Bhavik Patel.
(cherry picked from commit b4d97a8dc7
)
This commit is contained in:
parent
2d83369f6f
commit
9e1e89b1b9
|
@ -295,10 +295,8 @@ public class DelegationTokenAuthenticatedURL extends AuthenticatedURL {
|
|||
// delegation token
|
||||
Credentials creds = UserGroupInformation.getCurrentUser().
|
||||
getCredentials();
|
||||
if (LOG.isDebugEnabled()) {
|
||||
LOG.debug("Token not set, looking for delegation token. Creds:{},"
|
||||
+ " size:{}", creds.getAllTokens(), creds.numberOfTokens());
|
||||
}
|
||||
LOG.debug("Token not set, looking for delegation token. Creds:{},"
|
||||
+ " size:{}", creds.getAllTokens(), creds.numberOfTokens());
|
||||
if (!creds.getAllTokens().isEmpty()) {
|
||||
dToken = selectDelegationToken(url, creds);
|
||||
if (dToken != null) {
|
||||
|
|
|
@ -284,7 +284,7 @@ public class HttpFSServer {
|
|||
}
|
||||
});
|
||||
} catch (InterruptedException ie) {
|
||||
LOG.info("Open interrupted.", ie);
|
||||
LOG.warn("Open interrupted.", ie);
|
||||
Thread.currentThread().interrupt();
|
||||
}
|
||||
Long offset = params.get(OffsetParam.NAME, OffsetParam.class);
|
||||
|
@ -317,7 +317,7 @@ public class HttpFSServer {
|
|||
enforceRootPath(op.value(), path);
|
||||
FSOperations.FSHomeDir command = new FSOperations.FSHomeDir();
|
||||
JSONObject json = fsExecute(user, command);
|
||||
AUDIT_LOG.info("");
|
||||
AUDIT_LOG.info("Home Directory for [{}]", user);
|
||||
response = Response.ok(json).type(MediaType.APPLICATION_JSON).build();
|
||||
break;
|
||||
}
|
||||
|
@ -339,7 +339,7 @@ public class HttpFSServer {
|
|||
FSOperations.FSContentSummary command =
|
||||
new FSOperations.FSContentSummary(path);
|
||||
Map json = fsExecute(user, command);
|
||||
AUDIT_LOG.info("[{}]", path);
|
||||
AUDIT_LOG.info("Content summary for [{}]", path);
|
||||
response = Response.ok(json).type(MediaType.APPLICATION_JSON).build();
|
||||
break;
|
||||
}
|
||||
|
@ -347,7 +347,7 @@ public class HttpFSServer {
|
|||
FSOperations.FSQuotaUsage command =
|
||||
new FSOperations.FSQuotaUsage(path);
|
||||
Map json = fsExecute(user, command);
|
||||
AUDIT_LOG.info("[{}]", path);
|
||||
AUDIT_LOG.info("Quota Usage for [{}]", path);
|
||||
response = Response.ok(json).type(MediaType.APPLICATION_JSON).build();
|
||||
break;
|
||||
}
|
||||
|
@ -650,14 +650,11 @@ public class HttpFSServer {
|
|||
break;
|
||||
}
|
||||
case CONCAT: {
|
||||
System.out.println("HTTPFS SERVER CONCAT");
|
||||
String sources = params.get(SourcesParam.NAME, SourcesParam.class);
|
||||
|
||||
FSOperations.FSConcat command =
|
||||
new FSOperations.FSConcat(path, sources.split(","));
|
||||
fsExecute(user, command);
|
||||
AUDIT_LOG.info("[{}]", path);
|
||||
System.out.println("SENT RESPONSE");
|
||||
response = Response.ok().build();
|
||||
break;
|
||||
}
|
||||
|
|
|
@ -194,9 +194,11 @@ public class FileSystemAccessService extends BaseService implements FileSystemAc
|
|||
throw new ServiceException(FileSystemAccessException.ERROR.H11, ex.toString(), ex);
|
||||
}
|
||||
|
||||
LOG.debug("FileSystemAccess FileSystem configuration:");
|
||||
for (Map.Entry entry : serviceHadoopConf) {
|
||||
LOG.debug(" {} = {}", entry.getKey(), entry.getValue());
|
||||
if (LOG.isDebugEnabled()) {
|
||||
LOG.debug("FileSystemAccess FileSystem configuration:");
|
||||
for (Map.Entry entry : serviceHadoopConf) {
|
||||
LOG.debug(" {} = {}", entry.getKey(), entry.getValue());
|
||||
}
|
||||
}
|
||||
setRequiredServiceHadoopConf(serviceHadoopConf);
|
||||
|
||||
|
@ -262,7 +264,7 @@ public class FileSystemAccessService extends BaseService implements FileSystemAc
|
|||
LOG.warn("Error while purging filesystem, " + ex.toString(), ex);
|
||||
}
|
||||
}
|
||||
LOG.debug("Purged [{}} filesystem instances", count);
|
||||
LOG.debug("Purged [{}] filesystem instances", count);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue