From 988b25419ce58baa4a7391d4d076f0ee8b91f10e Mon Sep 17 00:00:00 2001
From: Anu Engineer <aengineer@apache.org>
Date: Thu, 1 Jun 2017 13:07:20 -0700
Subject: [PATCH] HDFS-11894. Ozone: Cleanup imports. Contributed by Weiwei
 Yang.

---
 .../hadoop/scm/storage/ChunkInputStream.java  |  6 ++----
 .../apache/hadoop/ozone/OzoneClientUtils.java |  4 +---
 .../common/impl/ContainerStorageLocation.java |  8 ++++++--
 .../web/exceptions/OzoneExceptionMapper.java  |  4 ++--
 .../web/handlers/BucketProcessTemplate.java   |  5 -----
 .../web/netty/ObjectStoreJerseyContainer.java |  8 ++++++--
 .../hadoop/ozone/web/request/OzoneQuota.java  |  3 +--
 .../ozone/web/response/VolumeOwner.java       |  5 ++---
 .../storage/DistributedStorageHandler.java    | 19 +++++++++++--------
 9 files changed, 31 insertions(+), 31 deletions(-)

diff --git a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/scm/storage/ChunkInputStream.java b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/scm/storage/ChunkInputStream.java
index 8ca008ca443..c690474991d 100644
--- a/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/scm/storage/ChunkInputStream.java
+++ b/hadoop-hdfs-project/hadoop-hdfs-client/src/main/java/org/apache/hadoop/scm/storage/ChunkInputStream.java
@@ -18,8 +18,6 @@
 
 package org.apache.hadoop.scm.storage;
 
-import static org.apache.hadoop.scm.storage.ContainerProtocolCalls.*;
-
 import java.io.IOException;
 import java.io.InputStream;
 import java.nio.ByteBuffer;
@@ -179,8 +177,8 @@ private synchronized void readChunkFromContainer(int readChunkOffset)
       throws IOException {
     final ReadChunkResponseProto readChunkResponse;
     try {
-      readChunkResponse = readChunk(xceiverClient, chunks.get(readChunkOffset),
-          key, traceID);
+      readChunkResponse = ContainerProtocolCalls.readChunk(xceiverClient,
+          chunks.get(readChunkOffset), key, traceID);
     } catch (IOException e) {
       throw new IOException("Unexpected OzoneException", e);
     }
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/OzoneClientUtils.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/OzoneClientUtils.java
index 8ba8dc5c5f1..db3c14f98e1 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/OzoneClientUtils.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/OzoneClientUtils.java
@@ -20,7 +20,6 @@
 import com.google.common.base.Optional;
 
 import com.google.common.net.HostAndPort;
-import org.apache.avro.reflect.Nullable;
 import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.classification.InterfaceStability;
 import org.apache.hadoop.conf.Configuration;
@@ -585,8 +584,7 @@ public static CloseableHttpClient newHttpClient() {
    * @param conf configuration
    * @return a {@link CloseableHttpClient} instance.
    */
-  public static CloseableHttpClient newHttpClient(
-      @Nullable Configuration conf) {
+  public static CloseableHttpClient newHttpClient(Configuration conf) {
     int socketTimeout = OzoneConfigKeys
         .OZONE_CLIENT_SOCKET_TIMEOUT_MS_DEFAULT;
     int connectionTimeout = OzoneConfigKeys
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/impl/ContainerStorageLocation.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/impl/ContainerStorageLocation.java
index 14dcc3d229a..17f027bdd9f 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/impl/ContainerStorageLocation.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/container/common/impl/ContainerStorageLocation.java
@@ -30,7 +30,11 @@
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import java.io.*;
+import java.io.File;
+import java.io.IOException;
+import java.io.FileNotFoundException;
+import java.io.OutputStreamWriter;
+import java.io.FileOutputStream;
 import java.net.URI;
 import java.nio.charset.StandardCharsets;
 import java.util.Scanner;
@@ -184,4 +188,4 @@ void saveScmUsed() {
       IOUtils.cleanup(null, out);
     }
   }
-}
\ No newline at end of file
+}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/exceptions/OzoneExceptionMapper.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/exceptions/OzoneExceptionMapper.java
index 5b27210e593..28a83772503 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/exceptions/OzoneExceptionMapper.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/exceptions/OzoneExceptionMapper.java
@@ -22,7 +22,7 @@
 import javax.ws.rs.core.Response;
 import javax.ws.rs.ext.ExceptionMapper;
 
-import org.apache.log4j.MDC;
+import org.slf4j.MDC;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
@@ -42,4 +42,4 @@ public Response toResponse(OzoneException exception) {
       .entity(exception.toJsonString()).build();
   }
 
-}
\ No newline at end of file
+}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/handlers/BucketProcessTemplate.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/handlers/BucketProcessTemplate.java
index 9c4e1df1754..d89dfa76bd5 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/handlers/BucketProcessTemplate.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/handlers/BucketProcessTemplate.java
@@ -18,7 +18,6 @@
 
 package org.apache.hadoop.ozone.web.handlers;
 
-
 import org.apache.hadoop.fs.StorageType;
 import org.apache.hadoop.ozone.web.exceptions.ErrorTable;
 import org.apache.hadoop.ozone.web.exceptions.OzoneException;
@@ -34,7 +33,6 @@
 import org.slf4j.LoggerFactory;
 import org.slf4j.MDC;
 
-
 import javax.ws.rs.core.HttpHeaders;
 import javax.ws.rs.core.Request;
 import javax.ws.rs.core.Response;
@@ -52,9 +50,6 @@
 import static org.apache.hadoop.ozone.OzoneConsts.OZONE_REQUEST;
 import static org.apache.hadoop.ozone.OzoneConsts.OZONE_USER;
 
-
-
-
 /**
  * This class abstracts way the repetitive tasks in
  * Bucket handling code.
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/netty/ObjectStoreJerseyContainer.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/netty/ObjectStoreJerseyContainer.java
index 95ecbe890ee..5b18e2eafbb 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/netty/ObjectStoreJerseyContainer.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/netty/ObjectStoreJerseyContainer.java
@@ -17,8 +17,12 @@
  */
 package org.apache.hadoop.ozone.web.netty;
 
-import static io.netty.handler.codec.http.HttpHeaders.Names.*;
-import static io.netty.handler.codec.http.HttpHeaders.Values.*;
+import static io.netty.handler.codec.http.HttpHeaders.Names.CONTENT_LENGTH;
+import static io.netty.handler.codec.http.HttpHeaders.Names.CONNECTION;
+import static io.netty.handler.codec.http.HttpHeaders.Names.TRANSFER_ENCODING;
+import static io.netty.handler.codec.http.HttpHeaders.Names.HOST;
+import static io.netty.handler.codec.http.HttpHeaders.Values.KEEP_ALIVE;
+import static io.netty.handler.codec.http.HttpHeaders.Values.CLOSE;
 import static io.netty.handler.codec.http.HttpVersion.HTTP_1_1;
 
 import java.io.IOException;
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/request/OzoneQuota.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/request/OzoneQuota.java
index c73bd6d85d5..f9ed45fe05e 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/request/OzoneQuota.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/request/OzoneQuota.java
@@ -18,11 +18,10 @@
 
 package org.apache.hadoop.ozone.web.request;
 
-
 import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.ozone.OzoneConsts;
 import org.apache.hadoop.ozone.web.headers.Header;
-import org.codehaus.jackson.annotate.JsonIgnore;
+import com.fasterxml.jackson.annotation.JsonIgnore;
 
 /**
  * represents an OzoneQuota Object that can be applied to
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/response/VolumeOwner.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/response/VolumeOwner.java
index 1dadf79079f..afb0460538c 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/response/VolumeOwner.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/response/VolumeOwner.java
@@ -18,9 +18,8 @@
 
 package org.apache.hadoop.ozone.web.response;
 
-
+import com.fasterxml.jackson.annotation.JsonInclude;
 import org.apache.hadoop.classification.InterfaceAudience;
-import org.codehaus.jackson.map.annotate.JsonSerialize;
 
 /**
  * Volume Owner represents the owner of a volume.
@@ -30,7 +29,7 @@
  */
 @InterfaceAudience.Private
 public class VolumeOwner {
-  @JsonSerialize(include=JsonSerialize.Inclusion.NON_NULL)
+  @JsonInclude(JsonInclude.Include.NON_NULL)
   private String name;
 
   /**
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/storage/DistributedStorageHandler.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/storage/DistributedStorageHandler.java
index 14d82030bce..ab2363e2ca9 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/storage/DistributedStorageHandler.java
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/ozone/web/storage/DistributedStorageHandler.java
@@ -53,7 +53,12 @@
 import org.apache.hadoop.ozone.web.handlers.ListArgs;
 import org.apache.hadoop.ozone.web.handlers.VolumeArgs;
 import org.apache.hadoop.ozone.web.interfaces.StorageHandler;
-import org.apache.hadoop.ozone.web.response.*;
+import org.apache.hadoop.ozone.web.response.ListVolumes;
+import org.apache.hadoop.ozone.web.response.VolumeInfo;
+import org.apache.hadoop.ozone.web.response.VolumeOwner;
+import org.apache.hadoop.ozone.web.response.ListBuckets;
+import org.apache.hadoop.ozone.web.response.BucketInfo;
+import org.apache.hadoop.ozone.web.response.ListKeys;
 import org.apache.hadoop.scm.XceiverClientSpi;
 import org.apache.hadoop.scm.storage.ChunkInputStream;
 import org.apache.hadoop.scm.storage.ChunkOutputStream;
@@ -74,9 +79,6 @@
 import java.util.List;
 import java.util.stream.Collectors;
 
-import static org.apache.hadoop.ozone.web.storage.OzoneContainerTranslation.*;
-import static org.apache.hadoop.scm.storage.ContainerProtocolCalls.getKey;
-
 /**
  * A {@link StorageHandler} implementation that distributes object storage
  * across the nodes of an HDFS cluster.
@@ -356,10 +358,11 @@ public LengthInputStream newKeyReader(KeyArgs args) throws IOException,
     try {
       LOG.debug("get key accessing {} {}",
           xceiverClient.getPipeline().getContainerName(), containerKey);
-      KeyData containerKeyData = containerKeyDataForRead(
-          xceiverClient.getPipeline().getContainerName(), containerKey);
-      GetKeyResponseProto response = getKey(xceiverClient, containerKeyData,
-          args.getRequestID());
+      KeyData containerKeyData = OzoneContainerTranslation
+          .containerKeyDataForRead(
+              xceiverClient.getPipeline().getContainerName(), containerKey);
+      GetKeyResponseProto response = ContainerProtocolCalls
+          .getKey(xceiverClient, containerKeyData, args.getRequestID());
       long length = 0;
       List<ChunkInfo> chunks = response.getKeyData().getChunksList();
       for (ChunkInfo chunk : chunks) {