HBASE-17008 Examples to make AsyncClient go down easy
This commit is contained in:
parent
f5768b4306
commit
cb88b64629
|
@ -0,0 +1,184 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hbase.client.example;
|
||||
|
||||
import java.util.concurrent.CompletableFuture;
|
||||
import java.util.concurrent.CountDownLatch;
|
||||
import java.util.concurrent.ExecutorService;
|
||||
import java.util.concurrent.Executors;
|
||||
import java.util.concurrent.atomic.AtomicReference;
|
||||
import java.util.stream.IntStream;
|
||||
|
||||
import org.apache.commons.io.IOUtils;
|
||||
import org.apache.commons.logging.Log;
|
||||
import org.apache.commons.logging.LogFactory;
|
||||
import org.apache.hadoop.conf.Configured;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.AsyncConnection;
|
||||
import org.apache.hadoop.hbase.client.AsyncTable;
|
||||
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||
import org.apache.hadoop.hbase.client.Get;
|
||||
import org.apache.hadoop.hbase.client.Put;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.hadoop.hbase.util.Threads;
|
||||
import org.apache.hadoop.util.Tool;
|
||||
import org.apache.hadoop.util.ToolRunner;
|
||||
|
||||
/**
|
||||
* A simple example shows how to use asynchronous client.
|
||||
*/
|
||||
public class AsyncClientExample extends Configured implements Tool {
|
||||
|
||||
private static final Log LOG = LogFactory.getLog(AsyncClientExample.class);
|
||||
|
||||
/**
|
||||
* The size for thread pool.
|
||||
*/
|
||||
private static final int THREAD_POOL_SIZE = 16;
|
||||
|
||||
/**
|
||||
* The default number of operations.
|
||||
*/
|
||||
private static final int DEFAULT_NUM_OPS = 100;
|
||||
|
||||
/**
|
||||
* The name of the column family. d for default.
|
||||
*/
|
||||
private static final byte[] FAMILY = Bytes.toBytes("d");
|
||||
|
||||
/**
|
||||
* For the example we're just using one qualifier.
|
||||
*/
|
||||
private static final byte[] QUAL = Bytes.toBytes("test");
|
||||
|
||||
private final AtomicReference<CompletableFuture<AsyncConnection>> future =
|
||||
new AtomicReference<>();
|
||||
|
||||
private CompletableFuture<AsyncConnection> getConn() {
|
||||
CompletableFuture<AsyncConnection> f = future.get();
|
||||
if (f != null) {
|
||||
return f;
|
||||
}
|
||||
for (;;) {
|
||||
if (future.compareAndSet(null, new CompletableFuture<>())) {
|
||||
CompletableFuture<AsyncConnection> toComplete = future.get();
|
||||
ConnectionFactory.createAsyncConnection(getConf()).whenComplete((conn, error) -> {
|
||||
if (error != null) {
|
||||
toComplete.completeExceptionally(error);
|
||||
// we need to reset the future holder so we will get a chance to recreate an async
|
||||
// connection at next try.
|
||||
future.set(null);
|
||||
return;
|
||||
}
|
||||
toComplete.complete(conn);
|
||||
});
|
||||
return toComplete;
|
||||
} else {
|
||||
f = future.get();
|
||||
if (f != null) {
|
||||
return f;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private CompletableFuture<Void> closeConn() {
|
||||
CompletableFuture<AsyncConnection> f = future.get();
|
||||
if (f == null) {
|
||||
return CompletableFuture.completedFuture(null);
|
||||
}
|
||||
CompletableFuture<Void> closeFuture = new CompletableFuture<>();
|
||||
f.whenComplete((conn, error) -> {
|
||||
if (error == null) {
|
||||
IOUtils.closeQuietly(conn);
|
||||
}
|
||||
closeFuture.complete(null);
|
||||
});
|
||||
return closeFuture;
|
||||
}
|
||||
|
||||
private byte[] getKey(int i) {
|
||||
return Bytes.toBytes(String.format("%08x", i));
|
||||
}
|
||||
|
||||
@Override
|
||||
public int run(String[] args) throws Exception {
|
||||
if (args.length < 1 || args.length > 2) {
|
||||
System.out.println("Usage: " + this.getClass().getName() + " tableName [num_operations]");
|
||||
return -1;
|
||||
}
|
||||
TableName tableName = TableName.valueOf(args[0]);
|
||||
int numOps = args.length > 1 ? Integer.parseInt(args[1]) : DEFAULT_NUM_OPS;
|
||||
ExecutorService threadPool = Executors.newFixedThreadPool(THREAD_POOL_SIZE,
|
||||
Threads.newDaemonThreadFactory("AsyncClientExample"));
|
||||
// We use AsyncTable here so we need to provide a separated thread pool. RawAsyncTable does not
|
||||
// need a thread pool and may have a better performance if you use it correctly as it can save
|
||||
// some context switches. But if you use RawAsyncTable incorrectly, you may have a very bad
|
||||
// impact on performance so use it with caution.
|
||||
CountDownLatch latch = new CountDownLatch(numOps);
|
||||
IntStream.range(0, numOps).forEach(i -> {
|
||||
CompletableFuture<AsyncConnection> future = getConn();
|
||||
future.whenComplete((conn, error) -> {
|
||||
if (error != null) {
|
||||
LOG.warn("failed to get async connection for " + i, error);
|
||||
latch.countDown();
|
||||
return;
|
||||
}
|
||||
AsyncTable table = conn.getTable(tableName, threadPool);
|
||||
table.put(new Put(getKey(i)).addColumn(FAMILY, QUAL, Bytes.toBytes(i)))
|
||||
.whenComplete((putResp, putErr) -> {
|
||||
if (putErr != null) {
|
||||
LOG.warn("put failed for " + i, putErr);
|
||||
latch.countDown();
|
||||
return;
|
||||
}
|
||||
LOG.info("put for " + i + " succeeded, try getting");
|
||||
table.get(new Get(getKey(i))).whenComplete((result, getErr) -> {
|
||||
if (getErr != null) {
|
||||
LOG.warn("get failed for " + i, putErr);
|
||||
latch.countDown();
|
||||
return;
|
||||
}
|
||||
if (result.isEmpty()) {
|
||||
LOG.warn("get failed for " + i + ", server returns empty result");
|
||||
} else if (!result.containsColumn(FAMILY, QUAL)) {
|
||||
LOG.warn("get failed for " + i + ", the result does not contain " +
|
||||
Bytes.toString(FAMILY) + ":" + Bytes.toString(QUAL));
|
||||
} else {
|
||||
int v = Bytes.toInt(result.getValue(FAMILY, QUAL));
|
||||
if (v != i) {
|
||||
LOG.warn("get failed for " + i + ", the value of " + Bytes.toString(FAMILY) +
|
||||
":" + Bytes.toString(QUAL) + " is " + v + ", exected " + i);
|
||||
} else {
|
||||
LOG.info("get for " + i + " succeeded");
|
||||
}
|
||||
}
|
||||
latch.countDown();
|
||||
});
|
||||
});
|
||||
});
|
||||
});
|
||||
latch.await();
|
||||
closeConn().get();
|
||||
return 0;
|
||||
}
|
||||
|
||||
public static void main(String[] args) throws Exception {
|
||||
ToolRunner.run(new AsyncClientExample(), args);
|
||||
}
|
||||
}
|
|
@ -0,0 +1,264 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hbase.client.example;
|
||||
|
||||
import com.google.common.base.Preconditions;
|
||||
import com.google.common.base.Throwables;
|
||||
|
||||
import io.netty.bootstrap.ServerBootstrap;
|
||||
import io.netty.buffer.ByteBuf;
|
||||
import io.netty.channel.Channel;
|
||||
import io.netty.channel.ChannelHandlerContext;
|
||||
import io.netty.channel.ChannelInitializer;
|
||||
import io.netty.channel.ChannelOption;
|
||||
import io.netty.channel.EventLoopGroup;
|
||||
import io.netty.channel.SimpleChannelInboundHandler;
|
||||
import io.netty.channel.group.ChannelGroup;
|
||||
import io.netty.channel.group.DefaultChannelGroup;
|
||||
import io.netty.channel.nio.NioEventLoopGroup;
|
||||
import io.netty.channel.socket.nio.NioServerSocketChannel;
|
||||
import io.netty.channel.socket.nio.NioSocketChannel;
|
||||
import io.netty.handler.codec.http.DefaultFullHttpResponse;
|
||||
import io.netty.handler.codec.http.FullHttpRequest;
|
||||
import io.netty.handler.codec.http.HttpHeaderNames;
|
||||
import io.netty.handler.codec.http.HttpObjectAggregator;
|
||||
import io.netty.handler.codec.http.HttpResponseStatus;
|
||||
import io.netty.handler.codec.http.HttpServerCodec;
|
||||
import io.netty.handler.codec.http.HttpVersion;
|
||||
import io.netty.handler.codec.http.QueryStringDecoder;
|
||||
import io.netty.util.concurrent.GlobalEventExecutor;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.net.InetSocketAddress;
|
||||
import java.nio.charset.StandardCharsets;
|
||||
import java.util.Optional;
|
||||
import java.util.concurrent.ExecutionException;
|
||||
|
||||
import org.apache.hadoop.conf.Configuration;
|
||||
import org.apache.hadoop.hbase.HBaseConfiguration;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.client.AsyncConnection;
|
||||
import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||
import org.apache.hadoop.hbase.client.Get;
|
||||
import org.apache.hadoop.hbase.client.Put;
|
||||
import org.apache.hadoop.hbase.client.RawAsyncTable;
|
||||
import org.apache.hadoop.hbase.ipc.NettyRpcClientConfigHelper;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
|
||||
/**
|
||||
* A simple example on how to use {@link RawAsyncTable} to write a fully asynchronous HTTP proxy
|
||||
* server. The {@link AsyncConnection} will share the same event loop with the HTTP server.
|
||||
* <p>
|
||||
* The request URL is:
|
||||
*
|
||||
* <pre>
|
||||
* http://<host>:<port>/<table>/<rowgt;/<family>:<qualifier>
|
||||
* </pre>
|
||||
*
|
||||
* Use HTTP GET to fetch data, and use HTTP PUT to put data. Encode the value as the request content
|
||||
* when doing PUT.
|
||||
*/
|
||||
public class HttpProxyExample {
|
||||
|
||||
private final EventLoopGroup bossGroup = new NioEventLoopGroup(1);
|
||||
|
||||
private final EventLoopGroup workerGroup = new NioEventLoopGroup();
|
||||
|
||||
private final Configuration conf;
|
||||
|
||||
private final int port;
|
||||
|
||||
private AsyncConnection conn;
|
||||
|
||||
private Channel serverChannel;
|
||||
|
||||
private ChannelGroup channelGroup;
|
||||
|
||||
public HttpProxyExample(Configuration conf, int port) {
|
||||
this.conf = conf;
|
||||
this.port = port;
|
||||
}
|
||||
|
||||
private static final class Params {
|
||||
public final String table;
|
||||
|
||||
public final String row;
|
||||
|
||||
public final String family;
|
||||
|
||||
public final String qualifier;
|
||||
|
||||
public Params(String table, String row, String family, String qualifier) {
|
||||
this.table = table;
|
||||
this.row = row;
|
||||
this.family = family;
|
||||
this.qualifier = qualifier;
|
||||
}
|
||||
}
|
||||
|
||||
private static final class RequestHandler extends SimpleChannelInboundHandler<FullHttpRequest> {
|
||||
|
||||
private final AsyncConnection conn;
|
||||
|
||||
private final ChannelGroup channelGroup;
|
||||
|
||||
public RequestHandler(AsyncConnection conn, ChannelGroup channelGroup) {
|
||||
this.conn = conn;
|
||||
this.channelGroup = channelGroup;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void channelActive(ChannelHandlerContext ctx) {
|
||||
channelGroup.add(ctx.channel());
|
||||
ctx.fireChannelActive();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void channelInactive(ChannelHandlerContext ctx) {
|
||||
channelGroup.remove(ctx.channel());
|
||||
ctx.fireChannelInactive();
|
||||
}
|
||||
|
||||
private void write(ChannelHandlerContext ctx, HttpResponseStatus status,
|
||||
Optional<String> content) {
|
||||
DefaultFullHttpResponse resp;
|
||||
if (content.isPresent()) {
|
||||
ByteBuf buf =
|
||||
ctx.alloc().buffer().writeBytes(content.get().getBytes(StandardCharsets.UTF_8));
|
||||
resp = new DefaultFullHttpResponse(HttpVersion.HTTP_1_1, status, buf);
|
||||
resp.headers().set(HttpHeaderNames.CONTENT_LENGTH, buf.readableBytes());
|
||||
} else {
|
||||
resp = new DefaultFullHttpResponse(HttpVersion.HTTP_1_1, status);
|
||||
}
|
||||
resp.headers().set(HttpHeaderNames.CONTENT_TYPE, "text-plain; charset=UTF-8");
|
||||
ctx.writeAndFlush(resp);
|
||||
}
|
||||
|
||||
private Params parse(FullHttpRequest req) {
|
||||
String[] components = new QueryStringDecoder(req.uri()).path().split("/");
|
||||
Preconditions.checkArgument(components.length == 4, "Unrecognized uri: %s", req.uri());
|
||||
// path is start with '/' so split will give an empty component
|
||||
String[] cfAndCq = components[3].split(":");
|
||||
Preconditions.checkArgument(cfAndCq.length == 2, "Unrecognized uri: %s", req.uri());
|
||||
return new Params(components[1], components[2], cfAndCq[0], cfAndCq[1]);
|
||||
}
|
||||
|
||||
private void get(ChannelHandlerContext ctx, FullHttpRequest req) {
|
||||
Params params = parse(req);
|
||||
conn.getRawTable(TableName.valueOf(params.table)).get(new Get(Bytes.toBytes(params.row))
|
||||
.addColumn(Bytes.toBytes(params.family), Bytes.toBytes(params.qualifier)))
|
||||
.whenComplete((r, e) -> {
|
||||
if (e != null) {
|
||||
exceptionCaught(ctx, e);
|
||||
} else {
|
||||
byte[] value =
|
||||
r.getValue(Bytes.toBytes(params.family), Bytes.toBytes(params.qualifier));
|
||||
if (value != null) {
|
||||
write(ctx, HttpResponseStatus.OK, Optional.of(Bytes.toStringBinary(value)));
|
||||
} else {
|
||||
write(ctx, HttpResponseStatus.NOT_FOUND, Optional.empty());
|
||||
}
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
private void put(ChannelHandlerContext ctx, FullHttpRequest req) {
|
||||
Params params = parse(req);
|
||||
byte[] value = new byte[req.content().readableBytes()];
|
||||
req.content().readBytes(value);
|
||||
conn.getRawTable(TableName.valueOf(params.table)).put(new Put(Bytes.toBytes(params.row))
|
||||
.addColumn(Bytes.toBytes(params.family), Bytes.toBytes(params.qualifier), value))
|
||||
.whenComplete((r, e) -> {
|
||||
if (e != null) {
|
||||
exceptionCaught(ctx, e);
|
||||
} else {
|
||||
write(ctx, HttpResponseStatus.OK, Optional.empty());
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void channelRead0(ChannelHandlerContext ctx, FullHttpRequest req) {
|
||||
switch (req.method().name()) {
|
||||
case "GET":
|
||||
get(ctx, req);
|
||||
break;
|
||||
case "PUT":
|
||||
put(ctx, req);
|
||||
break;
|
||||
default:
|
||||
write(ctx, HttpResponseStatus.METHOD_NOT_ALLOWED, Optional.empty());
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void exceptionCaught(ChannelHandlerContext ctx, Throwable cause) {
|
||||
if (cause instanceof IllegalArgumentException) {
|
||||
write(ctx, HttpResponseStatus.BAD_REQUEST, Optional.of(cause.getMessage()));
|
||||
} else {
|
||||
write(ctx, HttpResponseStatus.INTERNAL_SERVER_ERROR,
|
||||
Optional.of(Throwables.getStackTraceAsString(cause)));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
public void start() throws InterruptedException, ExecutionException {
|
||||
NettyRpcClientConfigHelper.setEventLoopConfig(conf, workerGroup, NioSocketChannel.class);
|
||||
conn = ConnectionFactory.createAsyncConnection(conf).get();
|
||||
channelGroup = new DefaultChannelGroup(GlobalEventExecutor.INSTANCE);
|
||||
serverChannel = new ServerBootstrap().group(bossGroup, workerGroup)
|
||||
.channel(NioServerSocketChannel.class).childOption(ChannelOption.TCP_NODELAY, true)
|
||||
.childHandler(new ChannelInitializer<Channel>() {
|
||||
|
||||
@Override
|
||||
protected void initChannel(Channel ch) throws Exception {
|
||||
ch.pipeline().addFirst(new HttpServerCodec(), new HttpObjectAggregator(4 * 1024 * 1024),
|
||||
new RequestHandler(conn, channelGroup));
|
||||
}
|
||||
}).bind(port).syncUninterruptibly().channel();
|
||||
}
|
||||
|
||||
public void join() {
|
||||
serverChannel.closeFuture().awaitUninterruptibly();
|
||||
}
|
||||
|
||||
public int port() {
|
||||
if (serverChannel == null) {
|
||||
return port;
|
||||
} else {
|
||||
return ((InetSocketAddress) serverChannel.localAddress()).getPort();
|
||||
}
|
||||
}
|
||||
|
||||
public void stop() throws IOException {
|
||||
serverChannel.close().syncUninterruptibly();
|
||||
serverChannel = null;
|
||||
channelGroup.close().syncUninterruptibly();
|
||||
channelGroup = null;
|
||||
conn.close();
|
||||
conn = null;
|
||||
}
|
||||
|
||||
public static void main(String[] args) throws InterruptedException, ExecutionException {
|
||||
int port = Integer.parseInt(args[0]);
|
||||
HttpProxyExample proxy = new HttpProxyExample(HBaseConfiguration.create(), port);
|
||||
proxy.start();
|
||||
proxy.join();
|
||||
}
|
||||
}
|
|
@ -0,0 +1,57 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hbase.client.example;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
|
||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.hadoop.util.ToolRunner;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.experimental.categories.Category;
|
||||
|
||||
@Category({ ClientTests.class, MediumTests.class })
|
||||
public class TestAsyncClientExample {
|
||||
|
||||
private static final HBaseTestingUtility UTIL = new HBaseTestingUtility();
|
||||
|
||||
private static final TableName TABLE_NAME = TableName.valueOf("test");
|
||||
|
||||
@BeforeClass
|
||||
public static void setUp() throws Exception {
|
||||
UTIL.startMiniCluster(1);
|
||||
UTIL.createTable(TABLE_NAME, Bytes.toBytes("d"));
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
public static void tearDown() throws Exception {
|
||||
UTIL.shutdownMiniCluster();
|
||||
}
|
||||
|
||||
@Test
|
||||
public void test() throws Exception {
|
||||
AsyncClientExample tool = new AsyncClientExample();
|
||||
tool.setConf(UTIL.getConfiguration());
|
||||
assertEquals(0, ToolRunner.run(tool, new String[] { TABLE_NAME.getNameAsString() }));
|
||||
}
|
||||
}
|
|
@ -0,0 +1,106 @@
|
|||
/**
|
||||
* Licensed to the Apache Software Foundation (ASF) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. The ASF licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.hadoop.hbase.client.example;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
|
||||
import com.google.common.io.ByteStreams;
|
||||
|
||||
import java.nio.charset.StandardCharsets;
|
||||
|
||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||
import org.apache.hadoop.hbase.TableName;
|
||||
import org.apache.hadoop.hbase.testclassification.ClientTests;
|
||||
import org.apache.hadoop.hbase.testclassification.MediumTests;
|
||||
import org.apache.hadoop.hbase.util.Bytes;
|
||||
import org.apache.http.HttpStatus;
|
||||
import org.apache.http.client.entity.EntityBuilder;
|
||||
import org.apache.http.client.methods.CloseableHttpResponse;
|
||||
import org.apache.http.client.methods.HttpGet;
|
||||
import org.apache.http.client.methods.HttpPut;
|
||||
import org.apache.http.entity.ContentType;
|
||||
import org.apache.http.impl.client.CloseableHttpClient;
|
||||
import org.apache.http.impl.client.HttpClientBuilder;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Test;
|
||||
import org.junit.experimental.categories.Category;
|
||||
|
||||
@Category({ ClientTests.class, MediumTests.class })
|
||||
public class TestHttpProxyExample {
|
||||
|
||||
private static final HBaseTestingUtility UTIL = new HBaseTestingUtility();
|
||||
|
||||
private static final TableName TABLE_NAME = TableName.valueOf("test");
|
||||
|
||||
private static final String FAMILY = "cf";
|
||||
|
||||
private static final String QUALIFIER = "cq";
|
||||
|
||||
private static final String URL_TEMPLCATE = "http://localhost:%d/%s/%s/%s:%s";
|
||||
|
||||
private static final String ROW = "row";
|
||||
|
||||
private static final String VALUE = "value";
|
||||
|
||||
private static HttpProxyExample PROXY;
|
||||
|
||||
private static int PORT;
|
||||
|
||||
@BeforeClass
|
||||
public static void setUp() throws Exception {
|
||||
UTIL.startMiniCluster(1);
|
||||
UTIL.createTable(TABLE_NAME, Bytes.toBytes(FAMILY));
|
||||
PROXY = new HttpProxyExample(UTIL.getConfiguration(), 0);
|
||||
PROXY.start();
|
||||
PORT = PROXY.port();
|
||||
}
|
||||
|
||||
@AfterClass
|
||||
public static void tearDown() throws Exception {
|
||||
if (PROXY != null) {
|
||||
PROXY.stop();
|
||||
}
|
||||
UTIL.shutdownMiniCluster();
|
||||
}
|
||||
|
||||
@Test
|
||||
public void test() throws Exception {
|
||||
try (CloseableHttpClient client = HttpClientBuilder.create().build()) {
|
||||
HttpPut put = new HttpPut(
|
||||
String.format(URL_TEMPLCATE, PORT, TABLE_NAME.getNameAsString(), ROW, FAMILY, QUALIFIER));
|
||||
put.setEntity(EntityBuilder.create().setText(VALUE)
|
||||
.setContentType(ContentType.create("text-plain", StandardCharsets.UTF_8)).build());
|
||||
try (CloseableHttpResponse resp = client.execute(put)) {
|
||||
assertEquals(HttpStatus.SC_OK, resp.getStatusLine().getStatusCode());
|
||||
}
|
||||
HttpGet get = new HttpGet(
|
||||
String.format(URL_TEMPLCATE, PORT, TABLE_NAME.getNameAsString(), ROW, FAMILY, QUALIFIER));
|
||||
try (CloseableHttpResponse resp = client.execute(get)) {
|
||||
assertEquals(HttpStatus.SC_OK, resp.getStatusLine().getStatusCode());
|
||||
assertEquals("value",
|
||||
Bytes.toString(ByteStreams.toByteArray(resp.getEntity().getContent())));
|
||||
}
|
||||
get = new HttpGet(String.format(URL_TEMPLCATE, PORT, TABLE_NAME.getNameAsString(), "whatever",
|
||||
FAMILY, QUALIFIER));
|
||||
try (CloseableHttpResponse resp = client.execute(get)) {
|
||||
assertEquals(HttpStatus.SC_NOT_FOUND, resp.getStatusLine().getStatusCode());
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue