mirror of https://github.com/apache/druid.git
Add druid-rocketmq module
This commit is contained in:
parent
d0b10c29d7
commit
8eb332c1c4
|
@ -0,0 +1,50 @@
|
|||
<?xml version="1.0" encoding="UTF-8"?>
|
||||
<!--
|
||||
~ Druid - a distributed column store.
|
||||
~ Copyright 2012 - 2015 Metamarkets Group Inc.
|
||||
~
|
||||
~ Licensed under the Apache License, Version 2.0 (the "License");
|
||||
~ you may not use this file except in compliance with the License.
|
||||
~ You may obtain a copy of the License at
|
||||
~
|
||||
~ http://www.apache.org/licenses/LICENSE-2.0
|
||||
~
|
||||
~ Unless required by applicable law or agreed to in writing, software
|
||||
~ distributed under the License is distributed on an "AS IS" BASIS,
|
||||
~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
~ See the License for the specific language governing permissions and
|
||||
~ limitations under the License.
|
||||
-->
|
||||
<project xmlns="http://maven.apache.org/POM/4.0.0"
|
||||
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
|
||||
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
|
||||
<modelVersion>4.0.0</modelVersion>
|
||||
<parent>
|
||||
<artifactId>druid</artifactId>
|
||||
<groupId>io.druid</groupId>
|
||||
<version>0.9.0-SNAPSHOT</version>
|
||||
<relativePath>../../pom.xml</relativePath>
|
||||
</parent>
|
||||
|
||||
<properties>
|
||||
<rocketmq.version>3.2.6</rocketmq.version>
|
||||
</properties>
|
||||
|
||||
<artifactId>druid-rocketmq</artifactId>
|
||||
|
||||
<dependencies>
|
||||
|
||||
<dependency>
|
||||
<groupId>com.alibaba.rocketmq</groupId>
|
||||
<artifactId>rocketmq-client</artifactId>
|
||||
<version>${rocketmq.version}</version>
|
||||
</dependency>
|
||||
|
||||
<dependency>
|
||||
<groupId>io.druid</groupId>
|
||||
<artifactId>druid-api</artifactId>
|
||||
</dependency>
|
||||
|
||||
</dependencies>
|
||||
|
||||
</project>
|
|
@ -0,0 +1,46 @@
|
|||
/*
|
||||
* Licensed to Metamarkets Group Inc. (Metamarkets) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. Metamarkets licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing,
|
||||
* software distributed under the License is distributed on an
|
||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
* KIND, either express or implied. See the License for the
|
||||
* specific language governing permissions and limitations
|
||||
* under the License.
|
||||
*/
|
||||
package io.druid.firehose.rocketmq;
|
||||
|
||||
import com.fasterxml.jackson.databind.Module;
|
||||
import com.fasterxml.jackson.databind.jsontype.NamedType;
|
||||
import com.fasterxml.jackson.databind.module.SimpleModule;
|
||||
import com.google.common.collect.ImmutableList;
|
||||
import com.google.inject.Binder;
|
||||
import io.druid.initialization.DruidModule;
|
||||
|
||||
import java.util.List;
|
||||
|
||||
public class RocketMQDruidModule implements DruidModule {
|
||||
|
||||
@Override
|
||||
public List<? extends Module> getJacksonModules() {
|
||||
return ImmutableList.of(
|
||||
new SimpleModule("RocketMQFirehoseModule")
|
||||
.registerSubtypes(
|
||||
new NamedType(RocketMQFirehoseFactory.class, "RocketMQ-3.2.6")
|
||||
)
|
||||
);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void configure(Binder binder) {
|
||||
|
||||
}
|
||||
}
|
|
@ -0,0 +1,537 @@
|
|||
/*
|
||||
* Licensed to Metamarkets Group Inc. (Metamarkets) under one
|
||||
* or more contributor license agreements. See the NOTICE file
|
||||
* distributed with this work for additional information
|
||||
* regarding copyright ownership. Metamarkets licenses this file
|
||||
* to you under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance
|
||||
* with the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing,
|
||||
* software distributed under the License is distributed on an
|
||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
* KIND, either express or implied. See the License for the
|
||||
* specific language governing permissions and limitations
|
||||
* under the License.
|
||||
*/
|
||||
package io.druid.firehose.rocketmq;
|
||||
|
||||
import com.alibaba.rocketmq.client.consumer.DefaultMQPullConsumer;
|
||||
import com.alibaba.rocketmq.client.consumer.MessageQueueListener;
|
||||
import com.alibaba.rocketmq.client.consumer.PullResult;
|
||||
import com.alibaba.rocketmq.client.consumer.store.OffsetStore;
|
||||
import com.alibaba.rocketmq.client.consumer.store.ReadOffsetType;
|
||||
import com.alibaba.rocketmq.client.exception.MQBrokerException;
|
||||
import com.alibaba.rocketmq.client.exception.MQClientException;
|
||||
import com.alibaba.rocketmq.common.ServiceThread;
|
||||
import com.alibaba.rocketmq.common.message.MessageExt;
|
||||
import com.alibaba.rocketmq.common.message.MessageQueue;
|
||||
import com.alibaba.rocketmq.common.protocol.heartbeat.MessageModel;
|
||||
import com.alibaba.rocketmq.remoting.exception.RemotingException;
|
||||
import com.fasterxml.jackson.annotation.JsonCreator;
|
||||
import com.fasterxml.jackson.annotation.JsonProperty;
|
||||
import com.google.common.collect.Sets;
|
||||
import com.metamx.common.parsers.ParseException;
|
||||
import io.druid.data.input.ByteBufferInputRowParser;
|
||||
import io.druid.data.input.Firehose;
|
||||
import io.druid.data.input.FirehoseFactory;
|
||||
import io.druid.data.input.InputRow;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.nio.ByteBuffer;
|
||||
import java.util.*;
|
||||
import java.util.concurrent.ConcurrentHashMap;
|
||||
import java.util.concurrent.ConcurrentSkipListSet;
|
||||
import java.util.concurrent.CountDownLatch;
|
||||
|
||||
public class RocketMQFirehoseFactory implements FirehoseFactory<ByteBufferInputRowParser> {
|
||||
|
||||
private static final Logger LOGGER = LoggerFactory.getLogger(RocketMQFirehoseFactory.class);
|
||||
|
||||
/**
|
||||
* Passed in configuration for consumer client.
|
||||
*/
|
||||
@JsonProperty
|
||||
private final Properties consumerProps;
|
||||
|
||||
/**
|
||||
* Consumer group.
|
||||
*/
|
||||
@JsonProperty
|
||||
private final String consumerGroup;
|
||||
|
||||
/**
|
||||
* Topics to consume.
|
||||
* Multiple topics are separated by comma ",".
|
||||
*/
|
||||
@JsonProperty
|
||||
private final String feed;
|
||||
|
||||
|
||||
/**
|
||||
* Store messages that are fetched from brokers but not yet delivered to druid via fire hose.
|
||||
*/
|
||||
private ConcurrentHashMap<MessageQueue, ConcurrentSkipListSet<MessageExt>> messageQueueTreeSetMap = new ConcurrentHashMap<>();
|
||||
|
||||
/**
|
||||
* Store message consuming status.
|
||||
*/
|
||||
private ConcurrentHashMap<MessageQueue, ConcurrentSkipListSet<Long>> windows = new ConcurrentHashMap<>();
|
||||
|
||||
/**
|
||||
* Default pull batch size.
|
||||
*/
|
||||
private static final int PULL_BATCH_SIZE = 32;
|
||||
|
||||
@JsonCreator
|
||||
public RocketMQFirehoseFactory(@JsonProperty("consumerProps") Properties consumerProps,
|
||||
@JsonProperty("consumerGroup") String consumerGroup,
|
||||
@JsonProperty("feed") String feed) {
|
||||
this.consumerProps = consumerProps;
|
||||
for (Map.Entry<Object, Object> configItem : this.consumerProps.entrySet()) {
|
||||
System.setProperty(configItem.getKey().toString(), configItem.getValue().toString());
|
||||
}
|
||||
this.consumerGroup = consumerGroup;
|
||||
this.feed = feed;
|
||||
}
|
||||
|
||||
/**
|
||||
* Check if there are locally pending messages to consume.
|
||||
* @return true if there are some; false otherwise.
|
||||
*/
|
||||
private boolean hasMessagesPending() {
|
||||
|
||||
for (ConcurrentHashMap.Entry<MessageQueue, ConcurrentSkipListSet<MessageExt>> entry : messageQueueTreeSetMap.entrySet()) {
|
||||
if (!entry.getValue().isEmpty()) {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
@Override
|
||||
public Firehose connect(ByteBufferInputRowParser byteBufferInputRowParser) throws IOException, ParseException {
|
||||
|
||||
Set<String> newDimExclus = Sets.union(
|
||||
byteBufferInputRowParser.getParseSpec().getDimensionsSpec().getDimensionExclusions(),
|
||||
Sets.newHashSet("feed")
|
||||
);
|
||||
|
||||
final ByteBufferInputRowParser theParser = byteBufferInputRowParser.withParseSpec(
|
||||
byteBufferInputRowParser.getParseSpec()
|
||||
.withDimensionsSpec(
|
||||
byteBufferInputRowParser.getParseSpec()
|
||||
.getDimensionsSpec()
|
||||
.withDimensionExclusions(
|
||||
newDimExclus
|
||||
)
|
||||
)
|
||||
);
|
||||
|
||||
/**
|
||||
* Topic-Queue mapping.
|
||||
*/
|
||||
final ConcurrentHashMap<String, Set<MessageQueue>> topicQueueMap;
|
||||
|
||||
/**
|
||||
* Default Pull-style client for RocketMQ.
|
||||
*/
|
||||
final DefaultMQPullConsumer defaultMQPullConsumer;
|
||||
final DruidPullMessageService pullMessageService;
|
||||
|
||||
messageQueueTreeSetMap.clear();
|
||||
windows.clear();
|
||||
|
||||
try {
|
||||
defaultMQPullConsumer = new DefaultMQPullConsumer(this.consumerGroup);
|
||||
defaultMQPullConsumer.setMessageModel(MessageModel.CLUSTERING);
|
||||
topicQueueMap = new ConcurrentHashMap<>();
|
||||
|
||||
pullMessageService = new DruidPullMessageService(defaultMQPullConsumer);
|
||||
String[] topics = feed.split(",");
|
||||
for (String topic : topics) {
|
||||
topic = topic.trim();
|
||||
if (topic.isEmpty()) {
|
||||
continue;
|
||||
}
|
||||
defaultMQPullConsumer.fetchSubscribeMessageQueues(topic);
|
||||
topicQueueMap.put(topic, defaultMQPullConsumer.fetchMessageQueuesInBalance(topic));
|
||||
}
|
||||
DruidMessageQueueListener druidMessageQueueListener =
|
||||
new DruidMessageQueueListener(Sets.newHashSet(topics), topicQueueMap, defaultMQPullConsumer);
|
||||
defaultMQPullConsumer.setMessageQueueListener(druidMessageQueueListener);
|
||||
defaultMQPullConsumer.start();
|
||||
pullMessageService.start();
|
||||
} catch (MQClientException e) {
|
||||
LOGGER.error("Failed to start DefaultMQPullConsumer", e);
|
||||
throw new IOException("Failed to start RocketMQ client", e);
|
||||
}
|
||||
|
||||
return new Firehose() {
|
||||
|
||||
@Override
|
||||
public boolean hasMore() {
|
||||
boolean hasMore = false;
|
||||
DruidPullRequest earliestPullRequest = null;
|
||||
|
||||
for (Map.Entry<String, Set<MessageQueue>> entry : topicQueueMap.entrySet()) {
|
||||
for (MessageQueue messageQueue : entry.getValue()) {
|
||||
if (messageQueueTreeSetMap.keySet().contains(messageQueue)
|
||||
&& !messageQueueTreeSetMap.get(messageQueue).isEmpty()) {
|
||||
hasMore = true;
|
||||
} else {
|
||||
DruidPullRequest newPullRequest = new DruidPullRequest();
|
||||
newPullRequest.setMessageQueue(messageQueue);
|
||||
try {
|
||||
long offset = defaultMQPullConsumer.fetchConsumeOffset(messageQueue, false);
|
||||
newPullRequest.setNextBeginOffset(offset);
|
||||
} catch (MQClientException e) {
|
||||
LOGGER.error("Failed to fetch consume offset for queue: {}", entry.getKey());
|
||||
continue;
|
||||
}
|
||||
newPullRequest.setLongPull(!hasMessagesPending());
|
||||
|
||||
// notify pull message service to pull messages from brokers.
|
||||
pullMessageService.putRequest(newPullRequest);
|
||||
|
||||
// set the earliest pull in case we need to block.
|
||||
if (null == earliestPullRequest) {
|
||||
earliestPullRequest = newPullRequest;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Block only when there is no locally pending messages.
|
||||
if (!hasMore && null != earliestPullRequest) {
|
||||
try {
|
||||
earliestPullRequest.getCountDownLatch().await();
|
||||
hasMore = true;
|
||||
} catch (InterruptedException e) {
|
||||
LOGGER.error("CountDownLatch await got interrupted", e);
|
||||
}
|
||||
}
|
||||
return hasMore;
|
||||
}
|
||||
|
||||
@Override
|
||||
public InputRow nextRow() {
|
||||
for (Map.Entry<MessageQueue, ConcurrentSkipListSet<MessageExt>> entry : messageQueueTreeSetMap.entrySet()) {
|
||||
if (!entry.getValue().isEmpty()) {
|
||||
MessageExt message = entry.getValue().pollFirst();
|
||||
InputRow inputRow = theParser.parse(ByteBuffer.wrap(message.getBody()));
|
||||
|
||||
if (!windows.keySet().contains(entry.getKey())) {
|
||||
windows.put(entry.getKey(), new ConcurrentSkipListSet<Long>());
|
||||
}
|
||||
windows.get(entry.getKey()).add(message.getQueueOffset());
|
||||
return inputRow;
|
||||
}
|
||||
}
|
||||
|
||||
// should never happen.
|
||||
return null;
|
||||
}
|
||||
|
||||
@Override
|
||||
public Runnable commit() {
|
||||
return new Runnable() {
|
||||
@Override
|
||||
public void run() {
|
||||
OffsetStore offsetStore = defaultMQPullConsumer.getOffsetStore();
|
||||
Set<MessageQueue> updated = new HashSet<>();
|
||||
// calculate offsets according to consuming windows.
|
||||
for (ConcurrentHashMap.Entry<MessageQueue, ConcurrentSkipListSet<Long>> entry : windows.entrySet()) {
|
||||
while (!entry.getValue().isEmpty()) {
|
||||
|
||||
long offset = offsetStore.readOffset(entry.getKey(), ReadOffsetType.MEMORY_FIRST_THEN_STORE);
|
||||
if (offset + 1 > entry.getValue().first()) {
|
||||
entry.getValue().pollFirst();
|
||||
} else if (offset + 1 == entry.getValue().first()) {
|
||||
entry.getValue().pollFirst();
|
||||
offsetStore.updateOffset(entry.getKey(), offset + 1, true);
|
||||
updated.add(entry.getKey());
|
||||
} else {
|
||||
break;
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
offsetStore.persistAll(updated);
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
@Override
|
||||
public void close() throws IOException {
|
||||
defaultMQPullConsumer.shutdown();
|
||||
pullMessageService.shutdown(false);
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Pull request.
|
||||
*/
|
||||
class DruidPullRequest {
|
||||
private MessageQueue messageQueue;
|
||||
private String tag;
|
||||
private long nextBeginOffset;
|
||||
private int pullBatchSize;
|
||||
private boolean longPull;
|
||||
private CountDownLatch countDownLatch;
|
||||
private PullResult pullResult;
|
||||
private boolean successful;
|
||||
|
||||
public DruidPullRequest() {
|
||||
countDownLatch = new CountDownLatch(1);
|
||||
tag = "*";
|
||||
pullBatchSize = PULL_BATCH_SIZE;
|
||||
successful = false;
|
||||
}
|
||||
|
||||
public MessageQueue getMessageQueue() {
|
||||
return messageQueue;
|
||||
}
|
||||
|
||||
public void setMessageQueue(MessageQueue messageQueue) {
|
||||
this.messageQueue = messageQueue;
|
||||
}
|
||||
|
||||
public long getNextBeginOffset() {
|
||||
return nextBeginOffset;
|
||||
}
|
||||
|
||||
public String getTag() {
|
||||
return tag;
|
||||
}
|
||||
|
||||
public void setTag(String tag) {
|
||||
this.tag = tag;
|
||||
}
|
||||
|
||||
public void setNextBeginOffset(long nextBeginOffset) {
|
||||
this.nextBeginOffset = nextBeginOffset;
|
||||
}
|
||||
|
||||
public int getPullBatchSize() {
|
||||
return pullBatchSize;
|
||||
}
|
||||
|
||||
public void setPullBatchSize(int pullBatchSize) {
|
||||
this.pullBatchSize = pullBatchSize;
|
||||
}
|
||||
|
||||
public boolean isLongPull() {
|
||||
return longPull;
|
||||
}
|
||||
|
||||
public void setLongPull(boolean longPull) {
|
||||
this.longPull = longPull;
|
||||
}
|
||||
|
||||
public CountDownLatch getCountDownLatch() {
|
||||
return countDownLatch;
|
||||
}
|
||||
|
||||
public void setCountDownLatch(CountDownLatch countDownLatch) {
|
||||
this.countDownLatch = countDownLatch;
|
||||
}
|
||||
|
||||
public PullResult getPullResult() {
|
||||
return pullResult;
|
||||
}
|
||||
|
||||
public void setPullResult(PullResult pullResult) {
|
||||
this.pullResult = pullResult;
|
||||
}
|
||||
|
||||
public boolean isSuccessful() {
|
||||
return successful;
|
||||
}
|
||||
|
||||
public void setSuccessful(boolean successful) {
|
||||
this.successful = successful;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Pull message service for druid.
|
||||
*
|
||||
* <strong>Note: this is a single thread service.</strong>
|
||||
*/
|
||||
class DruidPullMessageService extends ServiceThread {
|
||||
|
||||
private volatile List<DruidPullRequest> requestsWrite = new ArrayList<DruidPullRequest>();
|
||||
private volatile List<DruidPullRequest> requestsRead = new ArrayList<DruidPullRequest>();
|
||||
|
||||
private final DefaultMQPullConsumer defaultMQPullConsumer;
|
||||
|
||||
public DruidPullMessageService(final DefaultMQPullConsumer defaultMQPullConsumer) {
|
||||
this.defaultMQPullConsumer = defaultMQPullConsumer;
|
||||
}
|
||||
|
||||
public void putRequest(final DruidPullRequest request) {
|
||||
synchronized (this) {
|
||||
this.requestsWrite.add(request);
|
||||
if (!hasNotified) {
|
||||
hasNotified = true;
|
||||
notify();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private void swapRequests() {
|
||||
List<DruidPullRequest> tmp = requestsWrite;
|
||||
requestsWrite = requestsRead;
|
||||
requestsRead = tmp;
|
||||
}
|
||||
|
||||
@Override
|
||||
public String getServiceName() {
|
||||
return getClass().getSimpleName();
|
||||
}
|
||||
|
||||
/**
|
||||
* Core message pulling logic code goes here.
|
||||
*/
|
||||
private void doPull() {
|
||||
for (DruidPullRequest pullRequest : requestsRead) {
|
||||
PullResult pullResult = null;
|
||||
try {
|
||||
if (!pullRequest.isLongPull()) {
|
||||
pullResult = defaultMQPullConsumer.pull(
|
||||
pullRequest.getMessageQueue(),
|
||||
pullRequest.getTag(),
|
||||
pullRequest.getNextBeginOffset(),
|
||||
pullRequest.getPullBatchSize());
|
||||
} else {
|
||||
pullResult = defaultMQPullConsumer.pullBlockIfNotFound(
|
||||
pullRequest.getMessageQueue(),
|
||||
pullRequest.getTag(),
|
||||
pullRequest.getNextBeginOffset(),
|
||||
pullRequest.getPullBatchSize()
|
||||
);
|
||||
}
|
||||
pullRequest.setPullResult(pullResult);
|
||||
pullRequest.setSuccessful(true);
|
||||
|
||||
if (!messageQueueTreeSetMap.keySet().contains(pullRequest.getMessageQueue())) {
|
||||
messageQueueTreeSetMap.putIfAbsent(pullRequest.getMessageQueue(),
|
||||
new ConcurrentSkipListSet<>(new MessageComparator()));
|
||||
}
|
||||
messageQueueTreeSetMap.get(pullRequest.getMessageQueue()).addAll(pullResult.getMsgFoundList());
|
||||
|
||||
} catch (MQClientException | RemotingException | MQBrokerException | InterruptedException e) {
|
||||
LOGGER.error("Failed to pull message from broker.", e);
|
||||
} finally {
|
||||
pullRequest.getCountDownLatch().countDown();
|
||||
}
|
||||
|
||||
}
|
||||
requestsRead.clear();
|
||||
}
|
||||
|
||||
/**
|
||||
* Thread looping entry.
|
||||
*/
|
||||
@Override
|
||||
public void run() {
|
||||
LOGGER.info(getServiceName() + " starts.");
|
||||
while (!isStoped()) {
|
||||
waitForRunning(0);
|
||||
doPull();
|
||||
}
|
||||
|
||||
// in case this service is shutdown gracefully without interruption.
|
||||
try {
|
||||
Thread.sleep(10);
|
||||
} catch (InterruptedException e) {
|
||||
LOGGER.error("", e);
|
||||
}
|
||||
|
||||
synchronized (this) {
|
||||
swapRequests();
|
||||
}
|
||||
|
||||
doPull();
|
||||
LOGGER.info(getServiceName() + " terminated.");
|
||||
}
|
||||
|
||||
@Override
|
||||
protected void onWaitEnd() {
|
||||
swapRequests();
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Compare messages pulled from same message queue according to queue offset.
|
||||
*/
|
||||
class MessageComparator implements Comparator<MessageExt> {
|
||||
@Override
|
||||
public int compare(MessageExt lhs, MessageExt rhs) {
|
||||
return lhs.getQueueOffset() < rhs.getQueueOffset() ? -1 : (lhs.getQueueOffset() == rhs.getQueueOffset() ? 0 : 1);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* Handle message queues re-balance operations.
|
||||
*/
|
||||
class DruidMessageQueueListener implements MessageQueueListener {
|
||||
|
||||
private Set<String> topics;
|
||||
|
||||
private final ConcurrentHashMap<String, Set<MessageQueue>> topicQueueMap;
|
||||
|
||||
private final DefaultMQPullConsumer defaultMQPullConsumer;
|
||||
|
||||
public DruidMessageQueueListener(final Set<String> topics,
|
||||
final ConcurrentHashMap<String, Set<MessageQueue>> topicQueueMap,
|
||||
final DefaultMQPullConsumer defaultMQPullConsumer) {
|
||||
this.topics = topics;
|
||||
this.topicQueueMap = topicQueueMap;
|
||||
this.defaultMQPullConsumer = defaultMQPullConsumer;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void messageQueueChanged(String topic, Set<MessageQueue> mqAll, Set<MessageQueue> mqDivided) {
|
||||
if (topics.contains(topic)) {
|
||||
topicQueueMap.put(topic, mqDivided);
|
||||
|
||||
// Remove message queues that are re-assigned to other clients.
|
||||
Iterator<ConcurrentHashMap.Entry<MessageQueue, ConcurrentSkipListSet<MessageExt>>> it =
|
||||
messageQueueTreeSetMap.entrySet().iterator();
|
||||
while (it.hasNext()) {
|
||||
if (!mqDivided.contains(it.next().getKey())) {
|
||||
it.remove();
|
||||
}
|
||||
}
|
||||
|
||||
StringBuilder stringBuilder = new StringBuilder();
|
||||
for (MessageQueue messageQueue : mqDivided) {
|
||||
stringBuilder.append(messageQueue.getBrokerName())
|
||||
.append("#")
|
||||
.append(messageQueue.getQueueId())
|
||||
.append(", ");
|
||||
}
|
||||
|
||||
if (LOGGER.isDebugEnabled() && stringBuilder.length() > 2) {
|
||||
LOGGER.debug(String.format("%s@%s is consuming the following message queues: %s",
|
||||
defaultMQPullConsumer.getClientIP(),
|
||||
defaultMQPullConsumer.getInstanceName(),
|
||||
stringBuilder.substring(0, stringBuilder.length() - 2) /*Remove the trailing comma*/));
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
}
|
1
pom.xml
1
pom.xml
|
@ -94,6 +94,7 @@
|
|||
<module>extensions/s3-extensions</module>
|
||||
<module>extensions/kafka-eight</module>
|
||||
<module>extensions/kafka-eight-simpleConsumer</module>
|
||||
<module>extensions/druid-rocketmq</module>
|
||||
<module>extensions/rabbitmq</module>
|
||||
<module>extensions/histogram</module>
|
||||
<module>extensions/mysql-metadata-storage</module>
|
||||
|
|
Loading…
Reference in New Issue