HBASE-14247 Separate the old WALs into different regionserver directories
This commit is contained in:
parent
9a27ac8781
commit
6db1ce1e91
|
@ -42,7 +42,7 @@ public class ReplicationQueueInfo {
|
||||||
private final String peerClusterZnode;
|
private final String peerClusterZnode;
|
||||||
private boolean queueRecovered;
|
private boolean queueRecovered;
|
||||||
// List of all the dead region servers that had this queue (if recovered)
|
// List of all the dead region servers that had this queue (if recovered)
|
||||||
private List<String> deadRegionServers = new ArrayList<>();
|
private List<ServerName> deadRegionServers = new ArrayList<>();
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* The passed znode will be either the id of the peer cluster or
|
* The passed znode will be either the id of the peer cluster or
|
||||||
|
@ -66,7 +66,7 @@ public class ReplicationQueueInfo {
|
||||||
* cases: 2-ip-10-46-221-101.ec2.internal,52170,1364333181125-<server name>-...
|
* cases: 2-ip-10-46-221-101.ec2.internal,52170,1364333181125-<server name>-...
|
||||||
*/
|
*/
|
||||||
private static void
|
private static void
|
||||||
extractDeadServersFromZNodeString(String deadServerListStr, List<String> result) {
|
extractDeadServersFromZNodeString(String deadServerListStr, List<ServerName> result) {
|
||||||
|
|
||||||
if(deadServerListStr == null || result == null || deadServerListStr.isEmpty()) return;
|
if(deadServerListStr == null || result == null || deadServerListStr.isEmpty()) return;
|
||||||
|
|
||||||
|
@ -85,7 +85,7 @@ public class ReplicationQueueInfo {
|
||||||
if (i > startIndex) {
|
if (i > startIndex) {
|
||||||
String serverName = deadServerListStr.substring(startIndex, i);
|
String serverName = deadServerListStr.substring(startIndex, i);
|
||||||
if(ServerName.isFullServerName(serverName)){
|
if(ServerName.isFullServerName(serverName)){
|
||||||
result.add(serverName);
|
result.add(ServerName.valueOf(serverName));
|
||||||
} else {
|
} else {
|
||||||
LOG.error("Found invalid server name:" + serverName);
|
LOG.error("Found invalid server name:" + serverName);
|
||||||
}
|
}
|
||||||
|
@ -103,7 +103,7 @@ public class ReplicationQueueInfo {
|
||||||
if(startIndex < len - 1){
|
if(startIndex < len - 1){
|
||||||
String serverName = deadServerListStr.substring(startIndex, len);
|
String serverName = deadServerListStr.substring(startIndex, len);
|
||||||
if(ServerName.isFullServerName(serverName)){
|
if(ServerName.isFullServerName(serverName)){
|
||||||
result.add(serverName);
|
result.add(ServerName.valueOf(serverName));
|
||||||
} else {
|
} else {
|
||||||
LOG.error("Found invalid server name at the end:" + serverName);
|
LOG.error("Found invalid server name at the end:" + serverName);
|
||||||
}
|
}
|
||||||
|
@ -112,7 +112,7 @@ public class ReplicationQueueInfo {
|
||||||
LOG.debug("Found dead servers:" + result);
|
LOG.debug("Found dead servers:" + result);
|
||||||
}
|
}
|
||||||
|
|
||||||
public List<String> getDeadRegionServers() {
|
public List<ServerName> getDeadRegionServers() {
|
||||||
return Collections.unmodifiableList(this.deadRegionServers);
|
return Collections.unmodifiableList(this.deadRegionServers);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -141,7 +141,12 @@ public abstract class CleanerChore<T extends FileCleanerDelegate> extends Schedu
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private void preRunCleaner() {
|
||||||
|
cleanersChain.forEach(FileCleanerDelegate::preClean);
|
||||||
|
}
|
||||||
|
|
||||||
public Boolean runCleaner() {
|
public Boolean runCleaner() {
|
||||||
|
preRunCleaner();
|
||||||
try {
|
try {
|
||||||
FileStatus[] files = FSUtils.listStatus(this.fs, this.oldFileDir);
|
FileStatus[] files = FSUtils.listStatus(this.fs, this.oldFileDir);
|
||||||
checkAndDeleteEntries(files);
|
checkAndDeleteEntries(files);
|
||||||
|
|
|
@ -44,4 +44,10 @@ public interface FileCleanerDelegate extends Configurable, Stoppable {
|
||||||
* this method is used to pass some instance into subclass
|
* this method is used to pass some instance into subclass
|
||||||
* */
|
* */
|
||||||
void init(Map<String, Object> params);
|
void init(Map<String, Object> params);
|
||||||
}
|
|
||||||
|
/**
|
||||||
|
* Used to do some initialize work before every period clean
|
||||||
|
*/
|
||||||
|
default void preClean() {
|
||||||
|
}
|
||||||
|
}
|
|
@ -43,6 +43,7 @@ import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
||||||
import org.apache.hadoop.hbase.util.FSUtils;
|
import org.apache.hadoop.hbase.util.FSUtils;
|
||||||
import org.apache.hadoop.hbase.util.HasThread;
|
import org.apache.hadoop.hbase.util.HasThread;
|
||||||
import org.apache.hadoop.hbase.util.Threads;
|
import org.apache.hadoop.hbase.util.Threads;
|
||||||
|
import org.apache.hadoop.hbase.wal.AbstractFSWALProvider;
|
||||||
import org.apache.hadoop.hbase.wal.FSHLogProvider;
|
import org.apache.hadoop.hbase.wal.FSHLogProvider;
|
||||||
import org.apache.hadoop.hbase.wal.WALEdit;
|
import org.apache.hadoop.hbase.wal.WALEdit;
|
||||||
import org.apache.hadoop.hbase.wal.WALFactory;
|
import org.apache.hadoop.hbase.wal.WALFactory;
|
||||||
|
@ -58,7 +59,6 @@ import org.apache.htrace.Span;
|
||||||
import org.apache.htrace.Trace;
|
import org.apache.htrace.Trace;
|
||||||
import org.apache.htrace.TraceScope;
|
import org.apache.htrace.TraceScope;
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
|
import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
|
||||||
|
|
||||||
import com.lmax.disruptor.BlockingWaitStrategy;
|
import com.lmax.disruptor.BlockingWaitStrategy;
|
||||||
|
@ -797,7 +797,11 @@ public class FSHLog extends AbstractFSWAL<Writer> {
|
||||||
}
|
}
|
||||||
|
|
||||||
final Path baseDir = FSUtils.getWALRootDir(conf);
|
final Path baseDir = FSUtils.getWALRootDir(conf);
|
||||||
final Path archiveDir = new Path(baseDir, HConstants.HREGION_OLDLOGDIR_NAME);
|
Path archiveDir = new Path(baseDir, HConstants.HREGION_OLDLOGDIR_NAME);
|
||||||
|
if (conf.getBoolean(AbstractFSWALProvider.SEPARATE_OLDLOGDIR,
|
||||||
|
AbstractFSWALProvider.DEFAULT_SEPARATE_OLDLOGDIR)) {
|
||||||
|
archiveDir = new Path(archiveDir, p.getName());
|
||||||
|
}
|
||||||
WALSplitter.split(baseDir, p, archiveDir, fs, conf, WALFactory.getInstance(conf));
|
WALSplitter.split(baseDir, p, archiveDir, fs, conf, WALFactory.getInstance(conf));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1141,10 +1145,10 @@ public class FSHLog extends AbstractFSWAL<Writer> {
|
||||||
System.err.println("Arguments:");
|
System.err.println("Arguments:");
|
||||||
System.err.println(" --dump Dump textual representation of passed one or more files");
|
System.err.println(" --dump Dump textual representation of passed one or more files");
|
||||||
System.err.println(" For example: "
|
System.err.println(" For example: "
|
||||||
+ "FSHLog --dump hdfs://example.com:9000/hbase/.logs/MACHINE/LOGFILE");
|
+ "FSHLog --dump hdfs://example.com:9000/hbase/WALs/MACHINE/LOGFILE");
|
||||||
System.err.println(" --split Split the passed directory of WAL logs");
|
System.err.println(" --split Split the passed directory of WAL logs");
|
||||||
System.err.println(
|
System.err.println(
|
||||||
" For example: " + "FSHLog --split hdfs://example.com:9000/hbase/.logs/DIR");
|
" For example: " + "FSHLog --split hdfs://example.com:9000/hbase/WALs/DIR");
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -26,22 +26,19 @@ import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.FileStatus;
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
import org.apache.hadoop.hbase.Abortable;
|
import org.apache.hadoop.hbase.Abortable;
|
||||||
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
|
import org.apache.hadoop.hbase.HBaseInterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
|
||||||
import org.apache.hadoop.hbase.master.cleaner.BaseLogCleanerDelegate;
|
import org.apache.hadoop.hbase.master.cleaner.BaseLogCleanerDelegate;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationException;
|
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationFactory;
|
import org.apache.hadoop.hbase.replication.ReplicationFactory;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationQueuesClient;
|
import org.apache.hadoop.hbase.replication.ReplicationQueuesClient;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationQueuesClientArguments;
|
import org.apache.hadoop.hbase.replication.ReplicationQueuesClientArguments;
|
||||||
import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
|
import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.Collections;
|
import java.util.Collections;
|
||||||
import java.util.List;
|
|
||||||
import java.util.Set;
|
import java.util.Set;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.shaded.com.google.common.base.Predicate;
|
import org.apache.hadoop.hbase.shaded.com.google.common.base.Predicate;
|
||||||
import org.apache.hadoop.hbase.shaded.com.google.common.collect.ImmutableSet;
|
|
||||||
import org.apache.hadoop.hbase.shaded.com.google.common.collect.Iterables;
|
import org.apache.hadoop.hbase.shaded.com.google.common.collect.Iterables;
|
||||||
import org.apache.hadoop.hbase.shaded.com.google.common.collect.Sets;
|
import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
|
||||||
import org.apache.zookeeper.KeeperException;
|
import org.apache.zookeeper.KeeperException;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -54,23 +51,31 @@ public class ReplicationLogCleaner extends BaseLogCleanerDelegate {
|
||||||
private ZooKeeperWatcher zkw;
|
private ZooKeeperWatcher zkw;
|
||||||
private ReplicationQueuesClient replicationQueues;
|
private ReplicationQueuesClient replicationQueues;
|
||||||
private boolean stopped = false;
|
private boolean stopped = false;
|
||||||
|
private Set<String> wals;
|
||||||
|
private long readZKTimestamp = 0;
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public Iterable<FileStatus> getDeletableFiles(Iterable<FileStatus> files) {
|
public void preClean() {
|
||||||
// all members of this class are null if replication is disabled,
|
readZKTimestamp = EnvironmentEdgeManager.currentTime();
|
||||||
// so we cannot filter the files
|
|
||||||
if (this.getConf() == null) {
|
|
||||||
return files;
|
|
||||||
}
|
|
||||||
|
|
||||||
final Set<String> wals;
|
|
||||||
try {
|
try {
|
||||||
// The concurrently created new WALs may not be included in the return list,
|
// The concurrently created new WALs may not be included in the return list,
|
||||||
// but they won't be deleted because they're not in the checking set.
|
// but they won't be deleted because they're not in the checking set.
|
||||||
wals = replicationQueues.getAllWALs();
|
wals = replicationQueues.getAllWALs();
|
||||||
} catch (KeeperException e) {
|
} catch (KeeperException e) {
|
||||||
LOG.warn("Failed to read zookeeper, skipping checking deletable files");
|
LOG.warn("Failed to read zookeeper, skipping checking deletable files");
|
||||||
|
wals = null;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public Iterable<FileStatus> getDeletableFiles(Iterable<FileStatus> files) {
|
||||||
|
// all members of this class are null if replication is disabled,
|
||||||
|
// so we cannot filter the files
|
||||||
|
if (this.getConf() == null) {
|
||||||
|
return files;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (wals == null) {
|
||||||
return Collections.emptyList();
|
return Collections.emptyList();
|
||||||
}
|
}
|
||||||
return Iterables.filter(files, new Predicate<FileStatus>() {
|
return Iterables.filter(files, new Predicate<FileStatus>() {
|
||||||
|
@ -85,8 +90,9 @@ public class ReplicationLogCleaner extends BaseLogCleanerDelegate {
|
||||||
LOG.debug("Didn't find this log in ZK, deleting: " + wal);
|
LOG.debug("Didn't find this log in ZK, deleting: " + wal);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return !logInReplicationQueue;
|
return !logInReplicationQueue && (file.getModificationTime() < readZKTimestamp);
|
||||||
}});
|
}
|
||||||
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
|
@ -36,6 +36,7 @@ import org.apache.hadoop.fs.FileStatus;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.hbase.Abortable;
|
import org.apache.hadoop.hbase.Abortable;
|
||||||
import org.apache.hadoop.hbase.HBaseConfiguration;
|
import org.apache.hadoop.hbase.HBaseConfiguration;
|
||||||
|
import org.apache.hadoop.hbase.ServerName;
|
||||||
import org.apache.hadoop.hbase.Stoppable;
|
import org.apache.hadoop.hbase.Stoppable;
|
||||||
import org.apache.hadoop.hbase.client.Admin;
|
import org.apache.hadoop.hbase.client.Admin;
|
||||||
import org.apache.hadoop.hbase.client.ClusterConnection;
|
import org.apache.hadoop.hbase.client.ClusterConnection;
|
||||||
|
@ -58,7 +59,6 @@ import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
|
||||||
import org.apache.hadoop.util.Tool;
|
import org.apache.hadoop.util.Tool;
|
||||||
import org.apache.hadoop.util.ToolRunner;
|
import org.apache.hadoop.util.ToolRunner;
|
||||||
import org.apache.zookeeper.KeeperException;
|
import org.apache.zookeeper.KeeperException;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.AtomicLongMap;
|
import org.apache.hadoop.hbase.shaded.com.google.common.util.concurrent.AtomicLongMap;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -356,7 +356,7 @@ public class DumpReplicationQueues extends Configured implements Tool {
|
||||||
|
|
||||||
StringBuilder sb = new StringBuilder();
|
StringBuilder sb = new StringBuilder();
|
||||||
|
|
||||||
List<String> deadServers ;
|
List<ServerName> deadServers;
|
||||||
|
|
||||||
sb.append("Dumping replication queue info for RegionServer: [" + regionserver + "]" + "\n");
|
sb.append("Dumping replication queue info for RegionServer: [" + regionserver + "]" + "\n");
|
||||||
sb.append(" Queue znode: " + queueId + "\n");
|
sb.append(" Queue znode: " + queueId + "\n");
|
||||||
|
@ -385,6 +385,7 @@ public class DumpReplicationQueues extends Configured implements Tool {
|
||||||
}
|
}
|
||||||
return sb.toString();
|
return sb.toString();
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* return total size in bytes from a list of WALs
|
* return total size in bytes from a list of WALs
|
||||||
*/
|
*/
|
||||||
|
|
|
@ -29,7 +29,8 @@ import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.FileStatus;
|
import org.apache.hadoop.fs.FileStatus;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.Stoppable;
|
import org.apache.hadoop.hbase.Server;
|
||||||
|
import org.apache.hadoop.hbase.ServerName;
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationEndpoint;
|
import org.apache.hadoop.hbase.replication.ReplicationEndpoint;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationPeers;
|
import org.apache.hadoop.hbase.replication.ReplicationPeers;
|
||||||
|
@ -51,10 +52,10 @@ public class RecoveredReplicationSource extends ReplicationSource {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
public void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
||||||
ReplicationQueues replicationQueues, ReplicationPeers replicationPeers, Stoppable stopper,
|
ReplicationQueues replicationQueues, ReplicationPeers replicationPeers, Server server,
|
||||||
String peerClusterZnode, UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
String peerClusterZnode, UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
||||||
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException {
|
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException {
|
||||||
super.init(conf, fs, manager, replicationQueues, replicationPeers, stopper, peerClusterZnode,
|
super.init(conf, fs, manager, replicationQueues, replicationPeers, server, peerClusterZnode,
|
||||||
clusterId, replicationEndpoint, walFileLengthProvider, metrics);
|
clusterId, replicationEndpoint, walFileLengthProvider, metrics);
|
||||||
this.actualPeerId = this.replicationQueueInfo.getPeerId();
|
this.actualPeerId = this.replicationQueueInfo.getPeerId();
|
||||||
}
|
}
|
||||||
|
@ -98,7 +99,7 @@ public class RecoveredReplicationSource extends ReplicationSource {
|
||||||
}
|
}
|
||||||
// Path changed - try to find the right path.
|
// Path changed - try to find the right path.
|
||||||
hasPathChanged = true;
|
hasPathChanged = true;
|
||||||
if (stopper instanceof ReplicationSyncUp.DummyServer) {
|
if (server instanceof ReplicationSyncUp.DummyServer) {
|
||||||
// In the case of disaster/recovery, HMaster may be shutdown/crashed before flush data
|
// In the case of disaster/recovery, HMaster may be shutdown/crashed before flush data
|
||||||
// from .logs to .oldlogs. Loop into .logs folders and check whether a match exists
|
// from .logs to .oldlogs. Loop into .logs folders and check whether a match exists
|
||||||
Path newPath = getReplSyncUpPath(path);
|
Path newPath = getReplSyncUpPath(path);
|
||||||
|
@ -107,12 +108,13 @@ public class RecoveredReplicationSource extends ReplicationSource {
|
||||||
} else {
|
} else {
|
||||||
// See if Path exists in the dead RS folder (there could be a chain of failures
|
// See if Path exists in the dead RS folder (there could be a chain of failures
|
||||||
// to look at)
|
// to look at)
|
||||||
List<String> deadRegionServers = this.replicationQueueInfo.getDeadRegionServers();
|
List<ServerName> deadRegionServers = this.replicationQueueInfo.getDeadRegionServers();
|
||||||
LOG.info("NB dead servers : " + deadRegionServers.size());
|
LOG.info("NB dead servers : " + deadRegionServers.size());
|
||||||
final Path walDir = FSUtils.getWALRootDir(conf);
|
final Path walDir = FSUtils.getWALRootDir(conf);
|
||||||
for (String curDeadServerName : deadRegionServers) {
|
for (ServerName curDeadServerName : deadRegionServers) {
|
||||||
final Path deadRsDirectory =
|
final Path deadRsDirectory =
|
||||||
new Path(walDir, AbstractFSWALProvider.getWALDirectoryName(curDeadServerName));
|
new Path(walDir, AbstractFSWALProvider.getWALDirectoryName(curDeadServerName
|
||||||
|
.getServerName()));
|
||||||
Path[] locs = new Path[] { new Path(deadRsDirectory, path.getName()), new Path(
|
Path[] locs = new Path[] { new Path(deadRsDirectory, path.getName()), new Path(
|
||||||
deadRsDirectory.suffix(AbstractFSWALProvider.SPLITTING_EXT), path.getName()) };
|
deadRsDirectory.suffix(AbstractFSWALProvider.SPLITTING_EXT), path.getName()) };
|
||||||
for (Path possibleLogLocation : locs) {
|
for (Path possibleLogLocation : locs) {
|
||||||
|
@ -189,4 +191,9 @@ public class RecoveredReplicationSource extends ReplicationSource {
|
||||||
public String getPeerId() {
|
public String getPeerId() {
|
||||||
return this.actualPeerId;
|
return this.actualPeerId;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ServerName getServerWALsBelongTo() {
|
||||||
|
return this.replicationQueueInfo.getDeadRegionServers().get(0);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -40,6 +40,8 @@ import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.HBaseConfiguration;
|
import org.apache.hadoop.hbase.HBaseConfiguration;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
|
import org.apache.hadoop.hbase.Server;
|
||||||
|
import org.apache.hadoop.hbase.ServerName;
|
||||||
import org.apache.hadoop.hbase.Stoppable;
|
import org.apache.hadoop.hbase.Stoppable;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
|
@ -59,7 +61,6 @@ import org.apache.hadoop.hbase.util.Pair;
|
||||||
import org.apache.hadoop.hbase.util.Threads;
|
import org.apache.hadoop.hbase.util.Threads;
|
||||||
import org.apache.hadoop.hbase.wal.AbstractFSWALProvider;
|
import org.apache.hadoop.hbase.wal.AbstractFSWALProvider;
|
||||||
import org.apache.hadoop.hbase.wal.WAL.Entry;
|
import org.apache.hadoop.hbase.wal.WAL.Entry;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;
|
import org.apache.hadoop.hbase.shaded.com.google.common.collect.Lists;
|
||||||
|
|
||||||
|
|
||||||
|
@ -94,7 +95,7 @@ public class ReplicationSource extends Thread implements ReplicationSourceInterf
|
||||||
// The manager of all sources to which we ping back our progress
|
// The manager of all sources to which we ping back our progress
|
||||||
protected ReplicationSourceManager manager;
|
protected ReplicationSourceManager manager;
|
||||||
// Should we stop everything?
|
// Should we stop everything?
|
||||||
protected Stoppable stopper;
|
protected Server server;
|
||||||
// How long should we sleep for each retry
|
// How long should we sleep for each retry
|
||||||
private long sleepForRetries;
|
private long sleepForRetries;
|
||||||
protected FileSystem fs;
|
protected FileSystem fs;
|
||||||
|
@ -139,7 +140,7 @@ public class ReplicationSource extends Thread implements ReplicationSourceInterf
|
||||||
* @param conf configuration to use
|
* @param conf configuration to use
|
||||||
* @param fs file system to use
|
* @param fs file system to use
|
||||||
* @param manager replication manager to ping to
|
* @param manager replication manager to ping to
|
||||||
* @param stopper the atomic boolean to use to stop the regionserver
|
* @param server the server for this region server
|
||||||
* @param peerClusterZnode the name of our znode
|
* @param peerClusterZnode the name of our znode
|
||||||
* @param clusterId unique UUID for the cluster
|
* @param clusterId unique UUID for the cluster
|
||||||
* @param replicationEndpoint the replication endpoint implementation
|
* @param replicationEndpoint the replication endpoint implementation
|
||||||
|
@ -148,10 +149,10 @@ public class ReplicationSource extends Thread implements ReplicationSourceInterf
|
||||||
*/
|
*/
|
||||||
@Override
|
@Override
|
||||||
public void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
public void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
||||||
ReplicationQueues replicationQueues, ReplicationPeers replicationPeers, Stoppable stopper,
|
ReplicationQueues replicationQueues, ReplicationPeers replicationPeers, Server server,
|
||||||
String peerClusterZnode, UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
String peerClusterZnode, UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
||||||
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException {
|
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException {
|
||||||
this.stopper = stopper;
|
this.server = server;
|
||||||
this.conf = HBaseConfiguration.create(conf);
|
this.conf = HBaseConfiguration.create(conf);
|
||||||
this.waitOnEndpointSeconds =
|
this.waitOnEndpointSeconds =
|
||||||
this.conf.getInt(WAIT_ON_ENDPOINT_SECONDS, DEFAULT_WAIT_ON_ENDPOINT_SECONDS);
|
this.conf.getInt(WAIT_ON_ENDPOINT_SECONDS, DEFAULT_WAIT_ON_ENDPOINT_SECONDS);
|
||||||
|
@ -330,7 +331,7 @@ public class ReplicationSource extends Thread implements ReplicationSourceInterf
|
||||||
public void uncaughtException(final Thread t, final Throwable e) {
|
public void uncaughtException(final Thread t, final Throwable e) {
|
||||||
RSRpcServices.exitIfOOME(e);
|
RSRpcServices.exitIfOOME(e);
|
||||||
LOG.error("Unexpected exception in " + t.getName() + " currentPath=" + getCurrentPath(), e);
|
LOG.error("Unexpected exception in " + t.getName() + " currentPath=" + getCurrentPath(), e);
|
||||||
stopper.stop("Unexpected exception in " + t.getName());
|
server.stop("Unexpected exception in " + t.getName());
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
@ -500,7 +501,7 @@ public class ReplicationSource extends Thread implements ReplicationSourceInterf
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public boolean isSourceActive() {
|
public boolean isSourceActive() {
|
||||||
return !this.stopper.isStopped() && this.sourceRunning;
|
return !this.server.isStopped() && this.sourceRunning;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -564,4 +565,9 @@ public class ReplicationSource extends Thread implements ReplicationSourceInterf
|
||||||
public WALFileLengthProvider getWALFileLengthProvider() {
|
public WALFileLengthProvider getWALFileLengthProvider() {
|
||||||
return walFileLengthProvider;
|
return walFileLengthProvider;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ServerName getServerWALsBelongTo() {
|
||||||
|
return server.getServerName();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -26,7 +26,8 @@ import org.apache.yetus.audience.InterfaceAudience;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.Stoppable;
|
import org.apache.hadoop.hbase.Server;
|
||||||
|
import org.apache.hadoop.hbase.ServerName;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationEndpoint;
|
import org.apache.hadoop.hbase.replication.ReplicationEndpoint;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationException;
|
import org.apache.hadoop.hbase.replication.ReplicationException;
|
||||||
|
@ -48,13 +49,13 @@ public interface ReplicationSourceInterface {
|
||||||
* @param manager the manager to use
|
* @param manager the manager to use
|
||||||
* @param replicationQueues
|
* @param replicationQueues
|
||||||
* @param replicationPeers
|
* @param replicationPeers
|
||||||
* @param stopper the stopper object for this region server
|
* @param server the server for this region server
|
||||||
* @param peerClusterZnode
|
* @param peerClusterZnode
|
||||||
* @param clusterId
|
* @param clusterId
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
*/
|
*/
|
||||||
void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
||||||
ReplicationQueues replicationQueues, ReplicationPeers replicationPeers, Stoppable stopper,
|
ReplicationQueues replicationQueues, ReplicationPeers replicationPeers, Server server,
|
||||||
String peerClusterZnode, UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
String peerClusterZnode, UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
||||||
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException;
|
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException;
|
||||||
|
|
||||||
|
@ -163,4 +164,11 @@ public interface ReplicationSourceInterface {
|
||||||
* @param batchSize entries size pushed
|
* @param batchSize entries size pushed
|
||||||
*/
|
*/
|
||||||
void postShipEdits(List<Entry> entries, int batchSize);
|
void postShipEdits(List<Entry> entries, int batchSize);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* The queue of WALs only belong to one region server. This will return the server name which all
|
||||||
|
* WALs belong to.
|
||||||
|
* @return the server name which all WALs belong to
|
||||||
|
*/
|
||||||
|
ServerName getServerWALsBelongTo();
|
||||||
}
|
}
|
||||||
|
|
|
@ -127,8 +127,10 @@ public class ReplicationSourceWALReader extends Thread {
|
||||||
public void run() {
|
public void run() {
|
||||||
int sleepMultiplier = 1;
|
int sleepMultiplier = 1;
|
||||||
while (isReaderRunning()) { // we only loop back here if something fatal happened to our stream
|
while (isReaderRunning()) { // we only loop back here if something fatal happened to our stream
|
||||||
try (WALEntryStream entryStream = new WALEntryStream(logQueue, fs, conf, currentPosition,
|
try (WALEntryStream entryStream =
|
||||||
source.getWALFileLengthProvider(), source.getSourceMetrics())) {
|
new WALEntryStream(logQueue, fs, conf, currentPosition,
|
||||||
|
source.getWALFileLengthProvider(), source.getServerWALsBelongTo(),
|
||||||
|
source.getSourceMetrics())) {
|
||||||
while (isReaderRunning()) { // loop here to keep reusing stream while we can
|
while (isReaderRunning()) { // loop here to keep reusing stream while we can
|
||||||
if (!checkQuota()) {
|
if (!checkQuota()) {
|
||||||
continue;
|
continue;
|
||||||
|
|
|
@ -32,12 +32,14 @@ import org.apache.hadoop.fs.FileStatus;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
|
import org.apache.hadoop.hbase.ServerName;
|
||||||
import org.apache.yetus.audience.InterfaceAudience;
|
import org.apache.yetus.audience.InterfaceAudience;
|
||||||
import org.apache.yetus.audience.InterfaceStability;
|
import org.apache.yetus.audience.InterfaceStability;
|
||||||
import org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader;
|
import org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader;
|
||||||
import org.apache.hadoop.hbase.util.CancelableProgressable;
|
import org.apache.hadoop.hbase.util.CancelableProgressable;
|
||||||
import org.apache.hadoop.hbase.util.FSUtils;
|
import org.apache.hadoop.hbase.util.FSUtils;
|
||||||
import org.apache.hadoop.hbase.util.LeaseNotRecoveredException;
|
import org.apache.hadoop.hbase.util.LeaseNotRecoveredException;
|
||||||
|
import org.apache.hadoop.hbase.wal.AbstractFSWALProvider;
|
||||||
import org.apache.hadoop.hbase.wal.WAL.Entry;
|
import org.apache.hadoop.hbase.wal.WAL.Entry;
|
||||||
import org.apache.hadoop.hbase.wal.WAL.Reader;
|
import org.apache.hadoop.hbase.wal.WAL.Reader;
|
||||||
import org.apache.hadoop.hbase.wal.WALFactory;
|
import org.apache.hadoop.hbase.wal.WALFactory;
|
||||||
|
@ -63,6 +65,8 @@ class WALEntryStream implements Closeable {
|
||||||
private final FileSystem fs;
|
private final FileSystem fs;
|
||||||
private final Configuration conf;
|
private final Configuration conf;
|
||||||
private final WALFileLengthProvider walFileLengthProvider;
|
private final WALFileLengthProvider walFileLengthProvider;
|
||||||
|
// which region server the WALs belong to
|
||||||
|
private final ServerName serverName;
|
||||||
private final MetricsSource metrics;
|
private final MetricsSource metrics;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -71,17 +75,19 @@ class WALEntryStream implements Closeable {
|
||||||
* @param fs {@link FileSystem} to use to create {@link Reader} for this stream
|
* @param fs {@link FileSystem} to use to create {@link Reader} for this stream
|
||||||
* @param conf {@link Configuration} to use to create {@link Reader} for this stream
|
* @param conf {@link Configuration} to use to create {@link Reader} for this stream
|
||||||
* @param startPosition the position in the first WAL to start reading at
|
* @param startPosition the position in the first WAL to start reading at
|
||||||
|
* @param serverName the server name which all WALs belong to
|
||||||
* @param metrics replication metrics
|
* @param metrics replication metrics
|
||||||
* @throws IOException
|
* @throws IOException
|
||||||
*/
|
*/
|
||||||
public WALEntryStream(PriorityBlockingQueue<Path> logQueue, FileSystem fs, Configuration conf,
|
public WALEntryStream(PriorityBlockingQueue<Path> logQueue, FileSystem fs, Configuration conf,
|
||||||
long startPosition, WALFileLengthProvider walFileLengthProvider, MetricsSource metrics)
|
long startPosition, WALFileLengthProvider walFileLengthProvider, ServerName serverName,
|
||||||
throws IOException {
|
MetricsSource metrics) throws IOException {
|
||||||
this.logQueue = logQueue;
|
this.logQueue = logQueue;
|
||||||
this.fs = fs;
|
this.fs = fs;
|
||||||
this.conf = conf;
|
this.conf = conf;
|
||||||
this.currentPosition = startPosition;
|
this.currentPosition = startPosition;
|
||||||
this.walFileLengthProvider = walFileLengthProvider;
|
this.walFileLengthProvider = walFileLengthProvider;
|
||||||
|
this.serverName = serverName;
|
||||||
this.metrics = metrics;
|
this.metrics = metrics;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -296,15 +302,27 @@ class WALEntryStream implements Closeable {
|
||||||
|
|
||||||
private Path getArchivedLog(Path path) throws IOException {
|
private Path getArchivedLog(Path path) throws IOException {
|
||||||
Path rootDir = FSUtils.getRootDir(conf);
|
Path rootDir = FSUtils.getRootDir(conf);
|
||||||
|
|
||||||
|
// Try found the log in old dir
|
||||||
Path oldLogDir = new Path(rootDir, HConstants.HREGION_OLDLOGDIR_NAME);
|
Path oldLogDir = new Path(rootDir, HConstants.HREGION_OLDLOGDIR_NAME);
|
||||||
Path archivedLogLocation = new Path(oldLogDir, path.getName());
|
Path archivedLogLocation = new Path(oldLogDir, path.getName());
|
||||||
if (fs.exists(archivedLogLocation)) {
|
if (fs.exists(archivedLogLocation)) {
|
||||||
LOG.info("Log " + path + " was moved to " + archivedLogLocation);
|
LOG.info("Log " + path + " was moved to " + archivedLogLocation);
|
||||||
return archivedLogLocation;
|
return archivedLogLocation;
|
||||||
} else {
|
|
||||||
LOG.error("Couldn't locate log: " + path);
|
|
||||||
return path;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Try found the log in the seperate old log dir
|
||||||
|
oldLogDir =
|
||||||
|
new Path(rootDir, new StringBuilder(HConstants.HREGION_OLDLOGDIR_NAME)
|
||||||
|
.append(Path.SEPARATOR).append(serverName.getServerName()).toString());
|
||||||
|
archivedLogLocation = new Path(oldLogDir, path.getName());
|
||||||
|
if (fs.exists(archivedLogLocation)) {
|
||||||
|
LOG.info("Log " + path + " was moved to " + archivedLogLocation);
|
||||||
|
return archivedLogLocation;
|
||||||
|
}
|
||||||
|
|
||||||
|
LOG.error("Couldn't locate log: " + path);
|
||||||
|
return path;
|
||||||
}
|
}
|
||||||
|
|
||||||
private void handleFileNotFound(Path path, FileNotFoundException fnfe) throws IOException {
|
private void handleFileNotFound(Path path, FileNotFoundException fnfe) throws IOException {
|
||||||
|
@ -316,6 +334,7 @@ class WALEntryStream implements Closeable {
|
||||||
throw fnfe;
|
throw fnfe;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
private void openReader(Path path) throws IOException {
|
private void openReader(Path path) throws IOException {
|
||||||
try {
|
try {
|
||||||
// Detect if this is a new file, if so get a new reader else
|
// Detect if this is a new file, if so get a new reader else
|
||||||
|
|
|
@ -40,7 +40,6 @@ import org.apache.hadoop.hbase.regionserver.wal.WALActionsListener;
|
||||||
import org.apache.hadoop.hbase.util.CancelableProgressable;
|
import org.apache.hadoop.hbase.util.CancelableProgressable;
|
||||||
import org.apache.hadoop.hbase.util.FSUtils;
|
import org.apache.hadoop.hbase.util.FSUtils;
|
||||||
import org.apache.hadoop.hbase.util.LeaseNotRecoveredException;
|
import org.apache.hadoop.hbase.util.LeaseNotRecoveredException;
|
||||||
|
|
||||||
import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
|
import org.apache.hadoop.hbase.shaded.com.google.common.annotations.VisibleForTesting;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -59,6 +58,10 @@ public abstract class AbstractFSWALProvider<T extends AbstractFSWAL<?>> implemen
|
||||||
|
|
||||||
private static final Log LOG = LogFactory.getLog(AbstractFSWALProvider.class);
|
private static final Log LOG = LogFactory.getLog(AbstractFSWALProvider.class);
|
||||||
|
|
||||||
|
/** Separate old log into different dir by regionserver name **/
|
||||||
|
public static final String SEPARATE_OLDLOGDIR = "hbase.separate.oldlogdir.by.regionserver";
|
||||||
|
public static final boolean DEFAULT_SEPARATE_OLDLOGDIR = false;
|
||||||
|
|
||||||
// Only public so classes back in regionserver.wal can access
|
// Only public so classes back in regionserver.wal can access
|
||||||
public interface Reader extends WAL.Reader {
|
public interface Reader extends WAL.Reader {
|
||||||
/**
|
/**
|
||||||
|
@ -272,6 +275,23 @@ public abstract class AbstractFSWALProvider<T extends AbstractFSWAL<?>> implemen
|
||||||
return dirName.toString();
|
return dirName.toString();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Construct the directory name for all old WALs on a given server. The default old WALs dir
|
||||||
|
* looks like: <code>hbase/oldWALs</code>. If you config hbase.separate.oldlogdir.by.regionserver
|
||||||
|
* to true, it looks like <code>hbase//oldWALs/kalashnikov.att.net,61634,1486865297088</code>.
|
||||||
|
* @param conf
|
||||||
|
* @param serverName Server name formatted as described in {@link ServerName}
|
||||||
|
* @return the relative WAL directory name
|
||||||
|
*/
|
||||||
|
public static String getWALArchiveDirectoryName(Configuration conf, final String serverName) {
|
||||||
|
StringBuilder dirName = new StringBuilder(HConstants.HREGION_OLDLOGDIR_NAME);
|
||||||
|
if (conf.getBoolean(SEPARATE_OLDLOGDIR, DEFAULT_SEPARATE_OLDLOGDIR)) {
|
||||||
|
dirName.append(Path.SEPARATOR);
|
||||||
|
dirName.append(serverName);
|
||||||
|
}
|
||||||
|
return dirName.toString();
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Pulls a ServerName out of a Path generated according to our layout rules. In the below layouts,
|
* Pulls a ServerName out of a Path generated according to our layout rules. In the below layouts,
|
||||||
* this method ignores the format of the logfile component. Current format: [base directory for
|
* this method ignores the format of the logfile component. Current format: [base directory for
|
||||||
|
@ -387,6 +407,14 @@ public abstract class AbstractFSWALProvider<T extends AbstractFSWAL<?>> implemen
|
||||||
public static Path getArchivedLogPath(Path path, Configuration conf) throws IOException {
|
public static Path getArchivedLogPath(Path path, Configuration conf) throws IOException {
|
||||||
Path rootDir = FSUtils.getRootDir(conf);
|
Path rootDir = FSUtils.getRootDir(conf);
|
||||||
Path oldLogDir = new Path(rootDir, HConstants.HREGION_OLDLOGDIR_NAME);
|
Path oldLogDir = new Path(rootDir, HConstants.HREGION_OLDLOGDIR_NAME);
|
||||||
|
if (conf.getBoolean(SEPARATE_OLDLOGDIR, DEFAULT_SEPARATE_OLDLOGDIR)) {
|
||||||
|
ServerName serverName = getServerNameFromWALDirectoryName(path);
|
||||||
|
if (serverName == null) {
|
||||||
|
LOG.error("Couldn't locate log: " + path);
|
||||||
|
return path;
|
||||||
|
}
|
||||||
|
oldLogDir = new Path(oldLogDir, serverName.getServerName());
|
||||||
|
}
|
||||||
Path archivedLogLocation = new Path(oldLogDir, path.getName());
|
Path archivedLogLocation = new Path(oldLogDir, path.getName());
|
||||||
final FileSystem fs = FSUtils.getCurrentFileSystem(conf);
|
final FileSystem fs = FSUtils.getCurrentFileSystem(conf);
|
||||||
|
|
||||||
|
|
|
@ -61,8 +61,9 @@ public class AsyncFSWALProvider extends AbstractFSWALProvider<AsyncFSWAL> {
|
||||||
@Override
|
@Override
|
||||||
protected AsyncFSWAL createWAL() throws IOException {
|
protected AsyncFSWAL createWAL() throws IOException {
|
||||||
return new AsyncFSWAL(FSUtils.getWALFileSystem(conf), FSUtils.getWALRootDir(conf),
|
return new AsyncFSWAL(FSUtils.getWALFileSystem(conf), FSUtils.getWALRootDir(conf),
|
||||||
getWALDirectoryName(factory.factoryId), HConstants.HREGION_OLDLOGDIR_NAME, conf, listeners,
|
getWALDirectoryName(factory.factoryId),
|
||||||
true, logPrefix, META_WAL_PROVIDER_ID.equals(providerId) ? META_WAL_PROVIDER_ID : null,
|
getWALArchiveDirectoryName(conf, factory.factoryId), conf, listeners, true, logPrefix,
|
||||||
|
META_WAL_PROVIDER_ID.equals(providerId) ? META_WAL_PROVIDER_ID : null,
|
||||||
eventLoopGroup.next(), channelClass);
|
eventLoopGroup.next(), channelClass);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -76,8 +76,9 @@ public class FSHLogProvider extends AbstractFSWALProvider<FSHLog> {
|
||||||
@Override
|
@Override
|
||||||
protected FSHLog createWAL() throws IOException {
|
protected FSHLog createWAL() throws IOException {
|
||||||
return new FSHLog(FSUtils.getWALFileSystem(conf), FSUtils.getWALRootDir(conf),
|
return new FSHLog(FSUtils.getWALFileSystem(conf), FSUtils.getWALRootDir(conf),
|
||||||
getWALDirectoryName(factory.factoryId), HConstants.HREGION_OLDLOGDIR_NAME, conf, listeners,
|
getWALDirectoryName(factory.factoryId),
|
||||||
true, logPrefix, META_WAL_PROVIDER_ID.equals(providerId) ? META_WAL_PROVIDER_ID : null);
|
getWALArchiveDirectoryName(conf, factory.factoryId), conf, listeners, true, logPrefix,
|
||||||
|
META_WAL_PROVIDER_ID.equals(providerId) ? META_WAL_PROVIDER_ID : null);
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
|
|
@ -225,6 +225,7 @@ public class TestLogsCleaner {
|
||||||
"testZooKeeperAbort-faulty", null)) {
|
"testZooKeeperAbort-faulty", null)) {
|
||||||
faultyZK.init();
|
faultyZK.init();
|
||||||
cleaner.setConf(conf, faultyZK);
|
cleaner.setConf(conf, faultyZK);
|
||||||
|
cleaner.preClean();
|
||||||
// should keep all files due to a ConnectionLossException getting the queues znodes
|
// should keep all files due to a ConnectionLossException getting the queues znodes
|
||||||
Iterable<FileStatus> toDelete = cleaner.getDeletableFiles(dummyFiles);
|
Iterable<FileStatus> toDelete = cleaner.getDeletableFiles(dummyFiles);
|
||||||
assertFalse(toDelete.iterator().hasNext());
|
assertFalse(toDelete.iterator().hasNext());
|
||||||
|
@ -235,6 +236,7 @@ public class TestLogsCleaner {
|
||||||
cleaner = new ReplicationLogCleaner();
|
cleaner = new ReplicationLogCleaner();
|
||||||
try (ZooKeeperWatcher zkw = new ZooKeeperWatcher(conf, "testZooKeeperAbort-normal", null)) {
|
try (ZooKeeperWatcher zkw = new ZooKeeperWatcher(conf, "testZooKeeperAbort-normal", null)) {
|
||||||
cleaner.setConf(conf, zkw);
|
cleaner.setConf(conf, zkw);
|
||||||
|
cleaner.preClean();
|
||||||
Iterable<FileStatus> filesToDelete = cleaner.getDeletableFiles(dummyFiles);
|
Iterable<FileStatus> filesToDelete = cleaner.getDeletableFiles(dummyFiles);
|
||||||
Iterator<FileStatus> iter = filesToDelete.iterator();
|
Iterator<FileStatus> iter = filesToDelete.iterator();
|
||||||
assertTrue(iter.hasNext());
|
assertTrue(iter.hasNext());
|
||||||
|
|
|
@ -25,7 +25,8 @@ import java.util.UUID;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
import org.apache.hadoop.hbase.Stoppable;
|
import org.apache.hadoop.hbase.Server;
|
||||||
|
import org.apache.hadoop.hbase.ServerName;
|
||||||
import org.apache.hadoop.hbase.TableName;
|
import org.apache.hadoop.hbase.TableName;
|
||||||
import org.apache.hadoop.hbase.replication.regionserver.MetricsSource;
|
import org.apache.hadoop.hbase.replication.regionserver.MetricsSource;
|
||||||
import org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceInterface;
|
import org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceInterface;
|
||||||
|
@ -47,7 +48,7 @@ public class ReplicationSourceDummy implements ReplicationSourceInterface {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
public void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
||||||
ReplicationQueues rq, ReplicationPeers rp, Stoppable stopper, String peerClusterId,
|
ReplicationQueues rq, ReplicationPeers rp, Server server, String peerClusterId,
|
||||||
UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
||||||
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException {
|
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException {
|
||||||
this.manager = manager;
|
this.manager = manager;
|
||||||
|
@ -142,4 +143,9 @@ public class ReplicationSourceDummy implements ReplicationSourceInterface {
|
||||||
public WALFileLengthProvider getWALFileLengthProvider() {
|
public WALFileLengthProvider getWALFileLengthProvider() {
|
||||||
return walFileLengthProvider;
|
return walFileLengthProvider;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public ServerName getServerWALsBelongTo() {
|
||||||
|
return null;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -18,6 +18,8 @@
|
||||||
*/
|
*/
|
||||||
package org.apache.hadoop.hbase.replication;
|
package org.apache.hadoop.hbase.replication;
|
||||||
|
|
||||||
|
import java.util.Arrays;
|
||||||
|
import java.util.List;
|
||||||
import java.util.NavigableMap;
|
import java.util.NavigableMap;
|
||||||
import java.util.TreeMap;
|
import java.util.TreeMap;
|
||||||
|
|
||||||
|
@ -36,10 +38,13 @@ import org.apache.hadoop.hbase.client.ConnectionFactory;
|
||||||
import org.apache.hadoop.hbase.client.Table;
|
import org.apache.hadoop.hbase.client.Table;
|
||||||
import org.apache.hadoop.hbase.client.replication.ReplicationAdmin;
|
import org.apache.hadoop.hbase.client.replication.ReplicationAdmin;
|
||||||
import org.apache.hadoop.hbase.util.Bytes;
|
import org.apache.hadoop.hbase.util.Bytes;
|
||||||
|
import org.apache.hadoop.hbase.wal.AbstractFSWALProvider;
|
||||||
import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster;
|
import org.apache.hadoop.hbase.zookeeper.MiniZooKeeperCluster;
|
||||||
import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
|
import org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher;
|
||||||
import org.junit.AfterClass;
|
import org.junit.AfterClass;
|
||||||
import org.junit.BeforeClass;
|
import org.junit.BeforeClass;
|
||||||
|
import org.junit.runners.Parameterized.Parameter;
|
||||||
|
import org.junit.runners.Parameterized.Parameters;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* This class is only a base for other integration-level replication tests.
|
* This class is only a base for other integration-level replication tests.
|
||||||
|
@ -82,6 +87,14 @@ public class TestReplicationBase {
|
||||||
protected static final byte[] row = Bytes.toBytes("row");
|
protected static final byte[] row = Bytes.toBytes("row");
|
||||||
protected static final byte[] noRepfamName = Bytes.toBytes("norep");
|
protected static final byte[] noRepfamName = Bytes.toBytes("norep");
|
||||||
|
|
||||||
|
@Parameter
|
||||||
|
public static boolean seperateOldWALs;
|
||||||
|
|
||||||
|
@Parameters
|
||||||
|
public static List<Boolean> params() {
|
||||||
|
return Arrays.asList(false, true);
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @throws java.lang.Exception
|
* @throws java.lang.Exception
|
||||||
*/
|
*/
|
||||||
|
@ -106,6 +119,9 @@ public class TestReplicationBase {
|
||||||
conf1.setFloat("replication.source.ratio", 1.0f);
|
conf1.setFloat("replication.source.ratio", 1.0f);
|
||||||
conf1.setBoolean("replication.source.eof.autorecovery", true);
|
conf1.setBoolean("replication.source.eof.autorecovery", true);
|
||||||
|
|
||||||
|
// Parameter config
|
||||||
|
conf1.setBoolean(AbstractFSWALProvider.SEPARATE_OLDLOGDIR, seperateOldWALs);
|
||||||
|
|
||||||
utility1 = new HBaseTestingUtility(conf1);
|
utility1 = new HBaseTestingUtility(conf1);
|
||||||
utility1.startMiniZKCluster();
|
utility1.startMiniZKCluster();
|
||||||
MiniZooKeeperCluster miniZK = utility1.getZkCluster();
|
MiniZooKeeperCluster miniZK = utility1.getZkCluster();
|
||||||
|
|
|
@ -21,11 +21,14 @@ import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.testclassification.ReplicationTests;
|
import org.apache.hadoop.hbase.testclassification.ReplicationTests;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.junit.experimental.categories.Category;
|
import org.junit.experimental.categories.Category;
|
||||||
|
import org.junit.runner.RunWith;
|
||||||
|
import org.junit.runners.Parameterized;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Runs the TestReplicationKillRS test and selects the RS to kill in the master cluster
|
* Runs the TestReplicationKillRS test and selects the RS to kill in the master cluster
|
||||||
* Do not add other tests in this class.
|
* Do not add other tests in this class.
|
||||||
*/
|
*/
|
||||||
|
@RunWith(Parameterized.class)
|
||||||
@Category({ReplicationTests.class, LargeTests.class})
|
@Category({ReplicationTests.class, LargeTests.class})
|
||||||
public class TestReplicationKillMasterRS extends TestReplicationKillRS {
|
public class TestReplicationKillMasterRS extends TestReplicationKillRS {
|
||||||
|
|
||||||
|
|
|
@ -21,11 +21,14 @@ import org.apache.hadoop.hbase.testclassification.LargeTests;
|
||||||
import org.apache.hadoop.hbase.testclassification.ReplicationTests;
|
import org.apache.hadoop.hbase.testclassification.ReplicationTests;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.junit.experimental.categories.Category;
|
import org.junit.experimental.categories.Category;
|
||||||
|
import org.junit.runner.RunWith;
|
||||||
|
import org.junit.runners.Parameterized;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Runs the TestReplicationKillRS test and selects the RS to kill in the slave cluster
|
* Runs the TestReplicationKillRS test and selects the RS to kill in the slave cluster
|
||||||
* Do not add other tests in this class.
|
* Do not add other tests in this class.
|
||||||
*/
|
*/
|
||||||
|
@RunWith(Parameterized.class)
|
||||||
@Category({ReplicationTests.class, LargeTests.class})
|
@Category({ReplicationTests.class, LargeTests.class})
|
||||||
public class TestReplicationKillSlaveRS extends TestReplicationKillRS {
|
public class TestReplicationKillSlaveRS extends TestReplicationKillRS {
|
||||||
|
|
||||||
|
|
|
@ -645,7 +645,7 @@ public abstract class TestReplicationSourceManager {
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
public void init(Configuration conf, FileSystem fs, ReplicationSourceManager manager,
|
||||||
ReplicationQueues rq, ReplicationPeers rp, Stoppable stopper, String peerClusterId,
|
ReplicationQueues rq, ReplicationPeers rp, Server server, String peerClusterId,
|
||||||
UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
UUID clusterId, ReplicationEndpoint replicationEndpoint,
|
||||||
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException {
|
WALFileLengthProvider walFileLengthProvider, MetricsSource metrics) throws IOException {
|
||||||
throw new IOException("Failing deliberately");
|
throw new IOException("Failing deliberately");
|
||||||
|
|
|
@ -22,6 +22,7 @@ import org.apache.hadoop.hbase.HBaseConfiguration;
|
||||||
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
import org.apache.hadoop.hbase.HBaseTestingUtility;
|
||||||
import org.apache.hadoop.hbase.HConstants;
|
import org.apache.hadoop.hbase.HConstants;
|
||||||
import org.apache.hadoop.hbase.Server;
|
import org.apache.hadoop.hbase.Server;
|
||||||
|
import org.apache.hadoop.hbase.ServerName;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationFactory;
|
import org.apache.hadoop.hbase.replication.ReplicationFactory;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationQueueInfo;
|
import org.apache.hadoop.hbase.replication.ReplicationQueueInfo;
|
||||||
import org.apache.hadoop.hbase.replication.ReplicationQueues;
|
import org.apache.hadoop.hbase.replication.ReplicationQueues;
|
||||||
|
@ -105,11 +106,11 @@ public class TestReplicationSourceManagerZkImpl extends TestReplicationSourceMan
|
||||||
String queue3 = rq3.claimQueue(serverName, unclaimed.get(0)).getFirst();
|
String queue3 = rq3.claimQueue(serverName, unclaimed.get(0)).getFirst();
|
||||||
rq3.removeReplicatorIfQueueIsEmpty(unclaimed.get(0));
|
rq3.removeReplicatorIfQueueIsEmpty(unclaimed.get(0));
|
||||||
ReplicationQueueInfo replicationQueueInfo = new ReplicationQueueInfo(queue3);
|
ReplicationQueueInfo replicationQueueInfo = new ReplicationQueueInfo(queue3);
|
||||||
List<String> result = replicationQueueInfo.getDeadRegionServers();
|
List<ServerName> result = replicationQueueInfo.getDeadRegionServers();
|
||||||
// verify
|
// verify
|
||||||
assertTrue(result.contains(server.getServerName().getServerName()));
|
assertTrue(result.contains(server.getServerName()));
|
||||||
assertTrue(result.contains(s1.getServerName().getServerName()));
|
assertTrue(result.contains(s1.getServerName()));
|
||||||
assertTrue(result.contains(s2.getServerName().getServerName()));
|
assertTrue(result.contains(s2.getServerName()));
|
||||||
|
|
||||||
server.stop("");
|
server.stop("");
|
||||||
}
|
}
|
||||||
|
|
|
@ -148,7 +148,7 @@ public class TestWALEntryStream {
|
||||||
log.rollWriter();
|
log.rollWriter();
|
||||||
|
|
||||||
try (WALEntryStream entryStream =
|
try (WALEntryStream entryStream =
|
||||||
new WALEntryStream(walQueue, fs, conf, 0, log, new MetricsSource("1"))) {
|
new WALEntryStream(walQueue, fs, conf, 0, log, null, new MetricsSource("1"))) {
|
||||||
int i = 0;
|
int i = 0;
|
||||||
while (entryStream.hasNext()) {
|
while (entryStream.hasNext()) {
|
||||||
assertNotNull(entryStream.next());
|
assertNotNull(entryStream.next());
|
||||||
|
@ -175,7 +175,7 @@ public class TestWALEntryStream {
|
||||||
appendToLog();
|
appendToLog();
|
||||||
long oldPos;
|
long oldPos;
|
||||||
try (WALEntryStream entryStream =
|
try (WALEntryStream entryStream =
|
||||||
new WALEntryStream(walQueue, fs, conf, 0, log, new MetricsSource("1"))) {
|
new WALEntryStream(walQueue, fs, conf, 0, log, null, new MetricsSource("1"))) {
|
||||||
// There's one edit in the log, read it. Reading past it needs to throw exception
|
// There's one edit in the log, read it. Reading past it needs to throw exception
|
||||||
assertTrue(entryStream.hasNext());
|
assertTrue(entryStream.hasNext());
|
||||||
WAL.Entry entry = entryStream.next();
|
WAL.Entry entry = entryStream.next();
|
||||||
|
@ -193,7 +193,7 @@ public class TestWALEntryStream {
|
||||||
appendToLog();
|
appendToLog();
|
||||||
|
|
||||||
try (WALEntryStream entryStream = new WALEntryStream(walQueue, fs, conf, oldPos,
|
try (WALEntryStream entryStream = new WALEntryStream(walQueue, fs, conf, oldPos,
|
||||||
log, new MetricsSource("1"))) {
|
log, null, new MetricsSource("1"))) {
|
||||||
// Read the newly added entry, make sure we made progress
|
// Read the newly added entry, make sure we made progress
|
||||||
WAL.Entry entry = entryStream.next();
|
WAL.Entry entry = entryStream.next();
|
||||||
assertNotEquals(oldPos, entryStream.getPosition());
|
assertNotEquals(oldPos, entryStream.getPosition());
|
||||||
|
@ -207,7 +207,7 @@ public class TestWALEntryStream {
|
||||||
appendToLog();
|
appendToLog();
|
||||||
|
|
||||||
try (WALEntryStream entryStream = new WALEntryStream(walQueue, fs, conf, oldPos,
|
try (WALEntryStream entryStream = new WALEntryStream(walQueue, fs, conf, oldPos,
|
||||||
log, new MetricsSource("1"))) {
|
log, null, new MetricsSource("1"))) {
|
||||||
WAL.Entry entry = entryStream.next();
|
WAL.Entry entry = entryStream.next();
|
||||||
assertNotEquals(oldPos, entryStream.getPosition());
|
assertNotEquals(oldPos, entryStream.getPosition());
|
||||||
assertNotNull(entry);
|
assertNotNull(entry);
|
||||||
|
@ -232,7 +232,7 @@ public class TestWALEntryStream {
|
||||||
appendToLog("1");
|
appendToLog("1");
|
||||||
appendToLog("2");// 2
|
appendToLog("2");// 2
|
||||||
try (WALEntryStream entryStream =
|
try (WALEntryStream entryStream =
|
||||||
new WALEntryStream(walQueue, fs, conf, 0, log, new MetricsSource("1"))) {
|
new WALEntryStream(walQueue, fs, conf, 0, log, null, new MetricsSource("1"))) {
|
||||||
assertEquals("1", getRow(entryStream.next()));
|
assertEquals("1", getRow(entryStream.next()));
|
||||||
|
|
||||||
appendToLog("3"); // 3 - comes in after reader opened
|
appendToLog("3"); // 3 - comes in after reader opened
|
||||||
|
@ -257,7 +257,7 @@ public class TestWALEntryStream {
|
||||||
public void testNewEntriesWhileStreaming() throws Exception {
|
public void testNewEntriesWhileStreaming() throws Exception {
|
||||||
appendToLog("1");
|
appendToLog("1");
|
||||||
try (WALEntryStream entryStream =
|
try (WALEntryStream entryStream =
|
||||||
new WALEntryStream(walQueue, fs, conf, 0, log, new MetricsSource("1"))) {
|
new WALEntryStream(walQueue, fs, conf, 0, log, null, new MetricsSource("1"))) {
|
||||||
entryStream.next(); // we've hit the end of the stream at this point
|
entryStream.next(); // we've hit the end of the stream at this point
|
||||||
|
|
||||||
// some new entries come in while we're streaming
|
// some new entries come in while we're streaming
|
||||||
|
@ -280,7 +280,7 @@ public class TestWALEntryStream {
|
||||||
long lastPosition = 0;
|
long lastPosition = 0;
|
||||||
appendToLog("1");
|
appendToLog("1");
|
||||||
try (WALEntryStream entryStream =
|
try (WALEntryStream entryStream =
|
||||||
new WALEntryStream(walQueue, fs, conf, 0, log, new MetricsSource("1"))) {
|
new WALEntryStream(walQueue, fs, conf, 0, log, null, new MetricsSource("1"))) {
|
||||||
entryStream.next(); // we've hit the end of the stream at this point
|
entryStream.next(); // we've hit the end of the stream at this point
|
||||||
appendToLog("2");
|
appendToLog("2");
|
||||||
appendToLog("3");
|
appendToLog("3");
|
||||||
|
@ -288,7 +288,7 @@ public class TestWALEntryStream {
|
||||||
}
|
}
|
||||||
// next stream should picks up where we left off
|
// next stream should picks up where we left off
|
||||||
try (WALEntryStream entryStream =
|
try (WALEntryStream entryStream =
|
||||||
new WALEntryStream(walQueue, fs, conf, lastPosition, log, new MetricsSource("1"))) {
|
new WALEntryStream(walQueue, fs, conf, lastPosition, log, null, new MetricsSource("1"))) {
|
||||||
assertEquals("2", getRow(entryStream.next()));
|
assertEquals("2", getRow(entryStream.next()));
|
||||||
assertEquals("3", getRow(entryStream.next()));
|
assertEquals("3", getRow(entryStream.next()));
|
||||||
assertFalse(entryStream.hasNext()); // done
|
assertFalse(entryStream.hasNext()); // done
|
||||||
|
@ -306,13 +306,13 @@ public class TestWALEntryStream {
|
||||||
appendEntriesToLog(3);
|
appendEntriesToLog(3);
|
||||||
// read only one element
|
// read only one element
|
||||||
try (WALEntryStream entryStream = new WALEntryStream(walQueue, fs, conf, lastPosition,
|
try (WALEntryStream entryStream = new WALEntryStream(walQueue, fs, conf, lastPosition,
|
||||||
log, new MetricsSource("1"))) {
|
log, null, new MetricsSource("1"))) {
|
||||||
entryStream.next();
|
entryStream.next();
|
||||||
lastPosition = entryStream.getPosition();
|
lastPosition = entryStream.getPosition();
|
||||||
}
|
}
|
||||||
// there should still be two more entries from where we left off
|
// there should still be two more entries from where we left off
|
||||||
try (WALEntryStream entryStream =
|
try (WALEntryStream entryStream =
|
||||||
new WALEntryStream(walQueue, fs, conf, lastPosition, log, new MetricsSource("1"))) {
|
new WALEntryStream(walQueue, fs, conf, lastPosition, log, null, new MetricsSource("1"))) {
|
||||||
assertNotNull(entryStream.next());
|
assertNotNull(entryStream.next());
|
||||||
assertNotNull(entryStream.next());
|
assertNotNull(entryStream.next());
|
||||||
assertFalse(entryStream.hasNext());
|
assertFalse(entryStream.hasNext());
|
||||||
|
@ -323,7 +323,7 @@ public class TestWALEntryStream {
|
||||||
@Test
|
@Test
|
||||||
public void testEmptyStream() throws Exception {
|
public void testEmptyStream() throws Exception {
|
||||||
try (WALEntryStream entryStream =
|
try (WALEntryStream entryStream =
|
||||||
new WALEntryStream(walQueue, fs, conf, 0, log, new MetricsSource("1"))) {
|
new WALEntryStream(walQueue, fs, conf, 0, log, null, new MetricsSource("1"))) {
|
||||||
assertFalse(entryStream.hasNext());
|
assertFalse(entryStream.hasNext());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -334,7 +334,7 @@ public class TestWALEntryStream {
|
||||||
// get ending position
|
// get ending position
|
||||||
long position;
|
long position;
|
||||||
try (WALEntryStream entryStream =
|
try (WALEntryStream entryStream =
|
||||||
new WALEntryStream(walQueue, fs, conf, 0, log, new MetricsSource("1"))) {
|
new WALEntryStream(walQueue, fs, conf, 0, log, null, new MetricsSource("1"))) {
|
||||||
entryStream.next();
|
entryStream.next();
|
||||||
entryStream.next();
|
entryStream.next();
|
||||||
entryStream.next();
|
entryStream.next();
|
||||||
|
@ -440,7 +440,7 @@ public class TestWALEntryStream {
|
||||||
long size = log.getLogFileSizeIfBeingWritten(walQueue.peek()).getAsLong();
|
long size = log.getLogFileSizeIfBeingWritten(walQueue.peek()).getAsLong();
|
||||||
AtomicLong fileLength = new AtomicLong(size - 1);
|
AtomicLong fileLength = new AtomicLong(size - 1);
|
||||||
try (WALEntryStream entryStream = new WALEntryStream(walQueue, fs, conf, 0,
|
try (WALEntryStream entryStream = new WALEntryStream(walQueue, fs, conf, 0,
|
||||||
p -> OptionalLong.of(fileLength.get()), new MetricsSource("1"))) {
|
p -> OptionalLong.of(fileLength.get()), null, new MetricsSource("1"))) {
|
||||||
assertTrue(entryStream.hasNext());
|
assertTrue(entryStream.hasNext());
|
||||||
assertNotNull(entryStream.next());
|
assertNotNull(entryStream.next());
|
||||||
// can not get log 2
|
// can not get log 2
|
||||||
|
|
Loading…
Reference in New Issue