mirror of https://github.com/apache/lucene.git
cleanup stale TODOs; fix precommit
This commit is contained in:
parent
5640470593
commit
027bc0e4d6
|
@ -4738,7 +4738,9 @@ public class IndexWriter implements Closeable, TwoPhaseCommit, Accountable {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @lucene.internal */
|
/** Record that the files referenced by this {@link SegmentInfos} are still in use.
|
||||||
|
*
|
||||||
|
* @lucene.internal */
|
||||||
public synchronized void incRefDeleter(SegmentInfos segmentInfos) throws IOException {
|
public synchronized void incRefDeleter(SegmentInfos segmentInfos) throws IOException {
|
||||||
ensureOpen();
|
ensureOpen();
|
||||||
deleter.incRef(segmentInfos, false);
|
deleter.incRef(segmentInfos, false);
|
||||||
|
@ -4747,7 +4749,10 @@ public class IndexWriter implements Closeable, TwoPhaseCommit, Accountable {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @lucene.internal */
|
/** Record that the files referenced by this {@link SegmentInfos} are no longer in use. Only call this if you are sure you previously
|
||||||
|
* called {@link #incRefDeleter}.
|
||||||
|
*
|
||||||
|
* @lucene.internal */
|
||||||
public synchronized void decRefDeleter(SegmentInfos segmentInfos) throws IOException {
|
public synchronized void decRefDeleter(SegmentInfos segmentInfos) throws IOException {
|
||||||
ensureOpen();
|
ensureOpen();
|
||||||
deleter.decRef(segmentInfos);
|
deleter.decRef(segmentInfos);
|
||||||
|
|
|
@ -285,6 +285,7 @@ public final class SegmentInfos implements Cloneable, Iterable<SegmentCommitInfo
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/** Read the commit from the provided {@link ChecksumIndexInput}. */
|
||||||
public static final SegmentInfos readCommit(Directory directory, ChecksumIndexInput input, long generation) throws IOException {
|
public static final SegmentInfos readCommit(Directory directory, ChecksumIndexInput input, long generation) throws IOException {
|
||||||
|
|
||||||
// NOTE: as long as we want to throw indexformattooold (vs corruptindexexception), we need
|
// NOTE: as long as we want to throw indexformattooold (vs corruptindexexception), we need
|
||||||
|
@ -479,6 +480,7 @@ public final class SegmentInfos implements Cloneable, Iterable<SegmentCommitInfo
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/** Write ourselves to the provided {@link IndexOutput} */
|
||||||
public void write(Directory directory, IndexOutput out) throws IOException {
|
public void write(Directory directory, IndexOutput out) throws IOException {
|
||||||
CodecUtil.writeIndexHeader(out, "segments", VERSION_CURRENT,
|
CodecUtil.writeIndexHeader(out, "segments", VERSION_CURRENT,
|
||||||
StringHelper.randomId(), Long.toString(generation, Character.MAX_RADIX));
|
StringHelper.randomId(), Long.toString(generation, Character.MAX_RADIX));
|
||||||
|
@ -725,8 +727,11 @@ public final class SegmentInfos implements Cloneable, Iterable<SegmentCommitInfo
|
||||||
this.counter = other.counter;
|
this.counter = other.counter;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/** Set the generation to be used for the next commit */
|
||||||
public void setNextWriteGeneration(long generation) {
|
public void setNextWriteGeneration(long generation) {
|
||||||
assert generation >= this.generation;
|
if (generation < this.generation) {
|
||||||
|
throw new IllegalStateException("cannot decrease generation to " + generation + " from current generation " + this.generation);
|
||||||
|
}
|
||||||
this.generation = generation;
|
this.generation = generation;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -843,6 +848,7 @@ public final class SegmentInfos implements Cloneable, Iterable<SegmentCommitInfo
|
||||||
return userData;
|
return userData;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/** Sets the commit data. */
|
||||||
public void setUserData(Map<String,String> data, boolean doIncrementVersion) {
|
public void setUserData(Map<String,String> data, boolean doIncrementVersion) {
|
||||||
if (data == null) {
|
if (data == null) {
|
||||||
userData = Collections.<String,String>emptyMap();
|
userData = Collections.<String,String>emptyMap();
|
||||||
|
|
|
@ -30,6 +30,7 @@ import org.apache.lucene.store.Directory;
|
||||||
import org.apache.lucene.store.IOContext;
|
import org.apache.lucene.store.IOContext;
|
||||||
import org.apache.lucene.util.IOUtils;
|
import org.apache.lucene.util.IOUtils;
|
||||||
|
|
||||||
|
/** Default implementation of {@link DirectoryReader}. */
|
||||||
public final class StandardDirectoryReader extends DirectoryReader {
|
public final class StandardDirectoryReader extends DirectoryReader {
|
||||||
|
|
||||||
final IndexWriter writer;
|
final IndexWriter writer;
|
||||||
|
@ -336,7 +337,9 @@ public final class StandardDirectoryReader extends DirectoryReader {
|
||||||
return segmentInfos.getVersion();
|
return segmentInfos.getVersion();
|
||||||
}
|
}
|
||||||
|
|
||||||
/** @lucene.internal */
|
/** Return the {@link SegmentInfos} for this reader.
|
||||||
|
*
|
||||||
|
* @lucene.internal */
|
||||||
public SegmentInfos getSegmentInfos() {
|
public SegmentInfos getSegmentInfos() {
|
||||||
return segmentInfos;
|
return segmentInfos;
|
||||||
}
|
}
|
||||||
|
|
|
@ -54,7 +54,6 @@ import org.junit.Test;
|
||||||
|
|
||||||
import com.carrotsearch.randomizedtesting.generators.RandomPicks;
|
import com.carrotsearch.randomizedtesting.generators.RandomPicks;
|
||||||
|
|
||||||
|
|
||||||
@SuppressWarnings("resource")
|
@SuppressWarnings("resource")
|
||||||
public class TestNumericDocValuesUpdates extends LuceneTestCase {
|
public class TestNumericDocValuesUpdates extends LuceneTestCase {
|
||||||
|
|
||||||
|
|
|
@ -29,7 +29,7 @@ import org.apache.lucene.util.IOUtils;
|
||||||
|
|
||||||
/** Copies one file from an incoming DataInput to a dest filename in a local Directory */
|
/** Copies one file from an incoming DataInput to a dest filename in a local Directory */
|
||||||
|
|
||||||
class CopyOneFile implements Closeable {
|
public class CopyOneFile implements Closeable {
|
||||||
private final DataInput in;
|
private final DataInput in;
|
||||||
private final IndexOutput out;
|
private final IndexOutput out;
|
||||||
private final ReplicaNode dest;
|
private final ReplicaNode dest;
|
||||||
|
|
|
@ -21,7 +21,7 @@ package org.apache.lucene.replicator.nrt;
|
||||||
*
|
*
|
||||||
* @lucene.experimental */
|
* @lucene.experimental */
|
||||||
|
|
||||||
class FileMetaData {
|
public class FileMetaData {
|
||||||
|
|
||||||
// Header and footer of the file must be identical between primary and replica to consider the files equal:
|
// Header and footer of the file must be identical between primary and replica to consider the files equal:
|
||||||
public final byte[] header;
|
public final byte[] header;
|
||||||
|
|
|
@ -21,6 +21,7 @@ import java.io.Closeable;
|
||||||
import java.io.EOFException;
|
import java.io.EOFException;
|
||||||
import java.io.FileNotFoundException;
|
import java.io.FileNotFoundException;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
|
import java.io.PrintStream;
|
||||||
import java.nio.file.NoSuchFileException;
|
import java.nio.file.NoSuchFileException;
|
||||||
import java.nio.file.Path;
|
import java.nio.file.Path;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
|
@ -78,6 +79,9 @@ abstract class Node implements Closeable {
|
||||||
/** When this node was started */
|
/** When this node was started */
|
||||||
public static final long localStartNS = System.nanoTime();
|
public static final long localStartNS = System.nanoTime();
|
||||||
|
|
||||||
|
/** For debug logging */
|
||||||
|
protected final PrintStream printStream;
|
||||||
|
|
||||||
// public static final long globalStartNS;
|
// public static final long globalStartNS;
|
||||||
|
|
||||||
// For debugging:
|
// For debugging:
|
||||||
|
@ -86,10 +90,11 @@ abstract class Node implements Closeable {
|
||||||
/** File metadata for last sync that succeeded; we use this as a cache */
|
/** File metadata for last sync that succeeded; we use this as a cache */
|
||||||
protected volatile Map<String,FileMetaData> lastFileMetaData;
|
protected volatile Map<String,FileMetaData> lastFileMetaData;
|
||||||
|
|
||||||
public Node(int id, Directory dir, SearcherFactory searcherFactory) {
|
public Node(int id, Directory dir, SearcherFactory searcherFactory, PrintStream printStream) {
|
||||||
this.id = id;
|
this.id = id;
|
||||||
this.dir = dir;
|
this.dir = dir;
|
||||||
this.searcherFactory = searcherFactory;
|
this.searcherFactory = searcherFactory;
|
||||||
|
this.printStream = printStream;
|
||||||
}
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
|
@ -99,37 +104,41 @@ abstract class Node implements Closeable {
|
||||||
|
|
||||||
public abstract void commit() throws IOException;
|
public abstract void commit() throws IOException;
|
||||||
|
|
||||||
public static void nodeMessage(String message) {
|
public static void nodeMessage(PrintStream printStream, String message) {
|
||||||
long now = System.nanoTime();
|
if (printStream != null) {
|
||||||
System.out.println(String.format(Locale.ROOT,
|
long now = System.nanoTime();
|
||||||
"%5.3fs %5.1fs: [%11s] %s",
|
printStream.println(String.format(Locale.ROOT,
|
||||||
(now-globalStartNS)/1000000000.,
|
"%5.3fs %5.1fs: [%11s] %s",
|
||||||
(now-localStartNS)/1000000000.,
|
(now-globalStartNS)/1000000000.,
|
||||||
Thread.currentThread().getName(),
|
(now-localStartNS)/1000000000.,
|
||||||
message));
|
Thread.currentThread().getName(),
|
||||||
|
message));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public static void nodeMessage(int id, String message) {
|
public static void nodeMessage(PrintStream printStream, int id, String message) {
|
||||||
long now = System.nanoTime();
|
if (printStream != null) {
|
||||||
System.out.println(String.format(Locale.ROOT,
|
long now = System.nanoTime();
|
||||||
"%5.3fs %5.1fs: N%d [%11s] %s",
|
printStream.println(String.format(Locale.ROOT,
|
||||||
(now-globalStartNS)/1000000000.,
|
"%5.3fs %5.1fs: N%d [%11s] %s",
|
||||||
(now-localStartNS)/1000000000.,
|
(now-globalStartNS)/1000000000.,
|
||||||
id,
|
(now-localStartNS)/1000000000.,
|
||||||
Thread.currentThread().getName(),
|
id,
|
||||||
message));
|
Thread.currentThread().getName(),
|
||||||
|
message));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
protected void message(String message) {
|
protected void message(String message) {
|
||||||
long now = System.nanoTime();
|
if (printStream != null) {
|
||||||
System.out.println(String.format(Locale.ROOT,
|
long now = System.nanoTime();
|
||||||
"%5.3fs %5.1fs: %7s %2s [%11s] %s",
|
printStream.println(String.format(Locale.ROOT,
|
||||||
(now-globalStartNS)/1000000000.,
|
"%5.3fs %5.1fs: %7s %2s [%11s] %s",
|
||||||
(now-localStartNS)/1000000000.,
|
(now-globalStartNS)/1000000000.,
|
||||||
state, name(),
|
(now-localStartNS)/1000000000.,
|
||||||
Thread.currentThread().getName(), message));
|
state, name(),
|
||||||
|
Thread.currentThread().getName(), message));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
public String name() {
|
public String name() {
|
||||||
|
|
|
@ -18,6 +18,7 @@ package org.apache.lucene.replicator.nrt;
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
|
import java.io.PrintStream;
|
||||||
import java.util.Collections;
|
import java.util.Collections;
|
||||||
import java.util.HashMap;
|
import java.util.HashMap;
|
||||||
import java.util.HashSet;
|
import java.util.HashSet;
|
||||||
|
@ -67,8 +68,9 @@ public abstract class PrimaryNode extends Node {
|
||||||
|
|
||||||
private final AtomicInteger copyingCount = new AtomicInteger();
|
private final AtomicInteger copyingCount = new AtomicInteger();
|
||||||
|
|
||||||
public PrimaryNode(IndexWriter writer, int id, long primaryGen, long forcePrimaryVersion, SearcherFactory searcherFactory) throws IOException {
|
public PrimaryNode(IndexWriter writer, int id, long primaryGen, long forcePrimaryVersion,
|
||||||
super(id, writer.getDirectory(), searcherFactory);
|
SearcherFactory searcherFactory, PrintStream printStream) throws IOException {
|
||||||
|
super(id, writer.getDirectory(), searcherFactory, printStream);
|
||||||
message("top: now init primary");
|
message("top: now init primary");
|
||||||
this.writer = writer;
|
this.writer = writer;
|
||||||
this.primaryGen = primaryGen;
|
this.primaryGen = primaryGen;
|
||||||
|
@ -115,7 +117,7 @@ public abstract class PrimaryNode extends Node {
|
||||||
|
|
||||||
} catch (Throwable t) {
|
} catch (Throwable t) {
|
||||||
message("init: exception");
|
message("init: exception");
|
||||||
t.printStackTrace(System.out);
|
t.printStackTrace(printStream);
|
||||||
throw new RuntimeException(t);
|
throw new RuntimeException(t);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -18,6 +18,7 @@ package org.apache.lucene.replicator.nrt;
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
|
import java.io.PrintStream;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Arrays;
|
import java.util.Arrays;
|
||||||
import java.util.Collection;
|
import java.util.Collection;
|
||||||
|
@ -54,7 +55,7 @@ import org.apache.lucene.util.IOUtils;
|
||||||
*
|
*
|
||||||
* @lucene.experimental */
|
* @lucene.experimental */
|
||||||
|
|
||||||
abstract class ReplicaNode extends Node {
|
public abstract class ReplicaNode extends Node {
|
||||||
|
|
||||||
ReplicaFileDeleter deleter;
|
ReplicaFileDeleter deleter;
|
||||||
|
|
||||||
|
@ -79,8 +80,8 @@ abstract class ReplicaNode extends Node {
|
||||||
/** Primary gen last time we successfully replicated: */
|
/** Primary gen last time we successfully replicated: */
|
||||||
protected long lastPrimaryGen;
|
protected long lastPrimaryGen;
|
||||||
|
|
||||||
public ReplicaNode(int id, Directory dir, SearcherFactory searcherFactory) throws IOException {
|
public ReplicaNode(int id, Directory dir, SearcherFactory searcherFactory, PrintStream printStream) throws IOException {
|
||||||
super(id, dir, searcherFactory);
|
super(id, dir, searcherFactory, printStream);
|
||||||
|
|
||||||
if (dir instanceof FSDirectory && ((FSDirectory) dir).checkPendingDeletions()) {
|
if (dir instanceof FSDirectory && ((FSDirectory) dir).checkPendingDeletions()) {
|
||||||
throw new IllegalArgumentException("Directory " + dir + " still has pending deleted files; cannot initialize IndexWriter");
|
throw new IllegalArgumentException("Directory " + dir + " still has pending deleted files; cannot initialize IndexWriter");
|
||||||
|
@ -98,7 +99,7 @@ abstract class ReplicaNode extends Node {
|
||||||
deleter = new ReplicaFileDeleter(this, dir);
|
deleter = new ReplicaFileDeleter(this, dir);
|
||||||
} catch (Throwable t) {
|
} catch (Throwable t) {
|
||||||
message("exc on init:");
|
message("exc on init:");
|
||||||
t.printStackTrace(System.out);
|
t.printStackTrace(printStream);
|
||||||
throw t;
|
throw t;
|
||||||
} finally {
|
} finally {
|
||||||
if (success == false) {
|
if (success == false) {
|
||||||
|
@ -307,7 +308,7 @@ abstract class ReplicaNode extends Node {
|
||||||
} catch (Throwable t) {
|
} catch (Throwable t) {
|
||||||
if (t.getMessage().startsWith("replica cannot start") == false) {
|
if (t.getMessage().startsWith("replica cannot start") == false) {
|
||||||
message("exc on start:");
|
message("exc on start:");
|
||||||
t.printStackTrace(System.out);
|
t.printStackTrace(printStream);
|
||||||
} else {
|
} else {
|
||||||
dir.close();
|
dir.close();
|
||||||
}
|
}
|
||||||
|
@ -522,7 +523,7 @@ abstract class ReplicaNode extends Node {
|
||||||
} catch (NodeCommunicationException nce) {
|
} catch (NodeCommunicationException nce) {
|
||||||
// E.g. primary could crash/close when we are asking it for the copy state:
|
// E.g. primary could crash/close when we are asking it for the copy state:
|
||||||
message("top: ignoring communication exception creating CopyJob: " + nce);
|
message("top: ignoring communication exception creating CopyJob: " + nce);
|
||||||
//nce.printStackTrace(System.out);
|
//nce.printStackTrace(printStream);
|
||||||
if (state.equals("syncing")) {
|
if (state.equals("syncing")) {
|
||||||
state = "idle";
|
state = "idle";
|
||||||
}
|
}
|
||||||
|
@ -560,7 +561,7 @@ abstract class ReplicaNode extends Node {
|
||||||
} catch (NodeCommunicationException nce) {
|
} catch (NodeCommunicationException nce) {
|
||||||
// E.g. primary could crash/close when we are asking it for the copy state:
|
// E.g. primary could crash/close when we are asking it for the copy state:
|
||||||
message("top: ignoring exception starting CopyJob: " + nce);
|
message("top: ignoring exception starting CopyJob: " + nce);
|
||||||
nce.printStackTrace(System.out);
|
nce.printStackTrace(printStream);
|
||||||
if (state.equals("syncing")) {
|
if (state.equals("syncing")) {
|
||||||
state = "idle";
|
state = "idle";
|
||||||
}
|
}
|
||||||
|
|
|
@ -24,6 +24,7 @@ import java.util.concurrent.atomic.AtomicInteger;
|
||||||
|
|
||||||
import org.apache.lucene.index.DirectoryReader;
|
import org.apache.lucene.index.DirectoryReader;
|
||||||
import org.apache.lucene.index.IndexReader;
|
import org.apache.lucene.index.IndexReader;
|
||||||
|
import org.apache.lucene.index.IndexWriter;
|
||||||
import org.apache.lucene.index.LeafReader;
|
import org.apache.lucene.index.LeafReader;
|
||||||
import org.apache.lucene.index.LeafReaderContext;
|
import org.apache.lucene.index.LeafReaderContext;
|
||||||
import org.apache.lucene.index.SegmentInfos;
|
import org.apache.lucene.index.SegmentInfos;
|
||||||
|
|
|
@ -0,0 +1,21 @@
|
||||||
|
/*
|
||||||
|
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||||
|
* contributor license agreements. See the NOTICE file distributed with
|
||||||
|
* this work for additional information regarding copyright ownership.
|
||||||
|
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||||
|
* (the "License"); you may not use this file except in compliance with
|
||||||
|
* the License. You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
/**
|
||||||
|
* <h1>Near-real-time replication framework</h1>
|
||||||
|
*/
|
||||||
|
package org.apache.lucene.replicator.nrt;
|
|
@ -116,7 +116,7 @@ class SimplePrimaryNode extends PrimaryNode {
|
||||||
|
|
||||||
public SimplePrimaryNode(Random random, Path indexPath, int id, int tcpPort, long primaryGen, long forcePrimaryVersion, SearcherFactory searcherFactory,
|
public SimplePrimaryNode(Random random, Path indexPath, int id, int tcpPort, long primaryGen, long forcePrimaryVersion, SearcherFactory searcherFactory,
|
||||||
boolean doFlipBitsDuringCopy, boolean doCheckIndexOnClose) throws IOException {
|
boolean doFlipBitsDuringCopy, boolean doCheckIndexOnClose) throws IOException {
|
||||||
super(initWriter(id, random, indexPath, doCheckIndexOnClose), id, primaryGen, forcePrimaryVersion, searcherFactory);
|
super(initWriter(id, random, indexPath, doCheckIndexOnClose), id, primaryGen, forcePrimaryVersion, searcherFactory, System.out);
|
||||||
this.tcpPort = tcpPort;
|
this.tcpPort = tcpPort;
|
||||||
this.random = new Random(random.nextLong());
|
this.random = new Random(random.nextLong());
|
||||||
this.doFlipBitsDuringCopy = doFlipBitsDuringCopy;
|
this.doFlipBitsDuringCopy = doFlipBitsDuringCopy;
|
||||||
|
|
|
@ -64,8 +64,9 @@ class SimpleReplicaNode extends ReplicaNode {
|
||||||
/** Changes over time, as primary node crashes and moves around */
|
/** Changes over time, as primary node crashes and moves around */
|
||||||
int curPrimaryTCPPort;
|
int curPrimaryTCPPort;
|
||||||
|
|
||||||
public SimpleReplicaNode(Random random, int id, int tcpPort, Path indexPath, long curPrimaryGen, int primaryTCPPort, SearcherFactory searcherFactory, boolean doCheckIndexOnClose) throws IOException {
|
public SimpleReplicaNode(Random random, int id, int tcpPort, Path indexPath, long curPrimaryGen, int primaryTCPPort,
|
||||||
super(id, getDirectory(random, id, indexPath, doCheckIndexOnClose), searcherFactory);
|
SearcherFactory searcherFactory, boolean doCheckIndexOnClose) throws IOException {
|
||||||
|
super(id, getDirectory(random, id, indexPath, doCheckIndexOnClose), searcherFactory, System.out);
|
||||||
this.tcpPort = tcpPort;
|
this.tcpPort = tcpPort;
|
||||||
this.random = new Random(random.nextLong());
|
this.random = new Random(random.nextLong());
|
||||||
|
|
||||||
|
@ -139,7 +140,7 @@ class SimpleReplicaNode extends ReplicaNode {
|
||||||
|
|
||||||
// Corrupt any index files not referenced by current commit point; this is important (increases test evilness) because we may have done
|
// Corrupt any index files not referenced by current commit point; this is important (increases test evilness) because we may have done
|
||||||
// a hard crash of the previous JVM writing to this directory and so MDW's corrupt-unknown-files-on-close never ran:
|
// a hard crash of the previous JVM writing to this directory and so MDW's corrupt-unknown-files-on-close never ran:
|
||||||
Node.nodeMessage(id, "top: corrupt unknown files");
|
Node.nodeMessage(System.out, id, "top: corrupt unknown files");
|
||||||
dir.corruptUnknownFiles();
|
dir.corruptUnknownFiles();
|
||||||
|
|
||||||
return dir;
|
return dir;
|
||||||
|
|
|
@ -50,6 +50,7 @@ import org.apache.lucene.search.TermQuery;
|
||||||
import org.apache.lucene.store.MockDirectoryWrapper;
|
import org.apache.lucene.store.MockDirectoryWrapper;
|
||||||
import org.apache.lucene.util.IOUtils;
|
import org.apache.lucene.util.IOUtils;
|
||||||
import org.apache.lucene.util.LineFileDocs;
|
import org.apache.lucene.util.LineFileDocs;
|
||||||
|
import org.apache.lucene.util.LuceneTestCase.Nightly;
|
||||||
import org.apache.lucene.util.LuceneTestCase.SuppressCodecs;
|
import org.apache.lucene.util.LuceneTestCase.SuppressCodecs;
|
||||||
import org.apache.lucene.util.LuceneTestCase.SuppressSysoutChecks;
|
import org.apache.lucene.util.LuceneTestCase.SuppressSysoutChecks;
|
||||||
import org.apache.lucene.util.LuceneTestCase;
|
import org.apache.lucene.util.LuceneTestCase;
|
||||||
|
@ -63,29 +64,17 @@ import com.carrotsearch.randomizedtesting.SeedUtils;
|
||||||
TODO
|
TODO
|
||||||
- fangs
|
- fangs
|
||||||
- sometimes have one replica be really slow at copying / have random pauses (fake GC) / etc.
|
- sometimes have one replica be really slow at copying / have random pauses (fake GC) / etc.
|
||||||
- graceful primary close
|
|
||||||
- why do we do the "rename temp to actual" all at the end...? what really does that buy us?
|
|
||||||
- replica should also track maxSegmentName its seen, and tap into inflateGens if it's later promoted to primary?
|
|
||||||
- test should not print scary exceptions and then succeed!
|
- test should not print scary exceptions and then succeed!
|
||||||
- since all nodes are local, we could have a different test only impl that just does local file copies instead of via tcp...
|
|
||||||
- are the pre-copied-completed-merged files not being cleared in primary?
|
- are the pre-copied-completed-merged files not being cleared in primary?
|
||||||
- hmm the logic isn't right today? a replica may skip pulling a given copy state, that recorded the finished merged segments?
|
- hmm the logic isn't right today? a replica may skip pulling a given copy state, that recorded the finished merged segments?
|
||||||
- beast & fix bugs
|
|
||||||
- graceful cluster restart
|
|
||||||
- better translog integration
|
|
||||||
- get "graceful primary shutdown" working
|
|
||||||
- there is still some global state we rely on for "correctness", e.g. lastPrimaryVersion
|
|
||||||
- clean up how version is persisted in commit data
|
|
||||||
- why am i not using hashes here? how does ES use them?
|
|
||||||
- get all other "single shard" functions working too: this cluster should "act like" a single shard
|
|
||||||
- SLM
|
|
||||||
- controlled nrt reopen thread / returning long gen on write
|
|
||||||
- live field values
|
|
||||||
- add indexes
|
|
||||||
- make cluster level APIs to search, index, that deal w/ primary failover, etc.
|
|
||||||
- must prune xlog
|
|
||||||
- refuse to start primary unless we have quorum
|
|
||||||
- later
|
- later
|
||||||
|
- since all nodes are local, we could have a different test only impl that just does local file copies instead of via tcp...
|
||||||
|
- get all other "single shard" functions working too: this cluster should "act like" a single shard
|
||||||
|
- SLM
|
||||||
|
- controlled nrt reopen thread / returning long gen on write
|
||||||
|
- live field values
|
||||||
|
- add indexes
|
||||||
|
- replica should also track maxSegmentName its seen, and tap into inflateGens if it's later promoted to primary?
|
||||||
- if we named index files using segment's ID we wouldn't have file name conflicts after primary crash / rollback?
|
- if we named index files using segment's ID we wouldn't have file name conflicts after primary crash / rollback?
|
||||||
- back pressure on indexing if replicas can't keep up?
|
- back pressure on indexing if replicas can't keep up?
|
||||||
- get xlog working on top? needs to be checkpointed, so we can correlate IW ops to NRT reader version and prune xlog based on commit
|
- get xlog working on top? needs to be checkpointed, so we can correlate IW ops to NRT reader version and prune xlog based on commit
|
||||||
|
@ -190,6 +179,7 @@ public class TestStressNRTReplication extends LuceneTestCase {
|
||||||
|
|
||||||
final Set<Integer> crashingNodes = Collections.synchronizedSet(new HashSet<>());
|
final Set<Integer> crashingNodes = Collections.synchronizedSet(new HashSet<>());
|
||||||
|
|
||||||
|
@Nightly
|
||||||
public void test() throws Exception {
|
public void test() throws Exception {
|
||||||
|
|
||||||
Node.globalStartNS = System.nanoTime();
|
Node.globalStartNS = System.nanoTime();
|
||||||
|
|
Loading…
Reference in New Issue