mirror of https://github.com/apache/lucene.git
Reverse merged revision(s) from lucene/dev/trunk up to 1237523
git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/branches/lucene2858@1237524 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
commit
6d9a3c301e
|
@ -852,6 +852,12 @@ Bug fixes
|
||||||
TermAllGroupsCollector or TermAllGroupHeadsCollector if instantiated with a
|
TermAllGroupsCollector or TermAllGroupHeadsCollector if instantiated with a
|
||||||
non default small size. (Martijn van Groningen, yonik)
|
non default small size. (Martijn van Groningen, yonik)
|
||||||
|
|
||||||
|
* LUCENE-3727: When writing stored fields and term vectors, Lucene
|
||||||
|
checks file sizes to detect a bug in some Sun JREs (LUCENE-1282),
|
||||||
|
however, on some NFS filesystems File.length() could be stale,
|
||||||
|
resulting in false errors like "fdx size mismatch while indexing".
|
||||||
|
These checks now use getFilePointer instead to avoid this.
|
||||||
|
(Jamir Shaikh, Mike McCandless, Robert Muir)
|
||||||
|
|
||||||
Optimizations
|
Optimizations
|
||||||
|
|
||||||
|
|
|
@ -22,7 +22,7 @@ import java.io.IOException;
|
||||||
import org.apache.lucene.index.AtomicReader.AtomicReaderContext;
|
import org.apache.lucene.index.AtomicReader.AtomicReaderContext;
|
||||||
import org.apache.lucene.search.SearcherManager; // javadocs
|
import org.apache.lucene.search.SearcherManager; // javadocs
|
||||||
import org.apache.lucene.store.*;
|
import org.apache.lucene.store.*;
|
||||||
import org.apache.lucene.util.ReaderUtil; // for javadocs
|
import org.apache.lucene.util.ReaderUtil;
|
||||||
|
|
||||||
/** IndexReader is an abstract class, providing an interface for accessing an
|
/** IndexReader is an abstract class, providing an interface for accessing an
|
||||||
index. Search of an index is done entirely through this abstract interface,
|
index. Search of an index is done entirely through this abstract interface,
|
||||||
|
|
|
@ -78,7 +78,8 @@ public class Builder<T> {
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Instantiates an FST/FSA builder without any pruning. A shortcut
|
* Instantiates an FST/FSA builder without any pruning. A shortcut
|
||||||
* to {@link #Builder(FST.INPUT_TYPE, int, int, boolean, boolean, int, Outputs, FreezeTail)} with
|
* to {@link #Builder(FST.INPUT_TYPE, int, int, boolean,
|
||||||
|
* boolean, int, Outputs, FreezeTail, boolean)} with
|
||||||
* pruning options turned off.
|
* pruning options turned off.
|
||||||
*/
|
*/
|
||||||
public Builder(FST.INPUT_TYPE inputType, Outputs<T> outputs) {
|
public Builder(FST.INPUT_TYPE inputType, Outputs<T> outputs) {
|
||||||
|
@ -122,9 +123,10 @@ public class Builder<T> {
|
||||||
* FSA, use {@link NoOutputs#getSingleton()} and {@link NoOutputs#getNoOutput()} as the
|
* FSA, use {@link NoOutputs#getSingleton()} and {@link NoOutputs#getNoOutput()} as the
|
||||||
* singleton output object.
|
* singleton output object.
|
||||||
*
|
*
|
||||||
* @param willPackFST Pass true if you will rewrite (compact) the FST before saving. This
|
* @param willPackFST Pass true if you will pack the FST before saving. This
|
||||||
* causes the FST to create additional data structures intenrally to facilitate rewriting, but
|
* causes the FST to create additional data structures internally to facilitate packing, but
|
||||||
* it means the resulting FST cannot be saved: it must first be rewritten using {@link FST#FST(FST,int[])}}
|
* it means the resulting FST cannot be saved: it must
|
||||||
|
* first be packed using {@link FST#pack(int, int)}}.
|
||||||
*/
|
*/
|
||||||
public Builder(FST.INPUT_TYPE inputType, int minSuffixCount1, int minSuffixCount2, boolean doShareSuffix,
|
public Builder(FST.INPUT_TYPE inputType, int minSuffixCount1, int minSuffixCount2, boolean doShareSuffix,
|
||||||
boolean doShareNonSingletonNodes, int shareMaxTailLength, Outputs<T> outputs,
|
boolean doShareNonSingletonNodes, int shareMaxTailLength, Outputs<T> outputs,
|
||||||
|
|
|
@ -19,7 +19,6 @@ package org.apache.solr.cloud;
|
||||||
|
|
||||||
import java.net.BindException;
|
import java.net.BindException;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Collection;
|
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
import java.util.Map;
|
import java.util.Map;
|
||||||
import java.util.Random;
|
import java.util.Random;
|
||||||
|
@ -297,13 +296,7 @@ public class ChaosMonkey {
|
||||||
JettySolrRunner jetty;
|
JettySolrRunner jetty;
|
||||||
if (chance <= 5 && aggressivelyKillLeaders) {
|
if (chance <= 5 && aggressivelyKillLeaders) {
|
||||||
// if killLeader, really aggressively go after leaders
|
// if killLeader, really aggressively go after leaders
|
||||||
Collection<CloudJettyRunner> leaders = shardToLeaderJetty.values();
|
jetty = shardToLeaderJetty.get(slice).jetty;
|
||||||
List<CloudJettyRunner> leadersList = new ArrayList<CloudJettyRunner>(leaders.size());
|
|
||||||
|
|
||||||
leadersList.addAll(leaders);
|
|
||||||
|
|
||||||
int index = random.nextInt(leadersList.size());
|
|
||||||
jetty = leadersList.get(index).jetty;
|
|
||||||
} else {
|
} else {
|
||||||
// get random shard
|
// get random shard
|
||||||
List<CloudJettyRunner> jetties = shardToJetty.get(slice);
|
List<CloudJettyRunner> jetties = shardToJetty.get(slice);
|
||||||
|
|
Loading…
Reference in New Issue