mirror of https://github.com/apache/lucene.git
Revert "Merge remote-tracking branch 'origin/master'"
This reverts commit26298f35df
, reversing changes made to65c6c576b7
. undoing the merge commit
This commit is contained in:
parent
26298f35df
commit
325cbf0018
|
@ -107,11 +107,6 @@ New Features
|
|||
SortedNumericSelector.Type can give a ValueSource view of a
|
||||
SortedNumericDocValues field. (Tomás Fernández Löbbe)
|
||||
|
||||
* LUCENE-7465: Add SimplePatternTokenizer and
|
||||
SimplePatternSplitTokenizer, using Lucene's regexp/automaton
|
||||
implementation for analysis/tokenization (Clinton Gormley, Mike
|
||||
McCandless)
|
||||
|
||||
Bug Fixes
|
||||
|
||||
* LUCENE-7630: Fix (Edge)NGramTokenFilter to no longer drop payloads
|
||||
|
@ -176,11 +171,6 @@ Build
|
|||
Jars are not downloaded; compilation is not updated; and Clover is not enabled.
|
||||
(Steve Rowe)
|
||||
|
||||
* LUCENE-7694: Update forbiddenapis to version 2.3. (Uwe Schindler)
|
||||
|
||||
* LUCENE-7693: Replace "org.apache." logic in GetMavenDependenciesTask.
|
||||
(Daniel Collins, Christine Poerschke)
|
||||
|
||||
Other
|
||||
|
||||
* LUCENE-7666: Fix typos in lucene-join package info javadoc.
|
||||
|
|
|
@ -1,258 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.lucene.analysis.pattern;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import org.apache.lucene.analysis.Tokenizer;
|
||||
import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
|
||||
import org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.AttributeFactory;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.CharacterRunAutomaton;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
|
||||
/**
|
||||
* This tokenizer uses a Lucene {@link RegExp} or (expert usage) a pre-built determinized {@link Automaton}, to locate tokens.
|
||||
* The regexp syntax is more limited than {@link PatternTokenizer}, but the tokenization is quite a bit faster. This is just
|
||||
* like {@link SimplePatternTokenizer} except that the pattern shold make valid token separator characters, like
|
||||
* {@code String.split}. Empty string tokens are never produced.
|
||||
*
|
||||
* @lucene.experimental
|
||||
*/
|
||||
|
||||
public final class SimplePatternSplitTokenizer extends Tokenizer {
|
||||
|
||||
private final CharTermAttribute termAtt = addAttribute(CharTermAttribute.class);
|
||||
private final OffsetAttribute offsetAtt = addAttribute(OffsetAttribute.class);
|
||||
|
||||
private final CharacterRunAutomaton runDFA;
|
||||
|
||||
// TODO: this is copied from SimplePatternTokenizer, but there are subtle differences e.g. we track sepUpto an tokenUpto;
|
||||
// find a clean way to share it:
|
||||
|
||||
// TODO: we could likely use a single rolling buffer instead of two separate char buffers here. We could also use PushBackReader but I
|
||||
// suspect it's slowish:
|
||||
|
||||
private char[] pendingChars = new char[8];
|
||||
private int tokenUpto;
|
||||
private int pendingLimit;
|
||||
private int pendingUpto;
|
||||
private int offset;
|
||||
private int sepUpto;
|
||||
private final char[] buffer = new char[1024];
|
||||
private int bufferLimit;
|
||||
private int bufferNextRead;
|
||||
|
||||
/** See {@link RegExp} for the accepted syntax. */
|
||||
public SimplePatternSplitTokenizer(String regexp) {
|
||||
this(DEFAULT_TOKEN_ATTRIBUTE_FACTORY, regexp, Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
}
|
||||
|
||||
/** Runs a pre-built automaton. */
|
||||
public SimplePatternSplitTokenizer(Automaton dfa) {
|
||||
this(DEFAULT_TOKEN_ATTRIBUTE_FACTORY, dfa);
|
||||
}
|
||||
|
||||
/** See {@link RegExp} for the accepted syntax. */
|
||||
public SimplePatternSplitTokenizer(AttributeFactory factory, String regexp, int maxDeterminizedStates) {
|
||||
this(factory, new RegExp(regexp).toAutomaton());
|
||||
}
|
||||
|
||||
/** Runs a pre-built automaton. */
|
||||
public SimplePatternSplitTokenizer(AttributeFactory factory, Automaton dfa) {
|
||||
super(factory);
|
||||
|
||||
// we require user to do this up front because it is a possibly very costly operation, and user may be creating us frequently, not
|
||||
// realizing this ctor is otherwise trappy
|
||||
if (dfa.isDeterministic() == false) {
|
||||
throw new IllegalArgumentException("please determinize the incoming automaton first");
|
||||
}
|
||||
|
||||
runDFA = new CharacterRunAutomaton(dfa, Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
}
|
||||
|
||||
private void fillToken(int offsetStart) {
|
||||
termAtt.setLength(tokenUpto);
|
||||
offsetAtt.setOffset(correctOffset(offsetStart), correctOffset(offsetStart+tokenUpto));
|
||||
}
|
||||
|
||||
@Override
|
||||
public boolean incrementToken() throws IOException {
|
||||
|
||||
int offsetStart = offset;
|
||||
|
||||
clearAttributes();
|
||||
|
||||
tokenUpto = 0;
|
||||
|
||||
while (true) {
|
||||
sepUpto = 0;
|
||||
|
||||
// The runDFA operates in Unicode space, not UTF16 (java's char):
|
||||
int ch = nextCodePoint();
|
||||
if (ch == -1) {
|
||||
if (tokenUpto > 0) {
|
||||
fillToken(offsetStart);
|
||||
return true;
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
int state = runDFA.step(0, ch);
|
||||
|
||||
if (state != -1) {
|
||||
// a token separator just possibly started; keep scanning to see if the token is accepted:
|
||||
int lastAcceptLength = -1;
|
||||
do {
|
||||
|
||||
if (runDFA.isAccept(state)) {
|
||||
// record that the token separator matches here, but keep scanning in case a longer match also works (greedy):
|
||||
lastAcceptLength = sepUpto;
|
||||
}
|
||||
|
||||
ch = nextCodePoint();
|
||||
if (ch == -1) {
|
||||
break;
|
||||
}
|
||||
state = runDFA.step(state, ch);
|
||||
} while (state != -1);
|
||||
|
||||
if (lastAcceptLength != -1) {
|
||||
// strip the trailing separater we just matched from the token:
|
||||
tokenUpto -= lastAcceptLength;
|
||||
// we found a token separator
|
||||
int extra = sepUpto - lastAcceptLength;
|
||||
if (extra != 0) {
|
||||
pushBack(extra);
|
||||
}
|
||||
if (tokenUpto > 0) {
|
||||
fillToken(offsetStart);
|
||||
return true;
|
||||
} else {
|
||||
// we matched one token separator immediately after another
|
||||
offsetStart = offset;
|
||||
}
|
||||
} else if (ch == -1) {
|
||||
if (tokenUpto > 0) {
|
||||
fillToken(offsetStart);
|
||||
return true;
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
} else {
|
||||
// false alarm: there was no token separator here; push back all but the first character we scanned
|
||||
pushBack(sepUpto-1);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void end() throws IOException {
|
||||
super.end();
|
||||
final int ofs = correctOffset(offset + pendingLimit - pendingUpto);
|
||||
offsetAtt.setOffset(ofs, ofs);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void reset() throws IOException {
|
||||
super.reset();
|
||||
offset = 0;
|
||||
pendingUpto = 0;
|
||||
pendingLimit = 0;
|
||||
sepUpto = 0;
|
||||
bufferNextRead = 0;
|
||||
bufferLimit = 0;
|
||||
}
|
||||
|
||||
/** Pushes back the last {@code count} characters in current token's buffer. */
|
||||
private void pushBack(int count) {
|
||||
tokenUpto -= count;
|
||||
assert tokenUpto >= 0;
|
||||
if (pendingLimit == 0) {
|
||||
if (bufferNextRead >= count) {
|
||||
// optimize common case when the chars we are pushing back are still in the buffer
|
||||
bufferNextRead -= count;
|
||||
} else {
|
||||
if (count > pendingChars.length) {
|
||||
pendingChars = ArrayUtil.grow(pendingChars, count);
|
||||
}
|
||||
System.arraycopy(termAtt.buffer(), tokenUpto - count, pendingChars, 0, count);
|
||||
pendingLimit = count;
|
||||
}
|
||||
} else {
|
||||
// we are pushing back what is already in our pending buffer
|
||||
pendingUpto -= count;
|
||||
assert pendingUpto >= 0;
|
||||
}
|
||||
offset -= count;
|
||||
}
|
||||
|
||||
private void appendToToken(char ch) {
|
||||
char[] buffer = termAtt.buffer();
|
||||
if (tokenUpto == buffer.length) {
|
||||
buffer = termAtt.resizeBuffer(tokenUpto + 1);
|
||||
}
|
||||
buffer[tokenUpto++] = ch;
|
||||
sepUpto++;
|
||||
}
|
||||
|
||||
private int nextCodeUnit() throws IOException {
|
||||
int result;
|
||||
if (pendingUpto < pendingLimit) {
|
||||
result = pendingChars[pendingUpto++];
|
||||
if (pendingUpto == pendingLimit) {
|
||||
// We used up the pending buffer
|
||||
pendingUpto = 0;
|
||||
pendingLimit = 0;
|
||||
}
|
||||
appendToToken((char) result);
|
||||
offset++;
|
||||
} else if (bufferLimit == -1) {
|
||||
return -1;
|
||||
} else {
|
||||
assert bufferNextRead <= bufferLimit: "bufferNextRead=" + bufferNextRead + " bufferLimit=" + bufferLimit;
|
||||
if (bufferNextRead == bufferLimit) {
|
||||
bufferLimit = input.read(buffer, 0, buffer.length);
|
||||
if (bufferLimit == -1) {
|
||||
return -1;
|
||||
}
|
||||
bufferNextRead = 0;
|
||||
}
|
||||
result = buffer[bufferNextRead++];
|
||||
offset++;
|
||||
appendToToken((char) result);
|
||||
}
|
||||
return result;
|
||||
}
|
||||
|
||||
private int nextCodePoint() throws IOException {
|
||||
|
||||
int ch = nextCodeUnit();
|
||||
if (ch == -1) {
|
||||
return ch;
|
||||
}
|
||||
if (Character.isHighSurrogate((char) ch)) {
|
||||
return Character.toCodePoint((char) ch, (char) nextCodeUnit());
|
||||
} else {
|
||||
return ch;
|
||||
}
|
||||
}
|
||||
}
|
|
@ -1,76 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.lucene.analysis.pattern;
|
||||
|
||||
import java.util.Map;
|
||||
|
||||
import org.apache.lucene.analysis.util.TokenizerFactory;
|
||||
import org.apache.lucene.util.AttributeFactory;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
|
||||
/**
|
||||
* Factory for {@link SimplePatternSplitTokenizer}, for producing tokens by splitting according to the provided regexp.
|
||||
*
|
||||
* <p>This tokenizer uses Lucene {@link RegExp} pattern matching to construct distinct tokens
|
||||
* for the input stream. The syntax is more limited than {@link PatternTokenizer}, but the
|
||||
* tokenization is quite a bit faster. It takes two arguments:
|
||||
* <br>
|
||||
* <ul>
|
||||
* <li>"pattern" (required) is the regular expression, according to the syntax described at {@link RegExp}</li>
|
||||
* <li>"maxDeterminizedStates" (optional, default 10000) the limit on total state count for the determined automaton computed from the regexp</li>
|
||||
* </ul>
|
||||
* <p>
|
||||
* The pattern matches the characters that should split tokens, like {@code String.split}, and the
|
||||
* matching is greedy such that the longest token separator matching at a given point is matched. Empty
|
||||
* tokens are never created.
|
||||
*
|
||||
* <p>For example, to match tokens delimited by simple whitespace characters:
|
||||
*
|
||||
* <pre class="prettyprint">
|
||||
* <fieldType name="text_ptn" class="solr.TextField" positionIncrementGap="100">
|
||||
* <analyzer>
|
||||
* <tokenizer class="solr.SimplePatternSplitTokenizerFactory" pattern="[ \t\r\n]+"/>
|
||||
* </analyzer>
|
||||
* </fieldType></pre>
|
||||
*
|
||||
* @lucene.experimental
|
||||
*
|
||||
* @see SimplePatternSplitTokenizer
|
||||
*/
|
||||
public class SimplePatternSplitTokenizerFactory extends TokenizerFactory {
|
||||
public static final String PATTERN = "pattern";
|
||||
private final Automaton dfa;
|
||||
private final int maxDeterminizedStates;
|
||||
|
||||
/** Creates a new SimpleSplitPatternTokenizerFactory */
|
||||
public SimplePatternSplitTokenizerFactory(Map<String,String> args) {
|
||||
super(args);
|
||||
maxDeterminizedStates = getInt(args, "maxDeterminizedStates", Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
dfa = Operations.determinize(new RegExp(require(args, PATTERN)).toAutomaton(), maxDeterminizedStates);
|
||||
if (args.isEmpty() == false) {
|
||||
throw new IllegalArgumentException("Unknown parameters: " + args);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public SimplePatternSplitTokenizer create(final AttributeFactory factory) {
|
||||
return new SimplePatternSplitTokenizer(factory, dfa);
|
||||
}
|
||||
}
|
|
@ -1,242 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.lucene.analysis.pattern;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
import org.apache.lucene.analysis.Tokenizer;
|
||||
import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
|
||||
import org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.AttributeFactory;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.CharacterRunAutomaton;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
|
||||
/**
|
||||
* This tokenizer uses a Lucene {@link RegExp} or (expert usage) a pre-built determinized {@link Automaton}, to locate tokens.
|
||||
* The regexp syntax is more limited than {@link PatternTokenizer}, but the tokenization is quite a bit faster. The provided
|
||||
* regex should match valid token characters (not token separator characters, like {@code String.split}). The matching is greedy:
|
||||
* the longest match at a given start point will be the next token. Empty string tokens are never produced.
|
||||
*
|
||||
* @lucene.experimental
|
||||
*/
|
||||
|
||||
// TODO: the matcher here is naive and does have N^2 adversarial cases that are unlikely to arise in practice, e.g. if the pattern is
|
||||
// aaaaaaaaaab and the input is aaaaaaaaaaa, the work we do here is N^2 where N is the number of a's. This is because on failing to match
|
||||
// a token, we skip one character forward and try again. A better approach would be to compile something like this regexp
|
||||
// instead: .* | <pattern>, because that automaton would not "forget" all the as it had already seen, and would be a single pass
|
||||
// through the input. I think this is the same thing as Aho/Corasick's algorithm (http://en.wikipedia.org/wiki/Aho%E2%80%93Corasick_string_matching_algorithm).
|
||||
// But we cannot implement this (I think?) until/unless Lucene regexps support sub-group capture, so we could know
|
||||
// which specific characters the pattern matched. SynonymFilter has this same limitation.
|
||||
|
||||
public final class SimplePatternTokenizer extends Tokenizer {
|
||||
|
||||
private final CharTermAttribute termAtt = addAttribute(CharTermAttribute.class);
|
||||
private final OffsetAttribute offsetAtt = addAttribute(OffsetAttribute.class);
|
||||
|
||||
private final CharacterRunAutomaton runDFA;
|
||||
|
||||
// TODO: we could likely use a single rolling buffer instead of two separate char buffers here. We could also use PushBackReader but I
|
||||
// suspect it's slowish:
|
||||
|
||||
private char[] pendingChars = new char[8];
|
||||
private int pendingLimit;
|
||||
private int pendingUpto;
|
||||
private int offset;
|
||||
private int tokenUpto;
|
||||
private final char[] buffer = new char[1024];
|
||||
private int bufferLimit;
|
||||
private int bufferNextRead;
|
||||
|
||||
/** See {@link RegExp} for the accepted syntax. */
|
||||
public SimplePatternTokenizer(String regexp) {
|
||||
this(DEFAULT_TOKEN_ATTRIBUTE_FACTORY, regexp, Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
}
|
||||
|
||||
/** Runs a pre-built automaton. */
|
||||
public SimplePatternTokenizer(Automaton dfa) {
|
||||
this(DEFAULT_TOKEN_ATTRIBUTE_FACTORY, dfa);
|
||||
}
|
||||
|
||||
/** See {@link RegExp} for the accepted syntax. */
|
||||
public SimplePatternTokenizer(AttributeFactory factory, String regexp, int maxDeterminizedStates) {
|
||||
this(factory, new RegExp(regexp).toAutomaton());
|
||||
}
|
||||
|
||||
/** Runs a pre-built automaton. */
|
||||
public SimplePatternTokenizer(AttributeFactory factory, Automaton dfa) {
|
||||
super(factory);
|
||||
|
||||
// we require user to do this up front because it is a possibly very costly operation, and user may be creating us frequently, not
|
||||
// realizing this ctor is otherwise trappy
|
||||
if (dfa.isDeterministic() == false) {
|
||||
throw new IllegalArgumentException("please determinize the incoming automaton first");
|
||||
}
|
||||
|
||||
runDFA = new CharacterRunAutomaton(dfa, Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
}
|
||||
|
||||
@Override
|
||||
public boolean incrementToken() throws IOException {
|
||||
|
||||
clearAttributes();
|
||||
tokenUpto = 0;
|
||||
|
||||
while (true) {
|
||||
|
||||
int offsetStart = offset;
|
||||
|
||||
// The runDFA operates in Unicode space, not UTF16 (java's char):
|
||||
|
||||
int ch = nextCodePoint();
|
||||
if (ch == -1) {
|
||||
return false;
|
||||
}
|
||||
|
||||
int state = runDFA.step(0, ch);
|
||||
|
||||
if (state != -1) {
|
||||
// a token just possibly started; keep scanning to see if the token is accepted:
|
||||
int lastAcceptLength = -1;
|
||||
do {
|
||||
|
||||
if (runDFA.isAccept(state)) {
|
||||
// record that the token matches here, but keep scanning in case a longer match also works (greedy):
|
||||
lastAcceptLength = tokenUpto;
|
||||
}
|
||||
|
||||
ch = nextCodePoint();
|
||||
if (ch == -1) {
|
||||
break;
|
||||
}
|
||||
state = runDFA.step(state, ch);
|
||||
} while (state != -1);
|
||||
|
||||
if (lastAcceptLength != -1) {
|
||||
// we found a token
|
||||
int extra = tokenUpto - lastAcceptLength;
|
||||
if (extra != 0) {
|
||||
pushBack(extra);
|
||||
}
|
||||
termAtt.setLength(lastAcceptLength);
|
||||
offsetAtt.setOffset(correctOffset(offsetStart), correctOffset(offsetStart+lastAcceptLength));
|
||||
return true;
|
||||
} else if (ch == -1) {
|
||||
return false;
|
||||
} else {
|
||||
// false alarm: there was no token here; push back all but the first character we scanned
|
||||
pushBack(tokenUpto-1);
|
||||
tokenUpto = 0;
|
||||
}
|
||||
} else {
|
||||
tokenUpto = 0;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void end() throws IOException {
|
||||
super.end();
|
||||
final int ofs = correctOffset(offset + pendingLimit - pendingUpto);
|
||||
offsetAtt.setOffset(ofs, ofs);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void reset() throws IOException {
|
||||
super.reset();
|
||||
offset = 0;
|
||||
pendingUpto = 0;
|
||||
pendingLimit = 0;
|
||||
tokenUpto = 0;
|
||||
bufferNextRead = 0;
|
||||
bufferLimit = 0;
|
||||
}
|
||||
|
||||
/** Pushes back the last {@code count} characters in current token's buffer. */
|
||||
private void pushBack(int count) {
|
||||
|
||||
if (pendingLimit == 0) {
|
||||
if (bufferNextRead >= count) {
|
||||
// optimize common case when the chars we are pushing back are still in the buffer
|
||||
bufferNextRead -= count;
|
||||
} else {
|
||||
if (count > pendingChars.length) {
|
||||
pendingChars = ArrayUtil.grow(pendingChars, count);
|
||||
}
|
||||
System.arraycopy(termAtt.buffer(), tokenUpto - count, pendingChars, 0, count);
|
||||
pendingLimit = count;
|
||||
}
|
||||
} else {
|
||||
// we are pushing back what is already in our pending buffer
|
||||
pendingUpto -= count;
|
||||
assert pendingUpto >= 0;
|
||||
}
|
||||
offset -= count;
|
||||
}
|
||||
|
||||
private void appendToToken(char ch) {
|
||||
char[] buffer = termAtt.buffer();
|
||||
if (tokenUpto == buffer.length) {
|
||||
buffer = termAtt.resizeBuffer(tokenUpto + 1);
|
||||
}
|
||||
buffer[tokenUpto++] = ch;
|
||||
}
|
||||
|
||||
private int nextCodeUnit() throws IOException {
|
||||
int result;
|
||||
if (pendingUpto < pendingLimit) {
|
||||
result = pendingChars[pendingUpto++];
|
||||
if (pendingUpto == pendingLimit) {
|
||||
// We used up the pending buffer
|
||||
pendingUpto = 0;
|
||||
pendingLimit = 0;
|
||||
}
|
||||
appendToToken((char) result);
|
||||
offset++;
|
||||
} else if (bufferLimit == -1) {
|
||||
return -1;
|
||||
} else {
|
||||
assert bufferNextRead <= bufferLimit: "bufferNextRead=" + bufferNextRead + " bufferLimit=" + bufferLimit;
|
||||
if (bufferNextRead == bufferLimit) {
|
||||
bufferLimit = input.read(buffer, 0, buffer.length);
|
||||
if (bufferLimit == -1) {
|
||||
return -1;
|
||||
}
|
||||
bufferNextRead = 0;
|
||||
}
|
||||
result = buffer[bufferNextRead++];
|
||||
offset++;
|
||||
appendToToken((char) result);
|
||||
}
|
||||
return result;
|
||||
}
|
||||
|
||||
private int nextCodePoint() throws IOException {
|
||||
|
||||
int ch = nextCodeUnit();
|
||||
if (ch == -1) {
|
||||
return ch;
|
||||
}
|
||||
if (Character.isHighSurrogate((char) ch)) {
|
||||
return Character.toCodePoint((char) ch, (char) nextCodeUnit());
|
||||
} else {
|
||||
return ch;
|
||||
}
|
||||
}
|
||||
}
|
|
@ -1,76 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
|
||||
package org.apache.lucene.analysis.pattern;
|
||||
|
||||
import java.util.Map;
|
||||
|
||||
import org.apache.lucene.analysis.util.TokenizerFactory;
|
||||
import org.apache.lucene.util.AttributeFactory;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
|
||||
/**
|
||||
* Factory for {@link SimplePatternTokenizer}, for matching tokens based on the provided regexp.
|
||||
*
|
||||
* <p>This tokenizer uses Lucene {@link RegExp} pattern matching to construct distinct tokens
|
||||
* for the input stream. The syntax is more limited than {@link PatternTokenizer}, but the
|
||||
* tokenization is quite a bit faster. It takes two arguments:
|
||||
* <br>
|
||||
* <ul>
|
||||
* <li>"pattern" (required) is the regular expression, according to the syntax described at {@link RegExp}</li>
|
||||
* <li>"maxDeterminizedStates" (optional, default 10000) the limit on total state count for the determined automaton computed from the regexp</li>
|
||||
* </ul>
|
||||
* <p>
|
||||
* The pattern matches the characters to include in a token (not the split characters), and the
|
||||
* matching is greedy such that the longest token matching at a given point is created. Empty
|
||||
* tokens are never created.
|
||||
*
|
||||
* <p>For example, to match tokens delimited by simple whitespace characters:
|
||||
*
|
||||
* <pre class="prettyprint">
|
||||
* <fieldType name="text_ptn" class="solr.TextField" positionIncrementGap="100">
|
||||
* <analyzer>
|
||||
* <tokenizer class="solr.SimplePatternTokenizerFactory" pattern="[^ \t\r\n]+"/>
|
||||
* </analyzer>
|
||||
* </fieldType></pre>
|
||||
*
|
||||
* @lucene.experimental
|
||||
*
|
||||
* @see SimplePatternTokenizer
|
||||
*/
|
||||
public class SimplePatternTokenizerFactory extends TokenizerFactory {
|
||||
public static final String PATTERN = "pattern";
|
||||
private final Automaton dfa;
|
||||
private final int maxDeterminizedStates;
|
||||
|
||||
/** Creates a new SimplePatternTokenizerFactory */
|
||||
public SimplePatternTokenizerFactory(Map<String,String> args) {
|
||||
super(args);
|
||||
maxDeterminizedStates = getInt(args, "maxDeterminizedStates", Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
dfa = Operations.determinize(new RegExp(require(args, PATTERN)).toAutomaton(), maxDeterminizedStates);
|
||||
if (args.isEmpty() == false) {
|
||||
throw new IllegalArgumentException("Unknown parameters: " + args);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public SimplePatternTokenizer create(final AttributeFactory factory) {
|
||||
return new SimplePatternTokenizer(factory, dfa);
|
||||
}
|
||||
}
|
|
@ -21,8 +21,6 @@ org.apache.lucene.analysis.ngram.EdgeNGramTokenizerFactory
|
|||
org.apache.lucene.analysis.ngram.NGramTokenizerFactory
|
||||
org.apache.lucene.analysis.path.PathHierarchyTokenizerFactory
|
||||
org.apache.lucene.analysis.pattern.PatternTokenizerFactory
|
||||
org.apache.lucene.analysis.pattern.SimplePatternSplitTokenizerFactory
|
||||
org.apache.lucene.analysis.pattern.SimplePatternTokenizerFactory
|
||||
org.apache.lucene.analysis.standard.ClassicTokenizerFactory
|
||||
org.apache.lucene.analysis.standard.StandardTokenizerFactory
|
||||
org.apache.lucene.analysis.standard.UAX29URLEmailTokenizerFactory
|
||||
|
|
|
@ -96,11 +96,7 @@ import org.apache.lucene.util.CharsRef;
|
|||
import org.apache.lucene.util.Rethrow;
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
import org.apache.lucene.util.Version;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
import org.apache.lucene.util.automaton.AutomatonTestUtil;
|
||||
import org.apache.lucene.util.automaton.CharacterRunAutomaton;
|
||||
import org.apache.lucene.util.automaton.Operations;
|
||||
import org.apache.lucene.util.automaton.RegExp;
|
||||
import org.junit.AfterClass;
|
||||
import org.junit.BeforeClass;
|
||||
import org.tartarus.snowball.SnowballProgram;
|
||||
|
@ -498,9 +494,6 @@ public class TestRandomChains extends BaseTokenStreamTestCase {
|
|||
if (random.nextBoolean()) return null;
|
||||
return DateFormat.getDateInstance(DateFormat.DEFAULT, randomLocale(random));
|
||||
});
|
||||
put(Automaton.class, random -> {
|
||||
return Operations.determinize(new RegExp(AutomatonTestUtil.randomRegexp(random()), RegExp.NONE).toAutomaton(), Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
});
|
||||
}};
|
||||
|
||||
static final Set<Class<?>> allowedTokenizerArgs, allowedTokenFilterArgs, allowedCharFilterArgs;
|
||||
|
@ -510,8 +503,7 @@ public class TestRandomChains extends BaseTokenStreamTestCase {
|
|||
allowedTokenizerArgs.add(Reader.class);
|
||||
allowedTokenizerArgs.add(AttributeFactory.class);
|
||||
allowedTokenizerArgs.add(AttributeSource.class);
|
||||
allowedTokenizerArgs.add(Automaton.class);
|
||||
|
||||
|
||||
allowedTokenFilterArgs = Collections.newSetFromMap(new IdentityHashMap<Class<?>,Boolean>());
|
||||
allowedTokenFilterArgs.addAll(argProducers.keySet());
|
||||
allowedTokenFilterArgs.add(TokenStream.class);
|
||||
|
|
|
@ -1,273 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.lucene.analysis.pattern;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.StringReader;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.lucene.analysis.Analyzer;
|
||||
import org.apache.lucene.analysis.BaseTokenStreamTestCase;
|
||||
import org.apache.lucene.analysis.CharFilter;
|
||||
import org.apache.lucene.analysis.TokenStream;
|
||||
import org.apache.lucene.analysis.Tokenizer;
|
||||
import org.apache.lucene.analysis.charfilter.MappingCharFilter;
|
||||
import org.apache.lucene.analysis.charfilter.NormalizeCharMap;
|
||||
import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
|
||||
import org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
|
||||
public class TestSimplePatternSplitTokenizer extends BaseTokenStreamTestCase {
|
||||
|
||||
public void testGreedy() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("(foo)+");
|
||||
t.setReader(new StringReader("bar foofoo baz"));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"bar ", " baz"},
|
||||
new int[] {0, 10},
|
||||
new int[] {4, 14});
|
||||
}
|
||||
|
||||
public void testBackToBack() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("foo");
|
||||
t.setReader(new StringReader("bar foofoo baz"));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"bar ", " baz"},
|
||||
new int[] {0, 10},
|
||||
new int[] {4, 14});
|
||||
}
|
||||
|
||||
public void testBigLookahead() throws Exception {
|
||||
StringBuilder b = new StringBuilder();
|
||||
for(int i=0;i<100;i++) {
|
||||
b.append('a');
|
||||
}
|
||||
b.append('b');
|
||||
Tokenizer t = new SimplePatternSplitTokenizer(b.toString());
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
|
||||
b = new StringBuilder();
|
||||
for(int i=0;i<200;i++) {
|
||||
b.append('a');
|
||||
}
|
||||
t.setReader(new StringReader(b.toString()));
|
||||
t.reset();
|
||||
assertTrue(t.incrementToken());
|
||||
assertEquals(b.toString(), termAtt.toString());
|
||||
assertFalse(t.incrementToken());
|
||||
}
|
||||
|
||||
public void testNoTokens() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer(".*");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
String s;
|
||||
while (true) {
|
||||
s = TestUtil.randomUnicodeString(random());
|
||||
if (s.length() > 0) {
|
||||
break;
|
||||
}
|
||||
}
|
||||
t.setReader(new StringReader(s));
|
||||
t.reset();
|
||||
assertFalse(t.incrementToken());
|
||||
}
|
||||
|
||||
public void testEmptyStringPatternNoMatch() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("a*");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
t.setReader(new StringReader("bbb"));
|
||||
t.reset();
|
||||
assertTrue(t.incrementToken());
|
||||
assertEquals("bbb", termAtt.toString());
|
||||
assertFalse(t.incrementToken());
|
||||
}
|
||||
|
||||
public void testSplitSingleCharWhitespace() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("[ \t\r\n]");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
t.setReader(new StringReader("a \tb c"));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"a", "b", "c"},
|
||||
new int[] {0, 3, 7},
|
||||
new int[] {1, 4, 8});
|
||||
}
|
||||
|
||||
public void testSplitMultiCharWhitespace() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("[ \t\r\n]*");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
t.setReader(new StringReader("a \tb c"));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"a", "b", "c"},
|
||||
new int[] {0, 3, 7},
|
||||
new int[] {1, 4, 8});
|
||||
}
|
||||
|
||||
public void testLeadingNonToken() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("[ \t\r\n]*");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
t.setReader(new StringReader(" a c"));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"a", "c"},
|
||||
new int[] {4, 6},
|
||||
new int[] {5, 7});
|
||||
}
|
||||
|
||||
public void testTrailingNonToken() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("[ \t\r\n]*");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
t.setReader(new StringReader("a c "));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"a", "c"},
|
||||
new int[] {0, 2},
|
||||
new int[] {1, 3});
|
||||
}
|
||||
|
||||
public void testEmptyStringPatternOneMatch() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("a*");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
t.setReader(new StringReader("bbab"));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"bb", "b"},
|
||||
new int[] {0, 3},
|
||||
new int[] {2, 4});
|
||||
}
|
||||
|
||||
public void testEndOffset() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("a+");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
OffsetAttribute offsetAtt = t.getAttribute(OffsetAttribute.class);
|
||||
t.setReader(new StringReader("aaabbb"));
|
||||
t.reset();
|
||||
assertTrue(t.incrementToken());
|
||||
assertEquals("bbb", termAtt.toString());
|
||||
assertFalse(t.incrementToken());
|
||||
t.end();
|
||||
assertEquals(6, offsetAtt.endOffset());
|
||||
}
|
||||
|
||||
public void testFixedToken() throws Exception {
|
||||
Tokenizer t = new SimplePatternSplitTokenizer("aaaa");
|
||||
|
||||
t.setReader(new StringReader("aaaaaaaaaaaaaaa"));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"aaa"},
|
||||
new int[] {12},
|
||||
new int[] {15});
|
||||
}
|
||||
|
||||
public void testBasic() throws Exception
|
||||
{
|
||||
String[][] tests = {
|
||||
// pattern input output
|
||||
{ "--", "aaa--bbb--ccc", "aaa bbb ccc" },
|
||||
{ ":", "aaa:bbb:ccc", "aaa bbb ccc" },
|
||||
{ ":", "boo:and:foo", "boo and foo" },
|
||||
{ "o", "boo:and:foo", "b :and:f" },
|
||||
};
|
||||
|
||||
for(String[] test : tests) {
|
||||
TokenStream stream = new SimplePatternSplitTokenizer(test[0]);
|
||||
((Tokenizer)stream).setReader(new StringReader(test[1]));
|
||||
String out = tsToString(stream);
|
||||
assertEquals("pattern: "+test[0]+" with input: "+test[1], test[2], out);
|
||||
}
|
||||
}
|
||||
|
||||
public void testNotDeterminized() throws Exception {
|
||||
Automaton a = new Automaton();
|
||||
int start = a.createState();
|
||||
int mid1 = a.createState();
|
||||
int mid2 = a.createState();
|
||||
int end = a.createState();
|
||||
a.setAccept(end, true);
|
||||
a.addTransition(start, mid1, 'a', 'z');
|
||||
a.addTransition(start, mid2, 'a', 'z');
|
||||
a.addTransition(mid1, end, 'b');
|
||||
a.addTransition(mid2, end, 'b');
|
||||
expectThrows(IllegalArgumentException.class, () -> {new SimplePatternSplitTokenizer(a);});
|
||||
}
|
||||
|
||||
public void testOffsetCorrection() throws Exception {
|
||||
final String INPUT = "Günther Günther is here";
|
||||
|
||||
// create MappingCharFilter
|
||||
List<String> mappingRules = new ArrayList<>();
|
||||
mappingRules.add( "\"ü\" => \"ü\"" );
|
||||
NormalizeCharMap.Builder builder = new NormalizeCharMap.Builder();
|
||||
builder.add("ü", "ü");
|
||||
NormalizeCharMap normMap = builder.build();
|
||||
CharFilter charStream = new MappingCharFilter( normMap, new StringReader(INPUT));
|
||||
|
||||
// create SimplePatternSplitTokenizer
|
||||
Tokenizer stream = new SimplePatternSplitTokenizer("Günther");
|
||||
stream.setReader(charStream);
|
||||
assertTokenStreamContents(stream,
|
||||
new String[] { " ", " is here" },
|
||||
new int[] { 12, 25 },
|
||||
new int[] { 13, 33 },
|
||||
INPUT.length());
|
||||
}
|
||||
|
||||
/**
|
||||
* TODO: rewrite tests not to use string comparison.
|
||||
*/
|
||||
private static String tsToString(TokenStream in) throws IOException {
|
||||
StringBuilder out = new StringBuilder();
|
||||
CharTermAttribute termAtt = in.addAttribute(CharTermAttribute.class);
|
||||
// extra safety to enforce, that the state is not preserved and also
|
||||
// assign bogus values
|
||||
in.clearAttributes();
|
||||
termAtt.setEmpty().append("bogusTerm");
|
||||
in.reset();
|
||||
while (in.incrementToken()) {
|
||||
if (out.length() > 0) {
|
||||
out.append(' ');
|
||||
}
|
||||
out.append(termAtt.toString());
|
||||
in.clearAttributes();
|
||||
termAtt.setEmpty().append("bogusTerm");
|
||||
}
|
||||
|
||||
in.close();
|
||||
return out.toString();
|
||||
}
|
||||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer a = new Analyzer() {
|
||||
@Override
|
||||
protected TokenStreamComponents createComponents(String fieldName) {
|
||||
Tokenizer tokenizer = new SimplePatternSplitTokenizer("a");
|
||||
return new TokenStreamComponents(tokenizer);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), a, 1000*RANDOM_MULTIPLIER);
|
||||
a.close();
|
||||
|
||||
Analyzer b = new Analyzer() {
|
||||
@Override
|
||||
protected TokenStreamComponents createComponents(String fieldName) {
|
||||
Tokenizer tokenizer = new SimplePatternSplitTokenizer("a");
|
||||
return new TokenStreamComponents(tokenizer);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), b, 1000*RANDOM_MULTIPLIER);
|
||||
b.close();
|
||||
}
|
||||
}
|
|
@ -1,218 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.lucene.analysis.pattern;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.StringReader;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
||||
import org.apache.lucene.analysis.Analyzer;
|
||||
import org.apache.lucene.analysis.BaseTokenStreamTestCase;
|
||||
import org.apache.lucene.analysis.CharFilter;
|
||||
import org.apache.lucene.analysis.TokenStream;
|
||||
import org.apache.lucene.analysis.Tokenizer;
|
||||
import org.apache.lucene.analysis.charfilter.MappingCharFilter;
|
||||
import org.apache.lucene.analysis.charfilter.NormalizeCharMap;
|
||||
import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
|
||||
import org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
import org.apache.lucene.util.automaton.Automaton;
|
||||
|
||||
public class TestSimplePatternTokenizer extends BaseTokenStreamTestCase {
|
||||
|
||||
public void testGreedy() throws Exception {
|
||||
Tokenizer t = new SimplePatternTokenizer("(foo)+");
|
||||
t.setReader(new StringReader("bar foofoo baz"));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"foofoo"},
|
||||
new int[] {4},
|
||||
new int[] {10});
|
||||
}
|
||||
|
||||
public void testBigLookahead() throws Exception {
|
||||
StringBuilder b = new StringBuilder();
|
||||
for(int i=0;i<100;i++) {
|
||||
b.append('a');
|
||||
}
|
||||
b.append('b');
|
||||
Tokenizer t = new SimplePatternTokenizer(b.toString());
|
||||
|
||||
b = new StringBuilder();
|
||||
for(int i=0;i<200;i++) {
|
||||
b.append('a');
|
||||
}
|
||||
t.setReader(new StringReader(b.toString()));
|
||||
t.reset();
|
||||
assertFalse(t.incrementToken());
|
||||
}
|
||||
|
||||
public void testOneToken() throws Exception {
|
||||
Tokenizer t = new SimplePatternTokenizer(".*");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
String s;
|
||||
while (true) {
|
||||
s = TestUtil.randomUnicodeString(random());
|
||||
if (s.length() > 0) {
|
||||
break;
|
||||
}
|
||||
}
|
||||
t.setReader(new StringReader(s));
|
||||
t.reset();
|
||||
assertTrue(t.incrementToken());
|
||||
assertEquals(s, termAtt.toString());
|
||||
}
|
||||
|
||||
public void testEmptyStringPatternNoMatch() throws Exception {
|
||||
Tokenizer t = new SimplePatternTokenizer("a*");
|
||||
t.setReader(new StringReader("bbb"));
|
||||
t.reset();
|
||||
assertFalse(t.incrementToken());
|
||||
}
|
||||
|
||||
public void testEmptyStringPatternOneMatch() throws Exception {
|
||||
Tokenizer t = new SimplePatternTokenizer("a*");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
t.setReader(new StringReader("bbab"));
|
||||
t.reset();
|
||||
assertTrue(t.incrementToken());
|
||||
assertEquals("a", termAtt.toString());
|
||||
assertFalse(t.incrementToken());
|
||||
}
|
||||
|
||||
public void testEndOffset() throws Exception {
|
||||
Tokenizer t = new SimplePatternTokenizer("a+");
|
||||
CharTermAttribute termAtt = t.getAttribute(CharTermAttribute.class);
|
||||
OffsetAttribute offsetAtt = t.getAttribute(OffsetAttribute.class);
|
||||
t.setReader(new StringReader("aaabbb"));
|
||||
t.reset();
|
||||
assertTrue(t.incrementToken());
|
||||
assertEquals("aaa", termAtt.toString());
|
||||
assertFalse(t.incrementToken());
|
||||
t.end();
|
||||
assertEquals(6, offsetAtt.endOffset());
|
||||
}
|
||||
|
||||
public void testFixedToken() throws Exception {
|
||||
Tokenizer t = new SimplePatternTokenizer("aaaa");
|
||||
|
||||
t.setReader(new StringReader("aaaaaaaaaaaaaaa"));
|
||||
assertTokenStreamContents(t,
|
||||
new String[] {"aaaa", "aaaa", "aaaa"},
|
||||
new int[] {0, 4, 8},
|
||||
new int[] {4, 8, 12});
|
||||
}
|
||||
|
||||
public void testBasic() throws Exception {
|
||||
String qpattern = "\\'([^\\']+)\\'"; // get stuff between "'"
|
||||
String[][] tests = {
|
||||
// pattern input output
|
||||
{ ":", "boo:and:foo", ": :" },
|
||||
{ qpattern, "aaa 'bbb' 'ccc'", "'bbb' 'ccc'" },
|
||||
};
|
||||
|
||||
for(String[] test : tests) {
|
||||
TokenStream stream = new SimplePatternTokenizer(test[0]);
|
||||
((Tokenizer)stream).setReader(new StringReader(test[1]));
|
||||
String out = tsToString(stream);
|
||||
|
||||
assertEquals("pattern: "+test[0]+" with input: "+test[1], test[2], out);
|
||||
}
|
||||
}
|
||||
|
||||
public void testNotDeterminized() throws Exception {
|
||||
Automaton a = new Automaton();
|
||||
int start = a.createState();
|
||||
int mid1 = a.createState();
|
||||
int mid2 = a.createState();
|
||||
int end = a.createState();
|
||||
a.setAccept(end, true);
|
||||
a.addTransition(start, mid1, 'a', 'z');
|
||||
a.addTransition(start, mid2, 'a', 'z');
|
||||
a.addTransition(mid1, end, 'b');
|
||||
a.addTransition(mid2, end, 'b');
|
||||
expectThrows(IllegalArgumentException.class, () -> {new SimplePatternTokenizer(a);});
|
||||
}
|
||||
|
||||
public void testOffsetCorrection() throws Exception {
|
||||
final String INPUT = "Günther Günther is here";
|
||||
|
||||
// create MappingCharFilter
|
||||
List<String> mappingRules = new ArrayList<>();
|
||||
mappingRules.add( "\"ü\" => \"ü\"" );
|
||||
NormalizeCharMap.Builder builder = new NormalizeCharMap.Builder();
|
||||
builder.add("ü", "ü");
|
||||
NormalizeCharMap normMap = builder.build();
|
||||
CharFilter charStream = new MappingCharFilter( normMap, new StringReader(INPUT));
|
||||
|
||||
// create SimplePatternTokenizer
|
||||
Tokenizer stream = new SimplePatternTokenizer("Günther");
|
||||
stream.setReader(charStream);
|
||||
assertTokenStreamContents(stream,
|
||||
new String[] { "Günther", "Günther" },
|
||||
new int[] { 0, 13 },
|
||||
new int[] { 12, 25 },
|
||||
INPUT.length());
|
||||
}
|
||||
|
||||
/**
|
||||
* TODO: rewrite tests not to use string comparison.
|
||||
*/
|
||||
private static String tsToString(TokenStream in) throws IOException {
|
||||
StringBuilder out = new StringBuilder();
|
||||
CharTermAttribute termAtt = in.addAttribute(CharTermAttribute.class);
|
||||
// extra safety to enforce, that the state is not preserved and also
|
||||
// assign bogus values
|
||||
in.clearAttributes();
|
||||
termAtt.setEmpty().append("bogusTerm");
|
||||
in.reset();
|
||||
while (in.incrementToken()) {
|
||||
if (out.length() > 0) {
|
||||
out.append(' ');
|
||||
}
|
||||
out.append(termAtt.toString());
|
||||
in.clearAttributes();
|
||||
termAtt.setEmpty().append("bogusTerm");
|
||||
}
|
||||
|
||||
in.close();
|
||||
return out.toString();
|
||||
}
|
||||
|
||||
/** blast some random strings through the analyzer */
|
||||
public void testRandomStrings() throws Exception {
|
||||
Analyzer a = new Analyzer() {
|
||||
@Override
|
||||
protected TokenStreamComponents createComponents(String fieldName) {
|
||||
Tokenizer tokenizer = new SimplePatternTokenizer("a");
|
||||
return new TokenStreamComponents(tokenizer);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), a, 1000*RANDOM_MULTIPLIER);
|
||||
a.close();
|
||||
|
||||
Analyzer b = new Analyzer() {
|
||||
@Override
|
||||
protected TokenStreamComponents createComponents(String fieldName) {
|
||||
Tokenizer tokenizer = new SimplePatternTokenizer("a");
|
||||
return new TokenStreamComponents(tokenizer);
|
||||
}
|
||||
};
|
||||
checkRandomData(random(), b, 1000*RANDOM_MULTIPLIER);
|
||||
b.close();
|
||||
}
|
||||
}
|
|
@ -2348,7 +2348,7 @@ ${ant.project.name}.test.dependencies=${test.classpath.list}
|
|||
<property name="forbidden-sysout-excludes" value=""/>
|
||||
|
||||
<target name="-install-forbidden-apis" unless="forbidden-apis.loaded" depends="ivy-availability-check,ivy-configure">
|
||||
<ivy:cachepath organisation="de.thetaphi" module="forbiddenapis" revision="2.3"
|
||||
<ivy:cachepath organisation="de.thetaphi" module="forbiddenapis" revision="2.2"
|
||||
inline="true" conf="default" transitive="true" pathid="forbidden-apis.classpath"/>
|
||||
<taskdef name="forbidden-apis" classname="de.thetaphi.forbiddenapis.ant.AntTask" classpathref="forbidden-apis.classpath"/>
|
||||
<property name="forbidden-apis.loaded" value="true"/>
|
||||
|
|
|
@ -362,13 +362,11 @@
|
|||
* </p>
|
||||
* <ol>
|
||||
* <li>Inhibiting phrase and proximity matches in sentence boundaries – for this, a tokenizer that
|
||||
* identifies a new sentence can add 1 to the position increment of the first token of the new sentence.</li>
|
||||
* <li>Injecting synonyms – synonyms of a token should be created at the same position as the
|
||||
* original token, and the output order of the original token and the injected synonym is undefined
|
||||
* as long as they both leave from the same position. As result, all synonyms of a token would be
|
||||
* considered to appear in exactly the same position as that token, and so would they be seen by
|
||||
* phrase and proximity searches. For multi-token synonyms to work correctly, you should use
|
||||
* {@code SynoymGraphFilter} at search time only.</li>
|
||||
* identifies a new sentence can add 1 to the position increment of the first token of the new sentence.</li>
|
||||
* <li>Injecting synonyms – here, synonyms of a token should be added after that token,
|
||||
* and their position increment should be set to 0.
|
||||
* As result, all synonyms of a token would be considered to appear in exactly the
|
||||
* same position as that token, and so would they be seen by phrase and proximity searches.</li>
|
||||
* </ol>
|
||||
*
|
||||
* <h3>Token Position Length</h3>
|
||||
|
|
|
@ -27,9 +27,9 @@ public class ByteRunAutomaton extends RunAutomaton {
|
|||
this(a, false, Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
}
|
||||
|
||||
/** expert: if isBinary is true, the input is already byte-based */
|
||||
/** expert: if utf8 is true, the input is already byte-based */
|
||||
public ByteRunAutomaton(Automaton a, boolean isBinary, int maxDeterminizedStates) {
|
||||
super(isBinary ? a : new UTF32ToUTF8().convert(a), 256, maxDeterminizedStates);
|
||||
super(isBinary ? a : new UTF32ToUTF8().convert(a), 256, true, maxDeterminizedStates);
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -36,7 +36,7 @@ public class CharacterRunAutomaton extends RunAutomaton {
|
|||
* it then a TooComplexToDeterminizeException is thrown.
|
||||
*/
|
||||
public CharacterRunAutomaton(Automaton a, int maxDeterminizedStates) {
|
||||
super(a, Character.MAX_CODE_POINT+1, maxDeterminizedStates);
|
||||
super(a, Character.MAX_CODE_POINT, false, maxDeterminizedStates);
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -29,17 +29,6 @@
|
|||
|
||||
package org.apache.lucene.util.automaton;
|
||||
|
||||
import java.util.ArrayDeque;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.BitSet;
|
||||
import java.util.Collection;
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
|
||||
import org.apache.lucene.util.ArrayUtil;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.BytesRefBuilder;
|
||||
|
@ -47,6 +36,17 @@ import org.apache.lucene.util.IntsRef;
|
|||
import org.apache.lucene.util.IntsRefBuilder;
|
||||
import org.apache.lucene.util.RamUsageEstimator;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.Arrays;
|
||||
import java.util.BitSet;
|
||||
import java.util.Collection;
|
||||
import java.util.HashMap;
|
||||
import java.util.HashSet;
|
||||
import java.util.LinkedList;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.Set;
|
||||
|
||||
/**
|
||||
* Automata operations.
|
||||
*
|
||||
|
@ -335,7 +335,7 @@ final public class Operations {
|
|||
Transition[][] transitions2 = a2.getSortedTransitions();
|
||||
Automaton c = new Automaton();
|
||||
c.createState();
|
||||
ArrayDeque<StatePair> worklist = new ArrayDeque<>();
|
||||
LinkedList<StatePair> worklist = new LinkedList<>();
|
||||
HashMap<StatePair,StatePair> newstates = new HashMap<>();
|
||||
StatePair p = new StatePair(0, 0, 0);
|
||||
worklist.add(p);
|
||||
|
@ -435,7 +435,7 @@ final public class Operations {
|
|||
// TODO: cutover to iterators instead
|
||||
Transition[][] transitions1 = a1.getSortedTransitions();
|
||||
Transition[][] transitions2 = a2.getSortedTransitions();
|
||||
ArrayDeque<StatePair> worklist = new ArrayDeque<>();
|
||||
LinkedList<StatePair> worklist = new LinkedList<>();
|
||||
HashSet<StatePair> visited = new HashSet<>();
|
||||
StatePair p = new StatePair(0, 0);
|
||||
worklist.add(p);
|
||||
|
@ -682,7 +682,7 @@ final public class Operations {
|
|||
// Create state 0:
|
||||
b.createState();
|
||||
|
||||
ArrayDeque<SortedIntSet.FrozenIntSet> worklist = new ArrayDeque<>();
|
||||
LinkedList<SortedIntSet.FrozenIntSet> worklist = new LinkedList<>();
|
||||
Map<SortedIntSet.FrozenIntSet,Integer> newstate = new HashMap<>();
|
||||
|
||||
worklist.add(initialset);
|
||||
|
@ -804,7 +804,7 @@ final public class Operations {
|
|||
return false;
|
||||
}
|
||||
|
||||
ArrayDeque<Integer> workList = new ArrayDeque<>();
|
||||
LinkedList<Integer> workList = new LinkedList<>();
|
||||
BitSet seen = new BitSet(a.getNumStates());
|
||||
workList.add(0);
|
||||
seen.set(0);
|
||||
|
@ -907,7 +907,7 @@ final public class Operations {
|
|||
if (numStates == 0) {
|
||||
return live;
|
||||
}
|
||||
ArrayDeque<Integer> workList = new ArrayDeque<>();
|
||||
LinkedList<Integer> workList = new LinkedList<>();
|
||||
live.set(0);
|
||||
workList.add(0);
|
||||
|
||||
|
@ -946,7 +946,7 @@ final public class Operations {
|
|||
}
|
||||
Automaton a2 = builder.finish();
|
||||
|
||||
ArrayDeque<Integer> workList = new ArrayDeque<>();
|
||||
LinkedList<Integer> workList = new LinkedList<>();
|
||||
BitSet live = new BitSet(numStates);
|
||||
BitSet acceptBits = a.getAcceptStates();
|
||||
int s = 0;
|
||||
|
@ -1010,6 +1010,22 @@ final public class Operations {
|
|||
return result;
|
||||
}
|
||||
|
||||
/**
|
||||
* Finds the largest entry whose value is less than or equal to c, or 0 if
|
||||
* there is no such entry.
|
||||
*/
|
||||
static int findIndex(int c, int[] points) {
|
||||
int a = 0;
|
||||
int b = points.length;
|
||||
while (b - a > 1) {
|
||||
int d = (a + b) >>> 1;
|
||||
if (points[d] > c) b = d;
|
||||
else if (points[d] < c) a = d;
|
||||
else return d;
|
||||
}
|
||||
return a;
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns true if the language of this automaton is finite. The
|
||||
* automaton must not have any dead states.
|
||||
|
|
|
@ -38,62 +38,13 @@ import java.util.Arrays;
|
|||
*/
|
||||
public abstract class RunAutomaton {
|
||||
final Automaton automaton;
|
||||
final int alphabetSize;
|
||||
final int maxInterval;
|
||||
final int size;
|
||||
final boolean[] accept;
|
||||
final int[] transitions; // delta(state,c) = transitions[state*points.length +
|
||||
// getCharClass(c)]
|
||||
final int[] points; // char interval start points
|
||||
final int[] classmap; // map from char number to class
|
||||
|
||||
/**
|
||||
* Constructs a new <code>RunAutomaton</code> from a deterministic
|
||||
* <code>Automaton</code>.
|
||||
*
|
||||
* @param a an automaton
|
||||
*/
|
||||
protected RunAutomaton(Automaton a, int alphabetSize) {
|
||||
this(a, alphabetSize, Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
}
|
||||
|
||||
/**
|
||||
* Constructs a new <code>RunAutomaton</code> from a deterministic
|
||||
* <code>Automaton</code>.
|
||||
*
|
||||
* @param a an automaton
|
||||
* @param maxDeterminizedStates maximum number of states that can be created
|
||||
* while determinizing a
|
||||
*/
|
||||
protected RunAutomaton(Automaton a, int alphabetSize, int maxDeterminizedStates) {
|
||||
this.alphabetSize = alphabetSize;
|
||||
a = Operations.determinize(a, maxDeterminizedStates);
|
||||
this.automaton = a;
|
||||
points = a.getStartPoints();
|
||||
size = Math.max(1,a.getNumStates());
|
||||
accept = new boolean[size];
|
||||
transitions = new int[size * points.length];
|
||||
Arrays.fill(transitions, -1);
|
||||
for (int n=0;n<size;n++) {
|
||||
accept[n] = a.isAccept(n);
|
||||
for (int c = 0; c < points.length; c++) {
|
||||
int dest = a.step(n, points[c]);
|
||||
assert dest == -1 || dest < size;
|
||||
transitions[n * points.length + c] = dest;
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* Set alphabet table for optimal run performance.
|
||||
*/
|
||||
classmap = new int[Math.min(256, alphabetSize)];
|
||||
int i = 0;
|
||||
for (int j = 0; j < classmap.length; j++) {
|
||||
if (i + 1 < points.length && j == points[i + 1]) {
|
||||
i++;
|
||||
}
|
||||
classmap[j] = i;
|
||||
}
|
||||
}
|
||||
final int[] classmap; // map from char number to class class
|
||||
|
||||
/**
|
||||
* Returns a string representation of this automaton.
|
||||
|
@ -112,7 +63,7 @@ public abstract class RunAutomaton {
|
|||
int min = points[j];
|
||||
int max;
|
||||
if (j + 1 < points.length) max = (points[j + 1] - 1);
|
||||
else max = alphabetSize;
|
||||
else max = maxInterval;
|
||||
b.append(" ");
|
||||
Automaton.appendCharString(min, b);
|
||||
if (min != max) {
|
||||
|
@ -152,19 +103,63 @@ public abstract class RunAutomaton {
|
|||
* Gets character class of given codepoint
|
||||
*/
|
||||
final int getCharClass(int c) {
|
||||
|
||||
// binary search
|
||||
int a = 0;
|
||||
int b = points.length;
|
||||
while (b - a > 1) {
|
||||
int d = (a + b) >>> 1;
|
||||
if (points[d] > c) b = d;
|
||||
else if (points[d] < c) a = d;
|
||||
else return d;
|
||||
}
|
||||
return a;
|
||||
return Operations.findIndex(c, points);
|
||||
}
|
||||
|
||||
/**
|
||||
* Constructs a new <code>RunAutomaton</code> from a deterministic
|
||||
* <code>Automaton</code>.
|
||||
*
|
||||
* @param a an automaton
|
||||
*/
|
||||
public RunAutomaton(Automaton a, int maxInterval, boolean tableize) {
|
||||
this(a, maxInterval, tableize, Operations.DEFAULT_MAX_DETERMINIZED_STATES);
|
||||
}
|
||||
|
||||
/**
|
||||
* Constructs a new <code>RunAutomaton</code> from a deterministic
|
||||
* <code>Automaton</code>.
|
||||
*
|
||||
* @param a an automaton
|
||||
* @param maxDeterminizedStates maximum number of states that can be created
|
||||
* while determinizing a
|
||||
*/
|
||||
public RunAutomaton(Automaton a, int maxInterval, boolean tableize,
|
||||
int maxDeterminizedStates) {
|
||||
this.maxInterval = maxInterval;
|
||||
a = Operations.determinize(a, maxDeterminizedStates);
|
||||
this.automaton = a;
|
||||
points = a.getStartPoints();
|
||||
size = Math.max(1,a.getNumStates());
|
||||
accept = new boolean[size];
|
||||
transitions = new int[size * points.length];
|
||||
Arrays.fill(transitions, -1);
|
||||
for (int n=0;n<size;n++) {
|
||||
accept[n] = a.isAccept(n);
|
||||
for (int c = 0; c < points.length; c++) {
|
||||
int dest = a.step(n, points[c]);
|
||||
assert dest == -1 || dest < size;
|
||||
transitions[n * points.length + c] = dest;
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* Set alphabet table for optimal run performance.
|
||||
*/
|
||||
if (tableize) {
|
||||
classmap = new int[maxInterval + 1];
|
||||
int i = 0;
|
||||
for (int j = 0; j <= maxInterval; j++) {
|
||||
if (i + 1 < points.length && j == points[i + 1]) {
|
||||
i++;
|
||||
}
|
||||
classmap[j] = i;
|
||||
}
|
||||
} else {
|
||||
classmap = null;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns the state obtained by reading the given char from the given state.
|
||||
* Returns -1 if not obtaining any such state. (If the original
|
||||
|
@ -173,8 +168,7 @@ public abstract class RunAutomaton {
|
|||
* transition function.)
|
||||
*/
|
||||
public final int step(int state, int c) {
|
||||
assert c < alphabetSize;
|
||||
if (c >= classmap.length) {
|
||||
if (classmap == null) {
|
||||
return transitions[state * points.length + getCharClass(c)];
|
||||
} else {
|
||||
return transitions[state * points.length + classmap[c]];
|
||||
|
@ -185,7 +179,7 @@ public abstract class RunAutomaton {
|
|||
public int hashCode() {
|
||||
final int prime = 31;
|
||||
int result = 1;
|
||||
result = prime * result + alphabetSize;
|
||||
result = prime * result + maxInterval;
|
||||
result = prime * result + points.length;
|
||||
result = prime * result + size;
|
||||
return result;
|
||||
|
@ -197,7 +191,7 @@ public abstract class RunAutomaton {
|
|||
if (obj == null) return false;
|
||||
if (getClass() != obj.getClass()) return false;
|
||||
RunAutomaton other = (RunAutomaton) obj;
|
||||
if (alphabetSize != other.alphabetSize) return false;
|
||||
if (maxInterval != other.maxInterval) return false;
|
||||
if (size != other.size) return false;
|
||||
if (!Arrays.equals(points, other.points)) return false;
|
||||
if (!Arrays.equals(accept, other.accept)) return false;
|
||||
|
|
|
@ -17,7 +17,7 @@
|
|||
under the License.
|
||||
-->
|
||||
<ivy-module version="2.0">
|
||||
<info organisation="org.apache.lucene" module="demo"/>
|
||||
<info organisation="org.apache.lucene" module="core-demo"/>
|
||||
<configurations defaultconfmapping="compile->master">
|
||||
<conf name="compile" transitive="false"/>
|
||||
</configurations>
|
||||
|
|
|
@ -367,7 +367,7 @@ class TermAutomatonScorer extends Scorer {
|
|||
|
||||
static class TermRunAutomaton extends RunAutomaton {
|
||||
public TermRunAutomaton(Automaton a, int termCount) {
|
||||
super(a, termCount);
|
||||
super(a, termCount, true);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -17,7 +17,7 @@
|
|||
under the License.
|
||||
-->
|
||||
<ivy-module version="2.0">
|
||||
<info organisation="org.apache.lucene" module="test-framework"/>
|
||||
<info organisation="org.apache.lucene" module="core-test-framework"/>
|
||||
|
||||
<configurations defaultconfmapping="compile->master">
|
||||
<conf name="compile" transitive="false"/>
|
||||
|
|
|
@ -40,7 +40,6 @@ import org.apache.lucene.util.Bits;
|
|||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.IOUtils;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.lucene.util.Rethrow;
|
||||
import org.apache.lucene.util.StringHelper;
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
|
||||
|
@ -233,7 +232,16 @@ public abstract class BasePointsFormatTestCase extends BaseIndexFileFormatTestCa
|
|||
dir.setRandomIOExceptionRateOnOpen(0.05);
|
||||
verify(dir, docValues, null, numDims, numBytesPerDim, true);
|
||||
} catch (IllegalStateException ise) {
|
||||
done = handlePossiblyFakeException(ise);
|
||||
if (ise.getMessage().contains("this writer hit an unrecoverable error")) {
|
||||
Throwable cause = ise.getCause();
|
||||
if (cause != null && cause.getMessage().contains("a random IOException")) {
|
||||
done = true;
|
||||
} else {
|
||||
throw ise;
|
||||
}
|
||||
} else {
|
||||
throw ise;
|
||||
}
|
||||
} catch (AssertionError ae) {
|
||||
if (ae.getMessage() != null && ae.getMessage().contains("does not exist; files=")) {
|
||||
// OK: likely we threw the random IOExc when IW was asserting the commit files exist
|
||||
|
@ -245,28 +253,23 @@ public abstract class BasePointsFormatTestCase extends BaseIndexFileFormatTestCa
|
|||
// This just means we got a too-small maxMB for the maxPointsInLeafNode; just retry w/ more heap
|
||||
assertTrue(iae.getMessage().contains("either increase maxMBSortInHeap or decrease maxPointsInLeafNode"));
|
||||
} catch (IOException ioe) {
|
||||
done = handlePossiblyFakeException(ioe);
|
||||
Throwable ex = ioe;
|
||||
while (ex != null) {
|
||||
String message = ex.getMessage();
|
||||
if (message != null && (message.contains("a random IOException") || message.contains("background merge hit exception"))) {
|
||||
done = true;
|
||||
break;
|
||||
}
|
||||
ex = ex.getCause();
|
||||
}
|
||||
if (done == false) {
|
||||
throw ioe;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// TODO: merge w/ BaseIndexFileFormatTestCase.handleFakeIOException
|
||||
private boolean handlePossiblyFakeException(Exception e) {
|
||||
Throwable ex = e;
|
||||
while (ex != null) {
|
||||
String message = ex.getMessage();
|
||||
if (message != null && (message.contains("a random IOException") || message.contains("background merge hit exception"))) {
|
||||
return true;
|
||||
}
|
||||
ex = ex.getCause();
|
||||
}
|
||||
Rethrow.rethrow(e);
|
||||
|
||||
// dead code yet javac disagrees:
|
||||
return false;
|
||||
}
|
||||
|
||||
public void testMultiValued() throws Exception {
|
||||
int numBytesPerDim = TestUtil.nextInt(random(), 2, PointValues.MAX_NUM_BYTES);
|
||||
int numDims = TestUtil.nextInt(random(), 1, PointValues.MAX_DIMENSIONS);
|
||||
|
|
|
@ -17,7 +17,7 @@
|
|||
under the License.
|
||||
-->
|
||||
<ivy-module version="2.0">
|
||||
<info organisation="org.apache.lucene" module="tools"/>
|
||||
<info organisation="org.apache.lucene" module="core-tools"/>
|
||||
<configurations defaultconfmapping="compile->master">
|
||||
<conf name="compile" transitive="false"/>
|
||||
</configurations>
|
||||
|
|
|
@ -54,7 +54,6 @@ import java.util.SortedMap;
|
|||
import java.util.SortedSet;
|
||||
import java.util.TreeMap;
|
||||
import java.util.TreeSet;
|
||||
import java.util.function.Consumer;
|
||||
import java.util.regex.Matcher;
|
||||
import java.util.regex.Pattern;
|
||||
|
||||
|
@ -119,7 +118,6 @@ public class GetMavenDependenciesTask extends Task {
|
|||
private final DocumentBuilder documentBuilder;
|
||||
private File ivyCacheDir;
|
||||
private Pattern internalJarPattern;
|
||||
private Map<String,String> ivyModuleInfo;
|
||||
|
||||
|
||||
/**
|
||||
|
@ -191,8 +189,6 @@ public class GetMavenDependenciesTask extends Task {
|
|||
internalJarPattern = Pattern.compile(".*(lucene|solr)([^/]*?)-"
|
||||
+ Pattern.quote(getProject().getProperty("version")) + "\\.jar");
|
||||
|
||||
ivyModuleInfo = getIvyModuleInfo(ivyXmlResources, documentBuilder, xpath);
|
||||
|
||||
setInternalDependencyProperties(); // side-effect: all modules' internal deps are recorded
|
||||
setExternalDependencyProperties(); // side-effect: all modules' external deps are recorded
|
||||
setGrandparentDependencyManagementProperty(); // uses deps recorded in above two methods
|
||||
|
@ -223,57 +219,11 @@ public class GetMavenDependenciesTask extends Task {
|
|||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Visits all ivy.xml files and collects module and organisation attributes into a map.
|
||||
*/
|
||||
private static Map<String,String> getIvyModuleInfo(Resources ivyXmlResources,
|
||||
DocumentBuilder documentBuilder, XPath xpath) {
|
||||
Map<String,String> ivyInfoModuleToOrganisation = new HashMap<String,String>();
|
||||
traverseIvyXmlResources(ivyXmlResources, new Consumer<File>() {
|
||||
@Override
|
||||
public void accept(File f) {
|
||||
try {
|
||||
Document document = documentBuilder.parse(f);
|
||||
{
|
||||
String infoPath = "/ivy-module/info";
|
||||
NodeList infos = (NodeList)xpath.evaluate(infoPath, document, XPathConstants.NODESET);
|
||||
for (int infoNum = 0 ; infoNum < infos.getLength() ; ++infoNum) {
|
||||
Element infoElement = (Element)infos.item(infoNum);
|
||||
String infoOrg = infoElement.getAttribute("organisation");
|
||||
String infoOrgSuffix = infoOrg.substring(infoOrg.lastIndexOf('.')+1);
|
||||
String infoModule = infoElement.getAttribute("module");
|
||||
String module = infoOrgSuffix+"-"+infoModule;
|
||||
ivyInfoModuleToOrganisation.put(module, infoOrg);
|
||||
}
|
||||
}
|
||||
} catch (XPathExpressionException | IOException | SAXException e) {
|
||||
throw new RuntimeException(e);
|
||||
}
|
||||
}
|
||||
});
|
||||
return ivyInfoModuleToOrganisation;
|
||||
}
|
||||
|
||||
/**
|
||||
* Collects external dependencies from each ivy.xml file and sets
|
||||
* external dependency properties to be inserted into modules' POMs.
|
||||
*/
|
||||
private void setExternalDependencyProperties() {
|
||||
traverseIvyXmlResources(ivyXmlResources, new Consumer<File>() {
|
||||
@Override
|
||||
public void accept(File f) {
|
||||
try {
|
||||
collectExternalDependenciesFromIvyXmlFile(f);
|
||||
} catch (XPathExpressionException | IOException | SAXException e) {
|
||||
throw new RuntimeException(e);
|
||||
}
|
||||
}
|
||||
});
|
||||
addSharedExternalDependencies();
|
||||
setExternalDependencyXmlProperties();
|
||||
}
|
||||
|
||||
private static void traverseIvyXmlResources(Resources ivyXmlResources, Consumer<File> ivyXmlFileConsumer) {
|
||||
@SuppressWarnings("unchecked")
|
||||
Iterator<Resource> iter = (Iterator<Resource>)ivyXmlResources.iterator();
|
||||
while (iter.hasNext()) {
|
||||
|
@ -288,13 +238,15 @@ public class GetMavenDependenciesTask extends Task {
|
|||
|
||||
File ivyXmlFile = ((FileResource)resource).getFile();
|
||||
try {
|
||||
ivyXmlFileConsumer.accept(ivyXmlFile);
|
||||
collectExternalDependenciesFromIvyXmlFile(ivyXmlFile);
|
||||
} catch (BuildException e) {
|
||||
throw e;
|
||||
} catch (Exception e) {
|
||||
throw new BuildException("Exception reading file " + ivyXmlFile.getPath() + ": " + e, e);
|
||||
}
|
||||
}
|
||||
addSharedExternalDependencies();
|
||||
setExternalDependencyXmlProperties();
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -444,7 +396,7 @@ public class GetMavenDependenciesTask extends Task {
|
|||
}
|
||||
}
|
||||
}
|
||||
String groupId = ivyModuleInfo.get(artifactId);
|
||||
String groupId = "org.apache." + artifactId.substring(0, artifactId.indexOf('-'));
|
||||
appendDependencyXml(builder, groupId, artifactId, " ", "${project.version}", false, false, null, exclusions);
|
||||
}
|
||||
}
|
||||
|
@ -629,7 +581,7 @@ public class GetMavenDependenciesTask extends Task {
|
|||
continue; // skip external (/(test-)lib/), and non-jar and unwanted (self) internal deps
|
||||
}
|
||||
String artifactId = dependencyToArtifactId(newPropertyKey, dependency);
|
||||
String groupId = ivyModuleInfo.get(artifactId);
|
||||
String groupId = "org.apache." + artifactId.substring(0, artifactId.indexOf('-'));
|
||||
String coordinate = groupId + ':' + artifactId;
|
||||
sortedDeps.add(coordinate);
|
||||
}
|
||||
|
|
|
@ -74,12 +74,20 @@ Optimizations
|
|||
* SOLR-9584: Support Solr being proxied with another endpoint than default /solr, by using relative links
|
||||
in AdminUI javascripts (Yun Jie Zhou via janhoy)
|
||||
|
||||
* SOLR-9996: Unstored IntPointField returns Long type (Ishan Chattopadhyaya)
|
||||
|
||||
* SOLR-5944: In-place updates of Numeric DocValues. To leverage this, the _version_ field and the updated
|
||||
field must both be stored=false, indexed=false, docValues=true. (Ishan Chattopadhyaya, hossman, noble,
|
||||
shalin, yonik)
|
||||
|
||||
Other Changes
|
||||
----------------------
|
||||
* SOLR-8396: Add support for PointFields in Solr (Ishan Chattopadhyaya, Tomás Fernández Löbbe)
|
||||
|
||||
* SOLR-10011: Refactor PointField & TrieField to now have a common base class, NumericFieldType. The
|
||||
TrieField.TrieTypes and PointField.PointTypes are now consolidated to NumericFieldType.NumberType. This
|
||||
refactoring also fixes a bug whereby PointFields were not using DocValues for range queries for
|
||||
indexed=false, docValues=true fields. (Ishan Chattopadhyaya, Tomás Fernández Löbbe)
|
||||
|
||||
================== 6.5.0 ==================
|
||||
|
||||
|
@ -131,10 +139,6 @@ New Features
|
|||
* SOLR-9903: Stop interrupting the update executor on shutdown, it can cause graceful shutdowns to put replicas into Leader
|
||||
Initiated Recovery among other undesirable things. (Mark Miller)
|
||||
|
||||
* SOLR-8396: Add support for PointFields in Solr (Ishan Chattopadhyaya, Tomás Fernández Löbbe)
|
||||
|
||||
* SOLR-9987: Add support for MultiValued DocValues in PointFields using SortedNumericDocValues (Tomás Fernández Löbbe)
|
||||
|
||||
Bug Fixes
|
||||
----------------------
|
||||
|
||||
|
@ -157,8 +161,6 @@ Bug Fixes
|
|||
|
||||
* SOLR-10063: CoreContainer shutdown has race condition that can cause a hang on shutdown. (Mark Miller)
|
||||
|
||||
* SOLR-10104: BlockDirectoryCache release hooks do not work with multiple directories. (Mike Drob, Mark Miller)
|
||||
|
||||
Optimizations
|
||||
----------------------
|
||||
|
||||
|
@ -195,13 +197,6 @@ Other Changes
|
|||
|
||||
* SOLR-10072: The test TestSelectiveWeightCreation appears to be unreliable. (Michael Nilsson via Mark Miller)
|
||||
|
||||
* SOLR-9996: Unstored IntPointField returns Long type (Ishan Chattopadhyaya)
|
||||
|
||||
* SOLR-10011: Refactor PointField & TrieField to now have a common base class, NumericFieldType. The
|
||||
TrieField.TrieTypes and PointField.PointTypes are now consolidated to NumericFieldType.NumberType. This
|
||||
refactoring also fixes a bug whereby PointFields were not using DocValues for range queries for
|
||||
indexed=false, docValues=true fields. (Ishan Chattopadhyaya, Tomás Fernández Löbbe)
|
||||
|
||||
================== 6.4.1 ==================
|
||||
|
||||
Consult the LUCENE_CHANGES.txt file for additional, low level, changes in this release.
|
||||
|
|
|
@ -17,7 +17,7 @@
|
|||
under the License.
|
||||
-->
|
||||
<ivy-module version="2.0">
|
||||
<info organisation="org.apache.solr" module="cell"/>
|
||||
<info organisation="org.apache.solr" module="extraction"/>
|
||||
<configurations defaultconfmapping="compile->master;test->master">
|
||||
<conf name="compile" transitive="false"/>
|
||||
<conf name="test" transitive="false"/>
|
||||
|
|
|
@ -490,6 +490,7 @@ public class Overseer implements Closeable {
|
|||
this.zkController = zkController;
|
||||
this.stats = new Stats();
|
||||
this.config = config;
|
||||
assert ObjectReleaseTracker.track(this);
|
||||
}
|
||||
|
||||
public synchronized void start(String id) {
|
||||
|
@ -520,7 +521,6 @@ public class Overseer implements Closeable {
|
|||
updaterThread.start();
|
||||
ccThread.start();
|
||||
arfoThread.start();
|
||||
assert ObjectReleaseTracker.track(this);
|
||||
}
|
||||
|
||||
public Stats getStats() {
|
||||
|
|
|
@ -688,12 +688,7 @@ public class RealTimeGetComponent extends SearchComponent
|
|||
|
||||
if (sf != null && sf.multiValued()) {
|
||||
List<Object> vals = new ArrayList<>();
|
||||
if (f.fieldType().docValuesType() == DocValuesType.SORTED_NUMERIC) {
|
||||
// SORTED_NUMERICS store sortable bits version of the value, need to retrieve the original
|
||||
vals.add(sf.getType().toObject(f));
|
||||
} else {
|
||||
vals.add( f );
|
||||
}
|
||||
vals.add( f );
|
||||
out.setField( f.name(), vals );
|
||||
}
|
||||
else{
|
||||
|
|
|
@ -1,106 +0,0 @@
|
|||
/*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
package org.apache.solr.handler.component;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.util.Map;
|
||||
|
||||
import org.apache.lucene.index.DocValues;
|
||||
import org.apache.lucene.index.LeafReaderContext;
|
||||
import org.apache.lucene.index.SortedNumericDocValues;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.solr.common.util.NamedList;
|
||||
import org.apache.solr.schema.NumberType;
|
||||
|
||||
public class SortedNumericStatsValues implements StatsValues {
|
||||
|
||||
private final NumericStatsValues nsv;
|
||||
private final String fieldName;
|
||||
private final NumberType numberType;
|
||||
private SortedNumericDocValues sndv;
|
||||
|
||||
|
||||
public SortedNumericStatsValues(NumericStatsValues nsv, StatsField field) {
|
||||
this.nsv = nsv;
|
||||
this.fieldName = field.getSchemaField().getName();
|
||||
this.numberType = field.getSchemaField().getType().getNumberType();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void accumulate(NamedList stv) {
|
||||
nsv.accumulate(stv);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void accumulate(int docId) throws IOException {
|
||||
if (!sndv.advanceExact(docId)) {
|
||||
missing();
|
||||
} else {
|
||||
for (int i = 0 ; i < sndv.docValueCount(); i++) {
|
||||
nsv.accumulate(toCorrectType(sndv.nextValue()), 1);
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
private Number toCorrectType(long value) {
|
||||
switch (numberType) {
|
||||
case INTEGER:
|
||||
case LONG:
|
||||
return value;
|
||||
case FLOAT:
|
||||
return NumericUtils.sortableIntToFloat((int)value);
|
||||
case DOUBLE:
|
||||
return NumericUtils.sortableLongToDouble(value);
|
||||
default:
|
||||
throw new AssertionError("Unsupported number type");
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void accumulate(BytesRef value, int count) {
|
||||
nsv.accumulate(value, count);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void missing() {
|
||||
nsv.missing();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void addMissing(int count) {
|
||||
nsv.addMissing(count);
|
||||
}
|
||||
|
||||
@Override
|
||||
public void addFacet(String facetName, Map<String,StatsValues> facetValues) {
|
||||
nsv.addFacet(facetName, facetValues);
|
||||
}
|
||||
|
||||
@Override
|
||||
public NamedList<?> getStatsValues() {
|
||||
return nsv.getStatsValues();
|
||||
}
|
||||
|
||||
@Override
|
||||
public void setNextReader(LeafReaderContext ctx) throws IOException {
|
||||
sndv = DocValues.getSortedNumeric(ctx.reader(), fieldName);
|
||||
assert sndv != null;
|
||||
}
|
||||
|
||||
}
|
|
@ -416,7 +416,7 @@ public class StatsField {
|
|||
return StatsValuesFactory.createStatsValues(this);
|
||||
}
|
||||
|
||||
if (null != schemaField && !schemaField.getType().isPointField()
|
||||
if (null != schemaField
|
||||
&& (schemaField.multiValued() || schemaField.getType().multiValuedFieldCache())) {
|
||||
|
||||
// TODO: should this also be used for single-valued string fields? (should work fine)
|
||||
|
|
|
@ -66,12 +66,7 @@ public class StatsValuesFactory {
|
|||
if (TrieDateField.class.isInstance(fieldType)) {
|
||||
return new DateStatsValues(statsField);
|
||||
} else if (TrieField.class.isInstance(fieldType) || PointField.class.isInstance(fieldType)) {
|
||||
|
||||
NumericStatsValues statsValue = new NumericStatsValues(statsField);
|
||||
if (sf.multiValued()) {
|
||||
return new SortedNumericStatsValues(statsValue, statsField);
|
||||
}
|
||||
return statsValue;
|
||||
return new NumericStatsValues(statsField);
|
||||
} else if (StrField.class.isInstance(fieldType)) {
|
||||
return new StringStatsValues(statsField);
|
||||
} else if (sf.getType().getClass().equals(EnumField.class)) {
|
||||
|
|
|
@ -30,7 +30,6 @@ import org.apache.lucene.index.FilterNumericDocValues;
|
|||
import org.apache.lucene.index.LeafReaderContext;
|
||||
import org.apache.lucene.index.NumericDocValues;
|
||||
import org.apache.lucene.index.SortedDocValues;
|
||||
import org.apache.lucene.index.SortedNumericDocValues;
|
||||
import org.apache.lucene.index.SortedSetDocValues;
|
||||
import org.apache.lucene.search.DocIdSet;
|
||||
import org.apache.lucene.search.DocIdSetIterator;
|
||||
|
@ -175,12 +174,8 @@ public class IntervalFacets implements Iterable<FacetInterval> {
|
|||
}
|
||||
|
||||
private void doCount() throws IOException {
|
||||
if (schemaField.getType().getNumberType() != null && (!schemaField.multiValued() || schemaField.getType().isPointField())) {
|
||||
if (schemaField.multiValued()) {
|
||||
getCountMultiValuedNumeric();
|
||||
} else {
|
||||
getCountNumeric();
|
||||
}
|
||||
if (schemaField.getType().getNumberType() != null && !schemaField.multiValued()) {
|
||||
getCountNumeric();
|
||||
} else {
|
||||
getCountString();
|
||||
}
|
||||
|
@ -246,36 +241,6 @@ public class IntervalFacets implements Iterable<FacetInterval> {
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
private void getCountMultiValuedNumeric() throws IOException {
|
||||
final FieldType ft = schemaField.getType();
|
||||
final String fieldName = schemaField.getName();
|
||||
if (ft.getNumberType() == null) {
|
||||
throw new IllegalStateException();
|
||||
}
|
||||
final List<LeafReaderContext> leaves = searcher.getIndexReader().leaves();
|
||||
|
||||
final Iterator<LeafReaderContext> ctxIt = leaves.iterator();
|
||||
LeafReaderContext ctx = null;
|
||||
SortedNumericDocValues longs = null;
|
||||
for (DocIterator docsIt = docs.iterator(); docsIt.hasNext(); ) {
|
||||
final int doc = docsIt.nextDoc();
|
||||
if (ctx == null || doc >= ctx.docBase + ctx.reader().maxDoc()) {
|
||||
do {
|
||||
ctx = ctxIt.next();
|
||||
} while (ctx == null || doc >= ctx.docBase + ctx.reader().maxDoc());
|
||||
assert doc >= ctx.docBase;
|
||||
longs = DocValues.getSortedNumeric(ctx.reader(), fieldName);
|
||||
}
|
||||
int valuesDocID = longs.docID();
|
||||
if (valuesDocID < doc - ctx.docBase) {
|
||||
valuesDocID = longs.advance(doc - ctx.docBase);
|
||||
}
|
||||
if (valuesDocID == doc - ctx.docBase) {
|
||||
accumIntervalWithMultipleValues(longs);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private void getCountString() throws IOException {
|
||||
Filter filter = docs.getTopFilter();
|
||||
|
@ -311,44 +276,6 @@ public class IntervalFacets implements Iterable<FacetInterval> {
|
|||
}
|
||||
}
|
||||
|
||||
private void accumIntervalWithMultipleValues(SortedNumericDocValues longs) throws IOException {
|
||||
// longs should be already positioned to the correct doc
|
||||
assert longs.docID() != -1;
|
||||
assert longs.docValueCount() > 0: "Should have at least one value for this document";
|
||||
int currentInterval = 0;
|
||||
for (int i = 0; i < longs.docValueCount(); i++) {
|
||||
boolean evaluateNextInterval = true;
|
||||
long value = longs.nextValue();
|
||||
while (evaluateNextInterval && currentInterval < intervals.length) {
|
||||
IntervalCompareResult result = intervals[currentInterval].includes(value);
|
||||
switch (result) {
|
||||
case INCLUDED:
|
||||
/*
|
||||
* Increment the current interval and move to the next one using
|
||||
* the same value
|
||||
*/
|
||||
intervals[currentInterval].incCount();
|
||||
currentInterval++;
|
||||
break;
|
||||
case LOWER_THAN_START:
|
||||
/*
|
||||
* None of the next intervals will match this value (all of them have
|
||||
* higher start value). Move to the next value for this document.
|
||||
*/
|
||||
evaluateNextInterval = false;
|
||||
break;
|
||||
case GREATER_THAN_END:
|
||||
/*
|
||||
* Next interval may match this value
|
||||
*/
|
||||
currentInterval++;
|
||||
break;
|
||||
}
|
||||
//Maybe return if currentInterval == intervals.length?
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
private void accumIntervalsMulti(SortedSetDocValues ssdv,
|
||||
DocIdSetIterator disi, Bits bits) throws IOException {
|
||||
// First update the ordinals in the intervals for this segment
|
||||
|
|
|
@ -32,14 +32,12 @@ import org.apache.lucene.index.FilterNumericDocValues;
|
|||
import org.apache.lucene.index.LeafReaderContext;
|
||||
import org.apache.lucene.index.NumericDocValues;
|
||||
import org.apache.lucene.index.ReaderUtil;
|
||||
import org.apache.lucene.index.SortedNumericDocValues;
|
||||
import org.apache.lucene.index.Terms;
|
||||
import org.apache.lucene.index.TermsEnum;
|
||||
import org.apache.lucene.queries.function.FunctionValues;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.CharsRefBuilder;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.lucene.util.PriorityQueue;
|
||||
import org.apache.lucene.util.StringHelper;
|
||||
import org.apache.solr.common.params.FacetParams;
|
||||
|
@ -63,18 +61,16 @@ final class NumericFacets {
|
|||
|
||||
long[] bits; // bits identifying a value
|
||||
int[] counts;
|
||||
int[] docIDs; //Will be null if HashTable is created with needsDocId=false
|
||||
int[] docIDs;
|
||||
int mask;
|
||||
int size;
|
||||
int threshold;
|
||||
|
||||
HashTable(boolean needsDocId) {
|
||||
HashTable() {
|
||||
final int capacity = 64; // must be a power of 2
|
||||
bits = new long[capacity];
|
||||
counts = new int[capacity];
|
||||
if (needsDocId) {
|
||||
docIDs = new int[capacity];
|
||||
}
|
||||
docIDs = new int[capacity];
|
||||
mask = capacity - 1;
|
||||
size = 0;
|
||||
threshold = (int) (capacity * LOAD_FACTOR);
|
||||
|
@ -103,23 +99,6 @@ final class NumericFacets {
|
|||
break;
|
||||
}
|
||||
}
|
||||
|
||||
void add(long value, int count) {
|
||||
if (size >= threshold) {
|
||||
rehash();
|
||||
}
|
||||
final int h = hash(value);
|
||||
for (int slot = h; ; slot = (slot + 1) & mask) {
|
||||
if (counts[slot] == 0) {
|
||||
bits[slot] = value;
|
||||
++size;
|
||||
} else if (bits[slot] != value) {
|
||||
continue;
|
||||
}
|
||||
counts[slot] += count;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
private void rehash() {
|
||||
final long[] oldBits = bits;
|
||||
|
@ -129,24 +108,14 @@ final class NumericFacets {
|
|||
final int newCapacity = bits.length * 2;
|
||||
bits = new long[newCapacity];
|
||||
counts = new int[newCapacity];
|
||||
if (oldDocIDs!= null) {
|
||||
docIDs = new int[newCapacity];
|
||||
}
|
||||
docIDs = new int[newCapacity];
|
||||
mask = newCapacity - 1;
|
||||
threshold = (int) (LOAD_FACTOR * newCapacity);
|
||||
size = 0;
|
||||
|
||||
if (oldDocIDs!= null) {
|
||||
for (int i = 0; i < oldBits.length; ++i) {
|
||||
if (oldCounts[i] > 0) {
|
||||
add(oldDocIDs[i], oldBits[i], oldCounts[i]);
|
||||
}
|
||||
}
|
||||
} else {
|
||||
for (int i = 0; i < oldBits.length; ++i) {
|
||||
if (oldCounts[i] > 0) {
|
||||
add(oldBits[i], oldCounts[i]);
|
||||
}
|
||||
for (int i = 0; i < oldBits.length; ++i) {
|
||||
if (oldCounts[i] > 0) {
|
||||
add(oldDocIDs[i], oldBits[i], oldCounts[i]);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -160,16 +129,7 @@ final class NumericFacets {
|
|||
}
|
||||
|
||||
public static NamedList<Integer> getCounts(SolrIndexSearcher searcher, DocSet docs, String fieldName, int offset, int limit, int mincount, boolean missing, String sort) throws IOException {
|
||||
final SchemaField sf = searcher.getSchema().getField(fieldName);
|
||||
if (sf.multiValued()) {
|
||||
// TODO: evaluate using getCountsMultiValued for singleValued numerics with SingletonSortedNumericDocValues
|
||||
return getCountsMultiValued(searcher, docs, fieldName, offset, limit, mincount, missing, sort);
|
||||
}
|
||||
return getCountsSingleValue(searcher, docs, fieldName, offset, limit, mincount, missing, sort);
|
||||
}
|
||||
|
||||
private static NamedList<Integer> getCountsSingleValue(SolrIndexSearcher searcher, DocSet docs, String fieldName, int offset, int limit, int mincount, boolean missing, String sort) throws IOException {
|
||||
boolean zeros = mincount <= 0;
|
||||
final boolean zeros = mincount <= 0;
|
||||
mincount = Math.max(mincount, 1);
|
||||
final SchemaField sf = searcher.getSchema().getField(fieldName);
|
||||
final FieldType ft = sf.getType();
|
||||
|
@ -177,11 +137,10 @@ final class NumericFacets {
|
|||
if (numericType == null) {
|
||||
throw new IllegalStateException();
|
||||
}
|
||||
zeros = zeros && !ft.isPointField() && sf.indexed(); // We don't return zeros when using PointFields or when index=false
|
||||
final List<LeafReaderContext> leaves = searcher.getIndexReader().leaves();
|
||||
|
||||
// 1. accumulate
|
||||
final HashTable hashTable = new HashTable(true);
|
||||
final HashTable hashTable = new HashTable();
|
||||
final Iterator<LeafReaderContext> ctxIt = leaves.iterator();
|
||||
LeafReaderContext ctx = null;
|
||||
NumericDocValues longs = null;
|
||||
|
@ -404,118 +363,4 @@ final class NumericFacets {
|
|||
return result;
|
||||
}
|
||||
|
||||
private static NamedList<Integer> getCountsMultiValued(SolrIndexSearcher searcher, DocSet docs, String fieldName, int offset, int limit, int mincount, boolean missing, String sort) throws IOException {
|
||||
// If facet.mincount=0 with PointFields the only option is to get the values from DocValues
|
||||
// not currently supported. See SOLR-10033
|
||||
mincount = Math.max(mincount, 1);
|
||||
final SchemaField sf = searcher.getSchema().getField(fieldName);
|
||||
final FieldType ft = sf.getType();
|
||||
assert sf.multiValued();
|
||||
final List<LeafReaderContext> leaves = searcher.getIndexReader().leaves();
|
||||
|
||||
// 1. accumulate
|
||||
final HashTable hashTable = new HashTable(false);
|
||||
final Iterator<LeafReaderContext> ctxIt = leaves.iterator();
|
||||
LeafReaderContext ctx = null;
|
||||
SortedNumericDocValues longs = null;
|
||||
int missingCount = 0;
|
||||
for (DocIterator docsIt = docs.iterator(); docsIt.hasNext(); ) {
|
||||
final int doc = docsIt.nextDoc();
|
||||
if (ctx == null || doc >= ctx.docBase + ctx.reader().maxDoc()) {
|
||||
do {
|
||||
ctx = ctxIt.next();
|
||||
} while (ctx == null || doc >= ctx.docBase + ctx.reader().maxDoc());
|
||||
assert doc >= ctx.docBase;
|
||||
longs = DocValues.getSortedNumeric(ctx.reader(), fieldName);
|
||||
}
|
||||
int valuesDocID = longs.docID();
|
||||
if (valuesDocID < doc - ctx.docBase) {
|
||||
valuesDocID = longs.advance(doc - ctx.docBase);
|
||||
}
|
||||
if (valuesDocID == doc - ctx.docBase) {
|
||||
long l = longs.nextValue(); // This document must have at least one value
|
||||
hashTable.add(l, 1);
|
||||
for (int i = 1; i < longs.docValueCount(); i++) {
|
||||
long lnew = longs.nextValue();
|
||||
if (lnew > l) { // Skip the value if it's equal to the last one, we don't want to double-count it
|
||||
hashTable.add(lnew, 1);
|
||||
}
|
||||
l = lnew;
|
||||
}
|
||||
|
||||
} else {
|
||||
++missingCount;
|
||||
}
|
||||
}
|
||||
|
||||
// 2. select top-k facet values
|
||||
final int pqSize = limit < 0 ? hashTable.size : Math.min(offset + limit, hashTable.size);
|
||||
final PriorityQueue<Entry> pq;
|
||||
if (FacetParams.FACET_SORT_COUNT.equals(sort) || FacetParams.FACET_SORT_COUNT_LEGACY.equals(sort)) {
|
||||
pq = new PriorityQueue<Entry>(pqSize) {
|
||||
@Override
|
||||
protected boolean lessThan(Entry a, Entry b) {
|
||||
if (a.count < b.count || (a.count == b.count && a.bits > b.bits)) {
|
||||
return true;
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
};
|
||||
} else {
|
||||
// sort=index
|
||||
pq = new PriorityQueue<Entry>(pqSize) {
|
||||
@Override
|
||||
protected boolean lessThan(Entry a, Entry b) {
|
||||
return a.bits > b.bits;
|
||||
}
|
||||
};
|
||||
}
|
||||
Entry e = null;
|
||||
for (int i = 0; i < hashTable.bits.length; ++i) {
|
||||
if (hashTable.counts[i] >= mincount) {
|
||||
if (e == null) {
|
||||
e = new Entry();
|
||||
}
|
||||
e.bits = hashTable.bits[i];
|
||||
e.count = hashTable.counts[i];
|
||||
e = pq.insertWithOverflow(e);
|
||||
}
|
||||
}
|
||||
|
||||
// 4. build the NamedList
|
||||
final NamedList<Integer> result = new NamedList<>(Math.max(pq.size() - offset + 1, 1));
|
||||
final Deque<Entry> counts = new ArrayDeque<>(pq.size() - offset);
|
||||
while (pq.size() > offset) {
|
||||
counts.addFirst(pq.pop());
|
||||
}
|
||||
|
||||
for (Entry entry : counts) {
|
||||
result.add(bitsToStringValue(ft, entry.bits), entry.count); // TODO: convert to correct value
|
||||
}
|
||||
|
||||
// Once facet.mincount=0 is supported we'll need to add logic similar to the SingleValue case, but obtaining values
|
||||
// with count 0 from DocValues
|
||||
|
||||
if (missing) {
|
||||
result.add(null, missingCount);
|
||||
}
|
||||
return result;
|
||||
}
|
||||
|
||||
private static String bitsToStringValue(FieldType fieldType, long bits) {
|
||||
switch (fieldType.getNumberType()) {
|
||||
case LONG:
|
||||
case INTEGER:
|
||||
return String.valueOf(bits);
|
||||
case FLOAT:
|
||||
return String.valueOf(NumericUtils.sortableIntToFloat((int)bits));
|
||||
case DOUBLE:
|
||||
return String.valueOf(NumericUtils.sortableLongToDouble(bits));
|
||||
//TODO: DATE
|
||||
default:
|
||||
throw new AssertionError("Unsupported NumberType: " + fieldType.getNumberType());
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -482,8 +482,8 @@ public class SimpleFacets {
|
|||
counts = getFacetTermEnumCounts(searcher, docs, field, offset, limit, mincount,missing,sort,prefix, termFilter, exists);
|
||||
break;
|
||||
case FCS:
|
||||
assert ft.isPointField() || !multiToken;
|
||||
if (ft.isPointField() || (ft.getNumberType() != null && !sf.multiValued())) {
|
||||
assert !multiToken;
|
||||
if (ft.getNumberType() != null && !sf.multiValued()) {
|
||||
// force numeric faceting
|
||||
if (prefix != null && !prefix.isEmpty()) {
|
||||
throw new SolrException(ErrorCode.BAD_REQUEST, FacetParams.FACET_PREFIX + " is not supported on numeric types");
|
||||
|
@ -494,10 +494,6 @@ public class SimpleFacets {
|
|||
throw new SolrException(ErrorCode.BAD_REQUEST, FacetParams.FACET_CONTAINS + " is not supported on numeric types");
|
||||
}
|
||||
}
|
||||
// We should do this, but mincount=0 is currently the default
|
||||
// if (ft.isPointField() && mincount <= 0) {
|
||||
// throw new SolrException(ErrorCode.BAD_REQUEST, FacetParams.FACET_MINCOUNT + " <= 0 is not supported on point types");
|
||||
// }
|
||||
counts = NumericFacets.getCounts(searcher, docs, field, offset, limit, mincount, missing, sort);
|
||||
} else {
|
||||
PerSegmentSingleValuedFaceting ps = new PerSegmentSingleValuedFaceting(searcher, docs, field, offset, limit, mincount, missing, sort, prefix, termFilter);
|
||||
|
|
|
@ -27,13 +27,10 @@ import org.apache.lucene.index.IndexableField;
|
|||
import org.apache.lucene.legacy.LegacyNumericType;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.queries.function.valuesource.DoubleFieldSource;
|
||||
import org.apache.lucene.queries.function.valuesource.MultiValuedDoubleFieldSource;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.SortField;
|
||||
import org.apache.lucene.search.SortedNumericSelector;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.BytesRefBuilder;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.solr.search.QParser;
|
||||
import org.apache.solr.uninverting.UninvertingReader.Type;
|
||||
import org.slf4j.Logger;
|
||||
|
@ -94,8 +91,6 @@ public class DoublePointField extends PointField implements DoubleValueFieldType
|
|||
if (val != null) {
|
||||
if (f.fieldType().stored() == false && f.fieldType().docValuesType() == DocValuesType.NUMERIC) {
|
||||
return Double.longBitsToDouble(val.longValue());
|
||||
} else if (f.fieldType().stored() == false && f.fieldType().docValuesType() == DocValuesType.SORTED_NUMERIC) {
|
||||
return NumericUtils.sortableLongToDouble(val.longValue());
|
||||
} else {
|
||||
return val;
|
||||
}
|
||||
|
@ -154,7 +149,8 @@ public class DoublePointField extends PointField implements DoubleValueFieldType
|
|||
@Override
|
||||
public Type getUninversionType(SchemaField sf) {
|
||||
if (sf.multiValued()) {
|
||||
return Type.SORTED_DOUBLE;
|
||||
throw new UnsupportedOperationException("MultiValued Point fields with DocValues is not currently supported");
|
||||
// return Type.SORTED_DOUBLE;
|
||||
} else {
|
||||
return Type.DOUBLE_POINT;
|
||||
}
|
||||
|
@ -165,11 +161,6 @@ public class DoublePointField extends PointField implements DoubleValueFieldType
|
|||
field.checkFieldCacheSource();
|
||||
return new DoubleFieldSource(field.getName());
|
||||
}
|
||||
|
||||
@Override
|
||||
protected ValueSource getSingleValueSource(SortedNumericSelector.Type choice, SchemaField f) {
|
||||
return new MultiValuedDoubleFieldSource(f.getName(), choice);
|
||||
}
|
||||
|
||||
@Override
|
||||
public LegacyNumericType getNumericType() {
|
||||
|
|
|
@ -626,7 +626,7 @@ public abstract class FieldType extends FieldProperties {
|
|||
|
||||
/**
|
||||
* Return the numeric type of this field, or null if this field is not a
|
||||
* numeric field.
|
||||
* numeric field.
|
||||
*/
|
||||
public NumberType getNumberType() {
|
||||
return null;
|
||||
|
|
|
@ -27,13 +27,10 @@ import org.apache.lucene.index.IndexableField;
|
|||
import org.apache.lucene.legacy.LegacyNumericType;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.queries.function.valuesource.FloatFieldSource;
|
||||
import org.apache.lucene.queries.function.valuesource.MultiValuedFloatFieldSource;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.SortField;
|
||||
import org.apache.lucene.search.SortedNumericSelector;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.BytesRefBuilder;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.solr.search.QParser;
|
||||
import org.apache.solr.uninverting.UninvertingReader.Type;
|
||||
import org.slf4j.Logger;
|
||||
|
@ -94,9 +91,7 @@ public class FloatPointField extends PointField implements FloatValueFieldType {
|
|||
if (val != null) {
|
||||
if (f.fieldType().stored() == false && f.fieldType().docValuesType() == DocValuesType.NUMERIC) {
|
||||
return Float.intBitsToFloat(val.intValue());
|
||||
} else if (f.fieldType().stored() == false && f.fieldType().docValuesType() == DocValuesType.SORTED_NUMERIC) {
|
||||
return NumericUtils.sortableIntToFloat(val.intValue());
|
||||
} else {
|
||||
} else {
|
||||
return val;
|
||||
}
|
||||
} else {
|
||||
|
@ -154,7 +149,8 @@ public class FloatPointField extends PointField implements FloatValueFieldType {
|
|||
@Override
|
||||
public Type getUninversionType(SchemaField sf) {
|
||||
if (sf.multiValued()) {
|
||||
return Type.SORTED_FLOAT;
|
||||
throw new UnsupportedOperationException("MultiValued Point fields with DocValues is not currently supported");
|
||||
// return Type.SORTED_FLOAT;
|
||||
} else {
|
||||
return Type.FLOAT_POINT;
|
||||
}
|
||||
|
@ -165,12 +161,6 @@ public class FloatPointField extends PointField implements FloatValueFieldType {
|
|||
field.checkFieldCacheSource();
|
||||
return new FloatFieldSource(field.getName());
|
||||
}
|
||||
|
||||
@Override
|
||||
protected ValueSource getSingleValueSource(SortedNumericSelector.Type choice, SchemaField f) {
|
||||
return new MultiValuedFloatFieldSource(f.getName(), choice);
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public LegacyNumericType getNumericType() {
|
||||
|
|
|
@ -26,10 +26,8 @@ import org.apache.lucene.index.IndexableField;
|
|||
import org.apache.lucene.legacy.LegacyNumericType;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.queries.function.valuesource.IntFieldSource;
|
||||
import org.apache.lucene.queries.function.valuesource.MultiValuedIntFieldSource;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.SortField;
|
||||
import org.apache.lucene.search.SortedNumericSelector;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.BytesRefBuilder;
|
||||
import org.apache.solr.search.QParser;
|
||||
|
@ -151,7 +149,8 @@ public class IntPointField extends PointField implements IntValueFieldType {
|
|||
@Override
|
||||
public Type getUninversionType(SchemaField sf) {
|
||||
if (sf.multiValued()) {
|
||||
return Type.SORTED_INTEGER;
|
||||
throw new UnsupportedOperationException("MultiValued Point fields with DocValues is not currently supported");
|
||||
// return Type.SORTED_INTEGER;
|
||||
} else {
|
||||
return Type.INTEGER_POINT;
|
||||
}
|
||||
|
@ -183,10 +182,5 @@ public class IntPointField extends PointField implements IntValueFieldType {
|
|||
protected StoredField getStoredField(SchemaField sf, Object value) {
|
||||
return new StoredField(sf.getName(), (Integer) this.toNativeType(value));
|
||||
}
|
||||
|
||||
@Override
|
||||
protected ValueSource getSingleValueSource(SortedNumericSelector.Type choice, SchemaField f) {
|
||||
return new MultiValuedIntFieldSource(f.getName(), choice);
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -26,7 +26,6 @@ import org.apache.lucene.index.IndexableField;
|
|||
import org.apache.lucene.legacy.LegacyNumericType;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.queries.function.valuesource.LongFieldSource;
|
||||
import org.apache.lucene.queries.function.valuesource.MultiValuedLongFieldSource;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.SortField;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
|
@ -150,7 +149,8 @@ public class LongPointField extends PointField implements LongValueFieldType {
|
|||
@Override
|
||||
public Type getUninversionType(SchemaField sf) {
|
||||
if (sf.multiValued()) {
|
||||
return Type.SORTED_LONG;
|
||||
throw new UnsupportedOperationException("MultiValued Point fields with DocValues is not currently supported");
|
||||
// return Type.SORTED_LONG;
|
||||
} else {
|
||||
return Type.LONG_POINT;
|
||||
}
|
||||
|
@ -161,12 +161,6 @@ public class LongPointField extends PointField implements LongValueFieldType {
|
|||
field.checkFieldCacheSource();
|
||||
return new LongFieldSource(field.getName());
|
||||
}
|
||||
|
||||
@Override
|
||||
protected ValueSource getSingleValueSource(org.apache.lucene.search.SortedNumericSelector.Type choice,
|
||||
SchemaField field) {
|
||||
return new MultiValuedLongFieldSource(field.getName(), choice);
|
||||
}
|
||||
|
||||
@Override
|
||||
public LegacyNumericType getNumericType() {
|
||||
|
|
|
@ -17,11 +17,9 @@
|
|||
package org.apache.solr.schema;
|
||||
|
||||
import org.apache.lucene.document.NumericDocValuesField;
|
||||
import org.apache.lucene.document.SortedNumericDocValuesField;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.search.MatchNoDocsQuery;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.solr.common.SolrException;
|
||||
import org.apache.solr.search.FunctionRangeQuery;
|
||||
import org.apache.solr.search.QParser;
|
||||
|
@ -51,36 +49,28 @@ public abstract class NumericFieldType extends PrimitiveFieldType {
|
|||
|
||||
protected Query getDocValuesRangeQuery(QParser parser, SchemaField field, String min, String max,
|
||||
boolean minInclusive, boolean maxInclusive) {
|
||||
assert field.hasDocValues() && (field.getType().isPointField() || !field.multiValued());
|
||||
assert field.hasDocValues() && !field.multiValued();
|
||||
|
||||
switch (getNumberType()) {
|
||||
case INTEGER:
|
||||
return numericDocValuesRangeQuery(field.getName(),
|
||||
min == null ? null : (long) Integer.parseInt(min),
|
||||
max == null ? null : (long) Integer.parseInt(max),
|
||||
minInclusive, maxInclusive, field.multiValued());
|
||||
minInclusive, maxInclusive);
|
||||
case FLOAT:
|
||||
if (field.multiValued()) {
|
||||
return getRangeQueryForMultiValuedFloatDocValues(field, min, max, minInclusive, maxInclusive);
|
||||
} else {
|
||||
return getRangeQueryForFloatDoubleDocValues(field, min, max, minInclusive, maxInclusive);
|
||||
}
|
||||
return getRangeQueryForFloatDoubleDocValues(field, min, max, minInclusive, maxInclusive);
|
||||
case LONG:
|
||||
return numericDocValuesRangeQuery(field.getName(),
|
||||
min == null ? null : Long.parseLong(min),
|
||||
max == null ? null : Long.parseLong(max),
|
||||
minInclusive, maxInclusive, field.multiValued());
|
||||
minInclusive, maxInclusive);
|
||||
case DOUBLE:
|
||||
if (field.multiValued()) {
|
||||
return getRangeQueryForMultiValuedDoubleDocValues(field, min, max, minInclusive, maxInclusive);
|
||||
} else {
|
||||
return getRangeQueryForFloatDoubleDocValues(field, min, max, minInclusive, maxInclusive);
|
||||
}
|
||||
return getRangeQueryForFloatDoubleDocValues(field, min, max, minInclusive, maxInclusive);
|
||||
case DATE:
|
||||
return numericDocValuesRangeQuery(field.getName(),
|
||||
min == null ? null : DateMathParser.parseMath(null, min).getTime(),
|
||||
max == null ? null : DateMathParser.parseMath(null, max).getTime(),
|
||||
minInclusive, maxInclusive, field.multiValued());
|
||||
minInclusive, maxInclusive);
|
||||
default:
|
||||
throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, "Unknown type for numeric field");
|
||||
}
|
||||
|
@ -114,32 +104,19 @@ public abstract class NumericFieldType extends PrimitiveFieldType {
|
|||
if ((minVal == null || minVal.doubleValue() < 0d || minBits == minusZeroBits) &&
|
||||
(maxVal != null && (maxVal.doubleValue() < 0d || maxBits == minusZeroBits))) {
|
||||
query = numericDocValuesRangeQuery
|
||||
(fieldName, maxBits, (min == null ? negativeInfinityBits : minBits), maxInclusive, minInclusive, false);
|
||||
(fieldName, maxBits, (min == null ? negativeInfinityBits : minBits), maxInclusive, minInclusive);
|
||||
} else { // If both max and min are positive, then issue range query
|
||||
query = numericDocValuesRangeQuery
|
||||
(fieldName, minBits, (max == null ? positiveInfinityBits : maxBits), minInclusive, maxInclusive, false);
|
||||
(fieldName, minBits, (max == null ? positiveInfinityBits : maxBits), minInclusive, maxInclusive);
|
||||
}
|
||||
}
|
||||
return query;
|
||||
}
|
||||
|
||||
protected Query getRangeQueryForMultiValuedDoubleDocValues(SchemaField sf, String min, String max, boolean minInclusive, boolean maxInclusive) {
|
||||
Long minBits = min == null ? NumericUtils.doubleToSortableLong(Double.NEGATIVE_INFINITY): NumericUtils.doubleToSortableLong(Double.parseDouble(min));
|
||||
Long maxBits = max == null ? NumericUtils.doubleToSortableLong(Double.POSITIVE_INFINITY): NumericUtils.doubleToSortableLong(Double.parseDouble(max));
|
||||
return numericDocValuesRangeQuery(sf.getName(), minBits, maxBits, minInclusive, maxInclusive, true);
|
||||
}
|
||||
|
||||
protected Query getRangeQueryForMultiValuedFloatDocValues(SchemaField sf, String min, String max, boolean minInclusive, boolean maxInclusive) {
|
||||
Long minBits = (long)(min == null ? NumericUtils.floatToSortableInt(Float.NEGATIVE_INFINITY): NumericUtils.floatToSortableInt(Float.parseFloat(min)));
|
||||
Long maxBits = (long)(max == null ? NumericUtils.floatToSortableInt(Float.POSITIVE_INFINITY): NumericUtils.floatToSortableInt(Float.parseFloat(max)));
|
||||
return numericDocValuesRangeQuery(sf.getName(), minBits, maxBits, minInclusive, maxInclusive, true);
|
||||
}
|
||||
|
||||
public static Query numericDocValuesRangeQuery(
|
||||
String field,
|
||||
Number lowerValue, Number upperValue,
|
||||
boolean lowerInclusive, boolean upperInclusive,
|
||||
boolean multiValued) {
|
||||
boolean lowerInclusive, boolean upperInclusive) {
|
||||
|
||||
long actualLowerValue = Long.MIN_VALUE;
|
||||
if (lowerValue != null) {
|
||||
|
@ -162,11 +139,6 @@ public abstract class NumericFieldType extends PrimitiveFieldType {
|
|||
--actualUpperValue;
|
||||
}
|
||||
}
|
||||
if (multiValued) {
|
||||
// In multiValued case use SortedNumericDocValuesField, this won't work for Trie*Fields wince they use BinaryDV in the multiValue case
|
||||
return SortedNumericDocValuesField.newRangeQuery(field, actualLowerValue, actualUpperValue);
|
||||
} else {
|
||||
return NumericDocValuesField.newRangeQuery(field, actualLowerValue, actualUpperValue);
|
||||
}
|
||||
return NumericDocValuesField.newRangeQuery(field, actualLowerValue, actualUpperValue);
|
||||
}
|
||||
}
|
||||
|
|
|
@ -24,17 +24,15 @@ import java.util.Collections;
|
|||
import java.util.List;
|
||||
|
||||
import org.apache.lucene.document.NumericDocValuesField;
|
||||
import org.apache.lucene.document.SortedNumericDocValuesField;
|
||||
import org.apache.lucene.document.StoredField;
|
||||
import org.apache.lucene.index.IndexableField;
|
||||
import org.apache.lucene.queries.function.ValueSource;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.SortedNumericSelector;
|
||||
import org.apache.lucene.search.SortedSetSelector;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.BytesRefBuilder;
|
||||
import org.apache.lucene.util.CharsRef;
|
||||
import org.apache.lucene.util.CharsRefBuilder;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.solr.common.SolrException;
|
||||
import org.apache.solr.response.TextResponseWriter;
|
||||
import org.apache.solr.search.QParser;
|
||||
|
@ -77,7 +75,7 @@ public abstract class PointField extends NumericFieldType {
|
|||
|
||||
// multivalued Point fields all use SortedSetDocValues, so we give a clean error if that's
|
||||
// not supported by the specified choice, else we delegate to a helper
|
||||
SortedNumericSelector.Type selectorType = choice.getSortedNumericSelectorType();
|
||||
SortedSetSelector.Type selectorType = choice.getSortedSetSelectorType();
|
||||
if (null == selectorType) {
|
||||
throw new SolrException(SolrException.ErrorCode.BAD_REQUEST,
|
||||
choice.toString() + " is not a supported option for picking a single value"
|
||||
|
@ -97,7 +95,9 @@ public abstract class PointField extends NumericFieldType {
|
|||
* @param field the field to use, guaranteed to be multivalued.
|
||||
* @see #getSingleValueSource(MultiValueSelector,SchemaField,QParser)
|
||||
*/
|
||||
protected abstract ValueSource getSingleValueSource(SortedNumericSelector.Type choice, SchemaField field);
|
||||
protected ValueSource getSingleValueSource(SortedSetSelector.Type choice, SchemaField field) {
|
||||
throw new UnsupportedOperationException("MultiValued Point fields with DocValues is not currently supported");
|
||||
}
|
||||
|
||||
@Override
|
||||
public boolean isTokenized() {
|
||||
|
@ -130,7 +130,7 @@ public abstract class PointField extends NumericFieldType {
|
|||
@Override
|
||||
public Query getRangeQuery(QParser parser, SchemaField field, String min, String max, boolean minInclusive,
|
||||
boolean maxInclusive) {
|
||||
if (!field.indexed() && field.hasDocValues()) {
|
||||
if (!field.indexed() && field.hasDocValues() && !field.multiValued()) {
|
||||
return getDocValuesRangeQuery(parser, field, min, max, minInclusive, maxInclusive);
|
||||
} else {
|
||||
return getPointRangeQuery(parser, field, min, max, minInclusive, maxInclusive);
|
||||
|
@ -203,8 +203,10 @@ public abstract class PointField extends NumericFieldType {
|
|||
fields.add(field);
|
||||
|
||||
if (sf.hasDocValues()) {
|
||||
final long bits;
|
||||
if (!sf.multiValued()) {
|
||||
if (sf.multiValued()) {
|
||||
throw new UnsupportedOperationException("MultiValued Point fields with DocValues is not currently supported. Field: '" + sf.getName() + "'");
|
||||
} else {
|
||||
final long bits;
|
||||
if (field.numericValue() instanceof Integer || field.numericValue() instanceof Long) {
|
||||
bits = field.numericValue().longValue();
|
||||
} else if (field.numericValue() instanceof Float) {
|
||||
|
@ -214,19 +216,8 @@ public abstract class PointField extends NumericFieldType {
|
|||
bits = Double.doubleToLongBits(field.numericValue().doubleValue());
|
||||
}
|
||||
fields.add(new NumericDocValuesField(sf.getName(), bits));
|
||||
} else {
|
||||
// MultiValued
|
||||
if (field.numericValue() instanceof Integer || field.numericValue() instanceof Long) {
|
||||
bits = field.numericValue().longValue();
|
||||
} else if (field.numericValue() instanceof Float) {
|
||||
bits = NumericUtils.floatToSortableInt(field.numericValue().floatValue());
|
||||
} else {
|
||||
assert field.numericValue() instanceof Double;
|
||||
bits = NumericUtils.doubleToSortableLong(field.numericValue().doubleValue());
|
||||
}
|
||||
fields.add(new SortedNumericDocValuesField(sf.getName(), bits));
|
||||
}
|
||||
}
|
||||
}
|
||||
if (sf.stored()) {
|
||||
fields.add(getStoredField(sf, value));
|
||||
}
|
||||
|
|
|
@ -56,7 +56,6 @@ import org.apache.lucene.index.NumericDocValues;
|
|||
import org.apache.lucene.index.PostingsEnum;
|
||||
import org.apache.lucene.index.ReaderUtil;
|
||||
import org.apache.lucene.index.SortedDocValues;
|
||||
import org.apache.lucene.index.SortedNumericDocValues;
|
||||
import org.apache.lucene.index.SortedSetDocValues;
|
||||
import org.apache.lucene.index.StoredFieldVisitor;
|
||||
import org.apache.lucene.index.StoredFieldVisitor.Status;
|
||||
|
@ -99,7 +98,6 @@ import org.apache.lucene.store.Directory;
|
|||
import org.apache.lucene.util.Bits;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
import org.apache.lucene.util.FixedBitSet;
|
||||
import org.apache.lucene.util.NumericUtils;
|
||||
import org.apache.solr.common.SolrDocumentBase;
|
||||
import org.apache.solr.common.SolrException;
|
||||
import org.apache.solr.common.SolrException.ErrorCode;
|
||||
|
@ -818,136 +816,110 @@ public class SolrIndexSearcher extends IndexSearcher implements Closeable, SolrI
|
|||
log.warn("Couldn't decorate docValues for field: [{}], schemaField: [{}]", fieldName, schemaField);
|
||||
continue;
|
||||
}
|
||||
FieldInfo fi = fieldInfos.fieldInfo(fieldName);
|
||||
if (fi == null) {
|
||||
continue; // Searcher doesn't have info about this field, hence ignore it.
|
||||
}
|
||||
final DocValuesType dvType = fi.getDocValuesType();
|
||||
switch (dvType) {
|
||||
case NUMERIC:
|
||||
final NumericDocValues ndv = leafReader.getNumericDocValues(fieldName);
|
||||
if (ndv == null) {
|
||||
continue;
|
||||
}
|
||||
Long val;
|
||||
if (ndv.advanceExact(localId)) {
|
||||
val = ndv.longValue();
|
||||
} else {
|
||||
continue;
|
||||
}
|
||||
Object newVal = val;
|
||||
if (schemaField.getType().isPointField()) {
|
||||
// TODO: Maybe merge PointField with TrieFields here
|
||||
NumberType type = schemaField.getType().getNumberType();
|
||||
switch (type) {
|
||||
case INTEGER:
|
||||
newVal = val.intValue();
|
||||
break;
|
||||
case LONG:
|
||||
newVal = val.longValue();
|
||||
break;
|
||||
case FLOAT:
|
||||
newVal = Float.intBitsToFloat(val.intValue());
|
||||
break;
|
||||
case DOUBLE:
|
||||
newVal = Double.longBitsToDouble(val);
|
||||
break;
|
||||
case DATE:
|
||||
newVal = new Date(val);
|
||||
break;
|
||||
default:
|
||||
throw new AssertionError("Unexpected PointType: " + type);
|
||||
|
||||
if (schemaField.multiValued()) {
|
||||
final SortedSetDocValues values = leafReader.getSortedSetDocValues(fieldName);
|
||||
if (values != null && values.getValueCount() > 0) {
|
||||
if (values.advance(localId) == localId) {
|
||||
final List<Object> outValues = new LinkedList<Object>();
|
||||
for (long ord = values.nextOrd(); ord != SortedSetDocValues.NO_MORE_ORDS; ord = values.nextOrd()) {
|
||||
final BytesRef value = values.lookupOrd(ord);
|
||||
outValues.add(schemaField.getType().toObject(schemaField, value));
|
||||
}
|
||||
} else {
|
||||
if (schemaField.getType() instanceof TrieIntField) {
|
||||
newVal = val.intValue();
|
||||
} else if (schemaField.getType() instanceof TrieFloatField) {
|
||||
newVal = Float.intBitsToFloat(val.intValue());
|
||||
} else if (schemaField.getType() instanceof TrieDoubleField) {
|
||||
newVal = Double.longBitsToDouble(val);
|
||||
} else if (schemaField.getType() instanceof TrieDateField) {
|
||||
newVal = new Date(val);
|
||||
} else if (schemaField.getType() instanceof EnumField) {
|
||||
newVal = ((EnumField) schemaField.getType()).intValueToStringValue(val.intValue());
|
||||
assert outValues.size() > 0;
|
||||
doc.addField(fieldName, outValues);
|
||||
}
|
||||
}
|
||||
} else {
|
||||
FieldInfo fi = fieldInfos.fieldInfo(fieldName);
|
||||
if (fi == null) {
|
||||
continue; // Searcher doesn't have info about this field, hence ignore it.
|
||||
}
|
||||
final DocValuesType dvType = fi.getDocValuesType();
|
||||
switch (dvType) {
|
||||
case NUMERIC:
|
||||
final NumericDocValues ndv = leafReader.getNumericDocValues(fieldName);
|
||||
if (ndv == null) {
|
||||
continue;
|
||||
}
|
||||
}
|
||||
doc.addField(fieldName, newVal);
|
||||
break;
|
||||
case BINARY:
|
||||
BinaryDocValues bdv = leafReader.getBinaryDocValues(fieldName);
|
||||
if (bdv == null) {
|
||||
continue;
|
||||
}
|
||||
BytesRef value;
|
||||
if (bdv.advanceExact(localId)) {
|
||||
value = BytesRef.deepCopyOf(bdv.binaryValue());
|
||||
} else {
|
||||
continue;
|
||||
}
|
||||
doc.addField(fieldName, value);
|
||||
break;
|
||||
case SORTED:
|
||||
SortedDocValues sdv = leafReader.getSortedDocValues(fieldName);
|
||||
if (sdv == null) {
|
||||
continue;
|
||||
}
|
||||
if (sdv.advanceExact(localId)) {
|
||||
final BytesRef bRef = sdv.binaryValue();
|
||||
// Special handling for Boolean fields since they're stored as 'T' and 'F'.
|
||||
if (schemaField.getType() instanceof BoolField) {
|
||||
doc.addField(fieldName, schemaField.getType().toObject(schemaField, bRef));
|
||||
Long val;
|
||||
if (ndv.advanceExact(localId)) {
|
||||
val = ndv.longValue();
|
||||
} else {
|
||||
doc.addField(fieldName, bRef.utf8ToString());
|
||||
continue;
|
||||
}
|
||||
}
|
||||
break;
|
||||
case SORTED_NUMERIC:
|
||||
final SortedNumericDocValues numericDv = leafReader.getSortedNumericDocValues(fieldName);
|
||||
NumberType type = schemaField.getType().getNumberType();
|
||||
if (numericDv != null) {
|
||||
if (numericDv.advance(localId) == localId) {
|
||||
final List<Object> outValues = new ArrayList<Object>(numericDv.docValueCount());
|
||||
for (int i = 0; i < numericDv.docValueCount(); i++) {
|
||||
long number = numericDv.nextValue();
|
||||
switch (type) {
|
||||
case INTEGER:
|
||||
outValues.add((int)number);
|
||||
break;
|
||||
case LONG:
|
||||
outValues.add(number);
|
||||
break;
|
||||
case FLOAT:
|
||||
outValues.add(NumericUtils.sortableIntToFloat((int)number));
|
||||
break;
|
||||
case DOUBLE:
|
||||
outValues.add(NumericUtils.sortableLongToDouble(number));
|
||||
break;
|
||||
case DATE:
|
||||
newVal = new Date(number);
|
||||
break;
|
||||
default:
|
||||
throw new AssertionError("Unexpected PointType: " + type);
|
||||
}
|
||||
Object newVal = val;
|
||||
if (schemaField.getType().isPointField()) {
|
||||
NumberType type = schemaField.getType().getNumberType();
|
||||
switch (type) {
|
||||
case INTEGER:
|
||||
newVal = val.intValue();
|
||||
break;
|
||||
case LONG:
|
||||
newVal = val.longValue();
|
||||
break;
|
||||
case FLOAT:
|
||||
newVal = Float.intBitsToFloat(val.intValue());
|
||||
break;
|
||||
case DOUBLE:
|
||||
newVal = Double.longBitsToDouble(val);
|
||||
break;
|
||||
case DATE:
|
||||
newVal = new Date(val);
|
||||
break;
|
||||
default:
|
||||
throw new AssertionError("Unexpected PointType: " + type);
|
||||
}
|
||||
assert outValues.size() > 0;
|
||||
doc.addField(fieldName, outValues);
|
||||
}
|
||||
}
|
||||
case SORTED_SET:
|
||||
final SortedSetDocValues values = leafReader.getSortedSetDocValues(fieldName);
|
||||
if (values != null && values.getValueCount() > 0) {
|
||||
if (values.advance(localId) == localId) {
|
||||
final List<Object> outValues = new LinkedList<Object>();
|
||||
for (long ord = values.nextOrd(); ord != SortedSetDocValues.NO_MORE_ORDS; ord = values.nextOrd()) {
|
||||
value = values.lookupOrd(ord);
|
||||
outValues.add(schemaField.getType().toObject(schemaField, value));
|
||||
} else {
|
||||
if (schemaField.getType() instanceof TrieIntField) {
|
||||
newVal = val.intValue();
|
||||
} else if (schemaField.getType() instanceof TrieFloatField) {
|
||||
newVal = Float.intBitsToFloat(val.intValue());
|
||||
} else if (schemaField.getType() instanceof TrieDoubleField) {
|
||||
newVal = Double.longBitsToDouble(val);
|
||||
} else if (schemaField.getType() instanceof TrieDateField) {
|
||||
newVal = new Date(val);
|
||||
} else if (schemaField.getType() instanceof EnumField) {
|
||||
newVal = ((EnumField) schemaField.getType()).intValueToStringValue(val.intValue());
|
||||
}
|
||||
assert outValues.size() > 0;
|
||||
doc.addField(fieldName, outValues);
|
||||
}
|
||||
}
|
||||
case NONE:
|
||||
break;
|
||||
doc.addField(fieldName, newVal);
|
||||
break;
|
||||
case BINARY:
|
||||
BinaryDocValues bdv = leafReader.getBinaryDocValues(fieldName);
|
||||
if (bdv == null) {
|
||||
continue;
|
||||
}
|
||||
BytesRef value;
|
||||
if (bdv.advanceExact(localId)) {
|
||||
value = BytesRef.deepCopyOf(bdv.binaryValue());
|
||||
} else {
|
||||
continue;
|
||||
}
|
||||
doc.addField(fieldName, value);
|
||||
break;
|
||||
case SORTED:
|
||||
SortedDocValues sdv = leafReader.getSortedDocValues(fieldName);
|
||||
if (sdv == null) {
|
||||
continue;
|
||||
}
|
||||
if (sdv.advanceExact(localId)) {
|
||||
final BytesRef bRef = sdv.binaryValue();
|
||||
// Special handling for Boolean fields since they're stored as 'T' and 'F'.
|
||||
if (schemaField.getType() instanceof BoolField) {
|
||||
doc.addField(fieldName, schemaField.getType().toObject(schemaField, bRef));
|
||||
} else {
|
||||
doc.addField(fieldName, bRef.utf8ToString());
|
||||
}
|
||||
}
|
||||
break;
|
||||
case SORTED_NUMERIC:
|
||||
throw new AssertionError("SORTED_NUMERIC not supported yet!");
|
||||
case SORTED_SET:
|
||||
throw new AssertionError("SORTED_SET fields should be multi-valued!");
|
||||
case NONE:
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -17,8 +17,6 @@
|
|||
package org.apache.solr.store.blockcache;
|
||||
|
||||
import java.nio.ByteBuffer;
|
||||
import java.util.List;
|
||||
import java.util.concurrent.CopyOnWriteArrayList;
|
||||
import java.util.concurrent.atomic.AtomicInteger;
|
||||
|
||||
import com.github.benmanes.caffeine.cache.Cache;
|
||||
|
@ -40,8 +38,8 @@ public class BlockCache {
|
|||
private final int numberOfBlocksPerBank;
|
||||
private final int maxEntries;
|
||||
private final Metrics metrics;
|
||||
private final List<OnRelease> onReleases = new CopyOnWriteArrayList<>();
|
||||
|
||||
private volatile OnRelease onRelease;
|
||||
|
||||
public static interface OnRelease {
|
||||
public void release(BlockCacheKey blockCacheKey);
|
||||
}
|
||||
|
@ -97,7 +95,7 @@ public class BlockCache {
|
|||
location.setRemoved(true);
|
||||
locks[bankId].clear(block);
|
||||
lockCounters[bankId].decrementAndGet();
|
||||
for (OnRelease onRelease : onReleases) {
|
||||
if (onRelease != null) {
|
||||
onRelease.release(blockCacheKey);
|
||||
}
|
||||
metrics.blockCacheEviction.incrementAndGet();
|
||||
|
@ -241,6 +239,6 @@ public class BlockCache {
|
|||
}
|
||||
|
||||
void setOnRelease(OnRelease onRelease) {
|
||||
this.onReleases.add(onRelease);
|
||||
this.onRelease = onRelease;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -23,7 +23,6 @@ import java.util.Map;
|
|||
import org.apache.lucene.document.BinaryDocValuesField; // javadocs
|
||||
import org.apache.lucene.document.NumericDocValuesField; // javadocs
|
||||
import org.apache.lucene.document.SortedDocValuesField; // javadocs
|
||||
import org.apache.lucene.document.SortedNumericDocValuesField;
|
||||
import org.apache.lucene.document.SortedSetDocValuesField; // javadocs
|
||||
import org.apache.lucene.document.StringField; // javadocs
|
||||
import org.apache.lucene.index.BinaryDocValues;
|
||||
|
@ -170,35 +169,7 @@ public class UninvertingReader extends FilterLeafReader {
|
|||
* Fields with this type act as if they were indexed with
|
||||
* {@link SortedSetDocValuesField}.
|
||||
*/
|
||||
SORTED_SET_DOUBLE,
|
||||
/**
|
||||
* Multi-valued Integer, (e.g. indexed with {@link org.apache.lucene.document.IntPoint})
|
||||
* <p>
|
||||
* Fields with this type act as if they were indexed with
|
||||
* {@link SortedNumericDocValuesField}.
|
||||
*/
|
||||
SORTED_INTEGER,
|
||||
/**
|
||||
* Multi-valued Float, (e.g. indexed with {@link org.apache.lucene.document.FloatPoint})
|
||||
* <p>
|
||||
* Fields with this type act as if they were indexed with
|
||||
* {@link SortedNumericDocValuesField}.
|
||||
*/
|
||||
SORTED_FLOAT,
|
||||
/**
|
||||
* Multi-valued Long, (e.g. indexed with {@link org.apache.lucene.document.LongPoint})
|
||||
* <p>
|
||||
* Fields with this type act as if they were indexed with
|
||||
* {@link SortedNumericDocValuesField}.
|
||||
*/
|
||||
SORTED_LONG,
|
||||
/**
|
||||
* Multi-valued Double, (e.g. indexed with {@link org.apache.lucene.document.DoublePoint})
|
||||
* <p>
|
||||
* Fields with this type act as if they were indexed with
|
||||
* {@link SortedNumericDocValuesField}.
|
||||
*/
|
||||
SORTED_DOUBLE
|
||||
SORTED_SET_DOUBLE
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -284,12 +255,6 @@ public class UninvertingReader extends FilterLeafReader {
|
|||
case SORTED_SET_DOUBLE:
|
||||
type = DocValuesType.SORTED_SET;
|
||||
break;
|
||||
case SORTED_INTEGER:
|
||||
case SORTED_FLOAT:
|
||||
case SORTED_LONG:
|
||||
case SORTED_DOUBLE:
|
||||
type = DocValuesType.SORTED_NUMERIC;
|
||||
break;
|
||||
default:
|
||||
throw new AssertionError();
|
||||
}
|
||||
|
|
|
@ -37,20 +37,20 @@
|
|||
<dynamicField name="*_i" type="int" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_i_dv" type="${solr.tests.intClass:pint}" indexed="false" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_is" type="int" indexed="true" stored="false" docValues="false" multiValued="true"/>
|
||||
<dynamicField name="*_is_dv" type="${solr.tests.intClass:pint}" indexed="false" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_is_dv" type="int" indexed="false" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_s" type="string" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_s_dv" type="string" indexed="false" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_ss" type="string" indexed="true" stored="false" docValues="false" multiValued="true"/>
|
||||
<dynamicField name="*_ss_dv" type="string" indexed="false" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_f" type="float" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_f_dv" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_fs_dv" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_fs_dv" type="float" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_l" type="long" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_l_dv" type="${solr.tests.longClass:plong}" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_ls_dv" type="${solr.tests.longClass:plong}" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_ls_dv" type="long" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_d" type="double" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_d_dv" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_ds_dv" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_ds_dv" type="double" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_dt" type="date" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_dt_dv" type="date" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_dts_dv" type="date" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
|
|
|
@ -34,9 +34,8 @@
|
|||
<dynamicField name="*_i" type="int" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_i_dv" type="int" indexed="false" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_i_p" type="pint" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_is" type="int" indexed="true" stored="true" docValues="false" multiValued="true"/>
|
||||
<dynamicField name="*_is" type="int" indexed="true" stored="false" docValues="false" multiValued="true"/>
|
||||
<dynamicField name="*_is_dv" type="int" indexed="false" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_is_p" type="pint" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_s" type="string" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_s_dv" type="string" indexed="false" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_ss" type="string" indexed="true" stored="false" docValues="false" multiValued="true"/>
|
||||
|
@ -46,19 +45,16 @@
|
|||
<dynamicField name="*_f_p" type="pfloat" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_fs" type="float" indexed="true" stored="false" docValues="false" multiValued="true"/>
|
||||
<dynamicField name="*_fs_dv" type="float" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_fs_p" type="pfloat" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_l" type="long" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_l_dv" type="long" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_l_p" type="plong" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_ls" type="long" indexed="true" stored="false" docValues="false" multiValued="true"/>
|
||||
<dynamicField name="*_ls_dv" type="long" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_ls_p" type="plong" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_d" type="double" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_d_dv" type="double" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_d_p" type="pdouble" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_ds" type="double" indexed="true" stored="false" docValues="false" multiValued="true"/>
|
||||
<dynamicField name="*_ds_dv" type="double" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_ds_p" type="pdouble" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_dt" type="date" indexed="true" stored="false" docValues="false"/>
|
||||
<dynamicField name="*_dt_dv" type="date" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_dts_dv" type="date" indexed="true" stored="false" docValues="true" multiValued="true"/>
|
||||
|
@ -71,18 +67,15 @@
|
|||
<copyField source="*_f" dest="*_f_dv"/>
|
||||
<copyField source="*_f" dest="*_f_p"/>
|
||||
<copyField source="*_is" dest="*_is_dv"/>
|
||||
<copyField source="*_is" dest="*_is_p"/>
|
||||
<copyField source="*_s" dest="*_s_dv"/>
|
||||
<copyField source="*_l" dest="*_l_dv"/>
|
||||
<copyField source="*_l" dest="*_l_p"/>
|
||||
<copyField source="*_d" dest="*_d_dv"/>
|
||||
<copyField source="*_d" dest="*_d_p"/>
|
||||
<copyField source="*_ss" dest="*_ss_dv"/>
|
||||
<copyField source="*_is" dest="*_is_dv"/>
|
||||
<copyField source="*_fs" dest="*_fs_dv"/>
|
||||
<copyField source="*_fs" dest="*_fs_p"/>
|
||||
<copyField source="*_ls" dest="*_ls_dv"/>
|
||||
<copyField source="*_ls" dest="*_ls_p"/>
|
||||
<copyField source="*_ds" dest="*_ds_dv"/>
|
||||
<copyField source="*_ds" dest="*_ds_p"/>
|
||||
<copyField source="id" dest="id_dv"/>
|
||||
</schema>
|
||||
|
|
|
@ -84,12 +84,6 @@
|
|||
<dynamicField name="*_p_d_ni_ns_dv" type="pdouble" indexed="false" stored="false" docValues="true" useDocValuesAsStored="true"/>
|
||||
<dynamicField name="*_p_f_ni_ns_dv" type="pfloat" indexed="false" stored="false" docValues="true" useDocValuesAsStored="true"/>
|
||||
|
||||
<dynamicField name="*_p_i_dv_ns_mv" type="pint" indexed="true" stored="false" docValues="true" useDocValuesAsStored="true" multiValued="true"/>
|
||||
<dynamicField name="*_p_d_dv_ns_mv" type="pdouble" indexed="true" stored="false" docValues="true" useDocValuesAsStored="true" multiValued="true"/>
|
||||
<dynamicField name="*_p_l_dv_ns_mv" type="plong" indexed="true" stored="false" docValues="true" useDocValuesAsStored="true" multiValued="true"/>
|
||||
<dynamicField name="*_p_f_dv_ns_mv" type="pfloat" indexed="true" stored="false" docValues="true" useDocValuesAsStored="true" multiValued="true"/>
|
||||
|
||||
|
||||
</fields>
|
||||
|
||||
<uniqueKey>id</uniqueKey>
|
||||
|
|
|
@ -571,17 +571,18 @@
|
|||
|
||||
<field name="timestamp" type="date" indexed="true" stored="true" docValues="true" default="NOW" multiValued="false"/>
|
||||
<field name="multiDefault" type="string" indexed="true" stored="true" default="muLti-Default" multiValued="true"/>
|
||||
<field name="intDefault" type="${solr.tests.intClass:pint}" indexed="true" stored="true" default="42" multiValued="false"/>
|
||||
<field name="intDvoDefault" type="${solr.tests.intClass:pint}" indexed="false" stored="false" multiValued="false"
|
||||
<field name="intDefault" type="int" indexed="true" stored="true" default="42" multiValued="false"/>
|
||||
<field name="intDvoDefault" type="int" indexed="false" stored="false" multiValued="false"
|
||||
useDocValuesAsStored="true" docValues="true" default="42" />
|
||||
<field name="intRemove" type="${solr.tests.intClass:pint}" indexed="true" stored="true" multiValued="true"/>
|
||||
<field name="intRemove" type="int" indexed="true" stored="true" multiValued="true"/>
|
||||
<field name="dateRemove" type="date" indexed="true" stored="true" multiValued="true"/>
|
||||
<field name="floatRemove" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true" multiValued="true"/>
|
||||
<field name="floatRemove" type="float" indexed="true" stored="true" multiValued="true"/>
|
||||
|
||||
<field name="nopositionstext" type="nopositions" indexed="true" stored="true"/>
|
||||
|
||||
<field name="tlong" type="tlong" indexed="true" stored="true"/>
|
||||
<field name="_version_" type="${solr.tests.longClass:plong}" indexed="false" stored="false" docValues="true" multiValued="false" useDocValuesAsStored="true"/>
|
||||
|
||||
<field name="_version_" type="long" indexed="false" stored="false" docValues="true" multiValued="false" useDocValuesAsStored="true"/>
|
||||
|
||||
<field name="title_stringNoNorms" type="string" omitNorms="true" indexed="true" stored="true"/>
|
||||
|
||||
|
@ -611,14 +612,14 @@
|
|||
|
||||
<dynamicField name="*_s" type="string" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_s1" type="string" indexed="true" stored="true" multiValued="false"/>
|
||||
<dynamicField name="*_l" type="${solr.tests.longClass:plong}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_l1" type="${solr.tests.longClass:plong}" indexed="true" stored="true" multiValued="false"/>
|
||||
<dynamicField name="*_l" type="long" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_l1" type="long" indexed="true" stored="true" multiValued="false"/>
|
||||
<dynamicField name="*_t" type="text" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_b" type="boolean" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_b1" type="boolean" indexed="true" stored="true" multiValued="false"/>
|
||||
<dynamicField name="*_f" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_f" type="float" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_f1" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true" multiValued="false"/>
|
||||
<dynamicField name="*_d" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_d" type="double" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_d1" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true" multiValued="false"/>
|
||||
<dynamicField name="*_dt" type="date" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_dt1" type="date" indexed="true" stored="true" multiValued="false"/>
|
||||
|
@ -678,10 +679,10 @@
|
|||
<dynamicField name="*_sev_enum" type="severityType" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
|
||||
<!-- With DocValues=true -->
|
||||
<dynamicField name="*_i_dv" type="${solr.tests.intClass:pint}" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_l_dv" type="${solr.tests.longClass:plong}" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_f_dv" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_d_dv" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_i_dv" type="int" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_l_dv" type="long" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_f_dv" type="float" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_d_dv" type="double" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_dt_dv" type="date" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_f1_dv" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true" docValues="true" multiValued="false"/>
|
||||
|
||||
|
@ -692,41 +693,21 @@
|
|||
useDocValuesAsStored="true"/>
|
||||
<dynamicField name="*_s_dvo" multiValued="false" type="string" docValues="true" indexed="false" stored="false"
|
||||
useDocValuesAsStored="true"/>
|
||||
|
||||
<dynamicField name="*_ii_dvo" multiValued="true" type="${solr.tests.intClass:pint}" docValues="true" indexed="false" stored="false"
|
||||
<dynamicField name="*_ii_dvo" multiValued="true" type="int" docValues="true" indexed="false" stored="false"
|
||||
useDocValuesAsStored="true"/>
|
||||
<dynamicField name="*_dd_dvo" multiValued="true" type="${solr.tests.doubleClass:pdouble}" docValues="true" indexed="false" stored="false"
|
||||
<dynamicField name="*_dd_dvo" multiValued="true" type="double" docValues="true" indexed="false" stored="false"
|
||||
useDocValuesAsStored="true"/>
|
||||
|
||||
<!-- Non-stored, DocValues=true, useDocValuesAsStored=false -->
|
||||
<field name="single_i_dvn" multiValued="false" type="${solr.tests.intClass:pint}" indexed="true" stored="true"/>
|
||||
<field name="single_d_dvn" multiValued="false" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true"/>
|
||||
<field name="single_s_dvn" multiValued="false" type="string" indexed="true" stored="true"/>
|
||||
<field name="copy_single_i_dvn" multiValued="false" type="${solr.tests.intClass:pint}" docValues="true" indexed="true" stored="false"
|
||||
<field name="copy_single_i_dvn" multiValued="false" type="int" docValues="true" indexed="true" stored="false"
|
||||
useDocValuesAsStored="false"/>
|
||||
<field name="copy_single_d_dvn" multiValued="false" type="${solr.tests.doubleClass:pdouble}" docValues="true" indexed="true" stored="false"
|
||||
<field name="copy_single_d_dvn" multiValued="false" type="double" docValues="true" indexed="true" stored="false"
|
||||
useDocValuesAsStored="false"/>
|
||||
<field name="copy_single_s_dvn" multiValued="false" type="string" docValues="true" indexed="true" stored="false"
|
||||
useDocValuesAsStored="false"/>
|
||||
|
||||
<!-- Test point fields explicitly -->
|
||||
<dynamicField name="*_i_p" type="pint" indexed="true" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_is_p" type="pint" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_i_ni_p" type="pint" indexed="false" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_is_ni_p" type="pint" indexed="false" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_l_p" type="plong" indexed="true" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_ls_p" type="plong" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_l_ni_p" type="plong" indexed="false" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_ls_ni_p" type="plong" indexed="false" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_f_p" type="pfloat" indexed="true" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_fs_p" type="pfloat" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_f_ni_p" type="pfloat" indexed="false" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_fs_ni_p" type="pfloat" indexed="false" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_d_p" type="pdouble" indexed="true" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_ds_p" type="pdouble" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_d_ni_p" type="pdouble" indexed="false" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_ds_ni_p" type="pdouble" indexed="false" stored="true" docValues="true" multiValued="true"/>
|
||||
|
||||
<copyField source="single_i_dvn" dest="copy_single_i_dvn"/>
|
||||
<copyField source="single_d_dvn" dest="copy_single_d_dvn"/>
|
||||
<copyField source="single_s_dvn" dest="copy_single_s_dvn"/>
|
||||
|
|
|
@ -322,12 +322,12 @@ valued. -->
|
|||
|
||||
<field name="text_no_analyzer" type="text_no_analyzer" indexed="true" />
|
||||
|
||||
<field name="_version_" type="${solr.tests.longClass:plong}" indexed="true" stored="true" multiValued="false" />
|
||||
<field name="_version_" type="long" indexed="true" stored="true" multiValued="false" />
|
||||
|
||||
<field name="cat" type="string" indexed="true" stored="true" multiValued="true"/>
|
||||
<field name="cat_docValues" type="string" indexed="true" stored="true" docValues="true" multiValued="true" />
|
||||
<field name="cat_intDocValues" type="${solr.tests.intClass:pint}" indexed="true" stored="true" docValues="true" multiValued="true" />
|
||||
<field name="cat_floatDocValues" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true" docValues="true" multiValued="true" />
|
||||
<field name="cat_intDocValues" type="int" indexed="true" stored="true" docValues="true" multiValued="true" />
|
||||
<field name="cat_floatDocValues" type="float" indexed="true" stored="true" docValues="true" multiValued="true" />
|
||||
<field name="cat_length" type="text_length" indexed="true" stored="true" multiValued="true"/>
|
||||
|
||||
|
||||
|
@ -346,14 +346,14 @@ valued. -->
|
|||
<dynamicField name="*_ss" type="string" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_sS" type="string" indexed="false" stored="true"/>
|
||||
<dynamicField name="*_i" type="${solr.tests.intClass:pint}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_ii" type="${solr.tests.intClass:pint}" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_ii" type="int" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_l" type="${solr.tests.longClass:plong}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_f" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_d" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true"/>
|
||||
|
||||
<dynamicField name="*_ti" type="tint" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_ti_dv" type="tint" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_ti_ni_dv" type="tint" indexed="false" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_ti_ni_dv" type="tint" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_tl" type="tlong" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_tl_dv" type="tlong" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_tl_ni_dv" type="tlong" indexed="false" stored="true" docValues="true"/>
|
||||
|
@ -382,24 +382,6 @@ valued. -->
|
|||
<dynamicField name="*_tdts" type="tdates" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_tdts_dv" type="tdates" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_tdts_ni_dv" type="tdates" indexed="false" stored="true" docValues="true"/>
|
||||
|
||||
<!-- Test point fields explicitly -->
|
||||
<dynamicField name="*_i_p" type="pint" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_is_p" type="pint" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_i_ni_p" type="pint" indexed="false" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_is_ni_p" type="pint" indexed="false" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_l_p" type="plong" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_ls_p" type="plong" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_l_ni_p" type="plong" indexed="false" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_ls_ni_p" type="plong" indexed="false" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_f_p" type="pfloat" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_fs_p" type="pfloat" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_f_ni_p" type="pfloat" indexed="false" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_fs_ni_p" type="pfloat" indexed="false" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_d_p" type="pdouble" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_ds_p" type="pdouble" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_d_ni_p" type="pdouble" indexed="false" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_ds_ni_p" type="pdouble" indexed="false" stored="true" docValues="true" multiValued="true"/>
|
||||
|
||||
<dynamicField name="*_t" type="text" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_b" type="boolean" indexed="true" stored="true"/>
|
||||
|
|
|
@ -458,8 +458,8 @@
|
|||
<field name="bind" type="boolean" indexed="true" stored="false"/>
|
||||
<field name="bsto" type="boolean" indexed="false" stored="true"/>
|
||||
<field name="bindsto" type="boolean" indexed="true" stored="true"/>
|
||||
<field name="isto" type="${solr.tests.intClass:pint}" indexed="false" stored="true"/>
|
||||
<field name="iind" type="${solr.tests.intClass:pint}" indexed="true" stored="false"/>
|
||||
<field name="isto" type="int" indexed="false" stored="true"/>
|
||||
<field name="iind" type="int" indexed="true" stored="false"/>
|
||||
<field name="ssto" type="string" indexed="false" stored="true"/>
|
||||
<field name="sind" type="string" indexed="true" stored="false"/>
|
||||
<field name="sindsto" type="string" indexed="true" stored="true"/>
|
||||
|
@ -554,20 +554,20 @@
|
|||
|
||||
<dynamicField name="*_sw" type="text_sw" indexed="true" stored="true" multiValued="true"/>
|
||||
|
||||
<dynamicField name="*_i" type="${solr.tests.intClass:pint}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_is" type="${solr.tests.intClass:pint}" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_i" type="int" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_is" type="int" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_i_dv" type="${solr.tests.intClass:pint}" indexed="true" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_is_dv" type="${solr.tests.intClass:pint}" indexed="true" stored="true" docValues="true" multiValued="true"/>
|
||||
<dynamicField name="*_is_dv" type="${solr.tests.intClass:pint}" indexed="true" stored="true" docValues="true" multiValued="false"/>
|
||||
<dynamicField name="*_s1" type="string" indexed="true" stored="true" multiValued="false"/>
|
||||
<!-- :TODO: why are these identical?!?!?! -->
|
||||
<dynamicField name="*_s" type="string" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_ss" type="string" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_l" type="${solr.tests.longClass:plong}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_l" type="long" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_t" type="text" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_tt" type="text" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_b" type="boolean" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_f" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_d" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_d" type="double" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_dt" type="date" indexed="true" stored="true"/>
|
||||
|
||||
<dynamicField name="*_pi" type="pint" indexed="true" stored="true" docValues="false" multiValued="false"/>
|
||||
|
|
|
@ -109,7 +109,7 @@
|
|||
<!-- If you remove this field, you must _also_ disable the update log in solrconfig.xml
|
||||
or Solr won't start. _version_ and update log are required for SolrCloud
|
||||
-->
|
||||
<field name="_version_" type="${solr.tests.longClass:plong}" indexed="false" stored="false" docValues="true"/>
|
||||
<field name="_version_" type="long" indexed="false" stored="false" docValues="true"/>
|
||||
|
||||
<!-- points to the root document of a block of nested documents. Required for nested
|
||||
document support, may be removed otherwise
|
||||
|
@ -130,9 +130,9 @@
|
|||
<field name="includes" type="text_general" indexed="true" stored="true" termVectors="true" termPositions="true"
|
||||
termOffsets="true"/>
|
||||
|
||||
<field name="weight" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true"/>
|
||||
<field name="price" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true"/>
|
||||
<field name="popularity" type="${solr.tests.intClass:pint}" indexed="true" stored="true"/>
|
||||
<field name="weight" type="float" indexed="true" stored="true"/>
|
||||
<field name="price" type="float" indexed="true" stored="true"/>
|
||||
<field name="popularity" type="int" indexed="true" stored="true"/>
|
||||
<field name="inStock" type="boolean" indexed="true" stored="true"/>
|
||||
|
||||
<field name="store" type="location" indexed="true" stored="true"/>
|
||||
|
@ -200,27 +200,27 @@
|
|||
a "*" only at the start or the end. -->
|
||||
|
||||
<!-- docvalues and stored are exclusive -->
|
||||
<dynamicField name="*_i" type="${solr.tests.intClass:pint}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_is" type="${solr.tests.intClass:pint}" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_i" type="int" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_is" type="int" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_id" type="${solr.tests.intClass:pint}" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_ids" type="${solr.tests.intClass:pint}" indexed="true" stored="false" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_ids" type="int" indexed="true" stored="false" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_s" type="string" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_s1" type="string" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_ss" type="string" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_sd" type="string" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_sds" type="string" indexed="true" stored="false" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_l" type="${solr.tests.longClass:plong}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_ls" type="${solr.tests.longClass:plong}" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_ld" type="${solr.tests.longClass:plong}" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_lds" type="${solr.tests.longClass:plong}" indexed="true" stored="false" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_f" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_fs" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_l" type="long" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_ls" type="long" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_ld" type="long" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_lds" type="long" indexed="true" stored="false" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_f" type="float" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_fs" type="float" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_fd" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_fds" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="false" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_d" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_ds" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_fds" type="float" indexed="true" stored="false" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_d" type="double" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_ds" type="double" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_dd" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="false" docValues="true"/>
|
||||
<dynamicField name="*_dds" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="false" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_dds" type="double" indexed="true" stored="false" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_dt" type="date" indexed="true" stored="true"/>
|
||||
<dynamicField name="*_dts" type="date" indexed="true" stored="true" multiValued="true"/>
|
||||
<dynamicField name="*_dtd" type="date" indexed="true" stored="false" docValues="true"/>
|
||||
|
@ -228,15 +228,15 @@
|
|||
|
||||
<!-- docvalues and stored (S suffix) -->
|
||||
<dynamicField name="*_idS" type="${solr.tests.intClass:pint}" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_idsS" type="${solr.tests.intClass:pint}" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_idsS" type="int" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_sdS" type="string" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_sdsS" type="string" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_ldS" type="${solr.tests.longClass:plong}" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_ldsS" type="${solr.tests.longClass:plong}" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_ldsS" type="long" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_fdS" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_fdsS" type="${solr.tests.floatClass:pfloat}" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_fdsS" type="float" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_ddS" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_ddsS" type="${solr.tests.doubleClass:pdouble}" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_ddsS" type="double" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
<dynamicField name="*_dtdS" type="date" indexed="true" stored="true" docValues="true"/>
|
||||
<dynamicField name="*_dtdsS" type="date" indexed="true" stored="true" multiValued="true" docValues="true"/>
|
||||
|
||||
|
@ -396,10 +396,10 @@
|
|||
<fieldType name="tdouble" class="solr.TrieDoubleField" precisionStep="8" positionIncrementGap="0"/>
|
||||
|
||||
<!-- Point Fields -->
|
||||
<fieldType name="pint" class="solr.IntPointField" docValues="true"/>
|
||||
<fieldType name="plong" class="solr.LongPointField" docValues="true"/>
|
||||
<fieldType name="pdouble" class="solr.DoublePointField" docValues="true"/>
|
||||
<fieldType name="pfloat" class="solr.FloatPointField" docValues="true"/>
|
||||
<fieldType name="pint" class="solr.IntPointField"/>
|
||||
<fieldType name="plong" class="solr.LongPointField"/>
|
||||
<fieldType name="pdouble" class="solr.DoublePointField"/>
|
||||
<fieldType name="pfloat" class="solr.FloatPointField"/>
|
||||
|
||||
|
||||
<!-- The format for this date field is of the form 1995-12-31T23:59:59Z, and
|
||||
|
|
|
@ -35,8 +35,6 @@
|
|||
<arr name="typeClass">
|
||||
<str>solr.TrieIntField</str>
|
||||
<str>solr.TrieLongField</str>
|
||||
<str>solr.IntPointField</str>
|
||||
<str>solr.LongPointField</str>
|
||||
</arr>
|
||||
</processor>
|
||||
<processor class="solr.MinFieldValueUpdateProcessorFactory">
|
||||
|
|
|
@ -76,8 +76,8 @@ public class TestDistributedSearch extends BaseDistributedSearchTestCase {
|
|||
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||
|
||||
String t1="a_t";
|
||||
String i1 = pickRandom("a_i1", "a_i_p", "a_i_ni_p");
|
||||
String nint = pickRandom("n_i", "n_is_p", "n_is_ni_p");
|
||||
String i1="a_i1";
|
||||
String nint = "n_i";
|
||||
String tint = "n_ti";
|
||||
String tlong = "other_tl1";
|
||||
String tdate_a = "a_n_tdt";
|
||||
|
@ -187,8 +187,7 @@ public class TestDistributedSearch extends BaseDistributedSearchTestCase {
|
|||
// these queries should be exactly ordered and scores should exactly match
|
||||
query("q","*:*", "sort",i1+" desc");
|
||||
query("q","*:*", "sort","{!func}testfunc(add("+i1+",5))"+" desc");
|
||||
query("q",i1 + "[* TO *]", "sort",i1+" asc");
|
||||
query("q","*:*", "sort",i1+" asc, id desc");
|
||||
query("q","*:*", "sort",i1+" asc");
|
||||
query("q","*:*", "sort",i1+" desc", "fl","*,score");
|
||||
query("q","*:*", "sort","n_tl1 asc", "fl","*,score");
|
||||
query("q","*:*", "sort","n_tl1 desc");
|
||||
|
@ -423,7 +422,6 @@ public class TestDistributedSearch extends BaseDistributedSearchTestCase {
|
|||
|
||||
query("q","*:*", "sort",i1+" desc", "stats", "true", "stats.field", "stats_dt");
|
||||
query("q","*:*", "sort",i1+" desc", "stats", "true", "stats.field", i1);
|
||||
query("q","*:*", "sort",i1+" desc", "stats", "true", "stats.field", nint);
|
||||
|
||||
handle.put("stddev", FUZZY);
|
||||
handle.put("sumOfSquares", FUZZY);
|
||||
|
|
|
@ -50,7 +50,6 @@ import org.noggit.ObjectBuilder;
|
|||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
@SolrTestCaseJ4.SuppressPointFields(bugUrl="https://issues.apache.org/jira/browse/SOLR-9992")
|
||||
public class TestGroupingSearch extends SolrTestCaseJ4 {
|
||||
|
||||
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||
|
|
|
@ -74,12 +74,6 @@ public class TestRandomDVFaceting extends SolrTestCaseJ4 {
|
|||
types.add(new FldType("small2_i",ZERO_ONE, new IRange(0,5+indexSize/3)));
|
||||
types.add(new FldType("small2_is",ZERO_TWO, new IRange(0,5+indexSize/3)));
|
||||
types.add(new FldType("small3_is",new IRange(0,25), new IRange(0,100)));
|
||||
|
||||
types.add(new FldType("foo_fs", new IRange(0,25), new FVal(0,indexSize)));
|
||||
types.add(new FldType("foo_f", ZERO_ONE, new FVal(0,indexSize)));
|
||||
types.add(new FldType("foo_ds", new IRange(0,25), new FVal(0,indexSize)));
|
||||
types.add(new FldType("foo_d", ZERO_ONE, new FVal(0,indexSize)));
|
||||
types.add(new FldType("foo_ls", new IRange(0,25), new IRange(0,indexSize)));
|
||||
|
||||
types.add(new FldType("missing_i",new IRange(0,0), new IRange(0,100)));
|
||||
types.add(new FldType("missing_is",new IRange(0,0), new IRange(0,100)));
|
||||
|
@ -162,9 +156,10 @@ public class TestRandomDVFaceting extends SolrTestCaseJ4 {
|
|||
Random rand = random();
|
||||
boolean validate = validateResponses;
|
||||
ModifiableSolrParams params = params("facet","true", "wt","json", "indent","true", "omitHeader","true");
|
||||
params.add("q","*:*"); // TODO: select subsets
|
||||
params.add("q","*:*", "rows","0"); // TODO: select subsets
|
||||
params.add("rows","0");
|
||||
|
||||
|
||||
SchemaField sf = req.getSchema().getField(ftype.fname);
|
||||
boolean multiValued = sf.getType().multiValuedFieldCache();
|
||||
boolean indexed = sf.indexed();
|
||||
|
@ -178,16 +173,12 @@ public class TestRandomDVFaceting extends SolrTestCaseJ4 {
|
|||
params.add("facet.offset", Integer.toString(offset));
|
||||
}
|
||||
|
||||
int limit = 100;
|
||||
if (rand.nextInt(100) < 20) {
|
||||
if(rarely()) {
|
||||
params.add("facet.limit", "-1");
|
||||
} else {
|
||||
int limit = 100;
|
||||
if (rand.nextBoolean()) {
|
||||
limit = rand.nextInt(100) < 10 ? rand.nextInt(indexSize/2+1) : rand.nextInt(indexSize*2);
|
||||
}
|
||||
params.add("facet.limit", Integer.toString(limit));
|
||||
if (rand.nextBoolean()) {
|
||||
limit = rand.nextInt(100) < 10 ? rand.nextInt(indexSize/2+1) : rand.nextInt(indexSize*2);
|
||||
}
|
||||
params.add("facet.limit", Integer.toString(limit));
|
||||
}
|
||||
|
||||
// the following two situations cannot work for unindexed single-valued numerics:
|
||||
|
@ -243,10 +234,7 @@ public class TestRandomDVFaceting extends SolrTestCaseJ4 {
|
|||
responses.add(strResponse);
|
||||
}
|
||||
// If there is a PointField option for this test, also test it
|
||||
// Don't check points if facet.mincount=0
|
||||
if (h.getCore().getLatestSchema().getFieldOrNull(facet_field + "_p") != null
|
||||
&& params.get("facet.mincount") != null
|
||||
&& params.getInt("facet.mincount").intValue() > 0) {
|
||||
if (h.getCore().getLatestSchema().getFieldOrNull(facet_field + "_p") != null) {
|
||||
params.set("facet.field", "{!key="+facet_field+"}"+facet_field+"_p");
|
||||
String strResponse = h.query(req(params));
|
||||
responses.add(strResponse);
|
||||
|
|
|
@ -22,7 +22,6 @@ import java.util.Map;
|
|||
import java.util.Set;
|
||||
import java.util.Random;
|
||||
|
||||
import org.apache.solr.SolrTestCaseJ4.SuppressPointFields;
|
||||
import org.apache.solr.client.solrj.SolrQuery;
|
||||
import org.apache.solr.client.solrj.impl.CloudSolrClient;
|
||||
import org.apache.solr.client.solrj.response.QueryResponse;
|
||||
|
@ -32,8 +31,7 @@ import org.apache.solr.common.params.CommonParams;
|
|||
import org.apache.solr.common.params.ShardParams;
|
||||
import org.apache.solr.common.util.SimpleOrderedMap;
|
||||
import org.apache.solr.response.SolrQueryResponse;
|
||||
// This test uses grouping requests, which doesn't work yet with PointFields
|
||||
@SuppressPointFields(bugUrl="https://issues.apache.org/jira/browse/SOLR-9992")
|
||||
|
||||
class SegmentTerminateEarlyTestState {
|
||||
|
||||
final String keyField = "id";
|
||||
|
|
|
@ -22,7 +22,6 @@ import java.util.HashMap;
|
|||
import java.util.Map;
|
||||
|
||||
import org.apache.lucene.util.TestUtil;
|
||||
import org.apache.solr.SolrTestCaseJ4.SuppressPointFields;
|
||||
import org.apache.solr.client.solrj.impl.CloudSolrClient;
|
||||
import org.apache.solr.client.solrj.request.CollectionAdminRequest;
|
||||
import org.apache.solr.client.solrj.request.schema.SchemaRequest.Field;
|
||||
|
@ -42,7 +41,6 @@ import org.junit.rules.TestName;
|
|||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
@SuppressPointFields(bugUrl="https://issues.apache.org/jira/browse/SOLR-9992")
|
||||
public class TestSegmentSorting extends SolrCloudTestCase {
|
||||
|
||||
private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
|
||||
|
|
|
@ -84,8 +84,7 @@ public class StatsComponentTest extends AbstractSolrTestCase {
|
|||
"stats_i","stats_l","stats_f","stats_d",
|
||||
"stats_ti","stats_tl","stats_tf","stats_td",
|
||||
"stats_ti_dv","stats_tl_dv","stats_tf_dv","stats_td_dv",
|
||||
"stats_ti_ni_dv","stats_tl_ni_dv","stats_tf_ni_dv","stats_td_ni_dv",
|
||||
"stats_i_ni_p","stats_l_ni_p","stats_f_ni_p","stats_d_ni_p",
|
||||
"stats_ti_ni_dv","stats_tl_ni_dv","stats_tf_ni_dv","stats_td_ni_dv"
|
||||
}) {
|
||||
|
||||
// all of our checks should work with all of these params
|
||||
|
@ -112,9 +111,7 @@ public class StatsComponentTest extends AbstractSolrTestCase {
|
|||
for (String f : new String[] {"stats_ii",
|
||||
"stats_tis","stats_tfs","stats_tls","stats_tds", // trie fields
|
||||
"stats_tis_dv","stats_tfs_dv","stats_tls_dv","stats_tds_dv", // Doc Values
|
||||
"stats_tis_ni_dv","stats_tfs_ni_dv","stats_tls_ni_dv","stats_tds_ni_dv", // Doc Values Not indexed
|
||||
"stats_is_p", "stats_fs_p", "stats_ls_p", "stats_ds_p", // Point Fields
|
||||
"stats_is_ni_p","stats_fs_ni_p","stats_ls_ni_p" // Point Doc Values Not indexed
|
||||
"stats_tis_ni_dv","stats_tfs_ni_dv","stats_tls_ni_dv","stats_tds_ni_dv" // Doc Values Not indexed
|
||||
}) {
|
||||
|
||||
doTestMVFieldStatisticsResult(f);
|
||||
|
|
|
@ -27,8 +27,6 @@ import java.util.regex.Pattern;
|
|||
*
|
||||
*
|
||||
**/
|
||||
// TermsComponent not currently supported for PointFields
|
||||
@SolrTestCaseJ4.SuppressPointFields
|
||||
public class TermsComponentTest extends SolrTestCaseJ4 {
|
||||
|
||||
@BeforeClass
|
||||
|
|
|
@ -17,9 +17,6 @@
|
|||
|
||||
package org.apache.solr.request;
|
||||
|
||||
import static junit.framework.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
|
||||
import org.apache.solr.request.SimpleFacets.FacetMethod;
|
||||
import org.apache.solr.schema.BoolField;
|
||||
import org.apache.solr.schema.IntPointField;
|
||||
|
@ -28,6 +25,8 @@ import org.apache.solr.schema.StrField;
|
|||
import org.apache.solr.schema.TrieIntField;
|
||||
import org.junit.Test;
|
||||
|
||||
import static junit.framework.Assert.assertEquals;
|
||||
|
||||
public class TestFacetMethods {
|
||||
|
||||
// TODO - make these public in FieldProperties?
|
||||
|
@ -215,12 +214,6 @@ public class TestFacetMethods {
|
|||
assertEquals(SimpleFacets.FacetMethod.FCS, SimpleFacets.selectFacetMethod(field, FacetMethod.ENUM, 0));
|
||||
assertEquals(SimpleFacets.FacetMethod.FCS, SimpleFacets.selectFacetMethod(field, FacetMethod.FC, 0));
|
||||
assertEquals(SimpleFacets.FacetMethod.FCS, SimpleFacets.selectFacetMethod(field, FacetMethod.FCS, 0));
|
||||
field = new SchemaField("fooMV", new IntPointField(), 0x00000200, "0"); //MultiValued
|
||||
assertTrue(field.multiValued());
|
||||
assertEquals(SimpleFacets.FacetMethod.FCS, SimpleFacets.selectFacetMethod(field, null, 0));
|
||||
assertEquals(SimpleFacets.FacetMethod.FCS, SimpleFacets.selectFacetMethod(field, FacetMethod.ENUM, 0));
|
||||
assertEquals(SimpleFacets.FacetMethod.FCS, SimpleFacets.selectFacetMethod(field, FacetMethod.FC, 0));
|
||||
assertEquals(SimpleFacets.FacetMethod.FCS, SimpleFacets.selectFacetMethod(field, FacetMethod.FCS, 0));
|
||||
}
|
||||
|
||||
}
|
||||
|
|
|
@ -105,11 +105,10 @@ public class TestIntervalFaceting extends SolrTestCaseJ4 {
|
|||
assertU(adoc("id", "10"));
|
||||
assertU(commit());
|
||||
int i = 11;
|
||||
while (getNumberOfReaders() < 2 && i < 20) {
|
||||
while (getNumberOfReaders() < 2 && i < 10) {
|
||||
//try to get more than one segment
|
||||
assertU(adoc("id", String.valueOf(i), "test_i_dv", String.valueOf(i)));
|
||||
assertU(commit());
|
||||
i++;
|
||||
}
|
||||
if (getNumberOfReaders() < 2) {
|
||||
// It is OK if for some seeds we fall into this case (for example, TieredMergePolicy with
|
||||
|
@ -245,11 +244,9 @@ public class TestIntervalFaceting extends SolrTestCaseJ4 {
|
|||
// All field values will be a number between 0 and cardinality
|
||||
int cardinality = 100000;
|
||||
// Fields to use for interval faceting
|
||||
String[] fields = new String[]{
|
||||
"test_s_dv", "test_i_dv", "test_l_dv", "test_f_dv", "test_d_dv",
|
||||
String[] fields = new String[]{"test_s_dv", "test_i_dv", "test_l_dv", "test_f_dv", "test_d_dv",
|
||||
"test_ss_dv", "test_is_dv", "test_fs_dv", "test_ls_dv", "test_ds_dv", "test_s", "test_i",
|
||||
"test_l", "test_f", "test_d", "test_ss", "test_is", "test_fs", "test_ls", "test_ds",
|
||||
"test_i_p", "test_is_p", "test_l_p", "test_ls_p", "test_f_p", "test_fs_p", "test_d_p", "test_ds_p"};
|
||||
"test_l", "test_f", "test_d", "test_ss", "test_is", "test_fs", "test_ls", "test_ds"};
|
||||
for (int i = 0; i < atLeast(500); i++) {
|
||||
if (random().nextInt(50) == 0) {
|
||||
//have some empty docs
|
||||
|
@ -746,11 +743,10 @@ public class TestIntervalFaceting extends SolrTestCaseJ4 {
|
|||
assertU(commit());
|
||||
|
||||
int i = 12;
|
||||
while (getNumberOfReaders() < 2 && i < 20) {
|
||||
while (getNumberOfReaders() < 2 && i < 10) {
|
||||
//try to get more than one segment
|
||||
assertU(adoc("id", String.valueOf(i), "test_s_dv", String.valueOf(i)));
|
||||
assertU(commit());
|
||||
i++;
|
||||
}
|
||||
if (getNumberOfReaders() < 2) {
|
||||
// It is OK if for some seeds we fall into this case (for example, TieredMergePolicy with
|
||||
|
|
|
@ -23,11 +23,15 @@ import java.util.TreeSet;
|
|||
|
||||
import org.apache.solr.SolrTestCaseJ4;
|
||||
import org.apache.solr.common.SolrException;
|
||||
import org.apache.solr.schema.DoublePointField;
|
||||
import org.apache.solr.schema.IntPointField;
|
||||
import org.apache.solr.schema.PointField;
|
||||
import org.apache.solr.schema.SchemaField;
|
||||
import org.junit.After;
|
||||
import org.junit.BeforeClass;
|
||||
import org.junit.Ignore;
|
||||
import org.junit.Test;
|
||||
|
||||
import com.google.common.collect.ImmutableList;
|
||||
import com.google.common.collect.ImmutableMap;
|
||||
|
||||
/**
|
||||
|
@ -56,7 +60,8 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
doTestIntPointFieldExactQuery("number_p_i_mv", false);
|
||||
doTestIntPointFieldExactQuery("number_p_i_ni_dv", false);
|
||||
doTestIntPointFieldExactQuery("number_p_i_ni_ns_dv", false);
|
||||
doTestIntPointFieldExactQuery("number_p_i_ni_mv_dv", false);
|
||||
// uncomment once MultiValued docValues are supported in PointFields
|
||||
// doTestIntPointFieldExactQuery("number_p_i_ni_mv_dv", false);
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -99,42 +104,110 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
public void testIntPointStats() throws Exception {
|
||||
testPointStats("number_p_i", "number_p_i_dv", new String[]{"0", "1", "2", "3", "4", "5", "6", "7", "8", "9"},
|
||||
0D, 9D, "10", "1", 0D);
|
||||
testPointStats("number_p_i", "number_p_i_mv_dv", new String[]{"0", "1", "2", "3", "4", "5", "6", "7", "8", "9"},
|
||||
0D, 9D, "10", "1", 0D);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testIntPointFieldMultiValuedExactQuery() throws Exception {
|
||||
testPointFieldMultiValuedExactQuery("number_p_i_mv", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedExactQuery("number_p_i_ni_mv_dv", getSequentialStringArrayWithInts(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testIntPointFieldMultiValuedReturn() throws Exception {
|
||||
testPointFieldMultiValuedReturn("number_p_i_mv", "int", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedReturn("number_p_i_ni_mv_dv", "int", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedReturn("number_p_i_dv_ns_mv", "int", getSequentialStringArrayWithInts(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testIntPointFieldMultiValuedRangeQuery() throws Exception {
|
||||
testPointFieldMultiValuedRangeQuery("number_p_i_mv", "int", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedRangeQuery("number_p_i_ni_mv_dv", "int", getSequentialStringArrayWithInts(20));
|
||||
}
|
||||
|
||||
//TODO MV SORT?
|
||||
@Test
|
||||
@Ignore("Enable once MultiValued docValues are supported in PointFields")
|
||||
public void testIntPointFieldMultiValuedFacetField() throws Exception {
|
||||
testPointFieldMultiValuedFacetField("number_p_i_mv", "number_p_i_mv_dv", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedFacetField("number_p_i_mv", "number_p_i_mv_dv", getRandomStringArrayWithInts(20, false));
|
||||
}
|
||||
|
||||
@Test
|
||||
@Ignore("Enable once MultiValued docValues are supported in PointFields")
|
||||
public void testIntPointFieldMultiValuedRangeFacet() throws Exception {
|
||||
doTestIntPointFieldMultiValuedRangeFacet("number_p_i_mv_dv", "number_p_i_mv");
|
||||
String docValuesField = "number_p_i_mv_dv";
|
||||
String nonDocValuesField = "number_p_i_mv";
|
||||
|
||||
for (int i = 0; i < 10; i++) {
|
||||
assertU(adoc("id", String.valueOf(i), docValuesField, String.valueOf(i), docValuesField, String.valueOf(i + 10),
|
||||
nonDocValuesField, String.valueOf(i), nonDocValuesField, String.valueOf(i + 10)));
|
||||
}
|
||||
assertU(commit());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).getType() instanceof IntPointField);
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='2'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='4'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='6'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='8'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='10'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='12'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='14'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='16'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='18'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='-10'][.='0']");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "dv"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='2'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='4'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='6'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='8'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='10'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='12'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='14'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='16'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='18'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='-10'][.='0']");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "0", "facet.range.end", "20", "facet.range.gap", "100"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0'][.='10']");
|
||||
|
||||
assertFalse(h.getCore().getLatestSchema().getField(nonDocValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).getType() instanceof IntPointField);
|
||||
// Range Faceting with method = filter should work
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", nonDocValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "filter"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='2'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='4'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='6'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='8'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='10'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='12'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='14'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='16'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='18'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='-10'][.='0']");
|
||||
|
||||
// this should actually use filter method instead of dv
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", nonDocValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "dv"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='2'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='4'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='6'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='8'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='10'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='12'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='14'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='16'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='18'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='-10'][.='0']");
|
||||
}
|
||||
|
||||
@Test
|
||||
@Ignore("Enable once MultiValued docValues are supported in PointFields")
|
||||
public void testIntPointMultiValuedFunctionQuery() throws Exception {
|
||||
testPointMultiValuedFunctionQuery("number_p_i_mv", "number_p_i_mv_dv", "int", getSequentialStringArrayWithInts(20));
|
||||
}
|
||||
|
@ -149,16 +222,6 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
testIntPointFieldsAtomicUpdates("number_p_i_dv_ns", "int");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testMultiValuedIntPointFieldsAtomicUpdates() throws Exception {
|
||||
if (!Boolean.getBoolean("enable.update.log")) {
|
||||
return;
|
||||
}
|
||||
testMultiValuedIntPointFieldsAtomicUpdates("number_p_i_mv", "int");
|
||||
testMultiValuedIntPointFieldsAtomicUpdates("number_p_i_ni_mv_dv", "int");
|
||||
testMultiValuedIntPointFieldsAtomicUpdates("number_p_i_dv_ns_mv", "int");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testIntPointSetQuery() throws Exception {
|
||||
doTestSetQueries("number_p_i", getRandomStringArrayWithInts(10, false), false);
|
||||
|
@ -175,7 +238,8 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
doTestFloatPointFieldExactQuery("number_p_d_mv");
|
||||
doTestFloatPointFieldExactQuery("number_p_d_ni_dv");
|
||||
doTestFloatPointFieldExactQuery("number_p_d_ni_ns_dv");
|
||||
doTestFloatPointFieldExactQuery("number_p_d_ni_mv_dv");
|
||||
// TODO enable once MuultiValued docValues are supported with PointFields
|
||||
// doTestFloatPointFieldExactQuery("number_p_d_ni_mv_dv");
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -228,41 +292,112 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
public void testDoublePointStats() throws Exception {
|
||||
testPointStats("number_p_d", "number_p_d_dv", new String[]{"-10.0", "1.1", "2.2", "3.3", "4.4", "5.5", "6.6", "7.7", "8.8", "9.9"},
|
||||
-10.0D, 9.9D, "10", "1", 1E-10D);
|
||||
testPointStats("number_p_d_mv", "number_p_d_mv_dv", new String[]{"-10.0", "1.1", "2.2", "3.3", "4.4", "5.5", "6.6", "7.7", "8.8", "9.9"},
|
||||
-10.0D, 9.9D, "10", "1", 1E-10D);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testDoublePointFieldMultiValuedExactQuery() throws Exception {
|
||||
testPointFieldMultiValuedExactQuery("number_p_d_mv", getRandomStringArrayWithDoubles(20, false));
|
||||
testPointFieldMultiValuedExactQuery("number_p_d_ni_mv_dv", getRandomStringArrayWithDoubles(20, false));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testDoublePointFieldMultiValuedReturn() throws Exception {
|
||||
testPointFieldMultiValuedReturn("number_p_d_mv", "double", getSequentialStringArrayWithDoubles(20));
|
||||
testPointFieldMultiValuedReturn("number_p_d_ni_mv_dv", "double", getSequentialStringArrayWithDoubles(20));
|
||||
testPointFieldMultiValuedReturn("number_p_d_dv_ns_mv", "double", getSequentialStringArrayWithDoubles(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testDoublePointFieldMultiValuedRangeQuery() throws Exception {
|
||||
testPointFieldMultiValuedRangeQuery("number_p_d_mv", "double", getSequentialStringArrayWithDoubles(20));
|
||||
testPointFieldMultiValuedRangeQuery("number_p_d_ni_mv_dv", "double", getSequentialStringArrayWithDoubles(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
@Ignore("Enable once MultiValued docValues are supported in PointFields")
|
||||
public void testDoublePointFieldMultiValuedFacetField() throws Exception {
|
||||
testPointFieldMultiValuedFacetField("number_p_d_mv", "number_p_d_mv_dv", getSequentialStringArrayWithDoubles(20));
|
||||
testPointFieldMultiValuedFacetField("number_p_d_mv", "number_p_d_mv_dv", getRandomStringArrayWithDoubles(20, false));
|
||||
}
|
||||
|
||||
@Test
|
||||
@Ignore("Enable once MultiValued docValues are supported in PointFields")
|
||||
public void testDoublePointFieldMultiValuedRangeFacet() throws Exception {
|
||||
doTestDoublePointFieldMultiValuedRangeFacet("number_p_d_mv_dv", "number_p_d_mv");
|
||||
String docValuesField = "number_p_d_mv_dv";
|
||||
String nonDocValuesField = "number_p_d_mv";
|
||||
|
||||
for (int i = 0; i < 10; i++) {
|
||||
assertU(adoc("id", String.valueOf(i), docValuesField, String.valueOf(i), docValuesField, String.valueOf(i + 10),
|
||||
nonDocValuesField, String.valueOf(i), nonDocValuesField, String.valueOf(i + 10)));
|
||||
}
|
||||
assertU(commit());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).multiValued());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).getType() instanceof DoublePointField);
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='2.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='4.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='6.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='8.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='10.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='12.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='14.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='16.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='18.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='-10.0'][.='0']");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "dv"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='2.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='4.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='6.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='8.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='10.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='12.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='14.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='16.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='18.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='-10.0'][.='0']");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "0", "facet.range.end", "20", "facet.range.gap", "100"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='10']");
|
||||
|
||||
assertFalse(h.getCore().getLatestSchema().getField(nonDocValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).multiValued());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).getType() instanceof DoublePointField);
|
||||
// Range Faceting with method = filter should work
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", nonDocValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "filter"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='2.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='4.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='6.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='8.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='10.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='12.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='14.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='16.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='18.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='-10.0'][.='0']");
|
||||
|
||||
// this should actually use filter method instead of dv
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", nonDocValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "dv"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='2.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='4.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='6.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='8.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='10.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='12.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='14.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='16.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='18.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='-10.0'][.='0']");
|
||||
}
|
||||
|
||||
@Test
|
||||
@Ignore("Enable once MultiValued docValues are supported in PointFields")
|
||||
public void testDoublePointMultiValuedFunctionQuery() throws Exception {
|
||||
testPointMultiValuedFunctionQuery("number_p_d_mv", "number_p_d_mv_dv", "double", getSequentialStringArrayWithDoubles(20));
|
||||
testPointMultiValuedFunctionQuery("number_p_d_mv", "number_p_d_mv_dv", "double", getRandomStringArrayWithFloats(20, true));
|
||||
|
@ -278,17 +413,6 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
doTestFloatPointFieldsAtomicUpdates("number_p_d_dv_ns", "double");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testMultiValuedDoublePointFieldsAtomicUpdates() throws Exception {
|
||||
if (!Boolean.getBoolean("enable.update.log")) {
|
||||
return;
|
||||
}
|
||||
testMultiValuedFloatPointFieldsAtomicUpdates("number_p_d_mv", "double");
|
||||
testMultiValuedFloatPointFieldsAtomicUpdates("number_p_d_ni_mv_dv", "double");
|
||||
testMultiValuedFloatPointFieldsAtomicUpdates("number_p_d_dv_ns_mv", "double");
|
||||
}
|
||||
|
||||
|
||||
private void doTestFloatPointFieldsAtomicUpdates(String field, String type) throws Exception {
|
||||
assertU(adoc(sdoc("id", "1", field, "1.1234")));
|
||||
assertU(commit());
|
||||
|
@ -330,6 +454,7 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
}
|
||||
|
||||
// Float
|
||||
|
||||
|
||||
@Test
|
||||
public void testFloatPointFieldExactQuery() throws Exception {
|
||||
|
@ -337,7 +462,7 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
doTestFloatPointFieldExactQuery("number_p_f_mv");
|
||||
doTestFloatPointFieldExactQuery("number_p_f_ni_dv");
|
||||
doTestFloatPointFieldExactQuery("number_p_f_ni_ns_dv");
|
||||
doTestFloatPointFieldExactQuery("number_p_f_ni_mv_dv");
|
||||
// doTestFloatPointFieldExactQuery("number_p_f_ni_mv_dv");
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -390,47 +515,23 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
public void testFloatPointStats() throws Exception {
|
||||
testPointStats("number_p_f", "number_p_f_dv", new String[]{"-10.0", "1.1", "2.2", "3.3", "4.4", "5.5", "6.6", "7.7", "8.8", "9.9"},
|
||||
-10D, 9.9D, "10", "1", 1E-6D);
|
||||
testPointStats("number_p_f_mv", "number_p_f_mv_dv", new String[]{"-10.0", "1.1", "2.2", "3.3", "4.4", "5.5", "6.6", "7.7", "8.8", "9.9"},
|
||||
-10D, 9.9D, "10", "1", 1E-6D);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFloatPointFieldMultiValuedExactQuery() throws Exception {
|
||||
testPointFieldMultiValuedExactQuery("number_p_f_mv", getRandomStringArrayWithFloats(20, false));
|
||||
testPointFieldMultiValuedExactQuery("number_p_f_ni_mv_dv", getRandomStringArrayWithFloats(20, false));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFloatPointFieldMultiValuedReturn() throws Exception {
|
||||
testPointFieldMultiValuedReturn("number_p_f_mv", "float", getSequentialStringArrayWithDoubles(20));
|
||||
testPointFieldMultiValuedReturn("number_p_f_ni_mv_dv", "float", getSequentialStringArrayWithDoubles(20));
|
||||
testPointFieldMultiValuedReturn("number_p_f_dv_ns_mv", "float", getSequentialStringArrayWithDoubles(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFloatPointFieldMultiValuedRangeQuery() throws Exception {
|
||||
testPointFieldMultiValuedRangeQuery("number_p_f_mv", "float", getSequentialStringArrayWithDoubles(20));
|
||||
testPointFieldMultiValuedRangeQuery("number_p_f_ni_mv_dv", "float", getSequentialStringArrayWithDoubles(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFloatPointFieldMultiValuedRangeFacet() throws Exception {
|
||||
doTestDoublePointFieldMultiValuedRangeFacet("number_p_f_mv_dv", "number_p_f_mv");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFloatPointFieldMultiValuedFacetField() throws Exception {
|
||||
testPointFieldMultiValuedFacetField("number_p_f_mv", "number_p_f_mv_dv", getSequentialStringArrayWithDoubles(20));
|
||||
testPointFieldMultiValuedFacetField("number_p_f_mv", "number_p_f_mv_dv", getRandomStringArrayWithFloats(20, false));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFloatPointMultiValuedFunctionQuery() throws Exception {
|
||||
testPointMultiValuedFunctionQuery("number_p_f_mv", "number_p_f_mv_dv", "float", getSequentialStringArrayWithDoubles(20));
|
||||
testPointMultiValuedFunctionQuery("number_p_f_mv", "number_p_f_mv_dv", "float", getRandomStringArrayWithFloats(20, true));
|
||||
}
|
||||
|
||||
|
||||
@Test
|
||||
public void testFloatPointFieldsAtomicUpdates() throws Exception {
|
||||
if (!Boolean.getBoolean("enable.update.log")) {
|
||||
|
@ -441,15 +542,6 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
doTestFloatPointFieldsAtomicUpdates("number_p_f_dv_ns", "float");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testMultiValuedFloatePointFieldsAtomicUpdates() throws Exception {
|
||||
if (!Boolean.getBoolean("enable.update.log")) {
|
||||
return;
|
||||
}
|
||||
testMultiValuedFloatPointFieldsAtomicUpdates("number_p_f_mv", "float");
|
||||
testMultiValuedFloatPointFieldsAtomicUpdates("number_p_f_ni_mv_dv", "float");
|
||||
testMultiValuedFloatPointFieldsAtomicUpdates("number_p_f_dv_ns_mv", "float");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testFloatPointSetQuery() throws Exception {
|
||||
|
@ -466,7 +558,7 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
doTestIntPointFieldExactQuery("number_p_l_mv", true);
|
||||
doTestIntPointFieldExactQuery("number_p_l_ni_dv", true);
|
||||
doTestIntPointFieldExactQuery("number_p_l_ni_ns_dv", true);
|
||||
doTestIntPointFieldExactQuery("number_p_l_ni_mv_dv", true);
|
||||
// doTestIntPointFieldExactQuery("number_p_i_ni_mv_dv", true);
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -512,43 +604,21 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
public void testLongPointStats() throws Exception {
|
||||
testPointStats("number_p_l", "number_p_l_dv", new String[]{"0", "1", "2", "3", "4", "5", "6", "7", "8", "9"},
|
||||
0D, 9D, "10", "1", 0D);
|
||||
testPointStats("number_p_l_mv", "number_p_l_mv_dv", new String[]{"0", "1", "2", "3", "4", "5", "6", "7", "8", "9"},
|
||||
0D, 9D, "10", "1", 0D);
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testLongPointFieldMultiValuedExactQuery() throws Exception {
|
||||
testPointFieldMultiValuedExactQuery("number_p_l_mv", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedExactQuery("number_p_l_ni_mv_dv", getSequentialStringArrayWithInts(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testLongPointFieldMultiValuedReturn() throws Exception {
|
||||
testPointFieldMultiValuedReturn("number_p_l_mv", "long", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedReturn("number_p_l_ni_mv_dv", "long", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedReturn("number_p_l_dv_ns_mv", "long", getSequentialStringArrayWithInts(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testLongPointFieldMultiValuedRangeQuery() throws Exception {
|
||||
testPointFieldMultiValuedRangeQuery("number_p_l_mv", "long", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedRangeQuery("number_p_l_ni_mv_dv", "long", getSequentialStringArrayWithInts(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testLongPointFieldMultiValuedFacetField() throws Exception {
|
||||
testPointFieldMultiValuedFacetField("number_p_l_mv", "number_p_l_mv_dv", getSequentialStringArrayWithInts(20));
|
||||
testPointFieldMultiValuedFacetField("number_p_l_mv", "number_p_l_mv_dv", getRandomStringArrayWithLongs(20, false));
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testLongPointFieldMultiValuedRangeFacet() throws Exception {
|
||||
doTestIntPointFieldMultiValuedRangeFacet("number_p_l_mv_dv", "number_p_l_mv");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testLongPointMultiValuedFunctionQuery() throws Exception {
|
||||
testPointMultiValuedFunctionQuery("number_p_l_mv", "number_p_l_mv_dv", "long", getSequentialStringArrayWithInts(20));
|
||||
}
|
||||
|
||||
@Test
|
||||
|
@ -561,16 +631,6 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
testIntPointFieldsAtomicUpdates("number_p_l_dv_ns", "long");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testMultiValuedLongPointFieldsAtomicUpdates() throws Exception {
|
||||
if (!Boolean.getBoolean("enable.update.log")) {
|
||||
return;
|
||||
}
|
||||
testMultiValuedIntPointFieldsAtomicUpdates("number_p_l_mv", "long");
|
||||
testMultiValuedIntPointFieldsAtomicUpdates("number_p_l_ni_mv_dv", "long");
|
||||
testMultiValuedIntPointFieldsAtomicUpdates("number_p_l_dv_ns_mv", "long");
|
||||
}
|
||||
|
||||
@Test
|
||||
public void testLongPointSetQuery() throws Exception {
|
||||
doTestSetQueries("number_p_l", getRandomStringArrayWithLongs(10, false), false);
|
||||
|
@ -850,14 +910,6 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + docValuesField +"']/int[@name='" + numbers[2] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + docValuesField +"']/int[@name='" + numbers[3] + "'][.='1']");
|
||||
|
||||
// assertU(commit());
|
||||
// assertQ(req("q", "id:0", "fl", "id, " + docValuesField, "facet", "true", "facet.field", docValuesField, "facet.mincount", "0"),
|
||||
// "//*[@numFound='1']",
|
||||
// "//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + docValuesField +"']/int[@name='" + numbers[0] + "'][.='1']",
|
||||
// "//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + docValuesField +"']/int[@name='" + numbers[1] + "'][.='0']",
|
||||
// "//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + docValuesField +"']/int[@name='" + numbers[2] + "'][.='0']",
|
||||
// "count(//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + docValuesField +"']/int))==10");
|
||||
|
||||
assertFalse(h.getCore().getLatestSchema().getField(nonDocValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).getType() instanceof PointField);
|
||||
assertQEx("Expecting Exception",
|
||||
|
@ -1005,25 +1057,7 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
for (int i=0; i < 10; i++) {
|
||||
assertU(adoc("id", String.valueOf(i), fieldName, numbers[i], fieldName, numbers[i+10]));
|
||||
}
|
||||
// Check using RTG before commit
|
||||
if (Boolean.getBoolean("enable.update.log")) {
|
||||
for (int i = 0; i < 10; i++) {
|
||||
assertQ(req("qt", "/get", "id", String.valueOf(i)),
|
||||
"//doc/arr[@name='" + fieldName + "']/" + type + "[1][.='" + numbers[i] + "']",
|
||||
"//doc/arr[@name='" + fieldName + "']/" + type + "[2][.='" + numbers[i+10] + "']",
|
||||
"count(//doc/arr[@name='" + fieldName + "']/" + type + ")=2");
|
||||
}
|
||||
}
|
||||
// Check using RTG after commit
|
||||
assertU(commit());
|
||||
if (Boolean.getBoolean("enable.update.log")) {
|
||||
for (int i = 0; i < 10; i++) {
|
||||
assertQ(req("qt", "/get", "id", String.valueOf(i)),
|
||||
"//doc/arr[@name='" + fieldName + "']/" + type + "[1][.='" + numbers[i] + "']",
|
||||
"//doc/arr[@name='" + fieldName + "']/" + type + "[2][.='" + numbers[i+10] + "']",
|
||||
"count(//doc/arr[@name='" + fieldName + "']/" + type + ")=2");
|
||||
}
|
||||
}
|
||||
String[] expected = new String[11];
|
||||
String[] expected2 = new String[11];
|
||||
expected[0] = "//*[@numFound='10']";
|
||||
|
@ -1103,9 +1137,6 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
for (int i = 0; i < 10; i++) {
|
||||
assertU(adoc("id", String.valueOf(i), dvFieldName, numbers[i], dvFieldName, numbers[i + 10],
|
||||
nonDocValuesField, numbers[i], nonDocValuesField, numbers[i + 10]));
|
||||
if (rarely()) {
|
||||
assertU(commit());
|
||||
}
|
||||
}
|
||||
assertU(commit());
|
||||
|
||||
|
@ -1128,85 +1159,12 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[3] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[10] + "'][.='1']");
|
||||
|
||||
assertU(adoc("id", "10", dvFieldName, numbers[1], nonDocValuesField, numbers[1], dvFieldName, numbers[1], nonDocValuesField, numbers[1]));
|
||||
assertU(commit());
|
||||
assertQ(req("q", "*:*", "fl", "id, " + dvFieldName, "facet", "true", "facet.field", dvFieldName, "facet.missing", "true"),
|
||||
"//*[@numFound='11']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[1] + "'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[2] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[3] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[10] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[not(@name)][.='0']"
|
||||
);
|
||||
|
||||
assertU(adoc("id", "10")); // add missing values
|
||||
assertU(commit());
|
||||
assertQ(req("q", "*:*", "fl", "id, " + dvFieldName, "facet", "true", "facet.field", dvFieldName, "facet.missing", "true"),
|
||||
"//*[@numFound='11']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[1] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[2] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[3] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[10] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[not(@name)][.='1']"
|
||||
);
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id, " + dvFieldName, "facet", "true", "facet.field", dvFieldName, "facet.mincount", "3"),
|
||||
"//*[@numFound='11']",
|
||||
"count(//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int)=0");
|
||||
|
||||
assertQ(req("q", "id:0", "fl", "id, " + dvFieldName, "facet", "true", "facet.field", dvFieldName),
|
||||
"//*[@numFound='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[0] + "'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + numbers[10] + "'][.='1']",
|
||||
"count(//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int)=2");
|
||||
|
||||
assertFalse(h.getCore().getLatestSchema().getField(nonDocValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).getType() instanceof PointField);
|
||||
assertQEx("Expecting Exception",
|
||||
"Can't facet on a PointField without docValues",
|
||||
req("q", "*:*", "fl", "id, " + nonDocValuesField, "facet", "true", "facet.field", nonDocValuesField),
|
||||
SolrException.ErrorCode.BAD_REQUEST);
|
||||
clearIndex();
|
||||
assertU(commit());
|
||||
|
||||
String smaller, larger;
|
||||
try {
|
||||
if (Long.parseLong(numbers[1]) < Long.parseLong(numbers[2])) {
|
||||
smaller = numbers[1];
|
||||
larger = numbers[2];
|
||||
} else {
|
||||
smaller = numbers[2];
|
||||
larger = numbers[1];
|
||||
}
|
||||
} catch (NumberFormatException e) {
|
||||
if (Double.valueOf(numbers[1]) < Double.valueOf(numbers[2])) {
|
||||
smaller = numbers[1];
|
||||
larger = numbers[2];
|
||||
} else {
|
||||
smaller = numbers[2];
|
||||
larger = numbers[1];
|
||||
}
|
||||
}
|
||||
|
||||
assertU(adoc("id", "1", dvFieldName, smaller, dvFieldName, larger));
|
||||
assertU(adoc("id", "2", dvFieldName, larger));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id, " + dvFieldName, "facet", "true", "facet.field", dvFieldName),
|
||||
"//*[@numFound='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + larger + "'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + smaller + "'][.='1']",
|
||||
"count(//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int)=2");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id, " + dvFieldName, "facet", "true", "facet.field", dvFieldName, "facet.sort", "index"),
|
||||
"//*[@numFound='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='" + smaller +"'][.='1']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int[@name='"+ larger + "'][.='2']",
|
||||
"count(//lst[@name='facet_counts']/lst[@name='facet_fields']/lst[@name='" + dvFieldName +"']/int)=2");
|
||||
|
||||
clearIndex();
|
||||
assertU(commit());
|
||||
|
||||
}
|
||||
|
||||
private void testPointMultiValuedFunctionQuery(String nonDocValuesField, String docValuesField, String type, String[] numbers) throws Exception {
|
||||
|
@ -1221,20 +1179,23 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).getType() instanceof PointField);
|
||||
String function = "field(" + docValuesField + ", min)";
|
||||
|
||||
// assertQ(req("q", "*:*", "fl", "id, " + function),
|
||||
// "//*[@numFound='10']",
|
||||
// "//result/doc[1]/" + type + "[@name='" + function + "'][.='" + numbers[0] + "']",
|
||||
// "//result/doc[2]/" + type + "[@name='" + function + "'][.='" + numbers[1] + "']",
|
||||
// "//result/doc[3]/" + type + "[@name='" + function + "'][.='" + numbers[2] + "']",
|
||||
// "//result/doc[10]/" + type + "[@name='" + function + "'][.='" + numbers[9] + "']");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id, " + docValuesField, "sort", function + " desc"),
|
||||
assertQ(req("q", "*:*", "fl", "id, " + function),
|
||||
"//*[@numFound='10']",
|
||||
"//result/doc[1]/str[@name='id'][.='9']",
|
||||
"//result/doc[2]/str[@name='id'][.='8']",
|
||||
"//result/doc[3]/str[@name='id'][.='7']",
|
||||
"//result/doc[10]/str[@name='id'][.='0']");
|
||||
|
||||
"//result/doc[1]/" + type + "[@name='" + function + "'][.='" + numbers[0] + "']",
|
||||
"//result/doc[2]/" + type + "[@name='" + function + "'][.='" + numbers[1] + "']",
|
||||
"//result/doc[3]/" + type + "[@name='" + function + "'][.='" + numbers[2] + "']",
|
||||
"//result/doc[10]/" + type + "[@name='" + function + "'][.='" + numbers[9] + "']");
|
||||
|
||||
// if (dvIsRandomAccessOrds(docValuesField)) {
|
||||
// function = "field(" + docValuesField + ", max)";
|
||||
// assertQ(req("q", "*:*", "fl", "id, " + function),
|
||||
// "//*[@numFound='10']",
|
||||
// "//result/doc[1]/int[@name='" + function + "'][.='10']",
|
||||
// "//result/doc[2]/int[@name='" + function + "'][.='11']",
|
||||
// "//result/doc[3]/int[@name='" + function + "'][.='12']",
|
||||
// "//result/doc[10]/int[@name='" + function + "'][.='19']");
|
||||
// }
|
||||
|
||||
assertFalse(h.getCore().getLatestSchema().getField(nonDocValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).multiValued());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).getType() instanceof PointField);
|
||||
|
@ -1258,86 +1219,6 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
SolrException.ErrorCode.BAD_REQUEST);
|
||||
}
|
||||
|
||||
private void testMultiValuedIntPointFieldsAtomicUpdates(String field, String type) throws Exception {
|
||||
assertU(adoc(sdoc("id", "1", field, "1")));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='1']",
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=1");
|
||||
|
||||
assertU(adoc(sdoc("id", "1", field, ImmutableMap.of("add", 2))));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='1']",
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='2']",
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=2");
|
||||
|
||||
assertU(adoc(sdoc("id", "1", field, ImmutableMap.of("remove", 1))));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='2']",
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=1");
|
||||
|
||||
assertU(adoc(sdoc("id", "1", field, ImmutableMap.of("set", ImmutableList.of(1, 2, 3)))));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='1']",
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='2']",
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='3']",
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=3");
|
||||
|
||||
assertU(adoc(sdoc("id", "1", field, ImmutableMap.of("removeregex", ".*"))));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=0");
|
||||
|
||||
}
|
||||
|
||||
private void testMultiValuedFloatPointFieldsAtomicUpdates(String field, String type) throws Exception {
|
||||
assertU(adoc(sdoc("id", "1", field, "1.0")));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='1.0']",
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=1");
|
||||
|
||||
assertU(adoc(sdoc("id", "1", field, ImmutableMap.of("add", 2.1f))));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='1.0']",
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='2.1']",
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=2");
|
||||
|
||||
assertU(adoc(sdoc("id", "1", field, ImmutableMap.of("remove", 1f))));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='2.1']",
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=1");
|
||||
|
||||
assertU(adoc(sdoc("id", "1", field, ImmutableMap.of("set", ImmutableList.of(1f, 2f, 3f)))));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='1.0']",
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='2.0']",
|
||||
"//result/doc[1]/arr[@name='" + field + "']/" + type + "[.='3.0']",
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=3");
|
||||
|
||||
assertU(adoc(sdoc("id", "1", field, ImmutableMap.of("removeregex", ".*"))));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "id:1"),
|
||||
"count(//result/doc[1]/arr[@name='" + field + "']/" + type + ")=0");
|
||||
|
||||
}
|
||||
|
||||
private void testIntPointFieldsAtomicUpdates(String field, String type) throws Exception {
|
||||
assertU(adoc(sdoc("id", "1", field, "1")));
|
||||
assertU(commit());
|
||||
|
@ -1611,152 +1492,4 @@ public class TestPointFields extends SolrTestCaseJ4 {
|
|||
}
|
||||
}
|
||||
}
|
||||
|
||||
private void doTestDoublePointFieldMultiValuedRangeFacet(String docValuesField, String nonDocValuesField) throws Exception {
|
||||
for (int i = 0; i < 10; i++) {
|
||||
assertU(adoc("id", String.valueOf(i), docValuesField, String.valueOf(i), docValuesField, String.valueOf(i + 10),
|
||||
nonDocValuesField, String.valueOf(i), nonDocValuesField, String.valueOf(i + 10)));
|
||||
}
|
||||
assertU(commit());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).multiValued());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).getType() instanceof PointField);
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='2.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='4.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='6.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='8.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='10.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='12.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='14.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='16.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='18.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='-10.0'][.='0']");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "dv"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='2.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='4.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='6.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='8.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='10.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='12.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='14.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='16.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='18.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='-10.0'][.='0']");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "0", "facet.range.end", "20", "facet.range.gap", "100"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='10']");
|
||||
|
||||
assertFalse(h.getCore().getLatestSchema().getField(nonDocValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).multiValued());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).getType() instanceof PointField);
|
||||
// Range Faceting with method = filter should work
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", nonDocValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "filter"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='2.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='4.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='6.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='8.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='10.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='12.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='14.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='16.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='18.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='-10.0'][.='0']");
|
||||
|
||||
// this should actually use filter method instead of dv
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", nonDocValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "dv"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='0.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='2.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='4.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='6.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='8.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='10.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='12.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='14.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='16.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='18.0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='-10.0'][.='0']");
|
||||
}
|
||||
|
||||
private void doTestIntPointFieldMultiValuedRangeFacet(String docValuesField, String nonDocValuesField) throws Exception {
|
||||
for (int i = 0; i < 10; i++) {
|
||||
assertU(adoc("id", String.valueOf(i), docValuesField, String.valueOf(i), docValuesField, String.valueOf(i + 10),
|
||||
nonDocValuesField, String.valueOf(i), nonDocValuesField, String.valueOf(i + 10)));
|
||||
}
|
||||
assertU(commit());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(docValuesField).getType() instanceof PointField);
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='2'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='4'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='6'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='8'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='10'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='12'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='14'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='16'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='18'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='-10'][.='0']");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "dv"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='2'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='4'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='6'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='8'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='10'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='12'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='14'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='16'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='18'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='-10'][.='0']");
|
||||
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", docValuesField, "facet.range.start", "0", "facet.range.end", "20", "facet.range.gap", "100"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + docValuesField + "']/lst[@name='counts']/int[@name='0'][.='10']");
|
||||
|
||||
assertFalse(h.getCore().getLatestSchema().getField(nonDocValuesField).hasDocValues());
|
||||
assertTrue(h.getCore().getLatestSchema().getField(nonDocValuesField).getType() instanceof PointField);
|
||||
// Range Faceting with method = filter should work
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", nonDocValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "filter"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='2'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='4'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='6'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='8'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='10'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='12'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='14'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='16'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='18'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='-10'][.='0']");
|
||||
|
||||
// this should actually use filter method instead of dv
|
||||
assertQ(req("q", "*:*", "fl", "id", "facet", "true", "facet.range", nonDocValuesField, "facet.range.start", "-10", "facet.range.end", "20", "facet.range.gap", "2", "facet.range.method", "dv"),
|
||||
"//*[@numFound='10']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='0'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='2'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='4'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='6'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='8'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='10'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='12'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='14'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='16'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='18'][.='2']",
|
||||
"//lst[@name='facet_counts']/lst[@name='facet_ranges']/lst[@name='" + nonDocValuesField + "']/lst[@name='counts']/int[@name='-10'][.='0']");
|
||||
}
|
||||
}
|
||||
|
|
|
@ -228,7 +228,7 @@ public class TestSolrQueryParser extends SolrTestCaseJ4 {
|
|||
assertEquals(26, ((TermInSetQuery)q).getTermData().size());
|
||||
|
||||
// large numeric filter query should use TermsQuery (for trie fields)
|
||||
qParser = QParser.getParser("foo_ti:(1 2 3 4 5 6 7 8 9 10 20 19 18 17 16 15 14 13 12 11)", req);
|
||||
qParser = QParser.getParser("foo_i:(1 2 3 4 5 6 7 8 9 10 20 19 18 17 16 15 14 13 12 11)", req);
|
||||
qParser.setIsFilter(true); // this may change in the future
|
||||
q = qParser.getQuery();
|
||||
assertEquals(20, ((TermInSetQuery)q).getTermData().size());
|
||||
|
|
|
@ -140,7 +140,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '3']");
|
||||
|
||||
|
||||
|
@ -153,7 +153,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '2']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -165,7 +165,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '1']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -175,7 +175,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:111", "indent", "true"), "//result[@numFound = '3']");
|
||||
}
|
||||
|
||||
|
@ -208,7 +208,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '3']");
|
||||
|
||||
|
||||
|
@ -221,7 +221,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '2']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -233,7 +233,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '1']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -243,7 +243,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:111", "indent", "true"), "//result[@numFound = '3']");
|
||||
}
|
||||
|
||||
|
@ -272,7 +272,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '3']");
|
||||
|
||||
|
||||
|
@ -285,7 +285,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '2']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -297,7 +297,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '1']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -307,7 +307,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:111", "indent", "true"), "//result[@numFound = '3']");
|
||||
}
|
||||
|
||||
|
@ -337,7 +337,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '3']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -349,7 +349,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '2']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -361,7 +361,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '1']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -371,7 +371,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:111", "indent", "true"), "//result[@numFound = '3']");
|
||||
}
|
||||
|
||||
|
@ -421,7 +421,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '3']");
|
||||
|
||||
|
||||
|
@ -434,7 +434,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '2']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -446,7 +446,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:222", "indent", "true"), "//result[@numFound = '1']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -456,7 +456,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:111", "indent", "true"), "//result[@numFound = '3']");
|
||||
}
|
||||
|
||||
|
@ -487,7 +487,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:22222222", "indent", "true"), "//result[@numFound = '3']");
|
||||
|
||||
|
||||
|
@ -500,7 +500,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:22222222", "indent", "true"), "//result[@numFound = '2']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -512,7 +512,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:22222222", "indent", "true"), "//result[@numFound = '1']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -522,7 +522,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "intRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "intRemove:11111111", "indent", "true"), "//result[@numFound = '3']");
|
||||
}
|
||||
|
||||
|
@ -708,7 +708,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "floatRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:\"222.222\"", "indent", "true"), "//result[@numFound = '3']");
|
||||
|
||||
|
||||
|
@ -722,7 +722,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "floatRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:\"222.222\"", "indent", "true"), "//result[@numFound = '2']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -734,7 +734,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "floatRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:\"222.222\"", "indent", "true"), "//result[@numFound = '1']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -744,7 +744,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "floatRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:\"111.111\"", "indent", "true"), "//result[@numFound = '3']");
|
||||
}
|
||||
|
||||
|
@ -776,7 +776,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "floatRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:\"222.222\"", "indent", "true"), "//result[@numFound = '3']");
|
||||
|
||||
|
||||
|
@ -790,7 +790,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "floatRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:\"222.222\"", "indent", "true"), "//result[@numFound = '2']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -802,7 +802,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "floatRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:\"222.222\"", "indent", "true"), "//result[@numFound = '1']");
|
||||
|
||||
doc = new SolrInputDocument();
|
||||
|
@ -812,7 +812,7 @@ public class AtomicUpdatesTest extends SolrTestCaseJ4 {
|
|||
assertU(adoc(doc));
|
||||
assertU(commit());
|
||||
|
||||
assertQ(req("q", "floatRemove:[* TO *]", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:*", "indent", "true"), "//result[@numFound = '4']");
|
||||
assertQ(req("q", "floatRemove:\"111.111\"", "indent", "true"), "//result[@numFound = '3']");
|
||||
}
|
||||
|
||||
|
|
|
@ -17,7 +17,7 @@
|
|||
under the License.
|
||||
-->
|
||||
<ivy-module version="2.0">
|
||||
<info organisation="org.apache.solr" module="test-framework"/>
|
||||
<info organisation="org.apache.solr" module="solr-test-framework"/>
|
||||
|
||||
<configurations defaultconfmapping="compile->master;junit4-stdalone->master">
|
||||
<conf name="compile" transitive="false" />
|
||||
|
|
|
@ -2474,8 +2474,4 @@ public abstract class SolrTestCaseJ4 extends LuceneTestCase {
|
|||
protected static void systemClearPropertySolrTestsMergePolicyFactory() {
|
||||
System.clearProperty(SYSTEM_PROPERTY_SOLR_TESTS_MERGEPOLICYFACTORY);
|
||||
}
|
||||
|
||||
protected <T> T pickRandom(T... options) {
|
||||
return options[random().nextInt(options.length)];
|
||||
}
|
||||
}
|
||||
|
|
|
@ -457,7 +457,7 @@ public class MiniSolrCloudCluster {
|
|||
}
|
||||
} finally {
|
||||
executor.shutdown();
|
||||
executor.awaitTermination(15, TimeUnit.SECONDS);
|
||||
executor.awaitTermination(2, TimeUnit.SECONDS);
|
||||
try {
|
||||
if (!externalZkServer) {
|
||||
zkServer.shutdown();
|
||||
|
|
Loading…
Reference in New Issue