rename tests

This commit is contained in:
kimchy 2011-01-25 14:40:04 +02:00
parent a5e02d559f
commit 00aad4c123
5 changed files with 79 additions and 17 deletions

View File

@ -423,6 +423,10 @@ public class SearchService extends AbstractLifecycleComponent<SearchService> {
private static final int[] EMPTY_DOC_IDS = new int[0]; private static final int[] EMPTY_DOC_IDS = new int[0];
/**
* Shortcut ids to load, we load only "from" and up to "size". The phase controller
* handles this as well since the result is always size * shards for Q_A_F
*/
private void shortcutDocIdsToLoad(SearchContext context) { private void shortcutDocIdsToLoad(SearchContext context) {
TopDocs topDocs = context.queryResult().topDocs(); TopDocs topDocs = context.queryResult().topDocs();
if (topDocs.scoreDocs.length < context.from()) { if (topDocs.scoreDocs.length < context.from()) {

View File

@ -98,6 +98,7 @@ public class SearchPhaseController {
int queueSize = queryResultProvider.queryResult().from() + queryResultProvider.queryResult().size(); int queueSize = queryResultProvider.queryResult().from() + queryResultProvider.queryResult().size();
if (queryResultProvider.includeFetch()) { if (queryResultProvider.includeFetch()) {
// if we did both query and fetch on the same go, we have fetched all the docs from each shards already, use them... // if we did both query and fetch on the same go, we have fetched all the docs from each shards already, use them...
// this is also important since we shortcut and fetch only docs from "from" and up to "size"
queueSize *= results.size(); queueSize *= results.size();
} }
PriorityQueue queue; PriorityQueue queue;
@ -173,6 +174,8 @@ public class SearchPhaseController {
return EMPTY; return EMPTY;
} }
// we only pop the first, this handles "from" nicely since the "from" are down the queue
// that we already fetched, so we are actually popping the "from" and up to "size"
ShardDoc[] shardDocs = new ShardDoc[resultDocsSize]; ShardDoc[] shardDocs = new ShardDoc[resultDocsSize];
for (int i = resultDocsSize - 1; i >= 0; i--) // put docs in array for (int i = resultDocsSize - 1; i >= 0; i--) // put docs in array
shardDocs[i] = (ShardDoc) queue.pop(); shardDocs[i] = (ShardDoc) queue.pop();

View File

@ -20,13 +20,11 @@
package org.elasticsearch.test.integration.search.embedded; package org.elasticsearch.test.integration.search.embedded;
import org.elasticsearch.client.Client; import org.elasticsearch.client.Client;
import org.elasticsearch.common.collect.Sets;
import org.elasticsearch.common.trove.ExtTIntArrayList; import org.elasticsearch.common.trove.ExtTIntArrayList;
import org.elasticsearch.common.unit.TimeValue; import org.elasticsearch.common.unit.TimeValue;
import org.elasticsearch.node.internal.InternalNode; import org.elasticsearch.node.internal.InternalNode;
import org.elasticsearch.search.Scroll; import org.elasticsearch.search.*;
import org.elasticsearch.search.SearchContextMissingException;
import org.elasticsearch.search.SearchService;
import org.elasticsearch.search.SearchShardTarget;
import org.elasticsearch.search.builder.SearchSourceBuilder; import org.elasticsearch.search.builder.SearchSourceBuilder;
import org.elasticsearch.search.controller.SearchPhaseController; import org.elasticsearch.search.controller.SearchPhaseController;
import org.elasticsearch.search.controller.ShardDoc; import org.elasticsearch.search.controller.ShardDoc;
@ -40,12 +38,14 @@ import org.elasticsearch.search.fetch.QueryFetchSearchResult;
import org.elasticsearch.search.internal.InternalSearchRequest; import org.elasticsearch.search.internal.InternalSearchRequest;
import org.elasticsearch.search.query.QuerySearchRequest; import org.elasticsearch.search.query.QuerySearchRequest;
import org.elasticsearch.search.query.QuerySearchResult; import org.elasticsearch.search.query.QuerySearchResult;
import org.elasticsearch.search.sort.SortOrder;
import org.elasticsearch.test.integration.AbstractNodesTests; import org.elasticsearch.test.integration.AbstractNodesTests;
import org.testng.annotations.AfterClass; import org.testng.annotations.AfterClass;
import org.testng.annotations.BeforeClass; import org.testng.annotations.BeforeClass;
import org.testng.annotations.Test; import org.testng.annotations.Test;
import java.util.Map; import java.util.Map;
import java.util.Set;
import static org.elasticsearch.client.Requests.*; import static org.elasticsearch.client.Requests.*;
import static org.elasticsearch.common.collect.Lists.*; import static org.elasticsearch.common.collect.Lists.*;
@ -58,25 +58,28 @@ import static org.hamcrest.Matchers.*;
/** /**
* @author kimchy (shay.banon) * @author kimchy (shay.banon)
*/ */
public class SingleInstanceEmbeddedSearchTests extends AbstractNodesTests { public class SingleShardEmbeddedSearchTests extends AbstractNodesTests {
private SearchService searchService; private SearchService searchService;
private SearchPhaseController searchPhaseController; private SearchPhaseController searchPhaseController;
private Client client;
@BeforeClass public void createNodeAndInitWithData() throws Exception { @BeforeClass public void createNodeAndInitWithData() throws Exception {
startNode("server1"); startNode("server1");
client = client("server1");
client("server1").admin().indices().prepareCreate("test") client.admin().indices().prepareCreate("test")
.setSettings(settingsBuilder().put("number_of_shards", 1).put("number_of_replicas", 0)) .setSettings(settingsBuilder().put("number_of_shards", 1).put("number_of_replicas", 0))
.execute().actionGet(); .execute().actionGet();
client("server1").admin().cluster().prepareHealth().setWaitForGreenStatus().execute().actionGet(); client("server1").admin().cluster().prepareHealth().setWaitForGreenStatus().execute().actionGet();
index(client("server1"), "1", "test1", 1); index("1", "test1", 1);
index(client("server1"), "2", "test2", 2); index("2", "test2", 2);
index(client("server1"), "3", "test3", 2); index("3", "test3", 3);
index(client("server1"), "4", "test4", 2); index("4", "test4", 4);
index(client("server1"), "5", "test5", 2); index("5", "test5", 5);
client("server1").admin().indices().refresh(refreshRequest("test")).actionGet(); client.admin().indices().refresh(refreshRequest("test")).actionGet();
searchService = ((InternalNode) node("server1")).injector().getInstance(SearchService.class); searchService = ((InternalNode) node("server1")).injector().getInstance(SearchService.class);
searchPhaseController = ((InternalNode) node("server1")).injector().getInstance(SearchPhaseController.class); searchPhaseController = ((InternalNode) node("server1")).injector().getInstance(SearchPhaseController.class);
@ -125,6 +128,58 @@ public class SingleInstanceEmbeddedSearchTests extends AbstractNodesTests {
assertThat(fetchResult.hits().hits()[0].type(), equalTo("type1")); assertThat(fetchResult.hits().hits()[0].type(), equalTo("type1"));
} }
@Test public void testQueryThenFetchIterateWithFrom() throws Exception {
QuerySearchResult queryResult = searchService.executeQueryPhase(searchRequest(searchSource().query(matchAllQuery()).from(0).size(2).sort("age", SortOrder.DESC)));
assertThat(queryResult.topDocs().totalHits, equalTo(5));
Set<String> idsLoaded = Sets.newHashSet();
ShardDoc[] sortedShardList = searchPhaseController.sortDocs(newArrayList(queryResult));
Map<SearchShardTarget, ExtTIntArrayList> docIdsToLoad = searchPhaseController.docIdsToLoad(sortedShardList);
assertThat(docIdsToLoad.size(), equalTo(1));
assertThat(docIdsToLoad.values().iterator().next().size(), equalTo(2));
FetchSearchResult fetchResult = searchService.executeFetchPhase(new FetchSearchRequest(queryResult.id(), docIdsToLoad.values().iterator().next()));
assertThat(fetchResult.hits().hits().length, equalTo(2));
for (SearchHit hit : fetchResult.hits()) {
idsLoaded.add(hit.id());
}
// iterate to the next 2
queryResult = searchService.executeQueryPhase(searchRequest(searchSource().query(matchAllQuery()).from(2).size(2).sort("age", SortOrder.DESC)));
assertThat(queryResult.topDocs().totalHits, equalTo(5));
sortedShardList = searchPhaseController.sortDocs(newArrayList(queryResult));
docIdsToLoad = searchPhaseController.docIdsToLoad(sortedShardList);
assertThat(docIdsToLoad.size(), equalTo(1));
assertThat(docIdsToLoad.values().iterator().next().size(), equalTo(2));
fetchResult = searchService.executeFetchPhase(new FetchSearchRequest(queryResult.id(), docIdsToLoad.values().iterator().next()));
assertThat(fetchResult.hits().hits().length, equalTo(2));
for (SearchHit hit : fetchResult.hits()) {
idsLoaded.add(hit.id());
}
// iterate to the next 2
queryResult = searchService.executeQueryPhase(searchRequest(searchSource().query(matchAllQuery()).from(4).size(2).sort("age", SortOrder.DESC)));
assertThat(queryResult.topDocs().totalHits, equalTo(5));
sortedShardList = searchPhaseController.sortDocs(newArrayList(queryResult));
docIdsToLoad = searchPhaseController.docIdsToLoad(sortedShardList);
assertThat(docIdsToLoad.size(), equalTo(1));
assertThat(docIdsToLoad.values().iterator().next().size(), equalTo(1));
fetchResult = searchService.executeFetchPhase(new FetchSearchRequest(queryResult.id(), docIdsToLoad.values().iterator().next()));
assertThat(fetchResult.hits().hits().length, equalTo(1));
for (SearchHit hit : fetchResult.hits()) {
idsLoaded.add(hit.id());
}
// verify all ids were loaded
Set<String> expectedIds = Sets.newHashSet("1", "2", "3", "4", "5");
assertThat(idsLoaded, equalTo(expectedIds));
}
@Test public void testQueryAndFetch() throws Exception { @Test public void testQueryAndFetch() throws Exception {
QueryFetchSearchResult result = searchService.executeFetchPhase(searchRequest(searchSource().query(termQuery("name", "test1")))); QueryFetchSearchResult result = searchService.executeFetchPhase(searchRequest(searchSource().query(termQuery("name", "test1"))));
FetchSearchResult fetchResult = result.fetchResult(); FetchSearchResult fetchResult = result.fetchResult();
@ -157,7 +212,7 @@ public class SingleInstanceEmbeddedSearchTests extends AbstractNodesTests {
.facet(FacetBuilders.queryFacet("age1", termQuery("age", 1))) .facet(FacetBuilders.queryFacet("age1", termQuery("age", 1)))
.facet(FacetBuilders.queryFacet("age2", termQuery("age", 2))) .facet(FacetBuilders.queryFacet("age2", termQuery("age", 2)))
)); ));
assertThat(queryResult.facets().facet(QueryFacet.class, "age2").count(), equalTo(4l)); assertThat(queryResult.facets().facet(QueryFacet.class, "age2").count(), equalTo(1l));
assertThat(queryResult.facets().facet(QueryFacet.class, "age1").count(), equalTo(1l)); assertThat(queryResult.facets().facet(QueryFacet.class, "age1").count(), equalTo(1l));
} }
@ -186,7 +241,7 @@ public class SingleInstanceEmbeddedSearchTests extends AbstractNodesTests {
return new InternalSearchRequest("test", 0, 1).source(builder.buildAsBytes()); return new InternalSearchRequest("test", 0, 1).source(builder.buildAsBytes());
} }
private void index(Client client, String id, String nameValue, int age) { private void index(String id, String nameValue, int age) {
client.index(indexRequest("test").type("type1").id(id).source(source(id, nameValue, age))).actionGet(); client.index(indexRequest("test").type("type1").id(id).source(source(id, nameValue, age))).actionGet();
} }

View File

@ -70,7 +70,7 @@ import static org.hamcrest.Matchers.*;
/** /**
* @author kimchy (shay.banon) * @author kimchy (shay.banon)
*/ */
public class TwoInstanceEmbeddedSearchTests extends AbstractNodesTests { public class ThreeShardsEmbeddedSearchTests extends AbstractNodesTests {
private ClusterService clusterService; private ClusterService clusterService;

View File

@ -74,7 +74,7 @@ import static org.hamcrest.Matchers.*;
/** /**
* @author kimchy (shay.banon) * @author kimchy (shay.banon)
*/ */
public class TwoInstanceUnbalancedShardsEmbeddedSearchTests extends AbstractNodesTests { public class ThreeShardsUnbalancedShardsEmbeddedSearchTests extends AbstractNodesTests {
private ClusterService clusterService; private ClusterService clusterService;
@ -84,7 +84,7 @@ public class TwoInstanceUnbalancedShardsEmbeddedSearchTests extends AbstractNode
@BeforeClass public void createNodeAndInitWithData() throws Exception { @BeforeClass public void createNodeAndInitWithData() throws Exception {
ImmutableSettings.Builder nodeSettings = ImmutableSettings.settingsBuilder() ImmutableSettings.Builder nodeSettings = ImmutableSettings.settingsBuilder()
.put("cluster.routing.operation.type", "org.elasticsearch.test.integration.search.embedded.TwoInstanceUnbalancedShardsEmbeddedSearchTests$UnevenOperationRoutingModule"); .put("cluster.routing.operation.type", "org.elasticsearch.test.integration.search.embedded.ThreeShardsUnbalancedShardsEmbeddedSearchTests$UnevenOperationRoutingModule");
startNode("server1", nodeSettings); startNode("server1", nodeSettings);
startNode("server2", nodeSettings); startNode("server2", nodeSettings);