tests: make TestHarness thread safe and add multi-threaded random add/commit/get test

git-svn-id: https://svn.apache.org/repos/asf/lucene/dev/trunk@1149437 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Yonik Seeley 2011-07-22 02:47:29 +00:00
parent cf1120f744
commit dcaca50a86
2 changed files with 262 additions and 7 deletions

View File

@ -0,0 +1,230 @@
/**
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.solr.search;
import org.apache.lucene.index.IndexReader;
import org.apache.lucene.index.IndexReader.AtomicReaderContext;
import org.apache.lucene.index.IndexReader.ReaderContext;
import org.apache.lucene.queries.function.DocValues;
import org.apache.lucene.queries.function.ValueSource;
import org.apache.lucene.util.ReaderUtil;
import org.apache.noggit.ObjectBuilder;
import org.apache.solr.SolrTestCaseJ4;
import org.apache.solr.client.solrj.SolrRequest;
import org.apache.solr.common.SolrException;
import org.apache.solr.request.SolrQueryRequest;
import org.apache.solr.schema.SchemaField;
import org.junit.BeforeClass;
import org.junit.Test;
import java.io.IOException;
import java.util.*;
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.atomic.AtomicInteger;
import java.util.concurrent.atomic.AtomicLong;
public class TestRealTimeGet extends SolrTestCaseJ4 {
@BeforeClass
public static void beforeClass() throws Exception {
initCore("solrconfig.xml","schema12.xml");
}
/***
@Test
public void testGetRealtime() throws Exception {
SolrQueryRequest sr1 = req("q","foo");
IndexReader r1 = sr1.getCore().getRealtimeReader();
assertU(adoc("id","1"));
IndexReader r2 = sr1.getCore().getRealtimeReader();
assertNotSame(r1, r2);
int refcount = r2.getRefCount();
// make sure a new reader wasn't opened
IndexReader r3 = sr1.getCore().getRealtimeReader();
assertSame(r2, r3);
assertEquals(refcount+1, r3.getRefCount());
assertU(commit());
// this is not critical, but currently a commit does not refresh the reader
// if nothing has changed
IndexReader r4 = sr1.getCore().getRealtimeReader();
assertEquals(refcount+2, r4.getRefCount());
r1.decRef();
r2.decRef();
r3.decRef();
r4.decRef();
sr1.close();
}
***/
private ConcurrentHashMap<Integer,Long> model = new ConcurrentHashMap<Integer,Long>();
private volatile Map<Integer,Long> committedModel = new HashMap<Integer,Long>();
volatile int lastId;
private final String field = "val_l";
private volatile Throwable ex;
@Test
public void testStressGetRealtime() throws Exception {
// update variables
final int commitPercent = 10;
final int softCommitPercent = 50; // what percent of the commits are soft
final int deletePercent = 8;
final int deleteByQueryPercent = 4;
final int ndocs = 100;
int nWriteThreads = 10;
final int maxConcurrentCommits = 2; // number of committers at a time... needed if we want to avoid commit errors due to exceeding the max
// query variables
final int percentRealtimeQuery = 0; // realtime get is not implemented yet
final AtomicLong operations = new AtomicLong(5000); // number of query operations to perform in total
int nReadThreads = 10;
for (int i=0; i<ndocs; i++) {
model.put(i, -1L);
}
committedModel.putAll(model);
final AtomicInteger numCommitting = new AtomicInteger();
List<Thread> threads = new ArrayList<Thread>();
for (int i=0; i<nWriteThreads; i++) {
Thread thread = new Thread("WRITER"+i) {
Random rand = new Random(random.nextInt());
@Override
public void run() {
while (operations.get() > 0) {
int oper = rand.nextInt(100);
int id = rand.nextInt(ndocs);
Long val = model.get(id);
long nextVal = Math.abs(val)+1;
// set the lastId before we actually change it sometimes to try and
// uncover more race conditions between writing and reading
boolean before = random.nextBoolean();
if (before) {
lastId = id;
}
if (oper < commitPercent) {
if (numCommitting.incrementAndGet() <= maxConcurrentCommits) {
if (rand.nextInt(100) < softCommitPercent)
assertU(h.commit("softCommit","true"));
else
assertU(commit());
}
committedModel = new HashMap<Integer,Long>(model); // take a snapshot
numCommitting.decrementAndGet();
} else if (oper < commitPercent + deletePercent) {
assertU("<delete><id>" + id + "</id></delete>");
model.put(id, -nextVal);
} else if (oper < commitPercent + deletePercent + deleteByQueryPercent) {
assertU("<delete><query>id:" + id + "</query></delete>");
model.put(id, -nextVal);
} else {
assertU(adoc("id",Integer.toString(id), field, Long.toString(nextVal)));
}
if (!before) {
lastId = id;
}
}
}
};
threads.add(thread);
}
for (int i=0; i<nReadThreads; i++) {
Thread thread = new Thread("READER"+i) {
Random rand = new Random(random.nextInt());
@Override
public void run() {
try {
while (operations.decrementAndGet() >= 0) {
int oper = rand.nextInt(100);
// bias toward a recently changed doc
int id = rand.nextInt(100) < 25 ? lastId : rand.nextInt(ndocs);
// when indexing, we update the index, then the model
// so when querying, we should first check the model, and then the index
boolean realTime = rand.nextInt(100) < percentRealtimeQuery;
long val;
if (realTime) {
val = model.get(id);
} else {
val = committedModel.get(id);
}
SolrQueryRequest sreq;
if (realTime) {
sreq = req("wt","json", "qt","/get", "ids",Integer.toString(id));
} else {
sreq = req("wt","json", "q","id:"+Integer.toString(id), "omitHeader","true");
}
String response = h.query(sreq);
Map rsp = (Map)ObjectBuilder.fromJSON(response);
List doclist = (List)(((Map)rsp.get("response")).get("docs"));
if (doclist.size() == 0) {
// there's no info we can get back with a delete, so not much we can check without further synchronization
} else {
assertEquals(1, doclist.size());
long foundVal = (Long)(((Map)doclist.get(0)).get(field));
assertTrue(foundVal >= Math.abs(val));
}
}
}
catch (Throwable e) {
ex = e;
operations.set(-1L);
SolrException.log(log,e);
}
}
};
threads.add(thread);
}
for (Thread thread : threads) {
thread.start();
}
for (Thread thread : threads) {
thread.join();
}
assertNull(ex);
}
}

View File

@ -71,8 +71,8 @@ import java.util.Map;
public class TestHarness { public class TestHarness {
protected CoreContainer container; protected CoreContainer container;
private SolrCore core; private SolrCore core;
private XPath xpath = XPathFactory.newInstance().newXPath(); private final ThreadLocal<DocumentBuilder> builderTL = new ThreadLocal<DocumentBuilder>();
private DocumentBuilder builder; private final ThreadLocal<XPath> xpathTL = new ThreadLocal<XPath>();
public XmlUpdateRequestHandler updater; public XmlUpdateRequestHandler updater;
public static SolrConfig createConfig(String confFile) { public static SolrConfig createConfig(String confFile) {
@ -145,15 +145,40 @@ public class TestHarness {
core = container.getCore(coreName); core = container.getCore(coreName);
if (core != null) if (core != null)
core.close(); core.close();
builder = DocumentBuilderFactory.newInstance().newDocumentBuilder();
updater = new XmlUpdateRequestHandler(); updater = new XmlUpdateRequestHandler();
updater.init( null ); updater.init( null );
} catch (Exception e) { } catch (Exception e) {
throw new RuntimeException(e); throw new RuntimeException(e);
} }
} }
private DocumentBuilder getXmlDocumentBuilder() {
try {
DocumentBuilder builder = builderTL.get();
if (builder == null) {
builder = DocumentBuilderFactory.newInstance().newDocumentBuilder();
builderTL.set(builder);
}
return builder;
} catch (Exception e) {
throw new RuntimeException(e);
}
}
private XPath getXpath() {
try {
XPath xpath = xpathTL.get();
if (xpath == null) {
xpath = XPathFactory.newInstance().newXPath();
xpathTL.set(xpath);
}
return xpath;
} catch (Exception e) {
throw new RuntimeException(e);
}
}
// Creates a container based on infos needed to create one core // Creates a container based on infos needed to create one core
static class Initializer extends CoreContainer.Initializer { static class Initializer extends CoreContainer.Initializer {
String coreName; String coreName;
@ -364,7 +389,7 @@ public class TestHarness {
Document document=null; Document document=null;
try { try {
document = builder.parse(new ByteArrayInputStream document = getXmlDocumentBuilder().parse(new ByteArrayInputStream
(xml.getBytes("UTF-8"))); (xml.getBytes("UTF-8")));
} catch (UnsupportedEncodingException e1) { } catch (UnsupportedEncodingException e1) {
throw new RuntimeException("Totally weird UTF-8 exception", e1); throw new RuntimeException("Totally weird UTF-8 exception", e1);
@ -374,7 +399,7 @@ public class TestHarness {
for (String xp : tests) { for (String xp : tests) {
xp=xp.trim(); xp=xp.trim();
Boolean bool = (Boolean) xpath.evaluate(xp, document, Boolean bool = (Boolean) getXpath().evaluate(xp, document,
XPathConstants.BOOLEAN); XPathConstants.BOOLEAN);
if (!bool) { if (!bool) {