mirror of https://github.com/apache/activemq.git
[AMQ-6606] - reset next batch so it does not get reused after error and refine fix to sync write batches b/c async locations will already be in the index
(cherry picked from commit 21ae1ef2e6
)
This commit is contained in:
parent
b8fc78ec6c
commit
7e3f344ea7
|
@ -355,7 +355,11 @@ class DataFileAppender implements FileAppender {
|
||||||
synchronized (enqueueMutex) {
|
synchronized (enqueueMutex) {
|
||||||
running = false;
|
running = false;
|
||||||
signalError(wb, error);
|
signalError(wb, error);
|
||||||
|
if (nextWriteBatch != null) {
|
||||||
signalError(nextWriteBatch, error);
|
signalError(nextWriteBatch, error);
|
||||||
|
nextWriteBatch = null;
|
||||||
|
enqueueMutex.notifyAll();
|
||||||
|
}
|
||||||
}
|
}
|
||||||
} finally {
|
} finally {
|
||||||
try {
|
try {
|
||||||
|
@ -402,12 +406,23 @@ class DataFileAppender implements FileAppender {
|
||||||
if (wb != null) {
|
if (wb != null) {
|
||||||
if (t instanceof IOException) {
|
if (t instanceof IOException) {
|
||||||
wb.exception.set((IOException) t);
|
wb.exception.set((IOException) t);
|
||||||
// revert batch increment such that next write is contiguous
|
// revert sync batch increment such that next write is contiguous
|
||||||
|
if (syncBatch(wb.writes)) {
|
||||||
wb.dataFile.decrementLength(wb.size);
|
wb.dataFile.decrementLength(wb.size);
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
wb.exception.set(IOExceptionSupport.create(t));
|
wb.exception.set(IOExceptionSupport.create(t));
|
||||||
}
|
}
|
||||||
signalDone(wb);
|
signalDone(wb);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// async writes will already be in the index so reuse is not an option
|
||||||
|
private boolean syncBatch(LinkedNodeList<Journal.WriteCommand> writes) {
|
||||||
|
Journal.WriteCommand write = writes.getHead();
|
||||||
|
while (write != null && write.sync) {
|
||||||
|
write = write.getNext();
|
||||||
|
}
|
||||||
|
return write == null;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -21,17 +21,27 @@ import org.apache.activemq.util.RecoverableRandomAccessFile;
|
||||||
import org.junit.Rule;
|
import org.junit.Rule;
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.junit.rules.TemporaryFolder;
|
import org.junit.rules.TemporaryFolder;
|
||||||
|
import org.slf4j.Logger;
|
||||||
|
import org.slf4j.LoggerFactory;
|
||||||
|
|
||||||
import java.io.File;
|
import java.io.File;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.ArrayList;
|
import java.util.ArrayList;
|
||||||
import java.util.Collections;
|
import java.util.Collections;
|
||||||
|
import java.util.HashSet;
|
||||||
import java.util.List;
|
import java.util.List;
|
||||||
|
import java.util.concurrent.ConcurrentLinkedQueue;
|
||||||
|
import java.util.concurrent.CountDownLatch;
|
||||||
|
import java.util.concurrent.TimeUnit;
|
||||||
|
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
|
import static org.junit.Assert.assertTrue;
|
||||||
import static org.junit.Assert.fail;
|
import static org.junit.Assert.fail;
|
||||||
|
|
||||||
public class DataFileAppenderNoSpaceNoBatchTest {
|
public class DataFileAppenderNoSpaceNoBatchTest {
|
||||||
|
|
||||||
|
private static final Logger LOG = LoggerFactory.getLogger(DataFileAppenderNoSpaceNoBatchTest.class);
|
||||||
|
|
||||||
@Rule
|
@Rule
|
||||||
public TemporaryFolder dataFileDir = new TemporaryFolder(new File("target"));
|
public TemporaryFolder dataFileDir = new TemporaryFolder(new File("target"));
|
||||||
|
|
||||||
|
@ -77,4 +87,62 @@ public class DataFileAppenderNoSpaceNoBatchTest {
|
||||||
assertEquals("offset is reused", seekPositions.get(0), seekPositions.get(1));
|
assertEquals("offset is reused", seekPositions.get(0), seekPositions.get(1));
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@Test(timeout = 10000)
|
||||||
|
public void testNoSpaceNextWriteSameBatchAsync() throws Exception {
|
||||||
|
final List<Long> seekPositions = Collections.synchronizedList(new ArrayList<Long>());
|
||||||
|
|
||||||
|
final DataFile currentDataFile = new DataFile(dataFileDir.newFile(), 0) {
|
||||||
|
public RecoverableRandomAccessFile appendRandomAccessFile() throws IOException {
|
||||||
|
|
||||||
|
return new RecoverableRandomAccessFile(dataFileDir.newFile(), "rw") {
|
||||||
|
|
||||||
|
public void seek(long pos) throws IOException {
|
||||||
|
seekPositions.add(pos);
|
||||||
|
}
|
||||||
|
|
||||||
|
public void write(byte[] bytes, int offset, int len) throws IOException {
|
||||||
|
if (seekPositions.size() == 2) {
|
||||||
|
throw new IOException("No space on device: " + seekPositions.size());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
underTest = new DataFileAppender(new Journal() {
|
||||||
|
@Override
|
||||||
|
public DataFile getCurrentDataFile(int capacity) throws IOException {
|
||||||
|
return currentDataFile;
|
||||||
|
};
|
||||||
|
|
||||||
|
@Override
|
||||||
|
public int getWriteBatchSize() {
|
||||||
|
// force multiple async batches
|
||||||
|
return 4*1024;
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
final ByteSequence byteSequence = new ByteSequence(new byte[1024]);
|
||||||
|
|
||||||
|
ConcurrentLinkedQueue<Location> locations = new ConcurrentLinkedQueue<Location>();
|
||||||
|
HashSet<CountDownLatch> latches = new HashSet<CountDownLatch>();
|
||||||
|
for (int i = 0; i <= 20; i++) {
|
||||||
|
Location location = underTest.storeItem(byteSequence, (byte) 1, false);
|
||||||
|
locations.add(location);
|
||||||
|
latches.add(location.getLatch());
|
||||||
|
}
|
||||||
|
|
||||||
|
for (CountDownLatch latch: latches) {
|
||||||
|
assertTrue("write complete", latch.await(5, TimeUnit.SECONDS));
|
||||||
|
}
|
||||||
|
|
||||||
|
LOG.info("Latches count: " + latches.size());
|
||||||
|
LOG.info("Seeks: " + seekPositions);
|
||||||
|
|
||||||
|
assertTrue("got more than on latch: " + latches.size(), latches.size() > 1);
|
||||||
|
assertTrue("got seeks: " + seekPositions, seekPositions.size() > 2);
|
||||||
|
assertEquals("no duplicates: " + seekPositions, seekPositions.size(), new HashSet<Long>(seekPositions).size());
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue