mirror of https://github.com/apache/activemq.git
[AMQ-6606] - reset next batch so it does not get reused after error and refine fix to sync write batches b/c async locations will already be in the index
This commit is contained in:
parent
0d824a8e68
commit
21ae1ef2e6
|
@ -355,7 +355,11 @@ class DataFileAppender implements FileAppender {
|
|||
synchronized (enqueueMutex) {
|
||||
running = false;
|
||||
signalError(wb, error);
|
||||
signalError(nextWriteBatch, error);
|
||||
if (nextWriteBatch != null) {
|
||||
signalError(nextWriteBatch, error);
|
||||
nextWriteBatch = null;
|
||||
enqueueMutex.notifyAll();
|
||||
}
|
||||
}
|
||||
} finally {
|
||||
try {
|
||||
|
@ -402,12 +406,23 @@ class DataFileAppender implements FileAppender {
|
|||
if (wb != null) {
|
||||
if (t instanceof IOException) {
|
||||
wb.exception.set((IOException) t);
|
||||
// revert batch increment such that next write is contiguous
|
||||
wb.dataFile.decrementLength(wb.size);
|
||||
// revert sync batch increment such that next write is contiguous
|
||||
if (syncBatch(wb.writes)) {
|
||||
wb.dataFile.decrementLength(wb.size);
|
||||
}
|
||||
} else {
|
||||
wb.exception.set(IOExceptionSupport.create(t));
|
||||
}
|
||||
signalDone(wb);
|
||||
}
|
||||
}
|
||||
|
||||
// async writes will already be in the index so reuse is not an option
|
||||
private boolean syncBatch(LinkedNodeList<Journal.WriteCommand> writes) {
|
||||
Journal.WriteCommand write = writes.getHead();
|
||||
while (write != null && write.sync) {
|
||||
write = write.getNext();
|
||||
}
|
||||
return write == null;
|
||||
}
|
||||
}
|
||||
|
|
|
@ -21,17 +21,27 @@ import org.apache.activemq.util.RecoverableRandomAccessFile;
|
|||
import org.junit.Rule;
|
||||
import org.junit.Test;
|
||||
import org.junit.rules.TemporaryFolder;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.util.ArrayList;
|
||||
import java.util.Collections;
|
||||
import java.util.HashSet;
|
||||
import java.util.List;
|
||||
import java.util.concurrent.ConcurrentLinkedQueue;
|
||||
import java.util.concurrent.CountDownLatch;
|
||||
import java.util.concurrent.TimeUnit;
|
||||
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertTrue;
|
||||
import static org.junit.Assert.fail;
|
||||
|
||||
public class DataFileAppenderNoSpaceNoBatchTest {
|
||||
|
||||
private static final Logger LOG = LoggerFactory.getLogger(DataFileAppenderNoSpaceNoBatchTest.class);
|
||||
|
||||
@Rule
|
||||
public TemporaryFolder dataFileDir = new TemporaryFolder(new File("target"));
|
||||
|
||||
|
@ -77,4 +87,62 @@ public class DataFileAppenderNoSpaceNoBatchTest {
|
|||
assertEquals("offset is reused", seekPositions.get(0), seekPositions.get(1));
|
||||
|
||||
}
|
||||
|
||||
|
||||
@Test(timeout = 10000)
|
||||
public void testNoSpaceNextWriteSameBatchAsync() throws Exception {
|
||||
final List<Long> seekPositions = Collections.synchronizedList(new ArrayList<Long>());
|
||||
|
||||
final DataFile currentDataFile = new DataFile(dataFileDir.newFile(), 0) {
|
||||
public RecoverableRandomAccessFile appendRandomAccessFile() throws IOException {
|
||||
|
||||
return new RecoverableRandomAccessFile(dataFileDir.newFile(), "rw") {
|
||||
|
||||
public void seek(long pos) throws IOException {
|
||||
seekPositions.add(pos);
|
||||
}
|
||||
|
||||
public void write(byte[] bytes, int offset, int len) throws IOException {
|
||||
if (seekPositions.size() == 2) {
|
||||
throw new IOException("No space on device: " + seekPositions.size());
|
||||
}
|
||||
}
|
||||
};
|
||||
};
|
||||
};
|
||||
|
||||
underTest = new DataFileAppender(new Journal() {
|
||||
@Override
|
||||
public DataFile getCurrentDataFile(int capacity) throws IOException {
|
||||
return currentDataFile;
|
||||
};
|
||||
|
||||
@Override
|
||||
public int getWriteBatchSize() {
|
||||
// force multiple async batches
|
||||
return 4*1024;
|
||||
}
|
||||
});
|
||||
|
||||
final ByteSequence byteSequence = new ByteSequence(new byte[1024]);
|
||||
|
||||
ConcurrentLinkedQueue<Location> locations = new ConcurrentLinkedQueue<Location>();
|
||||
HashSet<CountDownLatch> latches = new HashSet<CountDownLatch>();
|
||||
for (int i = 0; i <= 20; i++) {
|
||||
Location location = underTest.storeItem(byteSequence, (byte) 1, false);
|
||||
locations.add(location);
|
||||
latches.add(location.getLatch());
|
||||
}
|
||||
|
||||
for (CountDownLatch latch: latches) {
|
||||
assertTrue("write complete", latch.await(5, TimeUnit.SECONDS));
|
||||
}
|
||||
|
||||
LOG.info("Latches count: " + latches.size());
|
||||
LOG.info("Seeks: " + seekPositions);
|
||||
|
||||
assertTrue("got more than on latch: " + latches.size(), latches.size() > 1);
|
||||
assertTrue("got seeks: " + seekPositions, seekPositions.size() > 2);
|
||||
assertEquals("no duplicates: " + seekPositions, seekPositions.size(), new HashSet<Long>(seekPositions).size());
|
||||
}
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue