YARN-1984. LeveldbTimelineStore does not handle db exceptions properly. Contributed by Varun Saxena

(cherry picked from commit 1ce4d33c2d)

(cherry picked from commit 89ef49fb08)
This commit is contained in:
Jason Lowe 2014-11-24 22:36:59 +00:00 committed by Vinod Kumar Vavilapalli
parent c2e00e0012
commit 17a23b8191
3 changed files with 126 additions and 92 deletions

View File

@ -48,6 +48,9 @@ Release 2.6.1 - UNRELEASED
YARN-2964. RM prematurely cancels tokens for jobs that submit jobs (oozie) YARN-2964. RM prematurely cancels tokens for jobs that submit jobs (oozie)
(Jian He via jlowe) (Jian He via jlowe)
YARN-1984. LeveldbTimelineStore does not handle db exceptions properly
(Varun Saxena via jlowe)
Release 2.6.0 - 2014-11-18 Release 2.6.0 - 2014-11-18
INCOMPATIBLE CHANGES INCOMPATIBLE CHANGES

View File

@ -66,10 +66,10 @@ import org.apache.hadoop.yarn.conf.YarnConfiguration;
import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.VersionProto; import org.apache.hadoop.yarn.proto.YarnServerCommonProtos.VersionProto;
import org.apache.hadoop.yarn.server.records.Version; import org.apache.hadoop.yarn.server.records.Version;
import org.apache.hadoop.yarn.server.records.impl.pb.VersionPBImpl; import org.apache.hadoop.yarn.server.records.impl.pb.VersionPBImpl;
import org.apache.hadoop.yarn.server.utils.LeveldbIterator;
import org.fusesource.leveldbjni.JniDBFactory; import org.fusesource.leveldbjni.JniDBFactory;
import org.iq80.leveldb.DB; import org.iq80.leveldb.DB;
import org.iq80.leveldb.DBException; import org.iq80.leveldb.DBException;
import org.iq80.leveldb.DBIterator;
import org.iq80.leveldb.Options; import org.iq80.leveldb.Options;
import org.iq80.leveldb.ReadOptions; import org.iq80.leveldb.ReadOptions;
import org.iq80.leveldb.WriteBatch; import org.iq80.leveldb.WriteBatch;
@ -438,13 +438,15 @@ public class LeveldbTimelineStore extends AbstractService
.add(entityType).add(writeReverseOrderedLong(revStartTime)) .add(entityType).add(writeReverseOrderedLong(revStartTime))
.add(entityId).getBytesForLookup(); .add(entityId).getBytesForLookup();
DBIterator iterator = null; LeveldbIterator iterator = null;
try { try {
iterator = db.iterator(); iterator = new LeveldbIterator(db);
iterator.seek(prefix); iterator.seek(prefix);
return getEntity(entityId, entityType, revStartTime, fields, iterator, return getEntity(entityId, entityType, revStartTime, fields, iterator,
prefix, prefix.length); prefix, prefix.length);
} catch(DBException e) {
throw new IOException(e);
} finally { } finally {
IOUtils.cleanup(LOG, iterator); IOUtils.cleanup(LOG, iterator);
} }
@ -455,7 +457,7 @@ public class LeveldbTimelineStore extends AbstractService
* specified fields for this entity, return null. * specified fields for this entity, return null.
*/ */
private static TimelineEntity getEntity(String entityId, String entityType, private static TimelineEntity getEntity(String entityId, String entityType,
Long startTime, EnumSet<Field> fields, DBIterator iterator, Long startTime, EnumSet<Field> fields, LeveldbIterator iterator,
byte[] prefix, int prefixlen) throws IOException { byte[] prefix, int prefixlen) throws IOException {
if (fields == null) { if (fields == null) {
fields = EnumSet.allOf(Field.class); fields = EnumSet.allOf(Field.class);
@ -562,7 +564,7 @@ public class LeveldbTimelineStore extends AbstractService
o2.length); o2.length);
} }
}); });
DBIterator iterator = null; LeveldbIterator iterator = null;
try { try {
// look up start times for the specified entities // look up start times for the specified entities
// skip entities with no start time // skip entities with no start time
@ -606,7 +608,7 @@ public class LeveldbTimelineStore extends AbstractService
if (limit == null) { if (limit == null) {
limit = DEFAULT_LIMIT; limit = DEFAULT_LIMIT;
} }
iterator = db.iterator(); iterator = new LeveldbIterator(db);
for (iterator.seek(first); entity.getEvents().size() < limit && for (iterator.seek(first); entity.getEvents().size() < limit &&
iterator.hasNext(); iterator.next()) { iterator.hasNext(); iterator.next()) {
byte[] key = iterator.peekNext().getKey(); byte[] key = iterator.peekNext().getKey();
@ -623,6 +625,8 @@ public class LeveldbTimelineStore extends AbstractService
} }
} }
} }
} catch(DBException e) {
throw new IOException(e);
} finally { } finally {
IOUtils.cleanup(LOG, iterator); IOUtils.cleanup(LOG, iterator);
} }
@ -683,7 +687,7 @@ public class LeveldbTimelineStore extends AbstractService
String entityType, Long limit, Long starttime, Long endtime, String entityType, Long limit, Long starttime, Long endtime,
String fromId, Long fromTs, Collection<NameValuePair> secondaryFilters, String fromId, Long fromTs, Collection<NameValuePair> secondaryFilters,
EnumSet<Field> fields) throws IOException { EnumSet<Field> fields) throws IOException {
DBIterator iterator = null; LeveldbIterator iterator = null;
try { try {
KeyBuilder kb = KeyBuilder.newInstance().add(base).add(entityType); KeyBuilder kb = KeyBuilder.newInstance().add(base).add(entityType);
// only db keys matching the prefix (base + entity type) will be parsed // only db keys matching the prefix (base + entity type) will be parsed
@ -724,7 +728,7 @@ public class LeveldbTimelineStore extends AbstractService
} }
TimelineEntities entities = new TimelineEntities(); TimelineEntities entities = new TimelineEntities();
iterator = db.iterator(); iterator = new LeveldbIterator(db);
iterator.seek(first); iterator.seek(first);
// iterate until one of the following conditions is met: limit is // iterate until one of the following conditions is met: limit is
// reached, there are no more keys, the key prefix no longer matches, // reached, there are no more keys, the key prefix no longer matches,
@ -783,11 +787,24 @@ public class LeveldbTimelineStore extends AbstractService
} }
} }
return entities; return entities;
} catch(DBException e) {
throw new IOException(e);
} finally { } finally {
IOUtils.cleanup(LOG, iterator); IOUtils.cleanup(LOG, iterator);
} }
} }
/**
* Handle error and set it in response.
*/
private static void handleError(TimelineEntity entity, TimelinePutResponse response, final int errorCode) {
TimelinePutError error = new TimelinePutError();
error.setEntityId(entity.getEntityId());
error.setEntityType(entity.getEntityType());
error.setErrorCode(errorCode);
response.addError(error);
}
/** /**
* Put a single entity. If there is an error, add a TimelinePutError to the * Put a single entity. If there is an error, add a TimelinePutError to the
* given response. * given response.
@ -812,11 +829,7 @@ public class LeveldbTimelineStore extends AbstractService
entity.getStartTime(), events); entity.getStartTime(), events);
if (startAndInsertTime == null) { if (startAndInsertTime == null) {
// if no start time is found, add an error and return // if no start time is found, add an error and return
TimelinePutError error = new TimelinePutError(); handleError(entity, response, TimelinePutError.NO_START_TIME);
error.setEntityId(entity.getEntityId());
error.setEntityType(entity.getEntityType());
error.setErrorCode(TimelinePutError.NO_START_TIME);
response.addError(error);
return; return;
} }
revStartTime = writeReverseOrderedLong(startAndInsertTime revStartTime = writeReverseOrderedLong(startAndInsertTime
@ -883,11 +896,7 @@ public class LeveldbTimelineStore extends AbstractService
if (!domainId.equals(entity.getDomainId())) { if (!domainId.equals(entity.getDomainId())) {
// in this case the entity will be put, but the relation will be // in this case the entity will be put, but the relation will be
// ignored // ignored
TimelinePutError error = new TimelinePutError(); handleError(entity, response, TimelinePutError.FORBIDDEN_RELATION);
error.setEntityId(entity.getEntityId());
error.setEntityType(entity.getEntityType());
error.setErrorCode(TimelinePutError.FORBIDDEN_RELATION);
response.addError(error);
continue; continue;
} }
} }
@ -933,11 +942,7 @@ public class LeveldbTimelineStore extends AbstractService
if (entity.getDomainId() == null || if (entity.getDomainId() == null ||
entity.getDomainId().length() == 0) { entity.getDomainId().length() == 0) {
if (!allowEmptyDomainId) { if (!allowEmptyDomainId) {
TimelinePutError error = new TimelinePutError(); handleError(entity, response, TimelinePutError.NO_DOMAIN);
error.setEntityId(entity.getEntityId());
error.setEntityType(entity.getEntityType());
error.setErrorCode(TimelinePutError.NO_DOMAIN);
response.addError(error);
return; return;
} }
} else { } else {
@ -946,14 +951,14 @@ public class LeveldbTimelineStore extends AbstractService
entity.getDomainId().getBytes()); entity.getDomainId().getBytes());
} }
db.write(writeBatch); db.write(writeBatch);
} catch (DBException de) {
LOG.error("Error putting entity " + entity.getEntityId() +
" of type " + entity.getEntityType(), de);
handleError(entity, response, TimelinePutError.IO_EXCEPTION);
} catch (IOException e) { } catch (IOException e) {
LOG.error("Error putting entity " + entity.getEntityId() + LOG.error("Error putting entity " + entity.getEntityId() +
" of type " + entity.getEntityType(), e); " of type " + entity.getEntityType(), e);
TimelinePutError error = new TimelinePutError(); handleError(entity, response, TimelinePutError.IO_EXCEPTION);
error.setEntityId(entity.getEntityId());
error.setEntityType(entity.getEntityType());
error.setErrorCode(TimelinePutError.IO_EXCEPTION);
response.addError(error);
} finally { } finally {
lock.unlock(); lock.unlock();
writeLocks.returnLock(lock); writeLocks.returnLock(lock);
@ -983,15 +988,16 @@ public class LeveldbTimelineStore extends AbstractService
relatedEntity.getType(), relatedEntityStartTime), relatedEntity.getType(), relatedEntityStartTime),
writeReverseOrderedLong(relatedEntityStartAndInsertTime writeReverseOrderedLong(relatedEntityStartAndInsertTime
.insertTime)); .insertTime));
} catch (DBException de) {
LOG.error("Error putting related entity " + relatedEntity.getId() +
" of type " + relatedEntity.getType() + " for entity " +
entity.getEntityId() + " of type " + entity.getEntityType(), de);
handleError(entity, response, TimelinePutError.IO_EXCEPTION);
} catch (IOException e) { } catch (IOException e) {
LOG.error("Error putting related entity " + relatedEntity.getId() + LOG.error("Error putting related entity " + relatedEntity.getId() +
" of type " + relatedEntity.getType() + " for entity " + " of type " + relatedEntity.getType() + " for entity " +
entity.getEntityId() + " of type " + entity.getEntityType(), e); entity.getEntityId() + " of type " + entity.getEntityType(), e);
TimelinePutError error = new TimelinePutError(); handleError(entity, response, TimelinePutError.IO_EXCEPTION);
error.setEntityId(entity.getEntityId());
error.setEntityType(entity.getEntityType());
error.setErrorCode(TimelinePutError.IO_EXCEPTION);
response.addError(error);
} finally { } finally {
lock.unlock(); lock.unlock();
writeLocks.returnLock(lock); writeLocks.returnLock(lock);
@ -1072,6 +1078,7 @@ public class LeveldbTimelineStore extends AbstractService
private Long getStartTimeLong(String entityId, String entityType) private Long getStartTimeLong(String entityId, String entityType)
throws IOException { throws IOException {
EntityIdentifier entity = new EntityIdentifier(entityId, entityType); EntityIdentifier entity = new EntityIdentifier(entityId, entityType);
try {
// start time is not provided, so try to look it up // start time is not provided, so try to look it up
if (startTimeReadCache.containsKey(entity)) { if (startTimeReadCache.containsKey(entity)) {
// found the start time in the cache // found the start time in the cache
@ -1090,6 +1097,9 @@ public class LeveldbTimelineStore extends AbstractService
return l; return l;
} }
} }
} catch(DBException e) {
throw new IOException(e);
}
} }
/** /**
@ -1152,6 +1162,7 @@ public class LeveldbTimelineStore extends AbstractService
StartAndInsertTime startAndInsertTime = null; StartAndInsertTime startAndInsertTime = null;
// create lookup key for start time // create lookup key for start time
byte[] b = createStartTimeLookupKey(entity.getId(), entity.getType()); byte[] b = createStartTimeLookupKey(entity.getId(), entity.getType());
try {
// retrieve value for key // retrieve value for key
byte[] v = db.get(b); byte[] v = db.get(b);
if (v == null) { if (v == null) {
@ -1174,6 +1185,9 @@ public class LeveldbTimelineStore extends AbstractService
startAndInsertTime = new StartAndInsertTime(readReverseOrderedLong(v, 0), startAndInsertTime = new StartAndInsertTime(readReverseOrderedLong(v, 0),
readReverseOrderedLong(v, 8)); readReverseOrderedLong(v, 8));
} }
} catch(DBException e) {
throw new IOException(e);
}
startTimeWriteCache.put(entity, startAndInsertTime); startTimeWriteCache.put(entity, startAndInsertTime);
startTimeReadCache.put(entity, startAndInsertTime.startTime); startTimeReadCache.put(entity, startAndInsertTime.startTime);
return startAndInsertTime; return startAndInsertTime;
@ -1373,7 +1387,7 @@ public class LeveldbTimelineStore extends AbstractService
@VisibleForTesting @VisibleForTesting
List<String> getEntityTypes() throws IOException { List<String> getEntityTypes() throws IOException {
DBIterator iterator = null; LeveldbIterator iterator = null;
try { try {
iterator = getDbIterator(false); iterator = getDbIterator(false);
List<String> entityTypes = new ArrayList<String>(); List<String> entityTypes = new ArrayList<String>();
@ -1396,6 +1410,8 @@ public class LeveldbTimelineStore extends AbstractService
iterator.seek(lookupKey); iterator.seek(lookupKey);
} }
return entityTypes; return entityTypes;
} catch(DBException e) {
throw new IOException(e);
} finally { } finally {
IOUtils.cleanup(LOG, iterator); IOUtils.cleanup(LOG, iterator);
} }
@ -1406,7 +1422,7 @@ public class LeveldbTimelineStore extends AbstractService
* the given write batch. * the given write batch.
*/ */
private void deleteKeysWithPrefix(WriteBatch writeBatch, byte[] prefix, private void deleteKeysWithPrefix(WriteBatch writeBatch, byte[] prefix,
DBIterator iterator) { LeveldbIterator iterator) {
for (iterator.seek(prefix); iterator.hasNext(); iterator.next()) { for (iterator.seek(prefix); iterator.hasNext(); iterator.next()) {
byte[] key = iterator.peekNext().getKey(); byte[] key = iterator.peekNext().getKey();
if (!prefixMatches(prefix, prefix.length, key)) { if (!prefixMatches(prefix, prefix.length, key)) {
@ -1418,7 +1434,7 @@ public class LeveldbTimelineStore extends AbstractService
@VisibleForTesting @VisibleForTesting
boolean deleteNextEntity(String entityType, byte[] reverseTimestamp, boolean deleteNextEntity(String entityType, byte[] reverseTimestamp,
DBIterator iterator, DBIterator pfIterator, boolean seeked) LeveldbIterator iterator, LeveldbIterator pfIterator, boolean seeked)
throws IOException { throws IOException {
WriteBatch writeBatch = null; WriteBatch writeBatch = null;
try { try {
@ -1524,6 +1540,8 @@ public class LeveldbTimelineStore extends AbstractService
writeOptions.sync(true); writeOptions.sync(true);
db.write(writeBatch, writeOptions); db.write(writeBatch, writeOptions);
return true; return true;
} catch(DBException e) {
throw new IOException(e);
} finally { } finally {
IOUtils.cleanup(LOG, writeBatch); IOUtils.cleanup(LOG, writeBatch);
} }
@ -1542,8 +1560,8 @@ public class LeveldbTimelineStore extends AbstractService
try { try {
List<String> entityTypes = getEntityTypes(); List<String> entityTypes = getEntityTypes();
for (String entityType : entityTypes) { for (String entityType : entityTypes) {
DBIterator iterator = null; LeveldbIterator iterator = null;
DBIterator pfIterator = null; LeveldbIterator pfIterator = null;
long typeCount = 0; long typeCount = 0;
try { try {
deleteLock.writeLock().lock(); deleteLock.writeLock().lock();
@ -1583,13 +1601,14 @@ public class LeveldbTimelineStore extends AbstractService
} }
@VisibleForTesting @VisibleForTesting
DBIterator getDbIterator(boolean fillCache) { LeveldbIterator getDbIterator(boolean fillCache) {
ReadOptions readOptions = new ReadOptions(); ReadOptions readOptions = new ReadOptions();
readOptions.fillCache(fillCache); readOptions.fillCache(fillCache);
return db.iterator(readOptions); return new LeveldbIterator(db, readOptions);
} }
Version loadVersion() throws IOException { Version loadVersion() throws IOException {
try {
byte[] data = db.get(bytes(TIMELINE_STORE_VERSION_KEY)); byte[] data = db.get(bytes(TIMELINE_STORE_VERSION_KEY));
// if version is not stored previously, treat it as 1.0. // if version is not stored previously, treat it as 1.0.
if (data == null || data.length == 0) { if (data == null || data.length == 0) {
@ -1598,6 +1617,9 @@ public class LeveldbTimelineStore extends AbstractService
Version version = Version version =
new VersionPBImpl(VersionProto.parseFrom(data)); new VersionPBImpl(VersionProto.parseFrom(data));
return version; return version;
} catch(DBException e) {
throw new IOException(e);
}
} }
// Only used for test // Only used for test
@ -1726,6 +1748,8 @@ public class LeveldbTimelineStore extends AbstractService
writeBatch.put(domainEntryKey, timestamps); writeBatch.put(domainEntryKey, timestamps);
writeBatch.put(ownerLookupEntryKey, timestamps); writeBatch.put(ownerLookupEntryKey, timestamps);
db.write(writeBatch); db.write(writeBatch);
} catch(DBException e) {
throw new IOException(e);
} finally { } finally {
IOUtils.cleanup(LOG, writeBatch); IOUtils.cleanup(LOG, writeBatch);
} }
@ -1754,13 +1778,15 @@ public class LeveldbTimelineStore extends AbstractService
@Override @Override
public TimelineDomain getDomain(String domainId) public TimelineDomain getDomain(String domainId)
throws IOException { throws IOException {
DBIterator iterator = null; LeveldbIterator iterator = null;
try { try {
byte[] prefix = KeyBuilder.newInstance() byte[] prefix = KeyBuilder.newInstance()
.add(DOMAIN_ENTRY_PREFIX).add(domainId).getBytesForLookup(); .add(DOMAIN_ENTRY_PREFIX).add(domainId).getBytesForLookup();
iterator = db.iterator(); iterator = new LeveldbIterator(db);
iterator.seek(prefix); iterator.seek(prefix);
return getTimelineDomain(iterator, domainId, prefix); return getTimelineDomain(iterator, domainId, prefix);
} catch(DBException e) {
throw new IOException(e);
} finally { } finally {
IOUtils.cleanup(LOG, iterator); IOUtils.cleanup(LOG, iterator);
} }
@ -1769,12 +1795,12 @@ public class LeveldbTimelineStore extends AbstractService
@Override @Override
public TimelineDomains getDomains(String owner) public TimelineDomains getDomains(String owner)
throws IOException { throws IOException {
DBIterator iterator = null; LeveldbIterator iterator = null;
try { try {
byte[] prefix = KeyBuilder.newInstance() byte[] prefix = KeyBuilder.newInstance()
.add(OWNER_LOOKUP_PREFIX).add(owner).getBytesForLookup(); .add(OWNER_LOOKUP_PREFIX).add(owner).getBytesForLookup();
List<TimelineDomain> domains = new ArrayList<TimelineDomain>(); List<TimelineDomain> domains = new ArrayList<TimelineDomain>();
for (iterator = db.iterator(), iterator.seek(prefix); for (iterator = new LeveldbIterator(db), iterator.seek(prefix);
iterator.hasNext();) { iterator.hasNext();) {
byte[] key = iterator.peekNext().getKey(); byte[] key = iterator.peekNext().getKey();
if (!prefixMatches(prefix, prefix.length, key)) { if (!prefixMatches(prefix, prefix.length, key)) {
@ -1809,13 +1835,15 @@ public class LeveldbTimelineStore extends AbstractService
TimelineDomains domainsToReturn = new TimelineDomains(); TimelineDomains domainsToReturn = new TimelineDomains();
domainsToReturn.addDomains(domains); domainsToReturn.addDomains(domains);
return domainsToReturn; return domainsToReturn;
} catch(DBException e) {
throw new IOException(e);
} finally { } finally {
IOUtils.cleanup(LOG, iterator); IOUtils.cleanup(LOG, iterator);
} }
} }
private static TimelineDomain getTimelineDomain( private static TimelineDomain getTimelineDomain(
DBIterator iterator, String domainId, byte[] prefix) throws IOException { LeveldbIterator iterator, String domainId, byte[] prefix) throws IOException {
// Iterate over all the rows whose key starts with prefix to retrieve the // Iterate over all the rows whose key starts with prefix to retrieve the
// domain information. // domain information.
TimelineDomain domain = new TimelineDomain(); TimelineDomain domain = new TimelineDomain();

View File

@ -45,7 +45,8 @@ import org.apache.hadoop.yarn.conf.YarnConfiguration;
import org.apache.hadoop.yarn.server.records.Version; import org.apache.hadoop.yarn.server.records.Version;
import org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore; import org.apache.hadoop.yarn.server.timeline.LeveldbTimelineStore;
import org.apache.hadoop.yarn.server.timeline.NameValuePair; import org.apache.hadoop.yarn.server.timeline.NameValuePair;
import org.iq80.leveldb.DBIterator; import org.apache.hadoop.yarn.server.utils.LeveldbIterator;
import org.iq80.leveldb.DBException;
import org.junit.After; import org.junit.After;
import org.junit.Assert; import org.junit.Assert;
import org.junit.Before; import org.junit.Before;
@ -146,13 +147,15 @@ public class TestLeveldbTimelineStore extends TimelineStoreTestUtils {
private boolean deleteNextEntity(String entityType, byte[] ts) private boolean deleteNextEntity(String entityType, byte[] ts)
throws IOException, InterruptedException { throws IOException, InterruptedException {
DBIterator iterator = null; LeveldbIterator iterator = null;
DBIterator pfIterator = null; LeveldbIterator pfIterator = null;
try { try {
iterator = ((LeveldbTimelineStore)store).getDbIterator(false); iterator = ((LeveldbTimelineStore)store).getDbIterator(false);
pfIterator = ((LeveldbTimelineStore)store).getDbIterator(false); pfIterator = ((LeveldbTimelineStore)store).getDbIterator(false);
return ((LeveldbTimelineStore)store).deleteNextEntity(entityType, ts, return ((LeveldbTimelineStore)store).deleteNextEntity(entityType, ts,
iterator, pfIterator, false); iterator, pfIterator, false);
} catch(DBException e) {
throw new IOException(e);
} finally { } finally {
IOUtils.cleanup(null, iterator, pfIterator); IOUtils.cleanup(null, iterator, pfIterator);
} }