HADOOP-10005. No need to check INFO severity level is enabled or not. Contributed by Jackie Chang.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1532907 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
parent
f28f5ed628
commit
dea1b2e84a
|
@ -585,6 +585,9 @@ Release 2.1.1-beta - 2013-09-23
|
||||||
HADOOP-9977. Hadoop services won't start with different keypass and
|
HADOOP-9977. Hadoop services won't start with different keypass and
|
||||||
keystorepass when https is enabled. (cnauroth)
|
keystorepass when https is enabled. (cnauroth)
|
||||||
|
|
||||||
|
HADOOP-10005. No need to check INFO severity level is enabled or not.
|
||||||
|
(Jackie Chang via suresh)
|
||||||
|
|
||||||
Release 2.1.0-beta - 2013-08-22
|
Release 2.1.0-beta - 2013-08-22
|
||||||
|
|
||||||
INCOMPATIBLE CHANGES
|
INCOMPATIBLE CHANGES
|
||||||
|
|
|
@ -1315,15 +1315,13 @@ public class DataNode extends Configured
|
||||||
|
|
||||||
int numTargets = xferTargets.length;
|
int numTargets = xferTargets.length;
|
||||||
if (numTargets > 0) {
|
if (numTargets > 0) {
|
||||||
if (LOG.isInfoEnabled()) {
|
StringBuilder xfersBuilder = new StringBuilder();
|
||||||
StringBuilder xfersBuilder = new StringBuilder();
|
for (int i = 0; i < numTargets; i++) {
|
||||||
for (int i = 0; i < numTargets; i++) {
|
xfersBuilder.append(xferTargets[i]);
|
||||||
xfersBuilder.append(xferTargets[i]);
|
xfersBuilder.append(" ");
|
||||||
xfersBuilder.append(" ");
|
|
||||||
}
|
|
||||||
LOG.info(bpReg + " Starting thread to transfer " +
|
|
||||||
block + " to " + xfersBuilder);
|
|
||||||
}
|
}
|
||||||
|
LOG.info(bpReg + " Starting thread to transfer " +
|
||||||
|
block + " to " + xfersBuilder);
|
||||||
|
|
||||||
new Daemon(new DataTransfer(xferTargets, block,
|
new Daemon(new DataTransfer(xferTargets, block,
|
||||||
BlockConstructionStage.PIPELINE_SETUP_CREATE, "")).start();
|
BlockConstructionStage.PIPELINE_SETUP_CREATE, "")).start();
|
||||||
|
|
|
@ -972,12 +972,10 @@ public class MapTask extends Task {
|
||||||
maxRec = kvmeta.capacity() / NMETA;
|
maxRec = kvmeta.capacity() / NMETA;
|
||||||
softLimit = (int)(kvbuffer.length * spillper);
|
softLimit = (int)(kvbuffer.length * spillper);
|
||||||
bufferRemaining = softLimit;
|
bufferRemaining = softLimit;
|
||||||
if (LOG.isInfoEnabled()) {
|
LOG.info(JobContext.IO_SORT_MB + ": " + sortmb);
|
||||||
LOG.info(JobContext.IO_SORT_MB + ": " + sortmb);
|
LOG.info("soft limit at " + softLimit);
|
||||||
LOG.info("soft limit at " + softLimit);
|
LOG.info("bufstart = " + bufstart + "; bufvoid = " + bufvoid);
|
||||||
LOG.info("bufstart = " + bufstart + "; bufvoid = " + bufvoid);
|
LOG.info("kvstart = " + kvstart + "; length = " + maxRec);
|
||||||
LOG.info("kvstart = " + kvstart + "; length = " + maxRec);
|
|
||||||
}
|
|
||||||
|
|
||||||
// k/v serialization
|
// k/v serialization
|
||||||
comparator = job.getOutputKeyComparator();
|
comparator = job.getOutputKeyComparator();
|
||||||
|
@ -1180,10 +1178,8 @@ public class MapTask extends Task {
|
||||||
final int aligned = pos - (pos % METASIZE);
|
final int aligned = pos - (pos % METASIZE);
|
||||||
kvindex =
|
kvindex =
|
||||||
((aligned - METASIZE + kvbuffer.length) % kvbuffer.length) / 4;
|
((aligned - METASIZE + kvbuffer.length) % kvbuffer.length) / 4;
|
||||||
if (LOG.isInfoEnabled()) {
|
LOG.info("(EQUATOR) " + pos + " kvi " + kvindex +
|
||||||
LOG.info("(EQUATOR) " + pos + " kvi " + kvindex +
|
"(" + (kvindex * 4) + ")");
|
||||||
"(" + (kvindex * 4) + ")");
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -1198,10 +1194,8 @@ public class MapTask extends Task {
|
||||||
// set start/end to point to first meta record
|
// set start/end to point to first meta record
|
||||||
kvstart = kvend =
|
kvstart = kvend =
|
||||||
((aligned - METASIZE + kvbuffer.length) % kvbuffer.length) / 4;
|
((aligned - METASIZE + kvbuffer.length) % kvbuffer.length) / 4;
|
||||||
if (LOG.isInfoEnabled()) {
|
LOG.info("(RESET) equator " + e + " kv " + kvstart + "(" +
|
||||||
LOG.info("(RESET) equator " + e + " kv " + kvstart + "(" +
|
(kvstart * 4) + ")" + " kvi " + kvindex + "(" + (kvindex * 4) + ")");
|
||||||
(kvstart * 4) + ")" + " kvi " + kvindex + "(" + (kvindex * 4) + ")");
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -1456,15 +1450,13 @@ public class MapTask extends Task {
|
||||||
if (kvindex != kvend) {
|
if (kvindex != kvend) {
|
||||||
kvend = (kvindex + NMETA) % kvmeta.capacity();
|
kvend = (kvindex + NMETA) % kvmeta.capacity();
|
||||||
bufend = bufmark;
|
bufend = bufmark;
|
||||||
if (LOG.isInfoEnabled()) {
|
LOG.info("Spilling map output");
|
||||||
LOG.info("Spilling map output");
|
LOG.info("bufstart = " + bufstart + "; bufend = " + bufmark +
|
||||||
LOG.info("bufstart = " + bufstart + "; bufend = " + bufmark +
|
"; bufvoid = " + bufvoid);
|
||||||
"; bufvoid = " + bufvoid);
|
LOG.info("kvstart = " + kvstart + "(" + (kvstart * 4) +
|
||||||
LOG.info("kvstart = " + kvstart + "(" + (kvstart * 4) +
|
"); kvend = " + kvend + "(" + (kvend * 4) +
|
||||||
"); kvend = " + kvend + "(" + (kvend * 4) +
|
"); length = " + (distanceTo(kvend, kvstart,
|
||||||
"); length = " + (distanceTo(kvend, kvstart,
|
kvmeta.capacity()) + 1) + "/" + maxRec);
|
||||||
kvmeta.capacity()) + 1) + "/" + maxRec);
|
|
||||||
}
|
|
||||||
sortAndSpill();
|
sortAndSpill();
|
||||||
}
|
}
|
||||||
} catch (InterruptedException e) {
|
} catch (InterruptedException e) {
|
||||||
|
@ -1547,15 +1539,13 @@ public class MapTask extends Task {
|
||||||
kvend = (kvindex + NMETA) % kvmeta.capacity();
|
kvend = (kvindex + NMETA) % kvmeta.capacity();
|
||||||
bufend = bufmark;
|
bufend = bufmark;
|
||||||
spillInProgress = true;
|
spillInProgress = true;
|
||||||
if (LOG.isInfoEnabled()) {
|
LOG.info("Spilling map output");
|
||||||
LOG.info("Spilling map output");
|
LOG.info("bufstart = " + bufstart + "; bufend = " + bufmark +
|
||||||
LOG.info("bufstart = " + bufstart + "; bufend = " + bufmark +
|
"; bufvoid = " + bufvoid);
|
||||||
"; bufvoid = " + bufvoid);
|
LOG.info("kvstart = " + kvstart + "(" + (kvstart * 4) +
|
||||||
LOG.info("kvstart = " + kvstart + "(" + (kvstart * 4) +
|
"); kvend = " + kvend + "(" + (kvend * 4) +
|
||||||
"); kvend = " + kvend + "(" + (kvend * 4) +
|
"); length = " + (distanceTo(kvend, kvstart,
|
||||||
"); length = " + (distanceTo(kvend, kvstart,
|
kvmeta.capacity()) + 1) + "/" + maxRec);
|
||||||
kvmeta.capacity()) + 1) + "/" + maxRec);
|
|
||||||
}
|
|
||||||
spillReady.signal();
|
spillReady.signal();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -81,14 +81,12 @@ public class TestMultiFileInputFormat extends TestCase{
|
||||||
}
|
}
|
||||||
|
|
||||||
public void testFormat() throws IOException {
|
public void testFormat() throws IOException {
|
||||||
if(LOG.isInfoEnabled()) {
|
LOG.info("Test started");
|
||||||
LOG.info("Test started");
|
LOG.info("Max split count = " + MAX_SPLIT_COUNT);
|
||||||
LOG.info("Max split count = " + MAX_SPLIT_COUNT);
|
LOG.info("Split count increment = " + SPLIT_COUNT_INCR);
|
||||||
LOG.info("Split count increment = " + SPLIT_COUNT_INCR);
|
LOG.info("Max bytes per file = " + MAX_BYTES);
|
||||||
LOG.info("Max bytes per file = " + MAX_BYTES);
|
LOG.info("Max number of files = " + MAX_NUM_FILES);
|
||||||
LOG.info("Max number of files = " + MAX_NUM_FILES);
|
LOG.info("Number of files increment = " + NUM_FILES_INCR);
|
||||||
LOG.info("Number of files increment = " + NUM_FILES_INCR);
|
|
||||||
}
|
|
||||||
|
|
||||||
MultiFileInputFormat<Text,Text> format = new DummyMultiFileInputFormat();
|
MultiFileInputFormat<Text,Text> format = new DummyMultiFileInputFormat();
|
||||||
FileSystem fs = FileSystem.getLocal(job);
|
FileSystem fs = FileSystem.getLocal(job);
|
||||||
|
|
|
@ -122,11 +122,9 @@ public class DebugJobProducer implements JobStoryProducer {
|
||||||
// Add/remove excess
|
// Add/remove excess
|
||||||
recs[0] += totalrecs - tot_recs;
|
recs[0] += totalrecs - tot_recs;
|
||||||
bytes[0] += totalbytes - tot_bytes;
|
bytes[0] += totalbytes - tot_bytes;
|
||||||
if (LOG.isInfoEnabled()) {
|
LOG.info(
|
||||||
LOG.info(
|
"DIST: " + Arrays.toString(recs) + " " + tot_recs + "/" + totalrecs +
|
||||||
"DIST: " + Arrays.toString(recs) + " " + tot_recs + "/" + totalrecs +
|
" " + Arrays.toString(bytes) + " " + tot_bytes + "/" + totalbytes);
|
||||||
" " + Arrays.toString(bytes) + " " + tot_bytes + "/" + totalbytes);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private static final AtomicInteger seq = new AtomicInteger(0);
|
private static final AtomicInteger seq = new AtomicInteger(0);
|
||||||
|
|
Loading…
Reference in New Issue