MAPREDUCE-6054. native-task: Speed up tests. Contributed by Todd Lipcon.
This commit is contained in:
parent
fad4524c85
commit
bfd1d75d87
|
@ -16,4 +16,4 @@ MAPREDUCE-6006. native-task: add native tests to maven and fix bug in pom.xml (B
|
||||||
MAPREDUCE-6026. native-task: fix logging (Manu Zhang via todd)
|
MAPREDUCE-6026. native-task: fix logging (Manu Zhang via todd)
|
||||||
MAPREDUCE-6035. native-task: sources/test-sources jar distribution (Manu Zhang via todd)
|
MAPREDUCE-6035. native-task: sources/test-sources jar distribution (Manu Zhang via todd)
|
||||||
MAPREDUCE-5977. Fix or suppress native-task gcc warnings (Manu Zhang via todd)
|
MAPREDUCE-5977. Fix or suppress native-task gcc warnings (Manu Zhang via todd)
|
||||||
|
MAPREDUCE-6054. native-task: Speed up tests (todd)
|
||||||
|
|
|
@ -23,6 +23,9 @@ import com.google.common.primitives.Longs;
|
||||||
|
|
||||||
public class BytesUtil {
|
public class BytesUtil {
|
||||||
|
|
||||||
|
private static final char[] HEX_CHARS =
|
||||||
|
"0123456789abcdef".toCharArray();
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Converts a big-endian byte array to a long value.
|
* Converts a big-endian byte array to a long value.
|
||||||
*
|
*
|
||||||
|
@ -124,7 +127,9 @@ public class BytesUtil {
|
||||||
|| " `~!@#$%^&*()-_=+[]{}|;:'\",.<>/?".indexOf(ch) >= 0 ) {
|
|| " `~!@#$%^&*()-_=+[]{}|;:'\",.<>/?".indexOf(ch) >= 0 ) {
|
||||||
result.append((char)ch);
|
result.append((char)ch);
|
||||||
} else {
|
} else {
|
||||||
result.append(String.format("\\x%02X", ch));
|
result.append("\\x");
|
||||||
|
result.append(HEX_CHARS[(ch >> 4) & 0x0F]);
|
||||||
|
result.append(HEX_CHARS[ch & 0x0F]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return result.toString();
|
return result.toString();
|
||||||
|
|
|
@ -20,7 +20,11 @@ package org.apache.hadoop.mapred.nativetask.kvtest;
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.zip.CRC32;
|
import java.util.zip.CRC32;
|
||||||
|
|
||||||
|
import com.google.common.base.Stopwatch;
|
||||||
import com.google.common.primitives.Longs;
|
import com.google.common.primitives.Longs;
|
||||||
|
|
||||||
|
import org.apache.commons.logging.Log;
|
||||||
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
|
@ -36,6 +40,7 @@ import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat;
|
||||||
public class KVJob {
|
public class KVJob {
|
||||||
public static final String INPUTPATH = "nativetask.kvtest.inputfile.path";
|
public static final String INPUTPATH = "nativetask.kvtest.inputfile.path";
|
||||||
public static final String OUTPUTPATH = "nativetask.kvtest.outputfile.path";
|
public static final String OUTPUTPATH = "nativetask.kvtest.outputfile.path";
|
||||||
|
private static final Log LOG = LogFactory.getLog(KVJob.class);
|
||||||
Job job = null;
|
Job job = null;
|
||||||
|
|
||||||
public static class ValueMapper<KTYPE, VTYPE> extends Mapper<KTYPE, VTYPE, KTYPE, VTYPE> {
|
public static class ValueMapper<KTYPE, VTYPE> extends Mapper<KTYPE, VTYPE, KTYPE, VTYPE> {
|
||||||
|
@ -82,8 +87,9 @@ public class KVJob {
|
||||||
final TestInputFile testfile = new TestInputFile(Integer.valueOf(conf.get(
|
final TestInputFile testfile = new TestInputFile(Integer.valueOf(conf.get(
|
||||||
TestConstants.FILESIZE_KEY, "1000")),
|
TestConstants.FILESIZE_KEY, "1000")),
|
||||||
keyclass.getName(), valueclass.getName(), conf);
|
keyclass.getName(), valueclass.getName(), conf);
|
||||||
|
Stopwatch sw = new Stopwatch().start();
|
||||||
testfile.createSequenceTestFile(inputpath);
|
testfile.createSequenceTestFile(inputpath);
|
||||||
|
LOG.info("Created test file " + inputpath + " in " + sw.elapsedMillis() + "ms");
|
||||||
}
|
}
|
||||||
job.setInputFormatClass(SequenceFileInputFormat.class);
|
job.setInputFormatClass(SequenceFileInputFormat.class);
|
||||||
FileInputFormat.addInputPath(job, new Path(inputpath));
|
FileInputFormat.addInputPath(job, new Path(inputpath));
|
||||||
|
|
|
@ -20,32 +20,29 @@ package org.apache.hadoop.mapred.nativetask.kvtest;
|
||||||
import static org.junit.Assert.assertEquals;
|
import static org.junit.Assert.assertEquals;
|
||||||
|
|
||||||
import java.io.IOException;
|
import java.io.IOException;
|
||||||
import java.util.ArrayList;
|
import java.util.List;
|
||||||
import java.util.Arrays;
|
|
||||||
|
|
||||||
import org.apache.commons.logging.Log;
|
import org.apache.commons.logging.Log;
|
||||||
import org.apache.commons.logging.LogFactory;
|
import org.apache.commons.logging.LogFactory;
|
||||||
import org.apache.hadoop.conf.Configuration;
|
import org.apache.hadoop.conf.Configuration;
|
||||||
import org.apache.hadoop.fs.FileSystem;
|
import org.apache.hadoop.fs.FileSystem;
|
||||||
import org.apache.hadoop.fs.Path;
|
import org.apache.hadoop.fs.Path;
|
||||||
|
import org.apache.hadoop.io.LongWritable;
|
||||||
import org.apache.hadoop.mapred.nativetask.testutil.ResultVerifier;
|
import org.apache.hadoop.mapred.nativetask.testutil.ResultVerifier;
|
||||||
import org.apache.hadoop.mapred.nativetask.testutil.ScenarioConfiguration;
|
import org.apache.hadoop.mapred.nativetask.testutil.ScenarioConfiguration;
|
||||||
import org.apache.hadoop.mapred.nativetask.testutil.TestConstants;
|
import org.apache.hadoop.mapred.nativetask.testutil.TestConstants;
|
||||||
import org.junit.Before;
|
|
||||||
import org.junit.Test;
|
import org.junit.Test;
|
||||||
import org.junit.runner.RunWith;
|
import org.junit.runner.RunWith;
|
||||||
import org.junit.runners.Parameterized;
|
import org.junit.runners.Parameterized;
|
||||||
import org.junit.runners.Parameterized.Parameters;
|
import org.junit.runners.Parameterized.Parameters;
|
||||||
|
|
||||||
|
import com.google.common.base.Splitter;
|
||||||
|
import com.google.common.collect.Lists;
|
||||||
|
|
||||||
@RunWith(Parameterized.class)
|
@RunWith(Parameterized.class)
|
||||||
public class KVTest {
|
public class KVTest {
|
||||||
private static final Log LOG = LogFactory.getLog(KVTest.class);
|
private static final Log LOG = LogFactory.getLog(KVTest.class);
|
||||||
|
|
||||||
private static Class<?>[] keyclasses = null;
|
|
||||||
private static Class<?>[] valueclasses = null;
|
|
||||||
private static String[] keyclassNames = null;
|
|
||||||
private static String[] valueclassNames = null;
|
|
||||||
|
|
||||||
private static Configuration nativekvtestconf = ScenarioConfiguration.getNativeConfiguration();
|
private static Configuration nativekvtestconf = ScenarioConfiguration.getNativeConfiguration();
|
||||||
private static Configuration hadoopkvtestconf = ScenarioConfiguration.getNormalConfiguration();
|
private static Configuration hadoopkvtestconf = ScenarioConfiguration.getNormalConfiguration();
|
||||||
static {
|
static {
|
||||||
|
@ -53,50 +50,46 @@ public class KVTest {
|
||||||
hadoopkvtestconf.addResource(TestConstants.KVTEST_CONF_PATH);
|
hadoopkvtestconf.addResource(TestConstants.KVTEST_CONF_PATH);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
private static List<Class<?>> parseClassNames(String spec) {
|
||||||
|
List<Class<?>> ret = Lists.newArrayList();
|
||||||
|
Iterable<String> classNames = Splitter.on(';').trimResults()
|
||||||
|
.omitEmptyStrings().split(spec);
|
||||||
|
for (String className : classNames) {
|
||||||
|
try {
|
||||||
|
ret.add(Class.forName(className));
|
||||||
|
} catch (ClassNotFoundException e) {
|
||||||
|
throw new RuntimeException(e);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Parameterize the test with the specified key and value types.
|
||||||
|
*/
|
||||||
@Parameters(name = "key:{0}\nvalue:{1}")
|
@Parameters(name = "key:{0}\nvalue:{1}")
|
||||||
public static Iterable<Class<?>[]> data() {
|
public static Iterable<Class<?>[]> data() throws Exception {
|
||||||
final String valueclassesStr = nativekvtestconf
|
// Parse the config.
|
||||||
|
final String valueClassesStr = nativekvtestconf
|
||||||
.get(TestConstants.NATIVETASK_KVTEST_VALUECLASSES);
|
.get(TestConstants.NATIVETASK_KVTEST_VALUECLASSES);
|
||||||
LOG.info(valueclassesStr);
|
LOG.info("Parameterizing with value classes: " + valueClassesStr);
|
||||||
valueclassNames = valueclassesStr.replaceAll("\\s", "").split(";");// delete
|
List<Class<?>> valueClasses = parseClassNames(valueClassesStr);
|
||||||
// " "
|
|
||||||
final ArrayList<Class<?>> tmpvalueclasses = new ArrayList<Class<?>>();
|
final String keyClassesStr = nativekvtestconf.get(
|
||||||
for (int i = 0; i < valueclassNames.length; i++) {
|
TestConstants.NATIVETASK_KVTEST_KEYCLASSES);
|
||||||
try {
|
LOG.info("Parameterizing with key classes: " + keyClassesStr);
|
||||||
if (valueclassNames[i].equals("")) {
|
List<Class<?>> keyClasses = parseClassNames(keyClassesStr);
|
||||||
continue;
|
|
||||||
}
|
// Generate an entry for each key type.
|
||||||
tmpvalueclasses.add(Class.forName(valueclassNames[i]));
|
List<Class<?>[]> pairs = Lists.newArrayList();
|
||||||
} catch (final ClassNotFoundException e) {
|
for (Class<?> keyClass : keyClasses) {
|
||||||
e.printStackTrace();
|
pairs.add(new Class<?>[]{ keyClass, LongWritable.class });
|
||||||
}
|
|
||||||
}
|
}
|
||||||
valueclasses = tmpvalueclasses.toArray(new Class[tmpvalueclasses.size()]);
|
// ...and for each value type.
|
||||||
final String keyclassesStr = nativekvtestconf.get(TestConstants.NATIVETASK_KVTEST_KEYCLASSES);
|
for (Class<?> valueClass : valueClasses) {
|
||||||
LOG.info(keyclassesStr);
|
pairs.add(new Class<?>[]{ LongWritable.class, valueClass });
|
||||||
keyclassNames = keyclassesStr.replaceAll("\\s", "").split(";");// delete
|
|
||||||
// " "
|
|
||||||
final ArrayList<Class<?>> tmpkeyclasses = new ArrayList<Class<?>>();
|
|
||||||
for (int i = 0; i < keyclassNames.length; i++) {
|
|
||||||
try {
|
|
||||||
if (keyclassNames[i].equals("")) {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
tmpkeyclasses.add(Class.forName(keyclassNames[i]));
|
|
||||||
} catch (final ClassNotFoundException e) {
|
|
||||||
e.printStackTrace();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
keyclasses = tmpkeyclasses.toArray(new Class[tmpkeyclasses.size()]);
|
return pairs;
|
||||||
final Class<?>[][] kvgroup = new Class<?>[keyclassNames.length * valueclassNames.length][2];
|
|
||||||
for (int i = 0; i < keyclassNames.length; i++) {
|
|
||||||
final int tmpindex = i * valueclassNames.length;
|
|
||||||
for (int j = 0; j < valueclassNames.length; j++) {
|
|
||||||
kvgroup[tmpindex + j][0] = keyclasses[i];
|
|
||||||
kvgroup[tmpindex + j][1] = valueclasses[j];
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return Arrays.asList(kvgroup);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
private final Class<?> keyclass;
|
private final Class<?> keyclass;
|
||||||
|
@ -105,12 +98,10 @@ public class KVTest {
|
||||||
public KVTest(Class<?> keyclass, Class<?> valueclass) {
|
public KVTest(Class<?> keyclass, Class<?> valueclass) {
|
||||||
this.keyclass = keyclass;
|
this.keyclass = keyclass;
|
||||||
this.valueclass = valueclass;
|
this.valueclass = valueclass;
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
@Test
|
@Test
|
||||||
public void testKVCompability() {
|
public void testKVCompability() {
|
||||||
|
|
||||||
try {
|
try {
|
||||||
final String nativeoutput = this.runNativeTest(
|
final String nativeoutput = this.runNativeTest(
|
||||||
"Test:" + keyclass.getSimpleName() + "--" + valueclass.getSimpleName(), keyclass, valueclass);
|
"Test:" + keyclass.getSimpleName() + "--" + valueclass.getSimpleName(), keyclass, valueclass);
|
||||||
|
@ -135,11 +126,6 @@ public class KVTest {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@Before
|
|
||||||
public void startUp() {
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
private String runNativeTest(String jobname, Class<?> keyclass, Class<?> valueclass) throws IOException {
|
private String runNativeTest(String jobname, Class<?> keyclass, Class<?> valueclass) throws IOException {
|
||||||
final String inputpath = nativekvtestconf.get(TestConstants.NATIVETASK_KVTEST_INPUTDIR) + "/"
|
final String inputpath = nativekvtestconf.get(TestConstants.NATIVETASK_KVTEST_INPUTDIR) + "/"
|
||||||
+ keyclass.getName()
|
+ keyclass.getName()
|
||||||
|
|
|
@ -36,6 +36,7 @@ import org.apache.hadoop.io.LongWritable;
|
||||||
import org.apache.hadoop.io.SequenceFile;
|
import org.apache.hadoop.io.SequenceFile;
|
||||||
import org.apache.hadoop.io.VIntWritable;
|
import org.apache.hadoop.io.VIntWritable;
|
||||||
import org.apache.hadoop.io.VLongWritable;
|
import org.apache.hadoop.io.VLongWritable;
|
||||||
|
import org.apache.hadoop.io.Writable;
|
||||||
import org.apache.hadoop.mapred.nativetask.testutil.BytesFactory;
|
import org.apache.hadoop.mapred.nativetask.testutil.BytesFactory;
|
||||||
import org.apache.hadoop.mapred.nativetask.testutil.ScenarioConfiguration;
|
import org.apache.hadoop.mapred.nativetask.testutil.ScenarioConfiguration;
|
||||||
import org.apache.hadoop.mapred.nativetask.testutil.TestConstants;
|
import org.apache.hadoop.mapred.nativetask.testutil.TestConstants;
|
||||||
|
@ -123,7 +124,7 @@ public class TestInputFile {
|
||||||
}
|
}
|
||||||
|
|
||||||
public void createSequenceTestFile(String filepath, int base, byte start) throws Exception {
|
public void createSequenceTestFile(String filepath, int base, byte start) throws Exception {
|
||||||
LOG.info("create file " + filepath);
|
LOG.info("creating file " + filepath + "(" + filesize + " bytes)");
|
||||||
LOG.info(keyClsName + " " + valueClsName);
|
LOG.info(keyClsName + " " + valueClsName);
|
||||||
Class<?> tmpkeycls, tmpvaluecls;
|
Class<?> tmpkeycls, tmpvaluecls;
|
||||||
try {
|
try {
|
||||||
|
@ -178,6 +179,9 @@ public class TestInputFile {
|
||||||
int valuebytesnum = 0;
|
int valuebytesnum = 0;
|
||||||
int offset = 0;
|
int offset = 0;
|
||||||
|
|
||||||
|
Writable keyWritable = BytesFactory.newObject(null, keyClsName);
|
||||||
|
Writable valWritable = BytesFactory.newObject(null, valueClsName);
|
||||||
|
|
||||||
while (offset < buflen) {
|
while (offset < buflen) {
|
||||||
final int remains = buflen - offset;
|
final int remains = buflen - offset;
|
||||||
keybytesnum = keyMaxBytesNum;
|
keybytesnum = keyMaxBytesNum;
|
||||||
|
@ -202,9 +206,12 @@ public class TestInputFile {
|
||||||
|
|
||||||
System.arraycopy(databuf, offset, value, 0, valuebytesnum);
|
System.arraycopy(databuf, offset, value, 0, valuebytesnum);
|
||||||
offset += valuebytesnum;
|
offset += valuebytesnum;
|
||||||
|
|
||||||
|
BytesFactory.updateObject(keyWritable, key);
|
||||||
|
BytesFactory.updateObject(valWritable, value);
|
||||||
|
|
||||||
try {
|
try {
|
||||||
writer.append(BytesFactory.newObject(key, this.keyClsName), BytesFactory.newObject(value, this.valueClsName));
|
writer.append(keyWritable, valWritable);
|
||||||
} catch (final IOException e) {
|
} catch (final IOException e) {
|
||||||
e.printStackTrace();
|
e.printStackTrace();
|
||||||
throw new Exception("sequence file create failed", e);
|
throw new Exception("sequence file create failed", e);
|
||||||
|
|
|
@ -19,8 +19,10 @@ package org.apache.hadoop.mapred.nativetask.testutil;
|
||||||
|
|
||||||
import java.util.Random;
|
import java.util.Random;
|
||||||
|
|
||||||
|
import com.google.common.base.Preconditions;
|
||||||
import com.google.common.primitives.Ints;
|
import com.google.common.primitives.Ints;
|
||||||
import com.google.common.primitives.Longs;
|
import com.google.common.primitives.Longs;
|
||||||
|
|
||||||
import org.apache.hadoop.io.BooleanWritable;
|
import org.apache.hadoop.io.BooleanWritable;
|
||||||
import org.apache.hadoop.io.ByteWritable;
|
import org.apache.hadoop.io.ByteWritable;
|
||||||
import org.apache.hadoop.io.BytesWritable;
|
import org.apache.hadoop.io.BytesWritable;
|
||||||
|
@ -32,43 +34,58 @@ import org.apache.hadoop.io.Text;
|
||||||
import org.apache.hadoop.io.UTF8;
|
import org.apache.hadoop.io.UTF8;
|
||||||
import org.apache.hadoop.io.VIntWritable;
|
import org.apache.hadoop.io.VIntWritable;
|
||||||
import org.apache.hadoop.io.VLongWritable;
|
import org.apache.hadoop.io.VLongWritable;
|
||||||
|
import org.apache.hadoop.io.Writable;
|
||||||
import org.apache.hadoop.mapred.nativetask.util.BytesUtil;
|
import org.apache.hadoop.mapred.nativetask.util.BytesUtil;
|
||||||
|
|
||||||
|
|
||||||
public class BytesFactory {
|
public class BytesFactory {
|
||||||
public static Random r = new Random();
|
public static Random r = new Random();
|
||||||
|
|
||||||
public static Object newObject(byte[] seed, String className) {
|
public static void updateObject(Writable obj, byte[] seed) {
|
||||||
r.setSeed(seed.hashCode());
|
if (obj instanceof IntWritable) {
|
||||||
if (className.equals(IntWritable.class.getName())) {
|
((IntWritable)obj).set(Ints.fromByteArray(seed));
|
||||||
return new IntWritable(Ints.fromByteArray(seed));
|
} else if (obj instanceof FloatWritable) {
|
||||||
} else if (className.equals(FloatWritable.class.getName())) {
|
((FloatWritable)obj).set(r.nextFloat());
|
||||||
return new FloatWritable(r.nextFloat());
|
} else if (obj instanceof DoubleWritable) {
|
||||||
} else if (className.equals(DoubleWritable.class.getName())) {
|
((DoubleWritable)obj).set(r.nextDouble());
|
||||||
return new DoubleWritable(r.nextDouble());
|
} else if (obj instanceof LongWritable) {
|
||||||
} else if (className.equals(LongWritable.class.getName())) {
|
((LongWritable)obj).set(Longs.fromByteArray(seed));
|
||||||
return new LongWritable(Longs.fromByteArray(seed));
|
} else if (obj instanceof VIntWritable) {
|
||||||
} else if (className.equals(VIntWritable.class.getName())) {
|
((VIntWritable)obj).set(Ints.fromByteArray(seed));
|
||||||
return new VIntWritable(Ints.fromByteArray(seed));
|
} else if (obj instanceof VLongWritable) {
|
||||||
} else if (className.equals(VLongWritable.class.getName())) {
|
((VLongWritable)obj).set(Longs.fromByteArray(seed));
|
||||||
return new VLongWritable(Longs.fromByteArray(seed));
|
} else if (obj instanceof BooleanWritable) {
|
||||||
} else if (className.equals(BooleanWritable.class.getName())) {
|
((BooleanWritable)obj).set(seed[0] % 2 == 1 ? true : false);
|
||||||
return new BooleanWritable(seed[0] % 2 == 1 ? true : false);
|
} else if (obj instanceof Text) {
|
||||||
} else if (className.equals(Text.class.getName())) {
|
((Text)obj).set(BytesUtil.toStringBinary(seed));
|
||||||
return new Text(BytesUtil.toStringBinary(seed));
|
} else if (obj instanceof ByteWritable) {
|
||||||
} else if (className.equals(ByteWritable.class.getName())) {
|
((ByteWritable)obj).set(seed.length > 0 ? seed[0] : 0);
|
||||||
return new ByteWritable(seed.length > 0 ? seed[0] : 0);
|
} else if (obj instanceof BytesWritable) {
|
||||||
} else if (className.equals(BytesWritable.class.getName())) {
|
((BytesWritable)obj).set(seed, 0, seed.length);
|
||||||
return new BytesWritable(seed);
|
} else if (obj instanceof UTF8) {
|
||||||
} else if (className.equals(UTF8.class.getName())) {
|
((UTF8)obj).set(BytesUtil.toStringBinary(seed));
|
||||||
return new UTF8(BytesUtil.toStringBinary(seed));
|
} else if (obj instanceof MockValueClass) {
|
||||||
} else if (className.equals(MockValueClass.class.getName())) {
|
((MockValueClass)obj).set(seed);
|
||||||
return new MockValueClass(seed);
|
|
||||||
} else {
|
} else {
|
||||||
return null;
|
throw new IllegalArgumentException("unknown writable: " +
|
||||||
|
obj.getClass().getName());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public static Writable newObject(byte[] seed, String className) {
|
||||||
|
Writable ret;
|
||||||
|
try {
|
||||||
|
Class<?> clazz = Class.forName(className);
|
||||||
|
Preconditions.checkArgument(Writable.class.isAssignableFrom(clazz));
|
||||||
|
ret = (Writable)clazz.newInstance();
|
||||||
|
} catch (Exception e) {
|
||||||
|
throw new RuntimeException(e);
|
||||||
|
}
|
||||||
|
if (seed != null) {
|
||||||
|
updateObject(ret, seed);
|
||||||
|
}
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
public static <VTYPE> byte[] fromBytes(byte[] bytes) throws Exception {
|
public static <VTYPE> byte[] fromBytes(byte[] bytes) throws Exception {
|
||||||
throw new Exception("Not supported");
|
throw new Exception("Not supported");
|
||||||
|
|
|
@ -51,6 +51,14 @@ public class MockValueClass implements Writable {
|
||||||
txt = new Text(BytesUtil.toStringBinary(array));
|
txt = new Text(BytesUtil.toStringBinary(array));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void set(byte[] seed) {
|
||||||
|
a = seed.length;
|
||||||
|
array = new byte[seed.length];
|
||||||
|
System.arraycopy(seed, 0, array, 0, seed.length);
|
||||||
|
longWritable.set(a);
|
||||||
|
txt.set(BytesUtil.toStringBinary(array));
|
||||||
|
}
|
||||||
|
|
||||||
@Override
|
@Override
|
||||||
public void write(DataOutput out) throws IOException {
|
public void write(DataOutput out) throws IOException {
|
||||||
out.writeInt(a);
|
out.writeInt(a);
|
||||||
|
|
|
@ -53,4 +53,11 @@ public class TestBytesUtil extends TestCase {
|
||||||
|
|
||||||
Assert.assertEquals(d, BytesUtil.toDouble(doubleBytes));
|
Assert.assertEquals(d, BytesUtil.toDouble(doubleBytes));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
public void testToStringBinary() {
|
||||||
|
Assert.assertEquals("\\x01\\x02ABC",
|
||||||
|
BytesUtil.toStringBinary(new byte[] { 1, 2, 65, 66, 67 }));
|
||||||
|
Assert.assertEquals("\\x10\\x11",
|
||||||
|
BytesUtil.toStringBinary(new byte[] { 16, 17 }));
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue