HBASE-7904 Revert, due to concern over coprocessor exposure

git-svn-id: https://svn.apache.org/repos/asf/hbase/trunk@1463326 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Zhihong Yu 2013-04-01 22:29:33 +00:00
parent 1dbe6394c7
commit 2ec0f55f84
4 changed files with 5 additions and 17 deletions

View File

@ -484,7 +484,7 @@ public class HRegion implements HeapSize { // , Writable{
this.rsAccounting = this.rsServices.getRegionServerAccounting();
// don't initialize coprocessors if not running within a regionserver
// TODO: revisit if coprocessors should load in other cases
this.coprocessorHost = new RegionCoprocessorHost(this, rsServices, baseConf);
this.coprocessorHost = new RegionCoprocessorHost(this, rsServices, conf);
this.metricsRegionWrapper = new MetricsRegionWrapperImpl(this);
this.metricsRegion = new MetricsRegion(this.metricsRegionWrapper);
} else {

View File

@ -1,3 +1,5 @@
/**
*
* Licensed to the Apache Software Foundation (ASF) under one
@ -36,10 +38,8 @@ import java.util.Arrays;
import java.util.Collection;
import java.util.Collections;
import java.util.HashSet;
import java.util.Iterator;
import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
import java.util.NavigableSet;
import java.util.Random;
import java.util.Set;
@ -1648,8 +1648,6 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
forceChangeTaskLogDir();
conf.setBoolean("yarn.nodemanager.pmem-check-enabled", false);
conf.setBoolean("yarn.nodemanager.vmem-check-enabled", false);
// Allow the user to override FS URI for this map-reduce cluster to use.
mrCluster = new MiniMRCluster(servers,
FS_URI != null ? FS_URI : FileSystem.get(conf).getUri().toString(), 1,
@ -1658,7 +1656,6 @@ public class HBaseTestingUtility extends HBaseCommonTestingUtility {
if (jobConf == null) {
jobConf = mrCluster.createJobConf();
}
HBaseConfiguration.merge(this.conf, jobConf);
jobConf.set("mapred.local.dir",
conf.get("mapred.local.dir")); //Hadoop MiniMR overwrites this while it should not
LOG.info("Mini mapreduce cluster started");

View File

@ -228,11 +228,8 @@ public class TestImportExport {
OUTPUT_DIR
};
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(
cluster.getConfiguration()), args);
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
Configuration conf = opts.getConfiguration();
// copy or add the necessary configuration values from the map reduce config to the hbase config
copyConfigurationValues(UTIL.getConfiguration(), conf);
args = opts.getRemainingArgs();
assertEquals(conf.get(Export.EXPORT_BATCHING), EXPORT_BATCH_SIZE);
@ -353,8 +350,6 @@ public class TestImportExport {
GenericOptionsParser opts = new GenericOptionsParser(new Configuration(
cluster.getConfiguration()), args);
Configuration conf = opts.getConfiguration();
// copy or add the necessary configuration values from the map reduce config to the hbase config
copyConfigurationValues(UTIL.getConfiguration(), conf);
args = opts.getRemainingArgs();
Job job = Export.createSubmittableJob(conf, args);
@ -374,8 +369,6 @@ public class TestImportExport {
opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
conf = opts.getConfiguration();
// copy or add the necessary configuration values from the map reduce config to the hbase config
copyConfigurationValues(UTIL.getConfiguration(), conf);
args = opts.getRemainingArgs();
job = Import.createSubmittableJob(conf, args);
@ -399,8 +392,6 @@ public class TestImportExport {
opts = new GenericOptionsParser(new Configuration(cluster.getConfiguration()), args);
conf = opts.getConfiguration();
// copy or add the necessary configuration values from the map reduce config to the hbase config
copyConfigurationValues(UTIL.getConfiguration(), conf);
args = opts.getRemainingArgs();
job = Import.createSubmittableJob(conf, args);

View File

@ -733,7 +733,7 @@
<buildDate>${maven.build.timestamp}</buildDate>
<compileSource>1.6</compileSource>
<!-- Dependencies -->
<hadoop-two.version>2.0.4-SNAPSHOT</hadoop-two.version>
<hadoop-two.version>2.0.2-alpha</hadoop-two.version>
<hadoop-one.version>1.1.2</hadoop-one.version>
<commons-cli.version>1.2</commons-cli.version>
<commons-codec.version>1.7</commons-codec.version>