HADOOP-17280. Service-user cost shouldn't be accumulated to totalDecayedCallCost and totalRawCallCost. Contributed by Jinglun.
This commit is contained in:
parent
4c5ad57818
commit
a490d87eb7
|
@ -152,10 +152,15 @@ public class DecayRpcScheduler implements RpcScheduler,
|
|||
private final ConcurrentHashMap<Object, List<AtomicLong>> callCosts =
|
||||
new ConcurrentHashMap<Object, List<AtomicLong>>();
|
||||
|
||||
// Should be the sum of all AtomicLongs in decayed callCosts
|
||||
// Should be the sum of all AtomicLongs in decayed callCosts except
|
||||
// service-user.
|
||||
private final AtomicLong totalDecayedCallCost = new AtomicLong();
|
||||
// The sum of all AtomicLongs in raw callCosts
|
||||
// The sum of all AtomicLongs in raw callCosts except service-user.
|
||||
private final AtomicLong totalRawCallCost = new AtomicLong();
|
||||
// Should be the sum of all AtomicLongs in decayed callCosts of service-user.
|
||||
private final AtomicLong totalServiceUserDecayedCallCost = new AtomicLong();
|
||||
// The sum of all AtomicLongs in raw callCosts of service-user.
|
||||
private final AtomicLong totalServiceUserRawCallCost = new AtomicLong();
|
||||
|
||||
|
||||
// Track total call count and response time in current decay window
|
||||
|
@ -446,6 +451,8 @@ public class DecayRpcScheduler implements RpcScheduler,
|
|||
try {
|
||||
long totalDecayedCost = 0;
|
||||
long totalRawCost = 0;
|
||||
long totalServiceUserDecayedCost = 0;
|
||||
long totalServiceUserRawCost = 0;
|
||||
Iterator<Map.Entry<Object, List<AtomicLong>>> it =
|
||||
callCosts.entrySet().iterator();
|
||||
|
||||
|
@ -456,10 +463,15 @@ public class DecayRpcScheduler implements RpcScheduler,
|
|||
|
||||
|
||||
// Compute the next value by reducing it by the decayFactor
|
||||
totalRawCost += rawCost.get();
|
||||
long currentValue = decayedCost.get();
|
||||
long nextValue = (long) (currentValue * decayFactor);
|
||||
totalDecayedCost += nextValue;
|
||||
if (isServiceUser((String) entry.getKey())) {
|
||||
totalServiceUserRawCost += rawCost.get();
|
||||
totalServiceUserDecayedCost += nextValue;
|
||||
} else {
|
||||
totalRawCost += rawCost.get();
|
||||
totalDecayedCost += nextValue;
|
||||
}
|
||||
decayedCost.set(nextValue);
|
||||
|
||||
LOG.debug(
|
||||
|
@ -478,9 +490,13 @@ public class DecayRpcScheduler implements RpcScheduler,
|
|||
// Update the total so that we remain in sync
|
||||
totalDecayedCallCost.set(totalDecayedCost);
|
||||
totalRawCallCost.set(totalRawCost);
|
||||
totalServiceUserDecayedCallCost.set(totalServiceUserDecayedCost);
|
||||
totalServiceUserRawCallCost.set(totalServiceUserRawCost);
|
||||
|
||||
LOG.debug("After decaying the stored costs, totalDecayedCost: {}, " +
|
||||
"totalRawCallCost: {}.", totalDecayedCost, totalRawCost);
|
||||
LOG.debug("After decaying the stored costs, totalDecayedCost: {}, "
|
||||
+ "totalRawCallCost: {}, totalServiceUserDecayedCost: {},"
|
||||
+ " totalServiceUserRawCost: {}.", totalDecayedCost, totalRawCost,
|
||||
totalServiceUserDecayedCost, totalServiceUserRawCost);
|
||||
// Now refresh the cache of scheduling decisions
|
||||
recomputeScheduleCache();
|
||||
|
||||
|
@ -538,8 +554,13 @@ public class DecayRpcScheduler implements RpcScheduler,
|
|||
}
|
||||
|
||||
// Update the total
|
||||
totalDecayedCallCost.getAndAdd(costDelta);
|
||||
totalRawCallCost.getAndAdd(costDelta);
|
||||
if (!isServiceUser((String) identity)) {
|
||||
totalDecayedCallCost.getAndAdd(costDelta);
|
||||
totalRawCallCost.getAndAdd(costDelta);
|
||||
} else {
|
||||
totalServiceUserDecayedCallCost.getAndAdd(costDelta);
|
||||
totalServiceUserRawCallCost.getAndAdd(costDelta);
|
||||
}
|
||||
|
||||
// At this point value is guaranteed to be not null. It may however have
|
||||
// been clobbered from callCosts. Nonetheless, we return what
|
||||
|
@ -893,6 +914,14 @@ public class DecayRpcScheduler implements RpcScheduler,
|
|||
return totalRawCallCost.get();
|
||||
}
|
||||
|
||||
public long getTotalServiceUserCallVolume() {
|
||||
return totalServiceUserDecayedCallCost.get();
|
||||
}
|
||||
|
||||
public long getTotalServiceUserRawCallVolume() {
|
||||
return totalServiceUserRawCallCost.get();
|
||||
}
|
||||
|
||||
public long[] getResponseTimeCountInLastWindow() {
|
||||
long[] ret = new long[responseTimeCountInLastWindow.length()];
|
||||
for (int i = 0; i < responseTimeCountInLastWindow.length(); i++) {
|
||||
|
@ -922,6 +951,8 @@ public class DecayRpcScheduler implements RpcScheduler,
|
|||
addAvgResponseTimePerPriority(rb);
|
||||
addCallVolumePerPriority(rb);
|
||||
addRawCallVolume(rb);
|
||||
addServiceUserDecayedCallVolume(rb);
|
||||
addServiceUserRawCallVolume(rb);
|
||||
} catch (Exception e) {
|
||||
LOG.warn("Exception thrown while metric collection. Exception : "
|
||||
+ e.getMessage());
|
||||
|
@ -945,6 +976,20 @@ public class DecayRpcScheduler implements RpcScheduler,
|
|||
"incoming Call Volume"), getTotalRawCallVolume());
|
||||
}
|
||||
|
||||
// Key: ServiceUserDecayedCallVolume.
|
||||
private void addServiceUserDecayedCallVolume(MetricsRecordBuilder rb) {
|
||||
rb.addCounter(Interns.info("ServiceUserDecayedCallVolume",
|
||||
"Service-user Decayed Total incoming Call Volume"),
|
||||
getTotalServiceUserCallVolume());
|
||||
}
|
||||
|
||||
// Key: ServiceUserCallVolume.
|
||||
private void addServiceUserRawCallVolume(MetricsRecordBuilder rb) {
|
||||
rb.addCounter(Interns.info("ServiceUserCallVolume",
|
||||
"Service-user Raw Total incoming Call Volume"),
|
||||
getTotalServiceUserRawCallVolume());
|
||||
}
|
||||
|
||||
// Key: Priority.0.CompletedCallVolume
|
||||
private void addCallVolumePerPriority(MetricsRecordBuilder rb) {
|
||||
for (int i = 0; i < responseTimeCountInLastWindow.length(); i++) {
|
||||
|
|
|
@ -2588,7 +2588,8 @@
|
|||
<name>ipc.[port_number].decay-scheduler.service-users</name>
|
||||
<value></value>
|
||||
<description>Service users will always be scheduled into the highest-priority
|
||||
queue. They are specified as a comma-separated list.
|
||||
queue and won't be included in the priority computation of normal user
|
||||
calls. They are specified as a comma-separated list.
|
||||
</description>
|
||||
</property>
|
||||
|
||||
|
|
|
@ -92,7 +92,8 @@ provider simply uses the username of the client submitting the request. However,
|
|||
to performing throttling based on other groupings, or using an external identity provider.
|
||||
|
||||
If particular users submit important requests and you don't want to limit them, you can set them up as the
|
||||
**service-users**. They are always scheduled into the high-priority queue.
|
||||
**service-users**. They are always scheduled into the high-priority queue and won't be included in the priority
|
||||
computation of normal user calls.
|
||||
|
||||
### Cost-based Fair Call Queue
|
||||
|
||||
|
@ -141,7 +142,7 @@ omitted.
|
|||
| decay-scheduler.backoff.responsetime.enable | DecayRpcScheduler | Whether or not to enable the backoff by response time feature. | false |
|
||||
| decay-scheduler.backoff.responsetime.thresholds | DecayRpcScheduler | The response time thresholds, as time durations, for each priority queue. If the average response time for a queue is above this threshold, backoff will occur in lower priority queues. This should be a comma-separated list of length equal to the number of priority levels. | Threshold increases by 10s per level (e.g., for 4 levels: `10s,20s,30s,40s`) |
|
||||
| decay-scheduler.metrics.top.user.count | DecayRpcScheduler | The number of top (i.e., heaviest) users to emit metric information about. | 10 |
|
||||
| decay-scheduler.service-users | DecayRpcScheduler | Service users will always be scheduled into the highest-priority queue. They are specified as a comma-separated list. | |
|
||||
| decay-scheduler.service-users | DecayRpcScheduler | Service users will always be scheduled into the highest-priority queue and won't be included in the priority computation of normal user calls. They are specified as a comma-separated list. | |
|
||||
| weighted-cost.lockshared | WeightedTimeCostProvider | The weight multiplier to apply to the time spent in the processing phase which holds a shared (read) lock. | 10 |
|
||||
| weighted-cost.lockexclusive | WeightedTimeCostProvider | The weight multiplier to apply to the time spent in the processing phase which holds an exclusive (write) lock. | 100 |
|
||||
| weighted-cost.{handler,lockfree,response} | WeightedTimeCostProvider | The weight multiplier to apply to the time spent in the processing phases which do not involve holding a lock. See `org.apache.hadoop.ipc.ProcessingDetails.Timing` for more details on each phase. | 1 |
|
||||
|
|
|
@ -24,6 +24,7 @@ import java.util.Map;
|
|||
import org.eclipse.jetty.util.ajax.JSON;
|
||||
import org.junit.Test;
|
||||
|
||||
import static org.apache.hadoop.ipc.DecayRpcScheduler.IPC_DECAYSCHEDULER_THRESHOLDS_KEY;
|
||||
import static org.junit.Assert.assertEquals;
|
||||
import static org.junit.Assert.assertFalse;
|
||||
import static org.junit.Assert.assertNotEquals;
|
||||
|
@ -389,8 +390,11 @@ public class TestDecayRpcScheduler {
|
|||
return priority;
|
||||
}
|
||||
|
||||
/**
|
||||
* Test computing priorities and priority cache of users and service-users.
|
||||
*/
|
||||
@Test
|
||||
public void testServiceUsers() {
|
||||
public void testServiceUsersCase1() {
|
||||
Configuration conf = new Configuration();
|
||||
conf.setLong("ipc.19."
|
||||
+ DecayRpcScheduler.IPC_SCHEDULER_DECAYSCHEDULER_PERIOD_KEY, 999999);
|
||||
|
@ -421,4 +425,57 @@ public class TestDecayRpcScheduler {
|
|||
assertEquals(0L, summaryMap.get("service1"));
|
||||
assertEquals(0L, summaryMap.get("service2"));
|
||||
}
|
||||
|
||||
/**
|
||||
* Test the service users' calls are not included when computing user's call
|
||||
* priority.
|
||||
*/
|
||||
@Test
|
||||
public void testServiceUsersCase2() {
|
||||
final int level = 4;
|
||||
Configuration conf = new Configuration();
|
||||
conf.setLong("ipc.20."
|
||||
+ DecayRpcScheduler.IPC_SCHEDULER_DECAYSCHEDULER_PERIOD_KEY, 999999);
|
||||
conf.set("ipc.20." + DecayRpcScheduler.IPC_DECAYSCHEDULER_SERVICE_USERS_KEY,
|
||||
"service");
|
||||
conf.set(IPC_DECAYSCHEDULER_THRESHOLDS_KEY, "0.125,0.25,0.5");
|
||||
scheduler = new DecayRpcScheduler(level, "ipc.20", conf);
|
||||
|
||||
// test total costs.
|
||||
for (int i = 0; i < 10; i++) {
|
||||
getPriorityIncrementCallCount("user1");
|
||||
}
|
||||
for (int i = 0; i < 50; i++) {
|
||||
getPriorityIncrementCallCount("service");
|
||||
}
|
||||
assertEquals(10, scheduler.getTotalCallVolume());
|
||||
assertEquals(10, scheduler.getTotalRawCallVolume());
|
||||
assertEquals(50, scheduler.getTotalServiceUserCallVolume());
|
||||
assertEquals(50, scheduler.getTotalServiceUserRawCallVolume());
|
||||
// test priority of normal user.
|
||||
assertEquals(level - 1, scheduler.getPriorityLevel(mockCall("user1")));
|
||||
|
||||
// test total costs after decay.
|
||||
scheduler.forceDecay();
|
||||
assertEquals(5, scheduler.getTotalCallVolume());
|
||||
assertEquals(10, scheduler.getTotalRawCallVolume());
|
||||
assertEquals(25, scheduler.getTotalServiceUserCallVolume());
|
||||
assertEquals(50, scheduler.getTotalServiceUserRawCallVolume());
|
||||
// test priority of normal user.
|
||||
assertEquals(level - 1, scheduler.getPriorityLevel(mockCall("user1")));
|
||||
|
||||
// test total costs again.
|
||||
for (int i = 0; i < 10; i++) {
|
||||
getPriorityIncrementCallCount("user1");
|
||||
}
|
||||
for (int i = 0; i < 50; i++) {
|
||||
getPriorityIncrementCallCount("service");
|
||||
}
|
||||
assertEquals(15, scheduler.getTotalCallVolume());
|
||||
assertEquals(20, scheduler.getTotalRawCallVolume());
|
||||
assertEquals(75, scheduler.getTotalServiceUserCallVolume());
|
||||
assertEquals(100, scheduler.getTotalServiceUserRawCallVolume());
|
||||
// test priority of normal user.
|
||||
assertEquals(level - 1, scheduler.getPriorityLevel(mockCall("user1")));
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue