remove nb structures
This commit is contained in:
parent
05d07b036a
commit
4bb1baa9d4
|
@ -20,9 +20,6 @@
|
||||||
package org.elasticsearch.common.util.concurrent;
|
package org.elasticsearch.common.util.concurrent;
|
||||||
|
|
||||||
import org.elasticsearch.common.collect.MapBackedSet;
|
import org.elasticsearch.common.collect.MapBackedSet;
|
||||||
import org.elasticsearch.common.util.concurrent.highscalelib.NonBlockingHashMap;
|
|
||||||
import org.elasticsearch.common.util.concurrent.highscalelib.NonBlockingHashMapLong;
|
|
||||||
import org.elasticsearch.common.util.concurrent.highscalelib.NonBlockingHashSet;
|
|
||||||
|
|
||||||
import java.util.Set;
|
import java.util.Set;
|
||||||
import java.util.concurrent.ConcurrentHashMap;
|
import java.util.concurrent.ConcurrentHashMap;
|
||||||
|
@ -36,23 +33,23 @@ public abstract class ConcurrentCollections {
|
||||||
private final static boolean useNonBlockingMap = Boolean.parseBoolean(System.getProperty("elasticsearch.useNonBlockingMap", "false"));
|
private final static boolean useNonBlockingMap = Boolean.parseBoolean(System.getProperty("elasticsearch.useNonBlockingMap", "false"));
|
||||||
|
|
||||||
public static <K, V> ConcurrentMap<K, V> newConcurrentMap() {
|
public static <K, V> ConcurrentMap<K, V> newConcurrentMap() {
|
||||||
if (useNonBlockingMap) {
|
// if (useNonBlockingMap) {
|
||||||
return new NonBlockingHashMap<K, V>();
|
// return new NonBlockingHashMap<K, V>();
|
||||||
}
|
// }
|
||||||
return new ConcurrentHashMap<K, V>();
|
return new ConcurrentHashMap<K, V>();
|
||||||
}
|
}
|
||||||
|
|
||||||
public static <V> ConcurrentMapLong<V> newConcurrentMapLong() {
|
public static <V> ConcurrentMapLong<V> newConcurrentMapLong() {
|
||||||
if (useNonBlockingMap) {
|
// if (useNonBlockingMap) {
|
||||||
return new NonBlockingHashMapLong<V>();
|
// return new NonBlockingHashMapLong<V>();
|
||||||
}
|
// }
|
||||||
return new ConcurrentHashMapLong<V>();
|
return new ConcurrentHashMapLong<V>();
|
||||||
}
|
}
|
||||||
|
|
||||||
public static <V> Set<V> newConcurrentSet() {
|
public static <V> Set<V> newConcurrentSet() {
|
||||||
if (useNonBlockingMap) {
|
// if (useNonBlockingMap) {
|
||||||
return new NonBlockingHashSet<V>();
|
// return new NonBlockingHashSet<V>();
|
||||||
}
|
// }
|
||||||
return new MapBackedSet<V>(new ConcurrentHashMap<V, Boolean>());
|
return new MapBackedSet<V>(new ConcurrentHashMap<V, Boolean>());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,102 +0,0 @@
|
||||||
/*
|
|
||||||
* Licensed to Elastic Search and Shay Banon under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. Elastic Search licenses this
|
|
||||||
* file to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Written by Cliff Click and released to the public domain, as explained at
|
|
||||||
* http://creativecommons.org/licenses/publicdomain
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.elasticsearch.common.util.concurrent.highscalelib;
|
|
||||||
|
|
||||||
import java.util.Map;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A simple implementation of {@link java.util.Map.Entry}.
|
|
||||||
* Does not implement {@link java.util.Map.Entry.setValue}, that is done by users of the class.
|
|
||||||
*
|
|
||||||
* @author Cliff Click
|
|
||||||
* @param <TypeK> the type of keys maintained by this map
|
|
||||||
* @param <TypeV> the type of mapped values
|
|
||||||
* @since 1.5
|
|
||||||
*/
|
|
||||||
|
|
||||||
abstract class AbstractEntry<TypeK, TypeV> implements Map.Entry<TypeK, TypeV> {
|
|
||||||
/**
|
|
||||||
* Strongly typed key
|
|
||||||
*/
|
|
||||||
protected final TypeK _key;
|
|
||||||
/**
|
|
||||||
* Strongly typed value
|
|
||||||
*/
|
|
||||||
protected TypeV _val;
|
|
||||||
|
|
||||||
public AbstractEntry(final TypeK key, final TypeV val) {
|
|
||||||
_key = key;
|
|
||||||
_val = val;
|
|
||||||
}
|
|
||||||
|
|
||||||
public AbstractEntry(final Map.Entry<TypeK, TypeV> e) {
|
|
||||||
_key = e.getKey();
|
|
||||||
_val = e.getValue();
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return "key=val" string
|
|
||||||
*/
|
|
||||||
public String toString() {
|
|
||||||
return _key + "=" + _val;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return key
|
|
||||||
*/
|
|
||||||
public TypeK getKey() {
|
|
||||||
return _key;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return val
|
|
||||||
*/
|
|
||||||
public TypeV getValue() {
|
|
||||||
return _val;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Equal if the underlying key & value are equal
|
|
||||||
*/
|
|
||||||
public boolean equals(final Object o) {
|
|
||||||
if (!(o instanceof Map.Entry)) return false;
|
|
||||||
final Map.Entry e = (Map.Entry) o;
|
|
||||||
return eq(_key, e.getKey()) && eq(_val, e.getValue());
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Compute <code>"key.hashCode() ^ val.hashCode()"</code>
|
|
||||||
*/
|
|
||||||
public int hashCode() {
|
|
||||||
return
|
|
||||||
((_key == null) ? 0 : _key.hashCode()) ^
|
|
||||||
((_val == null) ? 0 : _val.hashCode());
|
|
||||||
}
|
|
||||||
|
|
||||||
private static boolean eq(final Object o1, final Object o2) {
|
|
||||||
return (o1 == null ? o2 == null : o1.equals(o2));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
|
@ -1,350 +0,0 @@
|
||||||
/*
|
|
||||||
* Licensed to Elastic Search and Shay Banon under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. Elastic Search licenses this
|
|
||||||
* file to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Written by Cliff Click and released to the public domain, as explained at
|
|
||||||
* http://creativecommons.org/licenses/publicdomain
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.elasticsearch.common.util.concurrent.highscalelib;
|
|
||||||
|
|
||||||
import sun.misc.Unsafe;
|
|
||||||
|
|
||||||
import java.io.Serializable;
|
|
||||||
import java.util.concurrent.atomic.AtomicLongFieldUpdater;
|
|
||||||
import java.util.concurrent.atomic.AtomicReferenceFieldUpdater;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* An auto-resizing table of {@code longs}, supporting low-contention CAS
|
|
||||||
* operations. Updates are done with CAS's to no particular table element.
|
|
||||||
* The intent is to support highly scalable counters, r/w locks, and other
|
|
||||||
* structures where the updates are associative, loss-free (no-brainer), and
|
|
||||||
* otherwise happen at such a high volume that the cache contention for
|
|
||||||
* CAS'ing a single word is unacceptable.
|
|
||||||
* <p/>
|
|
||||||
* <p>This API is overkill for simple counters (e.g. no need for the 'mask')
|
|
||||||
* and is untested as an API for making a scalable r/w lock and so is likely
|
|
||||||
* to change!
|
|
||||||
*
|
|
||||||
* @author Cliff Click
|
|
||||||
* @since 1.5
|
|
||||||
*/
|
|
||||||
|
|
||||||
|
|
||||||
public class ConcurrentAutoTable implements Serializable {
|
|
||||||
|
|
||||||
// --- public interface ---
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Add the given value to current counter value. Concurrent updates will
|
|
||||||
* not be lost, but addAndGet or getAndAdd are not implemented because the
|
|
||||||
* total counter value (i.e., {@link #get}) is not atomically updated.
|
|
||||||
* Updates are striped across an array of counters to avoid cache contention
|
|
||||||
* and has been tested with performance scaling linearly up to 768 CPUs.
|
|
||||||
*/
|
|
||||||
public void add(long x) {
|
|
||||||
add_if_mask(x, 0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* {@link #add} with -1
|
|
||||||
*/
|
|
||||||
public void decrement() {
|
|
||||||
add_if_mask(-1L, 0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* {@link #add} with +1
|
|
||||||
*/
|
|
||||||
public void increment() {
|
|
||||||
add_if_mask(1L, 0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Atomically set the sum of the striped counters to specified value.
|
|
||||||
* Rather more expensive than a simple store, in order to remain atomic.
|
|
||||||
*/
|
|
||||||
public void set(long x) {
|
|
||||||
CAT newcat = new CAT(null, 4, x);
|
|
||||||
// Spin until CAS works
|
|
||||||
while (!CAS_cat(_cat, newcat)) ;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Current value of the counter. Since other threads are updating furiously
|
|
||||||
* the value is only approximate, but it includes all counts made by the
|
|
||||||
* current thread. Requires a pass over the internally striped counters.
|
|
||||||
*/
|
|
||||||
public long get() {
|
|
||||||
return _cat.sum(0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Same as {@link #get}, included for completeness.
|
|
||||||
*/
|
|
||||||
public int intValue() {
|
|
||||||
return (int) _cat.sum(0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Same as {@link #get}, included for completeness.
|
|
||||||
*/
|
|
||||||
public long longValue() {
|
|
||||||
return _cat.sum(0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A cheaper {@link #get}. Updated only once/millisecond, but as fast as a
|
|
||||||
* simple load instruction when not updating.
|
|
||||||
*/
|
|
||||||
public long estimate_get() {
|
|
||||||
return _cat.estimate_sum(0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return the counter's {@code long} value converted to a string.
|
|
||||||
*/
|
|
||||||
public String toString() {
|
|
||||||
return _cat.toString(0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A more verbose print than {@link #toString}, showing internal structure.
|
|
||||||
* Useful for debugging.
|
|
||||||
*/
|
|
||||||
public void print() {
|
|
||||||
_cat.print();
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Return the internal counter striping factor. Useful for diagnosing
|
|
||||||
* performance problems.
|
|
||||||
*/
|
|
||||||
public int internal_size() {
|
|
||||||
return _cat._t.length;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Only add 'x' to some slot in table, hinted at by 'hash', if bits under
|
|
||||||
// the mask are all zero. The sum can overflow or 'x' can contain bits in
|
|
||||||
// the mask. Value is CAS'd so no counts are lost. The CAS is retried until
|
|
||||||
// it succeeds or bits are found under the mask. Returned value is the old
|
|
||||||
// value - which WILL have zero under the mask on success and WILL NOT have
|
|
||||||
// zero under the mask for failure.
|
|
||||||
|
|
||||||
private long add_if_mask(long x, long mask) {
|
|
||||||
return _cat.add_if_mask(x, mask, hash(), this);
|
|
||||||
}
|
|
||||||
|
|
||||||
// The underlying array of concurrently updated long counters
|
|
||||||
private volatile CAT _cat = new CAT(null, 4/*Start Small, Think Big!*/, 0L);
|
|
||||||
private static final AtomicReferenceFieldUpdater<ConcurrentAutoTable, CAT> _catUpdater =
|
|
||||||
AtomicReferenceFieldUpdater.newUpdater(ConcurrentAutoTable.class, CAT.class, "_cat");
|
|
||||||
|
|
||||||
private boolean CAS_cat(CAT oldcat, CAT newcat) {
|
|
||||||
return _catUpdater.compareAndSet(this, oldcat, newcat);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Hash spreader
|
|
||||||
|
|
||||||
private static final int hash() {
|
|
||||||
int h = System.identityHashCode(Thread.currentThread());
|
|
||||||
// You would think that System.identityHashCode on the current thread
|
|
||||||
// would be a good hash fcn, but actually on SunOS 5.8 it is pretty lousy
|
|
||||||
// in the low bits.
|
|
||||||
h ^= (h >>> 20) ^ (h >>> 12); // Bit spreader, borrowed from Doug Lea
|
|
||||||
h ^= (h >>> 7) ^ (h >>> 4);
|
|
||||||
return h << 2; // Pad out cache lines. The goal is to avoid cache-line contention
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- CAT -----------------------------------------------------------------
|
|
||||||
|
|
||||||
private static class CAT implements Serializable {
|
|
||||||
|
|
||||||
// Unsafe crud: get a function which will CAS arrays
|
|
||||||
private static final Unsafe _unsafe = UtilUnsafe.getUnsafe();
|
|
||||||
private static final int _Lbase = _unsafe.arrayBaseOffset(long[].class);
|
|
||||||
private static final int _Lscale = _unsafe.arrayIndexScale(long[].class);
|
|
||||||
|
|
||||||
private static long rawIndex(long[] ary, int i) {
|
|
||||||
assert i >= 0 && i < ary.length;
|
|
||||||
return _Lbase + i * _Lscale;
|
|
||||||
}
|
|
||||||
|
|
||||||
private final static boolean CAS(long[] A, int idx, long old, long nnn) {
|
|
||||||
return _unsafe.compareAndSwapLong(A, rawIndex(A, idx), old, nnn);
|
|
||||||
}
|
|
||||||
|
|
||||||
volatile long _resizers; // count of threads attempting a resize
|
|
||||||
static private final AtomicLongFieldUpdater<CAT> _resizerUpdater =
|
|
||||||
AtomicLongFieldUpdater.newUpdater(CAT.class, "_resizers");
|
|
||||||
|
|
||||||
private final CAT _next;
|
|
||||||
private volatile long _sum_cache;
|
|
||||||
private volatile long _fuzzy_sum_cache;
|
|
||||||
private volatile long _fuzzy_time;
|
|
||||||
private static final int MAX_SPIN = 2;
|
|
||||||
private long[] _t; // Power-of-2 array of longs
|
|
||||||
|
|
||||||
CAT(CAT next, int sz, long init) {
|
|
||||||
_next = next;
|
|
||||||
_sum_cache = Long.MIN_VALUE;
|
|
||||||
_t = new long[sz];
|
|
||||||
_t[0] = init;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Only add 'x' to some slot in table, hinted at by 'hash', if bits under
|
|
||||||
// the mask are all zero. The sum can overflow or 'x' can contain bits in
|
|
||||||
// the mask. Value is CAS'd so no counts are lost. The CAS is attempted
|
|
||||||
// ONCE.
|
|
||||||
|
|
||||||
public long add_if_mask(long x, long mask, int hash, ConcurrentAutoTable master) {
|
|
||||||
long[] t = _t;
|
|
||||||
int idx = hash & (t.length - 1);
|
|
||||||
// Peel loop; try once fast
|
|
||||||
long old = t[idx];
|
|
||||||
boolean ok = CAS(t, idx, old & ~mask, old + x);
|
|
||||||
if (_sum_cache != Long.MIN_VALUE)
|
|
||||||
_sum_cache = Long.MIN_VALUE; // Blow out cache
|
|
||||||
if (ok) return old; // Got it
|
|
||||||
if ((old & mask) != 0) return old; // Failed for bit-set under mask
|
|
||||||
// Try harder
|
|
||||||
int cnt = 0;
|
|
||||||
while (true) {
|
|
||||||
old = t[idx];
|
|
||||||
if ((old & mask) != 0) return old; // Failed for bit-set under mask
|
|
||||||
if (CAS(t, idx, old, old + x)) break; // Got it!
|
|
||||||
cnt++;
|
|
||||||
}
|
|
||||||
if (cnt < MAX_SPIN) return old; // Allowable spin loop count
|
|
||||||
if (t.length >= 1024 * 1024) return old; // too big already
|
|
||||||
|
|
||||||
// Too much contention; double array size in an effort to reduce contention
|
|
||||||
long r = _resizers;
|
|
||||||
int newbytes = (t.length << 1) << 3/*word to bytes*/;
|
|
||||||
while (!_resizerUpdater.compareAndSet(this, r, r + newbytes))
|
|
||||||
r = _resizers;
|
|
||||||
r += newbytes;
|
|
||||||
if (master._cat != this) return old; // Already doubled, don't bother
|
|
||||||
if ((r >> 17) != 0) { // Already too much allocation attempts?
|
|
||||||
// TODO - use a wait with timeout, so we'll wakeup as soon as the new
|
|
||||||
// table is ready, or after the timeout in any case. Annoyingly, this
|
|
||||||
// breaks the non-blocking property - so for now we just briefly sleep.
|
|
||||||
//synchronized( this ) { wait(8*megs); } // Timeout - we always wakeup
|
|
||||||
try {
|
|
||||||
Thread.sleep(r >> 17);
|
|
||||||
} catch (InterruptedException e) {
|
|
||||||
}
|
|
||||||
if (master._cat != this) return old;
|
|
||||||
}
|
|
||||||
|
|
||||||
CAT newcat = new CAT(this, t.length * 2, 0);
|
|
||||||
// Take 1 stab at updating the CAT with the new larger size. If this
|
|
||||||
// fails, we assume some other thread already expanded the CAT - so we
|
|
||||||
// do not need to retry until it succeeds.
|
|
||||||
master.CAS_cat(this, newcat);
|
|
||||||
return old;
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
// Return the current sum of all things in the table, stripping off mask
|
|
||||||
// before the add. Writers can be updating the table furiously, so the
|
|
||||||
// sum is only locally accurate.
|
|
||||||
|
|
||||||
public long sum(long mask) {
|
|
||||||
long sum = _sum_cache;
|
|
||||||
if (sum != Long.MIN_VALUE) return sum;
|
|
||||||
sum = _next == null ? 0 : _next.sum(mask); // Recursively get cached sum
|
|
||||||
long[] t = _t;
|
|
||||||
for (int i = 0; i < t.length; i++)
|
|
||||||
sum += t[i] & (~mask);
|
|
||||||
_sum_cache = sum; // Cache includes recursive counts
|
|
||||||
return sum;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Fast fuzzy version. Used a cached value until it gets old, then re-up
|
|
||||||
// the cache.
|
|
||||||
|
|
||||||
public long estimate_sum(long mask) {
|
|
||||||
// For short tables, just do the work
|
|
||||||
if (_t.length <= 64) return sum(mask);
|
|
||||||
// For bigger tables, periodically freshen a cached value
|
|
||||||
long millis = System.currentTimeMillis();
|
|
||||||
if (_fuzzy_time != millis) { // Time marches on?
|
|
||||||
_fuzzy_sum_cache = sum(mask); // Get sum the hard way
|
|
||||||
_fuzzy_time = millis; // Indicate freshness of cached value
|
|
||||||
}
|
|
||||||
return _fuzzy_sum_cache; // Return cached sum
|
|
||||||
}
|
|
||||||
|
|
||||||
// Update all table slots with CAS.
|
|
||||||
|
|
||||||
public void all_or(long mask) {
|
|
||||||
long[] t = _t;
|
|
||||||
for (int i = 0; i < t.length; i++) {
|
|
||||||
boolean done = false;
|
|
||||||
while (!done) {
|
|
||||||
long old = t[i];
|
|
||||||
done = CAS(t, i, old, old | mask);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (_next != null) _next.all_or(mask);
|
|
||||||
if (_sum_cache != Long.MIN_VALUE)
|
|
||||||
_sum_cache = Long.MIN_VALUE; // Blow out cache
|
|
||||||
}
|
|
||||||
|
|
||||||
public void all_and(long mask) {
|
|
||||||
long[] t = _t;
|
|
||||||
for (int i = 0; i < t.length; i++) {
|
|
||||||
boolean done = false;
|
|
||||||
while (!done) {
|
|
||||||
long old = t[i];
|
|
||||||
done = CAS(t, i, old, old & mask);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (_next != null) _next.all_and(mask);
|
|
||||||
if (_sum_cache != Long.MIN_VALUE)
|
|
||||||
_sum_cache = Long.MIN_VALUE; // Blow out cache
|
|
||||||
}
|
|
||||||
|
|
||||||
// Set/stomp all table slots. No CAS.
|
|
||||||
|
|
||||||
public void all_set(long val) {
|
|
||||||
long[] t = _t;
|
|
||||||
for (int i = 0; i < t.length; i++)
|
|
||||||
t[i] = val;
|
|
||||||
if (_next != null) _next.all_set(val);
|
|
||||||
if (_sum_cache != Long.MIN_VALUE)
|
|
||||||
_sum_cache = Long.MIN_VALUE; // Blow out cache
|
|
||||||
}
|
|
||||||
|
|
||||||
String toString(long mask) {
|
|
||||||
return Long.toString(sum(mask));
|
|
||||||
}
|
|
||||||
|
|
||||||
public void print() {
|
|
||||||
long[] t = _t;
|
|
||||||
System.out.print("[sum=" + _sum_cache + "," + t[0]);
|
|
||||||
for (int i = 1; i < t.length; i++)
|
|
||||||
System.out.print("," + t[i]);
|
|
||||||
System.out.print("]");
|
|
||||||
if (_next != null) _next.print();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
|
@ -1,58 +0,0 @@
|
||||||
/*
|
|
||||||
* Licensed to Elastic Search and Shay Banon under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. Elastic Search licenses this
|
|
||||||
* file to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Written by Cliff Click and released to the public domain, as explained at
|
|
||||||
* http://creativecommons.org/licenses/publicdomain
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.elasticsearch.common.util.concurrent.highscalelib;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A simple high-performance counter. Merely renames the extended {@link
|
|
||||||
* org.cliffc.high_scale_lib.ConcurrentAutoTable} class to be more obvious.
|
|
||||||
* {@link org.cliffc.high_scale_lib.ConcurrentAutoTable} already has a decent
|
|
||||||
* counting API.
|
|
||||||
*
|
|
||||||
* @author Cliff Click
|
|
||||||
* @since 1.5
|
|
||||||
*/
|
|
||||||
|
|
||||||
public class Counter extends ConcurrentAutoTable {
|
|
||||||
|
|
||||||
// Add the given value to current counter value. Concurrent updates will
|
|
||||||
// not be lost, but addAndGet or getAndAdd are not implemented because but
|
|
||||||
// the total counter value is not atomically updated.
|
|
||||||
//public void add( long x );
|
|
||||||
//public void decrement();
|
|
||||||
//public void increment();
|
|
||||||
|
|
||||||
// Current value of the counter. Since other threads are updating furiously
|
|
||||||
// the value is only approximate, but it includes all counts made by the
|
|
||||||
// current thread. Requires a pass over all the striped counters.
|
|
||||||
//public long get();
|
|
||||||
//public int intValue();
|
|
||||||
//public long longValue();
|
|
||||||
|
|
||||||
// A cheaper 'get'. Updated only once/millisecond, but fast as a simple
|
|
||||||
// load instruction when not updating.
|
|
||||||
//public long estimate_get( );
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
|
@ -1,134 +0,0 @@
|
||||||
/*
|
|
||||||
* Licensed to Elastic Search and Shay Banon under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. Elastic Search licenses this
|
|
||||||
* file to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Written by Cliff Click and released to the public domain, as explained at
|
|
||||||
* http://creativecommons.org/licenses/publicdomain
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.elasticsearch.common.util.concurrent.highscalelib;
|
|
||||||
|
|
||||||
import java.io.Serializable;
|
|
||||||
import java.util.AbstractSet;
|
|
||||||
import java.util.Iterator;
|
|
||||||
import java.util.Set;
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A simple wrapper around {@link NonBlockingHashMap} making it implement the
|
|
||||||
* {@link Set} interface. All operations are Non-Blocking and multi-thread safe.
|
|
||||||
*
|
|
||||||
* @author Cliff Click
|
|
||||||
* @since 1.5
|
|
||||||
*/
|
|
||||||
|
|
||||||
|
|
||||||
public class NonBlockingHashSet<E> extends AbstractSet<E> implements Serializable {
|
|
||||||
private static final Object V = "";
|
|
||||||
|
|
||||||
private final NonBlockingHashMap<E, Object> _map;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Make a new empty {@link NonBlockingHashSet}.
|
|
||||||
*/
|
|
||||||
public NonBlockingHashSet() {
|
|
||||||
super();
|
|
||||||
_map = new NonBlockingHashMap<E, Object>();
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Add {@code o} to the set.
|
|
||||||
*
|
|
||||||
* @return <tt>true</tt> if {@code o} was added to the set, <tt>false</tt>
|
|
||||||
* if {@code o} was already in the set.
|
|
||||||
*/
|
|
||||||
public boolean add(final E o) {
|
|
||||||
return _map.putIfAbsent(o, V) != V;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @return <tt>true</tt> if {@code o} is in the set.
|
|
||||||
*/
|
|
||||||
public boolean contains(final Object o) {
|
|
||||||
return _map.containsKey(o);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Remove {@code o} from the set.
|
|
||||||
*
|
|
||||||
* @return <tt>true</tt> if {@code o} was removed to the set, <tt>false</tt>
|
|
||||||
* if {@code o} was not in the set.
|
|
||||||
*/
|
|
||||||
public boolean remove(final Object o) {
|
|
||||||
return _map.remove(o) == V;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Current count of elements in the set. Due to concurrent racing updates,
|
|
||||||
* the size is only ever approximate. Updates due to the calling thread are
|
|
||||||
* immediately visible to calling thread.
|
|
||||||
*
|
|
||||||
* @return count of elements.
|
|
||||||
*/
|
|
||||||
public int size() {
|
|
||||||
return _map.size();
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Empty the set.
|
|
||||||
*/
|
|
||||||
public void clear() {
|
|
||||||
_map.clear();
|
|
||||||
}
|
|
||||||
|
|
||||||
public Iterator<E> iterator() {
|
|
||||||
return _map.keySet().iterator();
|
|
||||||
}
|
|
||||||
|
|
||||||
// ---
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Atomically make the set immutable. Future calls to mutate will throw an
|
|
||||||
* IllegalStateException. Existing mutator calls in other threads racing
|
|
||||||
* with this thread and will either throw IllegalStateException or their
|
|
||||||
* update will be visible to this thread. This implies that a simple flag
|
|
||||||
* cannot make the Set immutable, because a late-arriving update in another
|
|
||||||
* thread might see immutable flag not set yet, then mutate the Set after
|
|
||||||
* the {@link #readOnly} call returns. This call can be called concurrently
|
|
||||||
* (and indeed until the operation completes, all calls on the Set from any
|
|
||||||
* thread either complete normally or end up calling {@link #readOnly}
|
|
||||||
* internally).
|
|
||||||
* <p/>
|
|
||||||
* <p> This call is useful in debugging multi-threaded programs where the
|
|
||||||
* Set is constructed in parallel, but construction completes after some
|
|
||||||
* time; and after construction the Set is only read. Making the Set
|
|
||||||
* read-only will cause updates arriving after construction is supposedly
|
|
||||||
* complete to throw an {@link IllegalStateException}.
|
|
||||||
*/
|
|
||||||
|
|
||||||
// (1) call _map's immutable() call
|
|
||||||
// (2) get snapshot
|
|
||||||
// (3) CAS down a local map, power-of-2 larger than _map.size()+1/8th
|
|
||||||
// (4) start @ random, visit all snapshot, insert live keys
|
|
||||||
// (5) CAS _map to null, needs happens-after (4)
|
|
||||||
// (6) if Set call sees _map is null, needs happens-after (4) for readers
|
|
||||||
public void readOnly() {
|
|
||||||
throw new RuntimeException("Unimplemented");
|
|
||||||
}
|
|
||||||
}
|
|
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
|
@ -1,555 +0,0 @@
|
||||||
/*
|
|
||||||
* Licensed to Elastic Search and Shay Banon under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. Elastic Search licenses this
|
|
||||||
* file to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Written by Cliff Click and released to the public domain, as explained at
|
|
||||||
* http://creativecommons.org/licenses/publicdomain
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.elasticsearch.common.util.concurrent.highscalelib;
|
|
||||||
|
|
||||||
import sun.misc.Unsafe;
|
|
||||||
|
|
||||||
import java.io.IOException;
|
|
||||||
import java.io.Serializable;
|
|
||||||
import java.lang.reflect.Field;
|
|
||||||
import java.util.AbstractSet;
|
|
||||||
import java.util.Iterator;
|
|
||||||
import java.util.NoSuchElementException;
|
|
||||||
import java.util.concurrent.atomic.AtomicInteger;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A multi-threaded bit-vector set, implemented as an array of primitive
|
|
||||||
* {@code longs}. All operations are non-blocking and multi-threaded safe.
|
|
||||||
* {@link #contains(int)} calls are roughly the same speed as a {load, mask}
|
|
||||||
* sequence. {@link #add(int)} and {@link #remove(int)} calls are a tad more
|
|
||||||
* expensive than a {load, mask, store} sequence because they must use a CAS.
|
|
||||||
* The bit-vector is auto-sizing.
|
|
||||||
* <p/>
|
|
||||||
* <p><em>General note of caution:</em> The Set API allows the use of {@link Integer}
|
|
||||||
* with silent autoboxing - which can be very expensive if many calls are
|
|
||||||
* being made. Since autoboxing is silent you may not be aware that this is
|
|
||||||
* going on. The built-in API takes lower-case {@code ints} and is much more
|
|
||||||
* efficient.
|
|
||||||
* <p/>
|
|
||||||
* <p>Space: space is used in proportion to the largest element, as opposed to
|
|
||||||
* the number of elements (as is the case with hash-table based Set
|
|
||||||
* implementations). Space is approximately (largest_element/8 + 64) bytes.
|
|
||||||
* <p/>
|
|
||||||
* The implementation is a simple bit-vector using CAS for update.
|
|
||||||
*
|
|
||||||
* @author Cliff Click
|
|
||||||
* @since 1.5
|
|
||||||
*/
|
|
||||||
|
|
||||||
public class NonBlockingSetInt extends AbstractSet<Integer> implements Serializable {
|
|
||||||
private static final long serialVersionUID = 1234123412341234123L;
|
|
||||||
private static final Unsafe _unsafe = UtilUnsafe.getUnsafe();
|
|
||||||
|
|
||||||
// --- Bits to allow atomic update of the NBSI
|
|
||||||
private static final long _nbsi_offset;
|
|
||||||
|
|
||||||
static { // <clinit>
|
|
||||||
Field f = null;
|
|
||||||
try {
|
|
||||||
f = NonBlockingSetInt.class.getDeclaredField("_nbsi");
|
|
||||||
} catch (java.lang.NoSuchFieldException e) {
|
|
||||||
}
|
|
||||||
_nbsi_offset = _unsafe.objectFieldOffset(f);
|
|
||||||
}
|
|
||||||
|
|
||||||
private final boolean CAS_nbsi(NBSI old, NBSI nnn) {
|
|
||||||
return _unsafe.compareAndSwapObject(this, _nbsi_offset, old, nnn);
|
|
||||||
}
|
|
||||||
|
|
||||||
// The actual Set of Joy, which changes during a resize event. The
|
|
||||||
// Only Field for this class, so I can atomically change the entire
|
|
||||||
// set implementation with a single CAS.
|
|
||||||
private transient NBSI _nbsi;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Create a new empty bit-vector
|
|
||||||
*/
|
|
||||||
public NonBlockingSetInt() {
|
|
||||||
_nbsi = new NBSI(63, new Counter(), this); // The initial 1-word set
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Add {@code i} to the set. Uppercase {@link Integer} version of add,
|
|
||||||
* requires auto-unboxing. When possible use the {@code int} version of
|
|
||||||
* {@link #add(int)} for efficiency.
|
|
||||||
*
|
|
||||||
* @return <tt>true</tt> if i was added to the set.
|
|
||||||
* @throws IllegalArgumentException if i is negative.
|
|
||||||
*/
|
|
||||||
public boolean add(final Integer i) {
|
|
||||||
return add(i.intValue());
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Test if {@code o} is in the set. This is the uppercase {@link Integer}
|
|
||||||
* version of contains, requires a type-check and auto-unboxing. When
|
|
||||||
* possible use the {@code int} version of {@link #contains(int)} for
|
|
||||||
* efficiency.
|
|
||||||
*
|
|
||||||
* @return <tt>true</tt> if i was in the set.
|
|
||||||
*/
|
|
||||||
public boolean contains(final Object o) {
|
|
||||||
return o instanceof Integer ? contains(((Integer) o).intValue()) : false;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Remove {@code o} from the set. This is the uppercase {@link Integer}
|
|
||||||
* version of remove, requires a type-check and auto-unboxing. When
|
|
||||||
* possible use the {@code int} version of {@link #remove(int)} for
|
|
||||||
* efficiency.
|
|
||||||
*
|
|
||||||
* @return <tt>true</tt> if i was removed to the set.
|
|
||||||
*/
|
|
||||||
public boolean remove(final Object o) {
|
|
||||||
return o instanceof Integer ? remove(((Integer) o).intValue()) : false;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Add {@code i} to the set. This is the lower-case '{@code int}' version
|
|
||||||
* of {@link #add} - no autoboxing. Negative values throw
|
|
||||||
* IllegalArgumentException.
|
|
||||||
*
|
|
||||||
* @return <tt>true</tt> if i was added to the set.
|
|
||||||
* @throws IllegalArgumentException if i is negative.
|
|
||||||
*/
|
|
||||||
public boolean add(final int i) {
|
|
||||||
if (i < 0) throw new IllegalArgumentException("" + i);
|
|
||||||
return _nbsi.add(i);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Test if {@code i} is in the set. This is the lower-case '{@code int}'
|
|
||||||
* version of {@link #contains} - no autoboxing.
|
|
||||||
*
|
|
||||||
* @return <tt>true</tt> if i was int the set.
|
|
||||||
*/
|
|
||||||
public boolean contains(final int i) {
|
|
||||||
return i < 0 ? false : _nbsi.contains(i);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Remove {@code i} from the set. This is the fast lower-case '{@code int}'
|
|
||||||
* version of {@link #remove} - no autoboxing.
|
|
||||||
*
|
|
||||||
* @return <tt>true</tt> if i was added to the set.
|
|
||||||
*/
|
|
||||||
public boolean remove(final int i) {
|
|
||||||
return i < 0 ? false : _nbsi.remove(i);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Current count of elements in the set. Due to concurrent racing updates,
|
|
||||||
* the size is only ever approximate. Updates due to the calling thread are
|
|
||||||
* immediately visible to calling thread.
|
|
||||||
*
|
|
||||||
* @return count of elements.
|
|
||||||
*/
|
|
||||||
public int size() {
|
|
||||||
return _nbsi.size();
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Empty the bitvector.
|
|
||||||
*/
|
|
||||||
public void clear() {
|
|
||||||
NBSI cleared = new NBSI(63, new Counter(), this); // An empty initial NBSI
|
|
||||||
while (!CAS_nbsi(_nbsi, cleared)) // Spin until clear works
|
|
||||||
;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Verbose printout of internal structure for debugging.
|
|
||||||
*/
|
|
||||||
public void print() {
|
|
||||||
_nbsi.print(0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Standard Java {@link Iterator}. Not very efficient because it
|
|
||||||
* auto-boxes the returned values.
|
|
||||||
*/
|
|
||||||
public Iterator<Integer> iterator() {
|
|
||||||
return new iter();
|
|
||||||
}
|
|
||||||
|
|
||||||
private class iter implements Iterator<Integer> {
|
|
||||||
NBSI _nbsi2;
|
|
||||||
int _idx = -1;
|
|
||||||
int _prev = -1;
|
|
||||||
|
|
||||||
iter() {
|
|
||||||
_nbsi2 = _nbsi;
|
|
||||||
advance();
|
|
||||||
}
|
|
||||||
|
|
||||||
public boolean hasNext() {
|
|
||||||
return _idx != -2;
|
|
||||||
}
|
|
||||||
|
|
||||||
private void advance() {
|
|
||||||
while (true) {
|
|
||||||
_idx++; // Next index
|
|
||||||
while ((_idx >> 6) >= _nbsi2._bits.length) { // Index out of range?
|
|
||||||
if (_nbsi2._new == null) { // New table?
|
|
||||||
_idx = -2; // No, so must be all done
|
|
||||||
return; //
|
|
||||||
}
|
|
||||||
_nbsi2 = _nbsi2._new; // Carry on, in the new table
|
|
||||||
}
|
|
||||||
if (_nbsi2.contains(_idx)) return;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
public Integer next() {
|
|
||||||
if (_idx == -1) throw new NoSuchElementException();
|
|
||||||
_prev = _idx;
|
|
||||||
advance();
|
|
||||||
return _prev;
|
|
||||||
}
|
|
||||||
|
|
||||||
public void remove() {
|
|
||||||
if (_prev == -1) throw new IllegalStateException();
|
|
||||||
_nbsi2.remove(_prev);
|
|
||||||
_prev = -1;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- writeObject -------------------------------------------------------
|
|
||||||
// Write a NBSI to a stream
|
|
||||||
|
|
||||||
private void writeObject(java.io.ObjectOutputStream s) throws IOException {
|
|
||||||
s.defaultWriteObject(); // Nothing to write
|
|
||||||
final NBSI nbsi = _nbsi; // The One Field is transient
|
|
||||||
final int len = _nbsi._bits.length << 6;
|
|
||||||
s.writeInt(len); // Write max element
|
|
||||||
for (int i = 0; i < len; i++)
|
|
||||||
s.writeBoolean(_nbsi.contains(i));
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- readObject --------------------------------------------------------
|
|
||||||
// Read a CHM from a stream
|
|
||||||
|
|
||||||
private void readObject(java.io.ObjectInputStream s) throws IOException, ClassNotFoundException {
|
|
||||||
s.defaultReadObject(); // Read nothing
|
|
||||||
final int len = s.readInt(); // Read max element
|
|
||||||
_nbsi = new NBSI(len, new Counter(), this);
|
|
||||||
for (int i = 0; i < len; i++) // Read all bits
|
|
||||||
if (s.readBoolean())
|
|
||||||
_nbsi.add(i);
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- NBSI ----------------------------------------------------------------
|
|
||||||
|
|
||||||
private static final class NBSI {
|
|
||||||
// Back pointer to the parent wrapper; sorta like make the class non-static
|
|
||||||
private transient final NonBlockingSetInt _non_blocking_set_int;
|
|
||||||
|
|
||||||
// Used to count elements: a high-performance counter.
|
|
||||||
private transient final Counter _size;
|
|
||||||
|
|
||||||
// The Bits
|
|
||||||
private final long _bits[];
|
|
||||||
// --- Bits to allow Unsafe access to arrays
|
|
||||||
private static final int _Lbase = _unsafe.arrayBaseOffset(long[].class);
|
|
||||||
private static final int _Lscale = _unsafe.arrayIndexScale(long[].class);
|
|
||||||
|
|
||||||
private static long rawIndex(final long[] ary, final int idx) {
|
|
||||||
assert idx >= 0 && idx < ary.length;
|
|
||||||
return _Lbase + idx * _Lscale;
|
|
||||||
}
|
|
||||||
|
|
||||||
private final boolean CAS(int idx, long old, long nnn) {
|
|
||||||
return _unsafe.compareAndSwapLong(_bits, rawIndex(_bits, idx), old, nnn);
|
|
||||||
}
|
|
||||||
|
|
||||||
// --- Resize
|
|
||||||
// The New Table, only set once to non-zero during a resize.
|
|
||||||
// Must be atomically set.
|
|
||||||
private NBSI _new;
|
|
||||||
private static final long _new_offset;
|
|
||||||
|
|
||||||
static { // <clinit>
|
|
||||||
Field f = null;
|
|
||||||
try {
|
|
||||||
f = NBSI.class.getDeclaredField("_new");
|
|
||||||
} catch (java.lang.NoSuchFieldException e) {
|
|
||||||
}
|
|
||||||
_new_offset = _unsafe.objectFieldOffset(f);
|
|
||||||
}
|
|
||||||
|
|
||||||
private final boolean CAS_new(NBSI nnn) {
|
|
||||||
return _unsafe.compareAndSwapObject(this, _new_offset, null, nnn);
|
|
||||||
}
|
|
||||||
|
|
||||||
private transient final AtomicInteger _copyIdx; // Used to count bits started copying
|
|
||||||
private transient final AtomicInteger _copyDone; // Used to count words copied in a resize operation
|
|
||||||
private transient final int _sum_bits_length; // Sum of all nested _bits.lengths
|
|
||||||
|
|
||||||
private static final long mask(int i) {
|
|
||||||
return 1L << (i & 63);
|
|
||||||
}
|
|
||||||
|
|
||||||
// I need 1 free bit out of 64 to allow for resize. I do this by stealing
|
|
||||||
// the high order bit - but then I need to do something with adding element
|
|
||||||
// number 63 (and friends). I could use a mod63 function but it's more
|
|
||||||
// efficient to handle the mod-64 case as an exception.
|
|
||||||
//
|
|
||||||
// Every 64th bit is put in it's own recursive bitvector. If the low 6 bits
|
|
||||||
// are all set, we shift them off and recursively operate on the _nbsi64 set.
|
|
||||||
private final NBSI _nbsi64;
|
|
||||||
|
|
||||||
private NBSI(int max_elem, Counter ctr, NonBlockingSetInt nonb) {
|
|
||||||
super();
|
|
||||||
_non_blocking_set_int = nonb;
|
|
||||||
_size = ctr;
|
|
||||||
_copyIdx = ctr == null ? null : new AtomicInteger();
|
|
||||||
_copyDone = ctr == null ? null : new AtomicInteger();
|
|
||||||
// The main array of bits
|
|
||||||
_bits = new long[(int) (((long) max_elem + 63) >>> 6)];
|
|
||||||
// Every 64th bit is moved off to it's own subarray, so that the
|
|
||||||
// sign-bit is free for other purposes
|
|
||||||
_nbsi64 = ((max_elem + 1) >>> 6) == 0 ? null : new NBSI((max_elem + 1) >>> 6, null, null);
|
|
||||||
_sum_bits_length = _bits.length + (_nbsi64 == null ? 0 : _nbsi64._sum_bits_length);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Lower-case 'int' versions - no autoboxing, very fast.
|
|
||||||
// 'i' is known positive.
|
|
||||||
|
|
||||||
public boolean add(final int i) {
|
|
||||||
// Check for out-of-range for the current size bit vector.
|
|
||||||
// If so we need to grow the bit vector.
|
|
||||||
if ((i >> 6) >= _bits.length)
|
|
||||||
return install_larger_new_bits(i). // Install larger pile-o-bits (duh)
|
|
||||||
help_copy().add(i); // Finally, add to the new table
|
|
||||||
|
|
||||||
// Handle every 64th bit via using a nested array
|
|
||||||
NBSI nbsi = this; // The bit array being added into
|
|
||||||
int j = i; // The bit index being added
|
|
||||||
while ((j & 63) == 63) { // Bit 64? (low 6 bits are all set)
|
|
||||||
nbsi = nbsi._nbsi64; // Recurse
|
|
||||||
j = j >> 6; // Strip off low 6 bits (all set)
|
|
||||||
}
|
|
||||||
|
|
||||||
final long mask = mask(j);
|
|
||||||
long old;
|
|
||||||
do {
|
|
||||||
old = nbsi._bits[j >> 6]; // Read old bits
|
|
||||||
if (old < 0) // Not mutable?
|
|
||||||
// Not mutable: finish copy of word, and retry on copied word
|
|
||||||
return help_copy_impl(i).help_copy().add(i);
|
|
||||||
if ((old & mask) != 0) return false; // Bit is already set?
|
|
||||||
} while (!nbsi.CAS(j >> 6, old, old | mask));
|
|
||||||
_size.add(1);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
public boolean remove(final int i) {
|
|
||||||
if ((i >> 6) >= _bits.length) // Out of bounds? Not in this array!
|
|
||||||
return _new == null ? false : help_copy().remove(i);
|
|
||||||
|
|
||||||
// Handle every 64th bit via using a nested array
|
|
||||||
NBSI nbsi = this; // The bit array being added into
|
|
||||||
int j = i; // The bit index being added
|
|
||||||
while ((j & 63) == 63) { // Bit 64? (low 6 bits are all set)
|
|
||||||
nbsi = nbsi._nbsi64; // Recurse
|
|
||||||
j = j >> 6; // Strip off low 6 bits (all set)
|
|
||||||
}
|
|
||||||
|
|
||||||
final long mask = mask(j);
|
|
||||||
long old;
|
|
||||||
do {
|
|
||||||
old = nbsi._bits[j >> 6]; // Read old bits
|
|
||||||
if (old < 0) // Not mutable?
|
|
||||||
// Not mutable: finish copy of word, and retry on copied word
|
|
||||||
return help_copy_impl(i).help_copy().remove(i);
|
|
||||||
if ((old & mask) == 0) return false; // Bit is already clear?
|
|
||||||
} while (!nbsi.CAS(j >> 6, old, old & ~mask));
|
|
||||||
_size.add(-1);
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
public boolean contains(final int i) {
|
|
||||||
if ((i >> 6) >= _bits.length) // Out of bounds? Not in this array!
|
|
||||||
return _new == null ? false : help_copy().contains(i);
|
|
||||||
|
|
||||||
// Handle every 64th bit via using a nested array
|
|
||||||
NBSI nbsi = this; // The bit array being added into
|
|
||||||
int j = i; // The bit index being added
|
|
||||||
while ((j & 63) == 63) { // Bit 64? (low 6 bits are all set)
|
|
||||||
nbsi = nbsi._nbsi64; // Recurse
|
|
||||||
j = j >> 6; // Strip off low 6 bits (all set)
|
|
||||||
}
|
|
||||||
|
|
||||||
final long mask = mask(j);
|
|
||||||
long old = nbsi._bits[j >> 6]; // Read old bits
|
|
||||||
if (old < 0) // Not mutable?
|
|
||||||
// Not mutable: finish copy of word, and retry on copied word
|
|
||||||
return help_copy_impl(i).help_copy().contains(i);
|
|
||||||
// Yes mutable: test & return bit
|
|
||||||
return (old & mask) != 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
public int size() {
|
|
||||||
return (int) _size.get();
|
|
||||||
}
|
|
||||||
|
|
||||||
// Must grow the current array to hold an element of size i
|
|
||||||
|
|
||||||
private NBSI install_larger_new_bits(final int i) {
|
|
||||||
if (_new == null) {
|
|
||||||
// Grow by powers of 2, to avoid minor grow-by-1's.
|
|
||||||
// Note: must grow by exact powers-of-2 or the by-64-bit trick doesn't work right
|
|
||||||
int sz = (_bits.length << 6) << 1;
|
|
||||||
// CAS to install a new larger size. Did it work? Did it fail? We
|
|
||||||
// don't know and don't care. Only One can be installed, so if
|
|
||||||
// another thread installed a too-small size, we can't help it - we
|
|
||||||
// must simply install our new larger size as a nested-resize table.
|
|
||||||
CAS_new(new NBSI(sz, _size, _non_blocking_set_int));
|
|
||||||
}
|
|
||||||
// Return self for 'fluid' programming style
|
|
||||||
return this;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Help any top-level NBSI to copy until completed.
|
|
||||||
// Always return the _new version of *this* NBSI, in case we're nested.
|
|
||||||
|
|
||||||
private NBSI help_copy() {
|
|
||||||
// Pick some words to help with - but only help copy the top-level NBSI.
|
|
||||||
// Nested NBSI waits until the top is done before we start helping.
|
|
||||||
NBSI top_nbsi = _non_blocking_set_int._nbsi;
|
|
||||||
final int HELP = 8; // Tuning number: how much copy pain are we willing to inflict?
|
|
||||||
// We "help" by forcing individual bit indices to copy. However, bits
|
|
||||||
// come in lumps of 64 per word, so we just advance the bit counter by 64's.
|
|
||||||
int idx = top_nbsi._copyIdx.getAndAdd(64 * HELP);
|
|
||||||
for (int i = 0; i < HELP; i++) {
|
|
||||||
int j = idx + i * 64;
|
|
||||||
j %= (top_nbsi._bits.length << 6); // Limit, wrap to array size; means we retry indices
|
|
||||||
top_nbsi.help_copy_impl(j);
|
|
||||||
top_nbsi.help_copy_impl(j + 63); // Also force the nested-by-64 bit
|
|
||||||
}
|
|
||||||
|
|
||||||
// Top level guy ready to promote?
|
|
||||||
// Note: WE may not be the top-level guy!
|
|
||||||
if (top_nbsi._copyDone.get() == top_nbsi._sum_bits_length)
|
|
||||||
// One shot CAS to promote - it may fail since we are racing; others
|
|
||||||
// may promote as well
|
|
||||||
if (_non_blocking_set_int.CAS_nbsi(top_nbsi, top_nbsi._new)) {
|
|
||||||
//System.out.println("Promote at top level to size "+(_non_blocking_set_int._nbsi._bits.length<<6));
|
|
||||||
}
|
|
||||||
|
|
||||||
// Return the new bitvector for 'fluid' programming style
|
|
||||||
return _new;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Help copy this one word. State Machine.
|
|
||||||
// (1) If not "made immutable" in the old array, set the sign bit to make
|
|
||||||
// it immutable.
|
|
||||||
// (2) If non-zero in old array & zero in new, CAS new from 0 to copy-of-old
|
|
||||||
// (3) If non-zero in old array & non-zero in new, CAS old to zero
|
|
||||||
// (4) Zero in old, new is valid
|
|
||||||
// At this point, old should be immutable-zero & new has a copy of bits
|
|
||||||
|
|
||||||
private NBSI help_copy_impl(int i) {
|
|
||||||
// Handle every 64th bit via using a nested array
|
|
||||||
NBSI old = this; // The bit array being copied from
|
|
||||||
NBSI nnn = _new; // The bit array being copied to
|
|
||||||
if (nnn == null) return this; // Promoted already
|
|
||||||
int j = i; // The bit index being added
|
|
||||||
while ((j & 63) == 63) { // Bit 64? (low 6 bits are all set)
|
|
||||||
old = old._nbsi64; // Recurse
|
|
||||||
nnn = nnn._nbsi64; // Recurse
|
|
||||||
j = j >> 6; // Strip off low 6 bits (all set)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Transit from state 1: word is not immutable yet
|
|
||||||
// Immutable is in bit 63, the sign bit.
|
|
||||||
long bits = old._bits[j >> 6];
|
|
||||||
while (bits >= 0) { // Still in state (1)?
|
|
||||||
long oldbits = bits;
|
|
||||||
bits |= mask(63); // Target state of bits: sign-bit means immutable
|
|
||||||
if (old.CAS(j >> 6, oldbits, bits)) {
|
|
||||||
if (oldbits == 0) _copyDone.addAndGet(1);
|
|
||||||
break; // Success - old array word is now immutable
|
|
||||||
}
|
|
||||||
bits = old._bits[j >> 6]; // Retry if CAS failed
|
|
||||||
}
|
|
||||||
|
|
||||||
// Transit from state 2: non-zero in old and zero in new
|
|
||||||
if (bits != mask(63)) { // Non-zero in old?
|
|
||||||
long new_bits = nnn._bits[j >> 6];
|
|
||||||
if (new_bits == 0) { // New array is still zero
|
|
||||||
new_bits = bits & ~mask(63); // Desired new value: a mutable copy of bits
|
|
||||||
// One-shot CAS attempt, no loop, from 0 to non-zero.
|
|
||||||
// If it fails, somebody else did the copy for us
|
|
||||||
if (!nnn.CAS(j >> 6, 0, new_bits))
|
|
||||||
new_bits = nnn._bits[j >> 6]; // Since it failed, get the new value
|
|
||||||
assert new_bits != 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Transit from state 3: non-zero in old and non-zero in new
|
|
||||||
// One-shot CAS attempt, no loop, from non-zero to 0 (but immutable)
|
|
||||||
if (old.CAS(j >> 6, bits, mask(63)))
|
|
||||||
_copyDone.addAndGet(1); // One more word finished copying
|
|
||||||
}
|
|
||||||
|
|
||||||
// Now in state 4: zero (and immutable) in old
|
|
||||||
|
|
||||||
// Return the self bitvector for 'fluid' programming style
|
|
||||||
return this;
|
|
||||||
}
|
|
||||||
|
|
||||||
private void print(int d, String msg) {
|
|
||||||
for (int i = 0; i < d; i++)
|
|
||||||
System.out.print(" ");
|
|
||||||
System.out.println(msg);
|
|
||||||
}
|
|
||||||
|
|
||||||
private void print(int d) {
|
|
||||||
StringBuffer buf = new StringBuffer();
|
|
||||||
buf.append("NBSI - _bits.len=");
|
|
||||||
NBSI x = this;
|
|
||||||
while (x != null) {
|
|
||||||
buf.append(" " + x._bits.length);
|
|
||||||
x = x._nbsi64;
|
|
||||||
}
|
|
||||||
print(d, buf.toString());
|
|
||||||
|
|
||||||
x = this;
|
|
||||||
while (x != null) {
|
|
||||||
for (int i = 0; i < x._bits.length; i++)
|
|
||||||
System.out.print(Long.toHexString(x._bits[i]) + " ");
|
|
||||||
x = x._nbsi64;
|
|
||||||
System.out.println();
|
|
||||||
}
|
|
||||||
|
|
||||||
if (_copyIdx.get() != 0 || _copyDone.get() != 0)
|
|
||||||
print(d, "_copyIdx=" + _copyIdx.get() + " _copyDone=" + _copyDone.get() + " _words_to_cpy=" + _sum_bits_length);
|
|
||||||
if (_new != null) {
|
|
||||||
print(d, "__has_new - ");
|
|
||||||
_new.print(d + 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -1,53 +0,0 @@
|
||||||
/*
|
|
||||||
* Licensed to Elastic Search and Shay Banon under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. Elastic Search licenses this
|
|
||||||
* file to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
package org.elasticsearch.common.util.concurrent.highscalelib;
|
|
||||||
|
|
||||||
import sun.misc.Unsafe;
|
|
||||||
|
|
||||||
import java.lang.reflect.Field;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Simple class to obtain access to the {@link Unsafe} object. {@link Unsafe}
|
|
||||||
* is required to allow efficient CAS operations on arrays. Note that the
|
|
||||||
* versions in {@link java.util.concurrent.atomic}, such as {@link
|
|
||||||
* java.util.concurrent.atomic.AtomicLongArray}, require extra memory ordering
|
|
||||||
* guarantees which are generally not needed in these algorithms and are also
|
|
||||||
* expensive on most processors.
|
|
||||||
*/
|
|
||||||
class UtilUnsafe {
|
|
||||||
private UtilUnsafe() {
|
|
||||||
} // dummy private constructor
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Fetch the Unsafe. Use With Caution.
|
|
||||||
*/
|
|
||||||
public static Unsafe getUnsafe() {
|
|
||||||
// Not on bootclasspath
|
|
||||||
if (UtilUnsafe.class.getClassLoader() == null)
|
|
||||||
return Unsafe.getUnsafe();
|
|
||||||
try {
|
|
||||||
final Field fld = Unsafe.class.getDeclaredField("theUnsafe");
|
|
||||||
fld.setAccessible(true);
|
|
||||||
return (Unsafe) fld.get(UtilUnsafe.class);
|
|
||||||
} catch (Exception e) {
|
|
||||||
throw new RuntimeException("Could not obtain access to sun.misc.Unsafe", e);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -1,23 +0,0 @@
|
||||||
/*
|
|
||||||
* Licensed to Elastic Search and Shay Banon under one
|
|
||||||
* or more contributor license agreements. See the NOTICE file
|
|
||||||
* distributed with this work for additional information
|
|
||||||
* regarding copyright ownership. Elastic Search licenses this
|
|
||||||
* file to you under the Apache License, Version 2.0 (the
|
|
||||||
* "License"); you may not use this file except in compliance
|
|
||||||
* with the License. You may obtain a copy of the License at
|
|
||||||
*
|
|
||||||
* http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
*
|
|
||||||
* Unless required by applicable law or agreed to in writing,
|
|
||||||
* software distributed under the License is distributed on an
|
|
||||||
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
||||||
* KIND, either express or implied. See the License for the
|
|
||||||
* specific language governing permissions and limitations
|
|
||||||
* under the License.
|
|
||||||
*/
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A copy of Cliff High-Scale-Lib version 1.1.2.
|
|
||||||
*/
|
|
||||||
package org.elasticsearch.common.util.concurrent.highscalelib;
|
|
Loading…
Reference in New Issue