mirror of https://github.com/apache/druid.git
try to make javadoc happy
This commit is contained in:
parent
7cfe86b7bc
commit
77c9afe92f
|
@ -1019,8 +1019,6 @@ public class ApproximateHistogram
|
|||
* @param count current size of the heap
|
||||
* @param heapIndex index of the item to be deleted
|
||||
* @param values values stored in the heap
|
||||
*
|
||||
* @return
|
||||
*/
|
||||
private static int heapDelete(int[] heap, int[] reverseIndex, int count, int heapIndex, float[] values)
|
||||
{
|
||||
|
|
|
@ -200,9 +200,6 @@ public abstract class HyperLogLogCollector implements Comparable<HyperLogLogColl
|
|||
/**
|
||||
* Checks if the payload for the given ByteBuffer is sparse or not.
|
||||
* The given buffer must be positioned at getPayloadBytePosition() prior to calling isSparse
|
||||
*
|
||||
* @param buffer
|
||||
* @return
|
||||
*/
|
||||
private static boolean isSparse(ByteBuffer buffer)
|
||||
{
|
||||
|
@ -636,8 +633,6 @@ public abstract class HyperLogLogCollector implements Comparable<HyperLogLogColl
|
|||
* @param position The position into the byte buffer, this position represents two "registers"
|
||||
* @param offsetDiff The difference in offset between the byteToAdd and the current HyperLogLogCollector
|
||||
* @param byteToAdd The byte to merge into the current HyperLogLogCollector
|
||||
*
|
||||
* @return
|
||||
*/
|
||||
private static int mergeAndStoreByteRegister(
|
||||
final ByteBuffer storageBuffer,
|
||||
|
|
|
@ -49,8 +49,6 @@ public interface HavingSpec
|
|||
* @param row A Row of data that may contain aggregated values
|
||||
*
|
||||
* @return true if the given row satisfies the having spec. False otherwise.
|
||||
*
|
||||
* @see Row
|
||||
*/
|
||||
public boolean eval(Row row);
|
||||
|
||||
|
|
|
@ -173,8 +173,6 @@ public class CompressedFloatsIndexedSupplier implements Supplier<IndexedFloats>
|
|||
|
||||
/**
|
||||
* For testing. Do not depend on unless you like things breaking.
|
||||
*
|
||||
* @return
|
||||
*/
|
||||
GenericIndexed<ResourceHolder<FloatBuffer>> getBaseFloatBuffers()
|
||||
{
|
||||
|
|
|
@ -184,7 +184,6 @@ public class CompressedLongsIndexedSupplier implements Supplier<IndexedLongs>
|
|||
|
||||
/**
|
||||
* For testing. Do not use unless you like things breaking
|
||||
* @return
|
||||
*/
|
||||
GenericIndexed<ResourceHolder<LongBuffer>> getBaseLongBuffers()
|
||||
{
|
||||
|
|
|
@ -673,10 +673,9 @@ public class IncrementalIndex implements Iterable<Row>
|
|||
falseIdsReverse = biMap.inverse();
|
||||
}
|
||||
|
||||
/**
|
||||
* Returns the interned String value to allow fast comparisons using `==` instead of `.equals()`
|
||||
* @see io.druid.segment.incremental.IncrementalIndexStorageAdapter.EntryHolderValueMatcherFactory#makeValueMatcher(String, String)
|
||||
*/
|
||||
|
||||
// Returns the interned String value to allow fast comparisons using `==` instead of `.equals()`
|
||||
// see io.druid.segment.incremental.IncrementalIndexStorageAdapter.EntryHolderValueMatcherFactory#makeValueMatcher(String, String)
|
||||
public String get(String value)
|
||||
{
|
||||
return value == null ? null : poorMansInterning.get(value);
|
||||
|
|
|
@ -532,10 +532,8 @@ public class IncrementalIndexStorageAdapter implements StorageAdapter
|
|||
}
|
||||
|
||||
for (String dimVal : dims[dimIndex]) {
|
||||
/**
|
||||
* using == here instead of .equals() to speed up lookups made possible by
|
||||
* {@link io.druid.segment.incremental.IncrementalIndex.DimDim#poorMansInterning}
|
||||
*/
|
||||
// using == here instead of .equals() to speed up lookups made possible by
|
||||
// io.druid.segment.incremental.IncrementalIndex.DimDim#poorMansInterning
|
||||
if (id == dimVal) {
|
||||
return true;
|
||||
}
|
||||
|
|
|
@ -169,7 +169,7 @@ public class ZkCoordinator extends BaseZkCoordinator
|
|||
catch (IOException e) {
|
||||
throw new SegmentLoadingException(e, "Failed to announce segment[%s]", segment.getIdentifier());
|
||||
}
|
||||
};
|
||||
}
|
||||
}
|
||||
catch (SegmentLoadingException e) {
|
||||
log.makeAlert(e, "Failed to load segment for dataSource")
|
||||
|
|
Loading…
Reference in New Issue