Part ofa series on |
Probabilistic data structures |
---|
Random trees |
Related |
HyperLogLog is an algorithm for thecount-distinct problem, approximating the number of distinct elements in amultiset.[1] Calculating theexactcardinality of the distinct elements of a multiset requires an amount of memory proportional to the cardinality, which is impractical for very large data sets. Probabilistic cardinality estimators, such as the HyperLogLog algorithm, use significantly less memory than this, but can only approximate the cardinality. The HyperLogLog algorithm is able to estimate cardinalities of > 109 with a typical accuracy (standard error) of 2%, using 1.5 kB of memory.[1] HyperLogLog is an extension of the earlier LogLog algorithm,[2] itself deriving from the 1984Flajolet–Martin algorithm.[3]
In the original paper by Flajoletet al.[1] and in related literature on thecount-distinct problem, the term "cardinality" is used to mean the number of distinct elements in a data stream with repeated elements. However in the theory ofmultisets the term refers to the sum of multiplicities of each member of a multiset. This article chooses to use Flajolet's definition for consistency with the sources.
![]() | This section includes a list ofgeneral references, butit lacks sufficient correspondinginline citations. Please help toimprove this section byintroducing more precise citations.(March 2014) (Learn how and when to remove this message) |
The basis of the HyperLogLog algorithm is the observation that the cardinality of a multiset of uniformly distributed random numbers can be estimated by calculating the maximum number of leading zeros in the binary representation of each number in the set. If the maximum number of leading zeros observed is n, an estimate for the number of distinct elements in the set is 2n.[1]
In the HyperLogLog algorithm, ahash function is applied to each element in the original multiset to obtain a multiset of uniformly distributed random numbers with the same cardinality as the original multiset. The cardinality of this randomly distributed set can then be estimated using the algorithm above.
The simple estimate of cardinality obtained using the algorithm above has the disadvantage of a largevariance. In the HyperLogLog algorithm, the variance is minimised by splitting the multiset into numerous subsets, calculating the maximum number of leading zeros in the numbers in each of these subsets, and using aharmonic mean to combine these estimates for each subset into an estimate of the cardinality of the whole set.[4]
The HyperLogLog has three main operations:add to add a new element to the set,count to obtain the cardinality of the set andmerge to obtain the union of two sets. Some derived operations can be computed using theinclusion–exclusion principle like thecardinality of the intersection or thecardinality of the difference between two HyperLogLogs combining the merge and count operations.
The data of the HyperLogLog is stored in an arrayM ofm counters (or "registers") that are initialized to 0. ArrayM initialized from a multisetS is calledHyperLogLogsketch of S.
The add operation consists of computing the hash of the input datav with a hash functionh, getting the firstb bits (whereb is), and adding 1 to them to obtain the address of the register to modify. With the remaining bits compute which returns the position of the leftmost 1, where leftmost position is 1 (in other words: number of leading zeros plus 1). The new value of the register will be the maximum between the current value of the register and.
The count algorithm consists in computing the harmonic mean of them registers, and using a constant to derive an estimate of the count:
The intuition is thatn being the unknown cardinality ofM, each subset will have elements. Then should be close to. The harmonic mean of 2 to these quantities is which should be near. Thus, should ben approximately.
Finally, the constant is introduced to correct a systematic multiplicative bias present in due to hash collisions.
The constant is not simple to calculate, and can be approximated with the formula[1]
The HyperLogLog technique, though, is biased for small cardinalities below a threshold of. The original paper proposes using a different algorithm for small cardinalities known as Linear Counting.[5] In the case where the estimate provided above is less than the threshold, the alternative calculation can be used:
Additionally, for very large cardinalities approaching the limit of the size of the registers ( for 32-bit registers), the cardinality can be estimated with:
With the above corrections for lower and upper bounds, the error can be estimated as.
The merge operation for two HLLs () consists in obtaining the maximum for each pair of registers
To analyze the complexity, the data streaming model[6] is used, which analyzes the space necessary to get a approximation with a fixed success probability. The relative error of HLL is and it needs space, wheren is the set cardinality andm is the number of registers (usually less than one byte size).
Theadd operation depends on the size of the output of the hash function. As this size is fixed, we can consider the running time for the add operation to be.
Thecount andmerge operations depend on the number of registersm and have a theoretical cost of. In some implementations (Redis)[7] the number of registers is fixed and the cost is considered to be in the documentation.
The HyperLogLog++ algorithm proposes several improvements in the HyperLogLog algorithm to reduce memory requirements and increase accuracy in some ranges of cardinalities:[6]
When the data arrives in a single stream, the Historic Inverse Probability or martingale estimator[8][9]significantly improves the accuracy of the HLL sketch and uses 36% less memory to achieve a given error level. This estimator is provably optimal for any duplicate insensitive approximate distinct counting sketch on a single stream.
The single stream scenario also leads to variants in the HLL sketch construction.HLL-TailCut+ uses 45% less memory than the original HLL sketch but at the cost of being dependent on the data insertion order and not being able to merge sketches.[10]