Tag: DoubleAccumulators


Spark Accumulators are shared variables which are only “added” through an associative and commutative operation and are used to perform counters (Similar to Map-reduce counters) or sum operations Spark by default supports to create an accumulators of any numeric type Read more…