I was comparing two algorithms computing the average of random numbers.
- First algorithm sums all numbers and divides by the items count in the end
- Second algorithm computes the average on every iteration and reuses the result when new data is received
I suppose there's nothing revolutionary here, and I'm not a mathematician so I can't put a name on those two algorithms.
Here is my code:
#include <iostream>
#include <iomanip>
#include <cstdlib>
class Average1
{
public:
Average1() : total( 0 ), count( 0 ) {}
void add( double value )
{
total += value;
count++;
}
double average()
{
return total/count;
}
private:
double total;
size_t count;
};
class Average2
{
public:
Average2() : av( 0 ), count( 0 ) {}
void add( double value )
{
av = (av*count + value)/(count+1);
count++;
}
double average()
{
return av;
}
private:
double av;
size_t count;
};
void compare()
{
Average1 av1;
Average2 av2;
double temp;
for ( size_t i = 0; i != 100000000; ++i )
{
temp = static_cast<double>(std::rand()) / static_cast<double>(RAND_MAX);
av1.add( temp );
av2.add( temp );
}
std::cout << std::setprecision(20) << av1.average() << std::endl;
std::cout << std::setprecision(20) << av2.average() << std::endl;
}
int main()
{
compare();
return 0;
}
Output is:
0.50001084285722707801
0.50001084285744978875
The difference is certainly due to double
type precision.
In the end, which one is the good method? Which one gives the real mathematical average (or closest to...)?