Using scipy, I'd like to get a measure of how likely it is that a random variable was generated by my log-normal distribution.
To do this I've considered looking at how far it is from the maximum of the PDF.
My approach so far is this: If the variable is r = 1.5
, and the distribution σ=0.5, find the value from the PDF, lognorm.pdf(r, 0.5, loc=0)
. Given the result, (0.38286..
), I would then like to look up what area of the PDF is below 0.38286..
.
How can this last step be implemented? Is this even the right way to approach this problem?
To give a more general example of the problem. Say someone tells me they have 126 followers on twitter. I know that Twitter followers are a log-normal distribution, and I have the PDF of that distribution. Given that distribution do I determine how believable this number of followers is?