Confusino regarding statistical tests for given distribution
4 vues (au cours des 30 derniers jours)
Afficher commentaires plus anciens
I am attempting to describe some data I have with a distribution but am experiencing some strange behavior.
I have a dataset which I would describe as "obviously normal", see the probplot and histfit for that
when running some tests I can confirm it is normal as well:
>> lillietest(data)
ans =
0
>> pd = fitdist(data, 'normal'); kstest(data, 'cdf', pd)
ans =
logical
0
>> pd = fitdist(data, 'normal'); chi2gof(data, 'cdf', pd)
ans =
1
I am just having a hard time understanding why this is failing the chi square goodness of fit test. I have >7000 data points and I simply cannot see how it can't be Gaussian.
Attached is the data.
0 commentaires
Réponses (2)
John D'Errico
le 8 Juin 2021
Modifié(e) : John D'Errico
le 8 Juin 2021
It looks FAIRLY normal. But you have a whole crapload of data. It needs to look more normal than that. When you have a lot of data, it had better be darned tootin normal. Said differently, what those tests did not tell you is how badly does it fail?
[H,P] = chi2gof(data, 'cdf', pd)
H =
1
P =
0.032397
The default tolerance for the test was probably 0.05, if I had to guess. So your data was just over the line. By the way, if I look at that probplot, it has a few squiggles, that apparently were just enough to push it over the line.
0 commentaires
dpb
le 8 Juin 2021
Modifié(e) : dpb
le 8 Juin 2021
You need to examine what the default parameters of the function return --
>> [h,p,stats]=chi2gof(data)
h =
1.00
p =
0.03
stats =
struct with fields:
chi2stat: 13.76
df: 6.00
edges: [-0.47 -0.28 -0.19 -0.10 -0.01 0.08 0.17 0.26 0.36 0.45]
O: [50.00 211.00 833.00 1856.00 2133.00 1461.00 554.00 89.00 13.00]
E: [35.88 225.21 855.97 1818.58 2162.62 1439.93 536.40 111.61 13.80]
>>
NB: there are only 6 DOF in the output statistic --
However, it does show what the normality plot shows to some degree, the LH tail is "heavy" with more observations at the lower extreme than expected. With the coarse binning, this is enough to reject the hypothesis at the default level of significance. (And, correspondingly, it is a little light on the RH end).
>> [h,p,stats]=chi2gof(data,"NBins",40)
h =
0
p =
0.57
stats =
struct with fields:
chi2stat: 29.04
df: 31.00
edges: [1×35 double]
O: [1×34 double]
E: [1×34 double]
>>
Let's look at how well the guess worked; one should have at least 5 observations in a bin(*)
>> [min(stats.O), max(stats.O);min(stats.E), max(stats.E)]
ans =
5.00 550.00
5.18 559.11
>>
As the NIST handbook notes, one of the weaknesses of the Chi-Square is there is no optimal binning algorithm so the results can be sensitive to the choice made.
John d' has a very valid point that lots of data means can reject more easily...depending on the use of the data, it's probably such that these deviations from normality will not be very significant unless, of course, you're doing something like estimating from the tails in which a normal approximation will likely underestimate/overestimate the observed data frequency somewhat in the left/right tails, respectively.
(*) My memory refresher is always the NIST handbook; what it says about Chi-Square GOF is at https://www.itl.nist.gov/div898/handbook/eda/section3/eda35f.htm
I tend to rely upon the Shapiro-Wilk test which I don't believe TMW has implemented; I've a homebrew version I coded 40 years ago...
0 commentaires
Voir également
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!