Instatistics, theJarque–Bera test is agoodness-of-fit test of whether sample data have theskewness andkurtosis matching anormal distribution. The test is named afterCarlos Jarque andAnil K. Bera.The test statistic is always nonnegative. If it is far from zero, it signals the data does not have a normal distribution.
Thetest statisticJB is defined as
wheren is the number of observations (or degrees of freedom in general);S is the sampleskewness,K is the samplekurtosis :
where and are the estimates of third and fourthcentral moments, respectively, is the samplemean, and is the estimate of the second central moment, thevariance.
If the data comes from a normal distribution, theJB statisticasymptotically has achi-squared distribution with twodegrees of freedom, so the statistic can be used totest the hypothesis that the data are from anormal distribution. Thenull hypothesis is a joint hypothesis of the skewness being zero and theexcess kurtosis being zero. Samples from a normal distribution have an expected skewness of 0 and an expected excess kurtosis of 0 (which is the same as a kurtosis of 3). As the definition ofJB shows, any deviation from this increases the JB statistic.
For small samples the chi-squared approximation is overly sensitive, often rejecting the null hypothesis when it is true. Furthermore, the distribution ofp-values departs from auniform distribution and becomes aright-skewedunimodal distribution, especially for smallp-values. This leads to a largeType I error rate. The table below shows somep-values approximated by a chi-squared distribution that differ from their true alpha levels for small samples.
| True α level | 20 | 30 | 50 | 70 | 100 |
|---|---|---|---|---|---|
| 0.1 | 0.307 | 0.252 | 0.201 | 0.183 | 0.1560 |
| 0.05 | 0.1461 | 0.109 | 0.079 | 0.067 | 0.062 |
| 0.025 | 0.051 | 0.0303 | 0.020 | 0.016 | 0.0168 |
| 0.01 | 0.0064 | 0.0033 | 0.0015 | 0.0012 | 0.002 |
(These values have been approximated usingMonte Carlo simulation inMatlab)
InMATLAB's implementation, the chi-squared approximation for the JB statistic's distribution is only used for large sample sizes (> 2000). For smaller samples, it uses a table derived fromMonte Carlo simulations in order to interpolatep-values.[1]
The statistic was derived by Carlos M. Jarque and Anil K. Bera while working on their Ph.D. Thesis at the Australian National University.
According to Robert Hall, David Lilien, et al. (1995) when using this test along with multiple regression analysis the right estimate is:
wheren is the number of observations andk is the number of regressors when examining residuals to an equation.