where n is the number of observations (or degrees of freedom in general); S is the sample skewness, K is the sample kurtosis :
where and are the estimates of third and fourth central moments, respectively, is the sample mean, and
is the estimate of the second central moment, the variance.
If the data comes from a normal distribution, the JB statistic asymptotically has a chi-squared distribution with two degrees of freedom, so the statistic can be used to test the hypothesis that the data are from a normal distribution. The null hypothesis is a joint hypothesis of the skewness being zero and the excess kurtosis being zero. Samples from a normal distribution have an expected skewness of 0 and an expected excess kurtosis of 0 (which is the same as a kurtosis of 3). As the definition of JB shows, any deviation from this increases the JB statistic.
For small samples the chi-squared approximation is overly sensitive, often rejecting the null hypothesis when it is true. Furthermore, the distribution of p-values departs from a uniform distribution and becomes a right-skewedunimodal distribution, especially for small p-values. This leads to a large Type I error rate. The table below shows some p-values approximated by a chi-squared distribution that differ from their true alpha levels for small samples.
Calculated p-values equivalents to true alpha levels at given sample sizes
In MATLAB's implementation, the chi-squared approximation for the JB statistic's distribution is only used for large sample sizes (> 2000). For smaller samples, it uses a table derived from Monte Carlo simulations in order to interpolate p-values.[1]
History
The statistic was derived by Carlos M. Jarque and Anil K. Bera while working on their Ph.D. Thesis at the Australian National University.
Jarque–Bera test in regression analysis
According to Robert Hall, David Lilien, et al. (1995) when using this test along with multiple regression analysis the right estimate is:
where n is the number of observations and k is the number of regressors when examining residuals to an equation.
Implementations
ALGLIB includes an implementation of the Jarque–Bera test in C++, C#, Delphi, Visual Basic, etc.
gretl includes an implementation of the Jarque–Bera test
Julia includes an implementation of the Jarque-Bera test JarqueBeraTest in the HypothesisTests package.[2]
MATLAB includes an implementation of the Jarque–Bera test, the function "jbtest".
Python statsmodels includes an implementation of the Jarque–Bera test, "statsmodels.stats.stattools.py".
R includes implementations of the Jarque–Bera test: jarque.bera.test in the package tseries,[3] for example, and jarque.test in the package moments.[4]
Wolfram includes a built in function called, JarqueBeraALMTest[5] and is not limited to testing against a Gaussian distribution.
Jarque, Carlos M.; Bera, Anil K. (1980). "Efficient tests for normality, homoscedasticity and serial independence of regression residuals". Economics Letters. 6 (3): 255–259. doi:10.1016/0165-1765(80)90024-5.
Jarque, Carlos M.; Bera, Anil K. (1981). "Efficient tests for normality, homoscedasticity and serial independence of regression residuals: Monte Carlo evidence". Economics Letters. 7 (4): 313–318. doi:10.1016/0165-1765(81)90035-5.
Jarque, Carlos M.; Bera, Anil K. (1987). "A test for normality of observations and regression residuals". International Statistical Review. 55 (2): 163–172. doi:10.2307/1403192. JSTOR1403192.
Judge; et al. (1988). Introduction and the theory and practice of econometrics (3rd ed.). pp. 890–892.
Hall, Robert E.; Lilien, David M.; et al. (1995). EViews User Guide. p. 141.