Diagnostic Test
Diagnostic Test
Diagnostic Test
Normality Test
The test statistic in a regression model should follow a probability distribution that
is easy to handle, such as the normal distribution. Normality can be assessed by
examining the residuals to determine if they follow a normal distribution (Jones,
2020).
The Shapiro-Wilk test for normality generates a p-value that determines whether
the data follows a normal distribution. All predictors and dependent variables are
tested using the Shapiro-Wilk test. If the p-value is greater than 0.05, the data is
considered to be normally distributed (Shapiro & Wilk, 1965). The table
demonstrates that all variables had p-values exceeding 0.05, indicating normal
distribution for each variable. Consequently, further statistical analyses can be
conducted on the data.
Shaphiro-wilk Test
Test of multicollinearity
Collinearity Statistics
Heteroscadacity
Heteroscedasticity occurs when the variance of errors changes across data, leading
to an inefficient yet unbiased OLS estimator (Long & Ervin, 2000). To assess
heteroscedasticity, the researchers utilized the Breusch-Pagan and Koenker tests,
which test the null hypothesis that the variances of the error terms are constant. If
the significant result is less than 5%, the null hypothesis is rejected. The findings
of the Breusch-Pagan and Koenker tests are presented in the table below, showing
that heteroscedasticity was not an issue, as all significant values were greater than
5%.
Tests LM Sig
BP 1.089 .580
Koenker 1.211 .546
Autocorrelation
Durbin-Watson results
The linearity test assesses whether a significant linear relationship exists between
two or more variables. In linear regression, the assumption is that there is a linear
relationship between the independent variable x and the dependent variable y. To
verify this assumption, a scatter plot of x versus y is created. This visualization
helps determine if the data points align along a straight line, indicating a linear
relationship between the variables and confirming that the assumption is met (Kusi
et al., 2014).
The results from these tests are used to guide decisions about selecting the
appropriate regression model. If the scatter plot of x and y suggests a lack of a
linear relationship, possible adjustments include applying a nonlinear
transformation to the variables. Common transformations involve taking the
logarithm, square root, or reciprocal of the variables. Alternatively, adding another
independent variable to the model might be beneficial. For instance, if the scatter
plot shows a parabolic pattern, including x2 as an additional independent variable
could improve the model (Kamar, 2019).
Linearity testing can be performed using the significance values for linearity and
deviation from linearity provided in the ANOVA table (Ngechu, 2020). The Sig.
linearity value indicates how well the independent variable follows a straight line.
If Sig. linearity is less than the significance level (α), this suggests that linear
regression is appropriate for explaining the relationship between the variables.
Conversely, Sig. deviation from linearity assesses whether the data can be
considered linear. If Sig. deviation from linearity is greater than the significance
level (α), linear regression remains suitable for analyzing the variables (Smith,
2019). Testing for linearity determines if there is a mathematical relationship of the
form y=cx, where 𝑐c is a constant. This is important because many statistical
methods assume linearity of the data (Cooper & Schindler, 2003).