Basic Statistical Techniques in Research
2009
1
2
F.A. Adesoji & M.A. Babatunde
Basic Statistical Techniques in Research
1
CHAPTER 1
BASIC STATISTICAL TECHNIQUES IN
RESEARCH
F.A. Adesoji and M.A. Babatunde
Introduction
The essence of research is to solve problem(s). In order to do this,
we have to select a plan for this study which is being referred to as
“design of the experiment”. In this case, research can be regarded
as the application of the scientific method to the study of a
problem. Question/hypotheses are thought of in an attempt to find
solution to the problem at hand. In order to gather information or
data, instruments must be in place. The data collected are not
useful because they are referred to as raw data or untreated data
until they are analyzed by using appropriate statistical tools.
The design of experiment is inseparable from the statistical
treatment of the results. If the design of an experiment is faulty, no
amount of statistical manipulation can lead to the drawing of valid
inference. Experimental design and statistical procedures are two
sides of the same coin. Any research that deals with the
manipulation of variables which are basically of two types. These
are, numerical and categorical. Numerical variables are recorded as
numbers such as height, age, scores, weight, etc. Categorical
variables could be dichotomy (for example, male or female),
trichotomy (for example, high, medium and low economic status)
or polychotomy (for example, birth places). Statistical techniques
have to do with data generation, manipulation and interpretation.
In order to generate data, measurement is necessary. Measurement
is the ascribing of symbols or figures to entities and it is thus basic
2
F.A. Adesoji & M.A. Babatunde
to data analysis, interpretation and research in general. This is done
with the aid of well validated instruments.
The study of data is called STATISTICS and in a more refined
way, it deals with scientific and analytical methods of collecting,
organizing, analyzing and presenting data in such a way that some
meanings and conclusions could be made out of something that
appears to be jungle of data. Statistics can be very broadly
classified into two categories, viz, descriptive and inferential
statistics. Descriptive statistics refers to the type of statistics, which
deal with collection, organizing, summarizing and describing
quantitative data. For example, the average score describe the
performance of the class but does not make any generalization
about other classes. Examples of descriptive statistics are graphs,
charts (pie charts, columinal charts, bar charts, histogram, etc),
Pictograms, tables and any form whereby data are displayed for
easier understanding. Other examples are measures of central
tendency (mode, mean, median), correlation coefficient (degree of
relationship), kurtosis, skewedness etc.
Inferential statistics deals with the methods by which
inferences are made on the population on the basis of the
observations made on the smaller sample. For example, a
researcher may want to estimate the average score of two or more
classes of an English course by making use of the average score of
one class. Any procedure of making generalization that goes
beyond the original data is called inferential statistics. The
statistics provide a way of testing the significance of results
obtained when data are collected. It thus uses probability, that is,
the chance of an event occurring. Examples of inferential statistical
tools are student t-test, Analysis of Variance, Analysis of
Covariance,
Correlation
Analysis,
Multiple
regression,
Multivariate Analysis of Variance etc.
An area of inferential statistics called hypothesis testing is a
decision making process for evaluating claims about a population
based on information obtained from samples. Relationship among
variables can also be determined. Also, by studying past and
Basic Statistical Techniques in Research
3
present, data and conditions; it is also possible to make predictions
based on this information. It would be observed that descriptive
statistics consists of the collection, organization, summarization,
and presentation of data while inferential statistics on the other
hand, consists of generalizing from samples to populations,
performing estimation and hypothesis testing, determining
relationships among variables, and making predictions.
Nature of Variables and Types of Data
Variables can be classified as qualitative or quantitative.
Qualitative variables are variables that can be placed into distinct
categories, according to some characteristics or attributes. For
example, categorization according to gender (male or female) then,
variable gender is qualitative and it takes categorical data, let us
say 1 or 2. Other examples are religious preference, ability level
and geographical locations. Quantitative variables are numerical
and can be ordered or ranked. For example, the variable age is
numerical and people can be ranked in order according to the value
of their ages. Other examples are heights, weights and body
temperatures.
Quantitative can be further classified into two groups: discrete
and continuous. Discrete variables can be assigned values such as
0,1,2,3 and are said to be countable. Examples of discrete variables
are the number of children in a family, the number of student in the
classroom etc. Thus, discrete variables assume values that can be
counted. Continuous variables, by comparison can assume all
value in an interval between any two specific values. Temperature
is a continuous variable since it can assume all values between any
two given temperatures. Data could also be categorized as purely
numerical and not purely numerical. Mean cannot be determined in
a not purely numerical data. For example, data involving the
number of people and their mode of collecting information.
4
F.A. Adesoji & M.A. Babatunde
Qualitative
Quantitative
Data
discrete
continuous
The mean can not be determined in this case compare with the
scores of group of students in a course of study. Data can also
assume nominal level (assigning A, B, C), or ordinal (ordered or
ranked), interval (precise difference exist) or ratio.
For example, scores of 50% and 51%- a meaningful one
point difference exists, but there is no true zero point. For example,
0oF does not mean no heat at all, and the ratio- here, in addition to
the difference between units, there is a time zero point and true
ratio between values. Note that there is no complete agreement
among statisticians about the classification of data into one of the
four categories. Also, data can be altered so that they fit into a
different category. For example, if you categorize income of
workers into low, average and high, then a ratio variable becomes
an ordinal variable.
Parametric and Non-Parametric Tests
The distribution of many test statistics can be said to be normal or
follows some form that can be derived from the normal
distribution. A characteristic property of the normal distribution is
that 68% of all of its observations fall within a range of ±1
standard deviation from the mean, and a range of ±2 standard
deviations includes 95% of the scores. In other words, in a normal
distribution, observations that have a standardized value of less
Basic Statistical Techniques in Research
5
than -2 or more than +2 have a relative frequency of 5% or less.1
The exact shape of the normal distribution (the characteristic "bell
curve") is defined by a function which has only two parameters:
mean and standard deviation.
Figure 1: Probability Density Function. The left hand side of
the graph represents a Standard Normal Distribution Function
The attempt to choose the right test to compare measurements
may however be a bit difficult, since we must choose between two
families of tests: parametric and nonparametric. Many statistical
tests are based upon the assumption that the data are sampled from
a normal distribution. These tests are referred to as parametric
tests. Parametric statistics are statistics where the population is
assumed to fit any parametrized distributions (that is, most
typically the normal distribution).2 Commonly used parametric
1
Standardized value means that a value is expressed in terms of its difference
from the mean, divided by the standard deviation.
2
The normal distribution, also called the Gaussian distribution, is an important
family of continuous probability distributions, applicable in many fields. Each
member of the family may be defined by two parameters, the mean ("average",
μ) and variance (standard deviation squared) σ2, respectively. The standard
normal distribution is the normal distribution with a mean of zero and a variance
of one.
6
F.A. Adesoji & M.A. Babatunde
tests include the Mean, Standard deviation, the t-test (One-sample t
test, Unpaired t test, Paired t test), One-way ANOVA, Pearson
product moment correlation, Simple linear regression, Multiple
linear regression. For example, Analysis of Variance (ANOVA)
assumes that the underlying distributions are normally distributed
and that the variances of the distributions being compared are
similar. The Pearson product-moment correlation coefficient also
assumes normality.
Although parametric techniques are robust, that is, they
often retain considerable power to detect differences or similarities
even when these assumptions are violated, some distributions
violate the assumptions so markedly that a non-parametric
alternative is more likely to detect a difference or similarity.3
Hence, tests that do not make assumptions about the population
distribution are referred to as nonparametric-tests.
Specifically, nonparametric methods were developed to be
used in cases when the researcher knows nothing about the
parameters of the variable of interest in the population (hence the
name nonparametric). In more technical terms, nonparametric
methods do not rely on the estimation of parameters (such as the
mean or the standard deviation) describing the distribution of the
variable of interest in the population. Therefore, these methods are
also sometimes (and more appropriately) called parameter-free
methods or distribution-free methods. All commonly used
nonparametric tests rank the outcome variable from low to high
and then analyze the ranks.
For many variables of interest, we simply do not know for
sure that they are normally distributed. For example, we can not
state categorically that incomes of Universities staff in Nigeria are
distributed normally in the population. In addition, the incidence
rates of rare diseases are not normally distributed in the population,
the number of car accidents is also not normally distributed, and
neither are very many other variables in which a researcher might
3
The power of a statistical test is the probability that the test will reject a false
null hypothesis.
Basic Statistical Techniques in Research
7
be interested. Examples of nonparametric tests include Median,
Interquartile range, Wilcoxon test, Mann-Whitney test, KruskalWallis test, Friedman test for dependent samples, Chi square test,
Spearman correlation, Coefficient of concordance.
Many observed variables actually are normally distributed,
which is why the normal distribution represents an empirical
discussion in the literature. The problem may occur when one tries
to use a normal distribution-based test to analyze data from
variables that are themselves not normally distributed. Another
factor that often limits the applicability of tests based on the
assumption that the sampling distribution is normal is the size of
the sample of data available for the analysis (sample size; n). We
can assume that the sampling distribution is normal even if we are
not sure that the distribution of the variable in the population is
normal, as long as our sample is large enough (for example, 100 or
more observations). However, if our sample is very small, then
those tests can be used only if we are sure that the variable is
normally distributed, and there is no way to test this assumption if
the sample is small.
Applications of tests that are based on the normality
assumptions are further limited by a lack of precise measurement.
For example, the assumption in most common statistical
techniques such as Analysis of Variance (and t- tests), regression is
that the underlying measurements are at least of interval, meaning
that equally spaced intervals on the scale can be compared in a
meaningful manner. However, this assumption is very often not
tenable because the data may represent a rank ordering of
observations (ordinal) rather than precise measurements. Choosing
between parametric and nonparametric tests is sometimes easy.
You should definitely choose a parametric test if you are sure that
your data are sampled from a population that follows a normal
distribution. Non-parametric tests could be selected under three
conditions:
8
F.A. Adesoji & M.A. Babatunde
The outcome is a rank or a score and the population is clearly
not Gaussian. Examples include class ranking of students.
Some values are off the scale, that is, too high or too low to
measure. Even if the population is Gaussian, it is
impossible to analyze such data with a parametric test since
you do not know all of the values. Using a nonparametric
test with these data is simple. Assign values too low to
measure an arbitrary very low value and assign values too
high to measure an arbitrary very high value. Then perform
a nonparametric test. Since the nonparametric test only
knows about the relative ranks of the values, it does not
matter that you did not know all the values exactly.
If the data are not sampled from normal distribution, we
can consider transforming the values to make the
distribution become normal. For example, you might take
the logarithm or reciprocal of all values.
Despite the three cases stated above, it is not always easy to decide
whether a sample comes from a normal distributed population. For
example:
If we collect many data points (maybe over a hundred or
so), we can look at the distribution of the data and it will be
fairly obvious whether the distribution is approximately
bell shaped. With few data points, it is difficult to tell
whether the data are Gaussian by inspection, and the formal
test has little power to discriminate between Gaussian and
non-Gaussian distributions.
We can look at previous data as well. What is important is
the distribution of the overall population, and not the
distribution of our sample. In deciding whether a
population is Gaussian, we should look at all available data,
not just data in the current experiment.
When in doubt, some people choose a parametric test (because
they are not sure the Gaussian assumption is violated), and others
choose a nonparametric test (because they are also not sure
Basic Statistical Techniques in Research
9
whether the Gaussian assumption is met). There are four cases to
consider in the answer whether one chooses between a parametric
or nonparametric test sample sizes:
1. The central limit theorem ensures that parametric tests
work well with large samples even if the population is nonGaussian. In other words, parametric tests are robust to
deviations from Gaussian distributions, so long as the
samples are large. The problem is that it is impossible to
say how large is large enough, as it depends on the nature
of the particular non-Gaussian distribution. Unless the
population distribution is quite distorted, you are probably
safe choosing a parametric test when there are at least two
dozen data points in each group.
2. Nonparametric tests work well with large samples from
Gaussian populations. The probability (P) values tend to be
a bit too large, but the discrepancy is small. In other words,
nonparametric tests are only slightly less powerful than
parametric tests with large samples.
3. In a small sample situation, we can not rely on the central
limit theorem if we use a parametric test with data from
non-Gaussian populations because the P value may be
inaccurate.
4. In addition, when we use a nonparametric test with data
from a Gaussian population, the P values also tend to be
quite high. The nonparametric tests lack statistical power
with small samples.
Thus, large data sets present no problems. It is usually easy to
tell if the data come from a Gaussian population, but it does not
really matter because the nonparametric tests are so powerful and
the parametric tests are so robust. However, small data sets present
a dilemma. It is difficult to tell if the data come from a Gaussian
population, but it matters a lot. The nonparametric tests are not
powerful and the parametric tests are not robust.
10
F.A. Adesoji & M.A. Babatunde
Basic Statistical Techniques
A multitude of different statistical tools is available, some of them
simple, some complicated, and often very specific for certain
purposes. In analytical work, the most important common
operation is the comparison of data, or sets of data, to quantify
accuracy (bias) and precision. The value of statistics lies with
organizing and simplifying data, to permit some objective estimate
showing that an analysis is under control or that a change has
occurred. Statistical techniques can be used to describe data,
compare two or more data sets, determine if a relationship exists
between variables, test hypotheses and make estimates about
population measures. Some well known statistical tests and
procedures for research observations are:
The t-test
The t-test is the most commonly used method to evaluate the
differences in means between two groups. For example, the t-test
can be used to test for a difference in test scores between a group
of patients who were given a drug and a control group who
received an injection. Theoretically, the t-test can be used even if
the sample sizes are very small (e.g., as small as 10) as long as the
variables are normally distributed within each group and the
variation of scores in the two groups is not reliably different.
The p-level reported with a t-test represents the probability of
error involved in accepting the research hypothesis about the
existence of a difference. It is the probability of error associated
with rejecting the hypothesis of no difference between the two
categories of observations (corresponding to the groups) in the
population when, in fact, the hypothesis is true. If the calculated pvalue is below the threshold chosen for statistical significance
(usually the 0.05 level), then the null hypothesis which usually
states that the two groups do not differ is rejected in favor of an
alternative hypothesis, which typically states that the groups do
differ.
Basic Statistical Techniques in Research
11
There are however specific assumptions underlying the use of the
t-test:
1. The sample data should be normally distributed.
2. The sample must be representative of the population so that
we can made generalizations at the end of the analysis.
3. Equality of variances. Equal variances are assumed when
two independent samples are used to test a hypothesis.
4. The dependent measurements involved in the calculation of
the means must come from either interval or ratio scales.
Since all calculations are carried out subject to the null
hypothesis, it may be very difficult to come up with a reasonable
null hypothesis that accounts for equal means in the presence of
unequal variances. Consequently, the null hypothesis is that the
different treatments have no effect which therefore makes unequal
variances untenable. A fundamental issue in the use of the t-test is
often whether the samples are independent or dependent.
Independent samples typically consist of two groups with no
relationship while dependent samples typically consist of a
matched sample (alternatively a paired sample) or one group that
has been tested twice (repeated measures).
Dependent t-tests are also used for matched-paired samples,
where two groups are matched on a particular variable. For
example, if we examined the heights of men and women in a
relationship, the two groups are matched on relationship status.
This would call for a dependent t-test because it is a paired sample
(one man paired with one woman). Alternatively, we might recruit
100 men and 100 women, with no relationship between any
particular man and any particular woman; in this case we would
use an independent samples test. Another example of a matched
sample would be to take two groups of students, match each
student in one group with a student in the other group based on a
continous assessment result, then examine how much each student
reads. An example pair might be two students that score 70 and 71
or two students that scored 55 and 50 on the same continous
assesment. The hypothesis would be that students that did well on
12
F.A. Adesoji & M.A. Babatunde
the test may or may not read more. Alternatively, we might recruit
students with low scores and students with high scores in two
groups and assess their reading amounts independently.
An example of a repeated measures t-test would be if one
group were pre- and post-tested. For example, if a teacher wanted
to examine the effect of a new set of textbooks on student
achievement, he/she could test the class at the beginning of the
year (pre-test) and at the end of the year (post-test). A dependent ttest would be used, treating the pre-test and post-test as matched
variables (matched by student).
Table 1: T-Test for Comparison of Mean Score of LPT and CT
groups
Treatment
N
SD
t-value
Sig.(t)
Groups
x
LPT
120
11.69
3.19
7.92
0.000
CT
120
10.27
3.37
In all the testing of this nature, the probability value is very
important when interpreting the print-out. If the probability is less
than 0.05 (the probaibility for committing type 1 error), the result
is significant and we reject the null hypothesis. The reverse is the
case if the probability is greater than 0.05.
ANOVA/ANCOVA /MCA
The purpose of analysis of variance (ANOVA) is to test
differences in means (for groups or variables) for statistical
significance. This is accomplished by analyzing the variance, that
is, by partitioning the total variance into the component that is due
to true random error and the components that are due to differences
between means. These latter variance components are then tested
for statistical significance, and, if significant, we reject the null
Basic Statistical Techniques in Research
13
hypothesis of no differences between means, and accept the
alternative hypothesis that the means (in the population) are
different from each other.
The variables that are measured (for example, a test score) are
called dependent variables. The variables that are manipulated or
controlled (for example, a teaching method or some other criterion
used to divide observations into groups that are compared) are
called factors or independent variables. Analysis of covariance
(ANCOVA) is a general linear model with one continuous
explanatory variable and one or more factors. ANCOVA is a
merger of ANOVA and regression for continuous variables.
ANCOVA tests whether certain factors have an effect after
removing the variance for which quantitative predictors
(covariates) account. If there is a significant difference among the
groups, pairwise comparison is carried out. The pairwise tests that
could be used in order to detect the pairs which are significantly
different include Scheffe, Duncan or Turkey.
Multiple Classification Analysis (MCA), which is part of
ANOVA or ANCOVA, is a technique for examining the
interrelationship between several predictor variables and one
dependent variable in the context of an additive model. The MCA
can handle predictors with no better than nominal measurements
and interrelationships of any form among the predictor variables or
between a predictor and dependent variable. It is however essential
that the dependent variable should be either an interval-scale
variable without extreme skewness or a dichotomous variable with
frequencies which are not extremely unequal. In addition, it gives
the percentage contributions of treatment and other categorical
variables to the variance of the dependent measure.
However, the overall percentage contribution of all the
variables (composite) to the variance of the dependent measure
could also be obtained from the Multiple R square, which is
0.207*100=20.7%.
14
F.A. Adesoji & M.A. Babatunde
Table 2: Summary of Analysis of Covariance of Subjects Posttest Achievement Scores in Narrative Composition
Source
Variation
of
Sum of Squares
Df
Mean
square
Covariates PRE
ECAT
591.274
1
591.274
33.019
000
Main
TRT
370.979
3
123.660
6.905
000*
Explained
962.273
4
240.568
13.434
000
Residual
3689.026
206
17.708
Total
4651.229
210
22.143
Effects
F
Sig of F
*Significant at p<0.05
Table 2 shows that there is significant main effects of treatment
since the probability is less than 0.05. In Table 3, if Adjusted
Deviation is added or subtracted to the Grand Mean, the result is
the mean score for each of the treatment groups. For example, the
mean score of WLBIS group is 9.42-0.09=9.33. Table 4 shows that
significant differences existed between groups 1, 2 and 3
separately, and group 4.
Table 3: Multiple Classification Analysis (MCA) of Posttest
Achievement Scores in Narrative Composition
Grand Mean=9.42
Variable Category
N
Unadjusted
Deviation
TRT
1.WLBIS
2.PLEBIS
3.WLPLEBIS
4.CS
55
52
53
50
-0.18
2.26
.45
-2.68
ETA
Adjusted for
Independents
+Covariates
Deviation
-.09
1.65
.52
-2.20
.37
Multiple R Square
Multiple R
BETA
.29
.207
.445
Basic Statistical Techniques in Research
15
Table 4: Scheffe Post-Hoc Analysis of Posttest Means of
Achievement in Narrative According to Treatment Groups
Achievement
GRP 4
GRP 3
GRP 2
GRP 1
Mean
Treatment
7.22
GRP 4
9.33
GRP 1
*
9.94
GRP 3
*
11.07
GRP 2
*
*
(*) Indicates significant differences that are shown in the lower triangle
Multivariate Analysis of Variance (MANOVA)
The Multrivariate Analysis of Variance (MANOVA) is designed to
test the significance of group differences. The only substantial
difference between the two procedures is that MANOVA can
include several dependent variable, wheras ANOVA can handle
only one dependent variable. MANOVA is based on the following
assumptions:
The observations within each sample must be randomly
sampled and must be independent of each other.
The observations on all dependent variables follow a
multivariate normal distribution in each group.
The population covariance matrices for the dependent
variables in each group must be equal (homogeneity of
covariance matrices).
The relationships among all pairs of dependent variables
for each cell in the data matrix must be linear.
Example
A research is interested in examining whether age of some set of
workers could affect their income and hours worked per week. The
two dependent variable being income and hours worked per
week.The three tables below show the Multivariate Analysis of
Variance.
16
F.A. Adesoji & M.A. Babatunde
Table 5: Box’s Test of Equality of Covariance Matrices
Box‟s M
6.936
F
0.766
δf1
9
δf2
2886561
sig
0.648
Table 6: Multivariate Tests for Income and Hours Worked by
Age Category
Effect
value
Intercept
Pillai‟s
Trace
Wilk‟s
Lamdba
Hotelling‟s
Trace
Roy‟s
Largest
root
Age
Pillai‟s
Trace
Wilk‟s
Lamdba
Hotelling‟s
Trace
Roy‟s
Largest
root
F
0.957
Hypothesis
δf
7507.272
2.00
Error
δf
680.0
Sig.
0.000
EtaSquared
0.957
0.043
7507.272
2.00
680.0
0.000
0.957
22.080 7507.272
2.00
680.0
0.000
0.957
22.880 7507.272
2.00
680.0
0.000
0.957
0.091
10.791
6.00
1362.00
0.000
0.045
0.909
11.035
6.00
1360.00
0.000
0.046
0.100
11.279
6.00
1358.00
0.000
0.047
0.099
22.457
3.00
681.00
0.000
0.090
Basic Statistical Techniques in Research
17
Table 5 reveals that there is no significant diefference in the
observed variance (F(9,289)=0.766). Therefore Wilk‟s Lambda f
ratio will be used (If there is a significant difference, Pillai‟s Trace
F ratio will be used). Table 6 reveals that there is a significant
difference among the age groups with respect to income income
and hours worked per week. (F(6,1360) = P<0.05; Z2 =0.046).
Table 7: Univariate ANOVA Summary Table
Source
Dependent
Variable
Sum
of
Squares
df
Mean
Square
F
Sig
EtaSquared
Corrected
Model
Income
1029.016
3
343.005
20.995
0.000
0.085
HRS
64.281
3
21.427
0.167
0.919
0.001
Income
128493.5
1
128493.5
7864.97
0.000
0.920
HRS
1410954
1
1410954
10972.71
0.000
0.942
Income
1029.016
3
343.005
20.995
0.000
0.085
HRS
64.281
3
21.427
0.167
0.919
0.001
Income
11125.807
681
16.337
HRS
87568.119
681
128.588
Income
149966.0
685
HRS
1575151
685
Income
12154.82
684
Intercept
Age
Error
Total
Corrected
Total
HRS
2
87632.4
2
684
2
R = 0.085 (Adjusted R =0.081) ; R = 0.001 (Adjusted R2 =0.004)
Table 7 shows that there is significant difference among the
age groups in income (F(3, 681) =20.995; P< 0.05; Z2 =0.001). Post
hoc analysis could be used to determine the source of significant
difference. This is done in the same way it is done in ANOVA or
ANCOVA.
18
F.A. Adesoji & M.A. Babatunde
Regression Analysis
Regression analysis is a statistical technique used for the modeling
and analysis of numerical data consisting of values of a dependent
variable (response variable) and of one or more independent
variables (explanatory variables). The dependent variable in the
regression equation is modeled as a function of the independent
variables, corresponding parameters (constants), and an error term.
The error term is treated as a random variable. It represents
unexplained variation in the dependent variable. The parameters
are estimated so as to give a best fit of the data. Most commonly
the best fit is evaluated by using the least squares method, but
other criteria have also been used. For some kinds of research
questions, regression can be used to examine how much a
particular set of predictors explains differences in some outcome.
In other cases, regression is used to examine the effect of some
specific factor while accounting for other factors that influence the
outcome.
Regression analysis requires assumptions to be made
regarding probability distribution of the errors. Statistical tests are
made on the basis of these assumptions. In regression analysis the
term model embraces both the function used to model the data and
the assumptions concerning probability distributions. Regression
can be used for prediction (including forecasting of time-series
data), inference, hypothesis testing, and modeling of causal
relationships. These uses of regression rely heavily on the
underlying assumptions being satisfied. Regression analysis has
been criticized as being misused for these purposes in many cases
where the appropriate assumptions cannot be verified to hold.
The set of underlying assumptions in regression analysis is that:
The sample must be representative of the population for the
inference prediction.
The dependent variable is subject to error. This error is
assumed to be a random variable, with a mean of zero.
The independent variable is error-free.
19
Basic Statistical Techniques in Research
The predictors must be linearly independent, that is, it must
not be possible to express any predictor as a linear
combination of the others.
The errors are uncorrelated, that is, the variance-covariance
matrix of the errors is diagonal and each non-zero element
is the variance of the error.
The variance of the error is constant (homoscedasticity).
The errors follow a normal distribution. If not, the
generalized linear model should be used.
The violation of any of these implied conditions could have
potentially negative effects for your research. In a simple linear
regression, we examine a causal relationship between a dependent
variable , Y, and one independent variable, X. Linear regression is
synonymous with equation of a straight line. This is because it is
important to make certain that a linear relationship exists between
the factors before running a regression model. For example, we can
express the relationship between the dependent variable y and the
independent variable x through the following formula:
y =m0 + m1x1 …
(1)
That is, equation (1) says that y can be expressed as a linear
function of x. The constant for the model is represented by the
parameter m0. Regression is a tool that allows us to take data from
some sample and use these data to estimate m0 and m1. These
values are then used to create predicted values of the outcome,
with the observed or true value from the data designated as y and
the predicted value as . Furthermore, in equation (1), the value m1
measures the causal effect of a one unit increase of X on the value
of Y. The parameter m1 is also referred to as the regression
coefficient for X, and is the average amount the dependent variable
increases when the independent variable increases one unit and
other independents are held constant. Thus, when independent
measure increases by 1, the dependent variable increases by m1 units.
20
F.A. Adesoji & M.A. Babatunde
In multiple linear regression, there are more than one
independent variable in the model. Multiple regression allows
researchers to examine the effect of many different factors on some
outcome at the same time. The general purpose of multiple
regression is to learn more about the relationship between several
independent or predictor variables and a dependent variable. This
is because the simple linear regression is a simplification of reality.
In real life, there are more than one variable that affects the
behaviour of the dependent variable. It can be specified as follows:
Y= m0 + m1x1 + m2x2 … + mnxn
(2)
Where Y is the dependent variable; x1 and x2 are the
independent variables; and c0, c1, c2, are the parameters. The
intercept of the regression is c0 while c1 and c2 are referred to as the
partial regression coefficients. The error term is represented by ui.
Hypothesis testing are conducted to show whether the parameters
that have been estimated are statistically significant or, whether the
independent variables contribute to the explanation of variation in
the dependent variable. If we are able to reject the null hypothesis
at an acceptable significance level, then we conclude that the
parameter is not statistically significant. The quality of the fitness
of the model is determined by the R2. The R2 has a value that is
between 0 and 1. High values of R2 will indicate that the model fits
the data well. A limitation in the use of R2 is that its value increases
with the number of explanatory variables. It does not usually
penalize for the consequence loss of degrees of freedom as the
number of explanatory variables increases. The power of the test is
therefore affected. Thus, the adjusted R-square was developed to
take care of the inadequacies. In addition, the F-statistic test for the
joint significance of all the parameters in the model.
The composite and relative contributions of independent
variables to the dependent variable are usually determined through
multiple regression. These are explained in Tables 8, 9 and 10
respectively.
21
Basic Statistical Techniques in Research
Table 8: Composite Effect of Independent Variables (School
Environment and Teacher Competency) on Dependent
Variable (achievement in Integrated Science).
Multiple
Correlation
(R)
R
Square
Adjusted
R-Square
0.962
0.926
0.923
Standard
F
Error of
The
Estimate
4.2834
307.025
Sign F
0.000*
Table 8 shows Multiple Correlation R, the square of this
correlation (R2), Standard Error and F value and the probability
value (Sig. F). The variables under consideration correlated
significantly with the dependent variable. R value is 0.926. If R is
squared and the result is multiplied by 100, the percentage
contribution of all the independent variables taken together to the
variance of the dependent variable is obtained. In this case,
R2=0.923, R2*100=92.3%.
Table 9: Analysis of Variance
Source
Variance
of
Sum
Squares
of
DF
Mean
F
Significance
307.025
0.000*
Regression
39432.31
7
5633.188
Residual
3155.80
172
18.348
Total
42588.11
179
*Significant of p<0.05.
Table 9 gives the Analysis of Variance. The F ratio here is
different from those of ANOVA and ANCOVA. The value here is
significant because the probability is less than 0.05. What this
means is that the R value earlier obtained is not due to chance.
22
F.A. Adesoji & M.A. Babatunde
Table 10: Relative Contributions (Beta Weights) of the Seven
Independent Variables to Students’ Achievement in Integrated
Science
Independent
Variables
(Predictors)
Unstandarized
Coefficients
Beta
Weight
(β)
Rank
T
Significant
12.187
0.687
Std.
Error
3.118
0.061
0.548
1st
3.908
11.273
0.000
0.000*
0.0097
0.122
0.020
5th
0.790
0.431
2.742
0.301
0.435
2nd
9.104
0.000*
0.335
0.159
0.054
3rd
2.102
0.037
0.0063
0.119
0.012
7th
0.539
0.591
0.0027
0.041
0.015
6th
0.670
0.504
0.0033
0.029
0.025
4th
1.131
0.260
B
Constant
Special
Learning
Environment
Special
Equipment
Speech
Theraphy
Knowledge
of Science
and
Education
Knowledge
of learners
Teachers‟
strategies
Counselling
of parents
*Significant at p<0.05
Table 10 gives the relative effects of the independent variables
on the dependent variable. The B is referred to as partial
correlation, the Beta weight (β) is the weight contribution of each
variable. We also have t values for all the variables and the
probability values (sig.t). Note the ranking of the variables
according to their weight contributions. You will notice that the
constant is under B and not under Beta weight (β).
Basic Statistical Techniques in Research
23
(.005)(.000
Fig.2: Hypothesized Recursive Model
Coefficient and Zero Order Correlations
Showing
Path
Figure 2 is a follow-up to Figure 1 after trimming of the
paths in Figure 1. The detail of the trimming could not be given but
it involves removing any path whose path-coefficient is less than
0.05. The model in Figure 2 is usually referred to as parsimonious
model. It shows the paths which have direct and indirect influence
on the dependent variable.
Path Analysis
This is an extension of multiple regression analysis. The use of
path analysis enables the researcher to calculate the direct and
indirect influence of independent variables on a dependent
variable. These influences are reflected on the path coefficients,
which are actually standardized regression coefficients (Beta
weights). Path analysis is one of the techniques for the study and
analysis of causal relations in ex-post facto research. It usually
24
F.A. Adesoji & M.A. Babatunde
starts with hypothesized model and end up with parsimonous
model. This is after carrying out trimming of the paths by using
structural equations. In the hypothesized model belo, variables X1,
X2, X3, X4, X5, X6, and X7 re called independent variables or
exogenous variables while variable X8 is referred to as dependent
variable or endogenous variable. The model is hypothetical.
Fig. 3: The Parsimonious Model
Factor Analysis
Factor analysis is useful in reducing a mass of information to a
manageable and economical description. For example, data on fifty
characteristics for 300 states are unwiedly to handle, descriptively
or analytically. Reducing them to their common factor patterns
facilitates the management, analysis and understanding of such
data. These factors concentrate and index characteristics without
much loss of information. States can be more easily discussed and
compared on economic, development, size and public dimensions
other than on the hundreds of characteristics each dimension
involves.
Basic Statistical Techniques in Research
25
It should be noted that standardizing one‟s variables before
applying factor analysis is not necessary because result of factor
analysis are not affected by the standardization, which is built into
the procedure.
A Hypothetical Situation
Suppose a 25 item questionnaire on students‟ attitude towards
Physics was administered to 40 students. Factor analysis could be
carried out to find out the commonalities of the test items such that
the 25 items would be reduced to a fewer number of items and the
instrument would still be able to measure validly and reliably the
construct attitudes towards Physics. Also, the scale items could be
sorted into their various components so that the items, which
correlate highly with themselves are group together.
Table 11 shows the initial eigen values which provide
information on the % variance explained by each of the variables.
It could be observed that out of the 25 items, the first 9 items
account for 76.75 of the total variance. The 25 items have been
reduced to 9 and the 9 items could be assumed to have measured
the construct, which the 25 items were designed to measure. This
shows that since the 9 items were found to account for 76.75 of the
total variance, if items 10-25 are explained, no serious harm would
be done to the scale of measurement.
The analysis was carried out to establish the number of
meaningful factors. Nine factors have thus been found to be
meaningful or nontrivial. These are the factors considered as
peculiar factors perceived by the students as their attitudes toward
Physics.
26
F.A. Adesoji & M.A. Babatunde
Table 11: Total Variance Explained Students Attitudes Towards
Physics
Items
Initial Eigen Values
Total
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
4.179
3.745
2.662
1.927
1.703
1.458
1.388
1.107
1.017
0.931
0.826
0.686
0.653
0.536
0.443
0.381
0.364
0.329
0.208
0.151
0.105
0.0830
0.05917
0.03782
0.01579
% of
Variance
16.718
14.991
10.647
7.709
6.813
5.834
5.552
4.427
4.069
2.724
2.306
2.746
2.613
2.144
1.771
1.525
1.456
1.317
0.830
0.604
0.422
0.333
0.237
0.151
0.06316
Cumulative
16.718
31.748
42.256
50.065
56.878
62.711
68.264
72.690
76.759
80.483
83.789
86.535
89.148
91.292
93.062
94.588
96.043
97.360
98.190
98.794
99.216
99.549
99.786
99.937
100.000
Extraction Sum of Squared
Loadings
Total
% of
Cumulative
Variance
4.179 16.718
16.718
3.748 14.991
31.709
2.662 10.647
42.356
1.927 7.709
50.065
1.703 6.813
56.878
1.458 5.834
62.711
1.388 5.552
68.264
1.107 4.427
72.690
1.017 4.069
76.759
Extraction Method: principal Component Analysis
Names are usually given to the isolated factors and different ietms
are loaded on each factors.
Correlations
Correlation is a measure of the relation between two or more
variables. It indicates the strength and direction of a linear
relationship between two random variables. The correlation is 1 in
Basic Statistical Techniques in Research
27
the case of an increasing linear relationship, −1 in the case of a
decreasing linear relationship, and some value in between in all
other cases, indicating the degree of linear dependence between the
variables. The closer the coefficient is to either −1 or 1, the
stronger the correlation between the variables. If the variables are
independent then the correlation is 0, but the converse is not true
because the correlation coefficient detects only linear dependencies
between two variables.
Pearson Product-Moment Correlation Coefficient
The most widely-used type of correlation coefficient is Pearson
product-moment correlation coefficient. Pearson's correlation
coefficient is a parametric statistic. It is a common measure of of
the correlation between two variables X and Y. When measured in
a population the Pearson Product Moment correlation is designated
by rho (ρ). When computed in a sample, it is designated by the
letter r and is sometimes called Pearson's r. Pearson's correlation
reflects the degree of linear relationship between two variables.
Pearson correlation, assumes that the two variables are measured
on at least interval scales, and it determines the extent to which
values of the two variables are proportional to each other.
However, the value of correlation (that is, correlation coefficient)
does not depend on the specific measurement units used. For
example, the correlation between height and weight will be
identical regardless of whether inches and pounds, or centimeters
and kilograms are used as measurement units.
Spearman's rank correlation coefficient
The Spearman's rank correlation coefficient is a non-parametric
measure of correlation – that is, it assesses how well an arbitrary
monotonic function could describe the relationship between two
variables, without making any assumptions about the frequency
distribution of the variables.
Spearman's rank correlation coefficient does not require the
assumption that the relationship between the variables is linear, nor
does it require the variables to be measured on interval scales; it
28
F.A. Adesoji & M.A. Babatunde
can be used for variables measured at the ordinal level unlike the
Pearson product-moment correlation coefficient. However,
Spearman's correlation coefficient does assume that subsequent
ranks indicate equidistant positions on the variable measured.
Choosing Appropriate Statistical Technique in a Research
Enterprise
Statistical techniques can be used to describe data, compare two or
more data sets, determine if a relationship exists between variables,
test hypothesis and make estimates about population measures. Not
only it is important to have a sample size that is large enough, but
also it is necessary to see how the subjects in the sample were
selected. Volunteers generally do not represent the population at
large.
However, student should realize that computer merely give
numerical answer and save time and effort of doing calculations by
hand. It is the duty of students to understand and interprets
computer print-out correctly. Note that data can be subjected to
parametric and nonparametric statistics depending on the nature of
the data. Purely numerical data like student score in a chemistry
test could be subjected to a parametric test. Note that this score has
a zero origin like a score of 60% for example. On the other hand,
the number of people who read newspapers in the morning is not a
purely numerical data and it can only be subjected to a
nonparametric test. In this case, you can not perform all the
mathematical operations with the data. You can add, subtract but
you can not determine mean score neither can you carry out
division.
The variable type determines to some extent the type of
statistical (descriptive or inferential) method that it will support.
For example, while trichotomous and polychotomous variables
allow for the use of ANOVA/ANCOVA statistics, the
dichotomous and monochotomous allow for t-test analysis.
Variable types also influences the language of the hypotheses and
hence, inferences that can be made from such hypotheses.
Basic Statistical Techniques in Research
29
In any study, you describe, compare data or determine if
relationship exists between variables. If the percentage score of
group A is higher than that of group B, you are describing the
performance of the groups. When you go a step further by finding
out whether the performance are significantly different you are at
the realm of making inference and this informed the usage of the
parametric statistics; in this case, t-test. When we compare more
than two mean scores, determination of F ratio is involved; in this
case the statistical tool could be ANOVA or ANCOVA depending
on the design employed for the study.
If a study involves determination of relationship, we can use
Spearman Rank order correlation, Pearson Product moment
correlation, Chi-square statistics or even multiple regression
analysis. All depends on the nature of the research. Chi-square
shows the degree of association between two different bases of
classification. It should be noted that the z-test is used only when
the population parameters are known and the variable of interest is
normally distributed in the parent population. If the two conditions
are met, the z-test is used as an exact test, even for small samples
(n<30). However, if the variable is not normally distributed, a large
sample permits the use of a z-test. In this case, the use of z-test is
regarded as an approximate test. In most research situations, the ztest for single mean is rarely encountered because the conditions of
normality and known parameters ( ) are rarely met. However ztest could be used for two means from independent samples.
Normally, z-test is used to test for the mean of a large sample, and
the t-test used for the mean of small sample.
Examples of Topics, Research Questions, Hypotheses and
Selection of Appropriate Statistical Tools
As pointed out earlier, the selection of statistical tool in a study is a
function of the design of the experiment. Also, the language of the
hypotheses is determined by the design. However, a researcher
need not shy away from the fact that the topic of a research is an
offshoot of the statement of the problem. How then shall we
determine the statistics to apply in the analysis of data based on the
30
F.A. Adesoji & M.A. Babatunde
language of the topics, design, and the hypotheses? The following
examples would throw more light on the answer to the question.
1. Topic: Student, teacher and school environment factors as
determinants of achievement in Senior Secondary School
Chemistry in Oyo State, Nigeria.
Questions
i. What is the most meaningful causal model for students‟
achievement in secondary school Chemistry?
ii. To what extent will the seven independent variables when
taken together, predict students‟ achievement?
Interpretation
The key word in the topic is „determinant‟. This is a situation
where a variable (independent) could determine another variable
(dependent), a kind of causal relationship. The independent
variables here are: Student, teacher and school environment
variables while the dependent variable is achievement in Senior
Secondary school Chemistry. It would be observed that hypothesis
is not necessary here. This is an ex-post facto study of the survey
type in which the researcher need not manipulate the independent
variables, they have already manifested.
The first question demands that the experimenter would
construct hypothesized causal model, which he has to trim in order
to get the parsimonoius model. The second question has to do with
the determination of the composite effect of the independent
variables on the dependent variable. The answer to this question
could be obtained through multiple regression analysis. Path
analysis is an extension of multiple regression. Path coefficients
are the beta-weights in multiple regression.
2. Topic: Effects of guided discovery and self-learning strategies
on secondary school students‟ learning outcome in
Chemistry.
Basic Statistical Techniques in Research
31
Hypotheses
1. There is no significant main effect of treatment on
i. Learning outcome in Chemistry
ii. Attitude to Chemistry
2. There is no significant main effect of ability on:
i. Learning outcome in Chemistry
ii. Attitude to Chemistry
3. There is no significant main effect of gender on
i. Learning outcome in Chemistry
ii. Attitude to Chemistry
4. There is no significant interaction effect of treatment and
gender on
i. Learning outcome in Chemistry
ii. Attitude to Chemistry
Interpretation
The topic shows effects of two methods of instruction
(independent) variables on two dependent variables (learning
outcome in Chemistry and attitude to Chemistry). There should be
a control group. The design then becomes: Pretest, Posttest, control
group experimental design. Therefore, the appropriate statistical
tool is Analysis of Covariance (ANCOVA) with pretest scores as
covarite. Two categorical varaibles are being investigated. These are
gender (dichotomous) and academic ability level (trichotomous). And
because of the interaction hypotheses, there must be factorial design.
Here it is 3 x 2 x`3, which is interpreted as follows: treatment at 3
levels, gender at 2 levels nd academic ability at 3 levels.
3.
Topic: A comparative analysis of leadership styles of male
and female managers in the banking industry in
southwestern Nigeria.
Hypothesis
There is no significant relationship in the leadership styles of
male and female managers in the banking industry.
32
F.A. Adesoji & M.A. Babatunde
Interpretation
Other hypotheses could be based on other vaiables investigated in
the study. Here, the nature of the data generated cannot be purely
numerical. Therefore, non-parametric statistics is the best bet. The
candiadte could use Chi-square statistic for analyzing te data
collected.
4.
Topic: An evaluation of extra-mural studies programmes of
the University of Ibadan 1989/90 and 1998/99
sessions.
Hypotheses:
1. The extra-mural studies significantly influenced
achievement of self-actualization and self-esteemvalues of
candidates.
2. There is no significant difference in the academic
achievement of candidates in the 1989/90 and 1998/99
sessions.
3. Ther is no significant difference in the cademic
achievement of candidates from three different locations in
the 1989/90 session.
Interpretation
It could be observed that there are different languages in the three
hypotheses. Hypothesis 1 could be tested using Chi-Square statistic
because of the word “influence”. Hypothesis 2 and 3 has to do with
purely numerical data (academic achievements). While hypothesis
2 is a test of significnce between two sets of candidates from two
sessions, hypothesis 3 is a test of significance among candidates
from three different locations. In this case, hypothesis 2 will be
tested using t-test while hypothesis 3 will be tested uning Analysis
of Varianace (ANOVA). Note that the word “Evaluation” in the
topic can accommodate virtually all statistical procedures; it all
depends on the nature of the hypothesis.
Basic Statistical Techniques in Research
5. Topic:
33
Impact of government labour integration policy on
Nigerian dockworkers‟ productivity in Nigerian Ports
Authority, Nigeria.
Hypotheses:
1. Ther is no significant impact of government labour
integration policy on Nigerian workers‟ productivity in
public and private sectors of the economy.
2. There is no significant impact of government labour
integration policy on Nigerian male and female workers‟
productivity.
3. There is no significant impact of government labour
integration policy on three different categories of workers‟
productivity.
Interpretation
It could be observed that this could also be a cause and effect
study. The word “impact” in the topic is synonymous with
“effect”. Thus, the study demands the usage of t-test and Analysis
of Variance (ANOVA) depending on the nature of the hypotheses.
The study is ex-post facto but of the experimental type. Hypothesis
1refers to public and private sectors therefore, t-test will be
appropriate. Hypothesis 2 will also be tested with t-test beause it is
between male and female workers while hypothesis 3 can only be
tested with Analsysi of Variance (ANOVA) because three
categories of workers are involved.
6. Topic: Relationship between some school factors and
Secondary School system efficiency in Ogun State,
Nigeria.
Hypotheses
1. There is no significant relationship between the school
related factors and school system efficiency.
2. There is no significant relationship between school location
and school system efficiency.
Other hypotheses could be generated based on other school-related
factors.
34
F.A. Adesoji & M.A. Babatunde
Interpretation
This is a descriptive survey research and based on the languages of
the topic and hypotheses, it is a relational study. The hypotheses
could be tested by Pearson Product Moment Correlation or Chisquare statistic. If ranking is involved the candidate could still
make use of Spearman rank order correlation.
Data Analysis and Interpretation
Unprocessed data are called raw data. Data have to be processed
by making use of computers to for analysis. This is because
manual procedures for estimating and computing relevant statistics
have become increasingly tedious or entirely impossible.
Computers are now applied in all aspects of statistical analyses,
from the calculation of simple sums to the estimation of large scale
stochastic models.
There are numerous statistical programs for analyzing data.
These are known as package or canned programs. The popular
programs in this class include SPSS, SAS, GAUSS, E-Views,
RATS, LIMDEP, and STATA. However, the most popular series
for the educational researcher is the Statistical Package for Social
Sciences (SPSS). SPSS contains many of the most common
statistical procedures needed by the students.
References
Adesoji, F.A. (2006). Statistical Methods for Data Analysis and
Data Interpretation In Alegbeleye, G.O, Mabawonku, I and
Fabunmi, M (eds.) Research Methods in Education,
University of Ibadan, Ibadan.
Bluman, A.G. (1990) Elementary Statistics. McGraw Hill, Higher
Education, New York.
Gbadegesin, A., R. Olopoenia, and A. Jerome (2005) “Statistics for
the Social Sciences”. Ibadan University Press.
Gujarati, D. N (1995) Basic Econometrics. 3rd Edition. McGraw
Hill, New York.