International Journal of Assessment Tools in Education
2019, Vol. 6, No. 4, 706–720
https://dx.doi.org/10.21449/ijate.548516
Published at http://www.ijate.net
http://dergipark.org.tr
Research Article
Developing a Scale to Measure Students’ Attitudes toward Science
Adem Akkuş
1
1,*
Mus Alparslan University, Education Faculty, Elementary Science Education, Mus, Turkey
ARTICLE HISTORY
Received: 03 April 2019
Revised: 31 October 2019
Accepted: 06 December 2019
KEYWORDS
Science Attitude,
Scale Development,
Scale validation
Abstract: The aim of this study is to develop a science attitude scale (SAS).
For that purpose, the literature review has been done for suggestions for creating
scales and a new draft scale developed. The draft scale was analyzed by
specialists and a pilot study is done after its approval by experts. The SAS is
prepared with 21 items and among these, 11 items are reverse-coded. The SAS
consists of Likert-type items. The sample of the study consists of 154 college
students studying at the Faculty of Education, Elementary Science Education,
and Elementary Education departments. Principal axis factoring with
orthogonal rotation (varimax) was used for exploratory factor analysis. Factor
eigenvalues were checked with respect to parallel analysis and numbers of the
factors were determined with respect to the analysis. Items that did not serve
the purpose of the scale were omitted from the SAS. The finalized SAS’
Cronbach alpha value is .953. For confirmatory factor analysis data were
collected from a different sample which consists of university students who
were studying at elementary science education, elementary education, and
electric electronic engineering departments. Number of sample is 201.
Confirmatory factor analyses run through Amos 24.0 software. It is believed
that SAS is a valuable contribution to the science education field since it has
unidimensional structure and proved its item discrimination power, and
alongside with an excellent internal consistency. SAS also offers opportunity to
develop multidimensional science attitude scale. For that purpose, original SAS
and English version of it are provided in appendixes.
1. INTRODUCTION
Attitude is defined as an individual’s positive or negative characteristics towards a subject
(Serin & Mohammadzadeh, 2008). Students with positive attitudes toward science are likely to
display more science-related attitudes and choose science-related professions. On the other
hand, recent studies indicate that there is a trend in that science-related departments attract
fewer students than social science-related departments (Shah & Mahmood, 2011). Therefore,
attitudes toward science and science-related subject areas are in focus of research studies. Even
science attitudes may be used for predicting science achievement (Adesoji, 2008). Factors
affecting attitudes are also among the subjects to be studied. For example, gender might be
suggested as one of the factors. Although both genders have closely similar attitude values
CONTACT: Adem AKKUŞ ademakkus@gmail.com
Elementary Science Education, Mus, TURKEY
ISSN-e: 2148-7456 /© IJATE 2019
706
Mus Alparslan University, Education Faculty,
Int. J. Asst. Tools in Educ., Vol. 6, No. 4, (2019) pp. 706–720
toward science, underlying factors might be different. Girls learn better in an organized
environment and boys’ attitudes are related to cohesiveness. Other factors might be listed as
instructional style, teaching strategies, classroom design, etc. (Bernardez, 1982). Thus, having
knowledge of students’ attitudes and encouraging them toward science is important and
attitudes of students must be known (Shah & Mahmood, 2011). Scales are useful for this
purpose and in this regard, researchers try to develop their own instruments for various purposes
or use a standardized version (Coll, Dalgety & Salter, 2002). Using standardized scales or other
means of standardized measurements could value the purpose, letting researchers have an
opinion on the attitudes of students and understand dimensions and their value within the
context (Demirbaş, 2009). On the other hand, standardized scales are mostly in English and
have different theoretical aspects, different cultural settings, psychometric properties and hence
may lack assess the right domain of interest, not be suitable for local use due to contextual
differences (Shah & Mahmood, 2011). Perhaps that is the reason why different researchers have
failed to confirm the Test of Science Related Attitudes scale (TOSRA) in their sample
population. Attitudes may be observed in different types of responses and even be affected by
curriculum changes (Cheung, 2007). As curriculum changes are made, the need for to measure
attitudes and to develop new scales also becomes a value of interest to observe the effect of the
curriculum. Even instructional techniques might affect students’ attitudes whose change could
value the future implications (Evrekli, İnel, Balım & Kesercioğlu, 2009). Since Turkey has
already announced that changes on the curriculum are done to promote active learning (TTK,
2017), it is important to observe the effects of curriculum changes on students’ attitudes. For
that aim, several researchers already tried to develop attitude scales or applied existing ones.
For example, Can and Şahin (2015) studied kindergarten teacher candidates’ attitudes toward
science and science teaching. Analyses were done to investigate the relationships of grades and
gender with science attitude and science teaching. Serin & Mohammadzadeh (2008) used a
scale to determine attitude and academic success relationships. Korkmaz, Şahin and Yeşil
(2011) tried to investigate attitude toward scientific research. For that reason, they developed a
scale with 30 items and four dimensions. Tortop (2013) adapted a scale into Turkish for
assessing scientific field trip attitude. The study reveals that a single attitude might have
different dimensions. For example, another study tried to investigate the relationships between
attitudes and science process skills (Dönmez & Azizoğlu, 2010). All these studies show that
scales might be used for collecting data (Deshpande, 2004; Hinkin, 1998; Wong & Lian, 2003;
Francis et. al., 2004) so that effective measures might be taken into account for this purpose
(Hinkin, 1998; Hinkin, Tracey, Enz, 1997). Thus, the purpose of this study is to develop a
science attitude scale (SAS). Attitudes may have different dimensions and scales may reflect
those dimensions. However, most scales determined the number of dimensions based on
eigenvalues through factor analysis. SAS also determined number of dimensions through
parallel analysis which reflects more accurate number of dimensions. Moreover, SAS is a
unidimensional scale but offers the chance for researchers to develop multidimensional scales
based on SAS.
2. METHOD
2.1. Research Design
In the method, to develop a scale, based on suggestions from the literature, the guidelines have
been determined (Brinkman, 2009; Deshpande, 2004; Hinkin, 1998; Hinkin et. al., 1997;
Johanson & Brooks, 2010; Ajzen, 2005b; Francis et. al., 2004; Cabrera-Nguyen, 2010; Hof,
2012). Those guidelines are:
a) Not to cause any bias, the items’ context must be within the students’ cultures
(schemes).
b) Respondents should place themselves at a position.
707
Akkuş
c) Items must assess a single behavior or response.
d) Items must not be interpreted in different ways.
e) Language and expected knowledge should be familiar with the target group.
f) Sensitive and double negative items should be avoided.
Among Thurstone’s method of equal-appearing intervals, Likert scale, semantic differential
scales; it is determined that a Likert type scale would be more beneficial for the purpose of the
study to ensure easy compilation and generalization (Lovelace & Brickman, 2013; Brinkman,
2009; Johanson & Brooks, 2010; Hof, 2012). To ensure content adequacy and avoid fatigue, a
maximum number of items is determined so that respondents will respond within attention time.
For this purpose, the sentences “Strongly agree” or “Strongly disagree” are given at the
beginning of the scale as information. By placing five levels of response for an item, it is
ensured that internal consistency is increased and sufficient variances are obtained (Hinkin,
1998; Lovelace & Brickman, 2013; Brinkman, 2009; Hinkin et al., 1997; Ajzen, 2005b; Francis
et al., 2004). Since there might be respondents tending to choose options at the edges or in the
same direction, reversed coded sentences are appropriately used to trigger their vigilance
(Hinkin, 1998; Francis et al., 2004; Hof, 2012)
The process of developing the science attitude scale (SAS): The item sentences were finalized
after determining SAS’ scope, content, items and their numbers. After that specialist views were
taken account. The draft science attitude scale (SAS) consisted of 31 items. However, items 25
and 27 were removed from the SAS since they were the same as items 6 and 5. Initial internal
reliability analysis was carried out and Cronbach’s α value was found as .861 (good according
to Kalaycı, 2010). The draft scale’s content and scope were analyzed by instructors who have
the experience of teaching and researches on related issues since specialists could value the
prepared scale on the content domain. Specialists work in the education faculty at science
education department (Hinkin et al., 1997). Specialists’ views’ on sentences and corrected itemtotal correlation values of the items were cross-checked, and the items regarded as problematic
were excluded from the scale. Item of 1 and 9 contradicted guideline f “Sensitive and double
negative items should be avoided”, items of 8, 11, 14, and 26 contradicted guideline c “Items
must assess a single behavior or response” and guideline d “Items must not be interpreted in
different ways”. Thus, those items were excluded from the SAS immediately for further
analysis; thus finalized SAS Cronbach’s α reliability value is .943 and with 23 items. The
developed SAS consists of twelve reversed questions (items) which are items 2, 5, 6, 7, 10, 16,
17, 20, 21, 22, 23, and 31.
2.2. The sample size and sampling method
The SAS was applied to 154 college students at the Faculty of Education, Elementary Science
Education, and Elementary Education departments. In order to ensure the privacy of personal
information, (i.e. avoiding conflict of interest) only the students' gender and age information
were demanded.
In literature, to determine a sample size has been a controversial issue. Some researchers argue
about arbitrary sampling which presents high communalities without cross-loadings. Thus,
sampling may be determined by the nature of the data i.e. More acceptable view, some
researchers claim that if data is strong enough then sample size might be small, while others
argue on item-ratio sampling. The debate on item-ratio suggests proportion from 1:2 to 1:10 for
item and sampling (Anthoine, Moret, Regnault, Sébille & Hardouin, 2014; Hinkin, 1998;
Hinkin et al., 1997; Cabrera-Nguyen, 2010). Since the item respondents’ ratio of the study is
1:7, it is believed that sampling is adequate for the study with respect to first view.
As for arbitrary sampling, several arguments might be stated. For example, Johanson & Brooks
(2010) suggests to social researchers that minimum participants are 100 people for sampling.
708
Int. J. Asst. Tools in Educ., Vol. 6, No. 4, (2019) pp. 706–720
For a comprehensive item analysis, sample size 100 to 200 person should be conducted since
standard errors for Cronbach’s alpha value increase as the sample size decreases. However, it
is also noted that regardless of the number of items, the mean inter-item correlation is nominal
between N= 30 and 200. Hinkin (1998) and Hinkin et al., (1997) suggest N=150 to obtain
sufficient data for exploratory factor analysis as long as item inter correlations are reasonably
strong. However, researchers also note the difference between statistical and practical
significance because attaining statistical difference chance increases as the sample size
increases. Larger samples are useful to detect small fluctuations. On the other hand, as sample
size increases the practical meaning of the results may distort, so the decision on sample size
must be made with caution. Francis et al. (2004) state that some researchers claim that N=25
would be enough for purposive sampling and sample size could be increased until it is believed
that data saturation has been achieved. Yet, researchers claim a sample size N=80 would
generally be enough. Cabrera-Nguyen (2010), while stating similar statements, also indicates
that some researchers argue on sampling size and claim sample size depends on the gathered
data, and adequacy of sampling is determined after analyzing the gathered data. In the same
paper, the researcher embraces a mixed approach based on communalities value. Hof (2012)
suggests 10-15 respondents per item, yet states a KMO value already signals whether the
sample size is enough or not. Based on the suggestions above, it is again regarded that the
sample of the study will be enough for the purpose.
3. RESULT
3.1. Reliability Analysis
The data analyzed with respect to internal consistency, communalities, and factor loadings.
Analyses were carried out together for better judgment of retaining factors. Corrected item-total
correlation values of the items were analyzed. It was observed that items 12 (.004) and 16’s
(.126) corrected item-total correlation values were below the desired value of .200 (Johnson &
Morgan, 2016), hence; these items were excluded from the scale. After this process the scale’s
Cronbach’s α value was found as .953 and “excellent” for the final version of SAS.
3.2. Exploratory Analysis
Exploratory Factor Analysis: A principal axis was conducted on the 21 items with orthogonal
rotation (varimax) through the SPSS program to reveal the factors within the developed scale
since it is suggested for more reliable scale evaluation (Field, 2013; Hof, 2012). The KaiserMeyer-Olkin (KMO) measure verified the sampling adequacy as “marvelous” (Kalaycı, 2010).
The KMO value is .951 and above the acceptable limit of .5 (Field, 2013). Bartlett’s test of
sphericity was found significant (X2(210) = 2392.067, p= .00 < .05). Hence, the KMO value
already signaled that the sample size might be enough, the analysis of each SAS item was
initiated. An initial analysis was run to obtain eigenvalues for each factor in the data. Three
factors emerged having eigenvalue over Kaiser’s criterion of 1 and in combination
explained %58,778 of the variance. Eigenvalue of the factors were 11,301; 1,340 and 1,017
respectively for factor 1, factor 2 and factor 3.
The parallel analysis offers a good interpretation of the number of retaining factors (Field, 2013;
Johnson & Morgan, 2016) thus; a Monte Carlo PCA for parallel analysis with 1000 replications
was run to confirm the eigenvalues (Watkins, 2000). Eigenvalues obtained were 1,7259; 1,5915
and 1,4902 for that reason, it was concluded that only the first eigenvalue was acceptable since
the second and third factor’s eigenvalue was not significant. Therefore, an EFA was rerun with
a one-factor solution. The variance shared by the factor was 51,937. The scree plot (Figure 1)
was obtained and it was decided that the scale has one factor with respect to the convergence
709
Akkuş
of the scree plot and Kaiser’s criterion on this value. Table 1 shows the factor loadings after
rotation and extracted communalities.
Figure 1. Scree Plot
Table 1. extracted communalities and factor loadings
Item
SAS2
SAS3
SAS4
SAS5
SAS6
SAS7
SAS10
SAS13
SAS15
SAS17
SAS18
SAS19
SAS20
SAS21
SAS22
SAS23
SAS24
SAS28
SAS29
SAS30
SAS31
h2
,215
,602
,568
,422
,474
,259
,174
,494
,787
,621
,295
,826
,773
,549
,671
,721
,624
,207
,817
,617
,190
Factor
,464
,776
,754
,650
,689
,509
,417
,703
,887
,788
,543
,909
,879
,741
,819
,849
,790
,455
,904
,785
,435
Since none of the items’ factor loadings were below .400 (Table 1) validation of scale’s internal
consistency reliability coefficient was made as suggested (Field, 2013; Francis et. al., 2004).
Cronbach’s α value was found as =.953 and it means “excellent” (Kalaycı, 2010). For a detailed
analysis of items on discrimination value, an independent samples t-test was run for each item.
Lower and upper %27 of the samples (N=42) were compared through independent samples ttest. This analysis shows items’ discrimination value of individuals between lower and
710
Int. J. Asst. Tools in Educ., Vol. 6, No. 4, (2019) pp. 706–720
upper %27 of the sample and is used by many scientists in scale developments (Moore & Foy,
1997). Reliability analysis of item-total correlation and tup-down(%27) results and items’ codes are
shown in Table 2.
Table 2. Item-total correlation and tup-down(%27) results
Item
Mean
SAS1
SAS2
SAS3
SAS4
SAS5
SAS6
SAS7
SAS8
SAS9
SAS10
SAS11
SAS12
SAS13
SAS14
SAS15
SAS16
SAS17
SAS18
SAS19
SAS20
SAS21
2,92
3,14
3,05
3,14
3,18
3,23
3,06
3,07
3,31
3,20
3,06
3,19
3,38
3,03
3,17
3,26
3,22
3,16
3,19
3,08
3,01
Standard
Deviation
1,603
1,500
1,522
1,650
1,602
1,475
1,423
1,417
1,619
1,479
1,538
1,555
1,500
1,434
1,564
1,385
1,515
1,380
1,443
1,412
1,186
Corrected-item
total correlation
,456
,757
,738
,643
,677
,506
,412
,680
,863
,765
,520
,889
,862
,722
,801
,829
,771
,444
,887
,764
,414
tup-down(%27)
5,767*
12,781*
12,796*
9,729*
14,399*
6,588*
4,944*
9,175*
26,213*
12,029*
7,381*
22,715*
23,059*
13,191*
16,176*
17,507*
13,456*
5,343*
20,686*
14,565*
6,093*
* p < .05
The finalized SAS consists of 21 items and 11 items are reverse coded items. The reversed
coded items are 1, 4, 5, 6, 7, 10, 13, 14, 15, 16 and 21 (Table A1). For international readers, an
English translation of SAS is given in Table A2. Translation was done by the researcher and to
ensure translation was done correctly and comprehension of the scale is easy, SAS was
presented to a professor to take account of specialist’s opinion. That professor was working at
university and had a formal education in English language. After that a retranslation and
crosscheck were done by another professor who also had a formal education in English
language and working at Education faculty.
3.3. Confirmatory Factor Analysis
For confirmatory factor analysis data were collected from a different sample. The sample
consists of university students who were studying at elementary science education, elementary
education, and electric electronic engineering departments. The number of sample is 201 in
total. Confirmatory factor analyses run through Amos 24.0 software. Initial analysis results
revealed that χ2/DF ratio is 2,635 RMSEA value is .09; GFI value is .802; CFI value is .627;
SRMR value is .0852; NFI value is .520; AGFI value is .758. RMSEA, GFI, AGFI values
showed model did not show a good fit with respect to indices values. However, it is noteworthy
that they are close to the desired value. CFI and NFI values did not fall into the categories of
well fit. Thus, in the review of the literature it is decided to examine the error terms and decide
whether some items correlate together or not for a better comprehension of the model. Decisions
to covariate items based on the rule of thumbs. First, covariated items should exhibit similar set
of idea/pattern/mindset in phrases. Second, number of covariated items should be restricted up
711
Akkuş
to 7. Third, each time a covariate is done then, the results would be checked to see if model fit
indices were changed dramatically or not. Fourth, covariances between the items should not be
lower than 10. Covariated items are 13 and 14; 9 and 11; 12 and 17; 18 and 19. Final data
analyses results revealed that χ2/DF ratio is 1.893; RMSEA value is .067; SRMR value is .0746;
IFI value is 0.806; GFI value is .852; CFI value is .800; NFI value is .662; PNFI value is 0.583;
NNFI (TLI) value is 0.773; AGFI value is .815. The confirmatory factor analysis result is shown
in Figure 2.
Figure 2. Confirmatory Analysis Result
χ2/DF ratio is 1.893 and it is regarded that a model has a good fit if Chi-square (χ2)/degree of
freedom (df) ratio is < 2. RMSEA value is .067 and model has a good fit since RMSEA ≤ 0.1.
IFI value is 0.806 and CFI value is .800 and it is accepted model has a good fit since CFI ≥ 0.8
and IFI ≥ 0.8. (Browne and Cudeck, 1993; Garson, 2006 as cited in Chinda, Techapreechawong
& Teeraprasert, 2012).
Many recommendations are being done on value of Root Mean Square Error of Approximation
(RMSEA). For example, Pedroso et. al. (2016) state if RMSEA is ≤ 0.05 then, it indicates good
fit and if RMSEA is ≤ 0.08 it indicates good fit with reasonable errors. Other recommendations
generally advise that if RMSEA ≤ 0.08 then a model has good fit and, if 0.08 < RMSEA ≤ 0.1
712
Int. J. Asst. Tools in Educ., Vol. 6, No. 4, (2019) pp. 706–720
then it indicates adequate fit (Carlback & Wong, 2018; Shadfar & Malekmohammadi, 2013).
The confirmatory analysis revealed that RMSEA value is .067 so SAS has good fit of model.
Since RMR has bias, SRMR used instead of RMR. SRMR value of the confirmatory factor
analysis result is 0.0746. SRMR value below ≤ 0.08 indicates good fit (Carlback & Wong,
2018; Kline, 2011 as cited in; Kaya & Altinkurt, 2018; Vassallo & Saba, 2015).
Kline (2011) mentions that most used fit indices (GFI, AGFI, NFI, NNFI, CFI and IFI) should
be ≥ 0.85 (cited in; Kaya & Altinkurt, 2018; Vassallo & Saba, 2015). Other researchers indicate
AGFI, GFI and CFI values should be ≥ 0.80 (Byrne & Campbell, 1999 as cited in Nayir, 2013).
However, GFI is affected by sample size and for that reason, AGFI is developed. GFI ≥ 0.85
and AGFI ≥ 0.80 is accepted as good fit (Sica & Ghisi, 2007). On the other hand, AGFI also is
sensitive to sample size. For that reason, it is advised by researchers to disregard them. Yet
papers still indicate both values. Reason for that is not for their importance but historical values.
TLI ≥ 0.85 indicates good fit and > 0.8 mediocre fit (Carlback & Wong, 2018; Shadfar &
Malekmohammadi, 2013). The confirmatory analysis revealed that IFI value is 0.806; GFI
value is .852; CFI value is .800; NNFI (TLI) value is 0.773; AGFI value is .815 so SAS has
good fit of model.
4. DISCUSSION and CONCLUSION
SAS item discrimination values show that it might be used to measure the attitude toward
science of college students since all the items yielded significant results between up-down%27.
On the other hand, a detailed analysis of t-test values might reveal the facts lying beneath. For
example, the highest t value (26,213) of item9 implies that students who have high positive
attitudes enjoy the experiments, however; the value of items11 and 21 (7,381 and 6,093) imply
students wait for confirmation on their experiment results or expect more guidance during
experiments. Fin (2012) states perceived learning increases as students get feedback from
instructors, thus; the expectation of students is meaningful and expected in this context. Similar
reports indicate that instructor confirmation has a positive effect on cognitive learning (Schrodt,
Witt, Turman, Myers, Barton & Jernberg, 2009). Therefore, instructors need to interact with
students and help them in cognitive development.
Students have a tendency of thinking that feelings might assist a scientist. Low t value (4,944)
of item 7 already implies that both upper-lower (%27) students think that facts may have
subjective aspects. Perhaps, students think that scientific facts may change, and some
unscientific factors act as catalysts for that change. In fact, t value (5,767) of SAS1 already
hints that both upper-lower (%27) students have a tendency of the idea that facts might be
subjective. From a different aspect, that approach might be seen through t value (6,588) of
SAS6. That item indicates faith plays an important role in students’ ideas about science’s role.
Conflicts between the dimensions may create gaps for the students on the nature of science. As
a result, students struggle between scientific facts and faith. Since faith requires believing in
what is being told without question, it may also create a barrier toward scientific approach and
hence, students may close themselves to new ideas. Students who have high positive attitudes
toward science are also open to new ideas and t value (22,715) of SAS12 already reveals this.
One of the purposes of science courses is learning how to distinguish faith from science and
instructors should make students aware that; two concepts actually are not related. In fact, faith
and science have different roles and do not need to cut out each other’s way. For instance, a
high value of SAS19 (t =20,686) clarifies that students who have higher positive attitudes also
acknowledge that scientific knowledge is essential as it is related to life itself. Embracing this
idea might eventually increase positive attitudes toward science. Similarly, Nuhoglu (2008)
who developed a science attitude scale in Turkey mentions that one of the factors is “new
knowledge and using it” and Ajzen (2005a) points out that changing the attitudes and behaviors
may be achieved through changing beliefs.
713
Akkuş
For a cross-cultural analysis, Scientific Attitude Inventory II (SAI-II) was also analyzed since
some statements are similar to the statements in SAI-II. However, researchers did not provide
an item t%27up-down values thus a comparison could not be made (Moore & Foy, 1997).
Similarly, another well-known scientific attitude scale TOSRA, which has seven dimensions,
also did not provide such an approach. On the other hand, the author states that having a low
score on one of the dimensions should not concern instructors because this information might
be valuable in facilitating to identify profiles and creating solutions. One of the most important
aims of science education is considered to develop a positive attitude towards science (Fraser,
1981).
Motivation towards science has a long-term effect on science learning and it is affected by
different things such as curriculum structures. For that reason, Foley & McPhee (2008)
investigated the effect of different curriculum approaches on their study and argued that
students’ experiences might be affected by different curriculum structures. Kurnaz and Yigit
(2010) report that Turkish students have tended to develop a negative attitude towards science
since 2005. Thus, it is important to assess the changes of attitudes caused by curriculum changes
in Turkey (TTK, 2017). Although this research was done on local scale, it might be said that
science attitude affects the scientific approach and the scientific approach is the same all around
the world. Moreover, attitudes might be affected by a similar insight context whose effect may
yield similar results. Perhaps through such studies, identifying and creating solutions will be
possible. Researchers and teachers might use the developed scale and observe their students’,
attitudes, use interventions and may offer solutions. Then, perhaps understanding students’
attitudes on different dimensions may also offer solutions for long term aims.
An advantage of this study is seen as using parallel analysis to confirm eigenvalues in
identifying dimensions. Since the traditional factor analysis determination procedure is based
on eigenvalue of 1 then, the obtained number of factors may not be accurate. It is believed
through such analysis such cases are avoided and true factor structures are determined in
creating the SAS (Hayton, Allen & Scarpello, 2004). Developing science attitude scales is
important, even restudying the existing ones may provide useful information. For that reason,
researchers either create their own scales or develop the existing ones (Moore & Foy, 1997).
Having an excellent internal consistency (α=.953) and a consisted structure (one factor
structure) measuring the related domain, it is believed that SAS is a valuable contribution to the
science education field.
Validity must be considered each time when an instrument is used since the instrument was
validated for a sample or population but was not validated for another sample or population i.e.
structure may show varying results from a sample to another. Validity is not property of a scale
but it means as an instrument of interpretation. There are arguments on cut off values of fit
indices since they may lead the decision of an acceptable model to be rejected. Thus, it is
important for researchers to conduct the analyses and use their own judgments with respect to
obtained values. Values of fit indices help to understand the structure of a model and thinking
all the fit values together will provide a better decision. In other words, fit indices will help to
understand relationships of the items among each other and within the model structure. Fit
indices should confirm the model but it should not be used for championing the model in every
possible indices which will cause an artificially approved model. It is advised that once items
and factors make sense in the theoretical aspect of the researcher then, the decision will be
based on that (Knekta, Runyon & Eddy, 2019). Hu & Bentler (1995) argue decisions based on
fit indices and reminds that strictly depending on fit indices values may result in rejecting true
models, especially for small sample sizes such as 250 or 500. GFI and AGFI tend to increase
when sample size increase same thing could be also said for RMR and RHO (Anderson &
Gerbing, 1984). NFI results could be problematic if sample size is < 200 thus usage of NNFI
(TLI) is recommended. However, it is also noted that even NNFI could still indicate poor fit if
714
Int. J. Asst. Tools in Educ., Vol. 6, No. 4, (2019) pp. 706–720
sample size is not enough. Similar arguments are proposed for TLI since it could report poor fit
due to similar reasons. For that reason, suggestions for TLI could be as low as 0.80 and, for
parsimony fit indices (PNFI) values ≥ 0.5 indicates good fitness (Hooper, Coughlan & Mullen,
2008). The confirmatory analysis revealed that PNFI value is .583 so SAS has good fit of model.
Although there are other authors who conclude that with sample size N=200 a reasonable
estimate could be obtained for CFI and TLI, still researchers are warned since decisions strictly
based on CFI could also cause wrong decisions because it also depends on sample size and
hence rejection of fit model. For example, a correct model simulated with a sample N=200 and
CFI value turned out to be .611 (poorly fitting model). It is noted that even with a relatively
large sample size (N=500) a conventional cut off value of TLI may cause a correct model to be
rejected (Shi, Lee, Maydeu-Olivares, 2018). Hu & Bentler (1999) notes sample size ≤ 250 could
cause problems in Maximum Likelihood (ML) analysis. Thus, warns researchers to be
cautionary on evaluation on model fit evaluation. Questions (items) assessing the same target
or different items having nearly same meaning with different words may be the cause of
correlated errors which in fact, may cause the correlate error terms (Meyer, n.d.). Since SAS
has correlated error terms, this also concludes the idea that model could provide a
multidimensional aspect if provided with enough number of items targeting the domain of
interest. Be that as it may, Ellis (2017) states that if p value is < 0.05 and 0.05 < RMSEA < 0.08
then, null hypothesis is not exactly true but model has acceptable fit. Although generally
accepted indice values are ≥ 0.80 for fit indices, a proposed common guideline for indice values
follows as; very good fit ≥ 0.90; adequate but marginal fit ≥ 0.80-0.89; poor fit ≥ 0.60-0.79;
very poor fit > 0.60 (Planing, 2014).
As a final thought, it is believed that the developed SAS might be used in different regions/states
to compare the results and validate its purpose. For that reason, SAS with different samples is
welcome. With this aim, both the created SAS in the original language and an English version
of SAS are given in the appendixes.
ORCID
Adem AKKUŞ
https://orcid.org/0000-0001-9570-3582
5. REFERENCES
Adesoji, F. A. (2008). Managing students’ attitude towards science through problem–solving
instructional strategy. The Anthropologist, 10 (1), 21-24.
Ajzen, I. (2005a). Behavioral interventions based on the theory of planned behavior: Brief
description of the theory of planned behavior. Retrieved from http://people.umass.edu/aizen
/pdf/tpb.intervention.pdf (accessed on 5 February 2019)
Ajzen, I. (2005b). Constructing a theory of planned behavior questionnaire: Brief description
of the theory of planned behavior. Retrieved from http://people.umass.edu/aizen/pdf/tpb.me
asurement.pdf (accessed on 5 February 2019)
Anderson, J. C., & Gerbing, D. W. (1984). The effect of sampling error on convergence, improper
solutions, and goodness-of-fit indices for maximum likelihood confirmatory factor analysis.
Psychometrika, 49(2), 155–173. DOI: 10.1007/bf02294170
Anthoine, E., Moret, L., Regnault, A., Sébille, V., & Hardouin, J.-B. (2014). Sample size used to
validate a scale: a review of publications on newly-developed patient reported outcomes
measures. Health and Quality of Life Outcomes, 12, 2. DOI:10.1186/s12955-014-0176-2
Bernardez, R.Q. (1982). Factors affecting attitudes to laboratory work. Unpublished Master Thesis,
Saint Louis University, Baguio, Philippines. [Abstract]
Brinkman, W-P. (2009). Design of a questionnaire instrument, handbook of mobile technology
research methods. ISBN 978-1-60692-767-0, pp. 31-57 Netherlands: Nova Publisher
Browne, M.W. & Cudeck, R., 1993. Alternative ways of assessing model fit. In: Bollen, K.A. and
Long, J. S. (Eds.) Testing structural equation models, Beverly Hills, CA: Sage
715
Akkuş
Byrne, B. M. &Campbell, T. L. (1999). Cross-cultural comparisons and the presumption of
equivalent measurement and theoretical structure: A look beneath the surface. Journal of
Cross-Cultural Psychology, 30, 557 - 576. DOI: https://doi.org/10.1177/002202219903000
5001
Cabrera-Nguyen, P. (2010). Author guidelines for reporting scale development and validation
results in the Journal of the Society for Social Work and Research. Journal of the Society for
Social Work and Research, 1(2), 99-103.
Can, M., & Şahin, Ç. (2015). Okul öncesi öğretmen adaylarının fene ve fen öğretimine yönelik
tutumlarının incelenmesi [Investigating Prospective Kindergarten Teachers' Science and
Science Teaching Attitudes]. Abant İzzet Baysal Üniversitesi Eğitim Fakültesi Dergisi, 15
(2), 13-26. DOI: 10.17240/aibuefd.2015.15.2-5000161311
Carlback, J. & Wong, A. (2018). A study on factors influencing acceptance of using mobile
electronic identification applications in Sweden. Retrieved from http://www.divaportal.org/smash/get/diva2:1214313/FULLTEXT01.pdf (accessed on 03 April 2019)
Cheung, D. (2007). Confirmatory factor analysis of the attitude toward chemistry lessons scale.
Paper presented at 2nd NICE symposium, Taipei, Taiwan, July 30-31, 2007.
Chinda, T., Techapreechawong, S., & Teeraprasert, S. (2012). An investigation of relationships
between employees’ safety and productivity. Retrieved from http://www.ppml.url.tw/EPPM/
conferences/2012/download/SESSON4_A/10%20E145.pdf (accessed on 12 October, 2019)
Coll, R.L., Dalgety, J. & Salter, D. (2002). The development of the chemistry attitudes and
experiences questionnaire (CAEQ). Chemistry Education Research and Practice in Europe,
3(1), 19-32.
Demirbaş, M. (2009). The relationships between the scientist perception and scientific attitudes of
science teacher candidates in Turkey: A case study. Scientific Research and Essays, 4(6),
565-576.
Deshpande, L. (2004). Challenges in measurement of scientific attitude. Paper presented at
epiSTEME-1: An International Conference to Review Research on Science Technology and
Mathematics Education (137-138), Goa, India, December 13-17, 2004.
Dönmez, F., & Azizoğlu, N. (2010). Investigation of the students science process skill levels in
vocational schools: a case of Balıkesir. Necatibey Faculty of Education Electronic Journal
of Science and Mathematics Education, 4 (2), 79-109.
Ellis, J.E. (2017). Factor analysis and item analysis. Applying Statistics in Behavioural
Research (pp. 11-59). Retrieved from https://www.applyingstatisticsinbehaviouralresearch.c
om/documenten/factor_analysis_and_item_analysis_version_11_.pdf (accessed on 22
December 2018)
Evrekli, E., İnel, D., Balım, A. G., & Kesercioğlu, T. (2009). The attitude scale of constructivist
approach for prospective science teachers: a study of validity and reliability. Journal of
Turkish Science Education, 6(2), 134-148.
Field, A. (2013). Discovering statistics using ibm spss statistics (4th Edition). London: SAGE
Finn, A.N. (2012) Teacher use of prosocial and antisocial power bases and students’ perceived
instructor understanding and misunderstanding in the college classroom. Communication
Education, 61(1), 67-79, DOI: 10.1080/03634523.2011.636450
Foley, B., & McPhee, C. (2008). Students’ attitudes towards science in classes using hands-on or
textbook based curriculum. AERA, 1-12.
Fraser, B.J. (1981). Test of science-related attitudes (TOSRA) handbook. Victoria: Allanby
Garson, G.D., 2006. Structural equation modelling. North Carolina: G. David Garson and Statistical
Associates Publishing
Hayton, J.C., Allen, D.G & Scarpello, V. (2004). Factor retention decisions in exploratory factor
analysis: A tutorial on parallel analysis. Organizational Research Methods, 7(2), 191-205.
DOI: 10.1177/1094428104263675
716
Int. J. Asst. Tools in Educ., Vol. 6, No. 4, (2019) pp. 706–720
Hinkin, T. R. (1998). A brief tutorial on the development of measures for use in survey
questionnaires. Organizational Research Methods, 2 (1), 104-121. DOI:10.1177/109442819
800100106
Hinkin, T. R., Tracey, J. B., & Enz, C. A. (1997). Scale construction: Developing reliable and valid
measurement instruments. Journal of Hospitality & Tourism Research, 21(1), 100-120.
DOI:10.1177/109634809702100108
Hof, M.W. (2012). Questionnaire Evaluation with Factor Analysis and Cronbach’ s Alpha:
An Example. Retrieved from http://www.let.rug.nl/nerbonne/teach/rema-stats-meth-seminar
/student-papers/MHof-QuestionnaireEvaluation-2012-Cronbach-FactAnalysis.pdf (accessed
on 02 May 2016).
Hooper, D., Coughlan, J. & Mullen, M. R. (2008). Structural Equation Modelling: Guidelines for
Determining Model Fit. The Electronic Journal of Business Research Methods, 6 (1), 53 –
60. DOI: 10.21427/D7CF7R
Hu, L.-T., & Bentler, P. M. (1995). Evaluating model fit: in Structural Equation Modeling Ed. Rick
H. Hoyle. London: Sage Publications
Hu, L.-T., & Bentler, P. M. (1999). Cut off criteria for fit indexes in covariance structure analysis:
Conventional criteria versus new alternatives. Structural Equation Modeling: A
Multidisciplinary Journal, 6(1), 1-55. DOI: https://dx.doi.org/10.1080/10705519909540118
Johanson, G.A., & Brooks, G.P. (2010). Initial scale development: Sample size for pilot studies.
Educational and Psychological Measurement, 70(3), 394-400. DOI: 10.1177/00131644093
55692
Johnson, R.L. & Morgan, G.B. (2016). Survey scales: Investigating scale quality. New York, NY:
The Guilford Press.
Kalaycı, Ş. (2010). Spss uygulamalı çok değişkenli istatistik teknikleri. (5. Baskı) [SPSS applied
various statistical techniques (5th Edition)]. Ankara: Asil Yayın Dağıtım Ltd. Şti.
Kaya, Ç. & Altinkurt, Y. (2018). Öğretmenlerin psikolojik sermayeleri ile tükenmişlik düzeyleri
arasındaki ilişkide psikolojik ve yapısal güçlendirmenin rolü [Role of Psychological and
Structural Empowerment in the Relationship between Teachers’ Psychological Capital and
Their Levels of Burnout]. Eğitim ve Bilim, 43 (193), 63-78, DOI:
http://dx.doi.org/10.15390/EB.2018.6961
Kline, R. B. (2011). Principles and practice of structural equation modeling (3rd edition) New York:
The Guilford Press
Knekta, E., Runyon, C., & Eddy, S. (2019). One Size Doesn’t Fit All: Using Factor Analysis to
Gather Validity Evidence When Using Surveys in Your Research. CBE—Life Sciences
Education, 18 (1), 1-17. DOI: https://doi.org/10.1187/cbe.18-04-0064
Kurnaz, M.A. & Yigit, N. (2010). Physics attitude scale: Development, validity and reliability.
Necatibey Faculty of Education Electronic Journal of Science and Mathematics Education,
4 (1), 29-49.
Korkmaz, Ö., Şahin, A. & Yeşil, R. (2011). Bilimsel araştırmaya yönelik tutum ölçeği geçerlik ve
güvenirlik çalışması [Study of Validity and Reliability of Scale of Attitude towards Scientific
Research]. Elementary Education Online, 10 (3), 961-973.
Lovelace, M. & Brickman, P. (2013). Best practices for measuring students’ attitudes toward
learning science. CBE-Life Sciences Education, 12(4), 606-617. DOI: 10.1187/cbe.12-110197
Meyer, J. (n.d.). Correlated errors in confirmatory factor analysis. Retrieved from
https://www.theanalysisfactor.com/correlated-errors-in-confirmatory-factor-analysis
(accessed on 17 September 2019).
Moore, R.W. & Foy, R.L.H. (1997). The scientific attitude inventory: A revision (SAI II). Journal
of Research in Science Teaching, 34(4), 327-336.
Nayir, F. (2013). “Algılanan örgütsel destek ölçeğinin” kısa form geçerlik güvenirlik çalışması
[“Perceived Organizational Support Scale”- Short Form Validity-Reliability Study].
Mehmet Akif Ersoy Üniversitesi Eğitim Fakültesi Dergisi, 28, 89-106.
717
Akkuş
Nuhoglu, H. (2008). The development of an attitude scale for science and technology course.
Elementary Education Online, 7(3), 627-639.
Pedroso, R., Zanetello, L., Guimaraes, L., Pettenon, M., Goncalves, V., Scherer, J., Kessler, F., &
Pechansky, F. (2016). Confirmatory factor anlaysis (CFA) of the crack use relapse scale
(CURS). Archives of Clinical Psychiatry, 43 (3), 37-40.
Planing, P. (2014). Innovation Acceptance: The Case of Advanced Driver-Assistance Systems:
Quantitative research approach (pp. 230-231). Stuttgart: Springer
Sica, C. & Ghisi, M. (2007). The Italian versions of the Beck Anxiety Inventory and the Beck
Depression Inventory-II: Psychometric properties and discriminant power. In M.A. Lange
(Ed.), Leading - Edge Psychological Tests and Testing Research (pp. 27-50). New York:
Nova
Schrodt, P., Witt, P.L., Turman, P.D., Myers, S.A., Barton, M.H & Jernberg, K.A. (2009) Instructor
credibility as a mediator of instructors’ prosocial communication behaviors and students’
learning
outcomes.
Communication
Education,
58(3),
350-371,
DOI:
10.1080/03634520902926851
Serin, O., & Mohammadzadeh, B. (2008). The relationship between primary school students’
attitudes towards science and their science achievement (sampling: Izmir). Journal of
Educational Sciences, 2 (6), 68-75.
Shadfar, M. & Malekmohammadi, I. (2013). Application of Structural Equation Modeling (SEM)
in restructuring state intervention strategies toward paddy production development.
International Journal of Academic Research in Business and Social Sciences, 3 (12), 576618. DOI: 10.6007/IJARBSS/v3-i12/472
Shah, Z.A. & Mahmood, N. (2011). Developing a Scale to Measure Attitude towards Science
Learning among School Students. Bulletin of Education and Research, 33 (1), 71-81.
Shi, D., Lee, T., & Maydeu-Olivares, A. (2018). Understanding the model size effect on SEM fit
indices. Educational and Psychological Measurement, 79(2), 310-334. DOI:
https://doi.org/10.1177%2F0013164418783530
Tortop, H. S. (2013). Bilimsel Alan Gezisi Tutum Ölçeği Adaptasyon Çalışması [Adaptation Study
of Attitude Scale towards Scientific Field Trips]. Bartın Üniversitesi Eğitim Fakültesi
Dergisi, 2(1), 228.
TTK. (2017, July 18). Müfredatta yenileme ve değişiklik çalışmalarımız üzerine [On
curriculum revision and changes work]. Retrieved from https://ttkb.meb.gov.tr/meb_iys_do
syalar/2017_07/18160003_basin_aciklamasi-program.pdf (accessed on 25 August 2018).
Watkins, M. W. (2000). Monte carlo PCA for parallel analysis [computer software]. State College,
PA: Ed & Psych Associates.
Wong., M. & Lian, S. (2003). Development of a self-efficacy scale for assessing secondary school
students’ science self-efficacy beliefs. Retrieved from http://citeseerx.ist.psu.edu/viewdoc/s
ummary?doi=10.1.1.472.8479 (accessed on 19 October 2014)
Vassallo, M., & Saba, A. (2015). Does money for grocery expenditure sway Italian consumers’
motivational values in predicting Attitude towards eco-sustainable food products?
Contemporary Management Research, 11(1), 3-22. DOI: doi:10.7903/cmr.13840
718
Int. J. Asst. Tools in Educ., Vol. 6, No. 4, (2019) pp. 706–720
6. APPENDIX
2
Bilim adamları gerçeğin/fikirlerin değişebileceğine
inanırlar/kabul ederler
Kesinlikle Katılıyorum
Bilimsel gerçekler değişmez
Katılıyorum
1
Kararsızım
Madde
Katılmıyorum
Bilimsel Tutum
Kesinlikle Katılmıyorum
Table A1. Bilimsel Tutum Ölçeği
1
2
3
4
5
1
2
3
4
5
3
Bilim adamları sorularının cevaplarını her zaman bulamazlar
1
2
3
4
5
4
Bilim adamları birbirlerinin çalışmalarını eleştirmemelidirler
1
2
3
4
5
5
Bilimsel çalışmalar bilim adamları içindir
1
2
3
4
5
6
Dinle çatışan konular çalışılmamalı/önemsenmemeli
1
2
3
4
5
7
Bir bilim adamının sahip olduğu en önemli araç hisleridir
1
2
3
4
5
8
Bilimsel gelişmeler daha sağlıklı yaşam sürmemizi sağlar
1
2
3
4
5
9
Deney yapmak derslerden daha zevklidir
1
2
3
4
5
10
Bilimsel keşifler faydadan çok zarar veriyor
1
2
3
4
5
11
Hocanın anlatmasındansa deney yaparak gerçekleri bulmayı
1
2
3
4
5
tercih ederim
12
Farklı fikirleri hoş karşılarım
1
2
3
4
5
13
Fen dersleri zaman kaybıdır
1
2
3
4
5
14
Fen konuları zevksizdir
1
2
3
4
5
15
Deney yapmaktansa teorik bilgiler daha faydalıdır
1
2
3
4
5
16
Fen deneylerine daha az vakit verilmeli
1
2
3
4
5
17
Deneyler grup çalışmasıyla daha zevkli geçer/geçiyor
1
2
3
4
5
18
Bilimin temel amaçlarından biri yeni ilaçlar ve tedaviler
1
2
3
4
5
bulmaktır
19
Yaşamı etkilediğinden İnsanlar bilimsel gerçekleri anlamalı
1
2
3
4
5
20
Bilim bir şeyin nasıl olduğunu açıklamaya çalışmaktır
1
2
3
4
5
21
Bilimsel çalışma bana zor gelir
1
2
3
4
5
719
Akkuş
Strongly Disagree
Disagree
Undecided
Agree
Strongly Agree
Table A2. Scientific Attitude Scale
1
Scientific facts do not change
1
2
3
4
5
2
Scientists acknowledge/accept that facts may change
1
2
3
4
5
3
Scientists cannot always find the answers
1
2
3
4
5
4
Scientists should not criticize each other’s work
1
2
3
4
5
5
Scientific works are for scientists
1
2
3
4
5
6
Topics contradicting with religion should not be
1
2
3
4
5
Scientific Attitude
Item
studied/cared
7
The most important tool for a scientist is her/his feelings
1
2
3
4
5
8
Scientific progress helps us to have more healthy life
1
2
3
4
5
9
Doing experiments is more fun than having lectures
1
2
3
4
5
10
Scientific progress outputs harm more than good
1
2
3
4
5
11
I prefer to find facts rather than told by the instructor
1
2
3
4
5
12
I welcome different ideas
1
2
3
4
5
13
Science courses are waste of time
1
2
3
4
5
14
Science courses are tasteless
1
2
3
4
5
15
Theoretical knowledge is more helpful than experimenting
1
2
3
4
5
16
Science course hours must be reduced
1
2
3
4
5
17
Experiments are more fun with group works
1
2
3
4
5
18
One of the main aims of science is to find new cures
1
2
3
4
5
19
People should understand scientific facts since it affects life
1
2
3
4
5
20
Science is trying to explain things
1
2
3
4
5
21
Scientific works are baffling for me
1
2
3
4
5
720