Assessing Speaking Skills: A: Workshop For Teacher Development Ben Knight
Assessing Speaking Skills: A: Workshop For Teacher Development Ben Knight
Assessing Speaking Skills: A: Workshop For Teacher Development Ben Knight
Speaking skills are often considered the most important part of an EFL
course, and yet the difficulties in testing oral skills frequently lead teachers
into using inadequate oral tests or even not testing speaking skills at all.
This article describes a workshop used in teacher development
programmes to help teachers with one aspect of the problem of oral
testing: what should we look for when we assess a student's ability to speak
English? The workshop looks first at the range of criteria that teachers
might use in such assessment. Then it examines how the selection and
weighting of those criteria should depend on the circumstances in which
the test takes place. The article also discusses issues raised by the
workshop, and considers its applicability to people working in different
circumstances.
Reasons for the Assessment of speaking skills often lags far behind the importance given
workshop to teaching those skills in the curriculum. We recognize the importance of
relevant and reliable assessment for providing vital information to the
students and teachers about the progress made and the work to be done.
We also recognize the importance of backwash (the effect of the test on
the teaching and learning during the course). Most teachers would accept
that 'if you want to encourage oral ability, then test oral ability' (Hughes,
1989:44). But the problems of testing oral ability make teachers either
reluctant to take it on or lacking in any confidence in the validity of their
assessments. Such problems include: the practical problem of finding the
time, the facilities and the personnel for testing oral ability; the problem of
designing productive and relevant speaking tasks; and the problem of
being consistent (on different occasions, with different testees and
between different assessors). Another problem, which is the focus of the
workshop framework described here, is deciding which criteria to use in
making an assessment. The workshop has two principal aims:
The workshop The workshop takes between about 1 Vi and 2Vi hours and requires two or
three short video clips of students talking.1 Making your own video clips
is preferable, as you can make the task and situation reflect the type of test
which the teachers you are addressing are most likely to use.
/. Conclusion (5 mins.)
The presenter can ask students for their comments on the workshop—how
useful it was, how it could have been more useful, whether they think they
would change the way they assess their students' speaking skills, and so
on.
Discussion There is still a great deal of subjectivity in a) the selection of criteria, and
7. Objective b) the way each criterion is measured (e.g. how exactly do you decide the
criteria? grammatical accuracy of a speaker's performance?). The workshop aims
only to improve the quality of those subjective decisions about selecting
criteria by making it more conscious and explicit, and by giving the
teachers a chance to discuss other points of view. It assumes that teachers
do not have the resources to carry out their own research. A kind of
collective subjectivity can be reached for how each criterion is measured
by 'training' or 'moderating' sessions for assessors. But for those who
have the time and resources to look closely and objectively at these
questions, the following will be of interest: Hinofotis (1980), Hieke
(1985), Fulcher (1987), and Lennon (1990).
Assessing speaking skills 29S»
2. Analytic or Several tests, such as the RSA Profile Certificate or the ILR (Interagency
holistic Language Roundtable) oral interview, use a different type of criterion to
assessment? this workshop. The speakers are observed in different speaking tasks and
they are simply judged for their degree of success in that task. This
holistic approach argues that, as we cannot observe directly mental
characteristics like grammatical knowledge or ability to maintain
conversations, it will be inaccurate to give (inferred) scores for them.
Rather we should simply assess the learner's (observable) success in
performing authentic language tasks.
The approach behind this workshop, however, is one which argues that it
is those mental abilities (which we must infer from the learner's
performance) that we are most interested in, for at least the following
reasons. Firstly, we cannot predict, let alone test for, every function and
situation which a learner might need English for. Therefore any claim
about general proficiency must involve a lot of inferences from the few
sample functions/situations we do test. Secondly, a lot of teachers' own
tests are partly diagnostic, and teachers need to know more about why a
particular learner performed badly in some situations and better in others.
This will usually come down to inferred component abilities, such as
range of vocabulary or fluency. For a detailed discussion of the two
approaches, see Bachman (1990: 41-42, 301-333). Hughes (1989: 110)
recommends using both approaches, with one as a check on the other.
3. When to do this This workshop on its own may seem peripheral as teachers often worry
workshop? more about such problems as making a practical test, setting fair tasks and
getting a fair sample of the students' language, and being consistent.
However, it is probably helpful to tackle the problem of assessment
criteria before these other questions, since we need to start by deciding
what we want to measure, before deciding what is the most reliable and
practical means to measure it. The danger, otherwise, is that we choose
reliable and practical tests (e.g. multiple-choice tests) which do not give
us the information we really want about our students' oral skills, and
which can have a negative effect on students' attitudes to developing
those skills during the course.
4. Too complicated? Considering the context in selecting assessment criteria does make the
discussion more complicated. So with teachers for whom this topic is
completely new, it would probably be better to leave such considerations
aside or condense them severely. I have found some untrained teachers
saying that they wished they could have come away from the workshop
with one fixed set of 'best criteria'. Gebhard (1990:158,160) reports that
handed-down direction is preferred by beginning teachers (quoting
research in Copeland, 1982) and by teachers in certain countries who feel
that 'if the teacher is not given direction by the supervisor, then the
supervisor is not considered qualified'.
However, taking the testing context into account is valuable, despite the
added complexity, in dealing with two common problems with teacher-
development workers.
300 Ben Knight
Firstly, it makes it easier for teachers to apply what they leam in the
workshop to their own situations, especially when they are working in
contexts very different from that of the presenter. This is also helped by
the final task .(which is an exercise in applying to a particular situation
principles learnt during the workshop).
Secondly, it helps resolves conflicts of opinion. Many of the
disagreements at the beginning of a workshop can be related to different
assumptions about the testing context and its effect on the selection of
criteria. Thus, it not only improves our understanding, but also improves
the conduct of the workshop: it avoids the 'anything goes' approach
which creates cynicism, and it reduces the ex cathedra judgements by the
presenter which can lead to resentment or passivity.
Conclusions The workshop works as a way of stimulating teachers to think about and
discuss the way they assess their students' speaking skills. It is rare for a
participant not to be absorbed by the tasks and the exchange of ideas. A
few participants have found it rather frustrating not to have a fixed set of
testing criteria at the end of the workshop, but most seemed to find the
process of relating criteria to context helpful in clarifying their own
positions. A large survey of teachers' testing experience found 'there is
evidence that most [teachers] prefer to use informal and flexible
approaches which can be adapted to different student populations'
(Brindley, 1989: 31). This workshop suits such preferences.
Received September 1991