ReadingInA2ndLanguage Sample Ch17
ReadingInA2ndLanguage Sample Ch17
352
challenges for reading assessment. The fourth section addresses a set of
further issues for reading assessment.
Reading-proficiency assessment
Assessment of reading proficiency is important as a way to understand
students’ overall reading abilities (based on some assumed construct of
reading) and to determine if students are appropriately prepared for further
learning and educational advancement. Commonly, this type of assessment
is referred to as standardized testing, although local groups and researchers
also develop proficiency tests of different types. In most respects,
proficiency assessment represents high-stakes testing because decisions are
often made about students’ future educational goals and opportunities.
354 Expanding reading comprehension skills
1. Cloze
2. Gap-filling formats (rational cloze formats)
3. C-tests (retain initial letters of words removed)
4. Cloze elide (remove extra word)
5. Text segment ordering
6. Text gap
7. Choosing from a “heading bank” for identified paragraphs
8. Multiple-choice
9. Sentence completion
10. Matching (and multiple matching) techniques
11. Classification into groups
12. Dichotomous items (T / F / not stated, Y / N)
13. Editing
14. Short answer
15. Free recall
16. Summary (1 sentence, 2 sentences, 5–6 sentences)
17. Information transfer (graphs, tables, flow charts, outlines, maps)
18. Project performance
19. Skimming
20. Scanning
C-tests are variants of cloze formats, but rather than deleting whole
words, the initial letter or syllable of a targeted word remains, and students
use this clue, along with the sentence context, to determine the missing
word. This option is less of a production task. Cloze elide, in which “extra”
words are meant to be struck out, have the advantage of not requiring any
production. In both cases, however, it is not easy to make persuasive
construct-validity arguments for these formats.
Text-segment ordering and text-gap formats involve the moving around
of whole sentences or paragraphs, or the selection of the right space in the
text to supply a sentence or paragraph. Text-gap formats can be tricky when
multiple gaps are created and a list of sentences or paragraphs is provided to
insert in the correct spaces. These formats amount to a type of multiple
matching task. Choosing from a heading bank to label identified paragraphs
is a similar type of task. The strength of these types of tasks is that they call
on knowledge of discourse signals and discourse structuring to be answered
successfully. They require several comprehension skills for appropriate task
completion.
A number of task formats in Table 17.3 are relatively uncommon. Free-
recall formats simply ask test takers to make a list of ideas they remember
from a text they have just read. These responses are matched up against a
list established by the test maker. Summary formats can be straightforward
though difficult to score. Alternative summary formats can include, for
example, choosing the best from among three summary options and
identifying the weaknesses of unacceptable options. Information-transfer
formats, especially visual representations of text information, have powerful
construct-validity support. However, they can be difficult to score and can
have very high item interdependence. Project-performance evaluation is a
newer task format that evaluates test takers as they read texts and then
perform in groups to carry out a larger project. It is an interesting option, but
is problematic on several validity grounds (giving individual scores based on
group interactions and a holistic task). Skimming and scanning tasks are
well-known to teachers but are not common as standardized reading
assessment tasks.
In closing this discussion of standardized assessment, it is important to
emphasize that all formats need to go through a careful validation process
that includes a theoretical justification, a feasibility study (Does the task
work the way that it is supposed to?), a piloting study for reliability and item
performance, and a fairness review. While individual teachers are seldom
called upon to create standardized tests, programs and schools are
sometimes expected to generate a standardized assessment. It is important
that medium- and high-stakes tests be constructed and used carefully and in
the fairest possible way.
Reading assessment 361
Classroomassessmentallowsforamuchwiderrangeoftasksandstudent
observations. Testing formats in classroom contexts can include all of the
assessment formats identified for standardized assessment (Table 17.3).
Classroom reading assessments can also make use of informal reading
inventories or miscue analysis (reading aloud one-on-one with an evaluator
who notes errors on a record sheet and then determines what progress a
student has made or what instructional support is needed by the student).
A number of other informal, or alternative, assessment types are also well
recognized and commonly used in classroom contexts. Almost any language
task that is a useful teaching task can be used as an assessment task. What
might be lost in the way of relatively weak validity or consistency for any
given reading task or measurement in the classroom setting is, in principle,
countered by the continual nature of assessment practices of all types in this
context. Several types of informal assessment options are available to
teachers, including the following:
1. Observations
2. Self-reporting measures
3. Progress charts
4. Performance inventories
5. Participation and engagement records
6. Portfolios
1. Teachers wait for 3–5 seconds after asking a question, without answeringthe
question when students are silent, or without switching to a strong student for
the answer.
2. Teachers ask students to generate good questions about reading texts thatcould
be included on a test of some sort. Students learn how to formulate good
questions about the texts they read, and their questions then appear on quizzes
and comprehension checks.
3. Teachers move from more traditional question-and-answer sequencesabout
reading passages to questions that begin a discussion with students about their
understanding of the text.
4. Teachers withhold grades on student work until students respond tocomments
on their assignments (e.g., multiple text syntheses, filled-in graphic
organizers). In many cases, teachers do not provide grades at all, only
comments.
5. Teachers deal with wrong answers or performance difficulties in ways
thatengage students in finding good answers and achieving task success.
Teachers provide feedback to encourage student learning.
6. Teachers engage in “why” questions and “why” follow-up questions
togenerate discussions about an answer.
7. Students engage in self-assessment, monitoring their comprehension andtheir
levels of success in carrying out during-reading and postreading activities.
8. Students work in groups to answer questions and are given time to workout
answers together.
9. Students set learning goals at the outset of instruction.
10. Students assess the work of other students, usually through commentsrather
than specific grades.
11. Students select a peer who will answer a question, for example, by drawinga
student’s name from a hat.
12. Students learn to monitor their noncomprehension of a text or task andsignal
for assistance when they have difficulty.
13. Students review their own learning effectiveness.
14. Students redo their answers on a reading task or a quiz, often with
peersupport.
15. Students use “traffic-light” discs (green on one side, red on the other)
tosilently signal to the teacher that they are doing well or having trouble when
working on their own or in groups.
366 Expanding reading comprehension skills
tests that require test takers to read over, or skim, longer texts in brief time
periods and answer general questions. In contrast, this option would ask test
takers actually to read for main ideas, key details, connections among sets of
information, and integration of information. Such items would need to
presume a minimum reading rate and level of vocabulary knowledge to
allow such a task with a longer text. A test option might involve, with
computers, reading a passage of 750 words (read at 150 wpm) that then
disappears after 5 minutes. Test takers might then answer questions, fill in
outlines, list major ideas, or click-and-drag 10 statements from a list of 20
statements about the text that are true. Alternatively, tests might have a
passage of 750 words, and when students have finished reading it, they press
a button; the text disappears and the questions appear (providing both rate
and comprehension). These types of tasks also impose expectations about
reading fluency as part of the subconstruct being measured.
Third, adding a time limit to the above tasks raises the more general topic
of assessing reading fluency. Should fluency be assessed as part of a reading
comprehension measure? Are there aspects of fluency that are supportable
as part of the construct of reading and that are feasible as assessment tasks?
Obvious candidates would include exploring wordidentification fluency
skills and simple reading-rate skills (word lists, oral reading for 1 minute,
silent reading on a computer, timed reading,
assessmentofrereading).TheseformatsmightbemoreusefulatlowerL2
proficiency levels. At higher levels, oral passage-reading measures might be
relevant and fairly simple as a task type under the right conditions and
scoring rubric (see Fuchs et al., 2001; Jenkins et al., 2003a). Such tests
would ultimately place less emphasis on task authenticity and direct
measurement of reading comprehension and more emphasis on prior
extensivereadingpractice. Perhapsataskthatdirectlysupports extensive
reading practice would be an entire test working with one or two long texts,
with a variety of assessment tasks being produced around the text(s). The
limitations of such a test format would also need to be addressed. Such a
performance test would almost be like a professional or clinical performance
assessment and would require a considerable amount of extensive reading as
general preparation.
Fourth, assessment of discourse structure is another option that can be
explored further. With respect to text factors, we need to know how reading-
assessment difficulty can vary by text choice (genre, length, complexity,
vocabulary, organization). Will literary texts vs. general narrative texts vs.
informational texts vs. persuasive texts generate very different assessment
outcomes? Discourse-structure knowledge might also be assessed
effectively through various types of discourse-awareness tasks (see Chapter
12), and especially through the use of extended graphic organizers (Jiang,
2007). This use of graphic organizers is currently employed in some tests to
370 Expanding reading comprehension skills
Consequences of assessment
All assessments have consequences. In a classroom setting, students may
feel that they did not perform well. Teachers may be disappointed in a
specific performance by a student and let it show even if no grade is
involved. Informal assessments may inadvertently be made based on one or
two salient experiences rather than a fair accumulation of evidence over
time and across tasks. Teachers and administrators may use grades, tests,
and evaluation reports to make decisions about students that should require
additional input. Summative achievement tests certainly have consequences,
as do standardized tests. Even in an assessment for learning contexts, a
steady diet of difficulties and poor performances can take a toll on both
students and teachers. The point is that assessment, no matter the context, is
serious business and must be handled with care, purpose, and expertise.
Most important, teachers cannot opt out of the consequences of
assessment.Evenwhenteachersstatethattheydonotbelieveinstandardized
assessment or grading, they are generating consequences of assessment.
They are leaving assessment to others who may be less informed about the
students, their progress, and their needs. When teachers do not take
assessment seriously as their responsibility, they give up their ability to
advocate for students in assessment contexts or support fair assessment
practices. Teachers need to develop expertise in assessment practices of all
types if they are to ensure fair uses of assessment and appropriate
consequences of assessments. Only when many teachers have true
assessment expertise will their views be heard by administrators and policy
consultants.
Assessment is sometimes treated as a bad word. But all of us engage in
assessment all the time, both in our daily lives and in the classroom. We
assess the value of most daily activities we engage in and most purchases we
make. We assess our friends, our work, our leisure activities, and what we
eat. We assess students’ homework, their group work, their responses to
pre-, during-, and postreading activities, and their level of engagement in
SSR in class. Assessments also provide information for continuous learning.
It should not be surprising, then, that assessments are essential to learning
institutions, teachers, and students. Honest, fair, and appropriate assessments
are needed by students if they are to learn most effectively. Taking this view
of assessment also highlights the consequences of assessment. It is
everyone’s responsibility to ensure that these consequences are not harmful
or unfair.
Reading assessment 373
or are having difficulties. They may involve student and teacher discussions
to identify sources of comprehension difficulty and strategies for addressing
these difficulties. Projects may also involve actual assessments that can be a
resource for further instruction with students. The point of assessment for
learning is not that common assessment formats are inappropriate (e.g,.
quizzes, comprehension questions, unit tests), but that assessment formats
should be means for learning about difficulties and then working to address
these difficulties. The key for teacher exploration is how to turn this
information from assessment formats into effective learning opportunities.
Informal assessments often provide important opportunities for student
learning. Usually there is a product or a performance that students and
teachers can work on together, pointing to specific skills and
strategiesthatcanimprovelearning.Studentscanexplorehowtoimprovetheir
performance for future cycles in concrete ways. Again, action research
projects specifically focused on how to use informal assessments in support
of learning would be a way to build teacher expertise and student learning.
Two final issues that deserve mention involve (a) the relationship
between learning activities and assessment activities in the classroom, and
(b) the relationship between informal assessment and formal assessment in
the classroom. In the case of the former, can all learning activities in a class
also become assessment opportunities? If so, how; if not, why not?
Conversely, can all assessment activities also become learning activities? In
the case of the link between informal and formal assessment, when might
this linkage be emphasized and how might the two be linked up in ways that
support effective feedback to students? These are both issues that would
make useful and productive action research projects for a group of teachers
to explore. Neither has been discussed extensively in either assessment or
teacher training literature, but they would be very useful topics to
incorporate into future teacher-development practices.