The Effect of Using Alternative Assessment Methods On EFL Students
The Effect of Using Alternative Assessment Methods On EFL Students
Misurata Branch
School of: Languages
Department of: English
By
Yasmine Farid Abudagga
Supervised by:
Dr. Abdulhamid Onaiba
Associate Professor
No part of this dissertation was previously submitted for another degree at the
Libyan Academy or any other institution.
Name:
Signature:
Date:
Dedication
III
Acknowledgments
I express my deepest gratitude to ALLAH, the most compassionate and merciful, for
granting me the strength, knowledge, abilities, and opportunity to embark on this study
Onaiba who made this work possible. His guidance and advice carried me through all
I would also like to thank Dr. Ibrahim Henaish for his invaluable guidance and advice
throughout my research and the completion of this project. I am truly grateful for his
moment, and for their brilliant comments and suggestions, thanks to you.
Finally, I would like to extend my special thanks to my beloved husband, Seraj, and my
entire loved family, with a particular mention of my cherished parents, and my dear
grateful for their presence in my life and the prayers they offered, which have been a
my precious friend, Mona, for her constant support and assistance during this process.
IV
Table of Contents
Declaration .................................................................................................................................... II
Dedication .................................................................................................................................... III
Acknowledgments........................................................................................................................ IV
List of Tables............................................................................................................................... VII
List of Abbreviations ................................................................................................................. VIII
Abstract ........................................................................................................................................ IX
Chapter One.................................................................................................................................. 1
1. Introduction .............................................................................................................................. 1
1.1. Assessment in the Libyan Context .............................................................................. 3
1.2. Statement of the Problem ............................................................................................ 5
1.3. Aims of the Study ......................................................................................................... 6
1.4. Research Hypotheses .................................................................................................... 6
1.5. Research Questions ...................................................................................................... 7
1.6. Scope of the Study ........................................................................................................ 7
1.7. The Significance of the Study ...................................................................................... 7
1.8. Definitions of Key Terms ............................................................................................. 8
1.9. Outline of the Study ................................................................................................... 10
Chapter Two ............................................................................................................................... 11
2. Literature Review ............................................................................................................... 11
2.1. Introduction ................................................................................................................ 11
2.2. Theoretical Background ............................................................................................ 11
2.2.1 Traditional and Alternative Assessment .................................................................... 13
2.2.2. Alternative Assessment Tools .................................................................................... 17
2.3. Empirical Studies ....................................................................................................... 20
2.3.1 Previous Research on the Effect of Alternative Assessment Methods on Students’
Test Performance................................................................................................................. 21
2.3.2 Students’ Attitudes Towards Alternative Assessment Methods ................................ 29
2.4. Summary .......................................................................................................................... 34
Chapter Three ............................................................................................................................ 36
3. Methodology ........................................................................................................................... 36
3.1. Introduction ..................................................................................................................... 36
V
3.2. Research Design............................................................................................................... 36
3.3. Data Collection Instruments........................................................................................... 37
3.3.1. Quasi-Experimental Design ...................................................................................... 38
3.3.2. Students’ Questionnaire ............................................................................................ 42
3.4. Population, Sample, and Sampling ................................................................................ 43
3.5. Validity and Reliability ................................................................................................... 45
3.5.1. The Pilot Study .......................................................................................................... 46
3.6. Data Collection Procedures ............................................................................................ 47
3.7. Data analysis .................................................................................................................... 49
3.8. Ethical considerations ..................................................................................................... 49
3.9. Summary .......................................................................................................................... 50
4. Results and Discussion ........................................................................................................... 51
4.1. Introduction ..................................................................................................................... 51
4.2. Results .............................................................................................................................. 51
4.2.1 The Quasi-Experimental Results ............................................................................... 51
4.2.2 Students’ Questionnaire Results ................................................................................ 52
4.3. Discussion ......................................................................................................................... 67
4.3.1. The Effect of Using Alternative Assessment Methods on Second-Grade Secondary
School Students’ Test Performance .................................................................................... 67
4.3.2. Students’ Attitudes Towards Using Alternative Assessment Methods ..................... 69
Chapter Five ............................................................................................................................... 73
5. Conclusion ............................................................................................................................... 73
5.1. Introduction ..................................................................................................................... 73
5.2. Summary of the Study .................................................................................................... 73
5.3. Pedagogical Implications ................................................................................................ 74
5.4. Study Limitations ............................................................................................................ 75
5.5. Recommendations for Further Research ...................................................................... 75
REFERENCES ............................................................................................................................. 77
Appendices .................................................................................................................................. 91
VI
List of Tables
VII
List of Abbreviations
Abbreviation Meaning
AA Alternative Assessment
TA Traditional Assessment
VIII
The Effect of Using Alternative Assessment Methods on EFL Students’
Test Performance and Their Attitudes Towards These Methods at a
Secondary School in Misurata
Abstract
Alternative assessment refers to different methods and strategies used to assess students
that differ from traditional assessments like multiple-choice tests. This study investigates
at Misurata Secondary School for boys in Misurata, Libya. The study also explores
students' attitudes towards these assessment tools. A quantitative approach with a quasi-
experimental design and a questionnaire was employed. The findings show that
alternative assessment (AA) methods did not significantly affect students' testing
positive attitudes towards both alternative and traditional assessment (TA) methods, with
a slightly higher preference for AA. These findings suggest that AA methods are
generally well-received by students, even though they may not have a significant effect
on test performance.
IX
هستخلص البحث
انرمٛٛى انثذٚم ٚشٛش إنٗ يجًٕعح يخرهفح يٍ األعانٛة ٔاالعرشاذٛجٛاخ انًغرخذيح نرمٛٛى انطالب ٔانر ٙذخرهف عٍ
انرمًٛٛاخ انرمهٛذٚح يثم االخرثاساخ راخ االخرٛاساخ انًرعذدجٓٚ .ذف ْزا انثحث إنٗ دساعح ذأثٛش طشق انرمٛٛى انثذٚهح
عهٗ أداء االخرثاس نـ 65طانثًا ف ٙانصف انثاَ ٙف ٙانمغى انعهً ٙف ٙفصٕل انهغح اإلَجهٛضٚح كهغح أجُثٛح ف ٙيذسعح
يصشاذح انثإَٚح نهثُ ٍٛف ٙيصشاذح ،نٛثٛا .كًا ٓٚذف انثحث إنٗ اعركشاف آساء انطالب ذجاِ ْزِ األدٔاخ انرمًٛٛٛح.
ذى اعرخذاو َٓج كً ٙترصًٛى شثّ ذجشٚثٔ ٙاعرثاَح فْ ٙزا انثحث .ذشٛش انُرائج إنٗ أٌ طشق انرمٛٛى انثذٚهح نى ذؤثش
تشكم كثٛش عهٗ أداء االخرثاساخ نهطالب يماسَح تاألعانٛة انرمًٛٛٛح انرمهٛذٚحٔ .يع رنكٚ ،حًم انطالب عًٕيًا آساء
إٚجاتٛح ذجاِ كم يٍ طشق انرمٛٛى انثذٚهح ٔانرمهٛذٚح ،يع ذفضٛم طفٛف نهرمٛٛى انثذٚم .ذشٛش ْزِ انُرائج إنٗ أٌ طشق
انرمٛٛى انثذٚهح عًٕيًا يمثٕنح تشكم جٛذ يٍ لثم انطالب ،عهٗ انشغى يٍ عذو ٔجٕد ذأثٛش كثٛش عهٗ أداء االخرثاس.
انكهًاخ انشئٛغٛح :انرمٛٛى ،انرمٛٛى انثذٚم ،انرمٛٛى انرمهٛذ٘ ،أداء االخرثاس ،فصٕل انهغح اإلَجهٛضٚح كهغح أجُثٛح
X
Chapter One
1. Introduction
learn in education. This tradition has been used to inform educational practices and
policies, and this process provides information that can be used to make future decisions.
knowledge, motivation and performance, then used to make informed decisions (Allan,
1999; Ekbatani& Pierson, 2000; Lambert & Lines, 2000; as cited in Tosuncuoglu, 2018).
Assessment is also essential for students because it helps them maintain their
learning progress and enhances their motivation to learn the language actively
(Phongsirikul, 2018). Wojtczak (2002) explained that once the students know their
strengths and weaknesses, they can design a plan to overcome any language-related
problems they may have. This encourages and motivates them to keep learning and
different methods to analyze students‘ performance (Norova & Haydarali, 2021). One
measurable data are collected about some aspects of reality in the form of a pencil-and-
administered (Lustgarten, 2022). This method implies that students are primarily focused
1
on memorizing information to perform well on the test, rather than actively constructing
knowledge in a way that aligns with their preferred mode of assessment (Letina, 2015).
The prevalence of TA in the classroom can lead to a situation where teachers prioritize
can lead to negative emotions, perceptions, and memories, which can adversely affect
students' performance. Students may experience anxiety, stress, and burnout when faced
with TA tools. As a result, Mansory (2020) highlights that researchers and educators
have recognized the need for AA tools that go beyond traditional tests. The aim is to
meet students' expectations, enhance their motivation, reduce test anxiety, and improve
their overall performance. Language testers have also acknowledged the limitations of
have begun exploring innovative methods to assessment (Rezaee et al., 2013). This
approach, known as AA, offers a promising solution to address the challenges associated
situations. The TA tools can have the elements of surprise, competition, enjoyment, and
fun (Mansory, 2020). Mansory also stated that teachers could use various non-traditional
assessment techniques to assess students‘ learning. These may include projects, concept
2
assessments, observation, drama, journals, diagnostic trees, teacher and student
comfortable and practical classroom learning atmosphere. If new methods are not
developed and adopted, students can become disinterested in learning the language,
which would ultimately defeat the purpose of the language learning program.
educational system, requiring not only a shift in teaching and learning but also a
assumption of a desire for reform, or at the very least, an intention for it. Elmahjoub
argues that despite the implementation of certain changes in the early 2000s, that is
approach, assessment is often overlooked in the Libyan education system as it has not
been given adequate attention and is still only utilized as a tool to determine the overall
students‘ dependence on memorization, the increased emphasis on final exams, and the
prioritization of precise information recall. The system is largely centered around the
autonomy: the freedom to inquire, the freedom to articulate their thoughts, and the
2007).
3
Several researchers (Alhmali, 2007; Aldabbus, 2008; Orafi & Borg 2009;
Shihiba, 2011; Tantani, 2012, Onaiba 2013) have revealed that the majority of Libyan
the learning process. Additionally, they seem to face difficulties in applying their
Despite the fact that one of the overarching aims of education in Libya is to
facilitate new forms of learning and empower students to explore their potential and
evident that Libyan English teachers in secondary schools are not effectively
that the incompetence of teaching and assessment methods used in Libyan schools could
involves students preparing for a mid-term and a final exam for every subject they study.
Regretfully, most of the time, students are not given the opportunity to receive feedback
on their exams, which means they are unaware of their strengths and weaknesses.
Consequently, teachers are unable to track their students' progress effectively, and their
performance is evaluated based on criteria that are arbitrary rather than explicit (Al-
Fourganee, 2018).
4
1.2. Statement of the Problem
In the Libyan context, some EFL teachers continue to use outdated teaching
enhance classroom learning (Shihiba, 2011). This is in accordance with Grada's (2014)
claim that students in Libya are still relying on their teachers and take a passive role in
the classroom. This conflicts with modern assessment principles and new teaching
their own learning process and encouraged to take ownership of their learning.
In accordance with these authors, Onaiba, (2014) has reported that most of the
to TA, solely introduced to students in the form of pen and paper tests. Onaiba (ibid) also
stated that none of the AA practices had been deployed by teachers when assessing their
students. This means little or no is known about such deployment, particularly its effects
on students' test performance, which inspired the researcher to carry out this study.
fails to accurately measure students' English proficiency or their ability to effectively use
English in real-life contexts. Ramadan and Dekheel (2020) noted that there is a prevalent
discontent among both students and educators regarding the use of exams as a valid
method for assessing students‘ performances. This dissatisfaction stems from the
observation that students may achieve high scores on exams despite having a limited
understanding of the material, known as test score pollution. Additionally, some argue
5
that assessing students based on a limited number of exams may not provide an accurate
In line with the previous study, the researcher of the current study, along with
other EFL Libyan teachers, has observed that many students perform poorly in their
exams. This may be due to various factors, one of which is that students are assessed
using only one method, namely TA through pen-and-paper tests. This limitation prevents
students from compensating for any gaps in their knowledge or skills that may have
arisen during the assessment. As a result, this study was conducted to explore the
potential effect of AA tools in bridging the gap where TA methods have been
insufficient.
Goodarzi, (2014); Heidarian, (2016); Aliasin & Amanlu, (2017); Kafipour &
Khoshnood, (2023), AA tools have yet to prove effective in Libyan EFL classrooms due
to absence of its implementation in this context. Thus, this study aimed to (1) explore the
effect of AA methods on students‘ testing performances in Libyan EFL classes; and (2)
investigate students‘ attitudes towards AA tools. A thorough investigation into these aims
educational progress. The following are the research hypotheses and questions addressed
in this study.
In order to achieve the aims of the study, the following hypotheses have been set:
6
Null Hypothesis (H0): Alternative Assessment methods have no effect on
To accomplish all the aims mentioned previously, the following questions were
tackled:
methods?
second-grade students in the scientific section at Misurata Secondary School for boys in
Misurata, Libya. The research was conducted in the second semester of the academic
questionnaire with closed-ended items. So, this study is limited to the above stated aims,
The current study gained its significance because, to the best of the researcher‘s
knowledge, no previous empirical research has been conducted in the context of the
7
study used both a quasi-experiment method and a questionnaire to explore the effect AA
of AA tools on students' learning outcomes using questionnaires and surveys. Only a few
studies, among which is the current study, have used a mixed-method approach to
Hopefully, the study would help teachers opt for the appropriate assessment tools
that motivate students to be more involved in their learning and to gain better learning
outcomes. Moreover, it would help teachers improve their teaching materials and
practices into their teaching, but there needs to be more knowledge about how to do so
student performance accurately. Thus, this research is needed to explore the effect of
skills, and based on these specific instances, the test administrator makes
objective of making informed decisions. On the other hand, Brown and Green
8
(2006) describe assessment as a means of examining the learning process,
implemented within the instructional setting and seamlessly integrated into the
2020).
suggests that students acquire information to be utilized for test purposes rather
Attitudes according to Bizer et al., (2006) are considered essential in the field of
appealing or unappealing.
9
1.9. Outline of the Study
This study is structured into five chapters. Firstly, this chapter has provided a
research aims, research questions and hypothesis, scope, aims, significance and
definitions of key terms. Chapter Two reviews related literature on the effect of different
various assessment tools. Chapter Three details the methodology used in the study,
procedures, data analysis, and ethical considerations. In Chapter Four, the data analysis
framework and procedures are outlined for analyzing both the quasi-experimental data
and the questionnaire data in a quantitative format. The chapter provides details on the
specific methods and techniques employed to analyze the collected data. Additionally,
the chapter provides a discussion of the results obtained from the data analysis. Finally,
10
Chapter Two
2. Literature Review
2.1. Introduction
problem (Research guide, 2023). It allows them to situate each study within the broader
examining prior works, researchers can identify gaps in knowledge, build upon existing
findings, and ensure their study addresses novel aspects or fills knowledge gaps in the
field. This chapter is structured into two primary sections: Theoretical Background and
assessment, examining its definition, purpose, and significance within the field of study.
section focuses on reviewing and summarizing existing published research and scholarly
language learning, as it strives to enhance the abilities of both educators and those being
11
Brown (2004) defined assessment as ―an ongoing process that encompasses a
tries out a new word or structure, the teacher subconsciously assesses the student‘s
Statement for Grades R-9 Schools, (2002), is seen as an ongoing, planned procedure of
collecting data on the progress of students evaluated against the assessment criteria of
Although numerous learning theories and models have been developed in the last
few decades, there has been minimal effort to reform assessment practices in Libyan
Most of the adopted frameworks emphasized designing the curriculum and teaching
methodologies while neglecting the assessment aspect. (Shinn, 2012; Suleiman &
students towards the end of a unit or course in order to rank them. Instead, it is regarded
as an integral element of teaching and learning (Khattri & Sweet, 1996; Nasab, 2015).
Several factors have contributed to the change that different perspectives on assessment
have undergone. The first factor is associated with the comprehensive perspective of
with the diverse facets of that language (Moqbel, 2020). In this regard, Hamayan (1995,
12
as cited in Moqbel, 2020) claimed that AA can mirror the comprehensive view of
language since its procedures are established on the concept that the various facets of
The second factor is related to the performance and proficiency of the language,
which largely hinges on the objective of language usage and the setting in which it is
utilized (Moqbel, 2020). Hence, offering students assessment prospects reflecting the
Coombe, 2000).
Birenbaum and Dochy (1996) and Segers et al., (2003) have characterized this
enhance learning (Dochy & McDowell, 1997; as cited in Moqbel, 2020). Assessment in
this context is viewed as a tool for learning and an integral component of classroom
activities, and it should take place simultaneously with learning activities (Butler, 1997;
must be aware of the extent to which the learners have grasped the subject matter
regardless of the assessment type (Nasab, 2015). Broadly speaking, there are two
which has been in use for a longer period, typically involves testing practices that gather
factual and quantifiable information about a particular subject (Fulcher, 2010). Nasab
13
(2015) also defined TA as a standardized test consisting of multiple choice, short answer,
important to acknowledge that they do possess advantages. One of the key strengths of
standardized tests and other forms of multiple-choice assessments (Law & Eckes, 2007).
Traditional tests, including standardized ones, are often designed with well-defined
scoring criteria and established norms, allowing for consistent and unbiased assessment
of student performance. Dikli (2003) explained that TA methods aim to inspire students
by offering them external rewards and favorable outcomes. This approach seeks to
environments, where students are frequently given tests that address limited portions of
the limited benefits of detailed and descriptive feedback regarding outcomes and
2015). Lustgarten (2022) highlights the criticisms frequently directed at TA, particularly
14
world relevance. These assessments often prioritize rote memorization and reproduction
classroom has led to the emergence of AA, which can overcome the limitations of TA by
providing opportunities for EFL students to utilize language in practical situations and
measure a subject in a less objective manner. AA, also known as formative assessment,
students (Aliasin & Amanlu, 2017; Barlow & Coombe, 2000; Gielen et al., 2003;
Hamayan, 1995, as cited in Moqbel, 2020). This assessment process not only measures
academic performance but also assesses student drive, involvement, and approach
examination score (Etobro, 2022). On the other hand, academic achievement refers to
the educational outcomes that students attain after completing a specific academic
program within a defined timeframe. It represents the level of success achieved at the
15
vital role in accurately assessing students' abilities and identifying their strengths and
information instead of selecting the right answer (Nasab, 2015). Consequently, educators
can obtain valuable insights into how to revise future lesson plans and address any
childhood and adolescence, can greatly profit from the benefits offered by alternative
flexible groupings in diverse classrooms, and provide opportunities for support and
students from different backgrounds can achieve success at their own pace and receive
Despite the advantages linked with AA, there are still reservations about this
scale. On the teacher's side, a more significant amount of planning and well-thought-out
goals are necessary to be provided to the students (Nasab, 2015). Nasab criticizes this
method noting that the AA tasks require a lot of effort from the teacher to have a well-
16
Despite this, AA offers indisputable benefits over conventional testing
also considering their cultural background and existing knowledge levels (Chirimbu,
2013).
Within the classroom, there are diverse learning styles, which has led to students
(Mohammed & Hayder, 2022). As a result, several investigations must be carried out to
2022), notes that the following AA tools are frequently employed to assess students:
without a specific framework in mind. On the other hand, structured situations are
created by teachers to provide a context where specific skills are more likely to be
utilized. The grades assigned to students often reflect the number of competencies they
have successfully demonstrated. This approach facilitates the collection and reporting of
perform specific tasks. The benefit of this new tool is that it enables a rapid gathering of
17
data, yet the assessment may not be entirely reliable. Typically, the reported scores
techniques.
over their own fate. A journal is a record or log of an individual's feelings, sentiments,
necessitates careful forethought and design, while the teacher monitors the students'
ability to finish the tasks. Nevertheless, as it only measures a limited set of skills, it is a
presenting examples of their work that showcase their skills. The advantage of such an
student's work, as very few students are capable of producing entirely authentic written
18
Assignments. An assignment refers to tasks given to pupils by their educators to
finish outside the classroom. Typical tasks assigned for homework may comprise a
abilities to be practiced.
writing-focused classes, have been a common practice for many years. These
assessment. Carefully formulated questions ensure reliable responses. In this case, the
teacher interviews the student for a specific assessment purpose (not referring to a
(Brown, 2004).
Peer and Self Assessments. Peer assessment and self-assessment are widely used
AA tools in education. Peer assessment involves students assessing the work of their
peers based on criteria and standards, providing feedback and making suggestions for
presentations, written work, tests, and portfolios. Self-assessment, on the other hand,
involves students recognizing and assessing their own performance, aiming to produce
19
reflect on their progress and assess their language abilities. Self-assessment provides
opportunities for learners to monitor their own learning and gauge their advancement.
Concept Map. As stated by Novak & Gowin (1984, as cited in Nasri et al.,
concepts or notions and their interconnections. The concepts are contained in circles or
boxes and are connected to other concepts through lines or arrows. Arrows are utilized to
indicate the direction of the connection while lines signify the flow of the relationship,
either from top to bottom or horizontally, originating from the primary concept. The
crucial aspect of a concept map is that the concepts are organized in a hierarchical
manner.
All in all, AA tools serve different purposes and objectives. They can help in
determining training requirements and selecting the optimal educational materials. They
are beneficial in tracking advancements and measuring the degree to which course goals
have been attained. Moreover, they are applicable in various foreign language programs,
For the purpose and the scope of this study, the researcher utilized several AA
tools and techniques. These included tasks such as projects, concept mapping,
already conducted on the research topic, highlighting key findings and methodologies.
20
2.3.1 Previous Research on the Effect of Alternative Assessment Methods on
assessment has led to several positive outcomes in teachers‘ practices in general and
students‘ learning in particular (Borko et al., 1993; Koretz et al., 1993; Stecher and
Mitchell, 1995; Newman et al., 1996; Brown, 2004; El-Koumy, 2009; Baniabdelrahman,
2010; Safa & Goodarzi, 2014; Heidarian, 2016; Aliasin & Amanlu, 2017; Kafipour &
Khoshnood, 2023).
In Safa and Goodarzi‘s (2014) study, for instance, the researchers used Task-
grammar skills. Both a control and an experimental group underwent identical grammar
instruction for ten sessions. However, the experimental group was given a TBLA
grammar assessment every three sessions, while the control group took traditional
grammar tests. The same grammar test was given to both groups as the post-test
following the treatment, just as it was administered as the pre-test. The analysis of the
post-test and the t-test comparing group means revealed a significant difference between
the two groups, indicating that the students in the experimental group experienced a
assessment.
Despite the positive effect observed in the previous study, it has a limited focus
on specific language skills i.e., grammar. By solely examining this skill, the findings
may not provide a comprehensive understanding of the overall language proficiency and
21
of the present study addressed this limitation by extending the focus to include different
language aspects in students' testing performances. This broader scope allows for a more
development.
Upon reviewing the existing literature, it was found that only a limited number of
Aliasin and Amanlu (2017) and Baniabdelrahman (2010) conducted studies to highlight
skills.
In Aliasin and Amanlu's study, the focus was on Iranian EFL Learners‘ reading
tasks targeting different aspects of reading comprehension skills, while the control group
received TA methods. The participants' performance was assessed using two equivalent
reading comprehension tests, administered as pretest and posttest. The findings showed
total of 136 students (67 male and 69 female). A reading test was created and validated
for data collection. Student self-assessment was employed using one-minute papers and
rating-scale sheets to gauge their progress. The study's results indicated a positive impact
22
While the studies mentioned in the previous context focused specifically on
other language skills can provide a more comprehensive understanding of their effect on
and familiar format. It provided a reliable means to assess the effect of the intervention
on students' reading skills and determine whether any significant changes occurred as a
result of the treatment. According to Sanders (2019), both pre and post-test
administrations should utilize the same assessment instrument, but with different
versions or forms that are comparable. This ensures that the results obtained from the
administrations can be compared, while also ensuring that the student does not take the
exact same test twice. In the same manner, the researcher in the current study employed
a consistent form of traditional pre and post-tests to assess the effect of the intervention
Beyond the realm of English Education, the field of general education has also
knowledge, think critically, and solve problems. Researchers and educators in general
education (Newmann et al., 1996; Kırıkkaya & Vurkaya, 2011; Öztürk & Şahin, 2014;
Pereira et al., 2017; Ayyoub et al., 2021; Lustgarten, 2022) have investigated the effect
23
In a study conducted by Öztürk and Şahin (2014), the researchers investigated
the effect of assessment and evaluation activities, including learner journals, self-
different schools in Çanakkale, as well as two control groups comprising 69 pupils. The
researcher collected data using various scales, such as the 5th-grade Mathematics
Scale, and Mathematics Self-efficacy Scales. The findings of the study indicated that the
evaluation methods.
Fostoria City Schools in Ohio with a mixed-method research approach. This study was
end of the three-week physics unit, students were assessed using both alternative and
traditional assessment methods each on a separate day. The results showed that a
24
Although the studies mentioned have made valuable contributions to
some concerns have been raised. In Lustgarten‘s study, for instance, the study does not
mention the inclusion of a control group for comparison. Without a control group, it
AA were solely due to the assessment procedure or other factors such as teaching
relying solely on a quasi-experimental design for data collection in the study conducted
by Öztürk and Şahin (2014) may limit the depth of insights gained. While quasi-
experimental designs have their merits, incorporating additional data collection tools can
By shifting the focus to EFL classes and including different language aspects, the
researcher of the current study addressed a broader scope of language proficiency and
the study allowed for the collection of quantitative data on students' attitudes, providing
insights into how these assessment methods are perceived and experienced by the
students themselves.
It is worth noting that the study has also utilized an experimental group and a
control group, which helps address the limitations mentioned in Lustgarten's study. By
including a control group, the study allowed for a comparison between the effects of AA
methods and TA methods. The control group provides a baseline against which the
25
experimental group can be compared, helping to determine the specific effect of the AA
It is worth bringing to attention that not all studies yielded positive outcomes
United States. The researcher in the study conducted three units of instruction for the
students and documented the process in a journal. The teaching methods used by the
researcher were the same for both the control and experimental groups, covering the
required vocabulary and grammar goals established by the department at the school. The
primary difference between the groups was in the assessment methods. Both the control
and experimental groups took the department's traditional, summative final exam.
However, the experimental group underwent a dynamic assessment for unit one, a TBLA
for unit two, and a formative assessment involving peer and self-evaluations through
writing assignments for unit three. while the control group solely took traditional exams.
The results revealed that there was no significant difference in the scores of the
experimental group‘s final exam compared to the scores of the final exam of the control
group.
Bachelor (2017) did not elaborate on this particular finding. Thus, one may argue
that the absence of a significant difference in the final exam scores between the
experimental and control groups could potentially be attributed to the students' lack of
familiarity with the AA methods used in the experimental group. In other words, the
students may not have had enough exposure or experience with the specific AA
26
employed as the final assessment method. To address this limitation, the researcher in
the current study took steps to mitigate the potential effect of students' unfamiliarity with
course for the experimental group. This approach provided students with repeated
exposure and practice with AA method, allowing them to become more familiar and
Additionally, both the experimental and control groups were assessed in the pre
and post-tests using TA methods. By including TA for both groups, the researcher of the
present study ensured that all students had a common baseline of experience with TA
methods. This helped to establish a fair comparison between the groups and mitigate the
multiple AA methods throughout the course and including traditional tests as pre and
post-tests for both groups, the researcher aimed to enhance the validity and reliability of
the study's findings while addressing the limitation of students' unfamiliarity with AA
methods.
Sahera et al., (2022) conducted a study to assess the effectiveness of using both
authentic and TA in the field of health and educational sciences at Amman Arab
students enrolled in two introductory psychology courses that were similar in difficulty
and content. The two courses utilized different types of assessments: traditional
assessments and authentic assessments. The participants in the study were the same
students who took both courses. The researchers compared the final grades of the
students in both courses to assess the effect of the assessment type on their overall
27
performance. The analysis of students' academic achievements using traditional and
Specifically, students achieved higher grades in the traditional exam compared to their
The key difference between Sahera et al.'s study and the researcher of the current
study lies in the presence of a control group in the latter, which allows for a more robust
comparison and assessment of the effect of the AA methods. Both studies (Bachelor,
2017; Sahera et al., 2022) did not extensively discuss potential confounding factors that
may have influenced the results. Factors such as teaching methods and students‘ prior
knowledge and skills, could have affected the outcomes but were not explicitly
addressed. The researcher of the current study took steps to address potential
confounding factors. By ensuring that both the experimental and control groups received
prior knowledge, and skills, improve the internal validity of the study. It strengthens the
researcher's ability to attribute any differences in the outcomes to the AA method rather
The purpose of the current study was to contribute to the existing body of
students' skills, engagement, motivation, and academic achievement. This study sought
students' testing performances in Libyan EFL classrooms. To accomplish this, the study
28
employed quasi-experimental designs and administered a questionnaire to gather
students' perspectives on these AA methods. By doing so, the study aimed to fill the gap
in the literature and provide valuable insights into the effect of AA method in the Libyan
EFL context.
A number of studies highlight the positive attitudes that students have towards
AA methods. They suggest that these approaches can enhance student engagement,
Fajarsari (2016) conducted a study at Satya Wacana Christian High School in Salatiga,
Central Java, Indonesia. This research utilized questionnaires to collect data from the
students. The findings revealed that the majority of students held positive attitudes
towards AA. They recognized the importance of AA in improving their English language
Moreover, Bachelor‘s (2017) study which was reviewed in the previous section
also investigated how the study participants perceived three distinct alternative
Assessment utilizing self- and peer assessments. The results indicated that students in
the experimental group held favorable views towards the AA types, with Formative
Assessment being their preferred choice and TBAL being the least favored option.
29
Students' attitudes towards TA and AA methods as tools for assessing their
questionnaire and conducted interviews. The participants in this study were students in
generally similar in terms of their responses and opinions. Moreover, most students
expressed positive views towards both TA and AA. However, the AA method was found
A more recent study conducted by Lustgarten (2022), which was also reviewed
AA. The study was conducted in a physical science classroom with 21 students, and a
Likert-scale survey was used to collect qualitative data on students' attitudes regarding
both types of assessments. The results of the survey indicated that the majority of
students expressed a preference for AA. They reported experiencing less stress or
pressure when engaging with AA compared to TA. Additionally, students believed that
they gained a greater amount of learning from the AA methods. However, it's important
to note that the study's sample size was relatively small, consisting of only 21 students in
one specific classroom. This limited sample size may affect the generalizability of the
favorable outlook towards peer assessment i.e., one of the AA tools. For instance, Peng
(2009) found that students at both high and low-intermediate levels reacted positively to
30
peer assessment and their attitudes improved after experiencing this type of assessment.
assessment techniques in foreign language classrooms. The findings indicated that the
classrooms. In the same vein, Burnaz (2011) contended that Turkish university learners
assessment because they believe that the latter involves time pressure, promotes
memorization, and does not effectively measure their speaking skills in English.
students' attitudes towards it. Likewise, Cornelius and Kinghorn (2014) reported that
first-year Japanese university learners of EFL held favorable attitudes towards self and
peer assessment.
2018). In these studies, students' attitudes towards different AA methods were examined
and the findings indicated that some students preferred traditional testing methods,
believed that traditional exams were a more efficient way to assess their knowledge. The
preference for TA was associated with feeling comfortable in the test environment,
having confidence in their ability to succeed in the test, and perceiving the test as
31
Caner (2010) conducted a study that focused on the investigation of students'
attitudes towards AA through the use of portfolios. Caner examined the attitudes of 140
conducted both with the participants and their instructors. Two open-ended questions
were also included as part of the research methodology. The findings of the study
indicated that, in general, the participants exhibited a preference for traditional pen and
believed that portfolio assessment contributed to their English learning process. The
research also revealed that some of the participants held negative attitudes towards
perspectives on portfolio assessment in their writing courses. The findings may not
students' attitudes towards portfolio assessment could differ when applied to different
towards portfolio assessment across various subject areas, future research should explore
the perspectives of students from different disciplines and skill areas. This would allow
for a broader understanding of the benefits, challenges, and preferences associated with
regarding traditional English language testing. A survey of 323 randomly selected Thai
32
freshmen from various programs was conducted. An open-ended questionnaire with four
main concerns—test fairness, test format, test content validity, and learning
opportunities from the tests—was used to collect data. The results of the study indicated
that the testing practices (i.e., traditional testing) were considered to be adequate and fair
for measuring students' English language ability. The test formats were generally
perceived as a reflection of the students' English language ability, although there was a
need for more diverse and innovative item types that could be suitable for testing all
language skills. Additionally, the current tests were viewed as being capable of
The researcher of the present study has noted that Jaturapitakkul‘s study focused
solely on the attitudes of traditional testing practices without eliciting students‘ attitudes
relative strengths and weaknesses of traditional testing and how it compares to other
comprehensive insights into the attitudes and preferences regarding different assessment
methods by highlighting the attitudes towards both TA and AA methods. This would
enhance the validity and applicability of the findings, as it allows for a more nuanced
In the context of the current study, Ramadan and Dekheel (2020) conducted a
students' attitudes towards assessment methods. The study‘s findings showcased that
those students generally preferred AA methods and showed a preference for modes of
33
throughout the course. The study has a limitation in that it solely relied on a
questionnaire as the research method for data collection. Other research methods, such
as interviews or focus groups, could have provided additional insights and a richer
2.4. Summary
This chapter has provided a detailed investigation of previous studies that are
relevant to the topic of assessment and the effect of AA method on students' test
assessment practices and their effects on students' academic performance. These studies
explored various aspects such as the use of AA method, the influence of these methods
The current study expanded on the existing literature by investigating the effect
students in Libyan EFL classes. It added to the knowledge base by focusing on the
Similar to some previous studies, this research also examined EFL classes.
design with control and experimental groups, whereas some previous studies only used a
34
single experimental group. Additionally, the current study incorporated a questionnaire
and academic achievement, while the present study specifically focused on the effect of
language skills. Thus, reviewing pertinent literature has beneficially contributed to the
choice and selection of the research design and methodology which is explicated in the
upcoming chapter.
35
Chapter Three
3. Methodology
3.1. Introduction
study and its structure. This includes the techniques employed for data collection and
analysis. The study instruments are also included and were formulated to generate a
experimental method and a questionnaire. Apart from the ethical concerns implicated in
the investigation and the validity and reliability of the research tools, a description of the
The study employed a quantitative research design, and it aimed to examine the
research hypotheses (see 1.4.) and answer the research questions (see 1.5.). Cohen et
al.‘s (2007) definition of quantitative research involves the use of empirical methods and
al. (2007), refers to a descriptive statement about the way things are in the real world, as
approach that aims to explain social phenomena by collecting numerical data, which are
In other words, quantitative research involves the use of empirical data to understand
and explain social phenomena, with an emphasis on numerical data and statistical
36
In deterministic research, as Creswell (2014) stated, the focus is on examining
the relationships between variables and understanding the extent to which changes in
one variable may cause changes in the other variable. This approach is often used in
dependent variables, researchers can draw conclusions about the causal relationships
between variables. Thus, a quasi-experiment and a questionnaire were used in this study
and to increase the validity and reliability of the results. Both instruments are discussed
in details below.
This section describes the data collection tools that were utilized in the current
study. The study employed two methods for data collection which are quasi-experiment
responses to the questionnaire designed to elicit the students' attitudes towards the use of
is often impractical in school settings due to preexisting and intact structures such as
classes and grade levels. Therefore, quasi-experimental designs are commonly used in
research conducted within school environments (Edmonds & Kennedy, 2016, as cited in
Sawad & Law, 2023). Additionally, the quasi-experimental research design was also
chosen for this study due to ―the inability to control for all variables in a middle school
setting‖ (Rivera, 2015, p.73). While it was not possible to control all variables, the
37
researcher aimed to address this limitation by incorporating a comparison group. By
doing so, the researcher anticipated that any main effects resulting from the
uncontrollable variables would affect both the experimental groups and the comparison
participants in distinct groups for the experiment, which is either randomized or non-
research design often used when a randomized controlled trial (RCT) is not feasible or
participants to different groups, but instead uses existing groups or naturally occurring
pre-test and post-test for both the control and treatment group is known as the
nonequivalent group pretest-posttest design (Campbell & Stanley, 1963). In this design,
participants were not randomly assigned to the control and treatment groups, and the
groups were not equivalent at the outset of the study. However, both groups were tested
before and after the intervention or treatment, allowing for a comparison of changes in
the dependent variable over time (Campbell & Stanley, 1963). This design can help
control for internal validity threats, such as history, maturation, and regression to the
38
mean. However, it is still subject to other threats to internal validity, such as selection
baseline prior to a research intervention or treatment (Kim & Willson, 2010). Pretesting
can also be used to group subjects based on their pretest scores, a technique known as
pretesting can help plan instruction by identifying students' current level of knowledge
or areas where they may need additional support or resources (Kim & Willson, 2010).
six randomly assigned classes of second-year students. The test consisted of five
different tasks, including true or false, multiple-choice, and matching questions, as well
as two essay tasks that required students to complete a grammar rule and provide
definitions for given words (See Appendix A). The pretest was timed, by giving students
two hours to complete the tasks. Following the administration of the traditional pretest,
the papers were manually scored and the results were subjected to a paired sample t-test
to identify two groups with similar academic levels among the six classes.
The researcher carefully selected two classes that were closely matched in terms
covariance was performed. The results indicate that the reported p-value of 0.070, which
is greater than 0.05, suggests there is no statistically significant difference between the
two groups. This approach was taken to minimize heterogeneity between the groups and
ensure that any observed differences in the results could be attributed to the intervention
39
Treatment. The treatment refers to the independent variable that is manipulated
by the researchers in an experiment (Kenny, 1975). In this study, after the administration
of the traditional pretest, the two groups were taught traditionally in unit 5 and 6 of their
course and work books. The instruction spanned across four weekly sessions, with each
session lasting approximately 45 minutes. The control group was assessed traditionally
i.e., pen and paper tests, while the experimental group was exposed to a treatment that
involved the use of various AA methods and techniques. These included presentations,
writing, reading and comprehension tasks, role plays, and observations (See Appendix
B). The scoring of these AA tools was done using rubrics since they provide clear and
assess students' work. It consists of a clear and organized set of criteria that describes the
different levels of performance quality expected on each criterion (Brookhart & Chen,
2014). Rubrics are often used in alternative forms of assessment and can be applied to
performances.
students to engage in self-assessment, facilitating feedback from teachers and peers, and
guiding students on how to improve their work. However, these purposes can only be
achieved if the rubric includes clear criteria and detailed descriptions of the different
performance levels. This has been highlighted by studies conducted by Andrade (2000)
40
In this study, the researcher employed varying rubrics which were adopted online
and were appropriate for different AA methods used to assess students throughout the
course. These rubrics consisted of clear criteria and detailed descriptions of different
understanding of their areas of strengths and weaknesses, the expectations set forth for
intervention, treatment, or other condition to measure any changes that have occurred
(APA Dictionary of Psychology, n.d.). Posttests are commonly used in research settings
along with pretests to isolate the effects of a variable of interest (Cohen et al, 2007). The
use of posttests in conjunction with pretests helps to strengthen the validity of the
research findings and provides a more comprehensive understanding of the effects of the
After the intervention occurred in this study, a posttest was administered to both
the experimental and control groups. The post-test was a two-hour traditional test that
included four different tasks, such as true or false, multiple-choice, essay questions with
three sections, and matching questions (See Appendix C). Both pre and post-tests were
checked by two Ph.D. holders and a secondary school English inspector to ensure the
tests‘ validity and reliability. To prevent subjectivity and bias, the papers were evaluated
by multiple English secondary school teachers, including the researcher. The results
were then analyzed by a statistician using a paired sample t-test to compare the
41
3.3.2. Students’ Questionnaire
As Mackey and Gass (2016) stated questionnaires are written instruments that present
options from a series of statements. Mackey and Gass further explained that
Williams (2003) noted that questionnaires have a wide range of uses in social
science research, and can be used to collect data on a variety of topics, including health,
education, social attitudes, and consumer behavior. They can be designed to elicit both
quantitative and qualitative data and can be tailored to specific research questions or
hypotheses.
attitudes towards the use of AA method within the classroom. It is important to ensure
that all participants in the experimental group study had exposure to these AA tools
during the experiment and before taking part in the questionnaire, which contributes to
the reliability and trustworthiness of their attitudes. By being familiar with these
assessment methods, the students were able to provide informed and knowledgeable
attitudes on the use of AA within the classroom setting. The questionnaire consists of 15
statements written in English and translated into Arabic, the students‘ native language,
by the researcher. Furthermore, two Ph.D. holders and a secondary school English
42
inspector reviewed the statements, made necessary revisions to correct any grammatical
errors, and ensured the accuracy of the questionnaire's content during the validation
process. The questionnaire, which was adopted from Irawan‘s study (2017), and
modified according to the aim of the study, presented respondents with a five-point scale
ranging from "Strongly agree" to "Strongly disagree," with each response option
into three main themes, the first theme was 7 statements about assessment in general, the
second theme was 3 statements about traditional assessment and the third theme was five
statements about alternative assessment (See Appendix D). The information gathered
from the questionnaire was subjected to analysis by a statistician utilizing the SPSS
software program, which is a statistical software package designed for social science
research.
To ensure that research findings are applicable to a specific group of interest and
clearly define and distinguish the population from other groups. Creswell (2014) states
that a population is a group of individuals who share at least one characteristic that sets
them apart from other groups. This characteristic is often referred to as the inclusion
criteria. By identifying the population and its defining characteristics, the researcher can
ensure that the sample is representative of the population and that the findings are
investigation by the researcher. As described by Dörnyei (2007), the sample is the group
43
of individuals who are actually examined and studied in a research study. This group of
larger population from which they are drawn. According to Creswell (2012) participants
in a study are able to answer the researcher's questions, provide personal or demographic
information, and complete tasks or assessments as part of the study. The sample is an
important part of any research study because it helps to ensure that the findings are
generalizable to the larger population from which the sample was drawn.
In the current study, the research population comprised second-year high school
students who were enrolled in the scientific section at Misurata Secondary School for
boys, situated in the city of Misurata, during the academic year 2022/2023. The selection
of second-year students in the scientific section for the research sample is due to several
students typically study subjects such as mathematics, physics, and chemistry, which
This makes them an appropriate group to study the effects of AA method, which often
emphasize these skills. Second, they may also be more motivated to participate in the
research since it is related to their field of study. This could lead to a higher response rate
convenience, rather than random selection (McCombes, 2019). So, purposive sampling,
44
because it involves intentionally selecting the specific individuals, events, or settings
that can provide crucial information that cannot be obtained through other sampling
methods. Therefore, this approach is used when researchers aim to target specific groups
researcher's judgment and expertise to identify the most relevant and informative cases
students at Misurata Secondary School for boys, and the researcher selected a sample of
among the six, based on predetermined criterion i.e., their pretest scores.
Reliability and validity are essential aspects of research that help establish the
credibility and trustworthiness of the research process and the resulting findings
Validity refers to the extent to which the collected data accurately represent and
cover the intended area of investigation. It assesses the degree to which the data
accurately reflects the concepts, variables, or phenomena that the researcher aims to
study (Ghauri and Gronhaug, 2005 as cited in Taherdoost, 2016). According to Field
45
no other factors have changed. It quantifies the extent to which the measurement
remains dependable and reproducible over time and across different settings or
In this study, in order to assess the validity of the research instruments i.e., the
pre and post-tests and the questionnaire, they underwent evaluation by two PhD. holders
and a secondary school English inspector. Their review focused on aspects such as the
instruments' structure, clarity, logical flow, length, and the order of statements. The
evaluators provided valuable comments and suggestions regarding the tests and the
questionnaire. They concluded that these instruments were valid and suitable for
achieving the study's aims. Taking their feedback into consideration, certain items were
A pilot study is typically the initial phase of the entire research protocol and
often involves a smaller-sized study that helps in the planning and modification of the
main study (In, 2017). Bell (2010) recommended piloting all data-gathering instruments
in order to assess factors such as the duration of completion, clarity of instructions, and
identification of any unclear or ambiguous items. In this study, only the students'
questionnaire was piloted due to time constraints and the limited number of participants.
attention given to the wording of the questions (Cohen et al., 2007). According to Cohen
et al. (2007), piloting serves several purposes, primarily to enhance the reliability,
46
took approximately 30 minutes to complete the questionnaire as observed by the
researcher. The pilot participants, who were second-year secondary school students from
the scientific section other than the main participants, were then asked to complete the
questionnaire to verify if the items were clear and easy to follow. This evaluation was
conducted prior to using the questionnaire in the main study to ensure its effectiveness
and usability. The feedback received from the students was taken into consideration to
On the 16th of January 2023, a traditional pretest was conducted for 167 second-
year scientific section students, at Misurata Secondary School for boys in the city of
Misurata, Libya, who were assigned to six classes by the school's principal, taking into
account their preferences and academic achievement during their first year of study. The
scoring of the students' papers was completed within a week, and an additional week
was allocated for analyzing and interpreting the data using a paired-sample t-test and
ANCOVA. Based on the results of the pre-test, the researcher then chose two classes
students as well. Throughout the experiment, the control group was taught and assessed
using traditional methods by a colleague, while the experimental group was taught
researcher. This approach was adopted to ensure the consistency of teaching methods for
both groups. Prior to implementing the treatment, the researcher inquired about the
47
students' enrollment in any external courses outside school. It was ensured that none of
themselves. This was done to ensure that any differences in learning outcomes, implied
in testing performances and observed between the two groups, could be attributed solely
to the implementation of AA tools in the experimental group and to obtain valid and
stable results.
May 7th. During this period, multiple AA tools were employed on different days to
assess the students' performance. The AA tools used were quite diverse, including
outcomes and provided multiple opportunities for the students to demonstrate their
to ensure consistency with the pretest and to allow for a direct comparison of the
learning outcomes between the control and experimental groups. Using a traditional
posttest allowed the researcher to assess the effectiveness of the intervention in the
experimental group relative to the control group using the same assessment tool. On
May 8th, the posttest was administered to 56 students, and the test took two hours to
complete. The papers were graded within a week, and the data obtained were analyzed
using a paired-sample t-test to compare the results of both the control and experimental
groups.
48
A questionnaire was administered to the experimental group on May 9th to
and collection. Based on the findings of the pilot study, students were provided 30
minutes to complete this questionnaire, allowing ample time for participants to provide
their responses and seek clarification if needed. The collected data were subsequently
The data was arranged and saved in a computer system by the researcher to
make it ready for the analysis conducted by the statistician. The analysis program was
SPSS software version 21. In the quasi-experiment, the results of both the pre-test and
post-test were analyzed using a paired-sampling t-test to compare the control and the
experimental groups post test results, while a descriptive analysis was used to analyze
the questionnaire data. Further information on the findings and outcomes of the study
To ensure that the privacy and safety of the research participants were protected,
ethical issues were addressed in this study, given that it involved human respondents.
These ethical issues included gaining informed consent from the participants and
misconduct, and manage new and challenging problems that may arise during the study
49
establish trust with them, and uphold the reputation of their organizations or institutions
(Israel & Hay, 2006). This is in line with the principles of research ethics, which aim to
The researcher took into account the ethical issues and constraints when
collecting the study's data. The researcher obtained permission to conduct the
experiment at Misurata Secondary School for boys with the assistance of the supervisor.
An official request letter was sent to the Ministry of Education in Misurata to obtain
permission for the study to be conducted at the school (See Appendix E). The
participants were fully informed about the study's aims, and their rights were ensured if
they agreed to participate. The participants were also informed that they could withdraw
from the study at any time. Additionally, the participants were assured that their data
would be used solely for research purposes, and to protect their privacy, their real
3.9. Summary
the research design and research instruments used, which included a quasi-experiment
and a questionnaire. The chapter also discussed the population, sample, and sampling
techniques used, as well as the piloting and the procedures employed for data collection.
Additionally, the chapter provided details on the data analysis used and the ethical
considerations that were taken into account during the research process. Overall, this
which will help to contextualize and interpret the findings presented in the subsequent
chapter.
50
Chapter Four
4.1. Introduction
This chapter reports the results obtained from the study's participants, integrating
the data collected through the quasi-experiment and the questionnaire to answer the
research questions. Additionally, this chapter discusses the study‘s results that were
obtained with reference to the research questions and hypotheses and discusses pertinent
literature.
4.2. Results
The results are presented in two distinct sections. The first section is dedicated to
the analysis of the quasi-experimental design referring to the first research question, i.e.,
How does using alternative assessment methods affect second-grade secondary school
students‘ testing performances? An independent sample t-test was used to compare the
answer the second question, i.e., What are students‘ attitudes towards using alternative
assessment methods?
After the post-test was conducted the scores were tabulated and analyzed using
an independent sample t-test to compare the results of the post-test of the control and the
51
Table 4.1.
Independent Sample T-Test of the Control and the Experimental Groups’ Post-
Test
The above table provides the results of the independent sample t-test analysis
comparing the mean scores of the post-test control group and the post-test experimental
group. The mean score of the post-test control group was 32.6071, and the standard
deviation was 10.06769. The mean score of the post-test for the experimental group was
31.2143, and the standard deviation was 8.57800. The p-value was (0.580) >0.05 for
both groups, which indicates that there is no statistically significant difference between
the mean scores of the two groups in terms of students‘ testing performances. Therefore,
based on this analysis, it can be concluded that the intervention did not have a significant
effect on the post-test scores of the experimental group compared to the control group.
In the questionnaire, the range of values starts from the highest positive pole
―Strongly Agree‖ with a value (5) and ends at the lowest negative pole ―Strongly
R = 5-1 = 4
52
Table 4.2.
Range of Values
Each section of the questionnaire will be analyzed below, along with a table showing the
4.2.2.1 Assessment
The first section of the questionnaire provides insights into students' attitudes
general, where the overall mean score of approximately 4.0385 indicates a positive
degree of approval. This suggests that students generally recognize the purpose and
value of assessment and view it as important for their learning and progress.
Item 1: Tasks are the main tool to measure the ability of students.
Table 4.3.
Table (4.3.) presents the results of data analysis for item 1, which focuses on the
degree of approval regarding the main tool used to measure the ability of students,
namely tasks. According to the table, the mean for item 1 is 4.2692, and the standard
53
deviation is 0.66679. The mean represents the average or central tendency of the
responses for item 1. In this case, the mean of 4.2692 suggests that, on average, the
statement that tasks are the main tool to measure the ability of students. The standard
deviation provides a measure of the dispersion or variability of the responses for item 1.
A standard deviation of 0.66679 indicates that the responses tend to be relatively close to
the mean, suggesting a relatively consistent level of agreement among the participants.
According to the data in the table, the majority of participants expressed positive
attitudes towards tasks being the main tool to measure their abilities, with (88.5% - the
combined percentage of "Strongly Agree" and "Agree") agreeing and strongly agreeing
with the statement. Additionally, the remaining (11.5%) had a neutral stance on the topic.
The absence of any disagreement suggests a high level of approval among the
participants.
give assignments.
Table 4.4
The data presented in Table (4.4) regarding item 2 provide insights into the
respondents' opinions and attitudes regarding the belief that frequent assignments given
by teachers can enhance students' test performance. According to the table, the mean for
54
item 2 is 3.8846, and the standard deviation is 1.53172, suggesting that, on average, the
statement.
Based on the table, majority of the respondents, comprising (73%) either agreed
in their stance. On the other hand, (15.4%) strongly disagreed, and only (7.7%)
disagreed.
Item 3: In giving the task, the teacher should adapt the tasks to the ability of the
students.
Table 4.5.
The data presented in Table (4.5) offers valuable insights into the perspectives
and attitudes of the respondents regarding the significance of tailoring tasks to match the
abilities of students by teachers. The data reveals a compelling pattern: the relatively
high mean of 4.6154 and low standard deviation of 0.57110 indicate a consistent
consensus among the respondents. A substantial majority, with (96.2%) agreeing and
abilities. However, only (3.8%) maintained a neutral stance, and none of the respondent
55
Item 4: The task given so far has been in accordance with what has been taught by the
teacher.
Table 4.6.
The table illuminates significant insights into the respondents' attitudes regarding
the alignment between the tasks assigned and the content taught by the teacher. The
mean for item 4 is 4.1154, and the standard deviation is 0.90893 indicating that, on
regarding the alignment between the tasks given and the content taught by the teacher.
The data reveals that (73.1%) of the respondents either agreed or strongly agreed,
(23.1%) were neutral, and (3.8%) disagreed with the statement. There were no
respondents who strongly disagreed with the notion of alignment between the tasks and
Table 4.7.
Degree of Approval
Std.
Strongly Agree Neutral Disagree Strongly Mean
Deviation
Agree Disagree
N. 4 13 5 1 3 3.5385 1.17408
% 15.4 50.0 19.2 3.8 11.5
56
The data from Table (4.7.) reveals that item 5 pertains to the level of excitement
and enthusiasm expressed by the respondents towards various tasks. The mean of 3.5385
indicates a moderate level of excitement and enthusiasm among the participants. Further
analysis shows that (65.4%) of the respondents either agreed or strongly agreed with the
statement, (19.2%) were neutral, (11.5%) strongly disagreed, and only (3.8%) disagreed.
These findings highlight the range of attitudes and levels of excitement exhibited by the
Item 6: The task that I have done so far has been in accordance with my ability.
Table 4.8.
Based on the data presented in Table (4.8.), item 6 focuses on the respondents'
attitudes toward the alignment between the tasks they have completed and their abilities.
The table provides information on item 6 where the mean is 3.7692 and the standard
alignment between the tasks they have completed and their own abilities. From the
results, it can be observed that a significant number of respondents (69.2%) either agreed
or strongly agreed with the statement. On the other hand, a notable number of
57
lack of strong agreement or disagreement. A smaller percentage of respondents (3.8%
Item 7: The Tasks could make me measure my abilities from what I have achieved.
Table 4.9.
The data from the previous table indicates that, to a moderate extent, the
respondents perceive that tasks allow them to measure their abilities based on their
achievements. The mean score of 4.0769 and standard deviation of 9.7665 suggest a
level of variability in responses. Furthermore, the results show that (84.6%) of the
respondents either agreed or strongly agreed with this statement, indicating a stronger
On the other hand, (7.7%) of the respondents took a neutral stance, indicating a
lack of strong agreement or disagreement. Only 3.8% disagreed, and an equal percentage
students' attitudes towards TA methods. These items aim to gather information about
58
respondents are asked to indicate their level of agreement, disagreement, or other
relevant responses. The overall mean score of 3.6395 falling within the agreeing range
of {3.4 – 4.20} suggests that students generally have a moderate agreement with the
Item 8: The traditional forms of tasks are easier for me to do because it is so familiar.
Table 4.10.
tasks due to their familiarity. The data from Table (4.10) indicates that the mean score
for this item is 4.0385, suggesting a moderate level of agreement with the statement that
familiarity with traditional tasks contributes to their ease. It is notable that a significant
proportion (73.1%) either agreed or strongly agreed that the ease of traditional tasks is
due to familiarity.
Furthermore, (11.5%) disagreed with the statement, suggesting that a number of the
Item 9: The format of tasks such as multiple-choice, essays and fill-in-the-blanks are
59
Table 4.11.
the representativeness of different task formats (such as multiple-choice, essays, and fill-
in-the-blanks) to their abilities. The mean for item 9 is 3.2800, and the standard
agreement indicates that a significant portion of the respondents believe that the given
Over half of the respondents (52.0%) either agreed or strongly agreed that the
measures their abilities. (28.0%) of the respondents took a neutral stance, indicating that
they neither agreed nor disagreed with the statement. (12.0%) of the respondents
disagreed with the statement, indicating that they perceive a misalignment between the
task formats and their abilities. Furthermore, (8.0%) of the respondents strongly
disagreed with the statement, where they strongly believed that the task formats are not
Item 10: The Traditional assessment makes me more confident in doing the task, even
Table 4.12.
60
The results of data analysis for item 10
The mean and the standard deviation of 3.6000 and 1.02144 indicate a moderate level of
agreement with the statement that traditional assessment makes students more confident
in doing the task, even when the results exceed their expectations. A sizable number
confidence. The table also showed that respondents may have mixed feelings or
situations where (28%) of the respondents held a Neutral attitude that traditional
assessment enhances their confidence even when the results are better than expected. On
the other hand, (8%) of the respondents disagreed with the notion, indicating that they
do not believe traditional assessment contributes to their confidence in cases where the
results exceed their expectations. Lastly, (4%) of the respondents strongly disagreed with
the statement, expressing a strong belief that traditional assessment does not make them
more confident in doing the task when the results surpass their expectations.
The third section of the questionnaire presents the outcomes of statements 11-15,
deviate from TA tools such as exams and standardized tests and often include project-
61
based assessments, presentations, portfolios, or other non-traditional forms of assessing
students‘ performance. The results of these statements provide valuable insights into
how students perceive and engage with AA methods. By examining their responses, we
can gain a better understanding of their attitudes and preferences regarding these
approaches.
The overall mean for the responses in the third section of the questionnaire is
Item 11: Alternative assessment makes me enthusiastic about doing the task because the
Table 4.13.
According to the information provided, the mean value for item 11 is 3.7600,
assessment methods are employed, mainly due to their appreciation of the innovative
held a favorable or positive stance, (20%) maintained a neutral stance, while only (12%)
62
Item 12: The task in innovation forms makes me more creative in doing the assignment.
Table 4.14.
The average score of 3.8800 and the standard deviation of 1.33292 indicate a
moderate level of agreement among respondents regarding the statement that the task in
percentages in the previous table, represent the proportion of respondents who selected
each option for item 12. Based on this information, we can interpret that a number of
respondents (68.0%) expressed agreement or strong agreement with the statement that
the task in innovation forms makes them more creative in doing the assignment.
(12%) expressed their strong disagreement while no respondents disagreed with the
statement.
Item 13: Tasks such as self and peer assessment, interviews, group projects, concept
Table 4.15.
63
The mean score of 3.9600 and the standard deviation of 1.09848 suggest that
respondents generally agree and perceive tasks such as self and peer assessment,
statement regarding the representation of tasks such as self and peer assessment,
participants, totaling (76.0%), expressed their agreement as they strongly believe that
these tasks accurately reflect their abilities. On the other hand, (12.0%) expressed a
participants, (8.0%), disagreed with the statement, indicating that they do not consider
suggesting a minority who strongly believe that these tasks are not at all representative
of their abilities.
Item 14: Alternative tasks should often be applied more to make me adapt to my
Table 4.16.
64
Based on the data presented in the table, it can be concluded that a significant
number of respondents (72.0%) agreed or strongly agreed with the notion that alternative
tasks should be frequently utilized to facilitate adaptation to their abilities and enhance
their testing performance. The relatively high mean score of 4.0400 and the standard
among the respondents regarding the importance of employing alternative tasks for
believe that alternative tasks are effective or necessary in this context with only (12%) of
respondents expressing their disagreement with the idea that alternative tasks should be
frequently employed to adapt to their abilities and improve their testing performance.
Item 15: Alternative assessment made me enthusiastic about doing the task and it makes
me face new things that can improve my test performance and my ability.
Table 4.17.
agreed or strongly agreed with the statement that alternative assessments generated
enthusiasm and exposed them to new aspects that could enhance their testing
65
performances and abilities. Moreover, (12.0%) maintained a neutral stance, implying
neither strong agreement nor disagreement. It is noteworthy that none of the respondents
selected the "Disagree" option, while a small percentage (4.0%) strongly disagreed with
the statement. The mean score of 4.1600 and the standard deviation of 0.94340 further
support the interpretation of a high level of agreement among the respondents regarding
assessment methods i.e., alternative and traditional, a one-sample test was conducted
Table 4.18.
One-Sample Test
Test value = 3
Type
N Mean Std. Deviation P-value
The results of the one-sample test indicate that for both the TA and AA methods,
the mean scores are slightly different from the test value of 3 (representing a neutral
attitude). The analysis of the mean values and the p-value of the TA and AA methods can
provide insights into students' attitudes. In this case, the mean for the TA is 3.6282,
while the mean for the AA is 3.9635 which suggests that on average students generally
66
4.3. Discussion
This section discusses the study‘s results that were obtained with reference to the
research questions and hypotheses. It is important to review the research questions and
Research Question (i): How does using alternative assessment methods affect
Research Question (ii): What are students' attitudes towards using alternative assessment
methods?
The quantitative data related to the first research question revealed that there was
no statistically significant difference in the post-test scores between the control group
and the experimental group, as indicated by the independent sample t-test (see Table
4.1.). This suggests that the AA methods used in the experimental group did not have a
methods used in the control group. As a result, the findings from the data analysis
support the null hypothesis, which states that alternative assessment methods do not
67
Consequently, the alternative hypothesis (H1) is rejected. The observed outcome could
teaching method. Teaching methods and assessment methods are interconnected and
should be considered together. If traditional teaching methods are maintained while only
situation is evident in the Libyan context, where curriculum reforms have been
implemented multiple times, but the continued use of TA methods has produced
These findings align with the study conducted by Bachelor (2017), reported in
chapter two, where the results of his study revealed that there was no significant
difference between the scores of the experimental group in the final exam and the scores
These findings contribute to the existing literature and suggest that AA methods
compared to TA methods. However, this lack of difference between the control and
experimental groups in the post-test scores is not consistent with some previous studies
performances (Borko et al., 1993; Koretz et al., 1993; Newman et al., 1996; Stecher and
Mitchell, 1995; Brown, 2004; El-Koumy, 2009; Baniabdelrahman, 2010; Safa &
Goodarzi, 2014; Heidarian, 2016; Aliasin & Amanlu, 2017; Kafipour & Khoshnood,
2023).
study conducted by Safa and Goodarzi (2014) who found that the students in the
68
experimental group improved their grammar skills significantly after they were assessed
through the TBA. Aliasin and Amanlu (2017) also supported Safa and Goodarzi‘s
findings, where they reported that the implementation of multiple AA methods led to an
contrasting results between the current study and these previous studies highlight the
population.
The lack of difference between the control and experimental groups in the post-
test scores of this study may also suggest that TA methods continue to be effective in
by Sahera et al. (2022), where it was found that students achieved higher grades in
study and this study highlight the continued relevance and effectiveness of TA methods
in evaluating student performance. It suggests that these methods, which often include
standardized tests or exams, may still provide a valid and reliable measure of students'
The analysis of the questionnaire yielded the following findings. In regards to the
first section of the questionnaire about students‘ attitudes towards assessment in general,
the findings demonstrate that some students generally showed positive attitudes towards
various assessment tasks, while other students generated mixed or divergent opinions.
69
This conclusion complies with Edelstein et al. (2000 as cited in Struyven, et al., 2005)
comments when they examined students‘ perceptions of the various types of assessment
and found that students valued each method according to its specific purposes.
results showed that participants' preferences for both TA and AA methods leaned
towards more positive attitudes. This finding is consistent with Irawa‘s (2017) findings
where students‘ perceptions of both assessment methods were generally positive with
On the one hand, participants of the current study viewed AA method as valuable
resource that offers them fresh experiences and chances for growth. The suggestion that
the integration of innovative and creative assessment formats has a positive effect on
regarding studies related to AA. These include (Peng, 2009; Burnaz, 2011; Elezovic,
2011; Azarnoosh, 2013; Cornelius & Kinghorn, 2014; Fajarsari, 2016; Bachelor, 2017;
Lustgarten, 2022), where they made references to students who showed a preference for
AA, perceiving them as less stressful and more conducive to their learning. In particular,
in one of the studies (Bachelor, 2017), students in the experimental group displayed
positive attitudes towards AA types. Additionally, Peng (2009) found that both high and
On the other hand, the results of the present study revealed that students
perceived TA tasks as easier due to their familiarity. Moreover, the findings suggest that
students believe the diverse formats of TA tasks effectively assess their skills and
70
knowledge. Previous studies undertaken are in line with the present study where the
findings indicated that students considered the current testing practices i.e., traditional
testing, to be adequate and fair for assessing their English language proficiency (Garside
et al., 2009; Caner, 2010; Jaturapitakkul, 2013; Phongsirikul, 2018). The tests in these
studies were viewed by students as capable of measuring the important content that
students learned in their courses. Additionally, the test formats were generally seen as a
reflection of students' language abilities, but there was a desire for more diverse and
innovative item types that could effectively assess all language skills (Jaturapitakkul,
2013).
students expressed dissatisfaction with the existing TA methods. This finding contrasts
with the current study's findings, where students showed positive attitudes towards both
TA and AA methods. One possible explanation for this discrepancy could be the
difference in the sampled participants. The participants in Ramadan and Dekheel's study
were undergraduate students, while the participants in the current study were secondary
school students.
4.4. Summary
In this chapter, both the data obtained from the quasi-experiment and the
independent sample t-test to compare the post-test results between the control and
experimental groups. The chapter also discussed the findings derived from the quasi-
experimental scores and the students' questionnaires. The main focus was on examining
71
students at Misurata Secondary School for Boys in Misurata, Libya. Additionally, the
chapter explored the students' perceptions towards the implemented AA tools. The
findings of the study were carefully examined and expounded upon to provide a
provide a concise summary of the most significant conclusions drawn from the study.
72
Chapter Five
5. Conclusion
5.1. Introduction
This chapter provides a comprehensive summary of the study and discusses the
relation to suggestions for further research and the limitations of the study.
School for Boys in Misurata, Libya. Additionally, it has attempted to elicit students‘
attitudes towards these methods. To this end, quantitative design was used i.e., a quasi-
experimental design and a questionnaire as research tools. The obtained data were
The post-test scores of the control and experimental groups did not show a
students compared to TA method. This contradicts some previous literature that supports
The findings indicate that the AA method employed in the experimental group of
the current study do not surpass the effectiveness of TA method in assessing students'
learning outcomes. This suggests that while AA method have shown benefits in certain
contexts and for specific skills, TA method continue to hold value in assessing students‘
performances.
73
Regarding students' attitudes towards assessment methods, the results showed a
mix of positive attitudes towards AA and TA methods. However, the AA method to some
resources that provided fresh experiences and opportunities for growth. Students also
perceived AA as less stressful and more conducive to their learning. Meanwhile, other
students also believed that TA tasks are easier due to their familiarity and that they
Overall, the study suggests that while AA method may not have a significant
effect on students‘ testing performances compared to TA, students hold positive attitudes
towards the use of AA in language classroom assessment, and recognize their value in
promoting motivation and involvement. TA method is still viewed as adequate and fair
but there is a growing demand for a wider range of innovative assessment methods that
performances, it could provide valuable learning experiences for students and enhance
Secondly, the study emphasizes the need for a balanced approach to assessment.
Indeed, incorporating a variety of assessment formats and item types can effectively
assess students' language skills and provide a more comprehensive picture of their
testing performances and learning outcomes. Additionally, the study suggests that
74
teachers should be aware of students' preferences and attitudes towards different
assessment methods. Understanding students‘ attitudes can help educators tailor their
environment.
Every research study has limitations and this study is not an exception. This
study was conducted within the context of Misurata Secondary School for boys, second-
year Scientific Section students, in Misurata, Libya, which may limit the generalizability
of the findings to other contexts. The limited number of participants can be seen as a
limitation. Moreover, due to the nature of the experiment being conducted at a secondary
school, where class selection is typically determined by the principal, and considering
the limited availability of students to volunteer for participation outside of regular school
hours, the researcher faced challenges in forming the experimental and control groups.
Therefore, replicating the study in different contexts and with diverse student
performances, and did not explore other potential aspects, such as student engagement,
explore the potential benefits of AA methods as they can be effectively integrated into
75
Furthermore, it would be beneficial to investigate the barriers to implementing
addressing various settings, primary and preparatory classrooms. Further research can
also be suggested to explore EFL Libyan teachers‘ perceptions and attitudes towards the
Finally, to gain a more holistic view of students‘ abilities and progress, it would
76
REFERENCES
http://sure.sunderland.ac.uk/id/eprint/9585
Glasgow]. https://eleanor.lib.gla.ac.uk/record=b2613025
Aliasin, S. H., & Amanlu, M. (2017). The effect of alternative assessment techniques on
case of Iranian junior high school students. Linguistics and Literature Studies,
https://doi.org/10.1007/978-981-13-0265-7_4
https://dictionary.apa.org/posttest
77
Ayyoub, A. A., Bsharat, A., & Suleiman, M. (2021). The impact of alternative authentic
Azarnoosh, M. (2013). Peer assessment in an EFL context: Attitudes and friendship bias.
https://doi.org/10.1080/00220671.1991.10702818
https://files.eric.ed.gov/fulltext/EJ899764.pdf
Barlow, L., & Coombe, C. (2000). Alternative Assessment Acquisition in the United
https://files.eric.ed.gov/fulltext/ED426622.pdf
Bell, J. (2010). Doing your research project: A guide for first-time researchers in
learning processes and prior knowledge. Springer Science & Business Media.
Bizer, G. Y., Barden, J. C., & Petty, R. E. (2006). Attitudes. In Encyclopedia of Cognitive
Ltd. https://doi.org/10.1002/0470018860.s00483
78
Borko, H., Flory, M., & Cumbo, K. (1993). Teachers' ideas and practices about
Brookhart, S. M., & Chen, F. (2014). The quality and effectiveness of descriptive
https://doi.org/10.1080/00131911.2014.929565
Brown, A. H., & Green, T. D. (2006). The essentials of instructional design: Connecting
fundamental principles with process and practice (3rd ed.). Prentice Hall.
Brown, J. D., & Hudson, T. (1998). The alternatives in language assessment. TESOL
thesis]. http://etd.lib.metu.edu.tr/upload/12613466/index.pdf
223–236.
79
Chirimbu, S. (2013). Using alternative assessment methods in foreign language
https://www.semanticscholar.org/paper/Using-Alternative-Assessment-Methods-
in-Foreign-of-Chirimbu/e808e7f555d01e7185a6adc1aa651c7e23468561
Cohen, L., Manion, L., & Morrison, K. (2007). Research methods in education (6th ed.).
Routledge.
Cornelius, S., & Kinghorn, O. (2014). Student Attitudes Towards Self And Peer
Department of Education. (2002). Revised national curriculum statement for grades R-9
Diep, A. N., Zhu, C., & Vo, M. H. (2019). The Use of Alternative Assessment in
80
Dikli, S. (2003). Assessment at a distance: Traditional vs. Alternative Assessments. The
Linguistics , 1058-1067.
(ERIC). https://doi.org/10.2139/ssrn.2364813
Etobro, B. (2022). Re: What is the difference between academic performance and
https://www.researchgate.net/post/What-is-the-difference-between-academic-
performance-and-academic-
achievement/63728bd5db471cd211073b7b/citation/download.
Field, A. (2009). Discovering statistics using SPSS (3rd ed.). SAGE Publications.
81
Fulcher, G. (2010). Practical language testing. Routledge.
Garside, J., Nhemachena, J. Z., Williams, J., & Topping, A. (2009). Repositioning
Gentova, C. S., & Madrigal, D. V. (2020). Classroom climate and student‘s academic
Gielen, S., Dochy, F., & Dierick, S. (2003). Evaluating the Consequential Validity of
Gravetter, F. J., & Forzano, L. B. (2012). Research methods for the behavioral sciences
Harris, A. D., McGregor, J. C., Perencevich, E. N., Furuno, J. P., Zhu, J., Peterson, D. E.,
82
Heidarian, N. (2016). Investigating the Effect of Using Self-Assessment on Iranian EFL
Ikwuka, O. I., Ejikeme, L. I., Olugbemi, P. O., & Okoye, A. C. (2021). Effect of Role-
In, J. (2017). Introduction of a pilot study. Korean Journal of Anesthesiology, 70(6), 601.
https://doi.org/10.4097/kjae.2017.70.6.601
Repository. https://repository.ar-raniry.ac.id/id/eprint/2603/
Israel, M., & Hay, I. (2006). Research ethics for social scientists. SAGE Publications.
https://doi.org/10.5901/ajis.2013.v2n3p445
Jonsson, A., & Svingby, G. (2007). The use of scoring rubrics: Reliability, validity and
https://doi.org/10.1016/j.edurev.2007.05.002
Kafipour, R., & Khoshnood, A. (2023). Effect of feedback through dynamic assessment
https://doi.org/10.3389/feduc.2023.1049680
83
Kenny, D. A. (1975). A quasi-experimental approach to assessing treatment effects in the
https://doi.org/10.1037/0033-2909.82.3.345
Khattri, N., & Sweet, D. (1996). Assessment Reform: Promises and Challenges. In M. B.
Kim, E. S., & Willson, V. L. (2010). Evaluating pretest effects in pre–post studies.
https://doi.org/10.1177/0013164410366687
Kırıkkaya, E. B., & Vurkaya, G. E. (2011). The Effect of Using Alternative Assessment
Koretz, D., Stecher, B., Klein, S., McCaffrey, D., & Deibert, E. (1993). Can portfolios
Law, B., & Eckes, M. (2007). Assessment and ESL: An alternative approach. Portage &
Main Press.
https://doi.org/10.15516/cje.v17i0.1496
84
Lustgarten, B. (2022). Impact of Traditional Versus Alternative Assessment on Student
Repository. https://scholarworks.bgsu.edu/honorsprojects/682/
Mackey, A., & Gass, S. M. (2016). Second language research (2nd ed.). Routledge.
The SAGE Handbook of Applied Social Research Methods (pp. 214-253). SAGE
McCombes, S. (2019, September 19). Sampling methods Types, techniques & examples.
Scribbr. https://www.scribbr.com/methodology/sampling-methods/
https://doi.org/10.54809/jkss.vi3.97
Moon, T. R., Brighton, C. M., Callahan, C. M., & Robinson, A. (2005). Development of
https://doi.org/10.34050/elsjish.v3i3.11077
85
Muñoz, P., Véliz Campos, M., & Veliz, L. (2019). Assessment in the English language
http://jallr.ir/index.php/JALLR/index
Nasri, N., Roslan, S. N., Sekuan, M. I., Bakar, K. A., & Puteh, S. N. (2010). Teachers‘
7, 37-42. https://doi.org/10.1016/j.sbspro.2010.10.006
Newmann, F. M., Marks, H. M., & Gamoran, A. (1996). Authentic pedagogy and student
https://doi.org/10.1086/444136
(buxdu.uz), 2(2).
86
Orafi, S. M., & Borg, S. (2009). Intentions and realities in implementing communicative
https://doi.org/10.1016/j.system.2008.11.004
Ornstein, A., & Hunkins, F. (2009). Curriculum: Foundations, principles, and issues.
Öztürk, Y. A., & Şahin, Ç. (2014). The effects of alternative assessment and evaluation
dissertation].
https://www.proquest.com/openview/d213f0aa1bf915ef5d87eab113e5abe4/1?pq-
origsite=gscholar&cbl=18750&diss=y
Pereira, D., Assunção Flores, M., & Barros, A. (2017). Perceptions of Portuguese
https://doi.org/10.1080/03055698.2017.1293505
160. https://doi.org/10.19062/2247-3173.2017.19.2.22
87
Ramadan, M. O., & Dekheel, H. (2020). Libyan Students‘ Perceptions of Traditional
https://journal.su.edu.ly/index.php/abhat/article/view/481
Research guide: Organizing your social sciences research paper: Purpose of guide.
California. https://libguides.usc.edu/writingguide
Rezaee, A. A., Alavi, S. M., & Shabani, E. (2013). Alternative assessment or traditional
DOI:10.22132/TEL.2013.54864
Roberts, P., Priest, H., & Traynor, M. (2006). Reliability and Validity in Research.
Safa, M. A., & Goodarzi, S. (2014). The Washback effects of task-based assessment on
Saher, A. S., Ali, A. M., Amani, D., & Najwan, F. (2022). Traditional versus authentic
88
Sawad, N. A., & Law, J. M. (2023). Evaluating the efficacy of an Arabic metalinguistic
332-373. https://doi.org/10.46827/ejes.v10i10.5035
Segers, M., Dochy, F., & Cascallar, E. (2003). The era of assessment engineering:
Changing perspectives on teaching and learning and the role of new modes of
Academic publishers.
Shadish, W. R., Cook, T. D., & Campbell, D. T. (2002). Experimental and quasi-
Sharifi, A., & Hassaskhah, J. (2011). The Role of Portfolio Assessment and Reflection
dissertation]. http://etheses.dur.ac.uk/878/
Stears, M., & Gopal, N. (2010). Exploring alternative assessment strategies in science
https://doi.org/10.15700/saje.v30n4a390
89
Struyven, K., Dochy, F., & Janssens, S. (2005). Students‘ perceptions about evaluation
Taherdoost, H. (2016). Validity and reliability of the research instrument; How to test the
dissertation]. http://sure.sunderland.ac.uk/id/eprint/3306
curriculum-statement-grades-r-9-schools/
https://doi.org/10.1080/0969594x.2017.1401526
https://doi.org/10.1080/01421590220145699
90
Appendices
Appendix A
Traditional Pretest
(13marks )
91
2. Match a word from each list to form compound nouns:
1 Microwave A Phone 1
2 Alarm B Oven 2
3 Social C Clock 3
4 Remote D Media 4
5 Smart E Control 5
(5 marks)
3. Choose the correct answer:
92
7. Cambridge Science Festival lasts for …………
A. fifteen days
B. thirteen days
C. sixteen days
D. fourteen days
8. Which school subjects are you most interested ……………..
A. on
B. with
C. in
D. of
9. It's such ………………………….place.
A. hot
B. an hot
C. a hot
D. the hot
10. It was ………………….………….frightening that I couldn't watch it.
A. so
B. so a
C. such an
D. such
11. Spend time visiting a lot of websites
A. go online
B. download
C. send email
D. surf the internet
12. Where's your jacket? I'm not sure but I think your brother ……….…taken it.
A. must have
B. can't have
C. might have
D. should have
13. A …………………..……….insect .
A. red small
B. small red
C. red and small
D. small and red
14. The ……………………………eats bamboo.
A. giant panda
B. mountain gorilla
C. right whale
D. black rhino
93
15. My cat ................................................... a tiger
A. look
B. look like
C. looks
D. looks lik
(15 Marks)
4. Write the words in the box to their definitions:
94
Appendix B
Alternative Tools
4- Role Play the following situation in pairs. You and a friend found a wallet in the
street. Discuss the best thing to do with it.
6- Read the text titled “DANA’S STORY” on your coursebook page 60.
7- As part of the interview process, the instructor would engage each student by
asking for their advice in various situations presented on page 68 of the workbook.
95
8- For self-assessment, the instructor ask the students to complete exercise E on
page 61 of the workbook. After completing the exercise, each student should
independently mark their own answers.
96
Appendix C
Traditional Post Test
97
b-Write sentences in the second conditional (5 marks).
1-If you (have) wings, where (you fly)?
…………………………………………………………………………………………..
2-You (not feel) well if you (swim) in a polluted river.
…………………………………………………………………………………………..
3-If everyone (buy) products with less packaging, there (be) less waste.
…………………………………………………………………………………………..
4-What (you do) if you (see) someone drop litter in the street?
…………………………………………………………………………………………..
5-The earth (be) a better place if we (look) after the environment.
…………………………………………………………………………………………..
c-Rewrite these sentences with the correct punctuation (5 marks).
1-you know nabil is coming in november
…………………………………………………………………………………………
2-my lebanese friend mohamed al mansour loves football
………………………………………………………………………………………….
3-my email address is [email protected]
…………………………………………………………………………………………..
4-do they play basketball in toubrek
…………………………………………………………………………………………..
5-i think the eshtar hotel is in al sa‘adon street
…………………………………………………………………………………………..
Q3- Choose the correct answer: (13marks)
1- I wish I had ………………. My mother with me.
a -took b-take c-taken
2- She doesn‘t have to wear a uniform.
a-Flight attendant b-student c- teacher
3- They grow a lot of …………….in Brazil.
a-coffees b-coffee
4- Your grades have been getting better all year.
a-instructed b-improving
98
5- I wish you didn‘t live so far away.
a-past b-present c-future
6- The reporting verb for a command sentence is …………
a-asked b-said c-told
7- You …………. use your phone in school. It is forbidden.
a-mustn‘t b-must c-don‘t have to
8- We use ………….. to say something is not necessary.
a-shouldn‘t b-don‘t have to c-have to
9- Make a ……………
a-cake b- the table c- fire
10- A lamb is a ………….
a-fruit b-drinks c-meat
11- Adjective to describe food…………..
a-honey b-boiled c- spicy
12- She mustn‘t………….. loudly in the library.
a-talks b-talk c-talked
13- Light a …………………………...
a-fire b- meat c-cake
Q4-Write the words in the box to their meaning : (10 marks)
1- Annoying a- Not awake 1- ………
2- Selfish b- Talk in a friendly way 2- ………
3- Room c- At the beginning 3-………
4- To start with d- For a long time 4-………
5- Argue e- Only thinking about yourself 5-………
6- Picnic f- Difficult, causing trouble 6-………
7- Breeze g- Disagree, fight with words 7-………
8- Unconscious h-Gentle wind 8-………
9- For ages i-Meal outside 9-………
10-
10- Chat j-Space
………
99
Appendix D
Students’ Questionnaire
Strongly Strongly
Statements Agree Neutral Disagree
No. Agree Disagree
اَاخٛانث أٔافك ذٚيحا ال أٔافك
أٔافك تشذج ال أٔافك تشذج
Tasks are the main tool to
measure the ability of
students
1.
األداجْٙ ثاخٚاألَشطح ٔانرذس
اط لذسج انطالبٛح نمٛغٛانشئ
To improve students'
achievements, teachers
should often give
2. assignments
جةٚ ،ٍ إَجاصاخ انطالبٛعهٗ نرحغ
ٍش يٛ كثٍٙ فًٛاٌ إعطاء انًعهٛاألح
ثاخٚأَشطح ٔ ذذس
In giving the task, the teacher
3. should adapt the tasks to the
ability of the students
100
Strongly Strongly
Statements Agree Neutral Disagree
No. Agree Disagree
اَاخٛانث أٔافك ذٚيحا ال أٔافك
أٔافك تشذج ال أٔافك تشذج
ٗجة عهٚ ،ثاخٚعُذ إعطاء انرذس
انًعهى يالئًح األَشطح يع لذسج
انطالب
Tasks given so far have been
in accordance with what have
been taught by the teacher
4.
ًٌٜٓا حرٗ اٚ ذى ذمذٙاألَشطح انر
غّ يٍ لثمٚكاَد ٔفما نًا ذى ذذس
.انًعهى
I am very excited and
enthusiastic about doing
5. various tasks
او تاألَشطحٛأَا يرحًظ جذا نهم
.ثاخ انًخرهفحٚٔانرذس
Tasks that I have done so far
have been in accordance with
6. my abilities.
ٌ كاَدٜ لًد تٓا حرٗ اٙاألَشطح انر
.ٙٔفما نمذساذ
The Tasks could make me
measure my abilities from
7. what I have achieved
ُٙثاخ أٌ ذجعهًٚكٍ نألَشطح ٔانرذسٚ
.ّ يًا حممرٙظ لذساذٛأل
The traditional forms of tasks
are easier for me to do
8. because it is so familiar.
ح أعٓمٚذٛثاخ انرمهٚاألَشطح ٔانرذس
.او تٓا ألَٓا يأنٕفح جذاٛ نهمٙتانُغثح ن
The format of tasks such as
multiple-choice, essays and
sfmilb-ehN-li-llff are very
representative of my ability
9. to do the tasks.
اسٛح يثم االخرٚذٛثاخ انرمهٚغ انرذسٛص
،يٍ يرعذد ٔانًماالخ ٔيمء انفشاغاخ
ِاو تٓزٛ عهٗ انمٙيًثهح جذا نمذسذ
.ثاخٚانرذس
The Traditional assessment
makes me more confident in
doing the tasks, even though
sometimes the results I got
10. were beyond my expectations
ٙ أكثش ثمح فُٙجعهٚ ٘ذٛى انرمهٛٛانرم
ٌثاخ عهٗ انشغى يٍ أٚاو تٓزِ انرذسٛانم
ٙٓا كاَد فٛ حصهد عهٙانُرائج انر
ٙاٌ ذرجأص ذٕلعاذٛتعض األح
101
Strongly Strongly
Statements Agree Neutral Disagree
No. Agree Disagree
اَاخٛانث أٔافك ذٚيحا ال أٔافك
أٔافك تشذج ال أٔافك تشذج
Alternative assessment
makes me enthusiastic about
doing the tasks because the
format of the assignments is
11.
innovative.
يرحًغاُٙهح ذجعهٚى انثذٛٛٔعائم انرم
ِغ ْزٛاو تٓزِ األَشطح ألٌ صٛنهم
.األَشطح يثركشج
102
Appendix E
103