Understanding and Applying Assessment in Education - Damian Murchan Gerry Shiel - Sage Publications, LTD - (Textbooks), London, 2017 - SAGE - 9781473913288 - Anna's Archive
Understanding and Applying Assessment in Education - Damian Murchan Gerry Shiel - Sage Publications, LTD - (Textbooks), London, 2017 - SAGE - 9781473913288 - Anna's Archive
Pete Veta.
 uN
BR7285
Sara Miller McCune founded SAGE Publishing in 1965 to support
the dissemination of usable knowledge and educate a global
community. SAGE publishes more than 1000 journals and over
800 new books each year, spanning a wide range of subject areas.
Our growing selection of library products includes archives, data,
case Studies and video. SAGE remains majority owned by our
founder and after her lifetime will become owned by acharitable
trust that secures the company’s continued independence.
            @SAGE
            Los Angeles |London |New Delhi
         Singapore |Washington DC |Melbourne
@SAGE
Los Angeles |London |New Delhi
Singapore |Washington DC |Melbourne
                                                                Pore
Indexer: Anne Solamito
Marketing manager: Lorna Patkai
Cover design: Sheila Tong
Typeset by: C&M Digitals (P) Ltd, Chennai, India
                                                                               |. 26
Printed and bound by
CPI Group (UK) Ltd, Croydon, CRO 4YY
                                                                              R44 | >
         TRINITY LIBRARY                              ISBN 978-1-4739-1328-8
                                                      ISBN 978-1-4739-1329-5 (pbk)
      O er
        11 JUL 2017
                   DUBLIN
                                 Soarwmetin
                                         ni   ena
At SAGE we take sustainability seriously. Most of our products are printed in the UK using FSC papers
                                                                                                       and boards.
When we print overseas we ensure sustainable papers are used as measured by the PREPS grading
                                                                                                      system.
We undertake an annual audit to monitor our sustainability.
      About the authors                                                                        Vii
      Acknowledgements                                                                        Viii
      Companion website                                                                         x
      Preface                                                                                  xi
11 Assessment planning and policy at class, school and system levels 184
References                                                             208
Index                                                                  223
Damian   Murchan     is Assistant Professor   in the School of Education    at Trinity
College Dublin, the University of Dublin, Ireland. With former roles as Director of
both undergraduate and postgraduate teaching and learning in the School of
Education, he has been involved in initial and continuing teacher education pro-
grammes for primary and secondary teachers for many years. Recent research
interests include: school-based assessment reform; incorporation of key skills into
the curriculum; diagnostic testing in mathematics, e-learning and assessment, and
teacher professional development. A former teacher and school principal, Damian
has emphasised practical application in his research through a range of school-
based projects. He has held a number of advisory roles in relation to development
of assessment policy and practice in Ireland and internationally.
Gerry Shiel is Research Fellow at the Educational Research Centre, Dublin, Ireland
where he directs a range of national and international assessment projects. His
responsibilities include the development of standardised tests of reading (English and
Irish), mathematics and science for primary and lower secondary students, and
implementation of and reporting on the Programme for International Student
Assessment (PISA) in Ireland. He teaches on a number of modules at the Institute of
Education, Dublin City University. He has delivered courses on assessment for the
World Bank in a number of developing countries. A former teacher, he has worked
                                                                                es |
with children with special education needs and children experiencing difficulti
with literacy and numeracy.
We are grateful to many people who, in a variety of ways, informed the development,
of our ideas for this book. Through our work on a range of national and international
assessment projects we realised the extent to which teachers and policy-makers some-
times look at assessment through different lenses. Closing that gap so that teachers
and policy-makers understand assessment more fully from each other’s perspective
became one of our aims. Discussions with colleagues and students in Trinity College
and the Educational Research Centre in Dublin provided rich stimulus to undertake
and sustain our writing. Our various involvements with schools in Ireland and inter-
nationally brought us into direct contact with many teachers and students whose
dedication and commitment to professional development and learning provided inspi-
ration for us to write an account of assessment that was even-handed and provided
ample practical illustration. Through our work with government officials, policy-
makers, local educational administrators, NGO personnel and educational publishers,
we recognise and cherish the passionate interest amongst this broad constituency in
making education systems work more effectively for students. We have sought to
represent this inclusive optimistic perspective and commitment to learning in the
book. The support and advice of the editorial team in SAGE was much appreciated
throughout the conceptualisation and development of the book, with particular thanks
to James Clark and Robert Patterson for their helpful conversations over an extended
period. The book is richer for the addition of a number of practical illustrations.
                                                                                  For
permission to use copyrighted materials we owe thanks to a number of agencies
identified in the text. We reserve final thanks to the close family members who have
watched, waited and supported us patiently as the writing project moved from dream
to draft to reality: Mary, Jonathan, Eleanor and Sarah Murchan; and Mairéad, Eoin and
Caoimhe Shiel.
                                                    Acknowledgements        ix
SAGE would like to thank the following reviewers whose comments   have helped
shape this book:
Key points. Contains a brief overview of main issues introduced in the chapter. se
One of our hopes is that readers will relate the ideas for enhancing assessment
described in this volume to their own classroom practices. We hope also that read-
ers will accommodate and implement approaches to assessment that will provide
evidence to help them gauge whether teaching and learning are effective.
   The volume is divided into twelve chapters. Chapter 1 considers the purposes
of assessment and outlines eight key principles of assessment, including the ethi-
cal responsibilities of assessors. In Chapter 2, we address three key features of
assessment ~ validity, reliability and fairness — that, like the principles of assess-
ment, are relevant to understanding the different eee of assessment described
    xii      Understanding and applying assessment in education
e         Chapter introductions and summaries. Each chapter begins with a brief intro-
          duction followed by a list of questions that will help you to engage successfully
          with the chapter. Key concepts are summarised at the end of each chapter.
°         Case studies. Illustrations of assessment techniques, practices and experiences
          are provided that assist readers develop a more concrete understanding of the
          reality of assessment both in classrooms and at the broader system level.
e         Structured formative practice. Try this out activities offer resources and
          prompts to help readers to apply their learning in practice, for example in daily
          work or on school placement. Individual and group reflection on practice is
          facilitated using additional resources downloadable from the companion website.
e         Join-the-debate. Reflective activities are included that alert readers to the con-
          tested nature of some topical issues in assessment. Contrasting perspectives on
          specific assessment policies or practices are presented and readers are encour-
          aged to form their own views individually or in groups.
e         Personalities in educational assessment. Brief biographical profiles of selected
          individuals associated with key developments in assessment over the years.
°         Questions for discussion. Two or three questions that may be discussed indi-
          vidually or in groups.
e         Further reading. Chapters conclude with a number of selected annotated trefer-
          ences (print and web-based) offering further details about key concepts.
While the content of this volume was up-to-date at the time of writing, we urge read-
ers to bear in mind that assessment is continually evolving in response to policy
changes. Consequently, new ideas and approaches are inttoduced on an almost-daily
basis. We invite readers to check official and other sources, including websites,
                                                                                   for
information on new developments, as well as changes to existing approaches.
         What you will learn in this chapter
         This chapter sets the context for the work of teachers in class and links that work
         with student learning and with professional, legislative and societal expectations on
         schools and teachers. Schools are amongst the most complex organisations in society,
         tasked with ever-expanding expectations and obligations by parents, administrators
         and policy-makers. One key competency expected of teachers relates to student
         assessment. Teachers who do not grasp clearly the many purposes to which assess-
         ment is put and the array of approaches used in modern education systems are less
         prepared to contribute effectively to the collaborative education ventures that are
         designed to serve both individual students and broader society.
             In this chapter, you will learn about the purposes of assessment, purposes for
         teachers and students and other purposes that are legitimately expected by stake-
         holders who are outside school but who are central to the education process. The
     ‘
         chapter, and indeed the book overall, emphasise the diversity of assessment pur-
bs
         poses. There are many stakeholders at different levels of remove from your work in
         class. Students, parents, colleagues and school management are in close proximity to
         your work. Teachers in schools where your students might transfer have an interest
         also, as do policy-makers who develop curriculum, assessments and codes of prac-
         tice for teachers. Governments, representing taxpayers and responsible for anticipat-
         ing future national needs, require information and assurance about the effectiveness
         of the system. Your work needs to be set and understood not only in the context of
         your class but also within broader systems and professional contexts. We conceptu-
         alise this plurality of information, needs and actions within two key themes that are
         reinforced throughout the book: (i) using information from assessments to evaluate
         and plan learning at student, class, school and system levels; and Gi) engaging with
         assessment as a collaborative process. We identify eight key principles that should
         inform educators’ thinking and practice in relation to assessment.
                    ¢
2     Understanding and applying assessment in education
When you finish the chapter, you should be able to answer these questions:
We can add a few other trends and implications. The policy drive aimed at embed-
ding strategies in curricula to enhance students’ cross-curricular skills, such as creativ-
ity, collaboration and problem solving, is quite a bit ahead of large-scale practical
means to assess such traits. At secondary level, the move towards teacher involvement
in formal certification is seen as one solution. There is likely to be an even greater
use of assessment information as one of many sources of data in judging educational
quality and justifying policies and expenditure. Finally, as assessment systems become
more complex we are likely to see further ceding of responsibility for administrative
and technical aspects of assessment by education ministries to specialised agencies
and the commercial or non-profit sectors. These trends suggest a corresponding com-
plexity for teachers in understanding and working within such systems.
    There is no’ one role for assessment in education. Rather, assessment provides
information that can be used by different people for different purposes. Many people
have an interest in the work of schools and all may require information that suits
their purpose. Teachers strive to promote student learning. This requires careful plan-
ning and implementation, often in relation to statutory curricular guidance. The use
of assessment is an intrinsic part of this teaching and learning environment. A fun-
damental role for assessment, therefore, lies right within the classroom, close to
teaching and learning. Teachers need to monitor how well students are learning, the
successes they are experiencing and the difficulties being encountered by them. Only
when teachers are aware of the learning profile and trajectory of students, can they
really help them succeed.
                                                        Assessment    in education     3
and thus schools and teachers frequently share data with parents, as discussed later
in Chapters 8 and 9. Yet there are many other stakeholders who have an interest in
information about the learning and progress of individual students and cohorts of
students in schools. It is important that teachers recognise other stakeholders’ needs
and ensure that these needs are met where appropriate. Table 1.1 highlights four
broad categories of assessment purpose, identifying a range of stakeholders within
and across categories.
   Many of these needs are obvious, as with those directly associated with supporting
learning at class and school level. Other needs exist at levels more removed from the
classroom: aggregate-level data might be used, for example, by local authorities to
consider resource allocation across schools; or inspectors might wish to consider
trends over time in the same school. Similarly, policy-makers such as ministry officials
frequently require data to justify existing budgets, argue for more funding or evaluate
curricula. Admissions officers use students’ results in secondary education to help
select and allocate students to third-level courses. Similarly, employers draw on
students’ assessment outcomes as part of recruitment processes. Although teachers
individually in class may sometimes not fully see their part in the overall educational
endeavour, the scale of educational expenditure means that politicians, the media and
the wider public have considerable interest in outcomes also. This is not surprising
given estimates of annual contributions to economies by educational activity: US$2
trillion worldwide, £28 billion in the UK and €900 million in Ireland (Lynch et al.,
2012). Spending on education institutions averaged 5.3% of Gross Domestic Product
across OECD countries (OECD, 2015a), over 6% in the UK and 5.6% in the Republic
of Ireland. This suggests a very significant investment by government, other agencies
and individual students and families in education. Little wonder then that there is
interest in how well the system and the schools within it are performing. Some of this
interest, especially from the perspective of policy-makers, has given rise to the estab-
lishment of accountability mechanisms in many countries; some in the form of league
tables of schools, despite criticisms of such approaches.
to just two: promoting learning and finding out what students have learned. As a
teacher, you will routinely monitor how students are doing in class and adjust your
goals, instructional approach, resources or. expectations accordingly. Traditionally,
three purposes for assessment are identified-in classroom and school settings: forma-
tive, diagnostic and summative. Assessments that serve formative purposes are
implemented while teaching is under way, during lessons or during particular units
of work. The key consideration is that the assessment takes place during instruction
or an instructional phase and that the information yielded is used to help students
learn. (Further detail is provided in Chapters 4 and 5.) Sometimes, teachers encoun-
ter student responses to formative assessments that do not immediately suggest how
best to proceed. In such cases, the use of diagnostic assessment may be appropriate.
Diagnostic assessment can help identify specific difficulties that students encounter,
often in relation to language and mathematics. Diagnostic tests are generally devel-
oped by commercial or research agencies and frequently administered by specialist
support teachers.
    Assessments can also have a summative purpose in summing up students’
achievement at a particular point in time, for example at the end of a unit of work
or of a term, or at the conclusion of a point in schooling such as the end of com-
pulsory education. Summative assessments can be designed by teachers or by
external agencies and usually focus on summarising students’ achievement across
a wide range of learning. Examples of externally developed summative assessments
in Britain and Ireland include      assessments      for GCSE, A Levels, Scottish National
Qualifications and the Irish Leaving Certificate. In your teaching, you will need to
become confident and proficient in developing, selecting and using assessments for
summative purposes.
   As highlighted in the previous section, educational planners and policy-makers
require information about the progress of students in school. Patterns of grades from
public examinations     and qualifications    such    as GCSE   (England,     Wales,   Northern
Ireland) and Junior Certificate (Republic of Ireland) offer some limited information,
focused on very particular points in schooling. Student achievement data are some-
times gathered at system level throughout the school years on a census basis,
whereby all students are assessed. Examples are the SATs in reading, grammar and
mathematics at KS2 in England and the Foundation Skills Assessment of students in
Grades 4 and 7 in British Columbia. These approaches are not without their critics,
many of whom highlight a tendency amongst some teachers to ‘teach to the test’,
focusing significant amounts of class time on the type of content and skills that are
contained in the tests. Alternative approaches to system monitoring involve sampling
                                                                                            in
surveys such as the National Assessments of English Reading and Mathematics
Ireland (2nd and 6th  Class) and  the National   Monitoring Study    of Student Achievement
                                                                                          you
in New Zealand (Years 4 and 8). It is likely that at some stage in your career,
will be asked to cooperate    with  such  a  national sampling   survey.
                                                                                            or
   Finally, an extension of national system monitoring is international surveys
                                   typically  administ ered in first  language,  mathemat ics
studies of student achievement,
                                                                             (PISA) aimed at
and science. The Programme of International Student Assessment
                                                                                            in
15-year-old students and now involving over 70 education systems is discussed
                                                                              onal surveys
 detail, along with other international studies, in Chapter 3. Most internati
                                                                              onal data
 go beyond the mere testing of students and also gather survey/observati
                                          s, school principals , teachers, students and
 from policy-makers, curriculum developer
             iy
6     Understanding and applying assessment in education
parents to form a more holistic picture of the educational inputs and processes that
might shape outputs such as student achievement. Results are used to inform policy
development in individual countries. PISA results in 2009, for example, prompted
significant proposed policy change in the assessment of students at the end of the
Junior Cycle in Ireland. In Wales, students performed below the OECD and UK aver-
age in all three areas tested in 2012, prompting the education secretary to declare
that ‘PISA results were not good enough. The 2012 results confirm my view and that
of my predecessor that standards in Wales are not high enough and must improve’
(Lewis, 2012: n.p.), comments more or less echoed in relation to results in Northern
Ireland (Northern Ireland Executive, 2013). As with national surveys, it is likely that
both you and the students in your classes will be asked to participate in such surveys
by completing tests, questionnaires or other inputs.
   As seen above, legal obligations provide one context for teachers’ engagement
with assessment. Another influential factor is standards established by statutory and
other professional organisations that regulate teachers and their work. An analysis of
teacher professional standards in five education systems reveals a common core of
standards evident to a greater or lesser degree across all systems, as summarised in
Table 1.2. The Irish, English and Scottish standards apply to all or the majority of
teachers in the education system. The Australian and US standards, though not bind-
ing on all states or teachers, represent professional codes that have considerable
support and influence nationally. All codes include criteria relevant to teachers’
knowledge and expertise in relation to assessment.
Table 1.2 Common elements of expected professional standards for teachers: general
Sources: Ireland, Teaching Council (2016); England, DfE (201 1a); Scotland, GTC for Scotland (2012); Australia, AITSL (2014);
USA, National Board for Professional Teaching Standards (1989)
In the USA, the Standards for educational and psychological testing (AERA et al.,
2014) is the latest edition in a series of standards published jointly by the three influ-
ential professional organisations since 1966. Like the European Framework, the US
Standards aim to establish expected norms of good practice in assessment and pro-
vide criteria for evaluating the quality of practice. The standards are aimed primarily
at test developers and users of assessment.
progress or at the macro level of national tests, the dual intent is fundamentally
the same: evaluate and plan. When we consider the overall professional standards
expected of teachers, the need for teachers to work collaboratively becomes
apparent. Teachers can usefully collaborate with other teachers and other stake-
holders in planning, teaching, assessing and reviewing their own and school and
system performance.
   Given the array of assessment instruments, approaches and advice available to
teachers, it is easy to get caught up in issues of administration, recording and report-
ing without quite understanding the place of assessment in teaching and learning.
Having a clear set of assessment principles to guide planning and practice helps
bring coherence to teachers’- and systems-level practice. Over a number of years,
the authors have been involved in supporting system-wide change in assessment in
a number of countries. This involved providing support to policy-makers, school
leaders, teachers and. teacher educators, in addition to our sustained involvement in
teacher education in Ireland, especially focused on assessment. The assessment
principles highlighted below draw from that broad experience at the system and
classroom levels to articulate principles that should inform the conceptualisation,
development and enactment of assessment in practice.
   Our assessment vision emphasises eight principles:
7.  Assessment involves the use of continually evolving approaches and tools. This
    requires informed, competent use so that assessment can provide information
    that is accurate, interpreted carefully and communicated appropriately.
8. Teaching, learning and education more broadly should be guided by evidence:
    evidence of what is happening, what is working, what is not working and how
    processes and experiences can be improved. Assessment offers information that
   can be interpreted and used in evidence-led practice and policy.
These eight principles underpin the successful use of assessment within educa-
tion systems, schools and classrooms. There is a particular onus on the teacher
to reach an appropriate standard of knowledge and competency so that assess-
ment can be used judiciously. What the principles emphasise also, however, is
that technical competency is not sufficient. Professionalism involves careful
judgement and the fair, ethical use of approaches that serve the intended pur-
pose, while retaining a focus on the needs and rights of the parties involved, most
notably learners.
   Situating these principles within broader educational structures and processes
is summarised in Figure 1.1. The model highlights the associations between
selected curriculum and assessment inputs and regulations, teacher standards and
competencies, assessment principles and stakeholder needs.
               Teacher
             professional
              standards
                               Assessment
                                regulations                                  Curriculum
                                                      Individual
                                                         and
                                                       System        )        Teaching
                                                       Needs
                               Assessment
                                 principles                                  Resources
                Teacher
              assessment
             competencies
Figure 1.1 Systemic overview of teacher professional and assessment standards in context
degrees, sometimes very tightly, sometimes with greater latitude left to the
teacher. Regardless of this, it is up to teachers to use professional judgement in
‘bringing’ the curriculum to students andin how to monitor student learning.
Assessment principles help ensure that system and teacher practices work coher-
ently to serve agreed, fair purposes, with the teacher as mediator, exercising
professional judgement.
Chapter summary
What will schools be like in 50 years’ time? What will the experiences of teachers and
students be like? Given the pace of change in the last two decades, it is difficult to
offer reliable predictions. What we can probably say is that information will remain
central, particularly information to facilitate learning and inform others about what is
happening in schools. This chapter emphasised the central place of information and
of stakeholders’ needs in relation to education and assessment in particular. Later in
the book, we will explore different interpretations of assessment: for learning, of
learning and even as learning. Teachers need to acquire the appropriate competency
in assessment in light of legislative and regulatory requirements and as part of their
wider professional role and identity. Various people and agencies have a legitimate
right to information yielded by assessments. If used appropriately and subject to
sound principles of use, students, teachers, parents and wider society are all the bet-
ter for the availability of this information. Teachers, with parents, are positioned clos-
est to students throughout their education. But the important work under way in
classrooms resonates     in the local and wider environment.       Little surprise, then, that
many people      have an interest in schools’ work and seek information about the out-
comes of that    work. This book is predicated on the concept of multiple, complemen-
tary functions   for assessment. The challenge for teachers is to acknowledge this and
to shape and      work with educational processes and structures for individual and
societal good.
   1   Identify any aspects of your assessment practice that are formally required of you as
       part of school or other regulations. Discuss with a colleague who it is that requires
       such action/information and why.
   >   Access the most relevant set of professional standards governing your role/future
       role as a teacher. Which specific standards or elements of them relate to assess-
       ment? How do you rate yourself in relation to any assessment-related standards?
       Which aspects do you need to work on?
   3   The psychologist Robert Ebel (1980) posed the question ‘what are the conse-
                                                                                     impacts
       quences of not testing?” Identify a range of possible positive and negative
       of not assessing students in the educational setting with which you are most directly
       involved.
12     Understanding and applying assessment in education
Further reading
McCormack, O., Lynch, R., & Hennessy, J. (2015). Plastic people in pinstripe suits: an explora-
   tion of the views of Irish parents on the publication of school league tables. Educational
   Studies, 41 (5), 513-33.
One use of assessment results in a number of countries is the compilation of school perfor-
mance or league tables, based on aggregated student test data. In Ireland, such practice is
prohibited by legislation, though proxy estimates are used by some media organisations.
McCormack and colleagues provide interesting and perhaps surprising insights into the views |
of Irish parents on the possible publication of school league tables.
Validity
At the outset, it is fair to say that validity is one of the most contested concepts in
educational assessment (Newton and Shaw, 2014, 2016; Geisinger, 2016). Definitions
of validity, and descriptions of the test validation process, have changed in significant —
ways over the past century, and will continue to evolve in the future.
    The Standards for educational and psychological testing (AERA et al., 2014), men-
tioned in Chapter 1, define validity as ‘the degree to which evidence and theory
support the interpretation of test scores for proposed uses of tests’ (p. 11). This defi-
nition highlights the view that it is the proposed uses of a test (or other assessment)
that are validated, rather than the test itself. It implies that evidence may need to be
gathered from a range of sources to justify what a score (or grade) means and how
it is interpreted in a particular situation.
    The Standards make it clear that validation is a joint responsibility of the test devel-
oper (often an organisation charged with developing tests) and the test user
(a teacher or counsellor who administers the test, interprets the outcomes and uses
the results to make decisions). The test developer is responsible for providing a
rationale and evidence to support users in interpreting test scores in particular ways.
For example, the developer of a test of reading comprehension might provide evi-
dence to support use of the test to identify students with high and low levels of read-
ing comprehension. The test user (teacher) is responsible for evaluating the validity
evidence as it relates to the particular setting in which the test is to be used. The test
user must decide if use of the test in his/her school or classroom can be justified on
the basis of information provided by the test developer and experience with using
the test. While subject teachers in secondary school may not have:a say in relation to
whether their students take or do not take examinations such as A levels, they will
have a keen interest in the content of such tests and how they are scored.
    Experts in test development have identified a number of types of validity evidence,
including evidence based on test content (sometimes referred to as content validity),
the internal structure of the test (construct validity), the predictive power of the test
(predictive validity) and the consequences of test use (consequential validity).
Validity is often regarded as a unitary concept and so we refer to different types of
evidence for the validity of test interpretations, rather than different types of validity,
It is not necessary to provide all these types of validity evidence for a particular test
interpretation. For example, the developers of an examination covering a two-year
course of work may only be interested in the extent to which performance on the
test represents how well students have mastered the course content, and may not be
concerned with predicting future performance or tracking the progress of cohorts of
students from year to year.
   In developing specifications such as this, the test developer will need to justify the
emphasis or weighting placed on each mathematical content area and process so that
the weighting given to each cell matches its weighting in the curriculum. Otherwise,
there is a risk that test users may misinterpret the meaning of test scores. For exam-
ple, if a teacher administers a test that includes number and algebra items only, and
another test user assumes that all the content areas are represented, the test user may
misinterpret what a high (or low) score on the test means.
   Assessments developed by teachers for use in their own classrooms typically
have good content validity. Teachers are familiar with what has been taught and
may be in a better position than an external test developer to ensure a direct match
between the content of the curriculum (as it has been taught) and the content of
the assessment. However, many curricula and syllabi specify standards or learning
outcomes and these will need to be translated into test questions and scoring
schemes especially for tests assessing national standards.
   An important element of test content is the stimulus that accompanies a test
question. In a test of reading comprehension, the stimulus might be a narrative or
informational text that is deemed suitable for the class level or age group being
assessed. Alternatively, it might include web pages and multiple texts (two or more
texts around a common theme or topic) associated with a computer-based test. In
a science test, the stimulus might be a diagram of the solar system or a chemical
formula. The test user will need to satisfy themselves that the stimuli that appear
on a test are appropriate for and fair to the students concerned. Similarly, the test
user will need to ensure that the content of the scoring key (marking scheme) that
is used to score student responses is appropriate and reasonable, and consistent
with the curriculum that students have been taught.
   Traditionally, curricula and syllabi have described the knowledge and skills that
students should learn in school. A curriculum may be described for a particular grade
level, key stage or course of study. Curricula may also include learning standards or
                ay
16     Understanding and applying assessment in education
outcomes (also known as attainment targets) that students are expected to achieve.
In England, for example, the curriculum document for state-supported primary
schools (DfE, 2014a) includes attainment targets for different key stages (grade
bands). The attainment targets for Number—Fractions (a component of mathematics)
for Year 2 (the second year of Key Stage 1) are as follows:
   Pupils should be taught to:
e    recognise, find, name and write fractions   ¥%, 4, % and    % of a length, shape,
     set of objects or quantity
e    write simple fractions, for example, “% of 6 = 3 and recognise the equivalence of
     % and hr.
An assessment based on Key Stage 1 mathematics, whether developed by a
teacher or by a test development agency, will need to. ensure that there is an
alignment between attainment targets like these and test questions. The test will
need to include questions designed to assess whether or not students have
achieved these attainment targets. By implication, the test should not include
more complex items (such as asking students to find % of 80) since these cannot
be linked in a systematic way to attainment targets for Year 2 and would be unfair
to the students concerned.
                                   A                 88 .              98
                                   B                 88                93
                                   Cc               105               103
                                   D                119               140
                                   =                 95                90
strong association. The number of subjects represented here (five students) is small.
Sometimes the test scores of thousands of students may be correlated in this way.
   Correlation coefficients range from —1.0 to 1.0. If a negative correlation (e.g. —0.6)
occurs, it means that, as performance on one test increases, it decreases on the other.
When the value of the correlation coefficient is 0, it means that scores on the two
tests are not related. Isaacs et al. (2013) provide the following guidelines for interpret-
ing positive correlation coefficients: 0.0 to 0.2: no relationship or very weak correla-
tion; 0.2 to 0.4: weak correlation; 0.4 to 0.7: moderate correlation; 0.7 to 0.9: strong
correlation; and 0.9 to 1.0: very strong correlation. Negative correlations can be
interpreted using the same criteria. For example, a correlation of —0.5 can be inter-
preted as a moderately strong negative correlation.
    A difficulty can arise if a test (for example, a test of mathematics achievement)
assesses a narrow range of processes, or a restricted set of content skills, when it
claims to assess the full range. If a published test claims to assess a mathematics cur-
riculum that includes problem solving, but instead includes questions based on lower-
order skills only, and this is not made clear in the test manual, test users (teachers)
may misinterpret what test scores mean and may make incorrect decisions based on
test scores. They might assume that students with high scores on the test are proficient
in all aspects of the curriculum, including problem solving. When test scores reflect a
reduced range of skills (relative to what is expected), we have what is described as
construct under-representation.
    Another difficulty can arise if students taking a test are required to demonstrate
knowledge or processes that can impact on performance in a manner that is not
                   sy
18    Understanding and applying assessment in education
intended. For example, students taking a mathematics test may be required to read
mathematical word problems embedded in complex vocabulary and sentence struc-
tures before they can demonstrate their problem-solving skills. Teachers may assume
that the resulting scores reflect performance on mathematical problem solving, when
in fact they also reflect reading proficiency. Since differences in higher-level language
proficiency are not relevant to the construct being assessed (i.e. mathematical prob-
lem solving), and may disadvantage certain subgroups (such as students with low
language proficiency), the reading element of the problems can be described as
construct-irrelevant variance. This doesn’t mean that it is inappropriate (invalid) to
assess students’ ability to solve word problems by asking them to read problems
embedded in text. Instead, test developers and test users (including teachers) should
ensure that grade-appropriate vocabulary and sentence structures are used in the
assessment tasks that are set for students, and hence that test scores reflect the pri-
mary construct being assessed. Construct-irrelevant variance also arises when assess-
ing students whose first language differs from the language of teaching in class. For
many education systems, this is increasingly relevant given the multi-cultural profile
of students, where many arrive in school speaking a different language to that of their
teacher and classmates. Further discussion on this is contained in Chapter 10, in the
context of inclusive assessment.
   Another area where construct-irrelevant variance could be an issue is e-assessment.
If access to the construct (for example, scientific reasoning) first requires a strong
level of computer skills, the scores achieved by students may reflect more than sci-
entific reasoning, and hence proficiency in computer skills could be described as
construct-irrelevant variance. A way to address this is to ensure that the computer-
specific skills students require to access the test and respond to questions are kept to
a minimum, and that, where relevant, students are provided with practice and feed-
back on their use before testing begins (for example, through completion of practice
items that highlight those skills).
measure (exam results) may not fully capture the relationship between the two meas-
ures, since it maybe based solely on those students who stay on at school.
In addition to the different sources of evidence considered here, the validity of test
score interpretations can be undermined by, for example, difficulties with the adminis-
tration of tests or with the scoring of test responses.
               iy
20     Understanding and applying assessment in education
    From time to time, stories in the media point to instances in which the validity of
test score interpretations could be challenged or where a programme is cancelled to
avoid misinterpretation of what scores mean. Examples include the following:
The focus of these concerns is not on content or on the constructs underlying the
tests (which journalists are generally not qualified to interrogate). Rather, it is on
issues such as test security, test administration and scoring or marking. Problems in
these areas can undermine the validity of test score interpretations and could lead to
incorrect conclusions about the meaning of test scores. For example, employers,
parents or other users of test scores might assume that an assessment has been
administered correctly and fairly, or that a script was scored carefully and accurately,
and accept the scores or grades achieved by candidates at face value.
   In your teaching, you will regularly develop assessments and use methods and
tools made by others. In all of this, validity issues are relevant. In this book, we pro-
vide an overview and practical application of some of the dominant views on validity,
a complex and contested concept even amongst assessment experts. The further
reading at the end of the chapter includes a special issue of one educational journal,
Assessment in Education: Principles, Policy and Practice (volume 23, 2016), edited by
Paul Newton and Jo-Anne Baird. Interested readers can explore divergent viewpoints
on validity in this fascinating professional joust.
Reliability
Reliability refers to the consistency of scores derived from an assessment (whether
a test, an examination or a task designed by a teacher). Reliability is a necessary
                                                Validity, reliability and fairness    21
condition for validity, but a test that is reliable may not be valid (that is, while
scores may be accurate and replicable, they may not measure what they are
intended to measure). In considering the reliability of an assessment, we may ask
questions such as:                             '
e   Are the results (scores, grades) consistent over time? Would the same results have
    arisen if the assessment had been administered at a different time? Could the
    student’s scores be replicated?
e   Is there consistency across assessment tasks? Would the same results have arisen
    if a different set of tasks had been administered?
e   Is there consistency across markers? Would the results have been the same if a
    different marker had scored the assessment?
In general, where there are high levels of consistency, we can say that scores or
grades are reliable. However, no scores are completely reliable, as there is some
random error associated with all grades or scores. The goal should be to reduce
sources of error over which we have control, such as error arising from differences
between markers, or error arising from a particular set of items within the test.
   Clearly, it is important for scores and grades to be reliable. Students taking high-
stakes examinations such as GCSEs, A levels, Scottish Nationals or the Irish Leaving
Certificate will want to be reassured that the grades they receive are a fair representa-
tion of their work and that their grades can be compared with those achieved by
other students taking the exam. Similarly, parents will want to be reassured that the
scores achieved by their children on achievement or ability tests are accurate, and
reflect knowledge and skills. Others, including the general public, will want to be
reassured that grades on high-stakes assessments are comparable from year to year
and that standards are not ‘dumbed down’ over time.
e   Test/retest reliability — the same test is administered to the same students at about
    the same point in time. Scores on the two tests are then correlated. If the cor-
    relation is strong or very strong (0.7 or higher), it can be concluded that test
    scores are consistent.
e   Alternate forms reliability — if there are multiple forms of the same assessment
    (for example, Form A and Form B), these forms are administered to the same
    students and their scores are correlated. If the correlation between forms is
    strong, it can be concluded that the different forms of the test are consistent with
    one another.
e    Split-half reliability — the responses to items taken by the same students on the
    same assessment are split into two halves, and scores on each half are correlated
               i
22     Understanding and applying assessment in education
A number of factors besides the internal structure of the test can contribute to the
strength of reliability coefficients reported in test manuals. Kubiszyn and Borich (2013)
provide the following general principles related to the evaluation of reliability:
e    Group variability affects the size of the reliability coefficient. Higher coefficients
     (correlations) result from heterogeneous groups (where there is a good spread
     of performance) than from homogenous groups (where scores cluster together,
     with small differences in performance between students).
e    Scoring reliability limits test score reliability. If tests are scored unreliably (.e.
     there are differences between markers), error is introduced that will limit the reli-
     ability of test scores (though, in the case of good multiple-choice questions, this
     source of error is minimised).
e    All other factors being equal, the more items included in a test, the higher the
     reliability of scores.
e    Reliability of test scores tends to decrease as tests become too easy or too difficult
     (again, a clustering effect with students clustering together at the top or bottom
     of the scale).
Finally, it might be noted that, even if a test has satisfactory reliability, there may still
be issues with the validity of test score interpretations. If the underlying construct is
undermined (for example, by the inclusion of questions that are not relevant to the
curriculum), the validity of test score interpretations will still be threatened.
Question: Why did the Cabinet Mission of 1946 fail to achieve a united India?
   Level 3 — Explains reasons (e.g., The Muslim League felt that the transfer of power from the British
   Raj to a League/Congress coalition would result in a Hindu Raj, with Indian Muslims suffering at
   the hands of a Hindu majority. The initial rejection of a proposed Hindu-majority India and a Muslim-
   majority Pakistan led to swaths of violence which, ultimately, led to the partition of India.)
   Level 2 — Identifies reasons (e.g., Congress and the Muslim League did not agree on a united
   government).
   Level 1 — Simplistic statement (e.g., the Cabinet Mission was a failure, describes the mission).
   Level 0 — No valid response
several hundred scripts are marked by teachers within a subject department. The term
‘inter-rater reliability’ is often used to describe the reliability achieved by two or more
raters scoring the same work.
   A number of approaches can be used in implementing moderation to ensure that
an assessment is marked reliably and in line with required standards (that is, high
levels of inter-rater reliability based on the marking scheme are achieved). These
include:
e    Trial marking, where markers mark a number of questions or papers, and marks
     are standardised (adjusted) by a marking team, in line with agreed criteria.
e    Double-blind marking, where two examiners mark or grade the same question
     or paper independently; a percentage agreement can then be calculated.
e    Sampling, where a moderator reviews a cross-section of a marker’s work to
     standardise the marks awarded. This may result in the statistical adjustment of
     marks if they are higher or lower than the required standard.
e    Group moderation, where a group of markers discuss examples of assessed
     work in order to share their understanding of the agreed criteria.
Gipps (1994) and others have noted the value of groups of teachers within a subject
department working together to establish standards for scoring students’ work. In
addition to enhancing the reliability of the marks awarded to students, group mod-
eration can enhance teachers’ understanding of the criteria that students need to
achieve, which, in turn, can impact in positive ways on teaching and learning (see
Klenowski and Wyatt-Smith, 2014 for more information on moderating teachers’
judgements of student performance).
Fairness
The Standards for educational and psychological testing (AERA et al., 2014) iden-
tify fairness as a fundamental validity issue that requires attention at all stages of
test development and use. A key issue is the extent to which unfairness or bias
could impact in a negative way on the interpretation of test scores. This, in
turn, relates to the construct being assessed and whether the construct is the
same for different groups of test takers, such as students with a learning disability
                                               Validity, reliability and fairness    25
(versus non-disabled   students), female (versus male) students, students with low
proficiency in the language of test (versus highly proficient students) or education-
ally disadvantaged (versus non-disadvantaged) students. Fairness does not imply
that students in different groups should achieve the same average scores. Rather,
it implies that assessment practice and the interpretation of assessment results must
be fair for all students (Gipps and Stobart, 2009). In a similar vein, Popham (2012)
argues that it is unfair to assess a student on a test of English reading if they have
not had a reasonable opportunity to acquire the language (that is, the resulting
score could not be considered valid).
In all cases, the goal is to ensure greater equality of access, without changing the
nature of the underlying construct: In cases where the construct is altered (for
example, by providing a waiver from the assessment of spelling on a language test
because a student has severe dyslexic difficulties), an explanatory note may accom-
pany the grades achieved by the student. Readers are referred to the Joint Council
for Qualifications (2016) and State Examinations Commission (2016) for additional
information on the types of accommodations that are available to students taking
examinations in the UK and Ireland. However, the issue of providing suitable
accommodations will arise for a broad range of assessments, including those
administered on an ongoing basis in the classroom. Hence, teachers will need to
take steps to ensure that their own classroom assessments — whether tests, port-
folios or performances — are fair to all students. We return to the issue of assessment
              4
26     Understanding and applying assessment in education
Ethical assessment
From the foregoing, it is clear that fairness has an ethical dimension, where test
developers and test users have an obligation (and sometimes a legal requirement) to
ensure that assessment is fair. Indeed, an emphasis on ethics/fairness and the rights
of the individual is one of five guiding principles underpinning the European
                                                  Validity, reliability and fairness      27
Chapter summary
The key issues discussed in this chapter — validity, reliability and fairness — are important,
whether the focus of assessment is improving students’ learning in the classroom
(assessment for learning) or summarising student performance at the end of a course or
school year (assessment of learning). Moreover, they relate to all aspects of assessment
and to all types of assessment.
   Validity was presented as a unitary concept, for which various sources of evi-
dence should be provided. The idea that validity evidence should focus on the
interpretation of assessment outcomes (scores), rather than on the assessment itself,
was emphasised, as was the obligation on test users to justify the uses they make
of assessment data, if evidence for those uses has not been provided by test devel-
opers. The concept of consequential validity, or the effect of an assessment on
teachers, students and other stakeholders, was introduced and linked to fairness in
assessment. Threats to the validity of test score interpretations, including construct
under-representation and construct-irrelevant variance, were examined. Reliability
was identified as a prerequisite for validity rather than being an end in itself. A
distinction was made between approaches to evaluating the reliability of fixed
marking schemes (such as those for tests consisting entirely of multiple-choice
questions) and of scoring schemes or rubrics (such as those used to mark essays or
projects). Moderation was identified as an approach to ensuring consistency (and
therefore reliability) when assessments are marked using a scoring scheme. The
need to také measurement error (a by-product of reliability) into account in inter-
preting scores or grades was considered.
   The concept of fairness in assessment was examined and it was argued that,
while access to assessment (and, in particular, to the construct being assessed) is
an important dimension of fairness, other dimensions should also be considered,
including the effect of assessment on different groups of students. Ethical dimensions
of fairness were also considered, including the responsibility on policy-makers, test
developers and test users to make decisions that always put the interests of the
student first.
28        Understanding and applying assessment in education
Further reading
Black, P., Harrison, C., Hodgen, J., Marshall, B., & Serret, N. (2010). Validity in teachers’ sum-
   mative assessments. Assessment in Education: Principles, Policy and Practice, 17 (2),
     215-32.
This article shows how moderation can support teachers in gaining a better understanding of
what is meant by validity. The focus is on the validity of teachers’ own summative assessments
(assessments of learning) to assess English and mathematics using portfolios at Key Stage 3.
The authors use the term ‘assessment literacy’ to describe teachers’ understanding of
assessment.
Elwood, J. (2013). Educational assessment policy and practice: a matter of ethics. Assessment
   in Education: Principles, Policy and Practice, 20 (2), 205-20.
The main focus of Elwood’s article is on ethical issues around the use of two different tests
used in the transfer of students from primary to post-primary schools in Northern Ireland, with
some schools accepting scores on either test for admission. In addition to issues around the
comparability of the two tests, Elwood considers the impact of the selection tests on students
and their parents.
Newton, P.E., & Baird, J. (eds) (2016). Assessment in Education: Principles, Policy and Practice,
   23 (2). Special Issue on Validity.
This special issue of the journal provides an opportunity for debate between scholars on the
topic of validity. A range of issues is explored spanning divergent understandings of validity
across time, geographic locations and philosophical and scientific perspectives.
What you will learn in this chapter
To arrive at this stage in your education or career, your learning will have been
assessed at several key points or transitions. For example, you may have sat
national tests such as the Standard Assessment Tests (SATs) at the end of a key
stage of learning. You may have completed an examination leading to certification
of your learning at the end of compulsory education such as the General Certificate
of Secondary Education (GCSE) in England, Wales and Northern Ireland, the
Scottish Qualifications Certificate or the Junior Certificate in Ireland. You and thou-
sands of your fellow students may, at some point, have been sampled to take part
in international assessments such as the Trends in International Mathematics and
Science Study (TIMSS) or the Programme of International Student Assessment
(PISA). You may have taken an aptitude test selected by a guidance teacher. The
process of assessing students (or indeed adults) at key points in their lives is often
referred to as ‘assessment of learning’ (AoL) or summative assessment (that is, a
summary of student learning). AoL is typically formal (that is, administered and
scored using standard procedures) and can include teacher judgement           of perfor-
mance as well as testing (Isaacs et al., 2013). A key factor distinguishing AoL from
AfL (for assessment for learning, see Chapter 4) is the use to which assessment
information is put. In AoL, assessment outcomes (results) are typically compared
with those achieved by others, whether individuals, schools, districts or even coun-
tries. In contrast, the outcomes of AfL typically relate to task performance and are
more likely to feed directly into future student learning. Typically, they do not
involve a direct comparison with other students but instead focus on the extent to
which key knowledge and skills have been acquired.                                :
    In this chapter, we look at the origins of assessment of learning and how it has
evolved over the past 100 years. We then examine examples of Aol and their main
features. We also consider the strengths and limitations of AoL.
              if
30        Understanding and applying assessment in education
When you finish the chapter, you should be able to answer these questions:
        What    is AoL?
        Why    is AoL important?
        What   procedures and technical requirements are associated with the use of AoL?
        How    can AoL inform teaching and learning?
        How    is AoL linked to the topic of Chapter 4, assessment for learning?
        What    are some strengths, weaknesses and pitfalls of AoL?
Student level —       Maintaining records and generating          End-of-term summative assessments conducted
internal to school/   reports for teachers, parents and           by teachers (teacher-made tests, portfolio
college               students                                    assessments)
Student level —       Certifying that a student has met           Key Stage 2 Assessments (England and Wales);
external to school/   course or graduation requirements           GCSE and A levels (England, Wales, Northern
college               Selection of students                       lreland); Nationals, Highers and Advanced Highers
                                                                  (Scotland); Junior and Leaving Certificate (Ireland);
                      Meeting statutory needs
                                                                  Baccalauréat (France)
School level          Monitoring trends in student                Use of standardised test or examination results to
                      achievement; Evaluation of teachers,        evaluate school improvement
                      schools or larger units, by school itself
                      or by national/local authority
System level          Monitoring of achievement — year-           Key Stage 2 assessments in England and Wales;
                      on-year or periodic comparisons of          Programme for International Student Assessment
                      students’ average performance at            (PISA)
                      regional, national or international level
                                          AoL: purposes, impacts and challenges               31
   While AoL often occurs in the context of formal tests and examinations, it can also
form a part of classroom assessment. Schools and teachers may conduct summative
assessments at the end of a term or school year, and report the outcomes to other
teachers, parents and students. Students may contribute to these reports by summa-
rising their own performance. The Assessment Reform Group (2006: 5) describes
summative assessment by teachers as:
   the process by which teachers gather evidence in a planned and systematic way in order
   to draw inferences about their students’ learning, based on their professional judgment,
   and to report at a particular time on their students’ achievements.
The Group notes that teachers can assess a broader range of achievement and learn-
ing outcomes than formal tests and examinations, and can also provide information
about students’ learning processes. It also notes that such information can be used
formatively to support student learning.
   AoL can be high-stakes or low-stakes, depending on the context and the actors
involved. End-of-term summative assessments that are internal to a school are
typically low-stakes for students; results may cause some discomfort when they
arrive home, but, generally, there are no long-term consequences for school,
teacher or student. Examinations, especially those that take place at the end of
secondary schooling and may determine the institutions that students attend and
courses they take in further and higher education, are high-stakes for students
and their parents. Other assessments, such as Key Stages 2 tests in England, may
be low-stakes for students but high-stakes for schools, because schools may be
judged on the outcomes. Finally, international assessments, such as the
Programme for International Student Assessment (PISA) (see below), may be
low-stakes for students and schools (since results are typically not reported at
student or school levels), but high-stakes for education systems (represented by
 national education departments).
 Examinations
The earliest recorded evidence of standardised examinations was in China, where
the imperial examinations were administered from the Song Dynasty (AD 960-1279)
onward, though their roots may date back to the imperial university established in
                                                                                 s for
the Han dynasty (206 BC-AD 220) (Encyclopaedia Britannica, n.d.). Candidate
                                                         primarily   with   Confucian
the civil service competed in a series of exams dealing
                                                                                  level.
texts, and, if successful, progressed from local to provincial and national
                                                              original  thinking,   and
Although the exams tended to emphasise rote learning over
                                                            grounded    in a common
form over substance, they produced an administrative corps
                                                                       deemed to be
body of teachings, based on a system of merit. As the system was
                    to be capable   of modernisation,    it was   finally abolished   in 1905.
 too   inflexible
32      Understanding and applying assessment in education
1 (a) (i) Use the Chemistry Data Sheet to help you to’answer this question.
Figure 1
  1 (a) (ii)   Complete the word equation for the oxidation of hydrogen.
                                                                                                            [1 mark]
Name two products of burning coal that have an impact on the environment.
               What impact does each of the producis you named have on the environment?
                                                                                                           [4 marks]
Question 7
                (ii) Write the letter X beside the unit of measurement for work.
                                                                                                       |_| Watt(w) |
                                                                                                       |__| soute(s) |
      Source: SEC, Junior Certificate Ordinary Level Science, June 2015
Standardised tests
The origin of standardised tests, as we know them today, can be traced back to a
number of assessment initiatives in the late 19th and early 20th centuries: written
essay-type examinations used for selection purposes (see above); early psycho-
logical testing, mostly designed to study individual differences in sensation, dis-
crimination and reaction time (associated with Francis Galton and James McKeen
Cattell); the development of statistical methods, in particular correlation methods
(associated with Karl Pearson); and testing to diagnose underperformance (associ-
ated with Alfred Binet and Theodore           Simon) (Du Bois, 1970). The tests of Binet
and Simon were particularly relevant to future developments as they consisted of
a wide range of separate items, using different types of material, and were
designed to assess higher mental processes (some of which are now considered
aspects of intelligence), such as memory span, problem solving and judgement
(see the profile of Alfred Binet below).
     While the earliest standardised tests, such as those of Binet and Simon, were
individually administered, it was only a matter of time until tests of achievement
that could be administered to groups were developed. A need for large-scale test-
ing arose in the USA during the First World War for the selection and placement
of personnel in the US army, leading to the development of the first group test —
the Army Alpha test. A parallel nonverbal group test (Army Beta) was developed
for use with individuals with literacy problems or whose first language was not
English — one interesting early illustration of the type of effort made towards more
inclusive assessment, discussed later in Chapter 10. Development of the tests was
facilitated by the invention of the multiple-choice format (attributed to American
Fredrick J. Kelly) in 1914. Through the 1920s, tests of achievement in a variety of
curriculum areas (arithmetic, English composition, spelling, handwriting) were
developed for use in American schools. Tests were designed primarily to assess
individual students, but test data were also aggregated to assess curricula and later
to evaluate the efficiency of teachers and school systems in delivering the curricu-
lum. Another key milestone was 1936 when the first automatic test scanner, IBM
805, was developed. This enabled the scanning and scoring of answer sheets, mak-
ing it easy to assess large numbers of students efficiently, provided the answers
were objective.
   A standardised test can be defined as:
                                                                        , scoring and
Some definitions of standardised tests specify only administration
                                              aspects  of constru ction   (how the test
interpretation (e.g. Popham, 201 1). However,
                                                they  are  relevant  to  underst anding
was put together) are also important, because
36         Understanding and applying assessment in education
the meaning of test scores (an aspect of test validity). While examinations can be
classified as standardised          tests, the focus in the remainder             of this section     is on
standardised norm-referenced tests, primarily consisting of objective test items such
as multiple-choice items, short-answer items, true—false items, or similar.
Construction
The construction of a standardised test involves a number of steps. These typically
include:
   The sample items in Figure 3.2 illustrate some of the item formats found in
standardised tests. They include: multiple choice, where the test taker selects the
correct response from the four options given (item 1); answer supply, where the
test taker provides a brief answer such as a‘number or a word (item 2); and short
answer, where the test taker responds to a question by writing two or three lines
of text (items 3 and 4). Items 3 and 4 show that even relatively short questions can
tap into the higher-level thinking required to respond to some texts, including an
understanding of complex language and causal relationships. These issues are dis-
cussed further in Chapter 6.
Administration
Standardised tests require a uniformity of procedure in how they are administered.
The materials used, instructions given to test takers, preliminary demonstrations
(practice questions) and ways of handling queries are all clearly specified.
Furthermore, the conditions under which a test is administered relating to comfort,
lighting, freedom from distraction and student interest, cooperation and motivation
should be the same for all candidates. Deviations in administration or in the condi-
tions of testing will affect the interpretation of examinees’ performance. The test
manual should provide clarity on any accommodations that can be offered to stu-
dents, including those with special educational needs (see Chapter 10 for further
discussion on assessment accommodations and the inclusion of students with special
educational needs in the assessment process).
Scoring
The instructions for scoring in the manual accompanying a test must be followed
exactly. Discretion on the part of the examiner is eliminated when selection-type
items (e.g. multiple-choice, true-false) are used. Tests with this type of item
can be, and frequently are, scored by machine, increasing the speed and reduc-
ing processing costs. A further advantage of selection-type items is that they
allow the wide sampling of a domain since responding to an item: requires very
little time.
    As illustrated in Figure 3.2, standardised tests may also include supply-type items
(e.g. short-answer items). While pre-set criteria are generally provided for scoring
such items (much like the marking scheme for an exam question), scoring will not
be as objective as is the case with regard to selection-type items, and steps may
need to be taken to establish the reliability of scoring (see Chapter 2). While such
items can tap into higher-order skills, there are also higher costs associated with
scoring them.
Interpretation of scores
Standardised test outcomes are usually presented with one of two interpretative
frameworks. In the first, it is possible to locate the position of a candidate’s score
38    Understanding and applying assessment in education
in a set of scores. In this case, the standard used in interpreting test performance
is a relative one, and the score given to an examinee is called a norm-referenced
measure, Often, student scores are reported on a scale with a mean (average)
score of 100 and a standard deviation (spread) of 15 (see Chapter 8). The second
interpretative framework is a criterion-referenced one, where performance is
interpreted in terms of the content or skills that a student has mastered or is likely
to successfully apply in taking a test. This idea is also taken up in more detail in
Chapter 8, where the use of proficiency levels as a method for reporting perfor-
mance, and as a basis for planning future learning, is considered.
International                       assessments
Many of the summative assessments that students complete at school will be for the
purposes of maintaining school or individual student records, reporting to parents
and students, or certifying students. However, students may also be asked to take part
in international assessments that are specifically designed to compare performance
across countries and educational systems. International assessments of educational
achievement compare the average performance of students in different countries.
Typically, they involve the administration of the same standardised tests to students
in different countries, and the scaling of performance so that scores can be compared
across countries. Table 3.2 provides summary information about the main interna-
tional assessments of educational achievement. As part of these assessments, ques-
tionnaires are also administered to generate contextual data that can be used to
interpret performance outcomes. The regular administration of the same tests allows
for monitoring of performance over time.
Progress in International Literacy        A study of reading literacy that is implemented every 5 years. In
Study (PIRLS):                            2016, PIRLS was administered in over 50 countries, including England
www.iea.nl/pirls_2016.html                (Year 5), Northern Ireland (Year 6) and ROI? (Fourth class)
Trends in International Mathematics       A study of mathematics and science that is implemented every four
and Science (TIMSS):                      years. In 2015, TIMSS was administered in 57 countries, including
www.iea.nl/timss_2015.html                England (Year 5 and Year 9),-Northern Ireland (Year 6 only) and RO!
                                          (Fourth class and Second year) (see Mullis et al., 2016 for outcomes)
Programme for International Student       A study of reading literacy, mathematics and science that is
Assessment (PISA):                        administered every three years to representative national samples
www.pisa.oecd.org                         of 15-year-olds. In 2015, PISA was administered in 72 countries/
                                          economies including the UK (with separate samples for England,
                                          Northern Ireland, Scotland and Wales) and ROI. (see OECD, 2016a;
                                          Shiel et al., 2016 for outcomes) Unlike other international studies,
                                          PISA aims to assess students’ preparedness for their future lives and
                                          study, rather than their performance on curriculum-based tests
International Civic and Citizenship       A study of students’ knowledge of and attitudes towards civics and
Study (ICCS):                             citizenship. In 2009, ICCS was administered in 38 countries, including
www.iea.nl/iccs_2016.html                 England and ROI. A further round was administered in 2016
Chapter summary
The focus of this chapter is on assessment of learning (AoL) or summative assess-
ment. AoL provides an outcome measure such as a mark or grade that is awarded
at the end of a unit or course of study. It incorporates a range of purposes,
                ¢
42        Understanding and applying assessment in education
including maintaining records and generating reports for teachers, parents and
students; certifying that a student has met course or graduation requirements,
evaluating teachers, schools or larger units; and monitoring (and raising) achieve-
ment from year to year, at school, regional or national levels. These purposes are
achieved by administering a range of different assessments, including examina-
tions, tests and moderated coursework, assessment tasks and teacher assess-
ments. The stakes associated with these instruments can be described as high or
low. Assessments may have high stakes for schools and teachers, and low (or no)
stakes for students, and vice versa.
     Examinations and tests include a broad range of tasks. These include multiple-
choice items, short-answer items, essay questions, interviews and, in some                  cases,
performances. Where these tasks are used for AoL, they must be administered in
systematic and uniform ways, and procedures for scoring and for reporting results
must be consistent within and, where relevant, across schools. If coursework or
teacher assessments are used as a basis for generating grades or marks, modera-
tion procedures should be put in place. As noted in Chapter 2, reliability, validity
and fairness are important attributes of assessment in general, and of AoL in
particular.
    AOL, and examinations and tests, in particular, have been the focus of strong criti-
cism in the literature, with concerns such as teaching to the test, predictability and
grade inflation often raised. Persistent differences in performance among ethnic and
socioeconomic groups have also been identified as a problem with some summative
assessments. Teachers may not derive sufficient information from examinations and
tests to allow them to plan instruction and intervene successfully in students’ learning.
Approaches associated with assessment for learning, outlined in Chapters 4 and 5, are
usually more suitable for this purpose.                  ;
     International      assessments   of educational    achievement     such   as PIRLS,    TIMSS
and PISA are used to compare performance across educational systems and to
monitor progress over time. These studies report performance using both norm-
referenced approaches (where performance is described in relative terms) and
criterion-referenced approaches (where the knowledge and skills of students
performing at different levels of proficiency are described).
     1.   Many aspects, features and practices associated with education and schooling have
          changed a great deal in the past few decades, but has the way we assess students
          changed as much? Try to locate a test or assessment used with students 20, 30 or
          4O years ago or more. You might source this in your school, in a book, online, or from
          someone you know. How different is the assessment from what is used nowadays?
          In what ways? What similarities exist?
     2.   How well informed are parents and the public about the breadth of summative
          assessments in use today within the education system? Write a brief explanatory
          memo for your school’s parents’ association that summarises the role of summative
          assessment.
                                            AoL: purposes, impacts and challenges              43
Further reading
Brown, G.T.L., & Hattie, J.A. (2012). The benefits of regular standardized assessment in childhood
   education: guiding improved instruction and learning. In S. Suggate & E. Reese (eds)
   Contemporary debates in child development and education (pp. 287-92). London: Routledge.
   Available at: www.academia.edu/1964802/The_benefits_of_regular_standardized_assessment_
   in_ childhood_education_Guiding_improved_instruction_and_learning
This book chapter outlines how standardised tests can be used to improve teaching and
learning.
Morris, A. (2011). Student standardised testing: Current practices in OECD countries and a
    literature review. OECD Education Working Papers No. 65. Paris: OECD Publishing.
   Available at: http://dx.doi.org/10.1787/5kg3rp9qbnr6-en
This short paper focuses on how standardised tests that have low stakes for students can be
used to improve teaching and learning. In particular, Sections 4 and 5 focus on the potential
positive impacts of standardised tests.
Murphy, C., Lundy, L., Emersen, L., & Kerr, K. (2013). Children’s perceptions of primary science
   assessment in England and Wales. British Educational Research Journal, 39 (3), 585-600.
This article considers how a range of assessments, including standardised tests, impact on
students, from the students’ point of view. In particular, it explores the effect of tests on
students’ attitudes. towards science.
What you will learn in this chapter .
You have already seen the various uses to which assessment can be put in school.
This chapter expands on some of those purposes, especially in relation to promoting
students’ learning in class. We first explore beliefs in education and psychology that
provide the theoretical foundation for assessment;for learning (A/L). The evolution of
this approach to assessment is outlined, highlighting similarities and differences with
other approaches to assessment with which you may be familiar. We also discuss
research and policy development in relation to AfL, while also highlighting some
recent cautionary comments that merit consideration.
    When you finish the chapter, you should be able to answer these questions:
Pre-cognitive theories
Considerable interest emerged during the late 19th and early 20th centuries in the
process of how people learn. Early behaviourist theories of learning by psycholo-
gists Ivan Pavlov, B.F. Skinner and others focused on modifying student behaviour.
Change in student behaviour, it was argued, could be engineered by the application
of reinforcement (for example, rewards that parents and teachers may be familiar
with, such as the systematic use of praise, stickers, tokens and toys) and through
shaping (directing students towards long-term goals through reinforcing short-term
related goals). Key to these behavioural theories of learning was the belief that
learners modify their behaviour in positive, predictable ways, to bring about a con-
sequence desirable to the learner.
   An alternative social learning theory highlights how learners observe and copy the
actions of others. The psychologist Albert Bandura stressed how learners relate their
behaviour to that of others and also understand reinforcement and consequences in
relation to others.
    Social cognitive theories of learning expand the range of factors that may influence
learning. Not only might learning be shaped by a person’s internal information-
processing skills and developmental stage, but the learner’s dispositions, emotional
state, beliefs about learning, perceptions and motivation may all interplay to impact
on learning. You may have seen this in school where some students feel that they
are no good at mathematics (a self-belief) and as a result are not motivated to engage
with new concepts or persist with difficult concepts.
e     Deep understanding is principled and supports the transfer from one learning
      context (topic, situation, subject) to another.
e     Intelligent thought involves ‘metacognition’ or self-monitoring of one’s own
      learning and thinking.
      Cognitive performance depends on dispositions and personal identity.
      New learning is shaped by prior knowledge and cultural factors and
      perspectives.
e     Learners construct knowledge and understanding within a social cultural context.
     If I had to reduce all of educational psychology to just one principle, I would say this:
     The most important single factor influencing learning is what the learner already knows.
     Ascertain this and teach him accordingly.
students’ learning throughout the instructional process, rather than waiting for
defined moments to administer more formal tests. Teachers also need .to help
students understand the very process by which they learn so that they acquire
the capacity over time to monitor their own’learning without a specific need for
direction by the teacher. Such self-monitoring and regulation of their own learn-
ing are considered key goals for students and have been promoted in national
and international educational policy (OECD, 2013). Finally, teachers need to rec-
ognise that individual dispositions such as emotional state, motivation and self-
efficacy impact on learning so that they can take this into account in planning
and implementing teaching.
    active, constructive process whereby learners set goals for their learning, and then
    attempt to monitor, regulate and control their cognition, motivation and behaviour,
    guided and constrained by their goals and the contextual features of the environment.
Barry Zimmerman (2002) suggests three distinct phases that self-regulated learners pass
through when learning new material or consolidating material already learned:
                  (
48       Understanding and applying assessment in education
1.    Forethought: the learner receives information about the task, sets some goals and
      plans strategies to realise the goals.
2.    Performance: this is the actual learning phase, where the student enacts the plans
      developed during the forethought phase and maintains the willpower to continue
      towards the goal. The learner needs to monitor learning and change strategies if
      goals are not being realised.
3.    Self-reflection: this is a crucial phase of learning. Students reflect on learning,
      think about specific reasons for success or failure and adjust the approach to the
      next learning task in light of such self-appraisal.
Such strategising requires the self-activation of students’ own internal cognitive traits,
emotions and motivation, approaches to learning and teaching that are quite different
from the behavioural and social learning theories outlined earlier.
    Effective AfL requires the active involvement of the learner. This assumes that the
student is motivated to learn and believes that success comes as a result of personal
effort rather than luck or some other uncontrollable factor. Students need to choose to
engage seriously with learning that is directed towards some specific goal. They need to
make an effort in response to that choice and maintain that effort over a period of time.
Without this, effective learning is difficult and students will not realise their potential.
     Activity 4.1
     Try this out - Implementing self-regulated learning
     Consider the four approaches to self-regulated learning listed above (cognitive, meta-
     cognition, willpower and motivation). Identify a number of cognitive strategies that you
     could use to revise the section on What do we know about how students learn earlier in
     this chapter. Devise a plan to gauge how successful your revisionis.The companion
     website to the book includes some prompt sheets to aid you in developing strategies.
     Each sheet contains prompts to help you try some strategies with students in class.
     Record your experience on the sheets.
Alex feels that the test was unfair and the teacher really did not reward him for what
was a good answer. Youcef, however, recognises that he hadn’t really revised well
for the test. Alex attributes his lack of successon the test to external factors outside
his control (unfair test, biased teacher), whereas Youcef takes responsibility, attribut-
ing his poor result to lack of preparation. He recognises his own capacity to deter-
mine the outcome and in this case believes that his lack of engagement, time and
effort resulted in a poor performance.
   The above scenario fits with Bernard Weiner’s (2000) attribution theory. Here,
learners perceive their success or failure in a task under three categories:
Depending on the manner in which a learner attributes success or failure, the moti-
vation to put in the effort to succeed is affected. Self-efficacy and attribution interact
with each other and influence motivation in complementary ways. For example, if
Youcef believes that tardy preparation (internal locus) caused his poor result, then
his self-efficacy is preserved. He can study better for the next test and succeed.
     task to be completed
     outcome expected
     resources used
     complexity of directions
     support offered
     pace of teaching/learning
@
®
6    choice (students select work appropriate to their own needs).
In England, Ofsted inspectors, while not expecting to observe any one standardised
instructional methodology in classrooms, nonetheless equate’ the outstanding quality of
teaching with teachers using ‘well-judged teaching strategies that ... match pupils’
needs accurately’ (Ofsted, 2015: 61). In Scotland, key aspects of the curriculum include
meeting the learning needs of students and embracing assessment for learning to
enable appropriate progress for all learners (HM Inspectorate of Education, 2007) and
ensuring ‘personalisation and choice’ in curriculum and learning (Scottish Government,
2008: 13). Whether by name or not, such indicators of sound teaching reflect in part
Carol Tomlinson’s (1999, 2000) categorisation of differentiation by:
   Practice in a classroom is formative to the extent that evidence about student achieve-
   ment is elicited, interpreted, and used. by teachers, learners, or their peers, to make
   decisions about the next steps in instruction that are likely to be better, or better
   founded, than the decisions they would have taken in the absence of the evidence that
   was elicited.
Assessment            purposes
Background to AfL
When reading about assessment for learning in books, articles, reports and online,
you will encounter a wide variety of related terms. In 1967 Michael Scriven coined
the term formative evaluation, indicating the process of critically reviewing the
effectiveness of educational processes, for example an alternative approach to
teaching mathematics or a new curriculum, while it is under way as a means to
improve its functioning. Rather than wait until the programme has finished or is
long-embedded, review is undertaken during the programme. This concept, later
applied to student assessment by Benjamin Bloom, suggested that teachers can
identify students’ learning levels during the teaching process ‘directly affecting
both the quality of the instruction and the student’s ability to understand the
instruction’ (Bloom, Hastings & Madaus, 1971: 53). Such in-class approaches to
assessment contrasted with a policy emphasis on measurement and external test-
ing, such as a minimum-competency approach to testing in the USA and term or
end-of-year exams in systems within the British Isles, continental Europe and Asia.
Interest in pedagogical uses of assessment and assessment information evolved,
accelerated by the work of the Assessment Reform Group (ARG) in England and
Wales from the mid-1990s. Researchers associated with the ARG identified specific
               {
52     Understanding and applying assessment in education
factors and features associated with assessment that has as its primary function the
improvement of learning, and sought to ensure that government policy, especially
in the UK, was    informed     by research evidence.
   Assessment for learning (AfL) contrasts with assessment of learning, or summative
assessment. This contrast occurs on a number of dimensions: the use to which assess-
ment information is put (AfL stresses the learning needs of the student); how fre-
quently this information is communicated (in AfL, students are constantly apprised on
how they are doing); and the role of the students where the task of self-assessment
in helping students regulate their own learning is central to AfL.
   The evolution of AfL can be understood by reference to the different functions
of assessment as outlined in Figure 4.1, Rather than describing assessment or
assessment formats, these terms highlight the main traditional purposes of
assessment — formative, diagnostic, summative and evaluative — and the uses to
which they are put. All purposes are relevant, to varying degrees, for learners in
school systems. Whether we look at the micro (student), meso (classroom and
school) or macro levels (educational system), the interplay between different
purposes of assessment is evident.
Formative
Evaluative assessment
As mentioned     already in Chapter       3, assessments   are sometimes      used to provide
descriptive information about the functioning of a system such as a school or a
broader educational structure (eg. LEA, region), Often, this information is based on
the assessment of samples of students, such as those selected for national monitoring
tests and international assessments, for example PISA and PIRLS. Summative assess-
ments are varied in form and scope, often aiming to provide a succinct snapshot of
student learning that has occurred during a defined time period, such as the learning
gained during a topic, a term or a course of study, Often, the results of summative
assessment are expressed in the form of grades or levels that relate student attain-
ment to expected levels represented in a curriculum or set of standards. Examples
include end-of-term grades compiled by teachers and results on public examinations.
Summative assessment may be based on specific tests, tasks and exams, but also on
the retrospective judgement of teachers in relation to a body of the student’s work
such as entries in a portfolio assessed summatively.
                                        AfL: purposes, impacts and challenges        53
Formative assessment
Formative differs in purpose from summative assessment, aiming to inform and form
the learning as it proceeds. Waiting until the unit or course has finished to assess
students cannot fulfil this purpose. Formative assessment occurs during learning and
offers the teacher and student information about how learning is progressing. Many
types of ‘informal’ classroom assessments (observation, questioning, self-assessment,
teacher-designed tasks) can provide this information. Timely interpretation of the
results of tests, portfolios and projects can also contribute useful formative informa-
tion that can help guide the next steps in learning. Formative assessment involves
using interactive assessment approaches to better understand student learning and
then intervening during the learning process to gather feedback, which is used to
guide teaching and learning in the direction of the intended learning outcomes. Key
to this process is the generation of information for the learner who actively uses this
to improve learning.
Diagnostic assessment
Figure 4.1 also highlights the diagnostic use of assessment. All teachers recognise
instances where a student’s difficulty with learning is hard to explain. This may occur,
for example, when a student consistently fails to answer simple number operations
in mathematics, perhaps due to an inadequate formation of the concept of place
value, or when a student makes frequent errors of reversal of letters in early writing.
Such scenarios are often first encountered and addressed through formative assess-
ment where teachers try to identify the problem and introduce some corrective meas-
ures. Sometimes the learning difficulty may not be amenable to easy identification by
teachers and diagnostic tests are employed to investigate further and try to pinpoint
the specific learning errors. Diagnostic tests are typically commercially produced and
reference a student’s performance on the test or task to specific pre-set criteria
(criterion-referenced) or to the performance of other students of the same age in the
population (norm-referenced), similar to the interpretation of standardised test scores
highlighted in Chapter 3. Purposes of diagnostic assessment include:
The overall purpose of diagnostic assessment is to help the teacher differentiate the
learning experience to best meet the needs of the student, through establishing a
learner’s specific knowledge, skills, strengths and weaknesses. This illustrates a simi-
larity of intention between diagnostic and formative assessment. Both have indi-
vidual student learning as the focus, both emphasise providing relevant, timely
information or feedback and both assume subsequent action by the teacher, in
              ‘
54     Understanding and applying assessment in education
conjunction with the student (Murchan, 2011). The main difference between the two
functions of assessment is the formality of the procedures, tests and assessment
instruments employed.
   Another related term that you might encounter in your reading on assessment is
classroom assessment, used especially in North American systems. In contrast to the
externally developed standardised tests, classroom assessments highlight the key role
of the teacher in developing, administering and interpreting assessments in the con-
text of normal day-to-day classroom activities. Clearly, there is a strong overlap
between   formative assessment,   classroom   assessment    and AfL, and Wiliam     (2011)
explores the nuanced relationship between different terms. Fundamentally, AfL high-
lights the potential of a wide variety of approaches to assessment that can be embed-
ded within teaching and learning practice in class. A voluminous and growing
literature describes the application of AfL in many different contexts and provides
evidence to support the practice.                           :
Feedback
A key feature of AfL is the provision of feedback to learners and how this feedback
can inform student (and teacher) actions. Feedback is ‘information provided by an
agent (e.g. teacher, peer, book, parent, self, experience) regarding aspects of one’s
performance or understanding’ (Hattie and Timperley, 2007: 81). Feedback is task-
specific and should provide specific information that fills the gap between what is
understood at present and what ought to be understood in relation to the learning
task (Sadler, 1989). A distinction is made between offering praise (not considered to
be particularly effective) and providing information about a student’s performance in
relation to a specific task, content or learning. Effective feedback provides cues or
reinforcement   to learners, whereas   praise, punishment    or extrinsic rewards   (stars,
stickers, grades) are less effective. The role of the student and the focus of the feed-
back are crucial: students must be committed to the goals of learning and the feed-
back must relate specifically to that learning, rather than to personal characteristics
of the students such as their effort or general orientation to study. Three questions
are often asked to reflect students’ engagement with AfL:
1.   Where am I now?
2.   Where am I going?
3.   How do I get there?
In answering these questions, a number of actions are necessary for the learner and
teacher. To gauge the present learning level (Question 1), the teacher and/or learner
needs to gather relevant evidence about students’ current level of functioning in rela-
tion to the intended learning, analogous to understanding prior to learning in the
context of differentiated teaching. To know where learning is going (Question 2),
students need to receive information that enables them to relate their current level of
learning to the intended learning outcome or state and understand also what success
on relevant learning tasks will look like. Typically, there is a gap between where
students are now and where they need to be. Such feedback provides information
to enable the student to self-regulate their learning. Once the nature of the gap
                                        AfL: purposes, impacts and challenges        55
between actual and intended learning is made clear, the learner needs to receive feed
forward in the form of information to help adapt learning approaches and processes
to generate enhanced learning. The above steps, while appearing straightforward,
require the adjustment of practice by teachers and students, rather than the mere
application of a few assessment techniques to traditional teaching. The blending of
specific assessment techniques with teaching and learning practice in class is dis-
cussed in the next chapter. Issues discussed include the articulating and sharing of
learning goals, the effectiveness of marking and feedback, the use of questioning
strategies in class and the deeper involvement of students in assessment through self-
and peer assessment.
Assessment as learning
Ideally, the student should be an active participant in assessment rather than some-
one to whom assessment is done. Motivation and willingness are required on the
part of students to actively engage with the process, share the development of learn-
ing goals, attend to the feedback and monitor their own learning (Dann, 2014).
Thus, engagement with assessment is itself learning, coining the term assessment as
learning (AdL). Normal classroom activities such as questioning, role-playing, com-
pleting a worksheet, working on projects and tasks or designing a model can all be
used to generate information of formative use to students and teachers alike. Given
the right conditions, students come to see assessment as a natural part of how they
learn in school, thus assessment as learning. This interpretation highlights the inter-
relationship between AfL and the concepts of motivation, attribution of cause and
self-regulated learning discussed earlier in the chapter. Whereas assessment can help
the teacher to evaluate and redirect the focus of the planning, teaching, organisa-
tional strategies and teaching methods, fundamentally formative assessment offers
feedback to students that can motivate and promote learning.
   Understanding assessment for learning as part of a wider pedagogical context is
central to proper understanding and to its successful implementation. AfL requires
fundamental adjustments to teaching practices and students’ engagement with learn-
ing, rather than merely attaching some novel assessment methods to teachers’ exist-
ing practice. One of the key elements in understanding assessment as formative is
that the results (feedback) are systematically used as part of the instructional system
in the class or school to promote further learning (Wiliam, 2011: 4). In this regard, it
is unhelpful and artificial to separate formative assessment from teaching pedagogy
and classroom management.
     Students should be fully engaged with their learning, contributing to the planning and
     organisation of lessons, having learning expectations communicated to them, assess-
     ing their learning and that of their peers, and benefitting from individualised support
     and differentiated learning. To become lifelong learners, students need to be able to
     assess their own progress, make adjustments to their understandings and take control
     of their own learning. (p. 13)
That position, strongly aligned with the basic tenets of AfL, is based on evidence for
the value of such practices. Research suggests that the effective implementation of AfL
has a positive impact on student learning and achievement, motivation, behaviour and
ownership by students of their own learning (Wiliam; 2011; Faragher, 2014), with
evidence of particular gains for low-achieving students. Benefits for teachers are also
evident, with more positive, dynamic and collaborative relationships with students,
                                            AfL: purposes, impacts and challenges               57
improved relationships in class, fewer behaviour problems and greater teacher control
over all aspects of professional practice, including assessment (in contexts where
assessment was often imposed from the outside).
   Some cautionary voices have been raised, however, about over-interpreting the
potential of AfL to provide achievement gains that have been described as
‘amongst the largest ever reported for educational interventions’ (Black and
Wiliam, 1998: 61). Bennett (2011), Dunn and Mulvenon (2009) and Kingston and
Nash (2011) explore the research evidence of the past few decades on AfL in a
wide variety of contexts, educational levels and formats. Their conclusions high-
light the complexity of implementing truly formative assessment in practice and
illustrate the need for focused research so that practice is evidence based. The
reservations do not dismiss the potential of AfL to positively impact on learning;
rather, they suggest that achieving such potential is perhaps more complicated
than it might appear. .
Chapter summary
Assessment that is designed to support learning needs to be framed and under-
stood within the many dimensions, features and practices of teaching and learning.
Theory on the nature of learning, learner attributes and the processes of teaching
and learning shape and are themselves shaped by formative assessment. Tailoring
instruction to the needs of students resonates well with policy in many modern
education systems and has spawned many instructional programmes in the past
few decades. Most initiatives require timely assessment of student learning to
reveal the optimal next steps in learning. Whether diagnostic or formative assess-
ment, it is Clear that focused just-on-time assessment intuitively complements
efficient and effective learning. For years, the use of assessment to measure what
students had learnt dominated policy, with effects evident in classroom practice,
not all of them as intended. More recent policy emphases support assessment by
teachers as part of their professional role and the use of these assessments to
direct teaching and learning. There is a considerable but not uncontested research
base for the policy shift. We encourage you to engage with that debate on AfL
with a view to incorporating evidence-based best practice in your professional
practice.
58        Understanding and applying assessment in education
     1.   Develop a lesson plan or strategy for a topic you hope to teach in one lesson. How
          will students know what is expected of them at the end of the lesson? What feed-
          back (oral and/or written) will students get from you? How does this feedback relate
          to what is expected of students?
     2.   Imagine you are asked to make a presentation to a staff meeting on the topic of AfL.
          Make five slides for a PowerPoint/visual presentation showing how ATL links to the
          following:
          e   the instructional approach/pedagogy
          e   the self-regulation of learning by students.
     3.   Table 41 presents a number of definitions related to assessment. Consider how
          each definition relates to the three main questions underpinning AfL: Where am |
          now? Where am | going? How do | get there? Work with a colleague if possible.
‘in the service of learning — the necessary intertwining of testing and teaching ...
woven into learning environments ... offering executable advice to both students and
teachers.’ (Glaser, 1990: 480)                                          .
‘a process of collecting data for the purpose of making decisions about individuals and
groups’ (Salvia and Ysseldyke, 2001: 5)
‘specifically intended to provide feedback on performance to improve and accelerate °
learning’ (Sadler, 1998: 77)
                                                                                          Where am | now?
‘use of a diverse set of data for a purpose. That purpose is the modification of the
learning work to adapt to the needs that are revealed by the evidence. Only when
assessment evidence is acted upon in this way does it become formative’
                                                                                          Where am | going?
(Black, 1998: 105)
‘any assessment for which the first priority is to serve the purpose of promoting
students’ learning ... it is usually informal, embedded in all aspects of teaching
and learning, and conducted by different teachers as part of their own diverse and        How do | get there?
individual teaching styles’ (Black et al., 2003: 2)
‘assessment that’supports the learning process in contrast to assessment that measures
the outcomes of learning’ (Shepard, 2006: 626)
‘the process of gathering, recording, interpreting, using, and reporting information
about a child’s progress and achievement in developing knowledge, skills and attitudes’
(NCCA, 2007a: 7)
eee                                                                                             eee
                                                 AfL: purposes, impacts and challenges                59
Further reading
Bennett,    R.E.   (2011).   Formative   assessment:   a critical   review.   Assessment   in Education.
   Principles, Policy and Practice, 18 (1), 5-25.
This is an analysis of assumptions, research and interpretations made about formative
assessment. While not dismissing the claims made for AfL, the article highlights the con-
testable nature of some of the policy narrative in relation to AfL.
Clarke, S. (2005). Formative assessment in action: Weaving the elements together, and (2008).
   Active learning through formative assessment. London: Hodder Murray.
These two books provide accessible illustrations of some of the key practices associated with
formative assessment, particularly in primary education. The work draws on collaborating
teachers in a number of schools who provide worked examples of lesson plans featuring
strong elements of formative assessment.
Swaffield, S. (ed.) (2008). Unlocking assessment: Understanding for reflection and application.
   Milton Park, Oxford: Routledge.
This edited volume contains chapters drawn from a range of authors in the field, many of
whom are associated with the Assessment Reform Group in the UK. The book highlights
formative approaches, issues of quality and the use of teacher judgements about students for
summative     purposes.
Wiliam, D. (2011). What is assessment for learning? Studies in Educational Evaluation, 37,
   3-14.
This article explores the nuanced differences in interpretations across terminology and prac-
tices related to formative assessment. The paper argues that clearly defined assessment in
support of teaching and learning can improve student engagement and attainment.
What you will learn in this chapter
The previous chapter highlighted the richness and complexity of learning and the need
for teachers to understand how to accommodate content, teaching methods and class-
room management to the needs of students in class. Understanding students’ previous
learning, selecting appropriate content and creating suitable learning experiences are
essential elements of teachers’ craft. Systematic approaches to planning, classroom
management, teaching and assessment help ensure that classrooms and the work of
teachers and students are directed, enjoyable and productive. Good planning requires
the consideration of curriculum guidelines, available resources, student characteristics
and specific instructional/learning intentions and methodologies. Evaluation of the suc-
cess of teaching is also important, both in terms of what and how the students have
learned and the teacher's reflection on the process. The successful implementation of
assessment for learning (AfL) requires that decisions about assessment permeate all
stages of planning, teaching and learning. There is no one AfL event in a lesson or in
a topic. Instead, we can look at assessment itself as learning, a seamless part of students’
experience in class. This chapter illustrates the interconnected links between AfL and
teaching. Readers who are enrolled in teacher education programmes may find some
of the approaches outlined similar to the instructional methods content of some of their
other modules and readings. This is the strength of AfL — it can and should be seen as
part of teaching, not as something appended at the end of the class, topic or term.
   When you finish the chapter, you should be able to answer these questions:
e    How can learners understand their learning better and act on this information
     effectively?
e    What feedback can assist both students ahd teachers in making learning more
     effective?
    Students learn better where          teachers are highly structured in their approach,
cover more content and allocate class time to relevant academic activities; where
students and teachers are on-task in active student-oriented teaching approaches;
and where students are enabled to achieve success in class. Teachers need to be
knowledgeable about the content to be taught and how to teach it, and to think
reflectively about the decisions to be made before, during and after instruction
(Harris, 1998). Assessment plays a central role in influencing what and how stu-
dents learn. The New Zealand educator, John Hattie, undertook a fascinating meta-
analysis of over 800 different studies that themselves reviewed other research
exploring what influences student achievement in school. Drawing on data based
on millions of. students worldwide, he explored variables identified as having an
influence      on   achievement    (most    influences    were     positive,     some      negative).      The
review found that practices associated with formative assessment are amongst the
most powerful correlates of student achievement. In a ranking of 138 potential
influences, Hattie (2009) found that the contributions of formative evaluation, of
feedback and of teaching meta-cognitive strategies were ranked third, tenth and
thirteenth respectively. Figure 5.1 presents effect sizes associated with the top 10
influences on student achievement.
   These influences are not trivial. For example, feedback was associated with a
‘high’ effect size of 0.73, indicating a three-quarters of a standard deviation increase
in achievement where teachers draw on feedback about students’ learning in class.
                     (
   62        Understanding and applying assessment in education
                                                                Northern                         New
                                              lreland'          Ireland?         Scotland?       Zealand‘          Norway*
                          (
64       Understanding and applying assessment in education
     Paul Black's work has spanned formative and summative assessment, though it is for the
     former that he is probably best known. An emeritus professor in the school of educa-
     tion, communication and society of at King’s College London, his work with Dylan Wiliam
     influenced a generation of assessment researchers and educational policy-makers. Their
     co-publication Assessment and Classroom Learning in 1998 collated existing literature
     on classroom formative assessment, concluding that such assessment had immense
     potential for the improvement of educational practice. Black was an active member of
     the Assessment Reform Group in the UK, a group dedicated to the formulation of edu-
     cational policy informed by research. He has been a champion of formative assessment
     through his research, writing and professional speaking worldwide and has been active
     in research and advisory roles since his retirement.
 This case study illustrates the importance of activating students’ awareness of, interest
 in and possibility of engaging with the learning by sharing the purpose of the lesson
 with them. This is important because students often feel that lessons happen and then
 end without their really understanding and appreciating the purpose or direction of the
 lesson (Earl, 2013). Learning requires the active engagement of students themselves and
 a willingness and motivation to persevere and succeed. The chances of this happening
 are increased when teachers clearly communicate to students what they are expected
 to know in language that they can understand. Furthermore, once students are aware
 of what they should learn, the teacher needs to scaffold the learning process and pro-
 vide explanations, illustrations or exemplars of the expected learning. This demon-
 strates an important concept in the use of learning intentions as part of AfL. Learning
 intentions are designed to help students understand what they will learn about during
 an instructional period. Because they are primarily aids to student learning rather than
 to teacher planning, learning intentions should be phrased in terms such as: We are
 learning to understand how electrical currents work; we are learning bow to use a cal-
 culator to solve complex calculations; we are learning about the importance of nutrition
for health; we are learning to use adverbs to improve our creative writing.
                j
66        Understanding and applying assessment in education
    In our own professional development work with teachers and also in working with
student teachers, participants report that their students do engage with beginning-of-
class discussions around learning intentions and that the intentions provide students
with a language and framework for discussing their learning at the end of the lesson.
The more the learning intentions describe the learning rather than the task, the richer
the post-lesson discussion and self-appraisal by students. Furthermore, the shorter the
learning intention and the more accessible the language used to describe it, the easier
it is for students to remember it.
     Activity 5.1
     Try this out — Identify and communicate learning intentions
     In this chapter, we provide some strategies to get you started using AfL. The first two are
     presented below. The companion website includes some prompt sheets to lead you
     through use of the strategies. Each sheet contains prompts to help you try the strategy with
     students in class. Record your experience and reaction to the prompts and follow your own
     learning. Good luck, and remember, there are no right answers to the prompt sheets. We've
     tried these activities out with student teachers and experienced teachers and know that it
     can take a few attempts to become comfortable with the strategies. What really works is if
     you can discuss your experiences with one or more colleagues and learn from each other.
     Convert teacher's learning objectives         Briefly revisit learning intention as part of lesson
     into a student-friendly format. Generally,    conclusion. Ask a small number of students to briefly tell
     present one or two clear learning             the class (share) something they learned in the lesson.
     intentions. If needed, clarify with a small   Teacher and other students can evaluate these comments
     number of sub-intentions. Write on one        in relation to the lesson intention written on the board.
     section of board and discuss with students    Be sensitive in dealing with any student misconceptions
     (2-5 minutes)                                 identified. If possible, address misconception on the spot.
                                                   Otherwise, note and address at start of next lesson
enable them to know to what extent they are successful in learning. At its simplest,
sharing success criteria (or, better still, involving students in the development of suc-
cess criteria) helps students know what to aim for. Students benefit from illustrations
of successful learning, whether through rich discussion, provision of exemplars of
student work or looking at the work of their peers in class. Many teachers use rubrics
to assess work, where dimensions of the work are assigned to different levels of
quality. Success criteria should then be visible to students, for example written on a
whiteboard. Teachers sometimes use strategies such as WALT (We Are Learning To...)
and WILF (What I’m Looking For) to simplify and aid students’ understanding of the
desired learning.
   Ideally, success criteria should not be interpreted or presented like a ‘solution’ to
a test question. Rather, you should offer students clarity about what constitutes
achievement of the learning intention without giving step-by-step guidelines to
follow. Success criteria provide more general statements about what strategies to use
and about what the evidence of success will look like. As highlighted in Chapter 4,
students need to monitor their own learning and take responsibility for that learning
in class. Simply telling students in minute detail what they must do is not the same
as fostering students’ independent learning. Through careful discussion of success
criteria, the teacher and students come to an understanding of the type of strategies
and performances that will demonstrate that students have learned.
   Teachers often identify the success criteria at the same time as the learning inten-
tion and both are communicated to students at about the same time, that is, at the
beginning of the lesson. It may be preferable sometimes to delay an extensive discus-
sion of the success criteria until students have had an opportunity to engage with the
learning through the context of the lesson and/or the assigned tasks. Sometimes only
when students have engaged with the task will they be able to comment more fully
on the evidence for successful learning. This provides a solid opportunity to involve
students more fully in the process of establishing the success criteria. One teacher we
worked with shared her rationale for delaying the introduction of success criteria in
a specific lesson developed as part of a set of lessons on a topic in Business Studies,
noting that ‘success criteria give students a goal and confidence. Students had already
covered the topic up to a certain point so they came up with the success criteria
themselves’.
(Continued)
     e    Identify a new club that would be of interest to some students in the school.
     e    Write at least one rule in Spanish forjoining the club.
     e    Write some text in Spanish for a short conversation about the club.
     The class listened to the recorded conversation and the teacher led a class discussion
     identifying the three success criteria evident. Student groups then engaged with the
     task, writing joining rules and conversational text in relation to clubs for baking, journal-
     ism, dancing, the environment, mathematics and singing.
     Activity 5.2
     Try this out — Explore success criteria
     Use the companion website prompts when trying out the following strategies in class.
     Reflect on your experience with a colleague. Remember, there are no right answers to the
     prompts, only experiences to reflect on. Avoid defining too many success criteria.
     Otherwise, you will essentially provide minute instructions to students in how to complete
     tasks. The responsibility for learning rests with students and they need to make the deci-
     sions to help them decide whether or not they have achieved the learning intention.
     For a lesson of your choice, develop one         Examples of work related to the learning intention and
     main task or learning activity to support        success criteria are shared with students. The examples
     the learning intention. What criteria can        might be newly developed by the teacher or, over time,
     students use to gauge their success in           sourced from students who have completed similar
     relation to this activity? Through discussion,   tasks previously. Try to provide examples that clearly
     help your students to understand the             show different quality levels in the work, perhaps using
     main criteria. Record criteria on the board      two or three levels of quality. In addition to providing
     alongside the learning intention. Express        examples, discuss the characteristics of the examples
     criteria simply and clearly to let students      with the class to ensure that they can identify
     know how they can demonstrate their              (i) the success criteria and (ii) the differences in quality
     learning in relation to the intention            evident in different examples
Wrong answers may be as useful as correct answers since they provide teachers
with opportunities to probe students’ understanding in greater depth and enable
scaffolding of learning. What we also know about questioning by teachers is, how-
ever, that in the form practised it may not be as effective as teachers think. Teachers
typically ask 200-400 questions per day (Wragg, 1997; Hattie, 2009), of which half
focus on students’ recall of facts and information and 20% focus on analysis
(Stiggins, Griswold & Wikelund, 1989). Nor do teachers wait long for students’
answers — some estimates being only about 1 second, despite the obvious fact that
waiting for answers conveys the impression to students that teachers are genuinely
curious about the question and the answer (Hattie, 2012; Kyriacou, 2014).
Increasing the ‘wait time’ by even a few seconds can result in an improved length
and quality of response by students. The challenges with questioning strategies are
also illustrated by research that shows that 70% of student answers are less than
5 seconds in duration, or only about three words. Wragg (1997) reported that in a
study of 1000 teachers’ questioning practices in the UK, one-third of students’
answers met with no response at all by teachers. All this does not suggest the
coherent application of a powerful pedagogical tool by teachers, and we can con-
clude that much questioning practice at present is not really a recipe for rich dia-
logue and debate in class.
   More systematic yet creative approaches to teacher questioning are required if it
is to provide usable information and feedback to students and teachers to optimise
the next steps in learning. For this reason, getting questioning right goes to the heart
of getting AfL right. Below, we identify a number of ways and approaches to max-
imise the chance that the questions that are posed in class are of high quality and
can contribute useful information to teachers and students about student learning.
   Teachers need to distinguish between ‘open’ and ‘closed’ questions. Closed (conver-
gent) questions typically require short ‘correct’ answers reflective of lower-order think-
ing such as memorisation and recall. Typically, there is only one correct answer, as in
‘Is seven a prime number?’ Sometimes we draw on the ‘W’ questions: what, who, when,
where, as in the examples: What is Guy Fawkes night? Who was Guy Fawkes? When
is he remembered each year? Where was the gunpowder stored? We need to realise,
however, that many of these are closed questions, encouraging the student to ‘con-
verge’ on one correct answer. While such recall is undoubtedly important as a first
stage in learning material, exposure only to lower-order questions will certainly restrict
students’ capacity to engage in deep and meaningful learning. In contrast, ‘open’ ques-
tions typically assume a number of possible correct answers, promote more elaborate
             iy
70       Understanding and applying assessment in education
responses and link with higher-order thinking such as organisation, analysis, problem
solving and speculation. For example, asking a student ‘Why is 7 an example of a
prime number?’ helps the teacher to explore better the student’s understanding.
     Mr Remous introduces the poem ‘The Snare’ to his P4 class. He   The Snare
     asks questions to the class and they have to use Think—Pair—
                                                                     | hear a sudden cry of pain!
     Share to address the questions. The teacher uses Anderson and          ,        Ft
     Krathwohl’s category system in developing some questions:       There is a rabbit in a snare:
                                                                     Now | hear the cry again,
     1. Remember: Where is the rabbit? What noise is
        ee   aie   abla’                                             But | cannot
                                                                                tell from where.
          e rabbit   making?
     2. Understand: Why is the rabbit crying? Why is                 But | cannot tell from where
        the rabbit wrinkling up his face?                            He is calling out for aid!
     3. Apply: What do you think will happen next?                   Crying on the frightened air,
     4. Analyse: What evidence in the poem suggests
        that the poet will be unable to help?                        Making everything afraid!
                                                                     Making everything afraid!
                                                                     Wrinkling up his little face!
                                                     implementing assessment for learning                               71
    5. Evaluate: How suitable is this poem for younger children such             As he cries again for aid;
       as those in P1? Why? What might they get from it? -
                                                                                 And | cannot find the place!
    6. Create: The rabbit can speak and calls out directionsto the
       poet to find him. Write up the conversation between the rabbit            And | cannot find the place
       and poet.                                                                 Where his paw is in the snare!
                                                                                  Little One! Oh, Little One!
                                                                                  | am searching everywhere!
James Stephens
Table 5.4 provides further information on how Anderson and Krathwohl’s taxonomy
can be of help to teachers at primary and at secondary school levels in developing
diverse questions to assess different types of thinking and learning.
Table 5.4     Higher- and lower-order question types based on Anderson and Krathwohl’s (2001)
taxonomy
                                                                                                                (Continued)
72       Understanding and applying assessment in education
e    Stage 1: Private — students individually write down their responses to the teacher’s
     oral or written question
e    Stage 2: Intimate — students compare responses with each other through pair or
     group discussion
°    Stage 3: Public — the answers become public, when the pairs or groups are
     invited to report back. Students can assess each other’s answers.
A number of variants to this approach are possible. You can ask pairs or groups to
provide an ‘agreed’ answer, or to pick the best one for communication to the teacher
or class. Think—Pair—Share is very similar in format to PIP. The Think—Pair—Square vari-
ant involves recording individual answers within each group of, say, four students on a
sheet of paper. The group then agrees the ‘best’ of the answers which is recorded in a
square drawn in the middle of the sheet. Whatever variant is used, the aim is the same:
                                                        implementing assessment for learning                         73
increase the amount of time that each student spends engaging with questions posed
in class, and thereby increase the quantity and quality of information available about
student learning. A number of additional strategies are provided below.
Activity 5.3
Peer assessment
Two key theories explored in Chapter 4 relate significantly to peer and self-assessment
practices.
    First, social constructivist interpretations of learning highlight the collaborative
nature of effective learning and the role that students as a cohesive group play in their
own development in school. Embedding constructivist principles in classroom prac-
tice is facilitated by approaches such as using knowledge of student pre-conceptions,
student inquiry, active learning, collaborative learning, and self-reflection by students
and teachers. Given the inter-related nature of planning, teaching, learning and
assessment, constructivist interpretations of learning need to be reflected in the
assessment’ approaches used by teachers and students. The worthwhile effort that
goes into creating communities of active inquiry-oriented learners in classrooms is
undermined if approaches to assessment do not involve students in collaboratively
engaging with assessment tasks, interpretations and reporting. Creating opportunities
for students to review the work of their classmates fosters coherence across teaching,
learning and assessment dimensions. If we view assessment as learning (Dann, 2014;
Hayward, 2015), then the work of students in assessing their peers is learning, and
thus, both assessor and partner should benefit from the process.
   Second, there is an obvious parallel between peer and self-assessment and self-
regulated learning and the processes by which students themselves increasingly take
on the role of their own teachers in monitoring and adjusting their own learning. In
separate studies of Dutch children in upper-primary classes and at lower vocational
level, Pat-El et al. (2011) and Baas et al. (2015) focused on two dimensions     of AfL:
its monitoring and scaffolding functions. In these studies, AfL practices helped stu-
dents to monitor their own learning and engage in appropriate subsequent steps
with the result that metacognitive strategies and cognitive outcomes were improved.
This ‘ceding [of] responsibility to students in taking control of their own learning’
(Baas et al., 2015: 33) can be facilitated through strategies such as peer and self-
assessment. This is a useful way to conceptualise peer assessment. On a practical
level, it can be very challenging and inefficient for teachers to act as the sole asses-
sor of learning in class. Sharing the assessment role with the students themselves
allows teachers to place more of the onus for learning and for monitoring learning
on the students themselves, enabling them to achieve greater levels of intrinsic moti-
vation and knowledge about and control over their learning. Transferring some
responsibility for assessment to students themselves also fits with the socio-cultural
perspective on assessment, as well as ensuring that individual students receive more
feedback more frequently and more swiftly (Topping, 2013).
    For peer assessment to work in class, the learning pairs or groups must be clear
about the learning intentions and associated success criteria. Peer assessment
should involve students in manageable chunks of assessment, whereby both the
specific aspect of learning to be assessed and the criteria and approach to assess-
ing it are clear.   A summary   of some   important points about peer assessment       is
presented below:
e    Peer assessment can provide students with the confidence, skills and tools to
     constructively monitor the learning of other students and also their own learning.
     It should benefit both assessor and assessed.
                                            Implementing assessment for learning             75
e   Students need support and structure for appraising their work. This support
    should provide clarity about the learning intention(s) and success criteria, along
    with advice on how to provide feedback’‘to their peers.
e   Feedback to peers should be specific, linked to the learning outcomes and suc-
    cess criteria, and should be fair and sensitive.
e   Students may need help in distinguishing between friendship groups and learn-
    ing groups. Peer assessment is most effective with learning groups where the
    purpose of the group is to learn.
e   Students need to see themselves as working in a positive learning environment
    that promotes a culture of mutual respect and learning.
e   Peer assessment offers an increased opportunity for on-task dialogue amongst all
    students in the class in relation to specific learning.
These criteria require considerable commitment and skill on the part of students
and we cannot assume that students will already possess such traits and expertise.
Depending on students’ prior familiarity with peer assessment, you may need to
model and coach sound peer-assessment practice. For example, ensure students
focus on the work rather than on the person; ensure critique is positive and
constructive; and ensure that learning groups rotate over time so that students
get the opportunity to work with several different peers and to see a range of
work. Topping (2013) recommends against putting together in pairs or groups
students who are very friendly or those who do not relate well to each other in
class or socially.
   Below, we introduce a few strategies to get you started on peer assessment,
beginning with a case study of one lower-secondary teacher who used a simple
checklist to enlist the help of her pupils in assessing each other’s learning of English
poetic technique.
The case above illustrates an interesting feature of peer assessment, identified by Black
et al. (2003) and Topping (2013). Students often quickly ‘withdraw’ from the teacher's
probing about whether they understand something or not, pretending that they under-
stand when, in fact, they do not, thus missing a rich opportunity for addressing mis-
conceptions. This can be the result of not wishing to reveal that they still don’t know
after a concept has been explained a second time. Ms Allen’s reflection on the checklist
assessment shows that genuine and sustained dialogue about learning and misconcep-
tions are possible in class, though it may happen more amongst students than between
students and teacher!
     Activity 5.4
     Try this out — Promoting peer assessment
     Use the companion website prompts when trying out the following strategies in class.
     Reflect on your experience with a colleague.
Students construct and apply a scoring rubric. The following steps may be used:
     1.   Students complete some learning or product (e.g. dramatic activity, speech, piece
          of writing, project, report on experiment).
     2.   Teacher and students together develop a scoring rubric. Students have some say in
          deciding what the components of the rubric are.
                                             implementing assessment for learning               77
  3.   Students break into cooperative learning pataitesot into pairs and apply the rubric
       to each other's work.
  4.   Students reflect and report on what they lesrned from applying the scoring rubric. For
       example, they might describe how they will improve their work on the next occasion.
There are many advantages associated with peer assessment, in terms of improved
student awareness, monitoring, confidence, self-esteem, motivation and self-regulation.
An additional benefit is the immediacy of the feedback for all students. The comment
by one learner in further education surely applies also in both primary and secondary
schooling: ‘our teacher takes so long to mark our work, we've forgotten what it was
about by the time we get it back’ (Armitage et al., 2012: 162). Peer assessment enables
students to remain actively on task in relation to both assessment and learning while
reviewing a peer’s work, providing feedback, receiving feedback and discussing the
work. While no class could exclusively use peer assessment, its judicious use releases
the responsibility from the teacher to review every piece of work and instead allows
more time to observe the process of assessment, with nobody ‘waiting’ to hear about
their work from the teacher. In your own teaching, you won't be long in also discover-
ing the challenges. Students can be reluctant and uncomfortable in assessing each
other’s work, perhaps expecting that this is the teacher's role. This suggests the need
for training and ongoing support for students, in part through the provision or co-
development of rubrics. Enabling students to see the value of an assessment culture in
classis vital. When students understand that sharing their work with other students is
a normal part of their learning, they come to see that they learn from the peer com-
ments and have the opportunity to actively discuss their work in class, much more than
if the assessment and monitoring role is reserved solely for the teacher.
Self-assessment
Teachers sometimes find it easier to introduce students to peer assessment before
self-assessment. This helps students to gauge the quality of other students’ work prior
to applying similar procedures and rubrics to their own work. Once embedded in
78    Understanding and applying assessment in education
with effective teaching and learning identified by education authorities in one state in
the USA. Review this list and note the importance attached to self-assessment:
      recognises good work and identifies steps for improving own work
      monitors progress towards reaching learning targets
      periodically assesses own work or that of peers
      uses feedback (teacher- and peer-generated) to improve own work
$s
6@    reflects on work and makes adjustments as learning occurs.
      (Kentucky Department of Education, 2015)
Activity 5.5
     Traffic lights
     This technique has the advantage of (i) promoting students’ self-reflection on their learning
     and understanding, while at the same time (ii) efficiently communicating information to the
     teacher in relation to the progress of each individual student and the class as a whole. Each
     student has three discs: Green signifies | understand; Yellow | am not sure; and Red | don't
     understand. At various points (not always at the end of the lesson), the teacher invites stu-
     dents to indicate, by a show of cards, their level of understanding of the topic or concept.
     Variations of this can be used, for example one large ‘set’ of traffic lights in the classroom
     which is used by the teacher to elicit student responses or smiley faces (very happy, con-
     fused, sad) for use with younger learners. The teacher needs to make concrete use of the
     information gleaned from the traffic lights. Try to address issues where difficulties in learning
     are indicated by the self-assessment process. Be aware of student apprehension about
     disclosing a lack of understanding of topics to their peers. Some teachers we have worked
     with occasionally ask students to close their eyes during traffic-lighting assessments. What
     will you do if you observe many red lights amongst your class?
(Continued)
     Take a ticket
     At the end of a lesson, ask students to respond to short prompts about the lesson.
     Answers can be written on a ticket/card containing space for a response and left with
     the teacher at the end of class. Prompts might include:
         Today | learned...
         One thing | could teach someone after today is...
         A new idea in the lesson is...
         | had a difficulty with...
         | tried to overcome the difficulty by...
     Occasional use of this method provides teachers with valuable feedback about student
     learning, thereby informing subsequent teaching. A variation involves the use of sticky
     notes, where students write a note about what they found easy/difficult and post it to a
     board at the end of class.
Name: Date:
Role of feedback
Scores and grades have a function in education. There are many contexts where it
is appropriate that the performance of students on tests and assessments is commu-
nicated to learners and others in the form of scores, marks or grades. Students in
many education systems routinely receive information about how they are achieving
in the form of public exam    grades, standardised test scores, levels of achievement
and other formats. The purpose and processes of grading are explored in detail in
Chapters 3 and 8, especially in terms of public exams, reporting to parents and sys-
tem accountability. This chapter focuses on a different function of assessment in
education and therefore we explore a different form of communication to students.
You will recall from earlier in the chapter that feedback is one of the most significant
factors influencing student achievement (see Figure 5.1).
   Feedback to students is sometimes viewed as a more user-friendly form of grad-
ing or marking. Instead of giving grades to students, teachers provide more qualita-
tive and helpful commentary, orally, in writing, or otherwise. There are two
difficulties with this interpretation. The burden on the teacher, as provider of all
formative information, is great and the value of feedback to the teacher about stu-
dents’ learning is underestimated. In fact, feedback to teachers about students’
learning, coupled with the appropriate teacher response, is key to student learning.
Hattie and Timperley (2007: 102) define feedback as ‘information provided by an
agent (e.g. teacher, peer, book, parent, experience) regarding aspects of one’s per-
formance or understanding’. One such agent is the student who, through informal
and formal mechanisms, provides feedback in relation to learning to the teacher
who can respond appropriately in terms of adjusted instruction and experiences
(Hattie, 2009).
    It is important to recognise that not all feedback is equally effective. More effec-
tive forms tend to focus students’ attention on specific aspects of the work or learn-
ing, as in commentary on the way in which a student made a persuasive argument
or the approach taken in relation to texture in a work of art. General statements of
affirmation do not constitute feedback at all. Table 5.5 highlights some ineffective
             \
82       Understanding and applying assessment in education
feedback with suggestions for how it might be improved. What is noticeable is that
the helpful feedback tends to be somewhat longer, though not excessively so. This
additional length, however, provides much more clarity for the learner about how
learning can develop.
   The approaches below and on the website highlight some of the ways that you
can create a feedback culture in class.
You've done a good job        It is interesting how you linked the Fairtrade logo on chocolate bars to equality
                              for farm workers in developing countries
You need to put more effort   There are sentences that are written in the future tense, but you used the past
into learning the verbs       tense form of many verbs instead. Can you spot where this happens?
Improve your diagrams         Can you work out a way to make your diagram more clearly understood for the
                              reader (e.g. add labels)?
You could play the part of    Can you think of suitable clothes to take from the costume box? How can you
Cinderella's sister better    make the sister's voice sound different to Cinderella's?
Come on! Respond to the       The beat to the song was syncopated in a number of places. Can you spot
music better                  where and add some additional shoulder-drop or other movements?
     Activity 5.6
     Try this out — Feedback
     Use the companion website prompts when trying out the following strategies in class.
     Reflect on your experience with a colleague.
feedback - the opposite is true. In the course of reviewing students’ work in class during
the day, take brief notes of key insights requiring action. Before the end of the lesson
(or day), prioritise the top two and address them in the conclusion, either with individu-
als or with the class as a whole. This will help ensure that your teaching is continually
informed by recent assessment information from students.
Use rubrics
Rubrics are rules (oftentimes illustrated as short statements) that teachers use to rate
the quality of separate elements of students’ work, such as the success criteria. For
simplicity initially, it may be helpful to present rubrics as tables, where the different
expected elements of the work (success criteria) are listed vertically and where differ-
ent quality levels are established (and listed horizontally). Rubrics can also be estab-
lished in relation to one dimension that attempts to capture the overall quality of a piece
of work.
(Continued)
most such films is the isolation of the innovative           teacher amongst    his or her
peers and the suspicion of school authorities. This illustrates one challenge of
formative assessment, namely ensuring that the culture, intention and approaches
are the mainstream rather than the remit of a few committed, charismatic but
isolated individual teachers. Teachers in many countries see this in the tension
that exists between the requirements for summative testing for certification or
school accountability (possibly in the form of league tables) and the wish to
empower students’ own love for learning and their progression through forma-
tive approaches.
   The recent educational journey of two exam-dominated education systems —
Singapore and Hong Kong - illustrates the challenges and possible solutions to
resolving the tension between formative and summative assessment. Ratnam-Lim and
Tan (2015) note how the implementation of ‘holistic assessment’ in Singapore was
predicated on the need for more constructive feedback to students to support all
aspects of their development. However, teachers had difficulty in providing a high
frequency of such feedback and couching it in terms that primary-age students could
understand. Teachers still favoured the use of frequent ‘bite-sized’ summative tests
that contributed to a cumulative grade per student. So whereas teachers were enthu-
siastic about formative assessment, even within a high-stakes examinations culture,
the need for a significant mindset change by all stakeholders, including parents,
should not be underestimated.
    Similarly in Hong Kong, where a strong tradition of summative assessment
exists, the Ministry of Education has been attempting to promote more AfL by
teachers.   However,   the ‘target-oriented   curriculum’,   introduced   in the 1990s, fell
foul of teachers who felt that the record keeping required for providing feedback
to students was too time-consuming (Berry, 2011). Further refinements and some
change in nomenclature from 2002 onwards saw greater emphasis on identifying
problems that students experience with learning and providing appropriate feed-
back to enable them to improve their work. This has required a more explicit
articulation of formative assessment policy by government, along with stronger
efforts to foster a greater balance between formative and summative approaches in
senior secondary school, strategies designed to ensure the active involvement of
students in the assessment process.
   Experience in Singapore, Hong Kong and many other education systems suggests
that changing assessment practices requires a shift in the teaching and learning culture
of classrooms and schools. Any educational change needs to be sufficiently close to
teachers’ existing practice to encourage engagement, but far enough away so that
some effort is required to implement it (Murchan et al., 2009). This presents a dilemma
for some teachers who may view formative assessment as a heavy burden that is
added to what they already do.
   Addressing and resolving the tensions and challenges associated with introducing
and maintaining effective AfL in schools requires the involvement of all key stake-
holders in education. Teachers on their own cannot ‘implement AfL without con-
comitant support from the other players, including students, parents and policy-makers.
A number of challenges are presented in Table 5.6, with suggestions for resolution.
What is obvious is that whereas teachers can effect some change, the system and
broader school and system cultures must change also.
8&6       Understanding and applying assessment in education
Time-consuming to               Teachers may feel they need to assess everything students do all the time. Try out
implement properly              specific AfL approaches with some targeted students on a rolling basis. You can
                                also implement specific AfL approaches for selected lessons or on selected days.
                                Create time within class to dialogue with specific students
Practicality of ‘recording’     Given the importance of transmitting feedback from students’ work back to the
assessments                     teacher, place more emphasis on students showing how they have learned from
                                teacher feedback. Foster peer and self-assessment and empower students with
                                checklists, rubrics and portfolios
Insufficient resources          School systems, schools and publishers can help. A lot of strategies and material
                                are available online
Teachers’ lack of belief        This is sometimes manifested in teachers not wishing to share a responsibility
in AfL                          for assessment with students. We need to embrace a new vision of teaching
                                that emphasises students’ responsibility and self-regulation. We also need to
                                recognise that AfL is not a panacea for all challenges in education and society.
                                Work with colleagues in your school to identify AfL approaches that are
                                particularly helpful and prioritise these
Teachers’ lack of               Share practice with colleagues whom you feel are ‘about the same’ as you and
confidence or                   with others whom you feel ‘know AfL’. Try out some ideas, just as you are invited
competency                      to do in this book, and reflect on your experience with others. Look for video
                                examples of AfL online, from your local educational or curriculum agency and
                                other providers; look at AfL as a combination of teaching, learning, assessment
                                and classroom-management practices and culture — it’s not only about specific
                                techniques
Social pressures to             Talk with colleagues within your school and with parents. Discuss AfL with your
provide grades                  students and explain about finding an appropriate balance in assessment. Create
                                a non-threatening, inquiry-based culture in class where students feel safe in
                                monitoring their own learning, in responding to and sharing what they find from
                                this process
Washback from exam-             Most systems now promote AfL, even those systems with high-stakes
dominated system                examinations or accountability systems. There are always opportunities to
                                practice AfL with your own students
Tension between                 Move away from a position where one purpose of assessment is better than
formative and                   another. There are different purposes that often, but not always, involve different
summative purposes of           methods. Include a blend of purposes and approaches in your professional
assessment                      practice
  ‘obstacles’ facing teachers, the collaborative in-school dialogue with teachers and
  school leaders provided solutions to many of the roadblocks. Teachers working within
  communities of practice can collaboratively solve problems that seem unsolvable to
  the individual.
Chapter summary
The history of educational change is littered with examples of initiatives that,
though theoretically sound, were difficult to implement in practice, or were modi-
fied to such an extent that classroom practice bore scant resemblance to the
intended policy. Assessment for learning certainly has the potential to transform
learning, but, to do so, it may also have to transform teaching. Most education
systems now include AfL as a significant component of practice in primary and
secondary education. For some, there is considerable tension between AfL and
other expectations for assessment. This chapter introduces some of the opportuni-
ties and challenges associated with embedding AfL in classroom practice. Six
dimensions of AfL are explored and the reader is invited to try out a range of
approaches associated with each one. A general theme throughout is the need to
make assessment practical, manageable and worthwhile for students and teachers.
The sustainability of incorporating AfL into your professional practice can be
enhanced by transferring some of the responsibility for assessment from the
teacher to the students themselves, individually and collectively.
  1.    Identify one subject or learning area from the curriculum you intend to teach and
        focus on a specific grade or class or year level. Note one student competency you
        wish to promote in one or a series of lessons. Develop a specific learning outcome or
        instructional objective (in adult wording) related to the targeted learning. Re-write
        this outcome as a learning intention in student-friendly language appropriate to the
        target age or grade level. Ensure that the statement describes what students will
        learn from engaging with any task or content. Ask a colleague to evaluate the learn-
       -_ ing intention.
  2.    Write criteria to show students how they can demonstrate success in relation to
        the learning intention(s) developed in question 1 above. Try not to simply repeat the
        learning intention or give the answer. Focus on evidence of learning. Your success
        criteria may be in the form of statements or rubrics. What do you notice about writ-
        ing learning intentions and success criteria?
  3.    See if you can obtain access to the assessment policy for a primary or secondary
        school. Describe the relative emphasis placed on formative and summative assess-
        ment. Identify some specific formative approaches to assessment that are articu-
         lated in the policy and relate them to the six dimensions of AfL presented earlier in
        Table 5.3.
                 4
88       Understanding and applying assessment in education
Further reading
Gershon, M. (2014). Assessment for learning toolkit. A website hosted by the Times Educational
   Supplement. Available at: www.tes.co.uk/teaching-resource/assessment-for-learning-tool
     kit-6020165
This is a free resource offering many practical classroom-based ideas for implementing AfL
within your class.
Krathwohl, D.R. (2002). A revision of Bloom’s taxonomy:      an overview.   Theory into Practice, |
     41 (4), 212-18.
Bloom’s taxonomy, developed in the 1950s by a group of educators in the USA, remains one
of the most widely used frameworks for categorising cognitive learning aims and objectives.
The taxonomy has influenced curriculum development, at national and at individual teacher
levels, in many countries worldwide. A revision of the taxonomy in the late 1990s sharpened
the focus on knowledge and cognitive processes within the taxonomy and provided greater
illustrations of applicability to primary and secondary education.
Torrance, H. (2007). Assessment as learning? How the use of explicit learning objectives, assess-
   ment criteria and feedback in post-secondary education and training can come to dominate
   learning. Assessment in Education: Principles, Policy and Practice, 14 (3), 281-94.
Set in the context of post-secondary education and training in the UK, this article offers an
alternative vision of assessment as something that may undermine rather than support genuine
learning and inquiry by students. It is a thought-provoking article that challenges some of the
dominant narrative around formative assessment.
     What you will learn in this chapter
     The previous chapter illustrated a range of techniques enabling you to embed
     assessments seamlessly and relatively informally in classroom practice. Many of
     them offer students a structured opportunity to learn through engagement with the
     assessment approaches: the assessments themselves promote enhanced learning.
     This chapter introduces some techniques more traditionally associated with sum-
     mative assessment, where the purpose is largely, though not exclusively, to gather
     information about students’ learning at a particular time. Chapter 5 illustrated some
     alternative ways in which students can demonstrate their learning. The emphasis
     in this chapter focuses largely on assessments that primarily call on students to
     write. Many assessments developed by teachers and contained in commercially-
     produced teaching material and textbooks involve writing. For the teacher, the
     judicious use of such methods reflects a key professional competency. The chapter
     explains and illustrates the role of written measures of student learning and helps
     you to learn how to construct and use such measures. You will hopefully also
     become a better user of tests produced by commercial publishers and national or
     statutory agencies.    »
ee
ae
ee
         When you finish the chapter, you should be able to answer these questions:
   In learning outcome 4, students are required to know and understand the termi-
nology related to plant metabolism, such as nutrition, transport, photosynthesis and
phototropism. This represents reasonably low skill levels in the subject area and
might therefore be assessed using simple assessment approaches such as short-
answer or matching exercises that can be administered and graded efficiently.
Outcome 5 also requires that students know specific information (plants take in
carbon dioxide and emit oxygen), but they must also use this information in a per-
suasive way to argue a point of view (the benefit of planting trees). This represents
more complex learning and necessitates a different form of assessment such as an
essay, a written speech or an active debate. Finally, we need to be careful in inter-
preting student understanding of grammar, particularly given the re-emphasis on
functional building blocks of language within initiatives to promote literacy in many
countries   (see DfES, 2006; DES, 2011b). Whereas                   students may   recognise     certain
rules (forming plurals of nouns, use of commas and phrasing in syntax), they may
lack the capacity to apply these principles consistently in practice in their writing.
Therefore, conflating the grammar and content when assessing a subject-based
essay (for example, on the merits of tolls on motorways) may lead to confusion in
                            Summative    written assessments for classroom use           91
interpreting what a grade means. Teachers need to be very clear with students and
others when they apportion weighting to grammar and writing mechanicsin essays.
   There is usually no one absolute way to assess any outcome. For practical reasons,
we often cannot use what might seem to be the best option. To assess a student’s
oral competency in a foreign language, creating an environment where the student
would converse with a native speaker would be ideal. However, this is rarely practi-
cable. Alternative methods are sought and compromises need to be made. However,
attention needs to be paid to the validity issue of how well inferences about test
scores reflect students’ knowledge, skills and ability on the construct of interest. All
assessment methods have advantages and disadvantages and you must weigh these
up in selecting suitable approaches. Look at the specific type of learning you are
trying to assess and match the learning to the approach most likely to let you make
sound judgements about pupils.
   Students can engage with most areas of learning from the perspective of any one
or a combination of these elements. For example, a child in primary school can learn
some basic facts about homelessness but also form attitudinal dispositions in relation
to people who find themselves in that situation. These are two separate but related
aspects of learning the same content, suggesting that approaches to assessment are
also likely to vary.
   Three achievement domains have been traditionally identified: cognitive learning,
focusing on knowledge, thinking and mental reasoning processes; affective learning,
related to attitudes, dispositions, feelings, interests, motives, values and beliefs; and
psychomotor learning, associated with students’ motor skills, perceptual processes,
movements          and skills. Usually a mixture of two, or even three, domains of achieve-
ment is reflected within the activities that students undertake in school, even though
there may be one dominant kind of behaviour required. Bloom’s taxonomy (Bloom
et al., 1956), mentioned in Chapter 5, represents one way of categorising cognitive
learning into six somewhat different dimensions, while a refinement by Anderson
and Krathwohl (2001) was also used in that chapter to illustrate questioning as part
of AfL. In reality, any reasonable taxonomy or categorisation system can be used to
conceptualise and structure learning so that students can engage with content in a
way that transcends the mere learning of isolated facts.
   Large-scale assessments such as standardised tests and national monitoring sys-
tems use test blueprints or tables of specification to ensure that items on a test reflect
and sample a wide range of learning in the domain of interest. Though the more
elaborate blueprints are beyond the needs of teachers in class, they illustrate how
such systematic approaches to test development work. For some samples, see the
Key Stage 1 English grammar, punctuation and spelling test framework (Standards
and Testing Agency, 2016b) or the Framework for the National Assessment of Second
Class Mathematics (Shiel et al., 2014: 17, Table 2.4). Nitko and Brookhart (2014) pro-
vide a good overview for developing tables of specifications. In general, such
approaches allow teachers to identify both the learning content and the different
ways in which students engage with this content (using dimensions such as those of
                                  Summative written assessments for classroom use                93
Table 6.3. Table of specifications for mathematics assessment: Operations and money
Stiggins, Bloom and Anderson). Table 6.3 presents a test blueprint for a mathematics
assessment developed for use with children in the early years of primary school. This
provides clarity for the teacher about the different topic areas included in the test and
also about the specific cognitive dimensions of students’ mathematics learning cov-
ered by the assessment. Eighteen items are presented to students, covering two topic
areas and three cognitive dimensions in relation to those topics, thus resulting in a
test with a known balance of items.
    Moving beyond the specification of content and process skill matrices in single
subjects, school systems increasingly promote complex cross-subject competencies
or skills. These initiatives stem from a number of large-scale efforts, such as the
OECD’s Definition and Selection of Competencies — DeSeCo initiative (OECD, 2001)
and the Partnership for 21st Century Skills (2015). These promise a more holistic
education for students across cognitive, social, emotional and physical dimensions.
Such changes in curriculum require a change in assessment, a point well illustrated
by the American psychologist Lorrie Shepard (2000), who highlighted a disjoint
between current theories of learning, curriculum and assessment. She made the com-
pelling argument that whereas our understanding about how students learn has led
to changes in what and how we teach them, approaches to assessment have been
much slower to change.
    A glance at mission statements from education ministries worldwide highlights the
widespread aim that schools would provide challenging subject matter that empha-
sises deep and meaningful learning and that students should acquire competencies
relevant to real-life contexts. These include the capacity to think creatively, flexibly
and critically; to identify, interpret and solve problems; to relate and communicate
effectively; and to use technology seamlessly in learning. Large-scale testing pro-
grammes struggle to adequately capture these complex skills. Accordingly, a greater
onus is placed on teachers to assess students at local level so that there is a great
opportunity to widen the types of learning that are promoted and rewarded in class
(Collins et al., 2010; Baird and Black, 2013).
                 \
94          Understanding and applying assessment in education
     Activity 6.1
     Try this out — Developing assessments related to key skills
     Table 6.4 presents a list of selected key competencies promoted in four national education
     systems across all subjects. For a particular lesson or topic you are going to teach at a
     particular age level, develop one or two learning outcomes related to any competency and
     outline how you would assess students in relation to it.
Table 6.4 Extracts from key competencies specified in selected education systems
     Sources: 1. Department for Education (2014a); 2. The Scottish Government (2008); 3. NCCA (2014); 4. Partnership for
     21st Century Skills (2015)
                                                                            learning
e   There should be an appropriate balance of easy-to-assess aspects of the
    and elements that are more difficult to assess.
e   Assessment items should be appropriate for the learning outcomes.
e   Directions to students should be clear.
             ¢
96         Understanding and applying assessment in education
In the remainder of the chapter, we present ideas for developing and using a number
of assessment formats suited to written tests in class, namely: fill in the blanks or short
answers; essays and other extended constructed response tests; and objective items
such as true—false, multiple-choice and matching.
Format                               Illustration
Question                             What is the common name for Beethoven's Sixth Symphony?
Completion                           The common name for Beethoven's Sixth Symphony is the
Identification/Association           On the lines opposite the numerals, write the popular names that identify
                                     these symphonies by Beethoven:
   With the question format, a direct question is posed and the student provides
a short answer. A benefit for students is the clarity provided by the question itself,
presented as a complete sentence. Provision of a clear, intact sentence in ques-
tion form reduces the potential reading difficulty for younger students or students
with a more limited reading ability, an important validity consideration when
interpreting student responses. Where the learning intention being assessed does
                                 Summative        written assessments for classroom use                          97
not relate directly to linguistic skill (e.g. in the case of the music knowledge
outcome in Figure 6.1), students’ probability of answering correctly should not
depend on their reading level. An incomplete sentence with a blank at the end
can increase the reading burden on students unnecessarily. Items that require the
identification of information or ideas offer an efficient way to assess a number of
concepts based on one prompt. A stimulus (such as a list of terms, a diagram, a
photograph) is presented to the student, who is required to respond appropri-
ately by writing labels, numbers, symbols or other terms. This format is frequently
used to assess basic knowledge in science and geography.
    You may have also taken some cloze tests in your own education, as in the
example in Figure 6.2. These formats are frequently employed for use in assessing
 language skills and reading comprehension (Wall, 2004). By filling in the missing
words or terms, students’ understanding of the content is assessed. Although they
 are widely used, some concerns atise about the comparability of results when dif-
 ferent rules for deleting words are used. In developing such a test, the teacher
 might delete every fifth word, every tenth, all nouns, all verbs, all key content
 words or function words, for example. However, student scores can vary depend-
ing on the deletion          method      chosen     (Alderson,      Clapham       and Wall, 1995; Henk,
1981). Some teachers use cloze tests in relation to subjects other than language,
sometimes to check for knowledge of specific vocabulary or terminology.
   Last Monday |         feeling ill. |         a headache and a sore stomach. | stayed in bed and my
   mother         me some drinks. | was very tired. My mother phoned the doctor who came at two
   o’clock. The doctor examined me and put a thermometer                  my mouth. | had a high temperature.
   She said | had a fever and gave my mother a prescription                medicine. Later that day, my mother
   got the medicine         the pharmacy      and  |      taking it. After three more days, | felt better but
   did not go back to school            the next week. Although | got a week off school, | was disappointed
   because my school won a football match and | should                 been on the team.
                                                                                       is the
    One positive feature of short-answer items such as cloze and completion
                                                                                rather   than
 requirement on students to produce a correct or acceptable answer
                                                                          of students  using
 merely recognise one from a list of possibilities. This offsets the risk
                                                                                      choice
 partial knowledge triggered by seeing plausible alternatives in, say, a multiple-
                                                                       teacher
 question. However, whereas the correct answer may seem obvious to the
                                                                              answers
 who crafts the item, scoring can be complicated by unexpected but plausible
                                                                        could  have a
 and even by mis-spellings. See, even in Figure 6.2, how some blanks
 number of acceptable answers.
                                                                      n of a word,
    Whereas short-answer questions typically involve the insertio
                                                                    write a sentence
 phrase or symbol, some varieties offer students limited scope to
                                                                    ing. Review the
 or two. However, the objectivity of scoring may be more challeng
                                                                   hy and consider
 items in the box overleaf, relating to secondary school geograp
                                                                item as written and
 the quality of each item in relation to any ambiguity in the
 the challenges in scoring.
                {
93       Understanding and applying assessment in education
         Develop the item to assess important knowledge, concepts or skills that students
         are expected to have learned.
         Word the question yourself rather than copying it directly from textbooks or other
         material. Students can recognise and remember such previously seen material.
         Be precise in phrasing the prompt so that only one answer is correct and students
         are clear about the desired nature of the response.
         Avoid providing unintended clues about the answer in the question through gram-
         matical structure, words used or length of blanks.
         Use direct questions rather than incomplete formats, where possible.
         Don't place too many blanks in a question. One is preferable, located at the end.
         Keep all blanks the same length when a series of questions is used.
         Anticipate mis-spellings by students and consider how important accurate spelling
         is in relation to the learning outcome being assessed.
     Activity 6.2
     Try this out — Scoring cloze tests
     Select a couple of paragraphs from an article in a magazine or newspaper. Delete every
     fifth word and have a colleague ‘replace’ each word exactly as it was written in the article.
                              Summative     written assessments for classroom use                99
  Repeat this exercise where only every noun is deleted. Now score both tests by giving
  one point for every word correctly ‘replaced’. exactly as it was written in the article.
  Convert the two scores to percentages. What do you notice about the results? What
  issues arose for you in scoring the tests?
A group of children in another school were asked to calculate the area of the shape below:
12m
8m
   Take a look at the answers some of the children gave. Check if each answer is correct. Give a reason
   why you think this for each one.
Is he correct?
films, games and TV programmes. The following outlines a restricted response item
designed to focus students’ responses on specific dimensions of one such digital text:
   Briefly describe a recent film that you have watched under the following headings:
      Title
      Genre and atmosphere
      Cinematography, including any special effects
      Use of soundtrack
                              Summative written assessments for classroom use                     101
Examine the defeat of Churchill’s government in the 1945 general election in Britain.
Developing essays
We will look now at the development and scoring of essays. Some key tips for devel-
opment are presented in the box.
  e   Develop the essay prompt to capture specific knowledge and skills that are iden-
      tifiable in the learning outcomes and that are not more easily assessed by other
      methods.
  e   Unless dealing with ambiguity is an aspect of skill to be assessed, phrase the essay
      prompt so that the meaning and intention are clear to students.
  e   Students should be given some guidance as to the scope of the task through an
      indication of time to spend on the essay.
  e   Avoid or minimise the use of choice in essay titles.
  e   Use an appropriate analytic or holistic scoring guide.
  e   Where there are a number of essays to be completed on a test, grade all the answers
      to one question for all students before moving on to the next essay.
  e   Conduct a check on your scoring consistency by re- Scone some essays yourself or
      enlisting the help of a colleague to do so.
  e   Score on the basis of writing mechanics only if this is warranted by the learning out-
      come and if students are aware that this is included in the rubric. In content-based
      essays, it is preferable to score analytically so that a separate score can be provided
      for writing mechanics, if wished.
Teachers sometimes assign essay questions that are under powered in relation to the
learning intention. For example, the extended essay question: Describe and explain
the functions associated with parts of a flowering plant, really only asks students to
remember specific facts and write them out in narrative form. It is littke more than
recall and translation into the students’ own words, indicating nothing more than basic
comprehension of a set of facts related to plant life. Depending on the guidelines,
students may spend pages and a whole-class period writing out these facts, which
might be much more easily assessed using a series of short-answer or objective items
(objective items will be discussed in the next section of the chapter). Many assess-
ments offer students some choice about what items to attempt. Typically, this is either
to provide scope to students to choose, or it reflects a curriculum so broad that there
is tacit understanding that students cannot ‘cover’ all the content and skills. However,
offering a choice of items may not necessarily advantage students in the way envis-
aged and may lower the validity of inferences made about the student results
(Murchan, 1993; Crookes, 2007; McMillan, 2011). Whereas this is less of an issue in
formative classroom assessment where the teacher is quite familiar with students’
work and has multiple opportunities to assess, it is more significant in summative
assessment where students’ performance is compared to their peers. If a wide choice
is offered to students, they will choose different combinations of questions, with
the result that no two students might take the same exam. This is reasonable if all the
items are of equal difficulty, which they may not be. Furthermore, grading students
                            Summative     written assessments      for classroom   use      103
where they have all essentially taken individual exams chosen by themselves makes
a comparison of overall results problematic — an issue if results are used to formally
certify students’ achievement or to select students.
  Activity 6.3
  Try this out — Evaluating essay tests
   Write a prompt for an essay designed to assess some aspect of students’ language or
   other subject-based learning. Either review the essay prompt yourself or get a colleague
   to review it, drawing on the checklist below. If necessary, re-develop the prompt based on
   the suggestions below. Prompt sheets on the companion website will guide you in this task.
   1.   Write out the specific learning outcome(s) that the essay is designed to assess.
   2.   To what extent is your essay the best way to assess the outcome?
   3.   Identify the dominant cognitive skills expected of students to complete the essay
        (knowledge, summarising, reasoning, creativity).
   4.   Are students aware of the relative importance of language competency in the essay
        (spelling, punctuation, syntax, flow)?
   5.   Is the task clear to the students? (Is the task appropriate for all students to whom
        it will be given? Is the prompt written in accessible language, pitched at students’
        reading level?)
   6.   Are the expected length of the essay and probable/available time limit clear to
        students?
   7.   If the essay is part of a wider assessment, are students aware of the weighting
        (marks) to be allocated to the essay?
This process of drafting and re-drafting the same essay mirrors the real-world writing
of reports in the workplace, where people are seldom expected to produce a final
document or report without having the opportunity to amend it. Chapters 4 and 5
explored comment-only grading, an AfL practice well suited to essays. If, prior to work-
ing on an essay, students are clear about the success criteria, teachers’ comments and
annotations on the essay can help students understand the extent to which their
response adequately meets these criteria, along with some suggestions for develop-
ment in their learning. Conferencing with students offers a further opportunity to dis-
cuss student work with the students themselves. Meeting individually with students in
busy classrooms can be challenging, but alternative group sessions can also be benefi-
cial, particularly where the comments are reasonably uniform across a group. Think of
student conferences as brief but frequent opportunities to affirm the quality of the work
of individual students, to help the students identify a small number of improvements
that can be made and to support them in achieving those goals. Such targeted sessions
offer a great opportunity to involve students in shared reflection on their own work
and in collaboratively planning the next steps in their learning.
    When used summatively, grades are frequently assigned to essays. Two main
approaches are typically used when grading essays. Analytic approaches focus sepa-
rately on a number of features within the essay, ‘rewarding’ performance on each
feature separately. In contrast, teachers using holistic methods focus more on the
essay as a whole, identifying an overall impression of the response. Key features of
the two approaches are summarised in Table 6.5.
Analytic Holistic
Fine-grained analysis of detail in essay             Offers an overview or impressionistic evaluation of the essay
Marks awarded for individual elements                May use a number of level descriptors
Frequently used for content-based essays (history,   Generic rubrics often applicable to a range of different essay
environmental studies)                               topics
Scoring rubric may be unique to the specific
prompt/content
   Two essay prompts and associated scoring rubrics are presented in Tables
6.6 and 6.7. The first example, relating to business studies, assesses students’
capacity to relate knowledge and concepts about consumer rights to a practical
skills-based scenario. An analytic scoring rubric is employed that provides detailed
feedback in relation to a number of key features expected in the response. Scores
for individual elements can be summed to get an overall score for the essay.
   The second example, aimed at students recently arrived in secondary school,
focuses on students’ narrative and reflective skills in English composition, combining
powers of observation within a reflective communication. A more holistic scoring
rubric is used that provides levels (grades) based on more aggregated criteria.
   The advantages of analytic rubrics include greater clarity for teachers and students
about what is rewarded in the scoring and an increased likelihood that the student
can draw on the mark to address specific lacunae in the response. Disadvantages
                                    Summative          written assessments for classroom use                    105
You purchased a second-hand smartphone from a local retailer. Within a few days you notice that some of
the features are not working properly. Write a detailed letter to the retailer explaining the problem in terms of
consumer rights and requesting that the matter be resolved.
1 2 3 4
Welcome to secondary school! Now that you have been in your new school for a few months, write an account
of your experience so far, describing how it is different to your primary school.
include the time involved for teachers in scoring, resulting in a possible delay in
returning scores to students. Holistic rubrics can often be applied to a range of essays
or/prompts and can speed up scoring. However, they generally offer little formative
feedback to students about how they can improve their performance. Another
approach, known as Adaptive Comparative Judgement (Pollitt, 2012), represents a
variation of holistic judgement, where a teacher or rater compares two pieces of work
(two essays) and decides simply which of them is better. This process is repeated to
include all essays in a set and enables a ranking of all the essays on a scale. While
encouraging levels of reliability have been yielded by this method, it is difficult to
provide formative feedback to students.
   Regardless of which scoring methods you use, it is good practice to review the
accuracy of scoring, with a colleague if possible. This can involve re-checking all or
a sample of essays that have been assigned the same score or category to see if they
all share similar characteristics in terms of quality. This is generally easier to do in
holistic scoring where broad score bands are used, as in Table 6.7, rather than where
percentages or a large number of narrower grade bands are used. The more grade
bands you use, the more likelihood there is that essays will be misclassified (receive
a B instead of a B+; a Cl instead of a C2) — a potentially significant reliability issue
(see Chapter 2). The recent reduction in grade bands in the Leaving Certificate in
Ireland from 14 to 8 reflects, in part, concerns about the ‘reliability of scoring near
the cut scores between grades, especially when many grades are used. Overall, it is
important to remember that some form of rubric is essential when scoring essays,
perhaps combined with sample answers at different levels of quality.
Objective items
The final type of assessment format we will look at in this chapter is objective items.
Objectivity in scoring is assumed in-the name of these items — it is expected that there
is one agreed answer to the question and that, therefore, different teachers or raters
can come easily to the same conclusion about the quality of a student’s response.
                           Summative     written assessments for classroom use              107
   Whereas constructed response items such as short answers, essays, orals and even
practical exercises have a long history, objective items have increased in popularity
over the past century or so, particularly since the development of IQ tests in the early
20th century (see Chapter 3). Studies have queried the consistency of the scoring of
essays, leading to an interest in forms of tests that might prove more objective, in that
they can provide higher levels of scoring precision than the older forms of testing,
judged to be ‘subjective’. Objective forms are often used with large cohorts of stu-
dents as they can usually be scored more efficiently and reliably. There is evidence
that objective forms of assessment can measure complex learning as well as more
subjective methods (Lukhele, Thissen and Wainer, 1994). A number of reviews by
Lukhele and colleagues concluded that essays are very expensive to score and that
the information yielded by such tests is often not very different to that which can be
obtained from far less-costly but carefully-crafted objective items.
True-—false items
This format may be useful if you wish to quickly assess some relatively simple knowl-
edge such as recognition of previously learned material, the recall of historical facts,
definitions in science or basic concepts in mathematics. True-false items are usually
presented to students in written form, on an exam paper, in a textbook or digitally but,
as indicated in Chapter 5, they can also be used orally as part of AfL to gauge students’
knowledge about a topic in advance of instruction. True—false tests can enable teachers
to assess a broad range of knowledge without exploring the depth of that knowledge.
Students can move swiftly through the test, thereby allowing the teacher to include
more items and obtain a broader snapshot of students’ learning. The writing burden
on the students is minimal and scoring by teachers or students themselves is quick. An
obvious disadvantage is the potential for student guessing, thus complicating teachers’
ability to draw reasonable inferences from the test results. We will return to this matter
in more depth later in the chapter. Activity 6.4 outlines some tips to consider when
developing true-false items, along with some scenarios to evaluate.
Activity 6.4
   1.   This item format is generally designed to assess lower levels of learning so keep the
        questions short and simple.
   2.   Ensure that the correct or best answer is unambiguous.
   ca   Avoid the tendency to use negative phrasing in questions to make them more difficult.
                                                                                   response
   4.   Ensure that there is a mix of true and false statements, without a correct
        pattern that is obvious to students.
   5.   Avoid the introduction of irrelevant clues in the questions through careless wording.
                                                                                  (Continued)
108     Understanding and applying assessment in education
(Continued)
   Now review the sample true-false items below. Evaluate them in relation to the tips high-
   lighted above, finding one major flaw in each item. Can you think of some ways to
   improve the items? Re-write them. Then develop three true-false items that you could
  use with a class to assess some aspect of learning. Prompts on the website will assist you.
  Candidates are elected to the European Parliament that meets in Strasbourg             True   False
  The Trade Winds do not blow from the north east                                        True   False
  Owls always hunt at night                                                              True   False
Multiple-choice items
Many of the strengths of true-false tests are retained when teachers use multiple-choice
questions (MCQ), while some key disadvantages are minimised. MCQs are widely used
in standardised tests such as the Drumcondra Primary Reading Test, SIGMA Mathematics
Test, Wide Range Achievement Test and Cognitive Abilities Test. They have been an
integral element of large-scale testing programmes in many countries for decades, and
are widely used, along with other item types, in international assessments such as PISA,
TIMSS and PIRLS. The basic architecture of a MCQ consists of three components, as
illustrated in Figure 6.4.
   MCQs encompass a wide variety of related formats that facilitate the assessment
of a range of learning from lower-level recall to quite complex cognitive processing.
They can of course be used to test the recognition and recall of facts and concepts,
but can also assess higher-level cognitive skills, such as comprehension, making
                                Summative       written assessments         for classroom      use       109
connections and inferences, the application of concepts and principles, analysis and
evaluation (Haladyna, 2004; Downing, 2006; Haladyna and Rodriguez, 2013). Many
of you may be familiar with the use of MCQs‘in testing on the rules of the road. For
example, the Driver and Vehicle Standards’ Agency in Northern Ireland requires
learner drivers to take two theory tests containing a total of 100 MCQs (visit www.
nidirect.gov.uk/theory-test if you would like to try out some sample items testing
your knowledge of the rules of the road). These formats can also be found at col-
lege level and in accreditation processes for a number of professions, for example
in medicine and accountancy (see certification by the Association of Chartered
Certified Accountants at www.accaglobal.com/uk).
     Students in primary and secondary education frequently encounter MCQs. The
format is flexible and enables the assessment of a wide range of cognitive outcomes.
MCQs are efficient in facilitating the assessment of a broad range of content as writing
by students is minimised, a factor that also helps minimise an adverse impact on stu-
 dents with writing difficulties. They can be scored quickly and accurately, a process
 that can be easily automated, especially in the case of computer-based assessment
 where students and teachers can receive results immediately after the test. When
 combined with the affordances of computer-based assessment, MCQs offer the poten-
 tial to provide complex questions based on multi-media stimuli that go far beyond the
 traditional perception of MCQs as merely assessing recognition and knowledge
 (Scalise and Gifford, 2006; Sireci and Zenisky, 2006; Kreiter et al., 2011). Multiple-
 choice items can be constructed to provide powerful diagnostic information about
 students’ learning misconceptions based on the careful analysis of specific distractors
 chosen by students (Eggen and Lampe, 2011; Murchan, Oldham and O'Sullivan, 2013).
     However, quality items are more difficult to develop than people generally think.
 Many of the criticisms of MCQs are on the basis of poorly constructed items where
 trivial material is frequently assessed, that ambiguity exists in the question or the
 options, or the best answer is not clear. The answer to an MCQ is presented to the
  student amongst a set of possible alternatives, meaning that students do not neces-
  sarily have to conceptualise an answer — it is there within the question. There are,
  of course, many aspects of learning, such as collaborative working, the organisation
  of knowledge and concepts, original thinking and effective communication that can-
  not be assessed with MCQs. In keeping with our view in this book that teachers need
  to use a variety of assessment methods to reflect the variety of learning, MCQs,
  therefore, like essays; portfolios and performance tasks, contribute to, but are not the
  totality of, teachers’ assessment repertoires.
      Criticisms of MCQs by teachers and others sometimes focus on the risk of obtaining
  high scores through guessing. This is a concern, though less so than with true—false
                                                                                               item, with an
  items where there is a 50% chance of guessing correctly on any given
                                     on  a  test   containin   g  only     true-false    items.  With multi-
  expected chance score of 50%
                                    of  obtaining      high    or even      moderate  ly     high   scores by
  ple-choice tests, the chances
                                           due    to  the   number      of   options    in  each   item,  typi-
  random guessing is much reduced
                                        chances      of   guessing     all  or  even    a  large  number     of
  cally four or five. Therefore, the
                                          If a  student     were   to  guess    at  random     on  a  ten-item
  items correctly on a test, is remote.
                                                                                     is 25% or two and a
  MCQ test with four options per item, the expected chance score
                                                      guessing     all  ten   items  correctly    on that test
  half questions. The probability of a student
                                                 test,  the   chance     of   guessing     40%   of  the items
  is one in a million. On a 20-item MCQ
                                                       To  obtain    80%     or better,   through     guessing
  correctly by random guessing is 10 in 100.
110        Understanding and applying assessment in education
the probability drops to one in a million. However, in reality, students are more likely
to use whatever information they have to work out the answer to each item, and, if
they are motivated at all, blind guessing is the exception rather than the norm. As
outlined in Chapter 2, procedures exist for estimating the reliability of objective tests.
High reliability of test scores is evidence that guessing by students is minimal.
   While Figure 6.4 highlighted the basic structure of an MCQ, variations exist for
different purposes. Some of these include Correct answer, Best answer, Multiple
response, Combined response, Negative variety and Incomplete statement. Use
of best answer options requires students to choose an answer where a number
of reasonable possibilities exist but where one is preferable. This enables the
item to tap into students’ capacity to know content but also to be able to judi-
ciously weigh alternatives to provide an optimum solution. Figure 6.5 illustrates
some common types of MCQs. Opinion differs on the ideal number of options
on MCQs. Many tests use four options, including the keyed answer. However,
Circle the shape that is divided in half.                       What was the basic intention behind the establishment
                                                                of the European Coal and Steel Community?
A. incorporation, mortgage
B. maturity, proposal
C. demarcation, premium
D. policy, deficit
AandB O BandC O
a                                                              AandD   O                      CandD     O
                                                                     a
Negative variety                                                Incomplete variety
Study the regional map provided. Which of the                   An object dropped from your hand falls to the ground
following transport options is NOT available to                 due to the effect of:
residents of the region?
                                                                A. Acceleration
A. Air                                                          B. Force
B. Rail                                                         C. Friction
C. Road                                                         D. Gravity
D. Water
CY         Serene ne ee                RE       IEC ee                 Mee PAPA Te we                 N,N              Y et          uy"
Figure 6.5      Selected MCQ variants
Note: * SEC, Junior Certificate Ordinary Level Geography, June 2008
                              Summative      written assessments for classroom use                      111
  Activity 6.5
  Try this out — Evaluating and improving multiple-choice items
   Here are some tips for developing and scoring multiple-choice items:
  1.    Match each item with an intended learning outcome and try to design items that
        measure higher-order learning outcomes as well as knowledge offacts.
   2. Present a single, clearly formulated problem in the stem.
        Put as much of the wording as possible in the stem: ideally, a capable student should
        be able to answer the question without seeing the options.
   4. Eliminate unnecessary wordiness in the question overall: over-elaborate questions
        containing unrelated content can confuse weaker readers.
   5. Avoid negatively worded stems, sometimes included to make the item more diffi-
        cult, for example ‘Which of the following is not ...?”.
   6. Lay out the item in a clear, logical, systematic manner: list options vertically where
     possible, in a logical order (alphabetical, chronological or by size).
   7. Vary the positioning of the correct answer in a set of items to avoid students spot-
        ting a pattern (perhaps designed to make correction easier for the teacher).
   8. Use homogeneous and plausible options. Sometimes teachers have difficulty in
        coming up with a last option and include a choice that most students immediately
        know to be unrealistic.
    9. Avoid the introduction of irrelevant clues in the questions through careless wording:
        make all options grammatically consistent with the stem and ensure consistent
        punctuation.
   Now review the sample multiple-choice items below. Evaluate them in relation to the
   tips highlighted above, finding at least one major flaw in each item. Can you think of
   some ways to improve the items? Re-write them. Then develop two other multiple-
   choice items that you could use with a class you intend to teach to assess some aspect
   of learning.
   The lungs in mammals are        Lara is 7 years old. She has lovely red hair. She has 4 big red apples.
   examples of:                    Jan has black hair. Jan has 7 green apples: How many more apples
   A. Breathing                    has Jan than Lara?
B. Organs Be 2
   C. Reptiles                     aoe
   D. Species
   Peper              se   ie      Se
                                                                       iel
   Visit the companion website at: https://study.sagepub.com/murchansh
                  4
112    Understanding and applying assessment in education
           (b) The diagram shows the human eye. Examine the diagram and answer                      (15)
               the questions that follow.
Allows light in
al Focuses light
                 (i) In the table write the letter A beside the name of the part labelled A.
                 (ii) In the table write the letter B beside the name of the part labelled B.
                (iii) In the table write the letter C beside the function of the part labelled A.
                (iv) In the table write the letter D beside the function of the pupil.
Chapter summary
The item formats explored in this chapter have been widely used in schools and assess-
ment systems for decades. As such, test developers and teachers have built up consider-
able knowledge and expertise about how they are best developed, used and interpreted.
Some formats, such as multiple choice and essays, are frequently criticised as being
traditional and incapable of assessing the type of skills required by students in the
21st century. However, this criticism misses the point that teachers need to use a diversity
of assessment approaches and formats. The chapter highlighted the varied dimensions
of learning that modern curricula and schools are expected to foster in students, indicat-
ing the need to match approaches to assessment with educational intentions. No one
approach to assessment can universally capture information on all aspects of students’
education and development. It would be as inappropriate for a teacher of primary or
secondary students to use only essays as it would be for that teacher to use only in-class
presentations. What is needed is a diverse mix of approaches that, together, can help
teachers, students and other stakeholders to acquire greater understanding of students’
level of achievement and, therefore, of possible next steps in their development.
    The chapter explores a number of traditional item types, including essays, short-
answer, true-false, multiple-choice and association items. Possibilities for use are iden-
tified and advantages and disadvantages are discussed. Readers are invited to review
guidelines for developing quality items and to apply these guidelines in evaluating
                                                                                  students. A
some items and in developing their own items for use with their own
                                                              for teachers  can   be   inferred
number of general implications and recommendations
                                                                    or assessmen t   approach
from the discussion in this chapter. There is no one perfect test
suited to all learners and occasions.                     Rather, teachers need to select and/or develop
                                                                    with the intended
assessments that align as reasonably and as practically as possible
114      Understanding and applying assessment in education
learning outcomes. In preparing assessments, you could usefully share your items with
colleagues, craft items using clear, simple language and feel free to use a mixture of
formats, even within one assessment. When using tests that require students to write,
such as essays, consider to what extent students’ language ability does and should
relate to the intended learning. When using test formats that require students to select
a pre-determined response, such as multiple choice, consider the possible effect of
guessing on scores and your interpretation of student performance on the test.
   1.   Make a list of the test formats that you experienced as a student in primary or second-
        ary school (all that you can remember). Group them under a number of format head-
        ings (written, oral, practical, traditional, alternative). What conclusions can you draw?
   2.   Authentic assessment is a term widely used to identify approaches to assessing stu-
        dents that emphasise depth, challenge and real-world contexts in relation to student
        learning. Frey et al. (2012: 14) identify key elements of authentic assessment. These
        include: a cognitively complex task, a task that interests the student, and one that
        relates to skills or abilities that have relevance beyond the assessment itself. To what
        extent are any of the item formats discussed in this chapter ‘authentic’? In particu-
        lar, can essays support the authentic assessment of 21st-century skills? What are the
        counter-points that critics might make?
   3.   Your teaching colleague, Mr Walsh, uses some true-false and multiple-choice tests
        to assess students. At a recent parent-teacher meeting, a parent told him that these
        types of tests are too simplistic and the students can probably do fine byjust guessing.
        What advice would you give your colleague in responding to the parent's comments?
Further reading
Haladyna, T.M., Downing, S.M., & Rodriguez, M.C. (2002). A review of multiple-choice item-writing
   guidelines for classroom assessment. Applied Measurement in Education, 15 (3), 309-33.
This article explores in detail the arguments for and against many of the guidelines for writing
MCQs that are frequently advanced in the literature. The paper also provides useful illustration
and evaluative commentary on a range of MCQ formats frequently found in teacher-made tests
and in textbooks.
OECD (2005). The definition and selection of key competencies: Executive summary. Available
  at: www.deseco.admin.ch/bfs/deseco/en/index/02.parsys.43469.downloadList.2296.
  DownloadFile.tmp/2005.dskcexecutivesummary.en.pdf
This article outlines the result of an OECD project lead by Switzerland to investigate the key
competencies required by individuals and by society to live successfully and ensure a well-
functioning society in the future. DeSeCo identified three broad competencies: using tools
effectively; interacting in heterogeneous groups; and acting autonomously. Though not
addressing assessment per se, the competencies identified by the OECD have influenced edu-
cational policy in many countries and have implications for the ways in which student learning
should be assessed. Proponents of traditional written tests need to evaluate how such
approaches can be used or need to be modified (perhaps using computer-based assessment
with multi-media stimuli) if these approaches are to remain relevant in the future.
What you will learn in this chapter
In Chapter 6, we saw how the choice of assessment formats can be linked to expected
learning outcomes. This chapter builds on that idea, illustrating many occasions when
traditional written tests cannot be directly aligned with learning. Many current cur-
ricula focus on developing transferable skills, such as using information interactively,
cooperating with others to solve problems and managing one’s own learning. Such
learning requires the adoption of more varied authentic or performance assessments.
This chapter focuses on planning, constructing, marking and evaluating performance
assessments. Topics include (a) the move towards ‘21st-century skills’; (b) the process
of conducting performance-based assessments; (c) approaches to scoring such assess-
ments, including the use of checklists, rating scales and rubrics; and (d) the format of
performance-based assessments (oral exams and presentations, projects and portfolios
and coursework).
   When you finish the chapter, you should be able to answer these questions:
   Performance assessments can vary in length from activities that take just a few
minutes to complete, such as a solving a basic mathematics problem, to projects that
may take several weeks and require students ‘to present their findings to audiences
inside and outside the school.
   Various authors have identified aspects of knowledge and dispositions that can
best be assessed using performance-based assessments, and some of these frame-
works overlap:
e       Habits of mind — according to Costa and Kallick (2008), these are problem-
        solving, life-related skills that are needed to operate effectively in society and
        include persisting, thinking flexibly, managing impulsivity, thinking about one’s
        thinking (metacognition), applying past knowledge to new situations, taking
        responsible risks, thinking independently and remaining open to continuous
        learning.
e       Collaborative problem solving — students are assessed as they work together to
        complete a project or another performance task (e.g. Von Davier and Halpin,
        2013). In judging the outcomes of cooperative learning, there may be learning
        outcomes relating to the overall success of the project (for example, the develop-
        ment of a model or report), as well as outcomes specifying the expected
        contributions of individuals.
e        Twenty-first century skills — these are skills that are deemed important for the
        world of work in the 21st century. Griffin and Care (2015) describe these as
         including: ways of thinking (creativity and innovation, critical problem solving,
         problem solving, megacognition), ways of working (communication, collabora-
         tion/teamwork); tools for working (information literacy, ICT literacy); and living
         in the world (citizenship, life and career, personal and social responsibility).
         Griffith and Care’s work highlights the increasingly important role of ICTs in
         teaching, learning and assessment.
    e    Higher-order thinking skills — these comprise the more advanced skills on
         Bloom’s revised taxonomy (Anderson and Krathwohl, 2001) and include: apply-
         ing (using information in new situations); analysing (drawing connections among
                                                                                        new
         ideas); evaluating Gustifying a stand or decision); and creating (producing
         or original work). For illustrations of the applicatio n of assessmen t tasks using
         Anderson and Krathwohl’s taxonomy, see Table 5.4 in Chapter 5.
                                                                                         across
    Other sources that focus on the development and assessment of key skills
                                                                                 Governm  ent,
    subject areas include the Curriculum for Excellence in Scotland (Scottish
                                                                            and    capabilit ies
    2008), the Junior Cycle framework in Ireland (DES, 2015a) and skills
    (including cross-curricular skills) in Northern Ireland (CCEA, 2016a).
                                                                                  ance-based
       It is notable that many countries seeking to implement perform
                                                                       seek   to  ensure that
    assessment, with a view to assessing 21st-century skills, also
                                                                         y.   In   Ireland, for
    students have acquired strong basic skills in literacy and numerac
                                                                         y   across    the cur-
    example, there is currently an emphasis on literacy and numerac
                                                                        with    others,   being
    riculum, even as key skills such as self-management, working
                                                                      emphasi   sed.    Hence,
    creative and managing information and thinking are being
                               at school   level may need  to include both  basic skills and
    assessment programmes
    complex 21st-century skills.
                  \
118         Understanding and applying assessment in education
a a Performance standards/
Understand content knowledge in                Given some clay, a drinking straw and         Darling-Hammond and
physics (e.g. flotation)                       paper, design a sailboat that will sail       Adamson (2010): from
                                               across a small body of water; students        Illinois State Assessment
                                               can test and retest their designs
                                               (Grade 8)
Demonstrate presentation skills in             Listen to a speech extract on                 Assessment and
a formal setting; listen and respond           television and represent its main             Qualifications Alliance
appropriately to spoken language; use          points and biases; alternatively, listen      (AQA, 2016): GCSE
spoken and standard English effectively        to a school assembly and re-present           controlled assessment
in speeches and presentations                  its main points and explain its
                                               methods of presentation
Recognise questions that are appropriate       Over a three-week period, formulate           National Council
for scientific investigation; design, plan     a scientific hypothesis, plan and             for Curriculum and
and conduct investigations; organise and       conduct an experimental investigation         Assessment (Ireland)
communicate research and investigative         to test the hypothesis, generate and          (NCCA, 2016): Junior
findings in a variety of ways, fit for         analyse primary data, and reflect on          Certificate science
purpose and audience, using relevant           the process, with support or guidance         extended experimental
scientific terminology and representations     from the teacher                              investigation
                                               Performance-based     assessment      119
implementing a performance-based
assessment
A performance-based assessment task can be developed and scored by an individual
teacher, a subject department, an external assessor or an examining board. A perfor-
mance task seeks to assess learning targets or objectives that are specified in curricu-
lum documents.
   Such tasks may be carried out by individuals or groups. They can be scored as
students work on the task and/or after it has been completed. Generally, scoring is
done with one of the tools described in the next section. Often, curriculum objectives
are expressed as standards or learning outcomes and these become the focus of a
rating scale or a rubric.
   A moderation process may be put in place, where a check on the quality of the
grades assigned by the teacher is undertaken. This could involve a different rater
taking a random sample of completed tasks and scoring them independently.
Discrepancies between two or more raters can then be addressed in a marking or
moderation conference. Sometimes, when moderation unearths a discrepancy, the
assessor may need to review the standards (earning outcomes) to achieve a better
understanding of them. This is not dissimilar to the moderation that occurs when
exam papers are being corrected and a random sample is re-scored to ensure
consistency (see Chapter 3).
                     (for example,   Klenowsky    and Wyatt-Smith,   2014; NCCA,    2016)
   Several sources
                                                                                meetings’,
advocate moderation meetings or ‘subject learning and assessment review
                                                                               share their
where teaching staff in a school can share examples of student work and
                                                                                 a greater
judgements with other colleagues. The purpose of this activity is to develop
                                                                       judgement about
understanding of learning standards and to ensure a consistency of
                                                                           of technology
student performance. Klenowsky and Wyatt-Smith point to the value
                                                                    student tasks CCT-
for sharing and reviewing the scores or levels assigned to
                                                                        that, in the case
moderated or e-moderated assessment). However, they also argue
                                                                       l pieces of work
of assessment for learning, the validity of judgements about individua
                                    with  diagnosti c use of the outcomes   taking prece-
takes precedence over reliability,
dence over the consistency in   scoring.
120     Understanding and applying assessment in education
   Another form of moderation is to provide raters with samples of student work that
exemplify each point on a rating scale or scoring rubric. This enables raters to achieve
a better understanding of the required standards and to mark more reliably as a result.
   A simple approach to assessing agreement across raters is to calculate the percent-
age agreement. Thus, if two raters assign the same overall scores to 80% of tasks
(papers, projects), we would say that there is 80% agreement. An agreement rating
below this level might signal a need to review the scoring guide and re-score the tasks.
   The final stage in assessing performance on a task is to assign a grade or mark.
This may take the form of a numerical score (for example, 1-5), a descriptor (for
example, approaching mastery) or a grade. More extensive feedback may be pro-
vided to the student who completed the task — such as comments, an indication of
areas in need of further improvement, or targets that the student should strive to
reach in the future.
                                     Performance standards
                                     and criteria linked to
         Teacher develops and        learning outcomes
         assigns performance         (via rating scale, rubric)
         task (prepared by
         teacher or examining
         board)
             Student or group of
             students completes
             and submits task
                                      Moderation (within or
                                      between schools);
                                      work samples to define
                                      meaning of standards
Anecdotal records
Anecdotal records are notes based on teachers’ observations about students as they
perform an assessment task. They allow teachers to document student strengths and
                                                Performance-based     assessment       121
weaknesses as they edit a text, solve a problem or search for information. Data
gleaned from anecdotal notes can be reviewed with other information (such as a
finished product) to arrive at an overall judgement of a student’s performance.
Observational checklists
A checklist consists of a list of behaviours, characteristics or activities and a place for
marking whether each is present or absent. For example, a checklist designed to
assess a student’s ability to proofread a text might include:
Rating scales
Rating scales are often used for aspects of a complex performance that do not
lend themselves to a yes—no or present-absent judgement. A rating scale assesses
the degree to which a student has attained the learning outcomes linked to a
performance task (such as an oral presentation). A rating scale can be used as a
teaching tool (to familiarise students with what is required to achieve a standard)
as well as an assessment tool. The end points of a rating scale are usually
anchored (‘always’, ‘never’), with intermediate points defining levels of perfor-
mance (‘seldom’, ‘occasionally’, ‘frequently’). In general, the more points on the
rating scale, the more reliable are the scores. An example of a rating scale is given
in the section on project work below.
Scoring rubrics
Scoring rubrics, as highlighted in Chapter 6, are a type of rating scale on which each
level has a complete description of performance and quality. Rubrics may be analytic,
where   each of several dimensions    is assessed, or holistic, where either a judgement
about overall quality or an overall judgement on performance is made. Rubrics may also
be general (i.e. the same rubric can be applied to different tasks) or task-specific (where
the rubric describes quality with respect to a particular task). An analytic rubric has the
potential to generate specific feedback on strengths and weaknesses on each dimension
of a task. For example, in the case of a written text, the rater might make judgements
122     Understanding and applying assessment in education
Projects
Projects are long-term structured activities completed by individuals or groups that
result in a product such as a model, a substantial report or a collection of artefacts.
According to Nitko and Brookhart (2014), projects can call on a broad range of skills
and knowledge,    including creativity, communication    skills, problem solving, critical
thinking and subject matter knowledge. While individual projects focus on the work
of one student, group projects focus on the activities of several students. In addition
to evaluating the final product, assessment may look at whether students work
together collaboratively. Indeed, collaboration is specified as a learning target in
many curricula. As noted earlier, it is considered to be an essential 21st-century skill.
   A scoring rubric for a group project may have several dimensions, each of which
is rated on a three- or five-point scale. The dimensions, which relate to group and
individual contributions can include:
    creating a newsletter
    researching an historic figure on the internet
    implementing and reporting on an attitudinal survey
    researching a national issue (e.g. taking steps to manage climate change) or a local
    one (e.g. granting planning permission for a new shopping centre) and giving both
    sides of the argument
e   designing a campaign to persuade teenagers to take more exercise.
   The rating scale for scoring the Heroes or Villains project (see Figure 7.3) could
include three components, corresponding to the three clusters of learning outcomes
described in the project task — Working as a historian, Collaborative learning and
                                                                                      to
Communication. The purpose of the rating scale in Table 7.2 is to assign scores
a group as a whole. Although a 5-point scale is  given, only  three points (5, 3, 1) are
                                                                       that a stu-
defined. However, scores of 2 or 4 can be awarded if the rater decides
dent’s work falls between a 1 and a 3, or between a 3 and a 5 respectively.
124        Understanding and applying assessment in education
    Some educators may question the inclusion of collaborative learning and commu-
nication on a rating scale designed to assess an understanding of history. However,
if the history curriculum includes these elements as goals or learning objectives, or
expresses them as learning targets, they should be included in an assessment.
   People in Ireland who took part in the 1916 Rising have been hailed as heroes by some and as villains
   by others. In cooperative groups, source at least two articles giving these contrasting viewpoints,
   summarise the information provided and prepare a PowerPoint presentation that seeks to explain these
   contrasting viewpoints. You have two weeks to complete the project.
Working as a historian:
Collaborative learning:
Communication:
Table 7.2 Rating scale for group history project on heroes or villains
Oral presentations
Many performance-based assessments involve an oral presentation. Such presenta-
tions allow students to demonstrate their knowledge and skills through interviews,
speeches and presentations, including those that are multi-modal. An oral presenta-
tion may be a stand-alone activity or may be part of a larger project which culminates
in the communication of key findings by participants.
   While sharing checklists and scoring rubrics with students can heighten their
awareness of the criteria associated with effective oral presentations, teachers can
also model effective presentations and draw attention to important strategies, such as
opening with one or more key questions, making eye contact with listeners and
using signalling terms such as ‘in summary’ and ‘in conclusion’.
   The following are some topics that could be addressed in student presentations:
    Explain why iPads and other tablets are excellent tools for learning.
    Does frequent use of social media lead to lower levels of concentration and focus?
    Should recycling be compulsory and why?
    Do advertisements on social media influence what we eat and drink?
126          Understanding and applying assessment in education
Portfolios
A portfolio is a collection of material designed to showcase a student’s best work or
to show the student’s growth and development over time (for example, over a term
or school year). Entries to the portfolio may be linked to learning targets and may
include self-reflections on the student’s own work. Portfolio entries may be annotated
by the student, allowing the teacher to track student thinking and explanations as
well as progress over time. Portfolios may also be used as a basis for diagnosing a
student’s learning difficulties in a subject area.
   Examples of tasks for which portfolios might be appropriate are given below. Each
is accompanied by specific or implied learning targets:
e     Visual Arts, where the student is facilitated in putting together a personal assess-
      ment portfolio to demonstrate their growth and learning. This may be linked to
      learning targets such as:
      o      gathering     samples   (assembling   completed   pieces   before   making   a final
             selection)
      o      selecting    and supporting (providing a rationale for selecting specific pieces)
      o      reflecting    (thinking about what has been learned and looking ahead)
      o      applying     and extending (indicating how new learning can be applied in the
             future).
                                                     aduUapIAa JO ejep
                   SUOISN|DUOD             uly} Alaa sapnjaul ‘Delqns            suo!sn|aUOD syioddns
          Jo seap! 10} Woddns                 aul Woddns Ajayenbape             yey} BduUAaPIAa JO e}ep                               SDUAPIAS YIM
 JUAaIDINSU! SAAIH “yDa[qns jo             JOU OP YDIYM ‘sd1}S172}S JO    awos sapnydu! ‘pelgns aul                   seapl/SuoIsn|dUOD sLioddns
yoddns ou 10 yeam sapiAoid                /pue se} ‘sajduexe yeam            yoddns jeu} soi}s17e}s JO                     ‘soiysiqeys JO/pue ‘S}De4
    ‘asodind pue palqns ay}           sapiaoid ‘asodind pue pelqns        /pue s\de} ‘sajduuexe awos                 ‘sajdwexa juaulped ‘jDefqns
       auljap AjJea|d JOU s80q                dU} BUIJAP O} S}dWAa}VY       sey /1defqns pue asodund                  pue asodind sea sapiAosd
                     pafqns                                 suolsanb           Jeaj> JEYMaWOS e SeH         °                       uoljejogeja pue
au} jnoge suolsanb Jamsue                  Ajejuauipns AjuO Jamsue                uolesogeja INOYWM                  suoljeue|dxe yyIM suosanb
                                          0} ajge Ss! pue UOIEWJOJU!        ‘suolisanb |je 0} SiaMsue                       ssejp jje Huuamsue Aq              uoljesiueblo
jOUUED PU UOMEWJOJU! 34}
   jo dsei6 e aney JOU se0q               BU} YIM ajqeYoOjwooun s|           pajpedxe yiIM asea Je S|               aHpajmouy ||N} Se}zesjSUOWAaG                                 fjua}
                                                                                                                         sjuiod Aay asiseydwa pue
                  apebuasip                                                                uOo!}Da}}UI                  jsaJa}Ul edualpne ule}UIeWW
  0] aduaipne sasned YIUM                                                    PUP ALUNJOA JO UOI}EHWeA                      0} UONDa}JUI PUL BLUNI|OA
       QUO} SNOUOJOUOW JO                       UOIDAa}JUI OU JO 9/}}1|      Asopejsizes ym syeeds           e           ul uoHeNyDNI|} YIM syeads
                                      ULIM BWNIOA UaAeUN Ul Syeads                                        O}
                                                                                                          S3}OU                     Sa}OU Je Hulyxoo|
 /pue auUNjOA Mo| Ul syeads
                                          sa10u Wod} Ajjsow BHuipess       SUINJAL {|S ING ‘eoualpne                     Wop]as ‘jDe}UOD eAa PaJIp
                                                                                                                                                                               Manijeaq-
   sa]0U WO} peal S| Ode
 auljua au} se ‘aduaipne 3y}                aIUM ‘adua!|pNe ay) YUM              SU} Y}IM 3De}UOD aAe                      $O 9SN au} YUM adua!pNe
  LIM 39e}U0D ae OU SP|OH              pejuod afe jeululw ske\dsiq           aIp fO asn jUa}sISuOD              e     alljUa dU} fO UO!Ua}e SP|OH
os    \ ee             a                          ee             eS             eS                                  Eee
           JUSWaAOJdUI! Speen                                                                        poo                                      yuUa]|a2x4        jyuauodwo,)
e
hoe ee Nn SS e
             SS 8 eS
                                                                                                       uolejUasaid je10 Ue Hulssasse JOJ WUGny                    €"Z a1qeL
128      Understanding and applying assessment in education
   The English literature portfolio includes printed texts and multi-media artefacts
such as video or audio recordings.
   In setting up a portfolio assessment, it is important to outline what learning out-
comes are to be assessed, what proportion of marks is available for each section and
when the portfolio should be submitted. The teacher will also need to decide what
level of support or scaffolding should be provided to the student.
   Portfolios can be assessed using checklists, rating scales or scoring rubrics. They
also allow for self-assessment, as students review their own work and comment         on
it, drawing on a rubric developed by the teacher. Where possible, a rubric can be
co-developed by teacher and students.
    An electronic portfolio or e-portfolio is a collection of student work, usually
saved on the Web. Evidence of learning may include texts, electronic files, images
and multi-media. Such a portfolio allows students to share their work with teachers,
parents, administrators and other students. In addition to contributing to the devel-
opment of target skills linked to the learning targets for a course, the e-portfolio
builds a student’s digital skills. As well as providing students with a chance to reflect
on their learning processes and progress, an e-portfolio facilitates the participation
of the student’s teacher, peers and parents in the learning process as feedback and
comment can be provided. Perhaps the key advantages of an e-portfolio are that it
enables students to track their learning over a lengthy period of time, and to reflect
on their learning at any time.
    When introducing e-portfolios, issues such as whether the student’s work can be
made public, and whether students can view and comment on one another’s work
may need to be addressed.
While experiments and investigations are appropriate for many content areas, the fol-
lowing are examples that are science-related:
                                                                 Performance-based           assessment               129
e       How does the air pressure of a soccer ball affect how far it travels when kicked?
        How does the temperature of water affect the time it takes to freeze into ice cubes?
e       In a blind taste test, can you tell the difference between non-fat, low-fat and
        whole milk?                              x
e       Which can support more weight: paper or plastic grocery bags?
e       How does the tension in a violin’s strings affect its pitch?
e       How does the time of day affect your body’s temperature?
e       What type of ground layer limits erosion the most: sand, gravel or soil?
e       Which has a longer life: an LED or an incandescent light bulb?
2       Does a person’s weight vary throughout the day?
    1. Make an electrical current using a battery, wire, a light bulb and switch
    2. Make a simple drawing of your circuit, labelling all parts
                                                                                                                nail,
    3. Use the circuit to test which of the following conduct electricity: plastic spoon, steel washer, string,
       5-pence/cent coin, rubber band
    4. Describe what the objects that conduct electricity have in common
    5. What have the objects that don’t conduct electricity in common?
    6. What broad conclusions can you draw?
Figure 7.4 Science investigation: Making and testing a simple electric circuit
               ie       a    a ee        i                8               SS       SS                  Se
    a
    Making and testing a simple electrical circuit
                                                                                         evidence; demonstrates a
    4 points: Draws a complete and accurate diagram of a circuit with supporting
            understand  ing  of the  concepts of electricity and  conductivity ; may use descriptive terms (conductor,
    strong
    insulator, flow, current, etc.); draws appropriate    conclusions
                                                                                             ing of the concepts of
    3 points: Includes a complete diagram of a circuit; demonstrates good understand
                and  conductivit y; identifies most of the objects that conduct electricity; conclusions generally
    electricity
    correct, if not always stated in scientific terms
                                                                            missing or incomplete; shows some
    2 points: Draws diagram of a circuit where some components are
                    of the concept of electricity but not conductivi ty; conclusio ns partially correct
    understanding
                                                                                 does not draw appropriate conclusions
    1 point: Does not include a diagram of a circuit; most answers incorrect;sarees EERE PaRECELE L TY
                                                       Te      ee aS
    | wn      Ee ABS          Ge            Sa
                    \
130     Understanding and applying assessment in education
Simulations
While teachers have used technology as part of the assessment process for a long
time (for example, by asking students to respond orally or in writing to a segment
from a film on DVD, and to compare it with the novel on which the film is based),
in recent years there has been an extensive use of simulations in science, mathemat-
ics and other subjects that are designed to assess students’ ability to manipulate a
number of variables simultaneously, and to draw conclusions about what they have
found. For example, in the OECD Programme for International Study Assessment
(PISA) in 2015 (see Chapter 3), 15-year-olds in most participating countries encoun-
tered a combination of traditional science items (multiple-choice and short answer)
and ‘interactive’ or simulation-based items, all delivered on computer (OECD, 2016a).
The simulation-based items described scenarios such as running in hot weather.
Here, the student can manipulate air temperature, air humidity and whether the
runner drinks water or not, to draw conclusions about the conditions in which heat-
stroke or dehydration are likely to occur. In addition to manipulating variables,
students are asked to draw on the relevant science knowledge (for example, sweating
cools the body, helping to regulate body temperature). Students are also asked to
answer questions where the answer cannot be found in the simulation, but can be
inferred (they were asked whether it was safe to run without water at 40°C and 50%
air humidity, when the slider for air humidity could only be set at either 40% and 60%).
Another simulation asks students about the effects of varying outdoor temperature
and roof colour on energy consumption (defined as maintaining room temperature at
a constant 23 degrees celsius).
    Simulations such as these have the potential to tap into aspects of knowledge and
understanding that are less assessable with more traditional stimuli and questions.
They may be categorised as performance-based assessments since students interact
with stimuli that represent real-world situations, and are required to communicate
their reasoning. Equally, simulations can be useful as teaching and learning tools as
they allow students to access real-life situations that would not otherwise be acces-
sible, and to engage in reasoning as they consider how variables are related to one
another and change over time.
DO
Chapter summary
                                                                                    it can be
The focus of this chapter was on performance-based assessment and how
                                                         cannot  be  assessed    using   other
deployed to assess a range of 21st-century skills that
methods such as written tests and exanis.
                                                                                 ent for AfL
   A distinction was drawn between the use of performance-based assessm
                                                         perform ance-ba   sed    assessment
and for AOL, and the strengths and weaknesses of
                                                               sed   assessme  nts,   particu-
were outlined. It was noted that the use of performance-ba
                                                        tivist  learning   theory.
larly in classroom contexts is consistent with construc
                                                                           outlined include
    Broad approaches to performance-based assessment that were
                                                              investig ations,   and simula-
projects, oral presentations, portfolios, experiments and
                                                               these   measure s,   including
tions. Tools that could be used to assess performance on
rating scales and rubrics, were described.
                                                                      students’ metacogni-
    The value of performance-based assessments in developing
                                                                  internal ise performance
tive knowledge was stressed. It was noted that students could
132      Understanding and applying assessment in education
standards by evaluating their own work, using an appropriate rubric in which out-
comes and criteria are clearly specified (for example, self-evaluation).
   Throughout the chapter, the key concepts of validity and reliability were stressed,
and it was noted that, in the context of using performance-based assessment for
AfL, reliability may be less important than validity, as teachers use the outcomes of
performance tasks to identify students’ strengths and weaknesses, and modify
instruction accordingly.
   1.   What are the possibilities and applications for the use of performance assessment
        with very young children and primary school students? To what extent can young
        students engage with the self-assessment aspects of performance assessment?
   2.   The essay format is sometimes characterised as a performance assessment, as
        we do in this chapter, yet it is one of the oldest and most traditional approaches
        to assessment. Do you believe that essays should be considered performance
        assessments? Does the not-uncommon practice of rote learning for examinations
        compromise the claim of the essay to be a performance assessment?
   3.   Two challenges in implementing performance assessments are time constraints and
        workload management for teachers and students. Think of a subject in which you
        could introduce a performance assessment such as a portfolio or project. What
        steps could you take to make it more manageable?
Further reading
Brookhart, S.M. (2011). The use of teacher judgement for summative      assessment in the United
   States: Weighed in the balance and (often) found wanting. Paper       presented at the invited
   research seminar on teachers’ judgments within systems of             summative assessment:
   Strategies for enhancing consistency. Oxford University Centre for   Educational Assessment.
  Available   at:   http://oucea.education.ox.ac.uk/wordpress/wp-content/uploads/2011/07/
   Brookhart-FINAL.
                 pdf
Harlen, W. (2005). Trusting teachers’ judgment: research evidence of the reliability and valid-
   ity of teachers’ assessment used for summative purposes. Research Papers in Education,
   20 (3), 297-313.
These articles review research on the reliability of teachers’ judgements about student perfor-
mance. What overall view do they convey about the use of teacher judgements to evaluate
student learning and what research evidence is provided to support that view?
What you will learn in this chapter
In Chapter 3, we looked at several tools that are typically associated with assessment
of learning (AoL). Two of these tools — standardised tests and public examinations —
feature again in this chapter. You will be familiar with the outcomes of these assess-
ments from your own schooling. Indeed, it is unlikely that you have reached this stage
in the education system without achieving test scores and/or examination grades that
were in the above-average range or higher. In the context of standardised tests, we
consider ways in which outcomes are interpreted, referring specifically to norm-
referenced and criterion-referenced interpretative frameworks, how test scores can be
interpreted (and misinterpreted), and how standardised tests can inform both assess-
ment of learning and assessment for learning. In the case of examinations, we con-
sider strategies for combining results from external tests and coursework (projects,
e-portfolios and work samples), potential sources of error in students’ examination
grades and the potential effects of examinations on teaching and learning. Throughout
the chapter, recent changes to tests and examinations in the UK and Ireland are con-
sidered. By the end of the chapter, you should have a good understanding of the
strengths and weaknesses of a range of approaches to scoring and grading tests.
   When you finish the chapter, you should be able to answer these questions:
                                                                              ance
e   What are the benchmarks against which I can interpret a student’s perform
    on an assessment?
©   How is students’ achievement on formal tests reported?
    How confident can I be that the results I report are accurate?
                                                                 nt information
This and subsequent chapters in the book focus on using assessme
                                                                         for a variety of
drawn from a variety of assessments, for a range of stakeholders and
                                                                       ng five chapters.
purposes. Figure 8.1 illustrates the relationships between the remaini
            ¢
134     Understanding and applying assessment in education
e     Normalised standard score — this score tells where a student’s raw score is
      located in relation to a normal distribution fitted to the norm group. For many
      tests, the mean score is set at 100 and the standard deviation at 15. As shown in
      Figure 8.2, this means that 68% of students in the norm group performed
      between one standard deviation below the mean and one standard deviation
      above the mean (85-115), while 95% performed between two standard devia-
      tions above the mean and two standard deviations below the mean (70-130).
      Scores below 70 and scores above 130 are relatively rare (they are achieved by
      fewer than 5% of students).!
e     Percentile rank — this is the percentage of students in a norm group scoring lower
      than a particular raw score. If a student achieves a percentile rank of 70, it means
      that he or she performed as well as or better than 70% of students in the norm
      group. Percentile ranks should not be confused with percentage correct scores.
      Percentile scores cannot be averaged to derive a mean percentile rank.
e     Sten score — this is a score on an equal-interval scale ranging from 1 (low) to 10
      (high) (see Figure 8.2). Sten scores are transformations of standard scores, with
      a mean of 5.5 and a standard deviation of 2. Parents and others may find these
      somewhat easier to interpret than standard scores or percentile ranks.
e     Age-equivalent score — an age-equivalent score gives an estimated age level in
      an attribute (such as reading) based on the average score of learners of that
      age. Thus, a child who achieves a reading age of 7 years and 6 months can be
      said to perform at the same level as the average child of that age in the norm
      group. A child aged 8 years and 6 months, who has a reading age of 7 years
* Raw scores can also be transformed into a non-normalised scale. This typically involves
transforming raw scores into standard scores (with, for example, a mean of 100 and a stand-
ard deviation of 15). However, a simple raw score transformation may lead to a spread of
scores that does not match that found in a normalised distribution, such as the one shown
in Figure 8.2.
                                              Standardised tests and examinations            137
       and 6 months, can be said to be about a year behind in the aspect of reading
       assessed by the test. Particular care needs to be exercised in interpreting read-
       ing ages at the extremes of a distribution, as these are often extrapolated or
       estimated, based on what the reading ages would be if a sufficient number of
       high or low achievers had been assessed. Another problem is that growth slows
       as children become older, meaning that reading ages are less useful for report-
       ing on the performance and development of older children.
e      Test-wide scale score — these scores may be provided when a test assesses stu-
       dents at a range of grade levels (Year 1, Year 2, Year 3, and so on) in the same
       subject area. Typically, one grade level is identified as an anchor level, and scores
       are distributed across grade levels, with common items included in tests admin-
       istered at adjacent grade levels. A test-wide scale can be set to have any mean
       and standard deviation (for example, 250/50 or 500/100). Test-wide scores can
       be used to check progress from year to year and to provide an alternative to
       using a child’s percentile rank for this purpose. Thus, if a child progresses by
        70 score points, and the average progress for his/her grade level is 50 points, we
        can say that the child made greater-than-average progress. Rates of progress may
        vary by grade level, with, for example, greater average progress in the lower
        grade levels and less progress later on.
                                                                                                                desea
                                                                                                                     14%
      Standard            Oe               210)       eo.       =10)       Ooi MEANT         Oron         kOe        l-Dieeee-O}         e2.O   mrolO
      Deviation Units
                        Slten       Ss    ie,         a          he        See.    Fel eee      ar          a                   iprete    ae    ll
Percentiles 2 16 50 84 98
Figure 8.2. Normal distribution, normalised standard scores, percentile ranks and sten scores
standard score point (and vice versa). A similar issue arises with examination grades
(see below), where a student may miss out on a higher grade by just one or two
points. Conversely, a student may avoid a lower grade by the same narrow margin.
Conversion tables
If a standardised test is administered in paper-based format, the accompanying test
manual will include a conversion table that enables test users to transform raw scores
(such as the number of questions answered correctly) into derived scores (such as nor-
malised standard scores or percentile ranks). Table 8.1 provides a table from a mathe-
matics test that converts raw scores into standard scores, percentile ranks and sten scores
when the test is administered in the spring (a different table of norms is available for
autumn administrations of the same test). An interesting feature of the table (on the left-
hand side) is that raw scores of 70 or higher convert into percentile scores of 98 or 99.
This suggests that relatively few students cluster at the top of the distribution and that
the test discriminates reasonably well among higher-achieving students (that is, very
high scores on'the test are relatively rare). Nowadays, conversions from raw scores into
derived scores are often done by computer software. This allows for the generation of
comprehensive reports in a short time frame, but the same underlying tables as those
used with paper-based tests underpin the software algorithms used.
   In Table 8.1, a raw score of 71 converts into a standard score of 132, a percentile
rank of 98 (indicating that just 2% of students achieved higher scores) and a sten score
of 10. Very few students achieve scores that are this high. A raw score of 35 converts
into a standard score of 97, a percentile rank of 42 (indicating that 58% did better)
and a sten score of 5. Qualitative descriptions of what these scores mean can be found
in Table 8.2. These descriptors may be useful in communicating what scores mean to
parents and indeed to students themselves.
                                                      Standardised tests and examinations                         139
Table 8.1 Partial conversion table for a mathematics test (spring administration)
75              139          99                10     .            37               98            45                 5
74              137          99                10                  36               98            45                 5
73              136            99               10                 35               97             42                5
72              134            99               10                 34               96             39                5
71 132 98 10 38 95 Sy, 5
70              (}331          98               10                 32               95             37                5
69              129            97                 g                Bil              94             55                5
68               128       “Sy                    9                30               93             32                5
67 126 96 9 29 92 30 4
66 125 95 9 28 92 30 4
Table 8.2 Qualitative descriptors corresponding to standard score and sten score ranges
                                                                              score of
      Based on this, a student who achieves a standard score of 132 or a sten
                                                                                             their
 10 can be considered to have scores in the ‘well above average’ range, and
 scores are in the top one-sixth of the normalised   distributi on.  A  student    who   achieves
                                                                                      ed to have
 4 standard score of 97 (corresponding to a sten score of 5) can be consider
                                                         third  of  the  distributi on.  The use
 4 score that is in the average range or the middle
                                                  can    perhaps     reduce     the  competiti ve
 of descriptors Ginstead of numerical scores)
                                                       standardi  sed    testing    programm es.
 atmosphere that is sometimes associated with
                                                                               some computer-
 Some standardised tests may have more than one form, including
                                                               tricians       (those   who scale
 based tests. Approaches to scaling tests allow psychome
 tests) to place multiple forms of the same test   on  a  common       scale.
scores (with a mean of 100 and, we assume, a standard deviation of 15), and a
statement on whether the student has attained the required national standard. In
Ireland, performance on English reading and mathematics in grades 2, 4 and 6 is
reported to parents and (in aggregated form) to national educational authorities
with reference to sten scores, though schools may choose to report standard scores
to parents instead, if they wish. In Northern Ireland, schools may report perfor-
mance on computer-based tests of literacy to parents using standard scores
(mean = 100 and standard deviation = 15) and/or age-related outcome ranges
(where raw scores are converted into age-related score intervals, such as between
9 years 3 months and 9 years 6 months). In both England and Ireland, parents may
also be provided with scores based on teacher judgements of performance as well
as scores based on standardised tests. Teacher assessments may focus on aspects
of learning that cannot be captured by current tests, such as oral expression,
aspects of writing and some of the key skills highlightéd in our discussion of
21st-century skills in Chapter 7.
Criterion-referenced                    framework
As noted earlier, some standardised tests allow for both norm-referenced and crite-
rion-referenced interpretations of performance. Such tests report on performance
with reference to:
e     standard scores, percentile ranks and other derived scores that allow for a com-
      parison between a student’s performance and that of other students in the norm
      group
e     proficiency levels, performance bands or other indicators of the content or pro-
      cesses that a student is likely to have mastered.
Figure 8.3 shows the skills that students performing at Band 3 on a standardised math-
ematics test are likely to demonstrate. These skills are based on the most difficult items
that students in Grade 4 performing at Band 3 are likely to answer correctly.> In plan-
ning instruction for these students, teachers may wish to reinforce the skills associated
with Band 3 and provide instruction on the skills associated with the next highest
level, Band 4.
    Readers may come across the term ‘criterion-referenced assessment’ outside
the context of standardised tests. Many teacher assessments, where teachers
assess their students against specific standards, learning outcomes or levels of
attainment, based on the available assessment information, can be described as
criterion-referenced (for example, teacher assessments under National Curriculum
Assessment in England or teacher assessments based on the progression continua
in the new language curriculum for junior primary classes in Ireland (NCCA,
2015).   Criterion-referenced      assessment      is also   common       in vocational     training
programmes where students may be asked to demonstrate their competency on
* The use of Item Response Theory (IRT) scaling allows test developers to place students and
test items on the same underlying scale. This facilitates the specification of the tasks that students
at each proficiency band are expected to accomplish.
                                                 Standardised tests and examinations                   141
   Number/Algebra — pupils at Band 3 can complete multiplication number sentences involving the
   associative property, express tenths in decimal form and identify place value in decimal numbers.
   They can multiply two-digit numbers by one-digit numbers, and divide two-digit numbers by one-digit
   numbers with remainders. They can order decimals on the number line. They can solve word problems
   involving place value in three-digit numbers, multi-digit subtraction and fractions as sets.
   Shape and space — they can identify the use-of 2-D shapes in the environment, describe properties
   of 3-D shapes, construct 3-D shapes from 2-D shapes and recognise an angle as a rotation.
   Measures — they can compare lengths (m, cm), rename units of length and compare areas of
   irregular shapes. They can solve word problems on subtraction of dates on a calendar, addition of
   money and addition of weights (kg and g).
   They can also do the tasks described in Bands 1 and 2 (not shown).
Figure 8.3      Skills associated with Band 3 ona     standardised test of mathematics
Source: ERC (2010)
some combination of the two. On the one hand, marking schemes may provide a
detailed description of the competencies that must be observed if a particular grade
is to be awarded (a criterion-referenced approach). On the other hand, examining
bodies often adhere to a distribution of grades that was similar to the distribution
obtained in previous years (a norm-referenced approach). When the GCSE was intro-
duced in 1988, it was intended to represent a criterion-referenced approach to the
marking and awarding of grades, meaning that the proportion achieving each grade
would be based on how well or how poorly students performed in a particular year.
However, over time, there has been increasing attention paid to the distribution of
grades, with the distribution for a particular year often very similar to the previous
year’s distribution. For example, the Joint Council for Qualifications website (http://
www.jcq.org.uk/examination-results/) shows that the proportion achieving grade A*
or grade A on English GCE A Levels in the UK ranged from 20.7% in 2013 to 19.1%
in 2016, while the proportion achieving these grades in mathematics ranged from
43% in 2013 to 41.8% in 2016. Efforts by examination boards to ensure that the pro-
portions of grades within a subject are similar from year to year are likely to arise
because of concerns in the media and among the general public about grade infla-
tion. Small differences are usually attributed to natural variation in performance.
    As noted in the description of ongoing changes to the GCSE exam, the proportions
of students achieving scores at each grade point Cevel) will be carefully controlled
from the outset, and specific steps will be taken to ensure that grade distributions
remain stable from year to year (that is, grade inflation will be kept to a minimum,
unless there is clear corroborating evidence that standards in a subject have risen).
This points to a norm-referenced interpretation of performance.
    Ongoing proposals to reform the Junior Certificate in Ireland (see Junior Certificate
case study later in the chapter), where students may sit classroom assessments
(scored at school level) and assessment tasks and exams (scored externally), repre-
sent a hybrid system. The outcomes of the classroom assessments are intended to be
largely criterion-referenced, with students provided with feedback on their perfor-
mance, well before sitting their final exams. On the other hand, it is likely that the
final exam scores will be interpreted with respect to criterion-referenced and norm-
referenced frameworks. The criteria for each grade or descriptor are likely to be
clearly specified and linked to learning outcomes (a criterion-referenced perspective),
while it is also likely that the proportion of students achieving each descriptor will
be carefully monitored from year to year (a norm-referenced perspective).
     In the past, the most common measure of success was 5 A-C grades, including
  English and mathematics. A new grading scale will use the numbers 1-9 only (with 9
  being the top level). Those who fail to meet the minimum standard will be graded with
  a U (ungraded or unclassified), as before. The wider 1-9 scale is intended to provide
  greater differentiation between students achieving the middle and high grades.
     Three subjects were introduced in 2015 (for first examination in 2017): English lan-
  guage, English literature and mathematics.
     The transition to the new scoring system will be achieved through the use of ‘anchor
  points’. Hence, about the same proportion of students will achieve a grade 7 and above
  as currently achieve a grade A and above, while the top 20% of these (the highest per-
  formers) will be awarded a grade 9. The bottom of grade 1 will be aligned with the bot-
  tom of the old grade G.
      Broadly the same proportion of students will achieve a grade 4 and above as currently
  achieve a grade C and above. Grade 5 will be positioned in the top third of marks for a
  current grade C and the bottom third for grade B. It is also intended to align grade 5 with
  average performance in high-scoring countries in the international PISA assessment
  (described in Chapter 3), though it is not clear how this will be accomplished.
     The new        mathematics   assessment will be tiered (ie. available at two levels), with
  grades 4 and 5 available through both tiers.
      In the years following 2017, it is intended to maintain the standards set in the first
  year of the new regime. Statistical evidence will be used to support examiners in estab-
  lishing boundaries between grades. National reference tests will be used to monitor
  standards (and, presumably, also inform the setting of boundaries between grades).
two grades (for example, if a cut-off point for a pass grade is 50%, markers may be
encouraged to look for additional evidence in the scripts of students achieving 48 or
49%, with a view to raising their scores to 50%). Sometimes students may believe that
their exam papers have been scored too harshly, and may apply for re-grading. This
practice is consistent with criterion-referenced approaches to assessment and gener-
ally follows the release of marking schemes to teachers and students, as well as
examination scripts, so that a case for re-grading can be made.
Efforts to address some of these issues are apparent in current reform efforts in
England and Ireland described in this chapter. A noticeable feature of the reformed
GCSE is the shift from modular to linear testing (where grades will almost always be
based on a final examination      result, rather than on scores achieved     on tests and
tasks administered at different points during lower-secondary schooling). If we com-
pare the current reforms in England and Ireland, we see a stronger focus on the final
examination in the new GCSE and a reduction in the weighting given to the final
examination in Junior Certificate subjects. We see the abandonment of letter grades
in both examinations, though these are replaced by numbers in GCSE and by
descriptors in the Junior Certificate.
   The different approaches to reform, even in two adjacent jurisdictions, clearly sug-
gest that there is no universally agreed approach to examining student performance
at lower-secondary schooling, or indeed at any level. Moreover, the form that assess-
ment in general, and examinations in particular, take may be influenced by the pre-
vailing political narrative (for example, the view espoused by some politicians and
                                        Standardised tests and examinations         145
media outlets that the pre-reform GCSE exams were ‘dumbed down’ and ‘not fit for
purpose’) and hence needed to become more challenging.
Here, we see that the exam is weighted towards practical activities (250 out of 400
marks, or 62.5%), with the balance being allocated to the written component (150 out
of 400, 37.5%). Language examinations often include a practical component. For exam-
ple, GCSE French in Wales allocates 25% of the marks to an oral test that includes role
play, a photo card discussion and a conversation, while 25% each are allocated to lis-
tening (non-verbal and written responses), reading and writing (WJEC/CBAC, 2016).
Where practical and written components contribute to a final grade, it is important that
both components make a balanced contribution. A practical component in which most
or all students achieve maximum marks would mean that students’ grades are mainly
based on the written component. This would raise issues about the valid interpretation
of exam scores, as parents and employers might assume that a grade represents a
broad indication of proficiency in all aspects of the subject. Chapter 9 discusses this
issue further and offers useful additional readings.
summarise students’ strengths and weaknesses at the system level, can also be a
useful source of information for teachers.
   Teachers within secondary school may be involved in marking classroom assess-
ments that are linked to public examinations or certification. Teachers’ engagement
in moderating marks assigned to student work can lead to a greater shared under-
standing of what is meant by standards or learning outcomes, and thus can lead to
more effective teaching and learning (see Klenowski and Wyatt-Smith, 2014).
   In-house exams, including the ‘mock exams’ taken by students in many education
systems several months before they sit end-of-cycle exams, can provide useful feed-.
back on strengths and weaknesses both to teachers and students. The outcomes of
such exams, and indeed any other in-house exams administered by schools, can be
used to inform planning for teaching and learning. However, there is a point beyond
which taking practice exams is counterproductive. In some contexts, the content of
teaching and learning may be defined by past examination ‘papers, and this can lead
to a focus on only those aspects of a subject that are likely to be examined, to the
neglect of other important elements of the subject.
with performance on an exam offered at the end of Junior Cycle, with combined per-
formance on the task and the exam contributing to a student's final grade. Letter grades
will no longer be awarded. Instead, students will be given a descriptor that reflects their
performance in each subject: Distinction (91% to 100%), Higher Merit (75-89), Merit
(55-74), Achieved (40-54), Partially Achieved (20-39) and Not Graded (under 20%)
(DES, 2016a). This system of grading is different to that used for the Leaving Certificate.
    Certification at the end of Grade 9 will comprise a Junior Certificate Profile of
Achievement that will include the outcomes of classroom-based assessments, exami-
nation results (where performance on the externally scored assessment task is com-
bined with performance on an exam), the outcomes of short courses and, where
relevant, Priority Units.
    Finally, whereas in the past, all subjects were offered at two levels (higher and ordi-
nary) at Junior Cycle, and at three levels for English, Irish and maths (higher, ordinary
and foundation), the reform envisages two levels (higher and ordinary) for English, Irish
and maths, and a common level for all other subjects.
   The changes in assessment, though not universally welcomed by teachers in Ireland,
are intended to place a stronger focus on assessment for learning (formative assess-
ment) and to reduce the pressure on students associated with a single final exam in each
subject. Moreover, it is intended that the Junior Certificate Profile of Achievement will
recognise a wider range of learning, while reducing the focus on a final examination. It
should be noted that the proposed reform ofjunior cycle assessment arrangements has
been controversial and, at the time of writing, it remains to be seen what the final out-
come of the reform will be.
(Continued)
  3.   Some examination systems publish the marking schemes (or criteria) that have
       been used to mark exam papers, so a student can find out what answers were
       required to achieve different point scores for each question, and, perhaps, mount
       an appeal. What is an advantage of making marking schemes available? What is a
       disadvantage?
Chapter summary
This chapter, one of several that looks at how assessment information can be
used, extended several of the ideas introduced in Chapter 3 relating to standard-
ised tests and examinations. In particular, a distinction was made between norm-
referenced and criterion-referenced interpretative frameworks, though it was
acknowledged that test scores and grades can sometimes be interpreted using
both frameworks. In the case of standardised tests, a range of derived scores was
described including standard scores, percentile ranks and sten scores. In the case
of examinations, different schemes for reporting on performance were outlined
including letter grades, number grades and descriptors of achievement. Some of
these schemes were presented in the context of current reforms to examination
systems in the UK and Ireland.
   Sources of error associated with test scores and examination marks were also con-
sidered. The need to take measurement error into account in interpreting standard-
ised test scores was emphasised. Issues unique to examinations, such as the
weighting of practical tasks and tests in contributing to grades and the use of exam
results to improve teaching and learning, were considered. It was concluded that the
differing directions in which reforms to examinations in the UK and Ireland have
gone in the last few years, indicates that there is no unique internationally agreed on
blueprint for developing the perfect examination system.
  1.   The chapter discusses two interpretative frameworks for assessments - norm ref-
       erencing and criterion referencing - that can be quite technical in nature. Can you
       identify examples when teachers in class use such frameworks in their work with
       students?
  2.   Report cards sent home to parents sometimes provide information about students’
       performance summarised as percentile ranks or standard scores. Consider the par-
       ent who approaches you, firmly believing that the score (say a low percentile in
       reading) is a precise description of the student's ability, With a colleague, role-play
       the conversation that you, as teacher, would have with the parent.
                                             Standardised tests and examinations              149
Further reading
Baird, J.-A., Hopfenbeck, T.N., Elwood, J., Caro, D.,   & Ahmed, A. (2015). Predictability in the
  Irish Leaving Certificate. Oxford: Oxford Centre for Educational Assessment. Available at:
  http://pure.qub.ac.uk/portal/en/publications/predictability-in-the-irish-leaving-
   certificate(8385b46c-826d-4801-9921-d9fd3bbd2259).html
This report addresses aspects of the predictability of examinations, including positive and
negative aspects of predictability. Conclusions are situated in the context of examination
practices in the UK.
Simpson, L., & Baird, J. (2013). Perceptions of trust in public examinations. Oxford Review of
   Education, 39 (A), 17-35.
This article looks at the perceptions (expectations and requirements) of different stakeholders
regarding the credibility of A level examinations in England, and the multiplicity of factors that
contribute to credibility. The article highlights the complexity of ensuring that public exams
are trustworthy.
  What you will learn in this chapter
  The previous chapter described the processes of gathering and using information
  derived from standardised tests and high-stakes certification examinations. In this
  chapter, we focus more on the collation, interpretation and communication of infor-
  mation generated by assessments undertaken as partof normal classroom practice.
  First, we consider the nature of feedback and the distinction between feedback for
  student learning and feedback for other purposes. Then, the chapter explores a
  number of contexts in which feedback is offered, largely within the classroom,
  through discussion, conferences and illustrations of student work. The final sections
  emphasise recording and look at approaches to reporting feedback at individual,
  in-class and school levels.
      When you finish the chapter, you should be able to answer these questions:
  Overview of feedback                                    :
  A key feature of assessment for learning is the provision of information about student
  learning to guide subsequent teaching and learning (Wiliam, 2011). Kulhavy and Stock |
  (1989) suggest that feedback takes two possible forms for students: (i) verification of
                                       Record-keeping, feedback and reporting           151
the standard of performance reached; and (ii) elaboration of what is required to reach
a desired standard. Without elaboration, information from an assessment leaves it up
to the student to work out how to proceed. Elaboration empowers students to moti-
vate themselves and enables them to regulate their own learning.
   Effective feedback can be characterised as being:
Oral conferencing
Given the demands on teachers’ time in class, they can spend surprisingly little time
with any one student. Similarly, contact with parents can be sporadic, even in situa-
tions where technology offers the prospect of enabling greater contact between home
and school. In your own teaching, it is worth exploring how time can be arranged to
briefly discuss progress with individual students or groups of students. Group confer-
ences might focus on a group’s progress on a portfolio or other task. Providing feed-
back orally offers natural interactive dialogue between the key partners in education:
teachers, students and parents. Conferences can be formal, semi-formal or informal
and as frequent as is manageable for the teacher. Informal conferences can occur, for
example, when parents of primary children drop off or collect their children at
school. As fleeting and unstructured as these contacts are, they are nevertheless
highly valuable in building relationships between home and school.
              (
152    Understanding and applying assessment in education
Parent—teacher     conferences
These formal encounters help provide feedback to parents about the progress of their
children, also fostering productive cooperation between home and school. The for-
mat, frequency and duration of parent-teacher conferences will vary but they share
a focus on ensuring clear communication. Sometimes such meetings may be teachers’
only direct point of contact with parents in the year. Some schools structure meetings
as three-way discussions where the teacher, parent(s) and student meet together to
discuss progress, necessitating careful preparation by teachers and students.
   Guidelines for conducting effective parent-teacher meetings include the
following:
  Activity 9.1
  Try this out — Creating a rubric for a portfolio
  Mid-way through secondary school in Ireland, many students enrol in a ‘transition year’
  during which alternative curricula are followed. Some weeks are spent on work experi-
  ence with local employers, community groups and other agencies, and students are
  required to keep a record of their experience. A portfolio could be used by students to
  document their experience. Table 9.1 presents a partly completed holistic rubric that
  could be used to assess the portfolio. Work with a partner to expand the rubric to
  assess student portfolios at the end of the work placement (say three weeks). What
  evidence would you expect to see in the portfolio submitted by a student placed with a
  local supermarket, community group or other provider? What ‘entries could the student
  include? How could they be assessed?
  Exceptional
  (4 points)
  Strong (3)
  Acceptable (2)
  Needs further         Portfolio exhibits little or no evidence that learning outcomes associated with the
  development (1)       placement have been reached; little evidence of self-reflection; several required
                        elements are missing from the portfolio
       Review what.is written in Chapter 7 and in some other textbooks about portfolios.
    If possible, talk to a teacher who uses them in class. You might like to access another
    report by Koretz et al. (1994b), listed in the further reading at the end of the chapter.
    How trustworthy is evidence from portfolios (i) for formative feedback to students and
    teachers in class and (ii) for summative reporting to parents or other stakeholders such
    as education ministries? The companion website provides some useful resources to help
    you engage with this debate.
Teachers should:
e    know and understand how to apply the principles of assessment, recording and
     reporting as an integral part of the teaching process
e    have extensive knowledge and a secure understanding of the principles of assess-
     ment, of methods of recording assessment information, of the use of assessment
     in reviewing progress, in improving teaching and learning and in identifying next
     steps, and of the need to produce clear, informed and sensitive reports
e    tecord assessment information in a systematic and meaningful way in order to
     enhance teaching and learning and fulfil the requirements of the curriculum and
     awarding bodies
e    produce clear and informed reports for parents and other agencies which dis-
     cuss learners’ progress and matters related to personal, social and emotional
     development in a sensitive and constructive way.
156     Understanding and applying assessment in education
  Activity 9.2
  Try this out — Formulating a grading plan for your class
  Review a time in your school placement or teaching when you sought to provide sum-
  mative grades for all or some students in your class. Draw up a retrospective ‘grading
  plan’ for that occasion by answering the questions below:
Detailed treatment of how to combine marks from different assessments (exams, port-
folio, coursework) can be found in Nitko and Brookhart (2014) and Miller et al. (2013).
                                        Record-keeping, feedback and reporting           157
Comment-only marking
Concerns about the limited and even negative effect of grading have led to-an inves-
tigation of alternatives such as comment-only ‘marking. Though grades can motivate
some students, for others grades can have the opposite effect. Bangert-Drowns et al.
(1991) undertook a meta-analysis of various forms of feedback provided to students
in primary, secondary and college settings. They concluded that of four possible
types of feedback (right-wrong, correct answer, repeat until correct and explanation),
the least effective form was simply indicating to the student that the response was
correct or incorrect.
   Rather, grades need to be framed carefully within an overall feedback system
if they are to have the motivating effects that many teachers and parents believe
they have.
in class every day. It is not possible to retain all information about students in your
head. Rather, you need to judiciously select the most salient information that needs to
be retained for further use and record this in an appropriate manner. Finding the cor-
rect balance between record keeping and the ‘many other pressing tasks is something
that comes with time, and can even prove challenging to experienced teachers!
    We need to consider both professional and statutory reasons why teachers need
to keep records and make reports. Statutory or other required obligations stem from
government legislation or guidelines issued periodically by education ministries or
local education agencies. In the main, these often relate more to summative purposes
of assessment and reporting outcomes. Professional obligations also inform and
 determine teachers’ recording practices. Meeting the expectations of professional
 bodies, such as teaching councils and accreditation boards, generally requires teach-
 ers to adopt systematic record keeping and reporting. For example, 27 teacher com-
 petencies are identified by the General Teaching Council for Northern Ireland (2011),
 four of which specifically relate to assessment. Consider the two examples below:
Student achievement in curriculum       Benchmarks, proficiency levels, grades, numerals, letter grades, numerical
areas                                   scales, percentages, written descriptors
Test results                            Percentages, percentile ranks, standard scores, sten scores, grade-
                                        equivalent scores or other derived scores
Behavioural and study traits            Comments on attendance, punctuality, behaviour, class interaction and
                                        attitude
Activity 9.3
    Try this out — Review the student report template from one school
    Request access to the report card template for a school you know. Use the questions
    below to analyse the report card. Compare your card and report with a colleague who
    worked with a different report.
  4.   Identify any words, phrases, symbols or educational jargon that parents may have
       difficulty understanding; and any that students would not understand.
  5.   What information would you like to see added to the report?
  e    How clear and attractive is the report card? How would you improve it?
  7.   \f possible, discuss the report card with a parent and compare your own evaluation
       of the card with the parent's view.
Increasingly, schools use electronic report cards that enable schools to automate
some of the entries. Whereas electronic report cards can make the compilation and
distribution of reports more efficient, teachers need to ensure that comments
selected from ‘comment banks’ do not lose the individuality and impact that come
with traditional narrative approaches. Some teachers make use of free online report
card writers such as www.schoolreportwriter.com/
 Chapter summary
                                                  ng is crucial for the implementa-
 Management of information about student learni
                                                       sment approaches are of lim-
 tion of effective assessment systems. Elaborate asses                           ed
 ited use if the resultant information       is not recorded, distilled and communicat
                                                     feedback can take a variety of
 effectively to the relevant patties. Information or
              \
162      Understanding and applying assessment in education
   1.   A teacher notes that a Key Stage 2 student who received a low scale score on the
        national curriculum test in reading had, in fact, worked very hard during the school
        year and did her best. How can the teacher (and school) acknowledge the effort
        that the student put in, without undermining the interpretation of the score, which is
        benchmarked against national standards? Identify the challenges and possible solu-
        tions in this scenario.
   2.   Teachers sometimes feel that too much time is spent maintaining records in school,
        at the expense of teaching. Are teachers unique in this? Talk to two people who
        work in other professions. Identify the types of records they are required to keep,
        who requires this and the purposes to which the records are put. What conclusions
        do you draw?
   3.   Review the characteristics of effective feedback highlighted in the chapter. With a
        colleague, evaluate the feedback you have received to date in a current or previous
        college module in relation to this list. How would you improve the feedback? What
        resource implications are there in relation to implementing your suggestions?
Further reading
Brooks, V. (2002). Assessment in secondary schools. Maidenhead: Open University Press.
Two chapters by Val Brooks provide useful additional information to complement the topic
of the present chapter: Chapter 6 on record keeping and reporting; Chapter 7 on assessment,
accountability and standards.
Koretz, D., Stecher, B., Klein, S., & McCaffrey, D. (1994b). The evolution of a portfolio program:
   The impact and quality of the Vermont program in its second year (1992-93). CSE Technical
  Report No. 385. Los Angeles,      CA: University of California. Available at: http://eric.ed.
  gov/?id=ED379301
                                      Record-keeping, feedback and reporting           163
This report draws on data collected from teachers and principals involved in implementing
Vermont’s large-scale portfolio programme in the period 1988-93. Findings reflect issues of
manageability, the impact on teaching and the reliability of scoring writing portfolios.
Shute, VJ. (2007). Focus on formative feedback. ETS Research Report No. RR-07-11.
   Princeton, NJ: Educational Testing Service. Available at: www.ets.org/Media/Research/
   pdf/RR-07-11.pdf
This report provides an overview of literature exploring various aspects of feedback, espe-
cially for formative purposes. Advice on ways to enhance feedback, along with tips on what
to avoid, is provided.
What you will learn in this chapter
The principles of assessment first explored in Chapter 1 and revisited throughout the
book come together especially when we work with students from diverse contexts
and presenting with atypical personal characteristics. The fundamental aims of educa-
tion generally emphasise both personal and societal development and well-being.
Assessment should contribute to this development while supporting students’ acquisi-
tion and use of a broad range of knowledge, skills and competencies. In Chapter 4,
we saw how our concept of learning continues to evolve and broaden. Similarly, the
profile of students enrolling in school has evolved significantly in recent years.
Policies promoting the inclusion of as many students as possible within ‘regular’ or
‘mainstream’ schooling has altered classroom dynamics and required teachers to be
more eclectic and innovative educators.                     .
   Teachers need to teach and assess fairly and ethically, responding to student
diversity so that everybody is provided with learning suited to and taking account
of individual needs. Alex’s disruptive behaviour in class may simply be because he
finds the pace of lessons too slow and the content lacking in challenge, something
that ongoing.monitoring should identify. Sandra’s diagnosis on the autistic spec-
trum may explain why unexpected changes in lesson content or teaching (for
example, a surprise test) leave her agitated and uncomfortable in class. Concluding
that Lara, recently arrived from another country, is weak at mathematics based on
her score on a standardised test may be a misjudged and invalid inference when
using a test that was not normed with students whose second language is English.
To realise the promise and potential of formative and summative assessment, pri-
marily helping students to learn, teachers need to gather relevant and credible
evidence using appropriate methods. They need to ensure that each student’s edu-
cational experience is framed around the most appropriate expectations, content
                     Differentiating assessment to accommodate learners’ needs       165
and pedagogical approaches so that assessment works for rather than against the
student. This chapter reflects calls for what is termed universal design in learning
and in assessment, requiring that learning spaces, curricula, classroom manage-
ment, pedagogy and assessment be optimised for all learners, not only for those
who already have little difficulty in school.
   When you finish the chapter, you should be able to answer these questions:
                 f
166      Understanding and applying assessment in education
Policy response
International treaties
Increasingly, in policy at least, education systems seek to provide inclusive learning
environments for almost all students within regular or mainstream school structures.
One driver for this trend has been binding United Nations treaties such as the UN
Convention on the Rights of the Child (1989), ratified by 195 countries, that have
focused national governments’ attention on ensuring that children are educated in
the most inclusive environment possible. Perhaps more significantly in an assessment
context, the Convention on the Rights of Persons with Disabilities (UN, 2000), ratified
by 172 countries, sets out the legal obligations on states to promote, protect and
ensure the human rights of persons with disabilities. Article 24 deals specifically with
education and Section 1 of that article is presented in Figure 10.1:
   States Parties recognize the right of persons with disabilities to education. With a view to realizing
   this right without discrimination and on the basis of equal opportunity, States Parties shall ensure an
   inclusive education system at all levels and lifelong learning directed to:
   (a) The full development of human potential and sense of dignity and self-worth, and the
       strengthening of respect for human rights, fundamental freedoms and human diversity
   (b) The development by persons with disabilities of their personality, talents and creativity, as well as
       their mental and physical abilities, to their fullest potential
   (c) Enabling persons with disabilities to participate effectively in a free society.
Figure 10.1 UN Convention on the Rights of Persons with Disabilities, Article 24.1
National policies
In framing national policy and guidelines in relation to inclusion, national govern-
ments and state agencies draw on elements of the UN and other, earlier interna-
tional agreements. For example, guidelines for the inclusion of students with
special educational needs in Irish secondary education (DES, 2007a) acknowledge
the importance of meeting international obligations when implementing the pro-
visions of relevant national legislation in the form of the Education for Persons
with Special Educational Needs (EPSEN) Act 2004. This.act defines special educa-
tional needs as ‘a restriction in the capacity of the person to participate in and
benefit from education on account of an enduring physical, sensory, mental
health or learning disability, or any other condition which results in a person
                   Differentiating assessment to accommodate           learners’ needs        167
Teachers’ response
A clear challenge for teachers is how to meet the needs of such a diverse set of
learners, Assessment, in its broadest sense, clearly has a role in helping to identify
the specific context and/or impairment underpinning students’ needs. Often, when
teachers and parents think of assessment in relation to special educational needs, the
focus is on the more formal processes involved in obtaining a diagnosis for a child.
                                                                                             2005),
Under Section 8 (7) of the Disability Act in Ireland (Government of Ireland,
                                           of need   if the  parents   feel   that their child  has
children are entitled to an assessment
                                  assessment   is often   the  first official   confirmati on   and
a disability. The result of this
                                                                                       assessment
recognition by the state authorities that the child has a disability. The
report typically identifies the  following:
                                                                                   y
 e   whether the child has a disability and the nature and extent of the disabilit
 ©   the health and education needs arising from the disability
               {
168      Understanding and applying assessment in education
e     appropriate services to meet those needs (though delivery of these are not guar-
      anteed in the legislation)
e     when a review of the assessment should be carried out.
Whereas the assessment may highlight the services that should be provided to the
child, this is tempered significantly in the statutory service plan that is developed in
response to the assessment. Section 11 (7) (d) of the Act notes that the service plan
should be framed, in part, by taking into account the financial ‘practicability of
providing the services identified in the assessment report’, thus removing any statu-
tory entitlement for the child and parents to service to address the needs identified
in the assessment.
    As children enter and progress through the school system, the term assessment in
relation to students with SEN is often also perceived in relation to obtaining support
for students. Results of psychological, emotional, speech therapy, occupational ther-
apy or other assessments can form the evidential basis for requests for support and
services, such as a special assistant, a specialist referral or assistive technology. Many
authors focus significantly on specific aspects of special needs and the range of asso-
ciated assessments, for example Meyen and Skrtic (1995), Groth-Marnat (2009) and
Taylor (2008). The scope of this book is, however, focused more on tapping the
affordances of formative and summative assessment to inform teaching and learning
in relation to all students, including those with a range of special needs and charac-
teristics. One way to do this is to frame assessment as a key element of differentiated
teaching and learning. This is the focus of the next section.
approaches help establish baseline information about learners so that teachers can
tailor experiences to specific students.
    It is important that the forms of assessment used are sufficiently sensitivetoenable
an accurate identification of needs so that teachers can tailor appropriate educational
experiences to all students. For example, questions accompanying mathematics text-
books at primary level frequently use colour illustrations in representing data via
histograms. Caution needs to be taken in interpreting responses to such material by
students with any degree of visual impairment. Similarly at secondary level, using
cloze-type question paragraphs (where students supply missing words or phrases) to
assess the geography knowledge of students with English as a second language, can
 lead to invalid inferences about students’ understanding of geography.
                                                                              support
 e   the names of individuals contributing to the plan (e.g. teacher, parent,
     worker)
                                                                     other)
     a summary of the student’s abilities and needs (educational and
     long-term educational aims for the student                    ,
     short-term targets for the student
                                                                           , assistive
     instructional approaches and other supports (e.g. a special assistant
     technology) to help the student reach their targets
                                                                      reviewing the
 e   approaches to evaluating student learning and progress and to
     plan.
                                                                 monitoring of stu-
 Formative approaches to assessment facilitate the continued
                                                     by  person nel  working most
 dent progress and the re-calibration of support
                                                        -up   formal  review after a
 closely with the student on a day-to-day basis. Follow
                                                           provide s  the necessary
 longer period involving the school team and parents
                                                            and  approa ches for the
 information to help re-frame the educational expectations
                                                        at www.se ss.ie/ resources/
 next cycle of learning. Samples of IEPs can be found
 teaching-methods-and-organisation
170     Understanding and applying assessment in education
Activity 10.1
(Daily Telegraph, 2015). Patterns of migration ensure that similar student cultural and
linguistic diversityisa feature of schools in cities, urban and even rural areas_in many
countries, signalling the need to take language experience and competency into
account when framing assessments.
plan appropriate learning outcomes. The use of recording aids such as checklists or
rating scales helps provide increased precision to the observations and can result in
more informed and effective intervention strategies.                       ;
Feedback
Particular attention needs to be given to the role of feedback with students present-
ing with SEN. Used effectively, feedback can help highlight for students their current
levels of learning and how targeted levels can be attained. For atypical students who
may not be as well able to take in and process the feedback as other students in
class, teachers need to provide more scaffolding, perhaps using alternative methods
to communicate feedback and more sustained follow-up. The information may need
to be communicated more concisely and be more explicitly stated than for students
without such needs, for example through use of visual representations and short
phrases. The teacher cannot assume that these students will recognise or strategise
for themselves about how they can close the gap and bring about improved learning.
Their metacognitive and self-regulation skills may be less well developed, thus
 requiring greater input from the teacher if policies of inclusion are to be realised and
 bring tangible benefits to students. The student may need to have learning broken
 down into more manageable chunks. For example, when preparing for a written
 activity in class, the teacher may guide a student with attention deficit hyperactivity
 disorder (ADHD) by emphasising discrete steps such as organising their materials
 (pen, paper), finding the relevant page/section in the workbook, giving directions
 when the time is nearly up, and helping with strategies to ignore distractions. Similar
                                                                                       on
 care and structure is needed in relation to any feedback offered to the student
 performance on the task.
 Assessment tasks
                                                                         effectively
 Differentiated teacher-developed tasks and assessments can also be used
                                                                                   curriculum in
 with all learners. For example, the Key Stage 2 language and literacy
                                                     that  pupils   should   be  enabled  to ‘use
 Northern Ireland requires teachers to ensure
                                               ng   to  improve    their writing,  includin g that
 the skills of planning, revising and redrafti
                                                  n.d.).  A  learning   outcome   to support  this
 which they have composed digitally’ (CCEA,
 objective might be:
                                                                        such as plot, charac-
    We will develop a creative story using a number of writing features
             n
    terisatio and setting.
statement of financial position’ (CCEA, 2016b:       17). An in-class written test might be
designed as a series of short paragraph-length      responses to determine who has mas-
tered initial principles underpinning financial    statements before proceeding to more
complex application using simulated figures          for a company. For a student with
motor-control difficulties, a discussion with the teacher while others in the class
develop their written answers is more likely to capture levels of understanding accu-
rately than asking the student to write the test. This is an example of providing
accommodations within the classroom setting in response to the content matter, the
purpose of the assessment and the teacher’s knowledge of the student.
    adaptations to test format or administration (such as changes in the way the test is
    presented, the setting for the test, or the way in which the student responds) that
    maintain the same construct and produce results that are comparable to those
    obtained by students who do not use accommodations.
A distinction is drawn between such accommodations (that don’t change the construct
of interest, for example reading comprehension) and modifications that do. Like
accommodations,     modifications   are intended   to make   assessments   accessible   to as
many students as possible, but they may change what is being assessed for those avail-
ing of the modification. Allowing a primary-level student with autism spectrum disorder
(ASD) to provide evidence of oral communicative competency in a second language,
using a series of email exchanges with the teacher instead of a face-to-face structured
conversation like the rest of the class, would result in two different language constructs
being measured and should, therefore, result in different interpretations about the stu-
dent’s facility with language. Had the child participated in the conversation, but with
one or two short breaks (an accommodation), we could be more confident that similar
constructs are being assessed.
   Examples of other informal adaptations to formative assessments include:
e   changing the look of the assessment (if the text of a test in the chapter is in small
    print, reproduce the questions in larger font with more space for students to
    record their responses)
e   highlighting key words in an assessment prompt, either visually or by oral
    emphasis
e   asking students to state the question or task in their own words and checking
    this interpretation for accuracy
    giving students some indication of the typical length of answer needed
    using a different form of assessment (a series of short questions instead of an
    expository essay on a topic — this may enable the student to demonstrate more
  _ comprehensively the extent of understanding)
e — letting students determine the medium of response, for example an answer using
    a laptop, an annotated diagram or a concept map.
                    Differentiating assessment to accommodate           learners’ needs         175
In much of your teaching, where the general purpose is to check on student learn-
ing to inform subsequent teaching and learning, accommodations and modifications
are valuable in trying to provide as inclusive ‘an educational environment as possi-
ble. Some teachers might be hesitant to offer alternative forms and routes to assess-
ment for students of different abilities and characteristics. Yet, such an approach is
simply another element of the broader concept of differentiation and is consistent
with a socio-cultural interpretation of education. The fact that one or more students
may show their understanding in a different way to many of the other students in
the class simply reflects the diversity and need for inclusion in schools. It may
become an issue where students’ performance is being compared for certification or
other reasons, but in the context of formative assessment there is immense value in
differentiating the form, representation          and demonstration       of assessment      to suit
individual students, where practicable.
                                                                             in certification tests,
 In applying this in more formal assessment contexts, for example
                                                  to   offer     accomm    odatio    ns to enable
 teachers and assessment agencies often seek
                                                      ment      on   a  level   playing   field with
 students to demonstrate their learning and achieve
                                                      ,    2014).    Typical     accomm  odatio   ns
 other students (Tindal and Fuchs, cited in Thurlow
                                                     time,    a   differen t   format   of test  (e.g.
 in summative high-stakes tests include additional
                                                                    the examinee), a different
 a braille version of an exam paper, test questions read to
                                                         and    adaptat   ions in computerised
 location (e.g. in a small group in a separate room)
                                                       Cambri     dge    Assessm    ent for admin-
 testing. As part of the examinations developed by
                                                   provide      d   to   eliminat   e unnecessary
 istration worldwide, ‘access arrangements’ are
                                                            for   their  attainm   ent  (Cambridge
 obstacles and enable examinees to receive recognition                                         tions
                                                        Ireland,       the    State   Examina
 Assessment, 2015). In public examinations in                                               present-
                                                         dations     ’   for  examine  es
 Commission (SEC) may provide ‘reasonable accommo
                                                           examinations. For example, the
 ing with specific disabilities to help them take the
                                                         a  candidate needs the help of a
 SEC allows 10 minutes extra time per hour when
 scribe to complete an examination.
                                                         access to a reader or a scribe, or
     If a student avails of an accommodation such as
                                                        or punctuation, this may be noted
  4 waiver from the assessment of grammar, spelling
                {
176          Understanding and applying assessment in education
on the student’s certificate. While the Equality Authority in Ireland has argued that
this is discriminatory, the courts have, up until now at least, upheld the practice. For
example, in 2010, a High Court judge asserted that removing the annotation would
‘tip the balance too far in favour [of the candidate] to the detriment of others with a
legitimate interest in the fair and equitable administration of the Leaving Certificate’
Urish Times, 11 June 2010).
Table 10.1 Accommodations                 and modifications to help students access assessments and
examinations
e     lack of prior knowledge — the student does not have prior knowledge of the
      topic of the text and therefore cannot understand the information that has been
      presented
e     comprehension problem — the student cannot make the inferences required to
      answer comprehension questions
e     poor metacognitive skills — the student cannot manage their comprehension of
      the text, by, for example, re-reading a segment of text that is not understood, in
      relation to the purpose of reading
e     low motivation — the student may have been poorly motivated and may not have
      exerted sufficient effort. In PISA 2009 in Ireland, it was concluded that secondary
      school students’ level of engagement with the test had dropped compared with
      earlier cycles (Perkins et al., 2012)
e     high levels of anxiety — a student’s anxiety levels are related to performance in
      subjects like mathematics, with highly anxious students performing, on average,
      less well than students with low levels of anxiety (e.g. Perkins et al., 2013).
It is for this reason that a standardised test can only be viewed as providing a broad
indication of whether a student may or may not have a learning difficulty. If a student
achieves a low score, the teacher will need to reflect on the score and relate it to the
broader context in which teaching and learning occur:
e     Does the student’s performance on the standardised test correlate with perfor-
      mance on similar classroom-based tasks?
e     Does the student have a history of performing poorly on standardised tests and
      other assessments?
e     Does the student perform less well in one curriculum area, compared with
      another (perhaps pointing to a learning difficulty in a particular subject)?
e     What factors can be ruled out in explaining poor performance (e.g. can weak
      decoding skills be ruled out as an explanation for poor reading performance)?
e     Does the student have the language skills necessary to perform well on the
      test?
e     Does the student belong to a group who, on average, may perform less well on
      standardised tests (e.g. some students in socio-economically disadvantaged con-
      texts may struggle to achieve their potential in the face of risk factors such as
      poverty, poor nutrition or health-related challenges)?
e     Does one gender perform less well on the test than the other (e.g. male students
      often perform less well on average than females in language-based tests)?
¢     Does the student have one or more additional difficulties that might impact on
      performance, such as dyspraxia (sensory motor difficulties) or speech and lan-
      guage difficulties?
e     reading fluency — the ability to read with speed, accuracy and expression. This
      is often measured by computing the average number of words read per minute,
      while note is also taken of whether or not the student reads with expression,
      rather than in a word-by-word manner
e     reading comprehension skills — the ability to respond correctly to a range of
      questions about a text that has just been read orally or silently. Skills assessed
      may include sequencing events, identifying causal relationships, relating pro-
      nouns to their references, capturing the main ideas (whether implicitly or
      explicitly stated) and making generalisations and drawing conclusions.
 Beyond this, and again, for a small proportion of students, there may be a need to
 investigate whether a specific learning difficulty in mathematics such as dyscalculia
 (a specific difficulty with arithmetical operations) is present.            7
  In general, EAL students who have been learning English for a short period of time
  will not be expected to sit standardised tests. In PISA, for example, students who
  have received more than one year of instruction in the assessment language are
  expected to take part in testing (OECD, 2015b). However, it is arguable whether the
  test results of students who have such limited experience with the test language are
  valid, especially in light of the research on conversational fluency and academic lan-
  guage discussed earlier in the chapter. Certainly, such results are unlikely to provide
  a strong basis for making decisions about teaching and learning at the individual
  student level. This points to a need to employ a broader range of assessment tools
                                                                                        for
  in working with EAL students, including tools that will provide useful information
  making informed instructional decisions.
      Fortunately, there is a range of tools available for this purpose. These include:
                                                                                         7, but
      e   performance-based assessments such as those described in Chapter
                                                                        modified      language
          modified as needed (see, for example, Abedi (2010) where
          is used)
                                                                                     , reading
      e   initial and ongoing assessment of language skills (speaking, listening
                                                                                  and  Training
          and writing) using schemes such as the Integrate Ireland Learning
                                                                          )  or  the  Europea  n
          primary and post-primary language toolkits (IILT, 2007a, 2007b
                                                                       of Europe,     2012)   for
          Common Framework of Reference for Languages (Council
                                                                                in the assign-
          pre-school, primary and post-primary levels. These toolkits result
                                                                             for   instructional
          ment of proficiency levels to students and can provide a basis
                                                                                  compone  nts,
          decision making. Some of these measures include self-assessment
          where students can assess their own proficiency
                                                                   e, such as the Test of
      e   standardised measures of English as a second languag
                                                                         which comprises a
          English Language Learning (TELL, Pearson Education, 2016)
                                                                   tic test.
          standardised screening measure and a follow-up diagnos
182     Understanding and applying assessment in education
Chapter summary
The broad focus of this chapter was on differentiating assessment to address the
needs of a broad range of learners who, for one reason or another, may struggle on
formative and summative tasks designed for the general population of students, or
may find them very easy. Among the groups considered were students with sensory,
emotional or learning disabilities, including specific learning difficulties; students for
whom English is an additional language; and gifted and talented students. The full
range of disabilities and special educational needs likely to be encountered by you
in your teaching is beyond the scope of this book, but more specialised sources are
provided in the text and on the companion website as part of Activity 10.1. It was
noted that, in many cases, tests designed for a general population may not be suitable
for students in these specialised populations, as their needs were not taken into
account in designing the assessments.
   The provision of reasonable accommodation was identified as one approach to dif-
ferentiating assessment for special populations of learners. However, it was noted that
accommodations should not alter the underlying construct that is being assessed. For
example, if a reading passage is read aloud to a student, the assessment is no longer
one of reading comprehension for that student. Instead, modifications such as this were
identified as changes to an assessment that alter the construct being assessed.
   Two broad approaches to assessment that are more often associated with AfL than
AoL are performance-based and diagnostic assessments. Both approaches also allow
                 Differentiating assessment to accommodate              learners’ needs         183
   1.   Review the assessment policy of one school with which you are familiar or access
        the planning of one teacher. To what extent is formal identification of special educa-
        tional or other needs incorporated into the planning? What are the steps employed
        in identifying such needs? How is the differentiation of formative and summative
        assessment evident in the plan in relation to atypical learners?
   2.   If possible, seek access to the individual education plan (IEP) for one student. Work
        openly with the school, parent and student so that your purpose in accessing the
        plan is clear and understood, and the confidentiality of the information is main-
        tained. Identify the nature of the disability or other circumstances leading to the
        development of the IEP. What are the main targets for the student? For each tar-
        get, identify (i) the approach to assessment and (ii) how the assessment information
         might be used to help the student.
 Further reading
                                                                     summative assessment for
 Cumming, JJ., & Maxwell, G.S. (2014). Expanding approaches to
    students with impairment. In L. Florian (ed.) The Sage handboo k of special education, Vol. 2
    (2nd edition, pp. 573-93). Los Angeles, CA: Sage.
                                                                       SEN in the school system in
 This chapter offers an insight into the inclusion of students with
                                                     assessme nt  and   critically analyses the provi-
 Queensland. It advocates more equitable forms of
                                       nt.  The  chapter also  focuses    on the role of moderated
 sion of accommodations in assessme
                                                                 iated assessments to students of
 school-based assessments in providing appropriate different
                                                                   tion purposes.
 different background and ability, even for high-stakes certifica
e   How is assessment planned and used to promote and prioritise key educational,
    social and economic aims?
                                                      Assessment planning and policy       185
                                                                                         ce
    The central role of government in such exams suggests a level of importan
    above that of other forms of assessment not prescribed to the same extent.
       A number of general characteristics of these exams can be identified. The develop-
                                                                                government
    ment, administration and scoring of the exams is tightly regulated by
                                                                                 at terminal
    agencies or licensed providers. Exams are typically taken by students
                                                                         secondary and, in
    points of schooling: at the end of compulsory, the end of upper
                                                                             The exams are
    the case of some developing countries, at the end of primary school.
                                                                         formats, including
    highly aligned with specific subject areas and use a variety of item
    essay   or constructed   response,   objective,    aural, oral and practical demonstration,
                                                                             rk or assessment
    some of which may be undertaken as part of moderated coursewo
                                                                            numerals that con-
    at school level. Performance is frequently expressed as grades or
                                                                            college admissions
    vey a clear meaning to stakeholders including students, parents,
                                                                             certifying student
    officers and employers. The prime functions of results include
                                                                      of education. Additional
    achievement and the selection of students for the next level
                                                                  practice, directing learning
    uses involve motivating students, monitoring educational
                                                                         accountability.
    and teaching in schools and facilitating individual and system
                                   notewo rthy  in relatio n   to  public   examinations. The
       Two planning issues are
                                                              of such tests is a highly com-
    development, administration, scoring and reporting
                                                                  control practices in educa-
    plex process subject to some of the most robust quality
                                                                 s and parents ensures that
    tion. In addition, their importance to students, teacher
                                                             terms of timetabling, planning
    very significant efforts are expended by schools in
    and teaching, especially at secondary level.
                   4
186    Understanding and applying assessment in education
National       assessments
Whereas assessments to certify student achievement offer some insight into patterns
of achievement within a country, they miss many important dimensions of learning.
Certification usually relates to terminal points in education, normally the end of com-
pulsory or upper-secondary schooling, thus failing to capture achievement levels
throughout primary schooling and at intermediate points in secondary schooling.
Additionally, although across OECD countries, sizeable majorities of students com-
plete upper-secondary education, some do not, with completion rates ranging from
45% (Mexico) to 98% (Korea). Therefore, policy-makers plan alternative national
assessments to monitor and promote achievement at multiple time points during
students’ education. In this way, assessment is highly consequential to hundreds of
millions of students and their teachers across the globe.
   National assessments are frequently designed around a small number of ‘core’
subject areas to identify the overall levels of achievement in a population, such as
a particular grade level. Inferences about population achievement are often based
on samples of students and not all students are assessed. For example, the 2014
National Assessments of English Reading and Mathematics in Ireland tested approx-
imately 4000 students at Second Class and 4000 students at Sixth Class (Shiel et al.,
2014). Results are aggregated at national level to create a summary of achievement
for the system as a whole. In contrast, census sampling in England, Wales and
Northern Ireland, after the introduction of the national curriculum in 1988, meant
that SATs or other assessments were administered to all students at particular grade
                                              Assessment planning and policy          187
levels (usually at the end of Key Stages 1, 2 and 3). In this case, results are com-
municated to parents and to schools, the latter of which are expected to draw on
the results in planning subsequent teaching,
   The types of questions that national assessments are designed to answer include
the following:
evolution has occurred in relation to reading and science, with a significant redefini-
tion of the reading expected in the framework for PISA 2018 to reflect students’
increasing engagement with digital text.
   There is much policy and research support for the value of international assess-
ments, for example Schleicher (2006) and Sahlberg (2015). Alongside this support are
the contrasting voices of those who argue that a range of challenges complicates the
interpretation of results, casting doubt on the appropriateness of using such surveys
of achievement at all. Issues include translation difficulties and simplistic interpreta-
tions of scales (Goldstein, 2004; Arffman, 2013). Nonetheless, there can be little doubt
about two points: first, enormous effort is associated with the planning, preparation,
implementation, analysis and reporting of such assessments;                         and   second,     such
assessments are likely to continue for the foreseeable future.
2006      Mathematical literacy is an individual's capacity to identify and understand the role that
and       mathematics plays in the world, to make well-founded judgements and to use and engage with
2009      mathematics in ways that meet the needs of that individual's life as a constructive, concerned and
          reflective citizen.
2012      Mathematical literacy is an individual’s capacity to formulate, employ and interpret mathematics
and       in a variety of contexts. It includes reasoning mathematically and using mathematical concepts,
2015      procedures, facts and tools to describe, explain and predict phenomena. It assists individuals to
          recognise the role that mathematics plays in the world and to make the well-founded judgements
          and decisions needed by constructive, engaged and reflective citizens.
Strands Elements
 Learning outcomes (for exploring and        Progression steps (for use the writing process when
 using language in writing)                  creating texts collaboratively or independently)
 Draw and write with a sense of purpose      Uses personal experiences and interests as stimulus for
 and audience while creating texts in        their texts.
 a range of genres, and develop an
                                             Begins to take part in collaborative writing with the teacher
 individual voice to share thoughts,
                                             as scribe.
 knowledge and experiences.
                                            _ Uses familiar topics as a stimulus for texts, while creating
 Use the writing process when creating
                                              texts jointly with the teacher.
 texts collaboratively or independently.
                                             Collaboratively plans texts orally; sequences and writes texts
 Elaborate on the meaning of own
                                             with other children; reads and talks about their writing.
 writing and discuss the texts of others,
 showing an emerging recognition of the      Independently plans, sequences and writes texts; reads and
 author's intent.                            talks about their writing with the teacher to check if it makes
 Write using cursive script                  sense and begins to suggest simple changes to improve it.
Figure 11.2 Illustration of selected standards in the primary English curriculum: Ireland,
Junior Infants to end Second Class
collaboratively or independently). For that one outcome, the curriculum also specifies
eight progressive steps to guide student learning (and assessment) up to the end of
Second Class (midway through primary school). As you can imagine, planning assess-
ment systems nationally and at school and class levels to capture the complexity of
such a standards framework is a demanding and, as yet, unfinished challenge.
   Curricula expressed in terms of standards have been around for some time, yet
assessment and reporting approaches have been slow to adapt to reflect the changed
emphases. Many schools, teachers and education systems continue to operate points
                                                                                       a
or grade-based reporting systems that attempt to represent student achievement in
subject in a single numeral,   grade or letter. This approach conflicts with standards-
                                                                      the grade or
based approaches (lamarino, 2014; Munoz and Guskey, 2015). In effect,
                                                                          curriculum is
numeral becomes the standard by which student achievement on the
judged. One’s progress in relation to the standard (very broadly defined) is established
                    iy
190      Understanding and applying assessment in education
using different grade points along the scale (A, B, C; or H1, H2, H3). A challenge with
this approach is establishing the criteria associated with any particular grade. What
does a B mean? What knowledge and skills does a student possess who earned a H3?
Standards-referenced approaches try to link grades or assessment outcomes more
overtly to the type of performance criteria illustrated in Figure 11.2. Thus, planning
assessment and reporting systems that are compatible with standards-based curricula
requires increased transparency about levels of student performance.
e         the use of screening and diagnostic assessment to identify learning strengths and
          difficulties
          the role of psychological assessments
          the recording and storing of assessment Outcomes
e         the personnel involved in assessment and their responsibilities. (PPDS, 2007)
        Students adopt reflective approach               Self-assessment by students helps them plan and
                                                         monitor their own learning, including development of
        Students’ knowledge, skills and overall
                                                         key skills
        achievement are improving or at an
        appropriate standard                             Students act on feedback received
        Set learning objectives based on student needs   Develop a relevant whole-school policy on
                                                         assessment including AoL and AfL and implement it
        Employ data-led differentiated instruction
                                                                                                  learning and
        Employ AoL and AfL focusing on knowledge,
                                                         Frame assessmen: t to: support students’
                                                         to measure their attainment
        skills and dispositions
                                                         Adopt shared approaches to providing feedback,
        Provide useful feedback to students
                                                         informed in part by collective review of students’ work
        Use success criteria, questioning, self- and
        peer assessment,   and feedback                  Communicate and collaborate with parents and other
           Se            ;                               relevant parties in relation to student learning
        Maintain appropriate records                        yee            ;
                 :                                       Maintain appropriate records
                                                                  ’            ;         ‘
        Engage in relevant CPD                           Engage in collaborative reflection and CPD
    Activity 11.1
    Try this out - Benchmark your assessment competencies
    against a school assessment policy
    Seek access to the assessment policy of a school with which you are familiar. Note the
    elements or headings of the policy. How much of the policy focuses on formative pur-
    poses of assessment? How much on summative purposes of assessment? To work effec-
    tively as a teacher in this school, what do you need to know about assessment? Using a
    green highlighter, identify elements in the policy that you understand fully and feel
    confident about implementing. Use another colour to highlight elements about which
    you would need more information, advice or experience.
Planning by teachers
Long-term planning
At school level, local or nationally developed templates generally inform expectations
about and approaches to planning. Readers with experience of teacher education
programmes will be familiar with planning guidelines and formats promoted at pro-
gramme level. This may take the form of long- and short-term planning processes,
perhaps with individual lesson planning, in the early career stages at least. Long-term
planning provides high-level direction to help ensure that learning is logically con-
nected so that your teaching (and students’ learning) is consistent with and builds on
previous learning. This facilitates consolidation, continuity and progression over a
defined period of time such as a year or a term. For teachers, long-term planning
involves the consideration of a number of curricular and pedagogical dimensions.
These might include learning aims for a subject, the specification of content and
skills, the characteristics of learners, the teaching and learning methods to be
employed, assessment approaches, resources, differentiation and personal reflection
in relation to teaching.
    Professional or regulatory agencies also offer suggestions to teachers about the
expected content of planning documentation. However, it is understood that plan-
ning is highly individualistic and that teachers will vary in the scope and detail of
planning, depending on their career stage and familiarity with the curriculum.
                                                    Assessment planning and policy               193
In Ireland, for example, novice primary teachers are encouraged to develop long-
term plans for each subject, where each plan covers the expected teaching and
learning over a term or half-year, resulting in perhaps two plans for each subject
over the course of a year (NIPT, 2013). Whether in fact you refer to such planning
as long-term or medium-term is not so important. Of more relevance is the use of
such planning to think strategically in advance about the scope, sequence and
resourcing of learning over an extended period so that students are provided with
a coherent learning experience. As an integral element of teaching and learning,
assessment requires careful consideration in such plans. In developing your per-
 sonal plan, you will draw on the school plan along with your own resources suited
 to the subject and age or year group.
     Whereas, traditionally, assessment might have focused on the conduct of school or
 statutory tests with associated reporting commitments, embedding AfL as discussed in
 Chapters 4 and 5 requires a deeper integration of assessment thinking and practices
 into your planning. Therefore, you could highlight in broad terms in the long-term plan
 where and how assessment purposes and approaches (formative and summative) will
 feature in each subject area across the period of the plan. Ideally, this should specify
 a combination and balance of formative approaches, such as the six dimensions high-
 lighted in Chapter 5 — use of learning intentions, success criteria, providing feedback,
  questioning, peer assessment and self-assessment. Considering such approaches at the
  design or planning stage in advance of teaching will help ensure that the progress and
  learning of individual students will be at the heart of your teaching, including the needs
  of students with special educational needs, as discussed in Chapter 10.
      Long-term planning also affords a crucial opportunity to specify a range of sum-
  mative assessment approaches for each curriculum area alongside and complemen-
  tary to the formative methods. You may find that many of the more summative
  approaches highlighted in Chapters 3, 6 and 7 are either required or advantageous
  in optimising learning experiences for your students. In the plan, you can specify
                                                                                                      s
  the nature and timing of standardised tests or tasks and highlight the procedure
                             assessment    s  of  students’   learning  (for  example,    at  the  end
  and timing for periodic
                                                                                                to dif-
  of topics or end of term). You can also link specific assessment approaches
                             or  topics   within   areas.  This   might  involve   planning    for  use
  ferent curriculum areas
                               the  first term  to  highlight   student  developme   nt  at  mid-year
  of a portfolio throughout
                                                                                      media and/or
  parent-teacher meetings, or identifying opportunities for using digital
                                        .   Throughou   t  this  book,  we   have   emphasise   d the
  digital approaches to assessment
                                                                                  ensuring that the
  value of enacting diverse assessment methods in your teaching,
                                                                                        on the basis
   monitoring of and judgements about students’ learning are undertaken
                                    multiple   approache   s   and  sources  of  evidence.   The best
   of frequent assessment using
                                           of assessment     purposes    and  approache  s    exists in
  way to ensure that such plurality
                                                                                 while allowing for
  your teaching is to build it in at the design stage from the outset,
                                                                          long-term planning, you
   flexibility and change as the year progresses. Through such
                                                                               will bring this about
   can clarify what you hope to achieve, strategise about how you
   and thus teach with more confidence.
 Short-term planning
                                                            r education programmes or
 The use of pro-forma templates is common within teache
                                                             frames. Cohen, Manion and
 school systems to guide teachers in planning to short time
                 f
194      Understanding and applying assessment in education
Morrison (2004: 126) characterise short-term planning as where teachers ‘set out what
they will be teaching on a week-by-week, day-by-day, and lesson-by-lesson basis’.
They contrast strategic long-term planning with more tactical short-term planning,
where teachers identify specific learning outcomes for students. Short-term planning
may take the form of fortnightly or weekly plans where some detail is provided about
what will be covered in each subject and, for student and novice teachers, lesson
plans developed to a variety of suggested formats. In either case, the need to be
flexible in the face of classroom reality is essential. With the best will, teachers can
plan for every eventuality but should be sensitive to the opportunities provided by
amending planned outcomes and learning experiences over the course of a week, day
or lesson. This is not to suggest that careful planning is unnecessary; rather, through
planning you will be emboldened with confidence to teach well, and through the
experience of teaching you will have the confidence to modify that plan in real time
in the classroom.                                              :
   Incorporating assessment into short-term planning involves focusing on what to
assess and how to use the information from that assessment for the intended pur-
poses. Such tactical planning is hugely important as it is likely to influence the extent
to which students benefit from teaching. Accordingly, the General Teaching Council
for Scotland (2012) requires registered teachers to:
These two standards come together when teachers begin to consider how to incorpo-
rate assessment into their planning and teaching.
    In framing a short-term plan, whether in weekly or lesson plan format, you need
to draw on the detail or direction contained in the long-term plan and identify more
specific learning outcomes and learning experiences for students. Monitoring stu-
dents’ progress vis-a-vis the intended outcomes is facilitated by the application of
appropriate assessment methods that are directly calibrated with the learning out-
comes. As indicated in Chapter 5, many teachers use learning intentions and success
criteria to activate students’ metacognitive awareness around lesson content. In
choosing specific assessment approaches, care should be taken to ensure that the
lesson outcomes or intentions are clear and relevant and that the assessment is the
most direct or at least an appropriate means to ascertain if students have been suc-
cessful in engaging with and achieving the intentions. Whether for a fortnightly/
weekly plan or for a lesson plan, comprehensive use is made of assessments by
posing, and acting on the answers to, three questions:
1.    To what extent have individual students successfully attained the intended learn-
      ing Outcomes?
2.    To what extent has the class or group as a whole successfully attained the
      intended learning outcomes?
3.    How can I use the assessment information to plan the next learning experiences
      for individuals and for the group?
                                               Assessment planning and policy        195
The last question in particular illustrates the power of assessment to facilitate reflec-
tion by teachers on their own practice and thus transform teaching. Teachers need
constantly updated information about student learning if they are to cater to the dif-
ferent needs of individuals. This may involve. modification to teachers’ expectations
of students, to the learning outcomes teachers specify, to teaching and learning strat-
egies and to resources. It is this cyclical process of planning, teaching, assessing,
reviewing and re-planning that offers the potential for real continuity and progression
in student learning, as illustrated in Figure 11.4.
                                                    2. Selection of
                         “if Specification       content, experiences,
                            of aims for              resources and
                             learning                  methods
                                                                         In framing
    The power of assessment is embedded in all four stages in the model.
                                                                through long-term or
 teaching and learning aims or outcomes (Stage 1), whether
                                                               expressed in tangible
 short-term planning or lesson planning, aims need to be
                                                                approaches available.
 form amenable to assessment through some of the varied
                                                                  hes, assessment is
 Adopting a combination of formative and summative approac
                                                             intention for an 11-year-
 inherently part of Stages 2 and 3. For example, a learning
                                                            persuading classmates of
 old student might be to learn how to give a short speech
                                                           experience (Stage 2), the
 the benefits of school uniforms. In planning for this
                                                                 g them to self-assess
 teacher can co-construct a simple rubric for students, enablin
                                                         video or audio. Part of the
 their own input, perhaps using digital capture with
                                                              by students of their own
 implementation of the lesson could include self-analysis
                                                              review by the teacher in
 performance using the rubric (Stage 3). After the lesson,
                                                       should help steer subsequent
 relation to the outcomes of the lesson Gtage 4)
                                                       thus feeding into specification
 engagement with the topic or development from it,
                                                      ant in this model is that assess-
 of the next learning outcomes. What is most signific
                                                        3, nor is it contained solely in
 ment does not come as a separate stage after Stage
                                                   d by assessment that has been
 Stage 4. That stage uses the information provide
                                                 learning sequence from the outset
 planned, designed and built into the lesson or
196      Understanding and applying assessment in education
to help teachers understand       more   about the learning and teaching environment
and processes in class.
Chapter summary
 Figure 11.4 encapsulates the essence of the chapter, highlighting the message that
 assessment is a planned activity that runs in tandem with and is not subsequent
 to planning, teaching and learning. Assessment is a significant endeavour, whether
 at the macro international level or at the micro level of in-class monitoring. At
 every level, care should be taken to ensure that the instructional time spent on
 assessment yields learning and enactable information dividends for students,
 teachers, parents, policy-makers and wider society. Information from assessment
 can result in significant changes to educational policy —+nationally and at school
 level. Information derived from assessment counts — ask the many education min-
 isters whose systems were rocked or just reshaped by students’ results on PISA.
 Or ask the student whose dyslexic difficulties were not identified until late in the
 primary education years. Assessment matters, so it is important to embed suitable
 approaches in teaching and learning from the outset. The scale of the task and the
 resources required may differ when planning for a national assessment or when
 an individual teacher undertakes short-term planning for a secondary school
 course in Art. This is different in scale, but the importance of assessment to the
 system or to the individual student is great. This book, overall, has illustrated many
 approaches to assessment and how they can be used. This chapter has focused on
 issues around planning so that the affordances of assessment can be harnassed for
 the benefit of the many stakeholders involved in. education. As the old adage goes,
fail to prepare and prepare to fail.
Further reading
Cohen,   L., Manion,   L., & Morrison,   K. (2004) A. guide to teaching practice   (5th edition).
   London: Routledge.
A well-established text that includes chapters on planning for teaching and on assessment.
This article explores the consistency with which teachers mark student performance on
assessments where there are high-stakes implications. The author identifies the need for a
greater investigation of the efficacy of consensus moderation across teachers.
What you will learn in this chapter
Throughout this book, we have advocated a balance between the needs of various
stakeholders in relation to assessment, with one primary stakeholder, namely the
learner. The learning, development and welfare of the individual student should be
the fundamental focus of assessment efforts. In practice, that individual’s development
is largely set in and dependant on contexts such as family, peers, school, education
system and wider society. Therefore, a comprehensive understanding of assessment
of the individual requires acknowledgement of other forms and purposes of assess-
ment that may focus on larger units such as schools and populations of students. A
number of principles, summarised in Chapter 1, guided our discussion, analysis and
suggestions in relation to formative and summative assessment. We pick these up
again in this final chapter where we offer our interpretation of a range of opportuni-
ties, pressures and trends that are likely to shape educational and assessment practice
in the coming years.
    When you finish the chapter, you should be able to answer these questions:
    How can'we ensure that the needs of the learner remain the focus of assessment?
    In what ways are those needs met by different individuals and groups?
    What are some of the issues relating to assessment that may come to the fore in
    the coming years?
                                                                          teacher
                 portfolio assessment                                     observation
              conferencing                                                         teacher-designed
                                                                                   tasks and tests
      self-                                                                             standardised
      assessment                                                                        testing
   As outlined in Chapters 3 and 11, great care is taken in the development, admin-
istration, scoring, reporting and review of standardised tests and other large-scale
certification tests. Credit must be given for the seriousness with which they are
planned and the quality controls established to provide a justifiable warrant for their
use. When we think of the question: what forms of assessment should I use? it is
useful to reflect on the aims of education, shared by most within society. We wish to
develop the individual and individuals as a group that we call society; we want to
push the boundaries of thinking so that individuals and society can flourish and
prosper; we want to provide the opportunity for today’s young people to develop
their capacities for thinking, for invention, for feeling, for caring and for learning that
will help them inherit, cherish, protect and enhance the world we leave behind. If
these are some of the aspirations for education, then formative assessment can assist;
sO can summative assessment. Teachers need to know about both and be confident
and competent in drawing on whatever is the most appropriate assessment at the
right time for the right purpose. Generally, that purpose will have to do with the
student, the individual. Sometimes it will relate to the group: class, school, LEA,
national education system.
Diversity of stakeholders
It is increasingly rare that centralised authorities stay apart from education or from
assessment. Governments get involved in education and, by and large, foot the bill
at primary and secondary level. There are, of course, personal costs associated with
education: in some instances, these are by choice where parents opt to purchase
educational services outside the state system, or where parents must contribute
towards their child’s education even though it is nominally ‘free’, perhaps through
book purchases, uniforms or trips. Exchequer funding for education is a major
component of the state spend in many countries, with over 5% of GDP internation-
ally channelled to education. Amongst other motivations, this is one reason why
                                   Conclusions about assessment in education             201
policy-makers seek information about the functioning of the education system. This
is One summative    purpose for assessment.
prompt action at national levels. PISA, like other international assessments, can
nudge national policy towards novel approaches, providing additional impetus for
initiatives that may well have been gestating locally. On the other hand, we wonder
about the authenticity of some PISA tasks designed to assess 21st-century skills.
Examples include collaborative problem solving in 2015, where collaboration means
engagement with pre-set dialogue and the selection of pre-set responses rather
than human-to-human problem solving (see OECD, 2016b and the illustration in
Figure 12.2). The withdrawal of a number of OECD countries from a proposed
assessment of global competencies in PISA 2018 also points to the tensions that can
arise when new assessment domains are introduced.
   Figure 12.2 presents an illustration of the type of digital interface and item pre-
sented to'students when addressing collaborative problem-solving tasks on the PISA
2015 test. In this field-trial example, the task for the examinee is to engage in decision
making with computer-simulated classmates and a teacher.to plan a trip for a group
of exchange students coming to visit the school. Each successive screen adds to the
simulation, prompting the examinee to make decisions indicative of their capacity to
work collaboratively with peers. Student ‘decisions’ are made by responding to
multiple-choice items for each scenario, under the broad umbrella of planning for
the exchange students’ trip.
Notepad
         ..          ieee              B
                Lot's asks. Cosmo what we should do.
                                                                       ag
         :    1 Maybe we should think eboutthis ‘or awhile and check        /
                           1?
Figure 12.2 Example of digitally administered task assessing collaborative problem solving
   In the illustration in Figure 12.2, the automated scoring algorithm credits the
fourth option (let’s discuss what’s necessary for a good visit) as it illustrates that the
student can identify and describe the tasks to be completed. The other options are
                                      Conclusions about assessment          in education        203
not credited in the scoring process as they do not illustrate sufficient levels of col-
laborative problem solving. Individual scoring rubrics are presented for each sce-
nario presented to examinees and a total scaled score for the task is derived from
the examinee’s responses across all elements’of the simulation.
being accumulated all the time — but perhaps the evidence is not used sufficiently.
We see this problem linked as much to communication as to research per se.
Principals, teachers and policy-makers need to know how students learn and how
assessment can assist learning. The What Works Clearinghouse (http://ies.ed.gov/
ncee/wwc/) is one illustration of how one education system has attempted to syn-
thesise the results of research to facilitate policy and classroom practice, though
caution is needed as methodological constraints may be in operation (for example,
the prioritisation of studies that used random controlled trials). The modern digital
environment offers fertile ground for sowing the seeds of research-led practice, in
assessment as in all aspects of education. Books, academic and professional jour-
nals, digital journals, online repositories, alternative media such as blogs, video
posts and illustrations all have a place in sharing not only good practice but
evidence-led practice. Schools are busy places and teachers’ time is increasingly
stretched to fit across a myriad of responsibilities. We need to find ways to share
high-quality research about assessment in a succinct manner compatible with
teachers’ lives and the reality of schools.
   In addition to keeping apprised of research findings about assessment, we
encourage teachers to actively investigate their own practice in class and at
school. For example, you might wish to introduce learning logs as part of your
assessment practice (see Chapter 5). Table 12.1 illustrates a few questions you
could usefully consider.
   One way (but not the only way) this call to investigate your practice can be
expressed is graphically using a logic model (see, for example, Kellogg
Foundation, 2004; Kekahio et al., 2014). Such models assume that in any initiative
(say, introducing learning logs into your class) the class context, certain inputs,
activities, outputs and outcomes are all highly relevant. Figure 12.3 presents a
simple logic model revealing a teacher’s assumptions about what will happen
when learning logs are introduced into the class as part of AfL (A in Figure 12.3).
It is assumed that a three-month use of the logs (B) will provide an opportunity
for students to describe and reflect on their learning, say in musical appreciation
(C). This in turn can lead to analysis and consideration                  of the student data by
teacher and students (D), further response and modification to teaching and
learning by both (E), with noticeable changes in the quality of student learning
outcomes as a result of the process (F). It is based on these assumptions about
what might happen that a teacher may decide to try out learning logs. Many vari-
ations of this process exist but all provide the opportunity for teachers to concep-
tualise in advance how an assessment should work in class, and offer a structure
for trialling the process.
                                                   Conclusions about assessment           in education       205
                     C.                            D.                  E:                      F.
                    Logs                     Reflection          Differentiation             Quality
                  provide                  and response            by teacher:                  of
                  data on                    to data by:            Enhance                  learning
                 levels and                    Teacher              students’               improves
                 processes                     Student              motivation
                 of student                                             and
                  learning                                             SRL!
Figure 12.3.       Possible logic model underpinning teacher's investigation of using learning
logs in class
Note: 1. Self-regulated learning (see Chapter 4)
    Activity 12.1
    Try this out - What are the implications of selected trends in
    assessment for AfL and AoL?
    Table 12.2 presents a selection of key issues in relation to assessment now and in the
                                                                                         a
    coming years. We present these not necessarily as predictions of the future but as
                                                                                                    (Continued)
206     Understanding and applying assessment in education
(Continued)
  summary of possible needs, trends and directions. Some have implications largely in
  relation to AfL, some to AoL and some to both. As the final activity in this book, tick
  the boxes as you see fit and reflect on what the implications might be. Further
  prompts available on the companion website may be helpful.
Chapter summary
In today’s education systems, helping hands abound, most notably those of the
teacher who, in partnership with parents, plays the central role in helping learners
to realise their potential. Therefore, the knowledge, disposition and skill set of the
teacher determines to a great extent the benefit that will accrue to students from their
attendance at school over the course of most of their first two decades of life. This
is a tremendous responsibility on teachers, shared with parents, and therefore it is
not surprising that codes of practice and expectations are increasingly the norm
internationally, both in education generally and in assessment specifically. Other
chapters in the book also reiterate that responsibility. This is evident in the planning,
development, implementation and use of a variety of assessments discussed in
Chapters 5, 6 and 7, following on from some context-setting about AoL and AfL in
Chapters 3 and 4. Though part of school teams generally, teachers enjoy considerable
autonomy in their interactions with students, so the competency of the teacher, in
assessment and other dimensions of practice, has a profound influence on students’
educational and, it can be argued, life chances. Teachers also need to be good com-
municators, in class with students and in how they mediate information about stu-
dents’ learning to students, parents and others. Chapters 8 and 9 touch on this theme,
exploring how information derived from assessments can and should be shared.
                                 Conclusions about assessment      in education      207
   Running alongside and through these chapters is also the recognition of those
other helping hands supporting the learner. Though district officials, national plan-
ners and researchers don’t stand in class, their responsibilities are ultimately to the
student and students as a group. They cannot offer help without accurate informa-
tion. Funding needs to be prioritised, curricula and pedagogy need to be relevant
and updated from time to time, good practice needs to be identified and dissemi-
nated, and challenges that impede learning need to be addressed. Assessment pro-
vides information to aid such decisions, decisions that ultimately impact on groups
and on individual students. Therefore, the needs of such stakeholders are legitimate
and must be incorporated into principals’ and teachers’ planning and practice. Such
issues are picked up throughout many of the chapters in the book, especially though
not exclusively in relation to summative    assessment, as in Chapters 1, 3, 8 and 11.
Education is a large-scale complex process that increasingly requires collaboration
between highly qualified individuals within and outside the teaching profession.
What they share is the desire to further the best interests of the student, though they
may come at this from different perspectives and using different approaches. For the
individual student and their family, the stakes are high. It is the responsibility of the
professional team — teachers, principals and educational support staff in school;
along with curriculum developers, teacher educators, CPD providers, educational
policy-makers, officials and the research community — to collaborate in the best inter-
ests of the student. Appropriate use of assessment represents an opportunity to have
a real bearing on how successful that collaboration will be on behalf of the learner.
It is an opportunity not to be missed.
Abedi, J. (2010). Performance assessments for English language learners. Stanford, CA: Stanford
   University, Stanford Center for Opportunity Policy in Education. Available at: https://scale.
   stanford.edu/system/files/performance-assessments-english-language-learners.    pdf
Alderson, J.C., Clapham, C., & Wall, D. (1995). Language test construction and evaluation.
   Cambridge: Cambridge University Press.
American Educational Research Association (AERA), American Psychological Association
   (APA), & National Council on Measurement in Education (NCME). (2014). Standards for
   educational and psychological testing. Washington, DC: AERA.
American Federation of Teachers (AFT), National Council on Measurement in Education
   (NCME), & National Education Association (NEA). (1990). Standards for teacher competence
   in educational assessment of students. Available at: http://buros.org/standards-teacher-
   competence-educational-assessment-students
Anderson, L.W., & Krathwohl, D.R. (eds). (2001). A taxonomy for learning, teaching and
   assessing: A revision of Bloom’s taxonomy of educational objectives. New York: Longman.
Arffman, I. (2013). Problems and issues in translating international educational achievement
   tests. Educational Measurement: Issues and Practice, 32 (2), 2-14.
Armitage, A., Evershed, J., Hayes, D., Hudson, A., Kent, J., Lawes, S., et al. (2012). Teaching
   and training in lifelong learning (4th edition). Maidenhead: Open University Press.
ASG South Ayrshire. (n.d.). Self- and peer assessment: Using rubrics for science and
   mentors of art. Available at: www.educationscotland.gov.uk/resources/practice/s/self
   andpeerassessment/selfandpeerassessmentrubrics.asp?strReferringChannel=education
   scotland&strReferringPageID=tcm:4-615801-64
Assessment and Qualifications Alliance (AQA). (2016). Unit 2: Speaking and listening. GSCE
   English language. Available at: www.aqa.org.uk/subjects/english/gcse/english-language-4705/
   subject-content/unit-2
Assessment Reform Group, The. (2006). The role of teachers in the assessment of learning.
   Cambridge: Cambridge University Faculty of Education.     ~
Association of Educational Assessment.(AEA) — Europe. (2012). European framework of stand-
   ards for educational assessment 1.0. Rome: Edizioni Nuova Cultura. Available at: www.           —
   aea-europe.net/index. php/standards-for-educational-assessment
                                                                                       References          209
Australian Curriculum, Assessment and Reporting Authority (ACARA). (2014). Australian cur-
   riculum: Work sample portfolio summary — Year 5 (Above satisfactory). Available at: www.
   acara.edu.au/curriculum/worksamples/Year_5_English_Portfolio_Above.pdf
Australian Institute for Teaching and School Leadership (AITSL). (2014). Australian profes-
   sional standards for teachers. Melbourne: AITSL. Available at: www.aitsl.edu.au/
    australian-professional-standards-for-teachers
Ausubel, D.P. (1968). Educational psychology: A cognitive view. New York: Holt, Rinehart and
   Winston.
Baas, D., Castelijns, J., Vermeulen,        M., Martens,     R., & Segers, M. (2015). The relationship
   between assessment for learning and elementary students’ cognitive and metacognitive
   strategy use. British Journal of Educational Psychology, 85, 33-46.
Baird, J., & Black, P. (2013). The reliability of public examinations. Research Papers in
   Education, 28 (1), 1-4.
Baird, J., Hopfenbeck, T.N., Elwood, J., Caro, D., & Ahmed, A. (2015). Predictability in the Irish
   Leaving Certificate. Oxford and Belfast: University of Oxford and Queen’s University Belfast.
Bandura, A. (1997). Self-efficacy: The exercise of control. New York: Freeman.
Bangert-Drowns, R.L., Kulik, C.C., Kulik, J.A., & Morgan, M-T. (1991). The instructional effect
   of feedback in test-like events. Review of Educational Research, 61 (2), 213-38.
BBC News. (2001, 19 June). Should the exam have been cancelled? Available at: http://news.
   bbc.co.uk/2/hi/talking_point/1388224.stm
Belfast Telegraph. (2012, 5 January). Pupil evaluation test scrapped over scoring blunders (by
    Lindsay Fergus). Available at: www.belfasttelegraph.co.uk/news/education/pupil-evaluation-
    test-scrapped-over-scoring-blunders-28699343.html
 Bennett, R.E. (2011). Formative assessment: a critical review. Assessment in Education: Principles,
    Policy and Practice, 18 (1), 5-25.
 Berry, R. (2011). Assessment trends in Hong Kong: seeking to establish formative assessment
                                                                               Practice, 18 (2),
    in an examination culture. Assessment in Education: Principles, Policy and
    199-211.
                                                                                    and testing.
 Black, P. (1998). Testing: Friend or foe? In The theory and practice of assessment
    London: The Falmer Press.
 Black; P., ‘Harrison,   C., Lee, C., Marshal,      B., & Wiliam, D. (2003). Assessment for learning:
    Putting it into practice. Maidenhead: Open University Press.
                                                                              t in Education:
 Black, P., & Wiliam, D. (1998). Assessment and classroom learning. Assessmen
    Principles, Policy and Practice, 5 (A), 7-74:
                                                                    assessment. Educational
 Black, P., & Wiliam, D. (2009). Developing the theory of formative
    Assessment, Evaluation and   Accountab ility, 21 (1), 5-31.
                                                                        (1956). Taxonomy of
 Bloom, B., Englehart, M.D., Furst, EJ., Hill, W.H., & Krathwohl, D.R.
                                                                                                 k I: Cognitive
    educational objectives: The classification of educational goals. Handboo
    domain. White Plains, NY: Longman.
                                                                                k on formative and summa-
 Bloom, B.S., Hastings, J.T., & Madaus, G.F. (eds.) (1971). Handboo
    tive evaluation   of student  learning .  New    York:   McGraw-  Hill     Book   Company.
          ar, S. (2012).  The policy  impact   of PISA:  An    explorat ion  of the normative effects of inter-
 Breakspe
                                                                                            n working papers,
    national benchmarking in school system performance. OECD educatio
                                               Available   at:  http://d x.doi.or  g/10.178  7/Sk9fdfqffr28-en
    number 71. Paris: OECD Publishing.
                            Experie nces    from   large-sc ale    compute  r-based     testing   in the USA. In
 Bridgeman, B. (2009).
                                                                                    r-based     assessme nt: New
    F Scheuermann & J. Bjornsson (eds) The transition to compute
                                                                     for  large-sc  ale   testing   (pp.   39-44).
    approaches to skills assessment and implications
                                                   ions   of   the  Europea  n   Communi  ties.
    Luxembourg: Office for Official Publicat
                                                     LD: determining eligibility, selecting ser-
 Brigham, FJ., & Bakken, J.P. (2013). Assessment and                                          g
                                                                 & A.F. Rotatori (eds) Learnin
    vices, and guiding instruction. In J.P. Bakken, FE. Obiakor
                                                              of student s with LD (pp. 55-74).
    disabilities: Identification, assessment, and instruction
    Bingley, UK: Emerald Publishing.
                  ‘
210     Understanding and applying assessment in education
Brookhart, S.M. (2007). Feedback that fits. Educational Leadership, 65 (4), 54-9.
Brookhart, $.M. (2011). Educational assessment knowledge and skills for teachers. Educational
   Measurement: Issues and Practice, 30 (A), 3-12.
Brophy, J.E., & Good, T.L. (1986). Teacher behaviour and student achievement. In M.C. Wittrock
   (ed.) Handbook of research on teaching Grd edition, pp. 328-75). New York: Macmillan.
Brown, G.T.L., & Harris, L.R. (2013). Student self-assessment. In J.H. McMillan (ed.) Sage hand-
   book of research on classroom assessment (pp. 367-93). Thousand Oaks, CA: Sage.
Buck, S., Ritter, G.W., Jensen, N.C., & Rose, C.P. (2010). Teachers     say the most   interesting
   things: an alternative view of testing. Phi Delta Kappan, 91 (6), 50-4.
Cambridge Assessment (2015). Cambridge handbook 2015 (international): Regulations for
   conducting Cambridge examinations. Cambridge: Author.
Council for the Curriculum Examinations and Assessment (CCEA). (2009). Assessment for
   learning: A practical guide. Belfast: CCEA.
CCEA. (2016a). Skills and capabilities at Key Stages 1-2. Belfast: CCEA. Available at: http://
   ccea.org.uk/curriculum/key_stage_1_2/skills_and_capabilities ,
CCEA. (2016b). CCEA GCSE specification in business studies. Belfast: CCEA.
CCEA. (n.d.). Statutory requirements for language and literacy at Key Stage 2. Available at:
   http://ccea.org.uk/curriculum/key_stage_1_2/areas_learning/language_and_literacy
Central Statistics Office (CSO). (2015). Foreign nationals: PPSN allocations, employment and
   social welfare activity. Dublin: CSO. Available at: www.cso.ie/en/releasesandpublications/
   er/fnaes/foreignnationalsppsnallocationsemployment andsocialwelfareactivity2014/
Clarke, S. (2005). Formative assessment in action: Weaving the elements together. London:
   Hodder Murray.
Clarke, S. (2008). Active learning through formative assessment. London: Hodder Murray.
Coe, R. (2010). Understanding comparability of examination standards. Research Papers in
   Education, 25 (3), 271-84.
Cohen, L., Manion, L., & Morrison, K. (2004). A guide to teaching practice (Sth edition).
    London: Routledge.                                         ‘
Collins, S., Reiss, M., & Stobart, G. (2010). What happens when high-stakes testing stops?
   Teachers’ perceptions of the impact of compulsory national testing in science of 11-year-
    olds in England and its abolition in Wales. Assessment in Education: Principles, Policy and
   Practice, 17 (3), 273-86.
Costa, A.L., & Kallick, B. (2008). Learning and leading with habits of mind: 16 characteristics for
   success. Alexandria, VA: Association for Supervision and Curriculum Development (ASCD).
Council of Europe. (2012). European common framework of reference for languages.
    Strasbourg: Council of Europe, Language Policy Unit. Available at: www.education-support.
    org.uk/teachers/ids/ceft/
Crooks, T. (2011). Assessment for learning in the accountability era: New Zealand. Studies in
   Educational Evaluation, 37 (1), 71-7.
Crookes, A. (2007). The effects of topic choice on performance outcomes: an analysis of stu-
    dent selected third year essays. Psychology learning & teaching, 6 (2), 85-90.
Cumming, J.J., & Maxwell, G.S. (2014). Expanding approaches to summative assessment for
    students with impairment. In L. Florian (ed.) The Sage handbook of special education, Vol. 2
   (2nd edition, pp. 573-93). Los Angeles, CA: Sage.
Cummins, J. (1991). Interdependence of first- and second-language proficiency in bilingual
   children. In E. Bialystok (ed.) Language processing in bilingual children (pp. 70-89).
   Cambridge: Cambridge University Press.
Cummins, J. (2000). Language, power and pedagogy: Bilingual children in the crossfire.
   Clevedon, UK: Multilingual Matters.                     :
Daily Telegraph, The. (2015, 24 July). More than 300 languages spoken in British schools,
   report says.   Available   at: www.telegraph.co.uk/education/education        news/11761250/
   More-than-300-different-languages-spoken-in-British-schools-report-says. html
                                                                                 References         211
Dann, R. (2014). Assessment as learning: blurring the boundaries of assessment and learning
   for theory, policy and practice. Assessment in Education: Principles, Policy and Practice, 21
                                                                                      a
   (2), 149-66.
Darling-Hammond, L., & Adamson, F. (2010). Beyond basic skills:         The role of performance
   assessment in achieving 21st century standards of learning. Stanford, CA: Stanford
   University, Standford Centre for Opportunity Policy in Education.
Darling-Hammond, L., & McCloskey, L. (2008). What would it mean to be internationally com-
   petitive? Phi Delta Kappan, 90 (4), 263-72.
Davidson, C.N., & Goldberg, D.T. (2010). The future of thinking: Learning institutions in a
   digital age. The John D. and Catherine T. MacArthur Foundation Reports on Digital Media
   and Learning. Cambridge, MA: The MIT Press.
Dede, C. (2010). Comparing frameworks for 21st century skills. In J. Bellanca & R. Brandt (eds)
   21st century skills: Rethinking bow students learn (pp. 51-75). Bloomington, IN: Solution
   Tree Press.
Department for Education (DfE). (2011a). Teachers’ standards. London: Crown. Available at:
   www. gov.uk/government/publications/teachers-standard
Department for Education (DfE). (2011b). Teachers’ standards: Guidance for school leaders,
   school staff and governing bodies. Available at: www.gov.uk/government/publications/
   teachers-standards
Department for Education (DfE). (2012). School information (England) (Amendment)
   Regulations 2012. Available at: www .legislation.gov.uk/uksi/2012/1124/pdfs/uksi_2012
   1124_en.pdf
Department for Education (DfE). (2014a). The national curriculum in England: Framework
   document. London: Crown. Available at: www.gov.uk/government/uploads/system/
    uploads/ attachment_data/file/381344/Master_final_national_curriculum_28_Nov. pdf
 Department for Education (DfE). (2014b). Reforming assessment and accountability for pri-
    mary schools: Government response to consultation on primary school assessment and
    accountability. London: DfE. Available at: www.gov.uk/government/uploads/system/
    uploads/attachment_data/file/297595/Primary_Accountability_and_Assessment_
    Consultation_Response.pdf
                                                                                   code of
 Department for Education (DfE). (2015). Special educational needs and disability
                                                                        uk/gover  nment/
    practice: O to 25 years. London: Crown. Available at: www.gov.
    publications
                                                                                   ts at key stage 2 in
 Department for Education (Df). (2016). National curriculum assessmen
    England, 2016     (provision al). London:    Crown.   Available    at: www.gov.      uk/government/
    uploads/system/uploads/ attachment_data/file/549432/SFR39_2016_t          ext.pdf
                                                                                        strategy: Primary
 Department for Education and Skills (DfES). (2006). Primary national
                                            ics.    London:    DfES.   Available     at:  www.education
   framework for literacy and mathemat
    england.org.uk/documents    /pdfs/ 2006-pri mary-nat ional-st rategy.p df
                                                                                      with special educa-
 Department of Education and Skills (DES). (2007a). Inclusion of students
    tional needs: Post primary guidelines. Dublin: DES.
                                                                             nal needs: A continuum
 Department of Education and Skills (DES). (2007b). Special educatio
    of support. Dublin: DES.
                                                                               Initial steps in the imple-
 Department of Education and Skills (DES). (2011a). Circular 0056/11:
                                                                     Availabl e   at: www.education.ie/
    mentation of the national literacy and numeracy strategy.
    en/Circulars-and-Forms/Active-Circulars/   cl0056_2  011.pdf
                                                                     and numeracy for learning and
 Department of Education and Skills (DES). (2011b). Literacy
                                                                     cy among children and young
    life: The national strategy to improve literacy and numera
    people 2011-2020. Dublin: DES.
                                                                    for Junior Cycle 2015. Dublin:
 Department of Education and Skills (DES). (2015a). Framework
                                      ie/en/P ublicat ions/Po licy-Re ports/Framework-for-Junior-
    DES. Available at: www.education.
    Cycle-2015.pdf
                 ii
212     Understanding and applying assessment in education
Department of Education and Skills (DES). (2015b). Circular 0027/2015: Information in rela-
   tion to actions under the literacy and numeracy strategy standardised testing, reporting,
   library support and other matters. Available at: http://education.ie/en/ Circulars-and-Forms/
   Active-Circulars/cl0027_2015.pdf
Department of Education and Skills (DES). (2015c). Digital strategy for schools 2015-2020:
   Enhancing teaching, learning and assessment. Dublin: DES. Available at: www.education.
   ie/en/Publications/Policy-Reports/Digital-Strategy-for-Schools-2015-2020.pdf
Department of Education and Skills (DES). (2015d). Key statistics 2014/2015. Available at:
   www.education.ie/en/Publications/Statistics/Key-Statistics/Key-Statistics-2014-2015.pdf
Department of Education and Skills (DES). (2015e). Primary language curriculum: English-
   medium schools. Dublin: DES & NCCA.
Department of Education and Skills (DES). (2016a). Arrangements for the implementation of
   the framework for Junior Cycle, with particular reference to school years 2015/16 and
   2016/17 (Circular Letter 0024/2016). Dublin: DES. Available at: www.education.ie/en/
   Circulars-and-Forms/Active-Circulars/cl0024_2016.pdf             1
Department of Education and Skills (DES). (2016b). Looking at our school 20106. A quality
  framework for primary schools. Dublin: DES.
Downing, S.M. (2006). Selected-response item formats in test development. In S.M. Downing
   & T.M. Haladyna (eds) Handbook of test development (pp. 287-301). New York: Routledge.
Du Bois, P.H. (1970). A history of psychological testing. Boston: Allyn & Bacon.
Dunn, K.E., & Mulvenon, S.W. (2009). A critical review of research on formative assessment:
    the limited scientific evidence of the impact of formative assessment in education. Practical
    Assessment, Research & Evaluation, 14 (7). Available at: http://pareonline.net/getvn.
    asp?v=14&n=7
Earl, L.M. (2013). Assessment as learning: Using classroom assessment to maximise student
    learning (2nd edition). Thousand Oaks, CA: Corwin.
Ebel, R.L. (1980). Practical problems in educational measurement. Lexington, MA: D.C. Heath
   and Company.                                             .
Education Council New Zealand. (2015). Practising teacher criteria. Available at: https://
   educationcouncil.org.nz/content/review-of-standards-teaching-profession
Education Scotland (2015). Literacy across learning in secondary schools 2015-16. Reading.
   Available at: http://dera.ioe.ac.uk/25630/1/PDFReadingwithAppendices_tem4-874285.pdf
Education Scotland (2016). Benchmarks: Literacy and English, Draft, August 2016. Available at:
   https://education. gov.scot/improvement/Pages/Curriculum-for-Excellence-Benchmarks-.aspx
Educational Research Centre (ERC). (2007). Drumcondra Primary Mathematics Test — Revised —
   Levels 3—6. Manual. Dublin: ERC.
Educational Research Centre (ERC). (2010). Drumcondra Primary Mathematics Test — Revised.
   Progress in mathematics: Using class-level performance bands and monitoring progress over
   time. Supplement to the manuals for the DPRT-R, Levels 1, 2 and 3-6. Dublin: ERC. Available
   at: www.erc.ie/documents/dpmt_combined_documents.         pdf
Educational Research Centre (ERC). (2014a). Sample maths items: Sixth Class — NA ’14 released.
   Available at: www.erc.ie/documents/na14_sample_items_maths6th.pdf
Educational Research Centre (ERC). (2014b). Sample reading items: Sixth Class — NA ’14 released.
   Available at: www.erc.ie/documents/na14_sample_items_reading6th.pdf
Educational Testing Service (ETS). (2009a). ETS international principles for fairness review of.
   assessments: A manual for developing locally appropriate fairness review guidelines in vari-
   ous countries. Princeton, NJ: ETS. Available at: www.ets.org/s/about/pdf/fairness_review_
   international.pdf
Educational Testing Service (ETS). (2009b). Guidelines for the assessment of English language
   learners. Princeton, NJ: ETS. Available at: www.ets.org/s/about/pdf/ell _guidelines.pdf
Eggen, TJ.H.M., & Lampe, T.T.M. (2011). Comparison of the reliability of scoring methods of
   multiple-response items, matching items and sequencing items. CADMO, 19 (2), 85-104.
                                                                                     References         213
 Elwood, J. (2013). Educational assessment policy and practice: a matter of ethics. Assessment
    in Education: Principles, Policy and Practice, 20 (2), 205-20.
 Elwood, J., Hopfenbeck, T., & Baird, J. (2015). Predictability in high stakes examinations: stu-
    dents’ perspectives on a perennial assessment, dilemma. Research Papers in Education,
    DOI: 10.1080/02671522.2015.1086015
 Encyclopaedia Britannica. (n.d.). Chinese examination system. Available at: www.britannica.
    com/topic/Chinese-examination-system
 European Union. (2006). Recommendation of the European Parliament and the Council of 18
    December 2006 on Key Competences for Lifelong Learning. Official Journal of the
    European Union 31.12.2006. Available at: http://eur-lex.europa.eu/legal-content/EN/TXT, /?
     uri=celex:32006H0962
  Eve, M.P. (2011, 22 December). Secondary schools are not adequately preparing students for
     higher education. Guardian Higher Education Network Blog. Available at: www.theguard-
      ian.com/higher-education-network/blog/2011/dec/22/ humanities-in-secondary-schools
_ Expert Advisory Group on Certificate Examinations. (2000). Arrangements for the assessment
      of candidates with special educational needs in certificate examinations: Report to the
     Minister for Education and Science. Dublin: Author. Available at: http://examinations.ie/
      schools/Expert_Advisory_Group_Report_ev.pdf
  Faragher, S. (2014). Understanding assessment in primary education. London: Sage.
  Florez, M.T., & Sammons, P. (2013). Assessment for learning: Effects and impacts. Reading, UK:
      CfBT Education Trust.
  Frey, B.B., Schmitt, V.L., & Allen, J.P. (2012). Defining authentic classroom assessment.
     Practical Assessment, Research and Evaluation, 17 (2), 1-18.
  Geisinger, K.F. (2016). Intended and unintended meanings of validity: some clarifying com-
      ments. Assessment in Education: Principles, Policy and Practice, 23 (2), 287-9.
  General Teaching Council (GTC) for Northern Ireland. (2011). Teaching: The reflective profes-
      sion. Belfast: GTC.
  General Teaching Council (GTC) for Scotland. (2012). The standards for registration:
      Mandatory requirements for registration with the general teaching council for Scotland.
      Edinburgh: GTC. Available at: www.gtcs.org.uk/web/files/the-standards/standards-for-
      registration-1212.pdf
                                                                                           Phi Delta
  Gillie, O. (1977). Did Sir Cyril Burt fake his research on heritability of intelligence?
     Kappan, 58 (6), 409-71.
                                                                                                     . London:
  Gipps, C.V. (1994). Beyond testing: Towards a theory of educational assessment
     Falmer.
                                                                                           th and J. Cumming
  Gipps, C.V., & Stobart, G. (2009). Fairness in assessment. In C. Wyatt-Smi
                                                                      105-19).    Dordrecht  : Springer.
     (eds) Educational assessment in the 21st century (pp.
                                                   assessmen  t.    Internati   onal   Journal   of Educational
  Glaser, R. (1990). Toward new models for
     Research, 14 (5), 475-83.
                                                                                       some issues arising from
  Goldstein, H. (2004). International comparisons of student attainment:
                                                                            and   Practice,   11 3), 319-30.
      the PISA study. Assessment in Education: Principles, Policy
                                                  act.  Dublin:     Statione  ry  Office.
  Government Of Ireland. (1998). Education
                                                                                            nal needs act 2004.
  Government of Ireland. (2004). Education for persons with special educatio
                        y  Office.                                                   .
      Dublin: Stationer
                                                                             ry Office.
  Government of Ireland. (2005). Disability act. Dublin: Statione
                                                           In  P.  Griffin   & E. Care (eds) Assessment and
  Griffin, P., & Care, E. (2015). The ATC21S method.
      teaching of 21st century skills: Methods    and   approac   h   (pp.   3-33). Dordrecht: Springer.
                                Handbo ok   of psychol ogical    assessm   ent   (Sth edition). Hoboken, NJ:
  Groth-Marnat, G. (2009).
      Wiley & Sons.
                                                                                   declares thousands of test
  Guardian, The. (2014, 24 June). English language tests inquiry
                                                    e  at: www.th     eguard   ian.co   m/uk-news/2014/jun/
      results invalid (by Alan Travis). Availabl
                                    eating- results -invali d-    oversea   s-stude  nts
      24/english-language-tests-ch
                   ¢
214      Understanding and applying assessment in education
Guardian, The. (2015, 20 August). Take it from an examiner, your students’ exam results could
    easily be wrong (anonymous). Available at: www.theguardian.com/teacher-network/2015/
    aug/13/a-levels-gcses-examiner-exam-results-wrong
Guskey, T. (2006). Making high school grades meaningful. Phi Delta Kappan, 87 (9), 670-5.
Haladyna, T.M. (2004). Developing and validating multiple-choice test items. Mahwah, NJ:
    Lawrence Erlbaum Associates.
Haladyna, T.M., & Rodriguez, M.C. (2013). Developing and validating test items. Abingdon,
    Oxon: Routledge.
Harlen, W. (2005). Trusting teachers’ judgment: research evidence of the reliability and validity
    of teachers’ assessment used for summative purposes. Research Papers in Education, 20
    G)e2Z97-313:.
Harlen, W. (2007). Assessment of learning. London: Sage.
Harris, A. (1998). Effective teaching: a review of the literature. School Leadership and
   Management, 18 (2), 169-83.
Hattie, J. (2009). Visible learning: A synthesis of over 800 meta-analyses relating to achieve-
    ment. Abingdon, Oxon: Routledge.
Hattie, J. (2012). Visible learning for teachers: Maximising impact on learning. Abingdon,
    Oxon: Routledge.
Hattie, J., & Timperley, H. (2007). The power of feedback. Review of Educational Research, 77
   (1), 81-112.
Hayward, L. (2015). Assessment is learning: the preposition vanishes. Assessment in Education:
   Principles, Policy and Practice, 22 (1), 27-43.
Heacox, D. (2002). Differentiating instruction in the regular classroom: How to reach and
   teach all learners, Grades 3-12. Minneapolis, MN: Free Spirit Publishing.
Heacox, D. (2012). Differentiating instruction in the regular classroom: How to reach and
   teach all learners. Updated anniversary edition. Minneapolis, MN: Free Spirit Publishing.
Heneghan, H., & Murchan, D. (2014). Enhancing teachers’ capacity to implement effective dif-
  ferentiated reading in Irish primary schools. Paper presented at the annual meeting of the
   Educational Studies Association of Ireland, Athlone, April.
Henk, W.A. (1981). Effects of modified deletion strategies and scoring procedures on cloze
   test performance. Journal of Reading Behavior, 13 (4), 347-357.
HM Inspectorate of Education (HMIE). (2007). How good is our school? Livingston: HMIE.
   Available at: www.educationscotland.gov.uk/Images/HowgoodisourschoolJtEpart3_tcem4-
   684258. pdf
House of Commons Children, Schools and Families Committee. (2008). Testing and assess-
   ment: Government and Ofsted responses to the Committee’s Third Report of Session 2007-08.
   Fifth special report of session 2007-08. London: The Stationery Office. Available at: www.
   publications.parliament.uk/pa/cm200708/cmselect/cmchilsch/1003/1003.pdf
Husbands, C. (1996). What is history teaching? Buckingham: Open University Press.
Iamarino, D.L. (2014). The benefits of standards-based grading: a critical evaluation of modern
   grading practices. Current Issues in Education, 17 (2), 1-11.
Integrate Ireland Language and Training (IILT). (2007a). EAL primary schools assessment kit.
   Dublin: NCCA. Available at: www.ncca.ie/en/Curriculum_and_Assessment/Inclusion/
   English_as_an_Additional_Language/IILT_Materials/Primary/
Integrate Ireland Language and Training CILT). (2007b). EAL post-primary assessment kit.
   Dublin:      NCCA.   Available   at:   www.ncca.ie/en/Curriculum_and_Assessment/Inclusion/
   English_as_an_Additional_Language/IILT_Materials/Post-primary/EAL_Post_Primary_
   Assessment_Kit/EAL_Post-Primary_Assessment_Kit.html
International     Reading   Association   and   National   Council   for Teachers   of English   (IRA/
   NCTE). (2013). Oral presentation rubric: Grades 3-12. Newark, DE: IRA. Available at:
   www .readwritethink.org/classroom-resources/printouts/oral-presentation-rubric-30700.
   html
                                                                                       References         215
 Irish Times. (2010, 11 June). Leaving Cert dyslexia case dismissed (by Mary Carolan).
     Available at: www. irishtimes.com/news/leaving-cert-dyslexia-case-dismiss     1.859452
                                                                                       ed-
 Irish Times. (2010, 19 June). The year Leaving Cert cheating stopped being a secret (by Louise
     Holden). Available at: www.irishtimes.com/news/education/the-year-leaving-certificate-
     cheating-stopped-being-a-secret-1.680700
 Irish Times (2015, 24 February). Census figures raise concerns of ethnic segregation in schools
     (by Pamela Duncan and Joe Humphreys). Available at: www. irishtimes.com/news/
     education/census-figures-raise-concerns-of-ethnic-segregation-in-schools-1.2114559
  Isaacs, T., Zara, C., & Herbert, G., with Coombs, S.J., & Smith, C. (2013). Key concepts in edu-
     cational assessment. London: Sage.
  Jennings, J.M., & Bearak, J.M. (2014). Teaching to the test in the NCLB era: how test predict-
     ability affects our understanding of student performance. Educational Researcher, 52,
     423-62.
  Johnson, R.B., & Onwuegbuzie, AJ. (2004). Mixed methods research: a research paradigm
      whose time has come. Educational Researcher, 33, 14-20.
" Johnson, S. (2013). On the reliability of high-stakes teacher assessment. Research Papers in
      Education, 28 (1), 91-105.
  Johnston, K., Conneely, C., Murchan, D., & Tangney, B. (2015). Enacting key skills-based cur-
      ticula in secondary education: lessons from a technology-mediated, group-based learning
      initiative. Technology, Pedagogy and Education, 24 (A), 423-42.
  Joint Council for Qualifications. (2016). Adjustments for candidates with disabilities and learn-
      ing difficulties. Access arrangements and reasonable adjustments. 1 September 2016 to 31
                                                                                               ge-
      August 2017. London: Author. Available at: www.jcq.org.uk/exams-office/access-arran
      ments-and-special-consideration/regulations-a         nd-guidan ce/access  -arrangem  ents-
      and-reasonable-adjustments-2016-2017
                                                                                 Available at:
  Joint Council for Qualifications. (Various Years). Examination results: GCSEs.
      www ,jcq.org.uk/examination-results/gcse
                                                                                 Boston, MA:
  Joyce, B.R., Weil, M., & Calhoun, E. (2000). Models of teaching (6th edition).
      Pearson.
                                                                                          Erlbaum Associates.
   Kamin, LJ. (1974). The science and politics of IQ. Potomac, MD: Lawrence
                                                                              Logic   models:  A tool for effec-
   Kekahio, W., Cicchinelli, L., Lawton, B., & Brandon, P.R. (2014).
                                                         monitori  ng  (REL    2014-025)  .  Washingt on, DC:
      tive program planning, collaboration, and
                                   ,   Institute  of  Education   Sciences,  National    Center  for Education
      US Department of Education
                                                                                       y Pacific. Available at:
      Evaluation and Regional Assistance, Regional Educational Laborator
      http://ies.ed.gov/ncee/edlabs
                                                                                 Battle Creek, MI: Author.
   Kellogg Foundation. (2004). Logic model development guide.
                                      esource   -direct  ory/res  ource/2 006/02/    wk-kellogg-foundation-
      Available at: www.wkkf.org/r
      logic-model-development-guide
                                                                              highly effective teaching and
   Kentucky Department of Education. (2015). Characteristics of
                                                           gov/cur  riculum  /standa   rds/teachtools/Pages/
      learning. Available at: http://education.ky.
                                          ive-Teac   hing-and -Learnin g-(CHETL   ).aspx          .
      Characteristics-of-Highly-Effect
                           & Yovanoff ,  P. (2009).   Diagnos tic  assessm  ent   in  mathema tics to support
   Ketterlin-Geller, L.R.,                                                                             14 (16),
      instructional       decision   making.   Practical Research, Assessment and Evaluation,
       1-9. Available at: http://pareonline.net/getvn.asp?v=14&n=16
                                                          a meta-analysis and a call for research.
   Kingston, N., & Nash, B. (2011). Formative assessment:
                                                              28-37.
      Educational Measurement: Issues and Practice, 30 (4),                                    ent
                                                                 A longitudinal study. Assessm
   Klapp, A. (2015). Does grading affect educational attainment?
                                                         (3), 302-23.
       in Education: Principles, Policy and Practice, 22
                                                             education: Standards, judgement and
   Klenowski, V., & Wyatt-Smith, C. (2014). Assessment for
      moderation. London: Sage.
                                                               g: A review of research and evidence.
   Ko, J., Sammons, P., & Bakkum, L. (2013). Effective teachin
      Reading, UK: CfBT Education Trust.
                      ‘
216     Understanding and applying assessment in education
Koretz, D. (2008). Measuring up: What educational testing really tells us. Cambridge, MA:
    Harvard University Press.
Koretz, D., Stecher, B., Klein, S., & McCaffrey, D. (1994a). The Vermont portfolio assessment
    program: findings and implications. Educational Measurement: Issues and Practice, 13 (3),
    5-106.
Koretz, D., Stecher, B., Klein, S., & McCaffrey, D. (1994b). The evolution of a portfolio program:
    The impact and quality of the Vermont program in its second year (1992-93). CSE Technical
   ‘Report No. 385. Los Angeles, CA: University of California. Available at: http://eric.ed.
    gov/?id=ED379301
Krathwohl, D.R. (2002). A revision of Bloom’s taxonomy:        an overview.   Theory into Practice,
   41 (4), 212-18.
Kreiter, C.D., Haugen, T., Leaven, T., Goerdt, C., Rosenthal, N., McGaghie, W.C., & Dee, F.
   (2011). A report on the piloting of a novel computer-based medical case simulation for
   teaching and formative assessment of diagnostic laboratory testing. Medical Education
   Online, 16. Available at: http://doi.org/10.3402/meo.v16i0.5646
Kubiszyn, T., & Borich, G. (2013). Educational testing and measurement (10th edition).
   Hoboken, NJ: Wiley & Sons.
Kulhavy, R.W., & Stock, W. (1989). Feedback in written instruction: the place of response certitude.
   Educational Psychology Review, 1 (4), 279-308.
Kyriacou, C. (2014). Essential teaching skills (4th edition). Oxford: Oxford University Press.
Lewis, H. (2012). Written statement: PISA 2012. Updated 3 December 2013. Welsh Government.
   Available at: http://gov.wales/about/cabinet/cabinetstatements/2013/pisa2012/?lang=en
Lukhele, R., Thissen, D., & Wainer, H. (1994). On the relative value of multiple-choice, con-
   structed response, and examinee-selected items on two achievement tests. Journal of
   Educational Measurement, 31 (3), 234-50.
Lynch, K., Grummell, B., & Devine, D. (2012). New managerialism in education:
   Commercialization, carelessness and gender. Basingstoke, UK: Palgrave Macmillan.
Lysaght, Z. (2012). Towards inclusive assessment. In T. Day & J. Travers (eds) Special and inclu-
   sive education: A research perspective (pp. 245-60). Oxford: Peter Lang.
Mac Aogain, E. (2005). The points system and grading of the Leaving Certificate examination.
   Irish Journal of Education, 36, 3-24.
McCormack, O., Lynch, R., & Hennessy, J. (2015). Plastic people in pinstripe suits: an explora-
   tion of the views of Irish parents on the publication of school league tables. Educational
   Studies, 41 (5), 513-33.
McMillan, J.H. (2011). Classroom assessment: Principles and practice for effective standards-
   based instruction (5th edition). Boston, MA: Pearson.
MacNeela, P., & Boland, J. (2013) Leaving Certificate teachers’ grading and marking practices:
   Report on the online consultation. Dublin: NCCA.
Marzano, RJ. (2009). Setting the record straight on ‘high-yield’ strategies. Phi Delta Kappan,
   91), 30-7.
Messick, S. (1989). Validity. In R. Linn (ed.) Educational measurement (3rd edition,
   pp. 13-100). Washington, DC: American Council on Education.
Meyen, E.L., & Skrtic, T.M. (eds) (1995). Special education and student disability: An introduc-
    tion (Ath edition). Denver, CO: Love Publishing Co.
Miller, M.D., Linn, R.L., & Gronlund, N. (2013). Measurement and assessment in teaching
    (11th edition/Pearson International Edition). Upper Saddle River, NJ: Pearson.
Morris, A. (2011). Student standardised testing: Current practices in OECD countries and
    a literature review. OECD Education Working Papers, No. 65. Paris: OECD Publishing.
    Available at: http://dx.doi.org/10.1787/5kg3rp 9qbnr6-en -
Mortimore, P. (1993). School effectiveness and the management of effective teaching and learning.
    Paper presented at the annual meeting of the international congress for school effectiveness
   and improvement. Norrkoping, Sweden: Available as: ERIC document ED358560
                                                                               References        217
Mullis, 1.V.S., Martin, M.O., Foy, P., & Hooper, M. (2016). TIMSS 2015 international results in
    mathematics. Chestnut Hill, MA: International Association for the Evaluation of Educational
    Achievement. Available at: http://timssandpirls.bc.edu/timss2015/international-results/wp-
   content/uploads/filebase/full%20pdfs/T15-Interniational-Results-in-Mathematics.pdf
Munoz, M.A., & Guskey, T.R. (2015). Standards-based grading and reporting will improve
   education. Phi Delta Kappan, 96 (7), 64-8.
Murchan, D. (1993). The reliability and validity of essay-based examinations in secondary
   education. Irish Educational Studies, 12, 41-50.
Murchan, D. (2011). The formative use of diagnostic assessments to support teaching and learn-
    ing in mathematics. Proceedings of the international conference Assessment for Learning
    in the 21st Century (pp. 147-56). Ohrid, Macedonia, 23 May. Skopje: USAID Macedonia.
                                                                                                    In
Murchan, D. (2014). Changing curriculum and assessment mindsets in higher education.
                                      (eds) Higher  education  in  Ireland:  Practices, policies  and
    A. Loxley, A. Seery & J. Walsh
   possibilities (pp. 186-97). Basingstoke, UK: Palgrave Macmillan.
                                                                                             selected
 Murchan, D., Loxley, A., & Johnston, K. (2009). Teacher learning and policy intention:
                                                               of professiona  l  developme nt  in the
    findings from an evaluation of a large-scale programme
    Republic of Ireland. Ewropean   Journal   of Teacher Education,   32  (4),  455-71.
                                                                         y to traditional error
Murchan, D., Oldham, E., & O’Sullivan, C. (2013). Applying CBA technolog
                                                                            presented at the
    analysis in elementary mathematics: Lessons from Maths Assist. Paper
                                                Research Associati on, San Francisco, April.
    annual meeting of the American Educational
                                                                   V. (2013). Vierésimi formative
 Murchan, D., Shiel, G., Vula, E., Gashi Bajgora, A., & Balidemaj,
                                                                              Available at: http://
    [Formative assessment guidebook]. Pristina: Basic Education Program.
                                                                                  f
    bep-ks.org/wp-content/uploads/2013/ 10/BEP-Vleresimi-formativ_shq.pd
                                                      s.  (1989). What   teachers should   know and be
 National Board for Professional Teaching Standard
                                      ts.org/ sites/d efault/ files/w hat_tea chers_s hould_k  now. pdf
    able to do. Available at: www.nbp
                                                                            Assessme nt in  the primary
 National Council for Curriculum and Assessment (NCCA). (2007a).
    school curriculum: Guidelines for schools. Dublin: NCCA.
                                                        (2007b). Exceptionally able students:
 National Council for Curriculum and Assessment (NCCA).
                                                                 www.ncca.ie/en/Publications/
    Draft guidelines for teachers. Dublin: NCCA. Available at:
                                                                    delines_for_Teachers.pdf
    Syllabuses_and_Guidelines/Exceptionally_Able_Students_Draft_Gui
                                                            (2014).  Key skills of Junior Cycle.
 National Council for Curriculum and Assessment (NCCA).
    Dublin: NCCA.
                                                          (2015). Junior Cycle English:
 National Council for Curriculum and Assessment (NCCA).
                                                        assessment tasks (1st edition).
    Guidelines for the classroom-based assessments and                                /
                                                      irculars-and-Forms/Active-Ci           rculars
    Dublin: NCCA. Available at: www.education.ie/en/C
    cl0024_2016.pdf
                                                        (NCCA). (2016). Junior Cycle science:
 National Council for Curriculum and Assessment
                                                               ent task (1st edition). Dublin:
    Guidelines for classroom-based assessments and assessm
                                                              153bc83f-9848-49f0-ad87-0a0d6b9
    NCCA. Available at: http://curriculumonline.ie/ getmedia/
    b596c/Specification-for-Jr-Cycle-Science-EV_20160126-().pdf
                                                         Annual report 2015. Trim, Ireland: NCSE.
 National Council for Special Education (NCSE). (2016).
                                                        (2013). Long-term planning: Draft guide-
 National Induction Programme for Teachers (NIPT).
                                                          ie/support/pdf/long_term_planning.pdf
    lines. Dublin: NIPT. Available at: www.nccaplanning
                                                    of   educat ional assessment. Assessment in
 Newton, PE. (2007). Clarifying the purposes
                                                    (2), 149-70.
     Education: Principles, Policy and Practice, 14
                                     Assess ment in Educat ion: Principles, Policy and Practice.
  Newton, P.E., & Baird, J. (2016).
     Special Issue on Validity. 23 2):
                                                  educational and psychological measurement.
  Newton, P.E., & Shaw, S.D. (2014). Validity in
     London: Sage.                                                                           ty’
                                                nt over the best way to use the word ‘validi
  Newton, P.E., & Shaw, S.D. (2016). Disagreeme                         Princi ples, Policy and
                                                  sment in Educat ion:
     and options for reaching consensus. Asses
     Practice, 23 (2), 178-97.
                 iy
218      Understanding and applying assessment in education
New Zealand Ministry of Education. (2011). Position paper: Assessment. Wellington: Learning
   Media.
Nitko, AJ., & Brookhart, S.M. (2014). Educational assessment of students (6th international
   electronic edition). Harlow: Pearson.
Northern Ireland Executive. (2013). PISA highlights need for continued reform — O'Dowd,
   3 December. Available at: www.northernireland.gov.uk/index/media-centre/news-departments/
   news-de/news-de-december-2013/news-de-031213-pisa-highlights-need.htm
Nusche, D., Earl, L., Maxwell, W., & Shewbridge, C. (2011). OECD reviews of evaluation and
   assessment in education: Norway. Available at: www.oecd.org/ norway/48632032.pdf
OECD. (2001). Definition and selection of competencies: Theoretical and conceptual founda-.
   tions (DeSeCo). Background paper, revised December. Neuchatel, Switzerland: OECD/
   DeSeCo Secretariat.
OECD. (2005). Formative        assessment:   Improving     learning   in secondary   classrooms.   Patis:
   OECD Publishing.
OECD. (2013). Synergies for better learning: An international perspective on evaluation and
   assessment. Paris: OECD Publishing.                                    ;
OECD. (2014). TALIS 2013 results. An international perspective on teaching and learning.
   Paris: OECD Publishing. Available at: www.oecd-ilibrary.org/education/talis-2013-results_
   9789264196261-en
OECD. (2015a). Education at a glance 2015: OECD indicators. Paris: OECD Publishing.
   Available at: http://dx.doi.org/10.1787/eag-2015-en
OECD. (2015b). PISA 2015 technical standards. Paris: OECD Publishing. Available at: www.
   oecd.org/pisa/pisaproducts/PISA-2015-Technical-Standards.    pdf
OECD. (2016a). PISA 2015 results (volume I): Excellence and equity in education. Paris: OECD
   Publishing. Available at: http://dx.doi.org/10.1787/9789264266490-en
OECD. (2016b). PISA 2015 released field trial cognitive items. Paris: OECD Publishing.
   Available at: www.oecd.org/pisa/pisaproducts/PISA2015-Released-FT-Cognitive-Items.pdf
Office of Qualifications and Examinations Regulation (Ofqual). (2013). Review of controlled
   assessment in GCSEs. Coventry: Ofqual. Available at: www.gov.uk/government/uploads/
   system/uploads/attachment_data/file/377903/2013-06-11-review-of-controlled-assessment-
   in-GCSEs. pdf                                                :
Ofsted. (1996). The annual report of her majesty’s chief inspector of schools. London: HMSO.
Ofsted. (2015). School inspection handbook. Available at: www.gov.uk/government/uploads/
   system/uploads/attachment_data/file/391531/School_inspection_handbook.pdf
O'Leary, M. (2006). Towards a balanced assessment system for Irish primary and secondary
   schools.   Oideas,   52,   7-24.   Available   at:   www.education.ie/en/Publications/Education-
    Reports/Oideas/Oideas-52.pdf
Painter, D.D. (2009). Providing differentiated learning experiences through multigenre pro-
    jects. Intervention in School and Clinic, 44, 288-93.
Partnership for 21st Century Skills. (2006). A state leader's action guide to 21st century skills:
    A new vision for education. Tucson, AZ: Author.
Partnership for 21st Century Skills. (2015). P21 framework definitions. Available at: www.p2l.
    org/storage/documents/docs/P21_Framework_Definitions_New_Logo_2015.pdf
Pat-El, RJ., Tillema, H., Segers, M., & Vedder, P. (2011). Validation of assessment for learning
    questionnaires for teachers and students. British Journal of Educational Psychology, 83, 98-113.
Pearson Education. (2016). Test of English language learning (TELL). Menlo Park, CA: Pearson
    Education.
Perkins, R., Cosgrove, J., Moran, G., & Shiel, G. (2012). PISA 2009: Results for Ireland and
    changes since 2000. Dublin: Educational Research Centre.
Perkins, R., Shiel, G., Merriman, B., Cosgrove, J., & Moran, G. (2013). Learning for life: The
    achievements of 15-year-olds in Ireland on mathematics, reading literacy and science in
    PISA 2012. Dublin: Educational Research Centre.
                                                                                References         219
                                                                                           difficulties.
Phillips, S., Kelly, K., & Symes, L. (2013). Assessment of learners with dyslexic-type
    London: Sage.
Pintrich, PR. (2000). The role of goal orientation in self-regulated learning. In M.       Boekaerts,
                                                                                           Diego, CA:
    PR. Pintrich & M. Zeidner (eds) Handbook of self-regulation (pp. 451-502). San
    Academic Press.
                                                                                     in    Education:
Pollitt, A. (2012). The method of adaptive comparative judgement. Assessment
    Principles, Policy & Practice, 19 (3), 281-300.
                                                                know (6th edition). Boston:
Popham, WJ. (2011). Classroom assessment: What teachers need to
   Pearson.
                                                                                  Boston: Pearson.
Popham, WJ. (2012). Assessment bias: How to banish it (2nd edition).
   Available at: ati. pearson.com/downloads/chapters/Popham_Bi       as_BK04.  pdf
                                                                        questions towards drafting
Primary Professional Development Service (PPDS). (2007). Prompt
                                      at: http://w ww.educa tetogeth er.ie/wo rdpress/wp-content/
   an assessment policy. Available
   uploads/2010/02/   assessme nt-1.doc
                                                                       ist, 21, 1026-9.
Putwain, D. (2008). Examination stress and test anxiety. The Psycholog
                                               Enhancing    comprehension      and test taking across
 Raphael, T.E., & Au, K:H. (2005). QAR:
    grades and content areas. The Reading Teacher, 59 (3), 206-21.
                                                                      of formative assessment
 Ratnam-Lim, C.T.L., & Tan, K.H.T. (2015). Large-scale implementation
    practices in an examination-oriented      culture. Assessment in Education: Principles, Policy
    and Practice, 22 (1), 61-78.
                                                                 00. Cambridge: Cambridge
 Roach, J. (1971). Public examinations in England 1850-19
    University Press.
                                                               e-choice items: a meta-analysis
 Rodriguez, M.C. (2005). Three options are optimal for multipl
                                                            and Practice, 24 (2), 3-13.
    of 80 years of research. Educational Measurement Issues
                                                                       y of the research on the effects
 Rogers, C.M., Lazarus, S.S., & Thurlow, M.L. (2016). A summar
                                                                   402. Minneapolis, MN: University of
     of test accommodations: 2013-201 4, NCEO Report No.
                                                              s.
     Minnesota, National Center on Educational Outcome
                                                              Project IRIS: Inclusive research in Irish
 Rose, R., Shevlin, M., Winter, E., & O’Raw, P. (2015).
                                       No.  20. Trim,  Ireland:  NCSE.
     schools. NCSE Research Report
                           M.  (2013).   Examin ing   formati ve  feedback in the classroom context:
 Ruiz-Primo, M.A., & Li,
                                                                 Handbook of research on classroom
     new research perspectives. In J.H. McMillan (ed.)
     assessment (pp. 215-32). Thousand Oaks,        CA:  Sage.
                                                                 of instructional systems. Instructional
 Sadler, D.R. (1989). Formative assessment and the design
     Science, 18 (2), 119-44.
                                                         ing the territory. Assessment in Education:
  Sadler, D.R. (1998). Formative assessment: revisit
     Principles, Policy and Practice, 5 (1), 77-84.
                                                      can the world learn from educational change
  Sahlberg, P. (2015). Finnish Lessons 2.0: What
                                                           College Press.
      in Finland? (2nd edition). New York: Teachers
                       yke, J.E. (2001) . Asses sment (8th  edition). Boston: Houghton Mifflin Co.
  Salvia, J., & Ysseld                                                                 tables. Slough:
                                                ell, S. (2013). INCA comparative
  Sargent, C., Foot, E., Houghton, E., & O’Donn                                               /
                                           Resear ch. Available at: www.nfer.ac.uk/what-we-do
     National Foundation for Educational
                                                        etablesMarch201 2.pdf
      information-and-reviews/inca/INCA comparativ                                                     for
                                                  ased assessment in e-learning: a framework
  Scalise, K., & Gifford, B. (2006). Computer-b                              ology   platfo rms. Journ al
                                                    ons  and tasks for techn
      constructing ‘intermediate constraint’ questi
                                                  4 (6).
      of Technology, Learning and Assessment,                                                          s.
                                    mics  of knowl edge: Why education is key for Europe's succes
  Schleicher, A. (2006). The econo                                   lisbo ncoun  cil.n et/co  mpone nt/
                                                    ble at: www
      Brussels: The Lisbon Council. Availa
      downloads/  ?id=2 19                                                                        reform:
                                                 through professional learning and policy
  Schleicher, A. (2016). Teaching excellence                                  ble   at: www.o  ecd.o rg/
                                                  OECD Publishing. Availa
      Lessons from around the world. Paris:                                   ng-an  d-pol icy-r  eform -
                                                      -professional-l  earni
      oublications/teaching-excellence-through
      9789264252059-en.htm
                \‘
220      Understanding and applying assessment in education
State Examinations Commission.        (SEC). (2008a). Junior Certificate ordinary level geography.
  Athlone,    Ireland:   Author.   Available    at:   https://www.examinations.ie/tmp/1486398113_
    2717970.pdf                                                                          <
State Examinations Commission. (SEC). (2008b). Junior Certificate ordinary level science. Athlone,
    Ireland: Author. Available at: https://www.examinations.ie/tmp/1486398448_2849951.pdf
State Examinations Commission. (SEC). (2015). Junior Certificate ordinary level science. Athlone,
    Ireland: Author. Available at: https://www.examinations.ie/tmp/1485961029_6941277.pdf
State Examinations Commission (SEC). (2016). Reasonable accommodations at the 2017
    certificate examinations. Instructions for schools. Athlone: Author. Available at: www.
    examinations.ie/schools/cs_view. php?q=746932c1559a1f720973fca784ec3417af46b5 16
Stiggins, R.J. (1992). High quality classroom assessment: What does it really mean? Educational
   Measurement Issues and Practice, 11 (2), 35-9.
Stiggins, R.J., Griswold, M.M., & Wikelund, K.R. (1989). Measuring thinking skills through
    classroom assessment. Journal of Educational Measurement, 26 (3), 233=240.
Stobart, G. (2008). Testing times: The uses and abuses of assessment. Abingdon, Oxon: Routledge.
Stobart, G., & Eggen, T. (2012). High-stakes testing: value, fairness and consequences.
   Assessment in Education: Principles, Policy and Practice, 19 (1), 1-6.
Stone, C.L. (1983). A meta-analysis of advance organizer studies. The Journal of Experimental
   Education, 51 (4), 194-9.
Sunday Times, The. (2016, 6 November). Standardised achievement tests queried as pupils’
    results improve (by Sean McCarthaigh). Available at: www.thetimes.co.uk/edition/ireland/
    standardised-achievement-tests-queried-as- pupils-results-improve-fpjj32x6r
Taylor, R.L. (2008). Assessment of exceptional students: Educational and psychological procedures
    (8th edition). Boston: Pearson.
Teaching Council, The. (2016). Code of professional conduct for teachers (2nd edition).
    Updated July 2016. Available at: www.teachingcouncil.ie/en/Fitness-to-Teach/Code-of-
    Professional-Conduct/
Telegraph, The. (2012, 31 October). Ofqual: 340,000 GCSE and A-level exams ‘marked up’.
    Available at: www.telegraph.co.uk/education/educationnews/9645848/Ofqual-340000-
    GCSE-and-A-level-exams-marked-up.html
Thompson, M., & Wiliam, D. (2008). Tight but loose: a conceptual framework for scaling up
    school reforms. In E.C. Wylie (ed.) Tight but loose: Scaling up teacher professional deve-
    lopment in diversé contexts. ETS Research Report No. RR-08-29 (pp. 1-43). Wiley Online
   Library.
Thurlow, M.L. (2014). Instructional and assessment accommodations in the 21st century. In
   L. Florian (ed.) The Sage handbook of special education, Vol. 2 2nd edition, pp. 597-612).
   Los Angeles, CA: Sage.
                                                                                                   learners.
Tomlinson, C. (1999). The differentiated classroom: Responding to the needs of all
   Alexandria, VA: Association for Supervisio  n   and  Curriculum   Developme     nt.
                                                                                                       , IL:
Tomlinson, C.A. (2000). Differentiation of instruction in the elementary grades. Champaign
                                                                                                of Illinois.
   ERIC Clearinghouse on Elementary and Early Childhood Education, University
                                                                                                   McMillan
Topping, K J. (2013). Peers as a source of formative and summative assessment. In J.H.
                                                       assessment  (pp.  395-412).       Thousand      Oaks,
   (ed.) Sage handbook of research on classroom
                                                                            :
   CA: Sage:
                                                                                        objectives, assess-
Torrance, H. (2007). Assessment as learning? How the use of explicit learning
                                                                     training    can   come    to dominate
   ment criteria and feedback in post-secondary education and
                                                 ,  Policy  and Practice,    14  (3),  281-94.
   learning. Assessment in Education: Principles
                                                                                    at: www.unicef.org/
United Nations. (1989). UN convention on the rights of the child. Available
   crc/
                                                                              disabilities. Available at:
United Nations. (2006). UN convention on the rights of persons with
   www.un.org/development/desa/disabilities/
222     Understanding and applying assessment in education
United Nations High Commissioner for Refugees (UNHCR). (2016). Global trends: Forced
    displacement in 2015. Geneva: UNHCR. Available at: www.unhcr.org
Van Bramer, J. (2011). Teacher talk and assistance and self-regulated learning within the
    context of Rtl and explicit and systematic teaching. The New England Reading Association
   Journal, 46 (2), 40-4.
Von Davier, A.A., & Halpin, P. (2013). Collaborative problem solving and assessment of cogni-
    tive skills: Psychometric considerations. Research Report No. ETS RR-13-41. Princeton, NJ:
    Education Testing Service.
Wall, E. (2004). MICRA-T level 4 test manual. Dublin: CJ Fallon.
Walsh, B., & Dolan, R. (2009). A guide to teaching practice in Ireland. Dublin: Gill & Macmillan.
Weiner, B. (2000). Intrapersonal and interpersonal theories of motivation from an attributional
   perspective. Educational Psychology Review, 12 (1), 1-14.
Wiliam, D.. (2011). What is assessment   for learning? Studies in Educational Evaluation, 37,
   3-14.
WJEC/CBAC. (2016). WJEC Eduqas GCSE in French. Specification. Teaching from 2016. For
   award from 2018. Cardiff: Author. Available at: http://www.eduqas.co.uk/qualifications/
   french/gcse/eduqas-gcse-french-spec-from-2016.pdfrlanguage_id=1&dotcache=no&dotcache
   =refresh
World Health Organization (WHO). (2016). Overa billion people live with some form of disability.
   Available at: www.who.int/features/factfiles/disability/facts/en/
Wragg, E.C. (1997). Assessment and learning. London: Routledge.
Zimmerman, B,J. (2002). Becoming a self-regulated learner: an overview. Theory into Practice,
   41 (2), 64-70.                                                    /
*—
 _
alal
        Added to a page number ‘f’ denotes a figure and ‘t’ denotes a table.
       A Levels 5, 13, 14, 20, 21, 28,30, 32, 142, 149, 186        analytic scoring 104-6, 121-2
        ability tests 35, 108, 182                                 anchor tests 135            ;
        academic language 171, 181                                 Anderson, L.W. 70, 71, 92, 117
         academic standards 32                                     anecdotal records 120-1
         accommodations 37, 174-5, 176, 177, 205                   answer supply format 37
           see also reasonable accommodations                      anxiety 144, 178
         accountability 7, 39, 85, 112, 122, 203                   aptitude tests 18, 35
         accuracy                                                  Army Alpha/Beta tests 34
           of associations, matching €XEFCISES -112                assessment
           of scoring 106           see                              21st-century skills 201-3
         achievement                                                 change/evolution 9, 10
ee
SS
ese
Sy                                                                   choosing forms to use 199-200
           categorisation 92
           curricula as a structured bans for 91                     contemporary approaches 3
           domains 92                                                continuum of methods 200f
           impact of AfL on 56, 57                                   culture 77
           levels, standardised tests 135                            definitions 58t
           meta-analysis 61-2         =                              and development of SEN policy 170
           self-assessment and 78                                    disjoint between learning theories, curriculum
                                                                          and   93                        ”
            tests of 34
         active engagement 65                                         ensuring fit between learning outcomes and
         active learning74
                        74                                                 format of 90-1
         Adamson, F. 118                                              importance of appropriate framing and use 203
         adaptations 168, 174                                         legislation and codes-of practice 6-8
                                                                      outcomes 13, 19, 118, 119
         Adaptive Comparative Judgement 106
                                                                      paradigm wars 198-9
         admissions officers, use of assessment information 4
         affective learning 92                                        personalities in 35, 64, 94, 125
                                                                      policies 9, 10-11
         affects, as an element of achievement 92t
         age-equivalent scores 136-7                                  principles 1, 9-10, 11, 164
       - aggregated scores 135                                        procedures and practices 8-9, 10-11, 84, 85
         aggregated test data 34                                      purposes 1, 3, 4-6, 4t, 9, 53, 95
         agreement rating 120
                                                                      role in differentiation 168-9
         alphabet knowledge 179                                       role in teaching and learning 2-3
                                                                      stakeholders 1, 3-4, 161, 200-1
         alternate forms of reliability 21
224         Understanding and applying assessment in education
school club (case study) 67-8                        Shepard, L. 46, 93, 94, 201
School Information (England) (Amendment)             Sheil, G. 34, 40, 63, 92, 186
     Regulations (2012) 6                            short answer formats 36, 37, 90, 96-9
school leaders 9                                     short-term planning 193-6
school policyGes)                                    SIGMA Mathematics Test 108
  assessment issues 190-1                            Simon, T. 34, 35
  role in encouraging differentiation 168            simulations 130
School Self-Evaluation (SSE) 191                     Singapore 85
school-based tasks                                   skills
  combining marks from formal tests and 145              project-related 123
  Junior Cycle 146-7                                     teachers’ assessment-related 7
school-level assessments                                 see also cross-curricular skills; transferable skills/
   planning 190-2                                              competencies; twenty-first century skills
   purposes 30t                                       skilis analysis, mathematical 180
   role in certification 186                          Skinner, B.F. 45
   skills in 117                                     Skrtic, T.M. 168
school-level progress, lower-. and upper-primary      social cognitive theories 46
      grades 135                                      social constructivism 74
 school-level record-keeping 158-9                    social learning theory 45
 schools, as social spaces for learning 94            social spaces, for learning 94
 Schumm, J.S. 49                                      socio-cultural theories 39, 46-7, 74
 science 9, 33f, 128-9, 188                           socio-economic bias 26
 scoring see test scoring; tests scores               socio-economic inequity 39
 Scotland                                             Spearman, C. 91
   assessment policy 56                               Spearman-Brown prophecy formula 22
                                                      Special education revisited 170
   Curriculum for Excellence 56, 99, 117
   differentiated instruction 50                      special educational needs
   grading, public exams 147                             administration of standardised tests 136
   key AfL strategies 62t                                assessment planning 95
   professional standards 7                              assessment and policy development 170
   selected key competencies 94t                         diagnostic assessment and identification of 179-81
                                                         educational inclusion 166-8
   self-assessment 78
                                                         inclusive school environments 172
   SEN and educational inclusion 167
   student profiles 153                                  individual education plans (IEPs) 169
                                                         observation and questioning 172-3
 Scottish    National Qualifications 5, 21, 32
                                                         students
 Scriven,    M. 51
                                                            numbers with 165, 205
 second     language acquisition 171
                                                             see also atypical students
 second     language learners 181
                                                          the term assessment in relation to 168
  selection, assessment for 4t, 32, 34, 35
  selection-type items 37                                 see also disability; learning difficulties
                                                       split-half reliability 21-2
  self-assessment 55, 56, 77-81
                                                       stability (task outcome) 49
     effective 81
                                                       stage theory cognitive models 45
     performance-based 116
                                                       stakeholders
     portfolios 128
                                                         diversity of 1, 200-1
     pre-flight checklist 47
                                                          need for assessment information 3-4
     reporting in-153
                                                         reporting to 161
  self-awareness 3
                                                       standard deviation(s) 36, 38, 41, 62, 136, 137, 140
  self-belief 46
                                                       standard error of measurement (SEM) 22-3, 137-8
  self-efficacy theory 48-9
                                                       Standard examinations (Scotland) 147
  self-esteem 77, 78
                                                       standard scores 36, 135, 136, 137, 138, 139, 140, 148
  self-monitoring, learning 3, 47, 67, WS
                                                       standardisation 25, 31, 171
  self-reflection 48, 65, 74
                                                78     standardisation studies 36
  self-regulated learning 47-8, 50, 54, 55, 74,
                                                       standardised tests
  self-regulation 3, 77, 173
                                                         administration 37
  sequential drafting, essays 103-4                       alpha or KR,, values 22
  service plans 168
                                                          atypical students 172, 177-8) 181, 182
  severity error 122
                                                          concerns about the effects of 39
  shaping (behaviour) 45
234        Understanding and applying assessment in education
This is essential reading for students on initial teacher education courses including
university-based (BEd, BA.w/QTS, PGCE, PGDE, PME) and school-based (School
Direct, SCITT, Teach First) routes into teaching, and newly qualified teachers. It
is relevant also to teachers at any stage in their career who are keen to enhaince
their understanding and skills.
DAMIAN MURCHAN            is Assistant Professor in the School of Education, Trinity College, Dublin.
GERRY SHIEL is Research                   Fellow at the Educational Research Centre, Dublin.
    compan
         nro                    https://study.sagepub.com/murchanshiel
                                                                                                         |
                                                                                                       MIM
QSAGE ™w.sagepublishing.com