0% found this document useful (0 votes)
114 views23 pages

Cep2 Content Module 13

This document provides an overview of assessment in learning, including alternative forms of assessment. It discusses using assessment to measure authentic learning and assess process and product-oriented learning outcomes as well as affective learning. Specific assessment tools discussed include rubrics, performance-based assessments, self-reports, rating scales, and tests in various formats to measure abilities, aptitudes, and achievement. The purpose of assessment is to evaluate student learning and make instructional decisions to improve teaching.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
114 views23 pages

Cep2 Content Module 13

This document provides an overview of assessment in learning, including alternative forms of assessment. It discusses using assessment to measure authentic learning and assess process and product-oriented learning outcomes as well as affective learning. Specific assessment tools discussed include rubrics, performance-based assessments, self-reports, rating scales, and tests in various formats to measure abilities, aptitudes, and achievement. The purpose of assessment is to evaluate student learning and make instructional decisions to improve teaching.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 23

COURSE ENHANCEMENT PROGRAM 2

PED 6: Assessment in Learning 2 Page 1 of 23

TOPIC PED 6: ASSESSMENT IN LEARNING 2


Test
Assessment of Affective Learning Outcome
SUBTOPICS Alternative Assessment
, Levels or Scales of Measurement Giving
Grades

Welcome BCPians!
I. INTRODUCTION:

This module focuses on the principles, development, and utilization of alternative forms of assessment in measuring authentic learning. It
emphasizes on how to assess process and product - oriented learning outcomes as well as affective learning, Students will experience how to
develop rubrics and other assessment tools for performance - based and product-based assessment

At the end of this module, you should be able to:


1. Apply the principles of assessment in conceptualizing techniques for assessing authentic learning
2. Determine performance - based assessment tools
3. Demonstrate skills in preparing and reporting grades
4. Derive information from alternative forms of assessment in making instructional decisions

Before we proceed with the lesson, let’s have a short activity.

Before we proceed with the lesson, let’s have a short activity.

\\III. PRELIMINARY ACTIVITIES:


Briefly answer the Following Questions.
What is a test?
A test is a process to assess the student what have they learned so far in
a particular lesson. _
What is alternative form of assessment?

An alternative form of assessment is an authentic assessment that cater the


interest and needs of the students. _
How can assessment can help you in your instructional decisions?

It is the most important tool to improve my teaching. It gives me ideas regarding in instruction and
curriculum that I will use for my students with different skills and abilities. It can also help me
make decisions for the next term or year.
Great!!! You may proceed to the main lesson.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 2 of 23

IV: LESSON PROPER:

TEST
When defined within an educational setting, assessment, evaluation, and testing are all used to measure how much of the
assigned materials students are mastering, how well student is learning the materials, and how well student is meeting the stated
goals and objectives. Although you may believe that assessments only provide instructors with information on which to base a
score or grade, assessments also help you to assess your own learning.

Education professionals make distinctions between assessment, evaluation, and testing. However, for the purposes of this
tutorial, all you really need to understand is that these are three different terms for referring to the process of figuring out how much
you know about a given topic and that each term has a different meaning. To simplify things, we will use the term "assessment"
throughout this tutorial to refer to this process of measuring what you know and have learned.

Definitions:

A test or quiz is used to examine someone's knowledge of something to determine what he or she knows or has learned.
Testing measures the level of skill or knowledge that has been reached.

Evaluation is the process of making judgments based on criteria and evidence.

Assessment is the process of documenting knowledge, skills, attitudes and beliefs, usually in measurable terms. The goal
of assessment is to make improvements, as opposed to simply being judged. In an educational context, assessment is the
process of describing, collecting, recording, scoring, and interpreting information about learning.

Why is Assessment Important?


Hopefully by this point in your life you have discovered that learning can be fun! You have probably also realized that you are
constantly learning, whether you are in a classroom, a car, or a kitchen.

Assessment helps you build confidence in your ability to learn.

Perhaps you have heard that the global work culture is changing. Unlike your grandfather, you will probably have a number of
different jobs and careers during your lifetime. In order to be successful, you will need to have confidence in your ability to learn
and you will need to become a lifelong learner. Assessment plays a key role in developing your confidence in your ability to learn,
as well as in developing your lifelong learning skills.

Classification of Tests According to Format

I. Standardized Tests – tests that have been carefully constructed by experts in the light of accepted objectives.

1.Ability tests – combine verbal and numerical ability, reasoning and computations

2.Aptitude Tests – measure potential in a specific field or area; predict that degree to which an individual will succeed in any
given area such art, music, mechanical task or academic studies.

II.Teacher-made Tests – constructed by classroom teacher which measure and appraise student progress in terms of specific
classroom and instructional objectives

1.Objective Type – answers are in the form of a single word or phrase or symbol.
2.Limited Response Type – requires the student to select the answer from a given number of alternatives or choices.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 3 of 23

i. Multiple Choice Test – consists of a stem each of which presents three to five alternatives or options in which
only one is correct or better than the other. The correct option choice or alternative in each item is merely called answer and
the rest of the alternatives are called distracters or decoys or foils.
ii. True – False or Alternative Response – consists of declarative statements that one has to respond or mark true or
false, right or wrong, correct or incorrect, yes or no, fact or opinion, agree or disagree and the like. It is a test made up of items
which allow dichotomous responses.
iii.Matching Type – consists of two parallel columns with each dichotomous responses. Word, number, or symbol in
one column being matched to a word sentence, or phrase in the other column. The items in Column l or A for which a match is
sought are called premises, and the items in Column Il or B from which the selection is made are called
iv.Free Response Type or Supply Test- requires the student to supply or give the correct answer
1. Short Answer – uses a direct question that can be answered by a word, phrase, number, or symbol.
2. Completion Test-consists of an incomplete statement that can

v. Essay Type- Essay questions provide freedom of response that responses. Supply or give the correct answer. A
word, phrase, number, or symbol. Also be answered by a word, phrase, number, or symbol is needed to adequately assess students'
ability to formulate, organize, integrate, and evaluate ideas and information or apply knowledge and skills.

Other Classification of Tests


1.Psychological Tests- aim to measure students' intangible aspects of behavior, i.e., intelligence, attitudes, interests 2.Educational

Tests-aim to measure the results/effects of instructions.

3.Survey Tests-measure general level of student's achievement over a broad range of learning outcomes and tend to emphasize
norm - referenced interpretation

4.Mastery Tests -measure the degree of mastery of a limited set of specific learning outcomes and typically use criterion
referenced interpretations.

5.Power T Tests-designed to measure level of performance time conditions, consist of items arranged in order increasing
difficulty

6.Verbal Tests-one in which words are very necessary and the examinee should be equipped with vocabulary in attaching
meaning to or responding to test items.

7. Speed Tests - designed to measure the number of items an individual or responding to test items

8. Non-Verbal Tests-one in which words are not that important student responds to test items in the form of drawings, pictures,
or design

ASSESSMENT OF AFFECTIVE LEARNING OUTCOME


http://logcero.blogspot.com/2014/01/assessment-tools-in-affective-domain

Assessment Tools in the Affective Domain

The assessment tools in the affective domain are those that are used to assess attitudes, interests, motivations, and self-efficacy.
We consider a few of the standard assessment tools in the affective domain.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 4 of 23

Self-Report. This is the most common measurement tool in the affective domain. It essentially requires an individual to provide
an account of his attitude or feelings toward a concept or idea or people. Self-reports are also sometimes called "written reflections".
In using this measurement tool, the teacher requires the student to write his/her thoughts on a subject matter, like, "Why I like or
dislike Mathematics". The teacher ensures that the students write something which would demonstrate the various levels of the
taxonomy (from receiving up to characterization)

Rating Scale. This is a set of categories designed to elicit information about a quantitative attribute in social science. Common
examples are the Likert scale and 1-10 rating scales for which a person selects the number which is considered to reflect the
perceived quality of a product. The basic feature of any rating scale is that it consists of several categories. These are usually
assigned integers. According to Nitko (2001), rating scales can be used for teaching purposes and assessment.

Rating scales help students understand the learning target/outcomes and to focus students' attention on performance.
completed rating scale gives specified feedback to students as far as their strengths and weaknesses with respect to the targets to
which they are measured students not only learn the standards but also may internalize the set of standards ratings help to show
each students' growth and progress

Example: Rating Scale (Attitude towards Mathematics)


Directions: Put the score on the column for each of the statement as it applies to you. Use 1 to 5, 1 being the lowest and 5 the
highest possible score.

Indicators Score
1. I am happy during Math class

2. I get tired doing board work and drills

3. I enjoy solving word problems

Types of Rating Scales


The most common type of rating scales is numerical rating scales and a descriptive graphic rating scale.

The numerical rating scale translate the judgment of quality or degree into numbers. To increase the objectivity and consistency
of results from numerical rating scales, a short verbal description of the quality level of each number may be provided.

Example:
Directions: Indicate the degree to which the student contributes to team activity by encircling the appropriate number. The
numbers represent the following values:
4 - constantly appropriate and effective; 3 -
generally appropriate and effective;
2 - needs improvement, may do other unrelated tasks; and
1 - unsatisfactory, disruptive and do other tasks not related to the activity.

To what extent does the student participate in team meetings and discussions? 1 2 3 4

Descriptive Graphic Rating Scales. A better format for the rating is this descriptive graphic rating scale that replaces
the ambiguous single word with short behavioral descriptions of the various points along the scale. Describing the points of
the scale by behavior descriptions leads to increased consistency of ratings across raters and students.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 5 of 23

Example:
Directions: Make your ratings on each of the following by placing X anywhere along the horizontal line under each item. In
the space for comments(s), include anything that helps clarify your rating.
To what extent does the student participate in team meetings and discussions?

Common Rating Scale Errors


The table below contains the common rating scale errors that teachers and students must be familiar with to avoid
committing such kind of errors during the assessment.

3. Semantic Differential Scales - tries to assess an individual's reaction to specific words, ideas or concepts in terms of
ratings on bipolar scales defined with contrasting adjectives at each end. An example is shown below:

4. Likert Scale - is a simple and widely used self-report method in assessing affect. It requires that individuals tick on a box to
report whether they "strongly agree", "agree", "undecided", "disagree", or "strongly disagree", in response to many items concerning
attitude object or stimulus. An example is shown below:

Constructing the Likert Scale Instrument


 Write a series of statements expressing positive and negative opinions toward attitude objects.
 Select the best statements (at least 10), with a balance of positive and negative opinions and edit as necessary.
 List the statements combining the positive and negative and put the letters of the five-point scale to the left of each
statement for easy marking.
 Add the directions, indicating how to mark the answer and include a key at the top of the page if letters are used for each
statement. Some prefer to drop the undecided category so that respondents will be forced to indicate agreement or
disagreement

5. Checklists - are the most common and perhaps the easiest instrument to construct in the affective domain. A checklist
consists of simple items that the student or teacher marks as "absent" or "present". Here are steps in constructing a
checklist:

 Enumerate all the attributes and characteristics you wish to observe relative to the concept being measured. For instance,
if the concept is "interpersonal relation", then you might want to identify those indicators or attributes which
constitute evidence of good interpersonal relation
 Arrange these attributes as a shopping list of characteristics
 Ask the students to mark those attributes or characteristics which are present and to leave blank those which are
not.

Personality Assessment is a proficiency in professional psychology that involves the administration, scoring, and
interpretation of empirically supported measures of personality traits and styles in order to:

Refine clinical diagnoses; Structure and inform psychological interventions; and increase the accuracy of behavioral prediction in a
variety of contexts and settings (e.g., clinical, forensic, organizational, educational).

Specialized Knowledge
Psychometric properties of test instruments which demonstrate their construct validity and clinical utility Elements of psychometric
theory, including issues of reliability, validity, reference group norms, limits of generalizability and test
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 6 of 23
construction Theories of intelligence and human cognition, including the role of race and ethnicity in intellectual
evaluation Theory, administration, and interpretation of

performance-based measures of personality such as the Rorschach Inkblot Test and other major free-response tests; and major self-
report inventories, including the applicability of specific population norms to individual clients.

Appropriate selection of instruments to answer specific referral questions including the construction of a test battery

Data integration from multiple data sources, including interview, psychometric tests and collateral sources

Communication of assessment results to different referring individuals and agencies and feedback to clients themselves The
relationship between assessment and intervention Ethics, laws and regulations pertaining to Personality Assessment in general
and in particular settings

projective test, in psychology, examination that commonly employs ambiguous stimuli, notably inkblots (Rorschach Test) and
enigmatic pictures (Thematic Apperception Test), to evoke responses that may reveal facets of the subject’s personality by
projection of internal attitudes, traits, and behavior patterns upon the external stimuli. Projective tests are also used, less
frequently, to study learning processes. Other projective methods involve requiring subjects to build wooden block
structures, complete sentences, paint with the fingers, or provide handwriting samples; additional methods include association
tests in which spoken words serve as the stimuli.

Non - projective

Interview method: The original method used for the assessment of personality and the one which is still found to be useful is the
interview method. In interview there is a direct face to face contact between the interviewer and the interviewee. The
interviewer is the psychologist, or the psychiatrist and the interviewee may be the patient or a subject. In this method the
psychologists listen to the subject’s own story, and he seldom interferes while the subject is talking. He is sympathetic in
listening breaks the subjects’ restraints and he begins to talk in more intimate manners. This draws out the subject’s true
personality. Sometimes the subject is questioned about his attitude and interests. The interview has got two forms namely guided
and unguided. In the guided form the psychologist has the pre arranged questions where as in the unguided form the questions are
not pre - arranged.

A highly skilled interview can gain a great deal of useful information about his subject from the interview. The subject’s
manner of speaking his hesitation averted glans (escape) provide important clues to his personality. But the interview method has
serious limitations people bearing greatly in their skill as interviewers. The results of the interview though clear to the
interviewer cannot be expressed in precise terms.

Behavior test

One of the behavioral tests is called Brook test which is designed to test group participation in problem solving and to reveal natural
leader. Here a group of 4 to 5 candidates are taken to a brook. They will be provided with a few materials like tins, ropes, and
bamboos. Then they are asked to imagine themselves to ba in a situation where they are faced with enemy force and their leader to
the other side of the river. Here how the group starts its work, how the leader emerges will be observed by a team of experts
including a psychologist, a doctor and an army officer based on the performance the fit person will be selected for the job.

Characteristics of an Assessment Tool


1. Reliability
The right assessment tool produces the same results over time. So there's a consistency, or accuracy in these results. Here, you
should consider whether the test can replicate results whenever it is used.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 7 of 23
For example, if the students perform the same tasks simultaneously, then such assessment passes as
reliable.

2. Validity
The validity of an assessment boils down to how well it measures the different criteria being tested. In other words, it is the
idea that the test measures what it intends to measure.

This means your assessment method should be relevant to the specific context. For example, if you're testing physical strength,
you shouldn't send out a written test. Instead, your tests should include physical exercises like pushups and weightlifting.

3. Equitable
A good assessment tool is equitable, which means it doesn't favor or disfavor any participant. Fair assessments imply that
students are tested using methods and procedures most appropriate to them. Every participant must be familiar with the test
context so they can put up an acceptable performance.

4. Standardization
Standardization means applying consistency to your testing methods. For example, if you're sending out a questionnaire, it
should have the same set of questions for all participants, and all the answers should be graded using the same criteria.

Other characteristics of assessment tools include:

 A good assessment tool should provide a window for high-quality feedback.


 It is feasible and accounts for equivalence.
 It should motivate participants to be involved in the testing.
 It should be transparent, non-discriminatory, and match expectations.

Types of Assessment Tools

Educational Assessment Tool

An educational assessment tool is used for evaluating students’ performance and their level of knowledge in a
particular subject. Educational assessment tools can be used during the learning process or on an ongoing basis.

For example, in formative assessment, the teacher deploys different methods to track a student’s performance as learning
progresses. This allows the instructor to identify and fix knowledge gaps as they move from one learning phase to another. But
in summative assessment, the teacher waits till the end to measure the student’s performance using a rubric of expected
learning outcomes.

Expectedly, different assessment tools work for different types of educational assessment. So, while quizzes and straw polls are
suitable for formative evaluation, end-of-term examinations with close-ended questions are better- suited for summative
assessments.

Examples of Educational Assessment Tools

1. ASSISTments: This is a free online platform that teachers can use to assign Maths tests to students, whether during the
learning process or outside the classroom. ASSISTments allow instructors to provide immediate feedback as students turn in
their tests.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 8 of 23
2. Formplus: Teachers can use Formplus to create surveys, questionnaires, and polls for classroom assessments.
Students can participate in these assessments even without an internet connection, and teachers can analyze responses
immediately.

3. Google Classroom Question Tool: Google classroom allows you to create and share questions with your students right in the app.
Teachers can each student’s answer separately or set up a discussion with different groups.

4. Quizizz Tool: If you’re conducting fact-based tests, then the Quizizz tool is a great option. In Quizizz, you can set up self-
graded tests, which helps you save time on assessments.

Pros of Educational Assessment Tools


 Educational assessment tools help teachers to track and evaluate progress throughout the learning process.
 They provide enough data for improving the overall teaching and learning experiences.
Cons of Educational Assessment Tools
 Educational assessment tools can cause anxiety or affect students’ motivation, mainly when they do not
perform as well as they wanted.
 It is not always an accurate reflection of how much progress the students have made.

Research Assessment Tool


A research assessment tool is used to measure the impact of a systematic investigation based on specific criteria. These criteria
could be the research results, level of participation from research subjects, and other similar metrics.

For example, after a systematic investigation, you can send out an internal survey to gather feedback from key players in the
process. A research assessment tool can also help you determine the extent to which the research outcomes solve the
identified problem.

Pros of Research Assessment Tools


 Research assessment tools help you to identify gaps in your process.
 They allow you to determine the extent to which your systematic investigation solves the problem.
Cons of Research Assessment Tools
 Research assessment tools are subjective and easily affected by bias.
 It is time-consuming.

Project Assessment Tool


A project assessment tool is a method of evaluating the success of your project. These tools can be used in different contexts, such
as launching a new product, introducing an additional feature in an existing product, and launching a new branch of your
business.

Before choosing a project assessment tool or method of evaluation, you should consider different factors like the project
scale, how much money you've invested into it, and if you have the resources to pull it off. For example, a simple project
might not require a complex evaluation with complex assessment tools.

You should also define the project's success parameters. If success is quantitative, you need a tool that can measure numerical data
like the number of people who signed up for an event. If success is qualitative, then the assessment tool should evaluate the
participants' perceptions and impressions of the project.

Examples of Project Assessment Tools


1. Surveys: Tools like Form plus allow you to send out surveys, questionnaires, and polls for opinion sampling and feedback.
Before launching your project, you can send out a baseline survey, and after the project, you can administer a post-event
survey. This allows you to compare project outcomes with the issues raised in the baseline survey.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 9 of 23

2. Observation: Participant and non-participant observation methods can be used to gather first-hand feedback from
participants after your project launch.

3. Case Studies: A case study is an in-depth exploration of a subject matter related to real-life occurrences.

Pros of Project Assessment Tools


 Project assessment tools help you identify when the goals of a project have been achieved.
 It helps you to measure the ROI of your project.

Con of Project Assessment Tools


 It is time-consuming.

Scales of Measurement / Level of Measurement


1. Nominal Scale. Nominal variables (also called categorical variables) can be placed into categories. They don’t
have a numeric value and so cannot be added, subtracted, divided or multiplied. They also have no order; if they
appear to have an order then you probably have ordinal variables instead.
2. Ordinal Scale. The ordinal scale contains things that you can place in order. For example, hottest to coldest, lightest to
heaviest, richest to poorest. Basically, if you can rank data by 1st, 2nd, 3rd place (and so on), then you have data
that’s on an ordinal scale.
3. Interval Scale. An interval scale has ordered numbers with meaningful divisions. Temperature is on the interval scale: a
difference of 10 degrees between 90 and 100 means the same as 10 degrees between 150 and 160. Compare that to
high school ranking (which is ordinal), where the difference between 1st and 2nd might be .01 and between 10th and
11th .5. If you have meaningful divisions, you have something on the interval scale.
4. Ratio Scale. The ratio scale is exactly the same as the interval scale with one major difference: zero is meaningful.
For example, a height of zero is meaningful (it means you don’t exist). Compare that to a temperature of zero, which
while it exists, it doesn’t mean anything in particular (although admittedly, in the Celsius scale it’s the freezing point
for water).

Shapes of Distributions: Definitions, Examples


When a data set is graphed, each point is arranged to produce one of dozens of different shapes. The distribution shape can
give you a visual which helps to show how the data is:
 Spread out (e.g., dispersion, variability, scatter),
 Where the mean lies,
 What the range of the data set is, and many other useful statistics. Shapes of distributions are defined by several
different factors:

1.Number of peaks
The peaks are usually called modes; Sometimes they are called “local maximums“, but that term is more commonly used in
calculus. The mode tells you that the data count is higher in these areas than in any other areas on the graph.

 A unimodal distribution has one mode. A single peak can take on many shapes (e.g. very tall and thin or very squat
and fat). One of the most common types of unimodal distributions is the normal distribution, sometimes called the
“bell curve” because its shape looks like a bell.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 10 of 23

 A bimodal distribution has two modes.

 A multimodal distribution has three or more modes.


COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 11 of 23

 If a data set has no clear peaks (i.e., the whole graph looks flat), it’s called a “uniform distribution.”

One of the most confusing aspects about statistics is the terminology, and the “mode” is no exception. It may seem at first glance
that the word has two meanings:

The most common number in a set. For example, the mode of 1, 2, 2, 3, 5 is the number “2”.
A word to describe a peak on a graph.
However, the two terms mean the same thing. The most common number in the above list of numbers is “2”. If you were to plot
that set of numbers on a graph, the peak would happen at x = 2, which means that the peak contains the most common number in
the set. That said, if you are asked to “Find the Mode “, that usually indicates you should look for the most frequent item in the
set; it doesn’t mean you have to create a graph.

2.Symmetry
A symmetric graph has two sides that are mirror images of each other. The normal distribution is one example of a symmetric
graph.

Another type of symmetric graph is the U-distribution, which—perhaps not surprisingly— looks like the letter “U”.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 12 of 23

A symmetric box plot has the “box” in the center of the graph:

3.Skewness
Shapes of distributions can differ in skewness; these distributions are not symmetrical distributions. Instead, they have more
points plotted on one side of the mean than on the other. This causes long tails either in the negative direction on the number
line (a negative, or left skew) or in the positive direction on the number line (a positive, or right skew). For more on how
skewness affects shapes of distributions, see: Skewed Distribution in Statistics.

The tails of a distribution (i.e. how thin or fat they are) can also be described by kurtosis, which is measured against the standard
normal distribution. A positive value for kurtosis means you have a large peak and little data in the tails. A negative value means
you have a flattened peak with lots of data in the tails.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 13 of 23
CITE THIS AS:
Stephanie Glen. "Shapes of Distributions: Definitions, Examples" From StatisticsHowTo.com: Elementary Statistics for the rest of
us! https://www.statisticshowto.com/shapes-of-distributions/

Descriptive statistics
Descriptive statistics are used to describe the basic features of the data in a study. They provide simple summaries about the sample
and the measures. Together with simple graphics analysis, they form the basis of virtually every quantitative analysis of
data.

Descriptive statistics are typically distinguished from inferential statistics. With descriptive statistics you are simply
describing what is or what the data shows. With inferential statistics, you are trying to reach conclusions that extend beyond the
immediate data alone. For instance, we use inferential statistics to try to infer from the sample data what the population might
think. Or we use inferential statistics to make judgments of the probability that an observed difference between groups is a
dependable one or one that might have happened by chance in this study. Thus, we use inferential statistics to make inferences
from our data to more general conditions; we use descriptive statistics simply to describe what’s going on in our data.

Descriptive Statistics are used to present quantitative descriptions in a manageable form. In a research study we may have lots of
measures. Or we may measure many people on any measure. Descriptive statistics help us to simplify large amounts of data in a
sensible way.

Measures of Central Tendency


A measure of central tendency is a single value that attempts to describe a set of data by identifying the central position within
that set of data. As such, measures of central tendency are sometimes called measures of central location. They are also classed
as summary statistics. The mean (often called the average) is most likely the measure of central tendency that you are most
familiar with, but there are others, such as the median and the mode.

The mean, median and mode are all valid measures of central tendency, but under different conditions, some measures of central
tendency become more appropriate to use than others. In the following sections, we will look at the mean, mode and median,
and learn how to calculate them and under what conditions they are most appropriate to be used.

Mean (Arithmetic)
The mean (or average) is the most popular and well-known measure of central tendency. It can be used with both discrete and
continuous data, although its use is most often with continuous data (see our Types of Variable guide for data types). The mean is
equal to the sum of all the values in the data set divided by the number of values in the data set.

Median
The median is the middle score for a set of data that has been arranged in order of magnitude. The median is less affected by
outliers and skewed data.

Mode
The mode is the most frequent score in our data set. On a histogram it represents the highest bar in a bar chart or histogram. You
can, therefore, sometimes consider the mode as being the most popular option.

Measures of Variability
Variability refers to how "spread out" a group of scores is.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 14 of 23

Range
The range is the simplest measure of variability to calculate, and one you have probably encountered many times in your life. The
range is simply the highest score minus the lowest score.

Interquartile Range
The interquartile range (IQR) is the range of the middle 50% of the scores in a distribution.

Variance
Variability can also be defined in terms of how close the scores in the distribution are to the middle of the distribution. Using the
mean as the measure of the middle of the distribution, the variance is defined as the average squared difference of the scores
from the mean.

Standard Deviation
The standard deviation is simply the square root of the variance. This makes the standard deviations of the two quiz distributions
1.257 and 2.203. The standard deviation is an especially useful measure of variability when the distribution is normal or
approximately normal

Measures of Relationship
are statistical measures which show a relationship between two or more variables or two or more sets of data. For example,
generally there is a high relationship or correlation between parent's education and academic achievement. On the other hand, there
is generally no relationship or correlation between a person's height and academic achievement. The major statistical measure of
relationship is the correlation coefficient.

 Correlation is the relationship between two or more variables or sets of data. It is expressed in the form of a coefficient
with +1.00 indicating a perfect positive correlation; -1.00 indicating a perfect inverse correlation;
0.00 indicating a complete lack of a relationship.

Note: A simplified method of determining the magnitude of a correlation is as follows:


• .00 - .20 Negligible
• .20 - .40 Low
.40 - .60 Moderate
.60 - .80 Substantial
.80 - 1.0 High

 Pearson's Product Moment Coefficient (r) is the most often used and most precise coefficient; and generally used with
continuous variables.

 Spearman Rank Order Coefficient (p) is a form of the Pearson's Product Moment Coefficient which can be used
with ordinal or ranked data.

 Phi Correlation Coefficient is a form of the Pearson's Product Moment Coefficient which can be used with
dichotomous variables (i.e. pass/fail, male/female).In published studies or reports, correlational coefficients are
usually displayed as a r values.
For example, r = .66, p < .01 where r is the correlational coefficient and p is the level (.01) of statistical significance.

Measures Of Relative Position


are conversions of values, usually standardized test scores, to show where a given value stands in relation to other values of the
same grouping. The most common example in education is the conversion of scores on standardized tests to show where a given
student stands in relation to other students of the same age, grade level, etc. Converted
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 15 of 23
scores are based on the standard deviation or distance of a raw score from the mean for a normal curve or
distribution.

Standard Deviation, Sigma (Z) Score, T Score, College Board Score


 Converted scores, also referred to as standard scores are based on the standard deviation or distance of a raw score
from the mean for a normal curve or distribution. Examples of standard scores used in educational research are: Sigma
(z) scores, T scores, College Board scores, percentiles, and stanines.
 Sigma (z) Score - the deviation from the mean divided by the standard deviation. It is used to describe a score's
deviation from the mean. The range of z scores is usually -3.00 to +3.00 with a value at the mean of
0.00. For example: a standardized test with a mean of 70 and a standard deviation of 10, a score of 60 would be -10
points from the mean (deviation), divided by the standard deviation (10) equals a z score of -
1.00. Try calculating the z scores for scores of 85, 90, 50.
Stanines
 Percentiles and stanines help give comparative meanings to raw scores.
 Percentile - the point in the distribution below which a given percentage of scores fall. For example, if a score of
65 is at the seventieth percentile, then 70 percent of the scores fall below 65.
 Stanine - a conversion of raw scores into nine bands or stanines, with the 5th stanine being at the mean, 1st stanine the
lowest, and 9th stanine the highest.

ALTERNATIVE ASSESSMENT
This is a method of evaluation that measures a student’s level of proficiency in a subject as opposed to the student’s level of
knowledge. The overall goal of alternative assessment is to allow students to demonstrate their knowledge and execute tasks.

Alternative assessment is also called a performance test or authentic assessment because it is deeply rooted in one’s ability to
do something by leveraging newly gained knowledge. As part of the assessment, the student will need to perform meaningful
tasks that reflect a clear understanding of the teaching and learning objectives.

You can ask your students to create a portfolio, work with others on specific projects or engage in any other type of activity that
shows they have a full grasp of what has been discussed in the class or training.

Characteristics of Alternative Assessment


To better understand what alternative assessment is, let’s first look at some of the things alternative assessment is
not.

 Alternative assessment is not a high-stake or standardized method of evaluating students’ knowledge.


 It is not a replacement for other methods of classroom assessment like formative and summative
assessment methods.
 Alternative assessment does not measure the learner’s performance.
 Results from alternative assessments are not presented as quantitative or numerical data.
 Now you know what should not be expected from alternative assessment, it’s time to get familiar with the
characteristics of alternative assessment.

 Alternative assessment is tailored to the specific needs and abilities of individual students. You should not think of
it as a one-size-fits-all process.
 It is an instructor-led assessment process. This means the teacher develops a unique method to measure
the student’s ability.
 Alternative assessment methods are flexible and continuously adaptable.
 Alternative assessment or authentic assessment explores different learning styles and preferences in a bid
to measure learners’ abilities.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 16 of 23
 Alternative assessment methods are seamlessly aligned with the learning goals and objectives of the
course, class, or training.

Examples of Alternative Assessment

What methods of alternative assessment should you try out in class? There are a number of choices you can explore;
however, your final decision should be in line with the unique learning needs of every student and your teaching goals. Here,
we will discuss a few examples of alternative assessments.

Portfolio
You can ask students to build out a portfolio that demonstrates their knowledge of what has been taught in a class or training. A
portfolio is a collection of the different tasks a student has executed in the course of the class or training.

If you're handling learners in beginner classes, you can ask them to create a paper portfolio using a notebook; for advanced
learners, an online portfolio is the best bet. You can create a simple submission form on Formplus to collect links to your
students' portfolios easily.

Performance Test
Since alternative assessment is all about putting knowledge to work, design effective performance tests to help you rate a student's
level of skill or knowledge. A performance test requires the learners to execute complex tasks while the instructor observes
how they go about them.

As an instructor, you should have well-defined scoring criteria to effectively measure each student's ability and arrive at a valid
conclusion. Depending on the type of performance test you choose to adopt, you can create a quiz on Formplus for this
purpose.

Open Tests
An open test is a method of assessment that allows learners to refer to course materials as they take on tasks or write tests and
examinations. The questions in an open test require the student to provide responses that show how well they understand the
course.

Instead of memorizing the content of the training guide or course materials, the student has to apply the knowledge in the material(s)
to provide the best response to the stated questions.

Crib Sheet
This is another form of open-book evaluation. Instead of bringing the entire course materials to a test or examination, the student
selects important information from these materials to create some sort of abridged version called the crib sheet.

While it's up to the student to decide what makes it into the crib sheet, you should provide some level of guidance as their
instructor. You can subtly suggest the type of information that will be helpful during the assessment but you should not
impose your ideas on the students; if you really want to know how much they know.

Take-Home Assessment
Another way to assess your students' abilities is to give them take-home exercises. Take-home exercises typically check 2
boxes. Firstly, they require multiple references, and secondly, there isn't enough time for learners to do them in class.

With take-home exercises, you really cannot tell if the student received some help from their guardian(s) but you can ask the
students to work with materials that are strictly relevant to the course or training. You can create an online form on Formplus
for this and send it to students via email invitations.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 17 of 23

Collaborative Testing
This happens when you put the students in groups and get them to work together on different tasks. Ideally, you should pair
them or place the students in small groups of 3 or 4, to get the best results from this exercise.

As the name suggests, collaborative testing empowers the students to brainstorm together, solve challenges, and execute
ideas. At the end of each brainstorming session, you can ask your students to make individual submissions or submit
collective responses as a group.

Summaries
Instead of making students go through tests and assignments, you can ask them to summarise class readings, lectures, and
discussions. A good summary tells you 1 thing—the student has an impressive understanding of key concepts and ideas
from the classroom.

Asides from helping you track a student’s knowledge in real-time, summary writing also does the magic for knowledge
retention. Summary writing is a powerful way to improve learning and memory in the classroom.

Reports
Reports work just like summaries; they require the student to show how well they understand key concepts from class
discussions. However, reports take things a step further as the student needs to communicate his or her knowledge in a way
that presents a clear picture to whoever reads the report; even if such a person wasn’t part of the class.

Interviews
Pair students and ask them to perform interviews about different subject matters discussed in your class. For instance, if you
had a class discussion on a historical event, one student can roleplay as a key event player while the other person becomes the
interviewer: asking questions about that event.

Concept Maps
A concept map is a visual representation of the relationship between ideas and concepts. To test your students’ level of
understanding, ask them to build concept maps from scratch to show their knowledge or fill out existing concept maps.

GIVING GRADES
https://sc.edu/about/offices_and_divisions/cte/teaching_resources/grading_assessment_toolbox/what_is_grade/index.

Grading is used to evaluate and provide feedback on student work. In this way, instructors communicate to students how they are
performing in the course and where they need more help to achieve the course’s goals. For instructors, grades help to assess what
information, concepts, and skills students have successfully understood and which ones they have not. This kind of information
helps you know what you may need to reiterate in class and what may require reworking in the course design. Grades also provide
a standardized way of communicating student performance to third parties, including the departments in which students are
enrolled, and students themselves.

To ensure that grades are fair and to motivate students to improve their performance, instructors should think about the alignment
of their assignments to the course’s overarching goals and communicate their expectations and grading practices in a
transparent manner. Students are generally highly motivated to improve their work when the instructions of an assignment are
clear and achievable, when the standards the instructor uses for grading are clear and fair, and when the feedback is timely and
well aligned with the assignment in question. This kind of transparency will also enable students to understand what skills and
content they have learned and what they are still struggling with in the course.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 18 of 23
It helps to consider grading as a process. It is not simply a matter of assigning number or letter grades. As a process,
grading may involve some or all of these activities:

 Setting expectations with students through a grading policy


 Designing assignments and exams that promote the course objectives
 Establishing standards and criteria
 Calibrating the application of a grading standard for consistency and fairness
 Making decisions about effort and improvement
 Deciding which comments would be the most useful in guiding each student’s learning
 Returning assignments and helping students understand their grades What
Purpose Do Grades Serve?
Grades are essentially a way to measure or quantify learning and intellectual progress using objective criteria. They can serve
many purposes:

 As an evaluation of student work, effort, understanding of course content, skill development, and progress;
 As a source of self-motivation to students for continued learning and improvement;
 As a means of communicating feedback to students on their performance;
 As a means of communicating to students, parents, graduate schools, professional schools, and future
employers about a student’s potential in college and predictor for further success;
 As a means of organizing a lesson, a unit, or a semester in that grades mark transitions in a course and bring
closure to it (i.e. a summative assessment).
 As feedback, grades can also inform:

Students as to their own learning, clarifying for them what they understand, what they don’t understand, and where
they can improve.
Instructors on their students’ learning to help inform future teaching decisions.

Grades vs. Learning Assessment


Differences Between Grades and Assessment
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 19 of 23

V. ANALYSIS, APPLICATION AND EXPLORATION:

PED 6: ASSESSMENT IN LEARNING 2


Activity
Name:
Year & Section:

Direction: Identify if the statement is True or False.

1. A concept map is a visual representation of the relationship between ideas and concepts.

2. Grading is used to evaluate and provide feedback on student work.

3. Self-reports are also sometimes called "written reflections". In using this measurement tool, the teacher
requires the student to write his/her thoughts on a subject matter.

4. Essay Type – consists of two parallel columns with each dichotomous response.

5. An educational assessment tool is used for evaluating students’ performance and their level of
knowledge in a particular subject.

6. The original method used for the assessment of personality and the one which is still found to be
useful is the checklist method.
7. The overall goal of grading is to allow students to demonstrate their knowledge and execute tasks.

8. The range is the simplest measure of variability to calculate, and one you have probably
encountered many times in your life.

9. The mode is the most frequent score in our data set.

10. A research assessment tool is used to measure the impact of a systematic investigation based on specific
criteria.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 20 of 23

You are amazing! You’re done with the activities. Now, let’s summarize the lesson of
what we had discussed.

VI. GENERALIZATION:

Directions: Read and analyze the questions. Be specific and use complete sentences.

1. What are the goals of Alternative Assessment?


_ _ _
_ _ _
_ _ _

2. What is the difference of between grading and assessment?


_ _ _
_ _ _
_ _ _
_ _ _
_ _ _
_ _ _
_ _ _
_ _ _
_ _ _

3. How can we effectively give grades to students?


_ _ _
_ _ _
_ _ _
_ _ _
_ _ _
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 21 of 23

KUDOS! You have come to an end of this Module. OOPS! Don’t forget that you
have still an assignment to do. Here it is.

Direction: Read each sentences/ situation carefully and select the BEST letter of the correct answer among the
choices.

1. Can paper-and-pencil tests diagnose students’ ability to organize thoughts?


A.Yes, by way of the essay test.
B. No, unless the multiple-choice type of tests is made situational
C. Yes, by way of the selected response type of test.
D.No, they are limited only to simple recall.

2. Which of the most reliable tool for seeing the development in your pupils ability to write?
A.Interview of pupils
B. Portfolio assessment
C. Scoring rubrics
D. Self-assessment

3. Which of the following does not serve as organizer in the preparation of a portfolio?
A.Envelope
B. Clear book
C. Objective test
D. Album

4. Which are direct measures of competencies?


A.Personality tests
B. Performance tests
C. Paper-and-pencil tests
D. Standardized tests

5. Why should a teacher NOT use direct instruction all the time?
A.It requires much time
B. It requires use of supplementary materials
C. It reduces student engagement in learning
D.It is generally effective only in the teaching of concepts and abstractions.
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 22 of 23

VII. EVALUATION:
Direction: Choose the letter of the correct answer. Write the answer in the blank provided before each number.
Name: Year & Section:

1. A class obtained a low standard deviation in the Division test, what can you say about the students in that class?
A.The students performed very well in the test
B. They performed poorly in the Division Test
C. Their scores in the test are far apart
D.It is a homogeneous class

2. In his second item analysis, Mr. Castro found out that more from the lower group got test item #30 correctly. This means
that the test item has a
A.lower validity.
B. high reliability.
C. negative discriminating power.
D. positive discriminating power

3. What kind of distribution is given, when: Median = 73, Mean-64, Mode-81?


A.Negatively skewed distribution
B. Positively skewed distribution
C. Normal distribution
D.Frequency distribution

4. Lorenzo’s score in the third quarterly exam is 90 out of 110, what can you say about him?
A.He answered 90 items in the test correctly.
B. Lorenzo obtained a grade of 90 in the test
C. He was able to answer 90% of the test items correctly.
D. Lorenzo performed better than 90% of his classmates

5. When Ms. Ibay administered a Pre- test to her class, she found out that her students’ scores followed a negatively
skewed distribution. What kind of students does Ms. Ibay have?
A.Her students are of different abilities.
B. Most of her students are just average.
C. She has a lot of very good students in her class.
D.Ms. Ibay has homogeneous low performing students.

For Items 6-7: Study the table below and then answer the questions
Subject Mean Standard Deviation Reese’ Score
English 50 12 43
Science 40 7 91
Mathematics 70 15 109

6. In what subjects) did Reese perform best in relation to the group's performance?
A.English
B. Science
C. Mathematics
D.Mathematics and Science
COURSE ENHANCEMENT PROGRAM 2
PED 6: Assessment in Learning 2 Page 23 of 23

7. In what subject(s) were the scores less varied?


A.English
B. Science
C. Mathematics
D.No correct answer

8. Which group of students would be more likely to have a distribution of scores very close to a normal curve? A very,
A.bright class of pupils
B. slow group of pupils
C. large group of unselected pupils
D.small group of unselected pupils

9. Which test item measures a wide range of educational outcomes?


A.Alternate response
B. Completion
C. Essay
D.Multiple choice

10. Mr. Villalon wants to know the variability of her students' scores in the 4th periodical test which followed a skewed
distribution. What measure of dispersion would she most likely use?
A.Mean
B. Quartile deviation
C. Range
D. Standard deviation

This evaluation will be submitted on _


CONGRATULATIONS on reaching the end of this module! You may now proceed to the next module. Don’t forget to submit all the exercises,
activities, and portfolio on .

KEEP UP THE GOOD WORK. Well, Done!!!

Prepared by:

RYAN CHRISTOPHER M. VILLALON


Part-time Instructor

You might also like