Kirkpatrick's Four Levels of Training Evaluation in Detail
This grid illustrates the Kirkpatrick's structure detail, and particularly the modern-day interpretation of the Kirkpatrick learning evaluation
model, usage, implications, and examples of tools and methods. This diagram is the same format as the one above but with more detail
and explanation:
evaluation
level and type
evaluation description and
characteristics
examples of evaluation tools and
methods
relevance and practicability
1. Reaction
Reaction evaluation is how the
delegates felt, and their personal
reactions to the training or
learning experience, for example:
Typically 'happy sheets'.
Can be done immediately the training ends.
Feedback forms based on subjective
personal reaction to the training
experience.
Very easy to obtain reaction feedback
Did the trainees like and enjoy the
training?
Verbal reaction which can be noted and
analyzed.
Did they consider the training
relevant?
Post-training surveys or questionnaires.
Was it a good use of their time?
Online evaluation or grading by delegates.
Did they like the venue, the style,
timing, domestics, etc.?
Subsequent verbal or written reports given
by delegates to managers back at their
jobs.
Feedback is not expensive to gather or to
analyze for groups.
Important to know that people were not upset
or disappointed.
Important that people give a positive
impression when relating their experience to
others who might be deciding whether to
experience same.
Level of participation.
Ease and comfort of experience.
Level of effort required to make the
most of the learning.
Perceived practicability and
potential for applying the learning.
2. Learning
Learning evaluation is the
measurement of the increase in
knowledge or intellectual
capability from before to after the
learning experience:
Typically assessments or tests before and
after the training.
Interview or observation can be used
before and after although this is timeconsuming and can be inconsistent.
Relatively simple to set up, but more
investment and thought required than reaction
evaluation.
Highly relevant and clear-cut for certain
training such as quantifiable or technical skills.
Did the trainees learn what was
intended to be taught?
Methods of assessment need to be closely
related to the aims of the learning.
Did the trainee experience what
was intended for them to
experience?
Measurement and analysis is possible and
easy on a group scale.
What is the extent of advancement
or change in the trainees after the
training, in the direction or area that
was intended?
3. Behavior
Behavior evaluation is the extent
to which the trainees applied the
learning and changed their
behavior, and this can be
immediately and several months
after the training, depending on the
situation:
Did the trainees put their learning
into effect when back on the job?
Were the relevant skills and
knowledge used
Was there noticeable and
measurable change in the activity
and performance of the trainees
when back in their roles?
Was the change in behavior and
new level of knowledge sustained?
Would the trainee be able to
transfer their learning to another
person?
Is the trainee aware of their change
in behavior, knowledge, skill level?
Reliable, clear scoring and measurements
need to be established, so as to limit the
risk of inconsistent assessment.
Less easy for more complex learning such as
attitudinal development, which is difficult to
assess.
Cost escalates if systems are poorly designed,
which increases work required to measure and
analyze.
Hard-copy, electronic, online or interview
style assessments are all possible.
Observation and interview over time are
required to assess change, relevance of
change, and sustainability of change.
Measurement of behavior change is less easy
to quantify and interpret than reaction and
learning evaluation.
Simple quick response systems unlikely to be
Arbitrary snapshot assessments are not
reliable because people change in different adequate.
ways at different times.
Cooperation and skill of observers, typically
line-managers, are important factors, and
Assessments need to be subtle and
ongoing, and then transferred to a suitable difficult to control.
analysis tool.
Management and analysis of ongoing subtle
assessments are difficult, and virtually
Assessments need to be designed to
reduce subjective judgment of the observer impossible without a well-designed system
from the beginning.
or interviewer, which is a variable factor
that can affect reliability and consistency of
Evaluation of implementation and application is
measurements.
an extremely important assessment - there is
little point in a good reaction and good
The opinion of the trainee, which is a
increase in capability if nothing changes back
relevant indicator, is also subjective and
unreliable, and so needs to be measured in in the job, therefore evaluation in this area is
vital, albeit challenging.
a consistent defined way.
360-degree feedback is useful method and
need not be used before training, because
respondents can make a judgment as to
change after training, and this can be
analyzed for groups of respondents and
trainees.
Behavior change evaluation is possible given
good support and involvement from line
managers or trainees, so it is helpful to involve
them from the start, and to identify benefits for
them, which links to the level 4 evaluation
below.
Assessments can be designed around
relevant performance scenarios, and
specific key performance indicators or
criteria.
Online and electronic assessments are
more difficult to incorporate - assessments
tend to be more successful when
integrated within existing management and
coaching protocols.
Self-assessment can be useful, using
carefully designed criteria and
measurements.
4. Results
Results evaluation is the effect on
the business or environment
resulting from the improved
performance of the trainee - it is the
acid test.
Measures would typically be
business or organizational key
performance indicators, such as:
Volumes, values, percentages,
timescales, return on investment,
and other quantifiable aspects of
organizational performance, for
instance; numbers of complaints,
staff turnover, attrition, failures,
wastage, non-compliance, quality
ratings, achievement of standards
and accreditations, growth,
retention, etc.
Individually, results evaluation is not
particularly difficult; across an entire
organization it becomes very much more
challenging, not least because of the reliance
The challenge is to identify which and how on line-management, and the frequency and
relate to the trainee's input and influence.
scale of changing structures, responsibilities
and roles, which complicates the process of
Therefore it is important to identify and
agree accountability and relevance with the attributing clear accountability.
trainee at the start of the training, so they
Also, external factors greatly affect
understand what is to be measured.
organizational and business performance,
which
cloud the true cause of good or poor
This process overlays normal good
results.
management practice - it simply needs
It is possible that many of these measures
are already in place via normal
management systems and reporting.
linking to the training input.
Failure to link to training input type and
timing will greatly reduce the ease by which
results can be attributed to the training.
For senior people particularly, annual
appraisals and ongoing agreement of key
business objectives are integral to
measuring business results derived from
training.