Evaluating Learning Factsheet
Evaluating Learning Factsheet
Evaluating Learning Factsheet
Introduction
Most learning and development (L&D) practitioners are concerned about their level of
understanding of the impact of learning. Effective learning and development evaluation
needs to be strongly linked with identified performance gaps. The L&D strategy will
outline the organisation’s evaluation approach and describe how the impact of any
interventions, will be measured.
The CIPD is at the heart of change happening across L&D, supporting practitioners in
providing insights and resources. Connect with us through our Leading in Learning
network.
A learning and development strategy driven by the organisation’s strategic goals and
© Copyright Chartered Institute of Personnel and Development 2021, 151 The Broadway, London SW19 1JQ, UK
Incorporated by Royal Charter, Registered Charity no. 1079797 123 Page 1 of 8
needs is widely recognised as important to business success. To effectively evaluate L&D,
it’s first necessary to have clearly identified organisational performance targets and
subsequent learning needs, and agree what measures of success will look like. Evaluation
covers the impact of learning provision, how that is transferred as well as the engagement
of employees undertaking L&D activities and the engagement of wider stakeholders in the
process.
Our learning cultures research gives advice on evaluating the learning environment
across the whole organisation, with particular teams and at an individual level.
Whilst the majority of organisations carry out some evaluation of learning activities, in
2019 our Professionalising learning and development report showed about a quarter of
respondents struggle to understand L&Ds impact. Our 2021 Learning and skills at work
report showed that one in four respondents don’t systematically evaluate L&D initiatives..
Impact – where L&D can work with the organisation to show how the learning
interventions have impacted on performance – these can include links to key
performance indicators (financial and operational).
Transfer – where L&D can work with the organisation to show how any learning
undertaken on L&D events has been transferred back into the employee’s role and
work area – these can include performance goals and how new skills and knowledge
have been used.
Engagement – where L&D can demonstrate how stakeholders are engaged with
learning, this can be at an organisational level where a positive learning environment
is the goal, at team levels or at an individual level (the ‘happy sheet’ is an individual
reaction to an individual event).
As L&D practice moves from solely offering formal training to embrace ‘social learning’,
where sharing, impact and kudos are less tangible measures, measurement can become
more challenging. See more in our factsheet on evolving L&D practice which also covers
the need to measure value not volume in L&D activities.
© Copyright Chartered Institute of Personnel and Development 2021, 151 The Broadway, London SW19 1JQ, UK
Incorporated by Royal Charter, Registered Charity no. 1079797 123 Page 2 of 8
The seminal model for L&D evaluation, first published in the 1950s by US academic Don
Kirkpatrick remains influential today. However, research conducted by Thalheimer
indicates this model was first introduced by Raymond Katzell.
This was helpful guidance when launched. However, in the 1980s Alliger and Janak found
that the relationships between the levels were weak because each level is not always
linked positively to the next.
Various surveys from the Association for Talent Development have found that most
attention is focused on evaluation of learning at the reactions level because of the
difficulties and time costs of measuring the other three levels. Thalheimer suggests eight
recognised levels of learning evaluation, including some listed above, but he argues that
some of these are highly ineffective.
Firstly, an SCM evaluation involves finding likely ‘success cases’ where individuals or teams
have benefited from the learning. These typically come from a survey, performance
reports, organisational data or the ‘information grapevine’. Those representing potential
‘success cases’ are interviewed and ‘screened’ to find out if they genuinely represent
verifiable success with corroborating evidence from other parties. Factors that contribute
to success beyond the learning intervention are also explored.
Secondly, an SCM evaluation looks at ‘non-success cases’ to discover those who have
found little or no value from the learning. Exploring the reasons why can be very
illuminating.
© Copyright Chartered Institute of Personnel and Development 2021, 151 The Broadway, London SW19 1JQ, UK
Incorporated by Royal Charter, Registered Charity no. 1079797 123 Page 3 of 8
What needs to be improved?
What organisational barriers stand in the way of performance improvement?
SCM should not be seen as comprehensive evaluation method because of the nature of
the sampling, but it offers a manageable, cost-effective approach to determine success
insights and areas for improvement.
The Learning Transfer Evaluation Model is divided into 8 tiers and colour-coded to work as
a kind of barometer using a traffic light system: green shows which methods are most
useful in validating learning results, while red shows those which are inadequate in
measuring learning. Yellow shows those in-between.
Easterby-Smith model
In the mid-1990s Easterby-Smith was able to draw together four main strands for the
purposes of training evaluation:
Philips and Philips built on the Kirkpatrick model by adding return on investment (ROI) as
a fifth level. However, much ROI evaluating is carried out post project and does not build
from a baseline. Another problem is that the arithmetic of ROI means that when a small
cost learning intervention is set against a big project cost, it can look superficially
© Copyright Chartered Institute of Personnel and Development 2021, 151 The Broadway, London SW19 1JQ, UK
Incorporated by Royal Charter, Registered Charity no. 1079797 123 Page 4 of 8
impressive.
Some commentators ask whether a financial model represents the best way to address
the effectiveness of learning. Does stating an ROI of x% help an organisation address its
performance gaps and allow the L&D team to communicate their impact.
Each evaluation model demonstrates a specific approach and some were developed to
assess the value of individual training programmes not an holistic approach to L&Ds
impact on the organisation. Our 2021 Learning and skills at work survey shows that only
one in four respondents make changes based on the evaluation feedback they receive.
L&D teams can be a credible business partner to the organisation when they take time to
use a range of evaluation approaches that are in line with performance data
Relevance: How existing or planned learning provision will meet new opportunities
and challenges for the business.
Alignment: If the L&D strategy takes an integrated blended approach, it’s critical for
L&D practitioners to work with stakeholders about what their performance needs
and how to achieve them. Aligning with broader organisational strategy gives focus,
purpose and relevance to L&D.
© Copyright Chartered Institute of Personnel and Development 2021, 151 The Broadway, London SW19 1JQ, UK
Incorporated by Royal Charter, Registered Charity no. 1079797 123 Page 5 of 8
Measurement: L&D teams effectively and consistently measure the impact,
engagement and transfer of learning activities as part of the evaluation process. It
may be helpful to use a mixture of evaluation methods and broader measures of
expected change and improvement such as return on expectation, and to link L&D
outcomes to key performance indicators.
The RAM approach focuses on the outcome, rather than the response to a learning event
(the focus of the majority of ‘happy sheets’). Our costing and benchmarking L&D factsheet
has further detail on measurement.
The 70:20:10 Institute suggest L&D take on ‘performance roles’. They explore a role for
‘performance detective’ and ‘performance tracker’, where the detective role remit is to
find out where data exists in an organisation, and tracker to provide insights from data
for effective L&D provision in meeting stakeholder needs.
‘Talent analytics’ is about ‘mining’ a whole range of data streams to gain insight into how
people learn and develop. Looking at the way we develop talent and provide future
capability is a challenging area. It provides opportunities for real time evaluation close to
the operational pulse of the organisation and is therefore more likely to be useful as a
decision tool. Essentially, it’s an evidence-based approach to demonstrate value. Read
more in our Talent analytics and big data research report.
Measuring the impact, transfer and engagement of L&D activities can’t be done just by an
end of course questionnaire, knowledge quiz or post-training survey. L&D practitioners
must work closely with stakeholders to agree success criteria for the whole L&D offering.
L&D practitioners also need to work with the organisation to prioritise the available
© Copyright Chartered Institute of Personnel and Development 2021, 151 The Broadway, London SW19 1JQ, UK
Incorporated by Royal Charter, Registered Charity no. 1079797 123 Page 6 of 8
resources.
L&D practitioners need to question the value of traditional happy sheets or quizzes along
with the standard default questions they contain. Is it the learner’s responsibility to ‘rate’
the facilitator or materials? To what degree does a value on a Likert scale apply to learner
reaction to the learning, the engagement of a learner or, arguably the most important
element, the impact on the learner’s performance? Brinkerhof uses an analogy that
measuring the satisfaction of a learning event is akin to predicting the satisfaction and
longevity of a marriage based on the quality of the wedding reception.
Further reading
Books and reports
BEEVERS, K., REA, A. and HAYDEN, D. (2019) Learning and development practice in the
workplace. 4th ed. London: CIPD and Kogan Page.
PARRY-SLATER, M. (2020) The learning and development handbook. London: Kogan Page.
PHILLIPS, J.J. and PHILLIPS, P. (2016) Handbook of training evaluation and measurement
methods. 4th ed. New York: Routledge.
Visit the CIPD and Kogan Page Bookshop to see all our priced publications currently in
print.
Journal articles
BASKA, M. (2019) Majority of L&D professionals feel ‘growing pressure’ to measure impact.
People Management (online). 14 February.
DERVEN, M. (2012) Building a strategic approach to learning evaluation. T+D. Vol 66, No 11,
November. pp54-57.
DIAMANTIDIS, A.D. and CHATZOGLOU, P.D. (2014) Employee post-training behaviour and
performance: evaluating the results of the training process. International Journal of
Training and Development. Vol 18, No 3, September. pp149-170.
FARAGHER, J. (2020) Why is calculating the ROI of L&D like finding a needle in a haystack?
People Management (online). 22 October.
© Copyright Chartered Institute of Personnel and Development 2021, 151 The Broadway, London SW19 1JQ, UK
Incorporated by Royal Charter, Registered Charity no. 1079797 123 Page 7 of 8
MATTOX, J.R. (2012) Measuring the effectiveness of informal learning methodologies. T+D.
Vol 66, No 2, February. pp48-53.
CIPD members can use our online journals to find articles from over 300 journal titles
relevant to HR.
Members and People Management subscribers can see articles on the People
Management website.
© Copyright Chartered Institute of Personnel and Development 2021, 151 The Broadway, London SW19 1JQ, UK
Incorporated by Royal Charter, Registered Charity no. 1079797 123 Page 8 of 8