Professional Documents
Culture Documents
Introduction
Over the last decades, the relevance of human resources (HRs) as an essential source of
business value has climbed up the management research agenda (e.g. Barney & Wright,
1998; Delery & Doty, 1996; Huselid & Becker, 2011; Mollick, 2012; Nyberg et al., 2014;
Quinn et al., 1996; Ulrich, 1989, 1997, 2005; Fulmer & Ployhart, 2014). Most of this
research has been guided by the resource-based view of the firm, arguing that sus-
tained competitive advantage derives from the resources and capabilities a firm con-
trols that are valuable, rare, imperfectly imitable and not substitutable (Barney, 1991;
Barney et al., 2001; Wright et al., 1994).
Stefano Baraldi, Full professor in Management Accounting, Universit Cattolica del Sacro Cuore,
Milan, Italy. Email: stefano.baraldi@unicatt.it. Antonella Cifalin, Adjunct professor in Management
Accounting, Universit Cattolica del Sacro Cuore, Milan, Italy. Email: antonella.cifalino@unicatt.it
Theoretical framework
Training evaluation
Starting from the seminal work by Kirkpatrick (1959a, 1959b, 1960a, 1960b), the training
literature includes many contributions on training evaluation based on hierarchical
models (for overviews, see Aragn-Sanchez et al., 2003; Arthur et al., 2003; Salas et al.,
2012; Salas & Cannon-Bowers, 2001; Tannenbaum & Yukl, 1992). Basically, the hierar-
chical approach outlines four categories (levels) to measure training effectiveness: (1)
the reactions of trainees towards TPs, such as enjoyment, perceived usefulness and
perceived difficulty (Warr & Bunce, 1995); (2) their learning, in terms of knowledge,
skill and/or attitude; (3) the degree of learning transfer into on-the-job behavioural
practices; and (4) the impact of this transfer on organizational results. Despite criticisms
regarding the limitations of the hierarchical approach to training evaluation as a com-
prehensive model (Alliger & Janak, 1989; Alliger et al., 1997; Bates, 2004), it has enjoyed
a widespread and enduring popularity in both academic and professional contexts (e.g.
Alvarez et al., 2004; Kennedy et al., 2014; Salas & Cannon-Bowers, 2001).
The most recent criticisms to the hierarchical approach are twofold. A first criticism
is about the deepness of its application in both research and practice. On the one hand,
even though the model suggests an evaluation framework at both individual and
organizational levels of analysis, most empirical research has focused primarily, if not
exclusively, on the former (Aguinis & Kraiger, 2009; Kozlowski et al., 2000; Tharenou
et al., 2007). On the other hand, various studies have demonstrated that training pro-
fessionals have acknowledged the relevance of conducting behaviour-based and
results-based evaluations, yet organizations do not frequently conduct them (for an
overview, see Kennedy et al., 2014). This is alarming because it has been demonstrated
that even though organizations are most likely to evaluate trainee reactions and learn-
ing, only behaviour and results criteria are significantly related to higher rates of
training transfer (Saks & Burke, 2012).
A second criticism is the lack of a strategic perspective when the hierarchical
approach is applied for managerial purposes. Indeed, hierarchical models are primarily
focused on the measurement of training effectiveness, whereas the key issue is not only
measuring the results achieved by TPs, but also managing the continuous alignment
of individual and team behaviour with organizational goals (Brinkerhoff, 2006;
Kirkpatrick & Kirkpatrick, 2005; Kraiger, 2002; Kraiger et al., 2004; Robinson &
Robinson, 1998; Salas et al., 2012; Tannenbaum, 2002).
Research methodology
The study presented in this paper is based on AR methodology. Deeply rooted in
pragmatism, AR originated from the work of Kurt Lewin in the mid-1940s (Lewin,
1946, 1947), was later advocated by many prominent organizational scientists and has
proved its potential in improving the practical pertinence of research (Argyris, 1993).
AR questions the capability of positivist science to cope with the complexity and the
uniqueness of the empirical context in management studies and, thus, to deliver a
general theory. Positivist researchers usually show little inclination to carry out inter-
disciplinary studies and tend to leave the practitioners (and their problems) with no
role other than that of willing subjects of their inquiries (Fendt & Kaminska-Labb,
2011).
Conversely, AR emphasizes the role of a researcher actively involved in some form
of engaged scholarship with the situation being studied (van de Ven & Johnson,
2006). Practitioners and researchers work together to solve practical problems and
generate scientific knowledge. Thus, the validation of the findings does not come from
the consistency of prediction and control, but from the knowledge that can be applied
and validated in action (Gummesson, 1991). Typically, AR studies lead scholars to get
involved in practical organizational work and take action to stimulate change (French
& Bell, 1984). The collaborative research for an applicable theory, its implementation
and the assessment of its impact on the organizational environment allows action
researchers to get back into the realm of academia with an enriched research agenda
and data set (French, 2009). On the one hand, this active role has been criticized
because of its potential lack of rigour. Closely involved with problem solving or change
processes like a consultant, the researcher may lose his/her independence because of
the need to satisfy clients expectations. On the other hand, AR provides the opportu-
nity to obtain very rich insights otherwise not accessible because practitioners are
involved in things which actually matter to them (Eden & Huxham, 1996), and turn out
to be more willing to share their ideas and information.
Moving from this pragmatism, AR has taken an increasingly wide range of forms
inspired by different philosophical stances (Cassel & Johnson, 2006). In our study, we
have followed an approach labelled as inductive AR. As argued by Cassel and Johnson
(2006, p. 793), theory in this particular approach is generated from the data and
concerns the development of thick descriptions of the patterns of subjective meanings
that organizational actors use to make sense of their worlds, rather than entailing the
testing of hypotheses deduced from a priori theory that causally explains what has been
observed by the action researcher. This usually occurs through interpretive under-
standing by seeking to inductively access research participants cultures, in their
natural contexts. As organizational change remains a key issue in this form of AR, the
aim is to reflexively engender single and double loop learning through the involvement
of organizational participants (Argyris & Schon, 1989). The action researcher retains a
pivotal expert role, in providing advice about, and encouraging through processual
interventions, the changes that necessarily need occur as an outcome of this interpre-
tive, yet diagnostic, process (Cassel & Johnson, 2006).
As such, inductive AR is not used to test a well-defined theory, but rather to elaborate
theory from practice (Coghlan, 2004; Jonsson & Lukka, 2006; Kaplan, 1998; Westbrook,
1995) when research aims at: (1) experimenting how different theoretical frameworks
can be actually used in the same organization; (2) evaluating their feasibility; (3) under-
standing the relationships between them; and (4) identifying and building a so-called
emergent theory.
Our study of training performance seems to fit quite well this setting. We have
already experimented both operational and SPM frameworks and evaluated their fea-
sibility at a programme level (Baraldi & Cifalin, 2009), but we have not found any
Participants
The study was carried out in the healthcare sector for several reasons. First, healthcare
organizations (HCOs) are widely recognized as knowledge-intensive organizations
where training is an essential driver of competency development. In developed coun-
tries, including Italy, health professionals are required to attend programmes of con-
tinuing medical education. Tian et al. (2007) offer a systematic review of evaluation in
formal continuing education, pointing out that the effectiveness of TPs in the
healthcare sector is not systematically evaluated, particularly in terms of behavioural
changes in trainees and patient outcomes, and therefore underlined the importance of
developing further research on this subject. Likewise, McAlearney (2010) argues that,
in spite of the huge amount of money spent on training, HCOs are generally struggling
in their attempts to evaluate the return on their investments. Hence, the measurement
of training performance is growing critical in healthcare.
Second, HCOs are experiencing unprecedented pressure on their performance due
to the challenging environment they must cope with, where growing competition,
constraining regulations and resource scarcity require them to state proper goals and to
assign clear responsibilities for results (Lapsley, 2004). Thus, a multi-dimensional
system like the BSC seems to fit with the need to balance the relationships between
costs, quality, access and consumer choices properly (Inamdar et al., 2002). Indeed, the
literature widely reported both the introduction of the BSC fundamentals in HCOs and
its application in different healthcare settings such as hospitals, hospitals systems,
long-term care facilities, psychiatric centres and university departments (e.g. Chan &
Seaman, 2009; Chiang, 2009; Zelman et al., 2003).
Finally, AR is certainly well known by HCOs, and received an additional boost from
the extensive acceptance of the evidence-based medicine principles (Bate, 2000).
As the theoretical framework and the research questions deal with the feasibility of
a strategic approach to training function evaluation, we selected HCOs which explicitly
recognized training as a strategic driver of performance. In particular, we focused on
the scientific institutes for research and care (SIRCs) whose mission institutionally
includes training in addition to research and clinical activity. We excluded public SIRCs
because the literature cites many problems when performance measurement systems
are implemented in public organizations (e.g. Bouckaert & Peters, 2002).
Relying on a systematic method of case selection, we contacted all the 13 private
SIRCs located in Lombardy, an administrative region with almost 10 million inhabit-
ants (more than 15 per cent of the overall Italian population). In the preparatory phase
(Labro & Tuomela, 2003), we sent each HCO a written research proposal detailing the
research objectives and methodology. Three HCOs accepted our proposal. We then
carried out 2-h structured interviews with each training manager in order to verify
whether: (1) the training function really played a strategic role for improving
organizational performance; and (2) the HCO could actually share the purposes of our
research and commit enough time to its delivery.
Research project
We worked for about 12 months with the three HCOs. The first phase of the project
(diagnosing) took place through a plenary start-up meeting in order to: (1) share the
Findings
The training scorecard
During the action planning phase, participants and researchers developed a common
definition of the training scorecard. Following the BSC model (Kaplan & Norton,
2004), each team initially drew a strategy map emphasizing: (1) the link between
organizational and training strategies (i.e. how can the training function actually
contribute to organizational success?); and (2) the few critical issues (so-called key
performance areas, KPAs) underlying the training strategy and their cause-and-
effect relationships (i.e. how can the training function succeed in delivering its
strategy?).
All teams then shared a common version of their strategy maps and recognized its
general validity (see Figure 1). According to Kaplan and Norton (2004, p. 9), a strategy
map is a visual representation of the cause-and-effect relationships among the compo-
nents of a strategy, that is to say the few critical issues (namely, the KPAs) in the
different (but interconnected) perspectives of evaluation. Moving from the top down,
the following perspectives were identified:
an organizational perspective (is training actually contributing to business results?);
a trainee perspective (is training actually improving trainee performance?);
an internal process perspective (is training flawlessly delivered to trainees?); and
a learning and growth perspective (are the drivers for future training performance
properly nurtured?).
For each perspective, participants and researchers identified the few critical issues
(KPAs) to be addressed for a successful implementation of their training strategy,
namely:
Four KPAs were selected in the organizational perspective (i.e. clinical outcomes,
innovation, quality and productivity).
Three KPAs were identified in the trainee perspective (i.e. satisfaction, learning and
behaviour).
Five KPAs were focused in the internal process perspective (i.e. creating consensus,
planning, delivering, evaluating and managing work environment).
Three KPAs were defined, as enabling factors, in the learning and growth perspec-
tive (i.e. competencies, technologies and partnerships).
Organizational perspective
As training functions are internal units providing shared services, they have to be
responsive to organizational strategies (Kaplan & Norton, 2001). Therefore, participants
and researchers felt the need to provide evidence within this perspective of the
organizational-level outcomes actually affected by TPs. HCOs identified four (notable)
types of business results concerning the organizational perspective:
clinical outcomes: increasing the survival of emergency department patients,
reducing hospital infections, controlling patients pain intensity, minimizing the
prevalence and incidence of surgical complications, etc.;
innovation: introducing new medical or surgical procedures, defining new clinical
pathways, re-engineering critical administrative processes, etc.;
quality: diffusing the treatment of hospital pains, applying clinical risk manage-
ment procedures, adopting clinical audits, etc.; and
productivity: improving the productivity of both clinical and administrative
operations, etc.
Trainee perspective
Participants and researchers acknowledged that training is for improving trainee per-
formance to fit the desired organizational goals. Three key performance areas were
identified as critical issues for trainees:
Similarly, the KPA skills in the learning perspective of the programme scorecard
are linked with the KPA learning in the trainee perspective of the functional
scorecard.
There is another linkage between the two KPAs labelled as satisfaction, respec-
tively represented in the learning perspective of the programme scorecard and in
the trainee perspective of the functional scorecard.
Finally, the KPAs in the delivery and design perspectives of the programme score-
card (i.e. time, quality, cost, consensus and training project) flow into the KPAs in
the internal process perspective of the functional scorecard (i.e. creating consensus,
planning, delivering, evaluating, managing work environment).
As a result, the programme scorecard turned out to be a sort of sub-system of the
functional scorecard, focusing on the strategic contribution of a given programme to
the successful implementation of the whole training strategy. Of course, the BSC at the
functional level copes with critical issues that go well beyond programme management
(for instance, developing innovation in terms of training competencies, technologies
and partnerships) and, not surprisingly, reflects a greater level of complexity. As such,
the learning and growth perspective of the functional scorecard is not linked with any
perspective of the programme scorecard.
Second, participants observed that many KPIs used to measure the performances
achieved over the various KPAs and perspectives of the functional scorecard can be
broken down by TPs (see Table 1, drill-down column). Said differently, as a same KPI
can be used to evaluate training at both the functional and programme levels of
analysis, many KPIs collected for evaluating TPs can be easily aggregated and adopted
in the functional scorecard. For instance, the immediate reaction score used to evaluate
the KPA satisfaction in the trainee perspective of the functional scorecard can be
obtained by aggregating the specific immediate reaction score reported for each train-
ing programme.
Finally, participants agreed that, to be effective, the BSC should focus just on a
selection of TPs that may flow into the functional scorecard. Many selection criteria
References
Agostino, D. and Arnaboldi, M. (2011), How the BSC implementation process shapes its
outcome, International Journal of Productivity and Performance Management, 60, 2, 99114.
Aguinis, H. (2013), Performance Management, 3rd edn (Upper Saddle River, NJ: Pearson Prentice
Hall).
Aguinis, H. and Kraiger, K. (2009), Benefits of training and development for individual and
teams, organizations, and society, Annual Review of Psychology, 60, 1, 45174.
Alliger, G. M. and Janak, E. A. (1989), Kirkpatricks levels of training criteria: thirty years later,
Personnel Psychology, 42, 2, 33142.
Alliger, G. M., Tannenbaum, S. I., Bennet, W., Trave, H. and Shotland, A. (1997), A meta-analysis
of the relations among training criteria, Personnel Psychology, 50, 2, 34158.
Alvarez, K., Salas, E. and Garofano, C. M. (2004), An integrated model of training evaluation and
effectiveness, Human Resource Development Review, 3, 4, 385416.
Aragn-Sanchez, A., Barba-Aragn, I. and Sanz-Valle, R. (2003), Effects of training on business
results, International Journal of Human Resource Management, 14, 6, 95680.
Argyris, C. (1993), Knowledge for Action: A Guide to Overcoming Barriers to Organizational Change (San
Francisco, CA: Jossey-Bass).
Argyris, C. and Schon, D. A. (1989), Participatory action research and action science compared,
American Behavioral Scientist, 32, 5, 61223.
Arthur, W., Bennett, W., Edens, P. S. and Bell, S. T. (2003), Effectiveness of training in organiza-
tions: a meta-analysis of design and evaluation features, Journal of Applied Psychology, 88, 2,
23445.
Baldwin, T. T., Ford, K. J. and ad Blume, B. D. (2009), Transfer of training 19882008: an updated
review and agenda for future research, International Review of Industrial and Organizational
Psychology, 24, 4170.
Baraldi, S. and Cifalin, A. (2009), Training programs and performance measurement: evidence
from healthcare organizations, Journal of Human Resource Costing and Accounting, 13, 4, 294
315.
Barney, J. B. (1991), Firm resources and sustained competitive advantage, Journal of Management,
17, 1, 99120.
Barney, J. B. and Wright, P. M. (1998), On becoming a strategic partner: the role of human
resources in gaining competitive advantage, Human Resource Management, 37, 1, 3146.
Barney, J. B., Wright, P. M. and Ketchen, D. J. Jr. (2001), The resource-based view of the firm: ten
years after 1991, Journal of Management, 27, 6, 62541.
Bate, P. (2000), Synthesizing research and practice: using the action research approach in health
care settings, Social Policy and Administration, 34, 4, 47893.
Bates, R. (2004), A critical analysis of evaluation practice: the Kirkpatrick model and the principle
of beneficence, Evaluation and Program Planning, 27, 3, 3417.