2015 - Almeida Et Al. - Multicriteria and Multiobjective Models For Risk, Reliability and Maintenance Decision Analysis
2015 - Almeida Et Al. - Multicriteria and Multiobjective Models For Risk, Reliability and Maintenance Decision Analysis
2015 - Almeida Et Al. - Multicriteria and Multiobjective Models For Risk, Reliability and Maintenance Decision Analysis
Multicriteria and
Multiobjective Models
for Risk,
Reliability and Maintenance
Decision Analysis
International Series in Operations Research
& Management Science
Volume 231
Series Editor
Camille C. Price
Stephen F. Austin State University, TX, USA
Multicriteria and
Multiobjective Models
for Risk,
Reliability and Maintenance
Decision Analysis
Adiel Teixeira de Almeida Cristiano Alexandre Virgínio Cavalcante
Universidade Federal de Pernambuco Universidade Federal de Pernambuco
Recife, PE Recife, PE
Brazil Brazil
vii
Preface
Many decision problems have more than one objective that need to be dealt with
simultaneously. Risk, Reliability and Maintenance (RRM) are contexts in which
decision problems with multiple objectives have been on the increase in recent
years. The importance of having a better structured decision process is essential
for the success of any organization. Additionally, decisions on RRM matters may
affect the strategic results of any organization, as well as, human life (e.g. safety)
and the environment.
RRM influences society and organizations in many ways, since companies and
governments must satisfy several expectations related to the everyday lifestyle
inherent in modern society, such as safeguarding the safety of their employees,
their customers and the community they are part of. Such a lifestyle includes new
paradigms for judging what level of risk is acceptable and this requires multi-
dimensional risks to be evaluated in order to meet society’s and regulatory bodies’
expectations. Reliability and maintenance have become more important also,
since such expectations are extended to the demands that services are constantly
available and that products are of a consistently high quality. Therefore,
companies strive to reduce costs and simultaneously improve their performance
with regard to meeting their strategic objectives. These are affected by reliability
and maintenance and include implications for risk, namely that the analysis of risk
and reliability demands a more conservative approach as do maintenance policies
since failures may have serious implications regarding safety and environmental
losses. As a result, MCDM/A approaches are becoming inevitable when modeling
strategic problems that involve the RRM context.
This book integrates multiple criteria concepts and methods for problems
within the RRM context. The concepts and foundations related to RRM are
considered for this integration with multicriteria approaches. In the book, a
general framework for building decision models is presented and this is illustrated
in various chapters by discussing many different decision models related to the
RRM context.
In general, a decision process or problem in the multicriteria context is related
to the acronyms MCDM (Multi-Criteria Decision Making) and MCDA (Multi-
Criteria Decision Aiding; also known as Multi-Criteria Decision Analysis). The
distinctions between these acronyms are not emphasized in this text. Without loss
of generality, the acronym MCDM/A is applied throughout the text to represent a
variety of approaches associated with MCDM and MCDA (decision making,
decision analysis and decision aiding).
The scope of the book is related to ways of how to integrate Applied
Probability and Decision Making. In Applied Probability, this mainly includes:
decision analysis and reliability theory, amongst other topics closely related to risk
analysis and maintenance. In Decision Making, it includes a broad range of topics
ix
x Preface
that this book explores: risk, reliability and maintenance, although they can be
applied to any other context. Therefore, an explanation is given as to why and how
MCDM/A arises in the RRM context.
Chap. 2 deals with MCDM/A methods and the decision process. A procedure
for building an MCDM/A decision model is presented. Some concerns on the
choice of MCDM/A methods are presented, discussing the compensatory and non-
compensatory approaches. Although this procedure may be applied to any context,
some particular considerations are given to the RRM one. A few MCDM/A
methods are presented, the focus being on deterministic additive methods (MAVT)
and methods for aggregation in probabilistic context, with a focus on MAUT.
Outranking methods are also presented, with some emphasis to ELECTRE and
PROMETHEE methods.
Chap. 3 presents concepts of RRM. These concepts should be considered when
building RRM decision models in order to indicate procedures and techniques that
can be used to calculate and estimate consequences. This allows aspects related to
the state of nature and particularities of RRM to be incorporated when modeling a
decision problem. Chap. 3 includes techniques for dealing with risk analysis
such as the HAZOP, FMEA, FTA, ETA, QRA and ALARP principle; cost effective-
ness; and risk visualization. Reliability and maintenance aspects presented
in Chap. 3 include random failure modeling, reliability and failure functions,
maintenance and reliability interactions, FMEA/FMECA, redundant systems,
repairable and non-repairable systems, maintenance goals and maintenance
management techniques (TPM, RCM). Additionally Chap. 3 presents techniques
for eliciting expert’s prior knowledge.
Chaps. 4 to 12 present an integration of the first and second part when con-
sidering RRM decision problems structured within an MCDM/A approach, for
which formulation and insights for decision problems are given. Chap. 4 presents
a multidimensional risk analysis perspective by introducing a general structure for
building a multidimensional risk analysis decision model. Based on the structure
provided, Chap. 4 presents examples of multidimensional risk evaluation models
for natural gas pipelines and an underground electricity distribution system. Other
contexts are discussed, the purpose of which is to offer insights on how to evaluate
multidimensional risks, such as in power electricity systems, for natural hazards,
counter-terrorism and nuclear power.
Preventive maintenance decisions are presented in Chap. 5 with regard to how
to go about selecting which is the most suitable time interval for scheduling
preventive maintenance actions. This chapter explores the classical optimization
approach for preventive maintenance modeling and gives insights on the
implications of considering an MCDM/A approach by discussing illustrative
applications of two kinds of MCDM/A approaches based on the general procedure
for building MCDM/A models presented in Chap. 2.
xii Preface
the first three Chapters. However, Chap. 2 is required, if the reader wants to use
the procedure for building an MCDM/A decision model even though the reader
has good knowledge of MCDM/A concepts.
We would like to thank our colleagues, students and professionals from
industry, who jointly worked with us on modeling MCDM/A problems in the
RRM context, integrated to the Center for Decision Systems and Information
Development (CDSID). In addition, we are grateful to our sponsors (especially
CNPq - the Brazilian Research Council) and the business organizations that have
supported our research and activities since the 1990s. We would also like to thank
the editors of Springer for their professional help and cooperation, and finally, but
most of all, our families, who constantly supported and encouraged us in our
research work.
xxiii
xxiv Acronyms
Abstract: The decision-making process for any organization may be a key factor
for its success. Many decision problems have more than one objective that need to
be dealt with simultaneously. This chapter introduces decision problems with
multiple objectives, with a description of the basic elements needed to build
decision models and focuses on multicriteria methods (MCDM; MCDA;
MCDM/A), in which the DM’s preference structure is considered. An overview
for classification of MCDM/A methods is given, including a discussion on the
DM’s compensatory and non-compensatory rationality and on multi-objective and
multicriteria approaches. The concepts and basic elements of MCDM/A methods
are presented, including preference structures in a multi-attribute context, and
intra-criterion and inter-criteria evaluation. The basic elements of a decision pro-
cess for building decision models and the actors in this process are also presented.
Differences between the descriptive, normative, prescriptive and constructivism
decision approaches are discussed, considering the decision process. Although
these concepts are presented in a general sense, this description deals mainly with
the main context that this book explores: Risk, Reliability and Maintenance
(RRM). Decision problems in a RRM context may affect the strategic results of
any organization, as well as, human life (e.g. safety) and the environment. There-
fore, an explanation is given as to why and how a MCDM/A arises in the RRM
context. In particular, some peculiarities of service producing systems for MCDM/
A models are presented, as well as for goods producing systems.
1.1 Introduction
These quotations are just two of many others on situations reported long ago
which offered this insight of making tradeoffs amongst criteria in order to evaluate
alternatives in the decision process. The optimization conception is in these views,
when considering the attempt to find the best action, is obtained by means of
combining several objectives.
Historical views and perspectives for the MCDM/A area may be found in
several texts (Koksalan et al. 2011; Edwards et al. 2007).
In this Chapter a first view is given on decision problems with multiple
objectives, with a description of the basic elements needed for building decision
models. This Chapter is directly integrated with Chap. 2, which focuses on the
decision process, MCDM/A methods and multiobjective approaches. Then, an
emphasis is given to the main problems and situations found in the context that
this book explores: risk, reliability and maintenance (RRM), although they can be
applied to any other context.
Most of the literature makes a distinction between the terms Multiobjective and
Multicriteria. Therefore, one can say that a problem with multiple objectives can
be approached by using either: MCDM/A method or a multi-objective
optimization approach.
An MCDM/A method considers the preference structure of a decision maker
(DM) and involves value judgment. The DM’s preferences will be incorporated in
the decision model in order to support the choice of the alternative, and by doing
so, the multiple criteria will be analyzed simultaneously.
Multiobjective optimization approaches identify the Pareto frontier, the set of
non-dominated alternatives, from the set of alternatives. An alternative A1 is said
to dominate another alternative A2, if the following conditions hold: i) alternative
A1 is not worse than A2 in all criteria, and ii) alternative A1 is better than A2 in at
least one criterion.
The set of non-dominated alternatives consists of those which are not
dominated by any other of the set of alternatives.
In this approach, the DM’s preferences are not taken into consideration. This
means that a specific final solution is not indicated, since a DM’s preferences are
not incorporated into the model for combining objectives.
On the other hand, using an MCDM/A method, the objectives are combined
based on the DM’s preferences. These preferences consist of the DM’s subjective
evaluation of the criteria. This subjectivity is an inherent part of the problem and
cannot be avoided. Otherwise, it means that the model is related to any other
4 Chapter 1 Multiobjective and Multicriteria Problems and Decision Models
problem, instead of the real problem faced by the DM. Thus, the methodological
issues for dealing with this subjectivity have been one of the main purposes of
research on MCDM/A.
The meaning of models and methods may vary amongst texts. In this text, an
important distinction is made between MCDM/A models and MCDM/A methods,
although slight variations may occur in our discussion because of particular
contexts.
As is well known, a model is a simplification of a real situation and it is
expected to deviate (err) to some extent from the real situation. Therefore, when
building a model there is a conflict between its precision and its simplicity. This
precision is related to how close the model is to the real situation (approximation
of the model).
An MCDM/A model is a formal representation of a real MCDM/A problem
faced by a DM. The MCDM/A model incorporates the DM’s preference structure
and particular issues for a specific decision problem. In general, an MCDM/A
model is built based on an MCDM/A method.
An MCDM/A method consists of a methodological formulation, which can be
applied so as to build specific MCDM/A models. A method may consist of a
theoretical formulation based on a well-defined axiomatic structure.
The MCDM/A method has a more general characteristic and may be applied in
order to build a class of MCDM/A models and may be applicable for a variety of
situations related to preference structures. On the other hand, a decision model
incorporates a preference structure of a specific DM. Some decision models may
be built for a specific and unique problem, while others may be built for a more
general and repetitive decision situation.
The use of the term model may appear to be an exception to the above
concepts, when referring to the ‘additive aggregation model’, which indicates a
group of MCDM/A methods. Here this term is associated with the kind of
mathematical model applied for aggregating the criteria in a particular class of
methods. The additive model for aggregation of criteria will be detailed in Chap. 2,
but it is presented below in (1.1) so as to give a first view of an MCDM/A model.
n
v ( ai ) ¦k V (x )
j 1
j j ij (1.1)
1.4 Decision Process 5
where:
v(ai) is the global value of the alternative ai;
kj is the parameter related to inter-criteria evaluation of criterion j; this parameter
is named either as “weight” or “scale constant” of criterion j;
Vj(xij) is the value of consequence for criterion j;
xij is the consequence or outcome of alternative i for criterion j.
A model for decision process is given by Simon (1960), and consists of three
stages. This model has been adapted, including the addition of new stages, by a
number of posterior contributions, most of them from the area of information
management and decision systems (Bidgoli 1989; Sprague and Watson 1989;
Davis and Olson 1985; Thierauf 1982; Polmerol and Barba-Romero 2000).
Fig. 1.1 shows this updated model. Stages 1 to 3 are in the initial model
proposed by Simon (1960) and consist of Intelligence, Design and Choice. Stages
4 and 5 were added later and are related to revising and implementing the decision
process.
The intelligence stage sets out to search for decision situations, by monitoring
the organization and its environment. This is not a conventional stage for most of
the operational research procedures. In some ways, this stage is related to the view
on structuring a problem given by Keeney (1992) with the Value Focusing Thinking
(VFT) approach, with particular regard to identifying a decision situation. This
stage is also correlated to the vision of strategic management, in which con-
tinuous monitoring and diagnosis of the organization and its environment has to
be done in order to anticipate decision situations in a proactive way (de Almeida
2013).
Conventionally, most operational research procedures consider that there is
already a decision problem to be faced and defining the problem is already part of
working towards finding a solution (Ackoff and Sasieni 1968). Therefore, in most
cases, the decision process starts with the second stage, that of design. This happens
in general in most contexts, especially in the RRM context. However, even in the
RRM context, the organization may derive great benefit by introducing a more
strategic view for dealing with its decision process regarding risk management and
maintenance. For instance, an inadequate maintenance model may affect the
competitive position of any organization, when its clients are adversely affected
by the effects of the unreliability of its products (goods or services).
6 Chapter 1 Multiobjective and Multicriteria Problems and Decision Models
Implementation of solution
Implementation
The main focus of the design stage is on building the decision model. This
stage includes generating alternatives and other ingredients of the decision model.
In this stage the feasibility of alternatives are evaluated. Problem Structuring
Methods (PSM) are very useful in this stage in order to ensure that the problem is
clearly defined (Rosenhead and Mingers 2004; Eden 1988; Eden and Ackermann
2004). The mathematical model is worked out in this stage and the parameters of
the model are estimated. The DM has an important role in this stage, with
particular regard to information given through his/her preferences. Also, it is in the
design stage that the MCDM/A method is chosen.
Therefore, this stage has a basic role in the decision process and the model
designed has to be seen to guarantee that it is related as closely as possible to the
real problem faced. As mentioned a model is an approximation of a real situation.
There is a provocative aphorism about models related to this issue: “All models
are wrong but some are useful” (Box 1979). In other words, the aphorism is saying
that all models are approximations of the real situation. In the practical context of
building models the following recommendation is relevant: “Remember that all
models are wrong; the practical question is: how wrong do they have to be to not
be useful?” (Box and Draper 1987).
1.4 Decision Process 7
In the choice stage the alternatives are evaluated according to the model built in
order to produce a recommendation to the DM. The form of this recommendation
depends on the problematic (Roy 1996), which may be, for instance, a selection of
one of the alternative, ranking all alternatives, etc.
Before the recommendation is presented to the DM a revision stage is
conducted, in order to evaluate the assumptions chosen and results obtained in
earlier stages, and also to check for any possible inconsistencies. In this stage the
model building process is evaluated, and takes a comprehensive view, before final
confirmation that is given that the model is in an adequate state. Also, this stage
incorporates an organizational learning process. Actually, this revision may be
done at any time during this whole process, which may be based on a new
perception about aspects dealt with in earlier steps (Davis and Olson 1985).
The implementation stage consists of applying the recommendation in the
organization or in its environment. Communicating the recommendation is an
important action in this stage.
In the decision process there are several actors who play different kinds of role
in the decision process. The literature presents a few possible views on who these
actors should be (Roy 1996; Vincke 1992; Belton and Stewart 2002; Figueira
et al. 2005; Polmerol and Barba-Romero 2000), some of whom are considered in
what follows. The decision maker (DM) plays the central role, but may be
influenced by other actors. The other actors may include: an analyst, a client,
experts, and stakeholders.
The decision analyst (most of the time simply referred to as ‘analyst’) gives
methodological support to the DM in all stages of the decision process, and works
on the problem structuring process and building the decision model.
The client is an actor who acts on behalf of the DM and interacts most of the
time with the analyst, as a surrogate of the DM. In general this actor is a senior
assistant of the DM, who is not available in many situations; or at least for many
steps in the decision process. Perhaps this use of the term ‘client’ came into being
as this person was seen as someone who sought the guidance of the analyst, who,
in most cases, is an external consultant.
There are other actors, called stakeholders, who try to influence the DM’s
behavior in order to obtain a satisfactory result, for themselves or those whom
they represent. In general, these stakeholders are affected by the decision that will
be made by the DM.
The expert is an actor who has specialized knowledge of some part of the
system, which is object of the decision process and who gives factual information
to be incorporated within the model (de Almeida 2013). This information may be
based on prior probabilities related to the state of nature, which represents
variables not under the DM’s control. This actor may be relevant for decision
problems in the context of RRM, since this requires many probabilistic issues to
be modeled, such as that done in the Bayesian Decision Theory framework (Raiffa
1968). This kind of actor is rarely mentioned in the MCDM/A literature, but is
often present in the literature on Decision Analysis (or Decision Theory).
8 Chapter 1 Multiobjective and Multicriteria Problems and Decision Models
The basic ingredients include the consequences and the set of alternatives.
Concepts related to the family of criteria, the consequence matrix and the
problematic are presented below.
A situation is a decision problem if the DM has at least two alternatives, one of
which he/she must choose. The set of alternatives may be continuous or discrete.
In organizations many managerial decision problems have a set of alternatives
consisting of a discrete set of elements ai, available to the DM. This set may be
represented by A = {a1, a2, a3, ..., an}. A continuous set of alternatives may
be found, in several situations, such as in maintenance planning, in which the
alternatives consist of the time interval tp, within which a preventive maintenance
action should be performed.
In some situations, a continuous set of alternatives may be adapted and
presented as a discrete set of alternatives, when this is an adequate approximation
for the problem. For instance, the time interval for preventive maintenance tp, may
be seen as calendar days, such that the set of alternatives becomes A = {d1, d2, d3,
..., dn}. For any organization, this model is more realistic, since there is no
meaning in considering precisely a continuous time tp, including any time of day
or night. Making a choice of any day di is a reasonable approximation for the
context of preventive maintenance, since a variation in 24 hours does not make
a relevant difference in the consequences related to the decision problem.
The concept of problematic is related to the format of recommendation to be
made for the set of alternatives, which is reflected in the algorithm to be applied
and which will produce the desired result. There are a few types of problematic
found in the literature (Roy 1996; Belton and Stewart 2002) and some of those,
considered the most relevant for this text, are presented below:
x Problematic of choice - In this problematic the result consists of a chosen
subset of alternatives, which should be as small as the procedure can make it.
Normally it is desired to have only one alternative chosen, the optimal one.
This is a particular situation of this problematic, called: optimization. If the sub-
set chosen has more than one alternative, such alternatives are considered
incomparable, since the procedure may not be able to find only one alternative.
Whatever is the size of this subset, only one alternative is implemented in the end.
1.5 Basic Elements and Concepts of Multiobjective and Multicriteria Problems 9
This has been quoted in many texts related to risk management. There are
many decision problems, in which the consequences are presented in a pro-
babilistic way or there is no information on the frequency of occurrence regarding
the elements of the set of consequences. These situations involve decision
problems under risk or under uncertainty.
Given the nature of the multicriteria problem, a vector of consequences is
considered, since each dimension of this vector is related to each criterion.
10 Chapter 1 Multiobjective and Multicriteria Problems and Decision Models
For each alternative i there is a possible consequence Xij, given the criterion j.
Let us assume that the set of alternatives is discrete, then, a consequence matrix
may be considered as illustrated in Table 1.1. This consequence may be
represented by a deterministic or probabilistic variable. The Table 1.1 assumes the
deterministic case, in which there is a specific outcome Xij, for each combination
of alternative and criterion. There are situations in which the consequence may be
presented in a probabilistic way. For instance, for repair time t, the consequence
may be represented by a probability density function f(t) over t.
x Weak Preference (Q) - xQy indicates that there is some doubt if either the DM
clearly prefers the x to y (xPy) or is indifferent between them (xIy), although it
is clear that not(yPx). Property applied: asymmetry.
x Incomparability (J) - xJy indicates that the DM is not able to compare the two
elements. Any of the following situations may apply, but the DM can not
differentiate amongst them: xIy, xPy, yPx. Properties applied: symmetry and
not reflexive (not(xJx)).
A system of preferences or a preference structure is a collection of preference
relations, applied to a set of consequences, such that, the two following conditions
hold:
1. For each pair of elements (x, y) of X, at least one of the preference relations of
the system of preferences is applied to (x, y);
2. For each pair of elements (x, y) of X, if one of those preference relations is
applied, no other may be applied.
Several preference structures are considered for preference modeling studies.
The following preference structures are the ones most applied in practice:
Structure (P,I);
Structure (P,Q,I);
Structure (P,Q,I,J).
Structure (P,I) has a symmetric preference relation (I) and the other relation is
asymmetric. In this structure it is possible to obtain a complete pre-order or a
complete order for the elements of X. In an order there are no ties (no relation I).
A pre-order may have ties (existence of relation I). For a complete order there is
no incomparability. The Structure (P,I) corresponds to the traditional preference
model, with which many MCDM/A methods are associated. For instance, the
additive model for aggregation of criteria, shown in (1.1) is related to this
structure. Let a and b be elements of X, then, the following conditions hold for this
structure:
aPb Ј v(a) > v(b).
aIb Ј v(a) = v(b).
Structure (P,Q,I) has a symmetric preference relation (I) and two asymmetric
relations (P,Q). In this structure it is possible to obtain a complete pre-order for
the elements of X. For this structure, the previous two conditions hold and the
following may be added:
aQb Ј v(a) v(b).
Structure (P,Q,I,J) has the incomparability relation, which leads to partial pre-
orders for the elements of X. This structure is relevant for situations in which the
DM is not able to give full preference information; for instance the DM may not
be able to compare two elements of X. This is not in agreement with one of the
axioms for the model in (1.1), which is the first axiom of Utility Theory, and
states that the DM is able to make a pre-order of all elements of X. This kind of
situation has been pointed out by Roy (1996) and Simon (1955), who emphasizes
12 Chapter 1 Multiobjective and Multicriteria Problems and Decision Models
that this may be relevant for MCDM/A situations, in which the DM has to face
several dimensions in a multicriteria evaluation.
An evaluation of the DM’s preference structure is essential for choosing an
MCDM/A method and for building an MCDM/A model.
An arbitrary adoption of any preference structure with a convenient relation for
elements of X, such as a complete pre-order or order, with no considerations for
the DM’s preference may be considered anti-ethical. A situation in which the DM
has any doubt about applying the preference relation P is not a justification to
assume the indifference relation I. For instance, if the DM declares that he/she is
not able to distinguish whether xPy or yPx, and the analyst assumes that this
means an indifference relation I between x and y, this may be a distortion in the
process. Actually, a few elicitation procedures, for obtaining the preference infor-
mation from the DM, may induce this kind of distortion. In this situation it should
be considered whether indifference or incomparability relation should be applied.
Since the intra-criterion information is available, the following step can be that of
evaluating the inter-criteria, in which all criteria are combined in order to have the
global evaluation of all alternatives. For this evaluation an MCDM/A should be
chosen and applied.
A classification of MCDM/A methods is presented in the next section and a
description of a few methods is given in Chap. 2, but first the concept of a family
of criteria has to be accounted for.
A family F of criteria gj(xj) is the set F = {g1(x1), g2(x2), ..., gm(xm)}. The model
building process should work for a consistent family of criteria (Roy 1996), in
which a few properties has to be followed, such as: being capable of representing
all objectives related to the decision problem and avoiding redundancies.
Since, for each criterion j, the value of the consequences gj(xj) can be produced
for all consequences xj, then, the value of alternatives gj(ai) can be obtained for
each alternative ai.
Given the family of criteria, a dominance relation D between two alternatives
a and b is defined, considering all criteria gj. Then, aDb if gj(a)≥gj(b), given all
j = 1, 2, 3, ..., m, and since the inequality is strict (>) for at least one of the criterion j.
The use of the dominance relation could make the use of an MCDM/A method
unnecessary. However, it is very rare for a solution to be found by applying the
dominance relation. Since, in most situations, many alternatives will not be
dominated by others, then, an MCDM/A method is required in order to evaluate
the inter-criteria.
In Chap. 2 a description of a few MCDM/A methods are given and the
following section gives an overview of their possible classifications.
14 Chapter 1 Multiobjective and Multicriteria Problems and Decision Models
There are four basic decision approaches, which represent perspectives for the
decision process, and which are supported by many methods found in the literature.
These methods may be classified and grouped according to their characteristics.
This grouping process enables common features of such methods to be understood
and facilitates the process of choosing them so as to build particular decision
models. Decision approaches on the other hand will give a perspective on the con-
cepts and the organization of systematic knowledge that supports the decision process.
The literature differentiates amongst a few decision approaches, which are pointed
out as perspectives for the study on the decision process. The literature on decision
analysis considers three approaches: descriptive, normative and prescriptive (Bell
1988; Edwards et al. 2007). The literature on MCDM/A also considers a fourth
perspective to the decision process: constructivism (Roy and Vanderpooten 1996).
The descriptive approach focuses on describing how people decide in a real
situation, the concern being to describe how the DM makes judgments and choices
in decision making. This approach is developed by the area of behavioral decision
making (Edwards et al. 2007).
The normative approach focuses on rational choice, based on normative
models, sustained by an axiomatic framework that aims to ensure a logical structure
for decision making. The model in (1.1) is an example of such a normative model,
which imposes a specific rational procedure which a DM may follow. The utility
theory also provides a rational decision model for decisions under uncertainty.
The prescriptive approach consists of procedures that use a model from the
normative perspective, and are structured to support a DM in the decision process.
The prescriptive approach may use the results obtained in the descriptive
approach, in order to deal with the limitation of human judgment. The errors and
inconsistencies examined in the area of behavioral decision are studied in order to
build procedures that can address a consistent way of interacting with DMs so as
to build the preference modeling process and prescribe appropriate models.
The constructivism approach (Roy and Vanderpooten 1996) consists of an
iterative process that uses a learning paradigm (Bouyssou et al. 2006), in which an
analyst interacts with the DM with the support of some method, in order to
construct the recommendation for the problem that the DM faces.
Whereas the prescriptive approach assumes that the DM has a well-defined
preference structure (for instance a utility function to be elicited), in the
constructive approach there is an interactive process that aims to help the DM
reach a more thorough understanding of his/her preference structure.
1.6 Decision Approaches and Classification of MCDM/A Methods 15
The methods may be also classified according to their form of compensation for
aggregating the criteria, which may be considered a kind of rationality. In this
case, two rationalities may be considered leading to: compensatory and non-
compensatory methods (Roy 1996; Vincke 1992; Figueira et al. 2005). Bouyssou
(1986) made remarks on the concepts related to compensation and non-
compensation.
A number of methods may be included in the first type, for instance: MAUT
for uncertainty situations and MAVT, such as the deterministic additive methods,
including AHP, SMARTS, MACBETH, among many others, embracing basic
elicitation procedures; for instance: tradeoff and swing methods (Figueira et al.
2005; Keeney and Raiffa 1976). For non-compensatory methods, lexicographical
and outranking methods, such as PROMETHEE and ELECTRE are included in
this group.
A preference relation P is non-compensatory if the preference between two
elements x and y only depends on the subset of criteria in favor of x and y
(Fishburn 1976). Let P(x,y) = {j: xjPjyj}. That is, P(x,y) is the collection of criteria
for which xjPjyj. Then:
P ( x, y ) P ( z , w) ½
® ¾ [ xPy zPw] (1.2)
¯ P( y, x ) P( w, z ) ¿
In this case, it does not matter what the level of the performance of x or y in
each criterion is. The only information necessary is if one is higher or lower than
the other.
That is, what the value is of the performance (vi(xij)), in decision matrix, of an
alternative for a particular criterion is not taken into account. It is enough to know
if the level of performance (vi(aj)) of an alternative is higher of lower than another.
That is, the only information needed is if vj(az) > vi(ay).This would mean that the
performance of az is higher than the performance of ay and az is preferred to ay.
This is the only information required in (1.2).
Conversely, for a compensatory relation P, it is not enough to know if the level
of performance (vi(aj)) of an alternative is greater or less than another for criterion j.
For the compensatory inter-criteria evaluation process, it matters what the value is
of the performance (vi(aj)) for that criterion j, since that amount will be con-
sidered, in the aggregation model, as the opposite of a non-compensatory model.
That is, for a compensatory method the disadvantage of one criterion may be
compensated for by the advantage in another criterion, as can be done in the
additive model in (1.1).
As remarked by Bouyssou (1986), a preference relation is compensatory if
there are tradeoffs amongst criteria and it is non-compensatory otherwise.
1.6 Decision Approaches and Classification of MCDM/A Methods 17
the majority of votes cast in that state. At the end of the process, the winner
candidate is the one who gets the states, whose votes sum up to the majority of
weights.
In the presidential election of the USA, the states are equivalent to criteria and
the number of votes obtained in each state corresponds to the score for that
criterion. The combination of criteria, with their weights, plays the role described
for the meaning of the weights in an outranking method (Vincke 1992), which are
combined as a coalition of criteria in order to evaluate the best alternative. The
winner is the one who gets the best coalition of criteria, with the greatest
summation of criteria weights.
It is interesting to note that this non-compensatory rationality means that the
presidential election in the USA is a system with a number of v elections, in which
v = number of states.
The context of risk, reliability and maintenance (RRM) are the focus of this book,
although all the concepts and methodological procedures of MCDM/A are
applicable to any context in general. For this reason a few issues regarding RRM
contexts are discussed below.
In a literature review on MCDM/A models in maintenance and reliability (de
Almeida et al. 2015), more than 180 papers published between 1978 and 2013
were found, which had received more than 4,000 citations. In those studies many
different criteria were found for modeling MCDM/A problems. Amongst the most
common are cost, reliability, availability, time, weight, safety and risk.
Two issues are emphasized in this section regarding MCDM/A models in the
RRM contexts:
x What happens when a decision model does not incorporate the DM’s
preferences;
x The need for MCDM/A models for different kinds of producing systems:
services and goods.
The issue related to whether or not incorporate the DM’s preferences within the
decision model is discussed in the last Section.
There are important issues for MCDM/A models in RRM contexts, which are
related to the peculiarities of two different kinds of producing systems: one for
services and the other for goods, which have different frequencies of demand for
MCDM/A models.
Whatever kind of product it may be, this distinction makes a great difference in
the way that maintenance in general (and preventive maintenance in particular) is
linked to the results of a business. For instance, a system that produces services
1.7 MCDM/A Models in the Context of Risk, Reliability and Maintenance 19
has a feature related to simultaneousness (Slack et al. 2010). This means that at the
time the system is producing the product itself, the customer is being served.
Evidently, in such a context, when a failure in the system occurs, maintenance
definitely has a direct and immediate impact on the competitiveness of the
business (Almeida and Souza 2001). Therefore, preventive maintenance planning
becomes a more strategic decision, linked to highest level of the hierarchical
organizational structure. For the mentioned decision context mentioned above, the
consequences are characterized by multiple and less tangible objectives, which
may require support from an MCDM/A model.
In service systems, the output is produced while the customer is being served.
That is, the main feature of this system is its simultaneousness (Slack et al. 2010).
Therefore, the perception of the quality of the service is being created as the
client/user is being served, unlike in goods systems, in which the quality is linked
to the characteristics of the product itself.
The objectives in service producing systems endeavor to reduce costs, when
considered as part of a mix with other objectives, such as: availability, reliability
of the system, time during which the system is interrupted and the quality of the
service.
In service systems, the interruption of the system can be immediately
perceived, since this affects its users. There are many examples of this kind of
system: energy, telecommunications, health, transport, and other public services
(security, defense, water supply).
For this kind of system, interruptions can lead to serious consequences.
Actually, the domain of such consequences is not well defined when compared to
the goods producing system. Another issue that has to be considered is related to
the actors involved in the process. In the case of the service system, the number of
people who are affected by the interruption may be huge. Also, the degree of
impact may vary widely per person. Moreover, it is extremely difficult for a
business organization to trace the totality of damage caused by the disruption of
this kind of product, which is a service.
All things considered, it is easier to understand that failures in these systems
are not only restricted to the financial dimension, so it is of paramount importance
to have MCDM/A support, in order to provide the DM with a broader view about
the problem, and to give to him/her the tools that best take into account the
preferential aspects related to this multidimensional consequence space.
Furthermore, there is an increasing share of service products in the goods
systems, so that the output of this kind of system turns out to be a combination of
goods and services.
20 Chapter 1 Multiobjective and Multicriteria Problems and Decision Models
In systems that produce goods, losses due to machine downtime can be mitigated
by increasing production beyond normal capacity or by taking some action to
avoid downtime being noticed by clients. In general, failure entails production
delays, re-works, inefficiencies, wastages, overtimes, and/or supply storage
problems, which are easily converted into costs. This would make the problem
change from being one that has multiple objectives to one that has the single
objective of minimizing the total costs. That is why most decision models related
to this context are not based on MCDM/A methods.
However, there are situations, in which, even for systems that produces goods,
the decision context requires an MCDM/A model so that subjective issues can be,
for evaluated. There are two main reasons for this:
x These are more strategic decision contexts which are linked to the highest level
of a hierarchical organizational structure.
x Failures in the production system affect human or social issues, such as safety,
and those to do with the environment.
Moreover, one should be concerned when no DM’s preference is incorporated
into the model, in the modeling process, as subsequently explained.
References
Ackoff RL, Sasinieni MW (1968) Fundamentals of operations research. John Wiley & Sons,
New York, p 455
Bell DE, Raiffa H, Tversky A (1988) Decision making: Descriptive, normative, and prescriptive
interactions. Cambridge, UK: Cambridge University Press.
Belton V, Stewart TJ (2002) Multiple Criteria Decision Analysis. Kluwer Academic Publishers
Bidgoli H (1989) Decision support systems: principles and practice. West Pub. Co.
Bouyssou D (1986) Some remarks on the notion of compensation in MCDM. Eur J Oper Res
26(1):150–160
Bouyssou D, Marchant T, Pirlot M, Tsoukis A, Vincke P (2006) Evaluation and decision models
with multiple criteria: Stepping stones for the analyst. Springer Science & Business Media
Box GEP (1979) Robustness in the strategy of scientific model building. Robustness Stat. pp
201–236
Box GEP, Draper NR (1987) Empirical model-building and response surfaces. John Wiley & Sons
Brans JP, Vincke Ph (1985) A preference ranking organization method: the Promethee method
for multiple criteria decision making, Manage Sci 31:647–656
Davis CB, Olson MH (1985) Management Information Systems: Conceptual Foundations,
Structure and Development. McGraw-Hill
de Almeida AT (2013) Processo de Decisão nas Organizações: Construindo Modelos de Decisão
Multicritério (Decision Process in Organizaions: Building Multicriteria Decision Models),
São Paulo: Editora Atlas
de Almeida AT, Ferreira RJP, Cavalcante CAV (2015) A review of multicriteria and multi-
objective models in maintenance and reliability problems. IMA Journal of Management
Mathematics 26(3):249–271
de Almeida AT, Souza FMC (2001) Gestão da Manutenção: na Direção da Competitividade
(Maintenance Management: Toward Competitiveness) Editora Universitária da UFPE. Recife
de Almeida AT, Vetschera R (2012) A note on scale transformations in the PROMETHEE V
method. Eur J Oper Res 219(1):198–200
de Almeida AT, Vetschera R, de Almeida J (2014) Scaling Issues in Additive Multicriteria
Portfolio Analysis. In: Dargam F, Hernández JE, Zaraté P, et al. (eds) Decis. Support Syst.
III - Impact Decis. Support Syst. Glob. Environ. SE - 12. Springer International Publishing,
pp 131–140
22 Chapter 1 Multiobjective and Multicriteria Problems and Decision Models
2.1 Introduction
In this Chapter two main issues are dealt with. First, considerations are given to
building an MCDM/A model. Then, an overview of MCDM/A methods and
multiobjective optimization approaches are set out.
There are many views for building decision models, since the first propositions
of operational research area. First, some specific issues are emphasized in this
subject in order to establish a basis for the process for building multicriteria
In the process for building models the main focus is on simplicity with a view to
finding a degree of approximation that is good enough to make the model useful.
Therefore, when aiming at making a model useful and simple to use several basic
factors have to be borne in mind.
Bouyssou et al. (2006) point out that the use of formal models evokes the
power of hermeneutics, associated with the facility with which a DM’s
preferences can be elicited. They state that the latter depends on the intellectual
and cultural background of the DM. The analyst should be very cautions with
regard to this issue.
On the other hand, the analyst should spend an additional effort in order to
work on the DM’s interpretation difficulties, which are commonly found in
the interactions for preference modeling. Then again, the analyst should avoid the
temptations of choosing easy approaches, that, although keeping away from these
difficulties, deviates the model from the real problem, which it should be
representing, first and foremost (de Almeida 2013a).
Wallenius (1975) states that DMs in general do not trust models, when they
find them to be complex. Considering the observation from Bouyssou et al.
(2006), it may be plausible that this resistance of a DM rather than that associated
with the complexity, it is caused by the DM’s intellectual background for dealing
with the model. A certain complexity of the model may be acceptable, the better
the DM’s intellectual and cultural background are.
Building models is a creative process in nature, which involves intuition and
other spontaneous actions by the analyst, some of them being inspirations driven
in conjunction with the progression of the model (de Almeida 2013a).
2.2 Building MCDM/A Models 25
In spite of the scientific basis of the models, the building process of which may
follow several well-structured steps, as shown in sequence a, in Fig. 2.1, its
creative side does not recommend a rigorously sequential procedure.
Step 1 Step 1
Step 2 Step 2
Step 3 Step 3
Step n Step n
Sequence a Sequence b
The rigid procedure, in sequence a with consecutive steps, leads to the same
result, if a repetition is made in the process.
A different vision is shown in sequence b of Fig. 2.1, in which the building
process follows a successive refinement procedure (Ackoff and Sasieni 1968). In
this procedure, the analyst can return, at any time, from one step back to any other
previous step, as often as necessary. This return may or not imply the revision of
subsequent steps. This sequence consists of a recursive procedure.
The successive refinement procedure allows any step to be taken in a non
conclusive way, so as to conclude it on returning back, after having a view and
information from subsequent steps. This return makes it possible to enrich the
process with better results for the whole process. Another benefit of this approach
is that the creative modeling process is improved, since this flexibility produces an
environment that is more susceptible to innovative results. In this process the
analyst may get new insights at any time and return to a previous step. In contrast,
the rigid approach of sequence a do not lead the creativity flows for innovation.
26 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
Moreover, it should be emphasized that this flexible and creative process does
not hinder the support of the scientific foundations that any model should have.
Also, the process for building models follows basic scientific patterns in order to
avoid misconceptions.
To build decision models there is a strong support from PSM (Rosenhead and
Mingers 2004), whose methods have become vital to understand decision
problems, thus leading to a much closer connection between these problems and
the models. By using PSM, the analyst has adequate support for organizing
information from the actors of the decision process (Franco et al. 2004).
This link between the “real world” and a “model world” is discussed by Keisler
and Noonan (2012). Fig. 2.2 illustrates these ideas, including an adaptation for
considering Simon’s model for a decision process.
Design the
model
There is a
problem Revision of
model
Action is
implemented Choice
Figure 2.1 shows that in the “real world”, after recognizing the problem, a
decision process is started, by building the decision model in the “model world”,
which will finally produce the implementation of an action. Comparing this view
with Simons’s model, the stages of design, revision and choice are in the “model
world” and the two other stages are in the “real world”. In this view, there is a
possibility of returning back to reformulate the model, after implementing the
action, since this can still provoke the step of problem recognition.
At this point, it should be observed that the model building process may lead to
many possibilities of models, as illustrated in Fig. 2.3. In Fig. 2.3, at the beginning
of the process, many models are possible. The models are represented by the black
spheres. However, during the modeling process, many modeling decisions are
taken, in which assumptions, choices of approaches and simplifications are
introduced, leading to the elimination (filtering) of some possible models.
2.2 Building MCDM/A Models 27
Funnel in the
modeling process,
selecting the model,
at each assumption Final
taken in the process model
MCDM/A
model
options
Filters of
model
selection
The filters in Fig. 2.3 indicate that new assumptions or model definitions are
taken, thus implying the elimination of some possible models, that would prevail
with different assumptions. These modeling decisions also include the preference
modeling information given by the DM. Therefore, during the process parameters
are assigned to the MCDM/A model, thereby reducing the number of alternative
models and leading the process to the final model, as indicated in Fig. 2.3.
A similar illustration with a funnel is given by Slack et al. (1995), for a project
management planning process.
It is interesting to observe that many models may not even be perceived by the
analyst, who eliminates them, by taking directions in the process for building
models. If the analyst has some kind of bias, then, this will be reflected in this
elimination process and perhaps more useful models may not be taken into account.
28 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
There are many propositions or general views presented in the literature for
building models in operational research, particularly when using PSM. A few of
these propositions and views have been made for MCDM/A model building
processes.
Roy (1996) presents a view with several stages for building an MCDM/A
model, which includes: establishing the objectives and format of the recommend-
ations; the analysis of consequences and development of criteria; comprehensive
preference modeling and operational aggregation of performances; investigating
and developing the recommendations.
Polmerol and Barba-Romero (2000) propose a few steps for MCDM/A model
building, including: understanding and acceptance of the decision context;
modeling alternatives and criteria; Discussion and model acceptance, refinements
and evaluation of alternatives with a decision matrix; discussion on the choice of
the method, gathering DM information; application of the method; recommend-
ation and sensitivity analysis. They state that this procedure has a linear sequence,
but can be done in a recursive way.
Belton and Stewart (2002) also present their view with the following steps:
identification of the decision problem; problem structuring; model building; use of
model to inform and challenge thinking; developing an action plan.
Preliminary phase
Finalization
Step 11 Step 12
Step 9 Step 10 Drawing up Implementing
Evaluating Sensitivity recommendations action
alternatives analysis
In the second phase the preference modeling is conducted and the MCDM/A
method is chosen. At the end of this second phase the decision model is ready to
be applied in the third phase, meaning the end of the funnel, illustrated in Fig. 2.3.
The second phase is the most flexible of all of them. In fact the three steps of this
second phase may be done almost at the same time, exploring a richer insightful
process. An already built MCDM/A model is an input to the third phase, although
it still may be changed, due to the possibility of returning to review previous steps
in the successive refinement process.
30 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
In the third phase, the choice and implementation stages of Simon’s model are
conducted, for the final resolution of the problem. However, it should be
remembered that it is still possible to return and make revisions and changes in the
built model. In this phase there is a key step of sensitivity analysis, in which this
revision decision is evaluated.
The following sections present details regarding the conception and implement-
ation of each step of this procedure.
In this step is important to describe and typify the DM and other actors in the
decision process. This procedure has an emphasis on decision problems with an
individual DM, although adaptations may easily be conducted in order to
contemplate the situation with a group of DMs.
In this step it has to be clear what the role of the analyst is going to be and the
DM’s participation should be. For instance, the DM may have a more direct or
indirect involvement in the decision process. For the latter, another actor, often
called the “client” may play some important roles in the process and may be very
active in some of the steps of this procedure.
It is relevant to identify how other actors will take part in the process. It is
important to characterize the role of each actor for each of the steps of this
procedure.
Even for a situation of an individual DM, it is the DM who will decide if
decision process may involve many other actors in some steps of the process in
order to collect insights and a broad view regarding some particular issues to be
included in the model. In this case, the analyst may play the role of a facilitator,
who holds meetings with group of actors for a structured discussion of some
issues. In general these meetings are supported by PSM approaches (Rosenhead
and Mingers 2004; Eden 1988; Eden and Ackermann 2004; Ackermann and Eden
2004; Franco et al. 2004).
This step may be considered the most important one, although this can be only
stated in general terms. The most important step for this kind of decision process
depends on the nature of the problem, which demands special attention to one
these steps of the procedure. It may be the intrinsic nature of the problem may
indicate that a particular step has the greatest influence in the quality of the final
decision model. Therefore, only in general terms, it may be stated that this step is the
most important, since the objectives are going to influence every step in this process.
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 31
The constructed attributes are applied when it is not possible to use natural
attributes. For instance, an objective that is concerned with improving the image
of a business organization, requires such a type of attribute. Whereas the natural
attributes may be used in any decision context, the constructed attributes are
adequate only for a particular decision context, for which they have been built.
These attributes require the construction of a qualitative scale for evaluation of
the associated objective. These attributes normally are on a discrete scale, which
may be called subjective indices or subjective scales. A Table should be drawn up,
indicating the meaning of each level of this scale, in a clear way (Keeney 1992).
This description should indicate one or several impacts on consequences
associated with each level, and specify the degree of achievement of that
objective. It is quite common to reach a situation in which constructed attributes
are necessary.
If the two previous attributes are not feasible, then a proxy attribute may be
tried. This kind of attribute is an indirect measure of the associated objective. In
general, the proxy attribute of a fundamental objective is the natural attribute of a
mean objective that comprises that fundamental objective.
The criteria should have some properties: measurability, operationality,
understandability (Keeney 1992). Measurability defines the objective with more
details, thereby allowing the value judgment, necessary in the decision process.
An attribute is operational if it describes the possible consequences and provides a
common basis for value judgment, and is thus suitable for the intra-criterion
evaluation. This property has a very close relationship with step 7, in which a
return to this step, for refinement, may be necessary, if the criterion is not properly
operational. Understandability means the attribute may not be ambiguous in the
description of the consequences.
The criterion or attribute may be considered in two ways, regarding its
variability and uncertainty: it may be deterministic or probabilistic. A deterministic
criterion is assumed to have a constant level of performance or fixed outcome.
A probabilistic criterion has a consequence x, which is a random variable and is
specified in terms of its probability density function (pdf): f(x). If a criterion is a
random variable, with a not relevant variability it may be assumed to be
deterministic. In this case, it is assumed that the standard deviation is so small,
that the mean of the variable may represent the consequence x.
For instance, let us consider the time for delivering a product. If the criterion is
assumed to be deterministic, then the establishment of the value function, in step
7, will be that of comparing delivering time such as of 2 or 3 hours, for instance.
Another similar decision context, associated with the maintenance of electricity
supply system, may consider the interruption time (t) of the energy supply. It is
not plausible to assume that this kind of criterion is deterministic, since its
variability is very high, thus it is clearly characterized as a random variable t.
Therefore, the DM has to evaluate this criterion, considering its pdf f(t), since, that
is what the DM gets, as a consequence in the decision process.
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 33
consequence 1
f(t)
consequence 2
interruption time t
In this case, the DM does not evaluate the difference in preference between 2 or
3 hours of interruption time (t) in the energy supply, since these two consequences
do not really exist. Actually, the comparison would be between the consequences
or alternatives shown in Fig. 2.5. Which of the two pdfs does the DM prefer? f(t1)
or f(t2)? That is, the DM evaluates the difference in preference between f(t1) and
f(t2), shown in Fig. 2.5, related to the interruption time (t) in hours. This may
appear more complicated, at first, although this is actually what the DM gets in the
end in this kind of decision context. Regarding the complexity of the question to
the DM, it should be pointed out that questions put to the DM in the elicitation
procedures are much simpler.
Many problems in the RRM context have this probabilistic characteristic to be
considered. A literature review on maintenance and reliability points out the
nature of MCDM/A models in this context (de Almeida et al. 2015) and the
plausibility of using deterministic representation for criteria, which is discussed in
Sect. 2.3.15.
Thus, the model building process in this step may include a probabilistic
modeling task for this kind of consequence, which goes together with the
preference modeling.
Regarding uncertainty, a criterion or attribute may be found ambiguous in the
representation of its value function, by the DM, and therefore fuzzy numbers
(Pedrycz et al. 2011) could be used to represent them. In this case, a fuzzy
approach may be considered for the decision model, which may influence the
choice of the MCDM/A method. This should be properly evaluated in step 7.
34 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
This step is related to the set of alternatives for solving the decision problem.
There are four topics to be approached in this step: a) establishing the structure of
the set of alternatives, b) establishing the problematic to be applied to this set,
c) generating the alternatives; and d) establishing the matrix of consequences.
The structure of the set of alternatives has a direct connection with the choice
of the MCDM/A method, since a discrete or continuous set implies completely
different types of methods. For a discrete set of elements ai, A = {a1, a2, a3, ..., an}.
This issue also includes the determination of other features for the set A, which
can be stable or evolutive (Vincke 1992). In the first case, it is known for the
modeling process that the set A is fixed and does not change during the building
process. For the latter, the analyst should be aware of the possibilities of changes
during the decision process, which may represent some kind of constraint.
The set A can be globalized or fragmented. In the former, each element of A
excludes other elements in the resolution process. In contrast, for a fragmented A,
the elements may be combined for the resolution. A portfolio problematic may be
associated to this kind of set. The use of this kind of set is illustrated in Chap. 10.
After establishing the structure of A, then the problematic to be applied to this
set A has to be identified. The problematic may influence the kind of method,
depending on the class of methods to be applied. Some methods may be applied in
more than one problematic; the case of ranking problematic may include the
solution for choice.
After establishment the previous conditions, the generation of the alternatives
can proceed. This is one of the most creative tasks of the whole process.
Analytical insight may be applied in this task, particularly those delineated by the
VFT approach. In this approach the creation of alternatives is based on the value
structure of the objectives. In general PSM can contribute in a considerable way to
this task, and involve a group of experts supported with the guidance of a
facilitator. Depending on the MCDM/A method chosen new alternatives may be
included afterwards, even in an advanced stage of phase three, of finalization.
Some MCDM/A methods assume a fixed set of alternatives and make pairwise
comparisons, for instance. Other MCDM/A methods, build the model and the
preference modeling in a consequence space and may introduce new alternatives
later on.
At this stage, with the criteria and the set of alternatives established the matrix
of consequences can be presented, which consists of the information shown in
Table 1.1. For some problems this matrix can be built very easily, since the
association of alternatives with the corresponding outcome for each criterion can
be made straightforward.
However, for other decision problems this association may not be so
straightforward for some of the criteria. In some cases, the outcome to be achieved
by the alternatives has to be worked out in more complicated procedures.
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 35
The state of nature corresponds to one of the ingredients of decision theory (Raiffa
1968; Berger 1985; Edwards et al. 2007; Goodwin and Wright 2004).
The state of nature consists of factors in the system that are not under the DM’s
control and may change randomly, influencing the outcomes of the decision
process. A variable T may represent the state of nature and may be a discrete or
continuous set of elements.
For instance, in a decision problem related to capital investment, regarding new
technologies or machines in an industrial unit, the alternatives are a discrete set of
elements ai, A = {a1, a2, a3, ..., an}, which is a factor under DM’s control. On the
other hand, the demand for the product is the state of nature Ts in this problem,
which is not under the DM’s control. Depending on the nature of the product, it
may be represented by a discrete set of states of nature, 4 = {T1, T2, T3, ..., Tt},
such as for units of computers. Otherwise, the set of Tis continuous, for instance:
liters of juice.
One should be careful with this ingredient of a decision problem, which is in
some situations may be understood as a consequence and represented as a criterion
within the model. This could be a critical modeling error, and affect the decision
process substantially, including a preference modeling on T. Natural consequences
of this kind of problem may lead to two criteria: C: the total cost of the technology
(considering the purchasing and operational costs); and I: the image of the
enterprise as a confident supplier for its costumers.
This ingredient T is integrated in the model, by its association with the
consequences. A consequence function (Berger 1985) makes this association and
may be represented by P(xŇT, a), which for a probabilistic association, such as in
the example of machine purchase. P(xŇT, a) means the probability of obtaining
the consequence x, given that the state of nature is T and the DM chooses the
alternative a.
For a discrete representation of Ts, considering the consequences C and I,
Table 2.1 shows the decision matrix with the state of nature. In this case, the Ts
may represent different scenarios for demand.
36 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
This is the first step of the second phase of this procedure In this phase the model
is built and the MCDM/A is chosen, although both may be changed by returning
to previous steps.
This step is very connected to the next two steps and all of them are
considerable relevant for choosing the final model, according to the funnel view,
given before.
The preference structure should be evaluated in this step. For instance, the
preference structure (P,I) should be checked with the DM, evaluating if this
structure is appropriate for representing the DM’s preference. If it is, a traditional
aggregation model may be applied, such as the additive model.
However, if (P,I) is not adequate, then, other structures should be checked,
such as the preference structure (P,Q,I,J), in which the incomparability relation is
considered.
The analyst may start this process by checking some basic properties of the
preference structure (P,I), such as the transitivity and if the DM is able to make a
complete pre-order or order in the consequence space. These properties are
essential to the structure (P,I) and can easily be evaluated with the DM, by check-
ing the relations P and I on the consequences. This format is more conceptual than
operational and could be checked as a preliminary procedure, since these
questions in many cases are included in the elicitation procedures of step 8.
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 37
That is, these steps 6, 7 and 8, of the second phase, may be conducted in a very
flexible sequence, even simultaneously and integrated. This process should be
conducted under a non-structured approach, in the sense of this management
information systems concept (Bidgoli 1989; Sprague and Watson 1989; Davis and
Olson 1985; Thierauf 1982). That is, the non-structured approach is due to
the extremely interactive nature of the process, which depends on the DM’s
characteristics and availability. The process is recursive, with many moves
forwards and backwards. This is beyond the view of successive refinement shown
in Fig. 2.1. For instance, the evaluation of relations P and I on the consequences,
at step 6, could be done as an anticipation of the elicitation process of steps 7 and 8.
Also, for some decision contexts, the three steps of this phase may be conducted in
a sequential way, with no repetitions or returns. Considering the nature of the
preference modeling process, everything depends on the DM and decision context.
An important issue to be evaluated in this step is the assessment of rationality
regarding compensation amongst criteria, which can be shown in Fig. 3.6.
Non-compensatory Compensatory
This intra-criterion evaluation consists of eliciting the value function vj(x) (may be
referred to as gi(x)), related to the value of different performances of outcomes in
the criterion j. The information given in the decision matrix should be produced in
this step.
This intra-criterion evaluation depends on the preliminary selection of an
MCDM/A method, in the previous section. On the other hand, the results of this
step may influence a revision on the pre-selection of the kind of MCDM/A
method made in the step 6.
Regarding the influence of the previous step, if a non-compensatory method is
found to be the most appropriate, then, an ordinal evaluation for the consequences
may be enough. Therefore, the intra-criteria evaluation may not necessary, if the
preferences of consequences in each criterion j are already ordered. In such a case,
only a normalization for a common scale may be necessary, which is not often the
case.
For a non-compensatory method, such as an outranking method, the
indifference and preference threshold consists of an intra-criterion evaluation and
is conducted in this step. Also the veto and discordance threshold, commonly part
of the ELECTRE method, are evaluated in this step. It should be observed that an
interval scale may be required, depending on the formulation required for veto and
discordance.
For a compensatory method, such as the unique criterion of synthesis type of
method, a cardinal evaluation of outcomes should be considered and so, an
elicitation procedure should be applied for obtaining the value function vj(x). This
procedure may produce either: linear or non-linear value functions vj(x).
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 39
In this step, the choice of the MCDM/A is made, at the beginning or it may
already have been made. The inter-criteria evaluation in this step leads to the
parameters of the MCDM/A model, involving the elicitation procedure for the
criteria weights. This evaluation depends strongly on the kind of method chosen.
Since the meaning of weights changes for different methods, the elicitation
procedure depends on the method.
Regarding the additive model, the meaning of the weights, normally called
scale constants kj, does not involve only the importance of the criteria and their
elicitation is related to the scales of the value function vj(x) in each criterion.
Actually, there are quite a few MCDM/A methods related to the additive model
for aggregation of criteria, in which the main differences amongst them are related
to the elicitation procedure applied for kj.
For the additive model there are also indirect procedures, in which an inference
is made, based on the DM’s global evaluation of some alternatives. This kind of
method is usually classified as a disaggregation method.
Regarding outranking methods, the elicitation of weights is completely
different from that for compensatory methods. In this case, the meaning of weights
is closely related to the importance of criteria and can be obtained considering this
issue.
In the group of methods classified as interactive methods, in which MOLP
methods are included, the intra-criteria evaluation is worked out by an interactive
process involving dialog with the DM and a system, in general a DSS (Decision
Support System). The DM gives preference information at each dialog action,
which is alternated by computation action by the system. The DM views the
problem by considering the consequence space related to the decision context in
question.
There are also many adaptations of classical elicitation procedures for the
additive model, in which partial information is required, using interactive
procedures.
For probabilistic consequences, using MAUT, there are very well structured
elicitation procedures for obtaining the scale constants for aggregation of the
utility functions of the criteria (Keeney and Raiffa 1976).
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 41
This step concludes the second phase of the process, with two important
results:
x the decision model has been built;
x the MCDM/A method has been chosen.
Now, the third phase is started in order to resolve the problem, recalling that a
return to and revision of previous steps may be made and the model may change.
This is the first step of the third phase of the procedure, the finalization. In this
step the set of alternatives is evaluated, according to the problematic proposed.
The decision model is finally applied.
This step is straightforward and consists basically of applying an algorithm in
the decision model in order to evaluating the set of alternatives.
This step will rarely produce a situation that requires a return to a previous step
and the successive refinement has no place in this step, although this may be
represented in the model as a vague possibility.
The output of this step is still not enough for an evaluation, required for
revision of previous steps. Actually the final result concerning the alternatives has
its final consolidation in the next step.
For each problematic, the following questions (changes in the model output) are
checked:
x For the choice problematic, the output may present alternatives other than those
of step 9, as a solution for the problem. If so, it is desirable to evaluate: how
many alternatives are presented; in which alternatives this happens; and in
which frequency this happens.
x For the ranking problematic, the output may change the position of some
alternatives in the ranking. If so, it is desirable to evaluate: how often this
happens; in which alternatives it happens; and the significance of these
changes.
x For the sorting problematic, the output may present some alternatives in a class
other than that found in step 9. If so, it is desirable to evaluate: how often this
happens; in which alternatives it happens; and the significance of these
changes.
x For the portfolio problematic, the output may present portfolios other than that
of step 9, as a solution. If so, it is desirable to evaluate: how many portfolios
are presented; and in which frequency this happens.
If no changes are observed this indicates that the model is robust for that
particular set of input data. It may happen that a model appears to be robust for a
set of input data and the opposite may happen for another set of input data. It is
important to check the model and its parameters and also the input data.
If changes happen in the model output, then, it is necessary to investigate how
unacceptable this is. Also, the particular input data or parameter that influences
this change is an important piece of information. This may be useful in order to
evaluate if the model should be revised, returning to some previous step. At this
point, it is worthy remembering that there is not a right model; there are useful
models.
The sensitivity analysis may be conducted based on either: analytical analysis
of the mathematical structure of the model or numerical analysis on the model, by
changing the input data. In spite of simplifications of the model, the complexity of
a model may require a numerical analysis.
Many procedures for sensitivity analysis are available in the literature and are
not detailed in this text, the main focus of which is to discuss the role of this
procedure in the model building process. Therefore, for this focus, the following
two kinds of sensitivity analysis are considered:
x for the evaluation of the overall model in a comprehensive process, including
all parameters and input data, at once.
x for a particular analysis of a specific parameter or input data.
The former procedure consists of an evaluation of the overall model, by
changing, simultaneously, a subset or all sets of input data and parameters of the
model. The Monte Carlo simulation procedure may be applied in this case. In this
procedure a random generation of the subset or all set of data is made and applied
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 43
in the model to check the results. This procedure is repeated a number of times
(may be hundreds of thousands of times) in order to compare the frequency at
which the output changes, considering the problematic in question. Other
information to be considered is how significant these changes are, by applying
some statistical hypothesis tests, as demonstrated in Daher and de Almeida (2012).
The changes in each piece of data are established, according to a range around
the nominal value considered for the model and applied in step 9. The range is
specified according to the considerations for assumptions and approximations,
given to that particular piece of data in the modeling process. In general, a
percentage around the nominal value is applied; for instance, plus and minus 30%,
20%, or 10 %. A probability distribution should be applied for the random
generation of data, according to the nature of imprecision observed in the modeling
process; for instance, uniform, triangular, normal probability distributions may be
applied.
This first procedure consists of an overall evaluation of the model and may
indicate whether of not there is a need to continue to the second procedure. The
result of this procedure is included in the recommendation to be given to the DM,
which is worked out in step 11.
The second procedure is very simple to implement and consists of changing the
particular variables of awareness. Each variable is evaluated one at a time, in order
to check its specific impact in the model. This procedure may have an important
managerial role in the process of building the model. During this process a
decision may be made to simplify some step in the procedure. This may be
motivated by the time available being limited or the high costs of collecting
information (preferential or factual data).
These simplifications may be made on the following issues:
x general assumptions for the model;
x the elicitation process, with approximations in parameters; for instance in the
criteria weights;
x assumptions regarding specific analytical structures inside the model;
x using partial information for approximate estimation of input data or model
parameters.
For instance, let us suppose that the elicitation process, in step 8, has
considered approximations in the criteria weights, due to limitations on the DM’s
time. Then, in this procedure, the particular impact of changes in the weights may
be evaluated, in order to check whether or not approximations in the criteria
weights were adequate. If there is no relevant variation, then the simplification in
the model may be considered harmless and the results may be accepted.
Otherwise, an evaluation should be made of the possibility of returning to step 8
and repeating the elicitation process.
The DM may consider that other solutions produced in this step are equivalent
to the nominal solution presented at step 9 and therefore, the results of model may be
accepted. The performance proximity of alternatives may lead to such a situation.
44 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
A second example may be given for input data, such as estimates for the cost of
implementing each alternative. Estimates of costs for implementing projects are
obtained with a high level of approximation in many situations. In this case this
second procedure for the sensitivity analysis may indicate if the impact of such
approximation is relevant and should be reevaluated, by returning to step 3.
The results obtained in this step are as relevant as the solution given in step 9.
The DM should know not only the alternative indicated by the model, but also the
impact of model simplifications on this result.
There is moreover, an insightful consideration of this step for the whole process
for building models. Since the sensitivity analysis can indicate how the model
simplifications can affect the results, then, this possibility may influence decisions
that the analyst will make as to on simplifying the modeling process. That is, the
possibility for successive refinement may indicate that any step, which is cost or
time consuming may be conducted with approximations in a preliminary way, and
is expected to be repeated, after evaluating the impact of these approximations on
the result of step 9.
This may reveal that a rigorous procedure for some steps may be useless, in the
context of a building process for producing a useful model, as a simplification of
the reality. Therefore, the analyst has to be careful when evaluating the DM and
the organizational contexts, when building models.
After the conclusion of the last step, if no return to revise previous steps is
necessary, then, the finalization is approached in this step by analyzing the final
results and producing the report for the DM, with the final recommendations.
The two previous steps produce the main topics to be included in the
recommendations to be given to the DM. Also, the main considerations on
assumptions and simplifications on the model should be included in the report to
the DM.
That is, the DM is not given only the solution indicated in step 9. This is only
part of the recommendation. The DM has to be aware of the simplifications in the
modeling process and its impact in the solution proposed. This kind of report may
be useful for future evaluations regarding results to be achieved by implementing
alternatives.
A good report indicates to the DM the extent to which the solution can be
trusted. The DM should be advised on the nature of the models. The DM should
understand that there is no right model and the usefulness of the model is the main
issue to be evaluated.
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 45
Finally, after the DM has received the recommendation and accepted the proposed
solution, then, its implementation process can start. This may be either: simple and
immediate or complex and time consuming. The latter situation, may require
special attention. Also, the way in which the decision is taken may influence the
implementation process (Brunsson 2007).
A complex implementation process may be as complex as the decision process
and may take much more time to accomplish than the decision process itself. In
such situations, occasionally the implementation process may be conducted by an
actor other than the DM, who may be afraid of changes in the expected outcomes.
For instance, the implementation process for decisions related to public policy
may be so complex and require so much time to be spent on them, that the
complex solution may change in format as time goes by, leading to outcomes that
are different from those expected at the time of the decision process.
Possible changes in the expected outcomes may happen, when the actor
conducting the implementation introduces modifications in the process that may
alter the format of the solution and its expected outcomes. In these cases, the DM
may be concerned with controlling the content of the solution, although in some
cases this cannot be done. The analyst should be aware of this, since this may
influence the DM’s perception on the relations between the consequences and the
alternatives, if the latter may be changed, during the implementation.
There is another issue of time, which is related to the time at which the
implementation process should be started. That is, the deadline for starting the
process may be considerable, compared with the time for the decision process.
This may appear to be controversial, since the time given for producing the
recommendation may be short, thus leading to a stressed model building process,
and at the end, a longer time is available before starting the implementation. This
may be required, when the organization needs to announce the decision made and
there is still some time available before initiating the action.
In this situation, a procrastination process may be introduced in this step. The
procrastination process consists of introducing and managing a delay before
implementing the solution, so that a re-evaluation of the decision may take place.
The procrastination (Partnoy 2012) takes place under the allegation that it is more
important to take the correct action, than to take it sooner. In this case, it would be
wise to procrastinate, taking time to think over the chosen solution. This thinking
time may allow the decision made to be revised and thereby to gain new insights
from the whole process already conducted.
For some situations, managing this delay is more important than other steps of
the decision process. This may suggest the introduction of a sub-step in step 12, in
which the implementation delay is managed.
46 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
Phase 3 (Finalization)
Step 9
Model built
Step 10
Solution
chosen
Step 11
Recommendation
Step 12
Deadline for controlling the
procrastination process
greater or lesser than other, but by how much cannot be measured. In such a case,
the additive model in (1.1) may not be applied. Therefore, the scales impose
constraints for the kind of method to be applied.
Familiarity with these scales and their associated normalization procedures
(Polmerol and Barba-Romero 2000; Munda 2008) are important issues for dealing
with MCDM/A problems.
First of all, two kinds of scales may be considered: a) a numerical scale; and
b) a verbal scale. Amongst the numerical scales the following are the main interest
of this text: ratio scale, interval scale, and ordinal scale.
The ordinal scale is the one that has a minimal degree of information. In this
scale the numbers only represent the order to be assigned to the elements in a set.
They do not have cardinality in the sense that one can say 4 is twice as much as 2.
Basic arithmetic operations, such as summation, are not allowed when using this
scale. If a decision problem is presented in such a way that some of the criteria are
presented in the ordinal scale, then an ordinal method should be applied. A careful
application of another method is possible, considering an approximation, in which
case one should be careful, when drawing conclusions from the results.
Many verbal and numerical scales are applied for outcomes of criteria,
represented by subjective scales, which in the end present information that is only
consistent with an ordinal scale. Actually, most pieces of information collected
from a DM, by subjective evaluation, using a verbal or numerical scale, are not
consistent with a cardinal scale, unless, an adequate procedure is applied to ensure
that they are.
The ratio scale is the scale with the greatest degree of information. As
suggested by the name, in this scale the cardinality is in the ratio between two
numbers. For instance, the weight of an object is presented on this scale. This
means that 4 kg is twice as much as 2 kg. The ratio scale has unity and e origin,
represented by the zero of the scale, which means absence of property. That is, 0
kg means absence of weight. In this scale a transformation of the following type
may be done and the scale properties are maintained: y = ax, with a > 0. In this
transformation the origin is kept and the unity is changed. That is what happens
when the weight scale is changed between kg and g. Length and the time are other
examples of ratio scales.
In the interval scale, the cardinality is in the interval between two outcomes. In
this scale, the following linear transformation may be applied, keeping the
properties of the interval scale: y = ax + b, with a > 0. In this transformation the
unity and the origin are changed, respectively by a and b. In this scale the zero
does not have the same meaning as in the ratio scale. The zero means just the
minimum value of the scale (as is usual in MCDM/A problems). Temperature is
an example of an interval scale. In this scale, considering the Celsius scale, one
cannot say that 40oC is twice as much than 20oC. On the other hand, one can say
that passing from 30oC to 10oC is twice as much as passing from 40oC to 30oC.
The above linear transformation may be applied for temperature, so that on
changing from Celsius (x) to Fahrenheit (y), one can apply y = (9/5)x + 32).
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 49
Verbal scales are applied in many MCDM/A problems and can be transformed
into a numerical scale in order to be incorporated into a decision model. This scale
may be ordinal or cardinal (ratio or interval), depending on the elicitation
procedure applied. However, a simple process of asking a DM to declare a verbal
scale for a set of consequences in most cases will produce an ordinal evaluation. A
verbal scale that is very often applied is the Likert scale (Likert 1932), in which
the number of levels for evaluation is limited to five (there are variations, such as
a four-level scale), due to the limited human cognitive capacity for making
evaluation in a scale of many levels, such as a ten-level scale, from 1 to 10, which
is often applied inadequately.
The type of scale for the consequences of a criterion, as represented in the
consequence matrix, causes constraints for choosing an MCDM/A method. Also,
the type of scale for a value function vj(x), shown in the decision matrix is chosen
according to the necessary degree of information required and the kind of
transformation to be done.
An interval scale is applied in many MCDM/A methods, such as in Utility
Theory, and it is in its axiomatic structure. This scale presents a piece of
information which has a particular relevance for comparing two alternatives. It
shows how much performance is added from one alternative to another. In many
situations the DM wants to know, how much is added to go from one position to
another. Of course the ratio scale also has interval cardinality and, therefore, gives
the same information as the interval scale.
The interval or ratio scale are both applied for methods, such as those based on
the additive model in (1.1). The interval scale includes an additional feature that
may lead it to be the scale preferred by many of those methods, based on the
additive model. In this scale, the minimum value (xmin) of an outcome for a
criterion j is set to be zero, so that the value of vj(xmin) = 0. Since the maximum
(xmax) outcome is set to be 1, so that the value of vj(xmax) = 1, in this scale the range
(xmax - xmin) is reduced to a minimum, for the scale 0 to 1. In contrast, for the ratio
scale the range (xmax - 0) is higher, for this scale of 0 to 1. This makes the interval
scale more precise for estimating subjective values in the preference modeling
process.
There is a specific situation for the model in (1.1), in which the interval scale is
not adequate. When using MCDM/A in the portfolio problematic, the interval
scale may not be applied, since it induces a wrong solution due to a size effect
caused by this scale. In this case a ratio scale should be applied (de Almeida et al.
2014). For other MCDM/A methods similar situations occur (de Almeida and
Vetschera 2012) and additional procedures should be implemented.
If the value functions vj(x) obtained in the intra-criterion evaluation are linear,
then, the information produced in the decision matrix can be obtained by a
normalization procedure. It should be observed that the term normalization in
MCDM/A does not have the same meaning as it has in statistical procedures of
normalization.
50 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
This section deals with a few specific issues for building MCDM/A models, such
as psychological traps, the choice of the method, compensation of criteria, and the
intelligence stage of Simon’s model.
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 51
Psychological Traps
There are some psychological traps, discussed in the behavioral decision making
literature that can affect the quality of the information obtained from the DM,
during the elicitation procedures for preference modeling. This is relevant, since
the DM’s preferences to be included in the model are items of subjective-based
information. Simon (1982) discusses the limitation on rationality that people in
general have.
A few of these psychological traps are briefly presented below (Hammond
et al. 1998a):
x Anchoring - People tend to give a strong weight to information received
(impressions, estimates, data) just before making any subjective evaluation.
This should be considered in the way that preference questions are put to the
DM or factual questions to an expert.
x Status Quo - There is a tendency of choosing actions that maintain the Status
Quo. This may lead to confirm and repeat past decisions.
x Estimating and Forecasting - In general people are skilled at making estimates
about time, distance, etc, in a deterministic way. However, making these
estimates considering uncertainty is different. On the other hand, DMs usually
have to make such kinds of estimates for their decisions.
x Overconfidence - DMs tend to be overconfident about their own accuracy, thus
naturally guiding them to errors of judgment in preference elicitation procedures.
This is one of the traps that affect the DM’s ability to assess probabilities
adequately.
With regard to the estimating and forecasting trap, Hammond et al (1998a)
state that DMs rarely get clear feedback about the accuracy of those estimates they
have to make. The feature of successive refinement in the decision procedure
described above can minimize this situation, combined with the results of the
sensitivity analysis, although this does not improve the accuracy for future estimates.
The way in which questions are put forward to the DM may induce errs, in any
of these traps. For instance, the more choices the elicitation procedure gives to the
DM, the more chance there is that the status quo will be chosen (Hammond et al.
1998a).
Suggestions to deal with these difficulties are given by Hammond et al
(1998a).They also present other psychological traps, which include: confirming
evidence, framing, and prudence.
In the literature there are not many studies dealing with the choice of a proper
MCDM/A method for a decision problem. However, this seems to be changing.
The concern with the matching between the method and problem has increased and
52 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
are given on how to deal with this. Indeed, there is still much research work to be
conducted on the evaluation of the DM’s willingness to make compensations,
even though this is an extremely relevant factor for the choice of methods.
The foregoing procedure for building an MCDM/A model does not include the
intelligence stage of Simon’s model for the decision process (Simon 1960). This pro-
cedure assumes that there is already a problem that has been identified at the start
of the design stage of Simon’s model. Fig 2.8 shows how this intelligence stage can
be integrated with the procedure described above for building a decision model.
This intelligence stage requires a continuous monitoring process on the status
of the organization or the decision context, in which attention to the decision
process is established, and also its external environment.
Is there any
External environment opportunity or
treats no
yes
Method Percentage
Pareto Front 48.39
MAUT 10.22
AHP 9.68
MACBETH or other MAVT 8.60
Goal Programming 3.23
ELECTRE 2.69
PROMETHEE 2.15
TOPSIS 1.08
2.3 A Procedure for Resolving Problems and Building Multicriteria Models 55
With regard to the risk context, there is a variety of concepts in the literature on
risk and also on its perception (Chap. 3 deals with this topic). Some of them
consider only the probability for a specific context. However, if a decision is being
made then the consequences should be considered. Also, the model should
incorporate the DM’s preferences over these consequences. In fact, a ‘decision
process’, in which the DM’s preference is not considered is not a process in which
a decision is actually being made, as discussed at the end of Chap. 1.
According to Cox (2012), the application of utility functions rather than simple
risk formulas – consisting of terms such as exposure, probability and consequence -
allows a DM’s risk attitudes to take into account, thereby improving the effective-
ness of the decision making process to reduce risks. Cox (2009) discusses many
issues related to the decision process in the risk context, including the limitations
of risk assessment using risk matrices and a normative decision framework.
Another classical problem within the risk context is the direct association
between the quality of a decision and the actual consequence obtained at the end.
In fact, at the time in which the decision is being made, the DM cannot assure the
best consequence, since there are uncertainties in the process. Therefore, only
expectations can be evaluated when making the decision. In general, this is
something difficult for many DMs to understood and the analyst should be aware
of how to deal with this by clarifying all these issues to the DM, instead of using
inadequate models for simplifying what is going to be shown. These clarifications
should be made in step 11, when drawing up the recommendations to the DM.
There are many concerns in the literature with regard to interpreting the scores for
the alternatives given by utility functions. This concern is extended in general to
56 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
any MCDM/A method that gives final scores for alternatives, thus representing a
global evaluation, based on the aggregation of multiple criteria. However, these
numbers can be interpreted according to the properties of the scale, for each
particular method, in order to compare alternatives.
If the method uses a ratio scale, it is relatively easy to produce a comparison of
alternatives, considering the ratio of their scores. For instance, in a choice
problematic, a first alternative may be twice as good as the second one, or it could
be 20% better than the second one.
Even for a specific scale, such as the ratio scale, the meaning of this ratio may
be explained, by taking the rationality behind the method into account. For
instance, in the PROMETHE II method, the scores are based on the summation of
criteria weights, within a non-compensatory rationality.
With regard to the interval scale, which is applied for the utility function of
many of the MAVT methods, the alternatives may be compared based on the
properties of this scale.
The interval scale allows an incremental comparison between alternatives. That
is, the differences of the scores of the alternatives are considered. However, a ratio
may also be considered between two differences, as shown in Chap. 4 (see Equations
(4.13) and (4.14)). Therefore, a difference ratio DR may be applied to interpret
v ( a p ) v ( a p 1 )
the values in relation to the alternatives, so that: DR , in
v ( a p 1 ) v ( a p 2 )
which p represents the position in the ranking obtained by alternative ap and v(ap)
represents the score of the alternative. By analyzing these DR results, the DM
can perceive the distance between the pairs of alternatives. This is illustrated in
Table 2.3.
Table 2.3 presents the position of the alternatives in the second column, their
scores in the third column and their comparisons by the increments of the scores in
the fourth column. The fifth column shows the DR, from which it can be observed
2.4 Multicriteria Decision Methods 57
that the increment of the scores from A1 to A5 is 6.50 times greater than that from
A3 to A1.
Another possible way to explain these results to the DM is to consider the ratio
of differences between two alternatives and the whole range, given by the range
between the best and the worst scores. This difference can be expressed as a
percentage of the whole range. That is, in Table 2.3, the whole range is v(A2)-
v(A6)=0.70-0.10=0.60. Therefore, the difference in scores between alternatives A5
and A1 is 22% of the whole range, while for alternatives A1 and A3 it is 3%.
Applications of these indices are given in Chap 4. The analyst may use any one
of these indices, after evaluating which of them is the most appropriate for a given
DM to understand.
With regard to the use the expected utility function for models in the risk context,
there are a few paradoxes with which the analyst should be aware of. These
paradoxes have been analyzed by behavioral decision making studies, in the
descriptive perspective context.
There are other approaches that deal with some particular situations, such as
Rank-Dependent Utility (RDU) and Prospective Theory (Edwards et al. 2007;
Wakker 2010).
There are many situations regarding risk which cannot be easily integrated into
decision models. The kind of event known as a ‘black swan’, related to the so
called ‘black swan theory’ may be an example of such a situation. This event is
related to a kind of occurrence that is very unexpected (very low probability), with
very undesirable consequences. These are rare events, which result in a great
damage. In general, their evaluation is not well accepted in the expected value
principle, since the multiplication of the value of such great damage is excessively
reduced by the value of an extremely low probability.
On the other hand, although many concerns with the use of the expected utility
function are clamorously announced in part of the literature, the analyst should be
aware that in many situations these behavioral issues do not matter for many
practical problems. It is necessary to understand their meaning and to evaluate
them when they are relevant. Unfortunately, in many situations, these matters are
inappropriately announced in order to justify other less adequate approaches.
First the methods related to unique criterion of synthesis are presented, then
some outranking methods are introduced. Interactive methods, related to MOLP
are very briefly mentioned, since most of the problems in the RRM contexts are
non-linear problems. The next section deals with heuristics and evolutionary
multiobjective algorithms for dealing with multiobjective models.
This is one of the most applied models for aggregating criteria and it is usually
classified as MAVT (Belton and Stewart 2002), being part of the group of
methods of unique criterion of synthesis. MAVT is distinguished because it
considers deterministic consequences, whereas MAUT (see next subsection) deals
with probabilistic consequences (Keeney and Raiffa 1976).
The additive model, also called a weighted sum model, is recalled from (1.1)
and reintroduced below for prompt reference in (2.1), in which the global value
(v(xi)) is considered for a consequence vector xi = (xi1, xi2, ..., xin), for the
alternative i, which is the same as the global value v(ai) for alternative ai, as
indicated in (1.1).
n
v ( xi ) ¦k v (x )
j 1
j j ij (2.1)
where:
kj is the scale constant (weights) for attribute or criterion j.
vj(xij) is the value of consequence for criterion j, for the alternative i.
xij is the consequence or outcome of alternative i for criterion j.
The scale constant is usually normalized as follows:
¦k
j 1
j 1. (2.2)
The additive model has a few properties that should be checked before making a
decision on its application. For practical modeling purposes the main properties
are briefly described.
This model follows the preference structure (P,I), in which it is possible to
obtain a complete pre-order or a complete order. For two consequences xz and
xy, the following conditions hold for this structure: a) xyPxz Ј v(xy) > v(xz);
2.4 Multicriteria Decision Methods 59
b) xyIxz Ј v(xy) = v(xz). Therefore, one of the assumptions of this model is that the
DM is able to compare all consequences and order them. Also the transitivity
property holds for the preference relation R, whether it is P or I, so that for three
consequences xw, xy and xz, if xwRxy and xyRxz Ј xwRxz.
Another property of this model is the mutual preference independence
condition amongst the criteria (Keeney and Raiffa 1976). Let Y and Z be two
criteria, the preference independence between Y and Z occurs if and only if the
conditional preference in the Y space (intra-criteria evaluation given, different
levels of y, such as y' and y''), given a certain level of z = z', does not depend on
the level of z. That is, (y',z')P(y'',z') Њ (y',z)P(y'',z), for all z, y' and y''.
This property may be formally presented in the following formulation (Vincke
1992). Let a, b, c and d be four vector of consequences in a consequence space
with two criteria Y and Z. Then, Y and Z are preferentially independent if the
following condition holds: If for criterion Y, vy(a) = vy(b), and vy(c) = vy(d), and
for criterion Z, vz(a) = vz(c), and vz(b) = vz(d), then, aPb Ј cPd. This is illustrated
in Fig. 2.9.
b d
vz(xz)
a c
vz(xz)
vy(xy) vy(xy) Y
Therefore, the validation of this model should be done by confirming that the
DM’s preference structure is according to these properties. In some practical
situations a DM may refuse to follow the final recommendation based on this kind
of model, when a violation of one of these properties occurs and there alternatives
close to the solution, in which it is obvious for a global evaluation that a property
is violated. The DM may not be able to perceive which property is being violated,
in such cases, but can recognize the inconsistency of the final result. Although, the
DM can distinguish this kind of inconsistency only in an obvious situation, this
shows that this may not be an issue to be ignored.
60 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
There are many elicitation procedures in the literature for the elicitation of the
scale constants (Weber and Borcherding 1993). Amongst these are the tradeoff
and the swing procedures which are described below.
The tradeoff procedure is presented in detail by Keeney and Raiffa (1976).
Weber and Borcherding (1993) consider that this is the procedure with the
strongest theoretical foundation.
This procedure is classified as an indirect procedure (Weber and Borcherding
1993), since the determination of the scale constants is based on inference from
information given by the DM. It is also classified as an algebraic procedure, since
it calculates the n scale constants from a set of n-1 judgments often using a simple
system of equations, which also includes (2.1).
This procedure is based on a sequence of structured questions (Keeney and
Raiffa 1976) put to the DM, in order to obtain preference information, based on
choices between two consequences. A first group of questions obtains the ordering
of the scale constants, then, other questions prepare the DM to understand better
the consequence space and finally, the DM makes choices between pairs of
consequences related to neighboring criteria, in order to make the tradeoffs for the
equations for the algebraic process.
Thus, the procedure is based on the DM making a comparison on two
consequences xb = (x1, x2, ..., xj, ..., xn), which is a vector with the consequences xj
for each criterion j. These consequences have the best outcome bj, for one of the
2.4 Multicriteria Decision Methods 61
criteria and the worst outcome wj for the other criteria. For instance, x2 = (w1, b2,
..., wj, ..., wn) has the best outcome for the criterion j = 2, and x3 = (w1, w2, b3, ...,
wj, ..., wn) has the best outcome for j = 3. If the DM’s preference is such that
x3Px2, then, v(x3) > v(x2). Based on (2.1), the value of v(xb) = kb, since v(bj) = 1
and v(wj) = 0. Therefore, if x3Px2, then, k3 > k2. Using these kinds of questions,
the order of the scale constants is obtained.
Next, another pair of consequences is compared in order to find indifference
between them, by decreasing the value of the outcome bj for criterion j which is
the preferred one. For instance, for x3Px2, the consequence b3, of x3, has the
outcome decreased to the level of x3, such that x3'Ix2, in which x3' = (w1, w2, x3, ...,
wj, ..., wn). If the DM can specify the outcome x3', such that x3'Ix2, then, v(x3') =
v(x2). Since, v(xb) = kb and v(xb') = kbvb(xb), by applying (2.1), this leads to k3v3(x3)
= k2. This equation is related to one of the n-1 judgments for the system of
equations necessary in this procedure, in order to obtain all the scale constants kj.
A critical judgment in this procedure is adjustment of the outcome in order to
obtain the indifference between the two consequences above (Weber and
Borcherding 1993).
The swing procedure is included in the SMARTS method (Edwards and Barron
1994). This procedure is classified as an algebraic procedure and also as a direct
procedure (Weber and Borcherding 1993), since the determination of the scale
constants are based on direct information given by the DM, taking the range of the
consequences into consideration.
This procedure is also based on a sequence of structured question (Edwards and
Barron 1994). The first question considers the following consequence w = (w1, w2,
..., wj, ..., wn)., in which all criteria have the worst outcome. Then, the DM is asked
to choose one of the j criterion to improve the outcome of wj to the best outcome
bj. That is, the DM may choose a criterion to ‘swing’ from the worst to the best
outcome. This indicates criterion j for which the scale constant kj has the greatest
value. Then, the DM is asked to choose the next criterion, and so on. At the end
the scale constants of the criteria are ordered. Then, in another step, the criterion
with the largest value of scale constant is arbitrarily assigned 100 points. The other
criteria are assigned points expressed as percentages of the criterion with the
largest scale constant value, considering their range. Finally, these percentages are
normalized to produce the final scale constants.
There are quite a few methods incorporating the additive model. The main
difference amongst them is in the elicitation procedures of the parameters,
including both the intra-criterion and inter-criteria evaluations, with emphasis on
the scale constants.
In many situations the use of the additive model is straightforward with the use
of one of the classical elicitation procedures, there being no explicit consideration
of an MCDM/A method. In other cases, an MCDM/A method is considered.
One of the most applied methods that incorporates the additive model is
SMARTS (Simple Multi-Attribute Rating Technique with Swing), in which the
swing procedure is applied (Edwards and Barron 1994). SMARTER (Simple
Multi-Attribute Rating Technique Exploiting Ranks) is a related method that
applies the first step of ordering the scale constants of the criteria and then, uses a
surrogate weight. In these methods the value function for each criterion is assumed to
be linear (Edwards and Barron 1994).
The AHP (Analytic Hierarchy Process) presents a particular procedure for
preference modeling, considering the possibility of a hierarchical structure of
objectives (Saaty 1980). The method uses the additive aggregation model, and
2.4 Multicriteria Decision Methods 63
Additive-Veto Model
The use of additive models for the portfolio problematic demands some concerns
with the scale to be applied, since there is a size effect that causes the wrong
64 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
solution to be selected in the interval scale, which is the one most applied for
elicitation procedures (de Almeida et al. 2014).
The portfolio problematic in the additive model is based on the selection of a
portfolio pr that maximizes the value V pr as given in (2.3).
m § n ·
V pr ¦ ¨¨ x ¦ k v (a ) ¸¸
i 1
i
j 1
j j i (2.3)
© ¹
m
subject to some constraints, such as a budget constraint of ¦x c d B .
i 1
i i
where:
pr >a1, ... , a m @ is the portfolio, which is a vector with the items (projects) ai.
1 if the item (project) x i is included in the portfolio
xi ® .
¯0 if the item (project) x i is not included in the portfolio
C represents the vector of item costs, C >c1 , c2 , ! , cm @ T .
B is the budget or the limit for total cost C.
For portfolio selection, based on additive models, as in (2.3), the interval scale
may not be applied. It has an impact on the result due to the size effect of the
portfolio in this kind of scale, thus causing the wrong portfolio to be selected.
What has been proved to be most appropriate is the ratio scale for this kind of
problem (de Almeida et al. 2014). Most weight elicitation procedures are based on
the interval scale that sets the worst outcome to zero, whereas using a ratio scale
for the portfolio selection, the weights to be applied with the scale should be
changed. The transformation of these scales can be seen at de Almeida et al.
(2014).
weights. SMARTER (Edwards and Barron 1994) uses this idea, based on the
partial information of the order of criteria weights. Another procedure (Danielson
et al, 2014) increases the precision for surrogate weights by adding numerically
imprecise cardinal information into rank-order methods, such as the ROC (Rank
Order Centroid), also applied in SMARTER.
Other approaches collect more information and use procedures based on
decision rules, formulating linear programming problems (LPP) or simulation
procedures in order to analyze the alternatives. Among these approaches are:
PAIRS (Salo and Hämäläinen, 1992), which uses interval judgments; VIP
Analysis (Dias and Climaco, 2000), based on the progressive reduction of the
number of alternatives; PRIME (Salo; Hämäläinen, 2001) which uses preference
information based on swing method or holistic information; and RICH (Salo and
Punkka, 2005) which uses incomplete ordinal preference statements. Mustajoki
and Hamalainen (2005) integrate preference elicitation in the partial information
framework for the SMART/SWING method.
A flexible elicitation procedure adapts the tradeoff elicitation procedure by
using partial information in an interactive way, and conducts analysis by means of
a set of LPPs (de Almeida 2014a; de Almeida 2014b).
2.4.2 MAUT
MAUT has been developed for MCDM/A problems, from Utility Theory (von
Neumann and Morgenstern 1944), keeping its axiomatic structure (Keeney and
Raiffa 1976). According to Edwards and Barron (1994), Howard Raiffa presented
the fundamental insight for MAUT in 1968, pointing out that there would be more
than one reason to value an object. Raiffa (1968) presented a few considerations
for a multicriteria view in the context of health problems.
This approach gives one of the most classical MCDM/A methods, in which the
most widely applied aggregation approach has been the additive model, for which
the axiomatic structure of the theory indicates a number of properties to be
considered. As mentioned, the main difference from the MAUT additive model to
the model in the previous section is that the probabilistic consequence is
approached in the utility function uj(xj) for each criterion j.
The decision models with MAUT may include the framework of Decision
Theory (Raiffa 1968; Berger 1985; Edwards et al. 2007), also called as Decision
Analysis, which may consider the Bayesian approach to dealing with uncertainties,
incorporating prior probabilities. Therefore, the uncertainties on the state of nature
(T) may be obtained from experts, in the form of prior probabilities S(T). Thus,
T is an additional ingredient to be considered with MAUT, although this may not
be explicit in some models.
For each Ts chosen by nature and each action ai chosen by the DM, a con-
sequence x may be obtained, according to a consequence function (Berger 1985)
66 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
Consequence Space
The whole evaluation process for the utility function is made over the
consequence space, with which the DM should be familiar. Fig. 2.10 shows the
consequence space for two criteria x and y.
2.4 Multicriteria Decision Methods 67
y
(x0,y*) (x*,y*)
*
y
y1 (x*,y1)
(x0,y0) (x*,y0)
0
y
x0 x* x
In the consequence space shown in Fig. 2.10, for each criterion, the most
desirable outcomes are x* and y*, while the least desirable outcomes are x0 and y0.
For the whole space the points (x*,y*) and (x0,y0) represented respectively the most
and least desirable outcomes for the multi-attribute space. The scale for the utility
is arbitrarily set in the interval 0 to 1, so that u(x*,y*) = 1, u(x0,y0) = 0, uj(x*) = 1,
uj(y*) = 1, uj(x0) = 0 and uj(y0) = 0.
The utility function uj(xj) for each criterion j, related to the intra-criterion
evaluation, is assessed considering a conditional utility function of criterion j,
which is conditioned to a fixed level of the outcomes in other criteria. For
instance, on the x axis of Fig. 2.10, there is a conditional utility function of
criterion x, given a fixed level of the outcome for criterion y = y0.
The intra-criterion evaluation consists of eliciting this single dimensional utility
function uj(xj). There are several procedures for this elicitation (Raiffa 1968;
Keeney and Raiffa 1976; Berger 1985), many of them use the concept of certain
equivalent of a lottery. This certain equivalent is the consequence B for which
there is a probability p, such that the DM is indifferent between B and a lottery
[A, p; C, 1–p], with consequences A and C.
In general, the consequences of the lottery are the least and the most desirable,
so that the probability p = u(B). Since, u(x*) = 1 and u(x0) = 0, and considering
the indifference between B and [x*,p;x0,1–p], then u(B)=pu(x*)+(1-p)u(x0). From
this, it follows that u(B) = p.
Therefore, the elicitation procedure consists of obtaining the indifference
between this kind of lottery and the consequences x, so that the utility function
68 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
u(x) can be obtained. Detailed elicitation procedures are provided in Keeney and
Raiffa (1976).
n
u( x) ¦k u ( x )
j 1
j j j (2.4)
where:
kj is the scale constant for attribute or criterion j;
uj(xj) is the utility function for criterion j;
xj is the consequence or outcome for criterion j.
The scale constant kj is usually normalized as in (2.1).
If the mutual utility independence condition is found between x and y, in the DM’s
preference structure, then the multilinear model, u(x,y)=kxu(x)+kyu(y)+kxyu(x)u(y),
may be applied. Similar to (2.4), a generalization may be made for a model with n
criteria.
whatever the value of p is, is the same for any other lottery [(x*,y),p;(x0,y),1–p],
whatever the outcome for y is.
It is interesting to note that for the strategically equivalent utility function
u(x,y0), a utility u(x,y) may be found by a linear transformation, such as
u(x,y)=a(y)u(x,y0)+b(y), where: a(y)>0 and b(y)>0 are constants, established for
any outcome for y.
Therefore, as shown with this utility independent condition, the utility function
u(x,y) depends only on the particular level of the outcome in criterion y, even so,
by a linear transformation. More details on this concept are given by Keeney and
Raiffa (1976).
B C
y2
A D
y1
y0
x0 x1 x2 x
It can be seen that u(x,y0) and u(x0,y) can be obtained based on the scale
constants kj, such that: u(x,y0)=kxux(x) and u(x0,y)=kyuy(y). This leads to the format
of (2.4). This concept and its development are given in detail by Keeney and
Raiffa (1976).
A complete and detailed procedure for the elicitation of the MAU function is
given by Keeney and Raiffa (1976). The elicitation of the scale constants kj is
based on the analytical model obtained, associated with the independence
conditions.
For instance, the scale constants kj for the additive model on the two criteria x
and y correspond to the utility of the two specific consequences (x*,y0) and (x0,y*),
shown in Fig. 2.10. That is, kx=u(x*,y0) and ky=u(x0,y*).
Therefore, the elicitation of kx consists of finding the probability p for which
(x*,y0) is the certain equivalent to the lottery [(x*,y*),p;(x0,y0),1–p]. A similar
evaluation may be made for ky.
Again, as can be seen the scale constants kj for an MAU function are not simply
the relative degree of importance of the criterion. They are related to the scale,
considering the limits for x and y, since the lottery [(x*,y*),p;(x0,y0),1–p] is the
basis for their elicitation.
There are quite a few paradoxes related to the use of the expected utility function,
which are presented in the literature. Many of these paradoxes have been analyzed
in a descriptive perspective within the context of behavioral decision making.
In many situations Rank-Dependent Utility (RDU) and Prospective Theory
(Edwards et al. 2007) have been considered as ways of dealing with such
situations (Wakker 2010).
MCDM/A models based on MAUT may be adapted with Rank-Dependent
Utility and Prospective Theory views on modeling risk preferences, which may
have particular relevance for the RRM context.
This kind of method has a completely different rationality from the methods in the
two previous subsections. These methods are non-compensatory and may be
applied to a preference structure (P,Q,I,J). The possibility of the incomparability
2.4 Multicriteria Decision Methods 71
relation is one of the issues distinguished in this kind of method, and therefore,
only partial pre-orders may be obtained.
Therefore, unlike MAVT and MAUT, this kind of method may be applied in a
situation for which the DM’s preferences are not in agreement with the first two
properties. That is, the DM is not able to compare all consequences and order
them. Also, the transitivity property may not be followed.
This section presents some basic elements of these methods and then,
introduces an overview on the two most widely applied outranking methods:
ELECTRE and PROMETEE.
These methods are based on pairwise comparison of the alternatives, by
exploring an outranking relation between the pairs of alternatives.
There is an important difference between outranking methods and those of
MAVT and MAUT that impacts the preference modeling process, namely the
different meaning for the inter-criteria parameters, which may be called weights.
The meaning of criteria weights corresponds directly to the degree of importance
of the criteria, for outranking methods.
This notion of importance amongst criteria may be compared with votes in a
voting process (Roy 1996; Vincke 1992). Let there be two subsets of criteria G
and H and two alternatives a and b. If the subset of criteria in G is more important
(has more votes) than the criteria in the subset in H, and the following conditions
hold (Vincke 1992):
x a is better than b for all criteria in the subset G;
x b is better than a for all criteria in the subset H; and
x a and b are indifferent for any other criteria;
Then: a is globally better than b.
If this importance (or votes) can be represented by the criteria weights, the
comparison between the subsets of criteria G and H can be based on the
summation of these weights.
That is, the summation of weights for criteria in favor of a is greater than those
in favor of b. This means that a makes a better coalition of criteria than b.
These methods are worked out in two main steps (Roy 1996; Vincke 1992):
x Building the outranking relation, by comparing all pair of alternatives in the set
of alternatives;
x Exploiting the outranking relation by applying an algorithm or procedure for
solving the problem, according to each particular problematic.
These methods may work with different kinds of criteria, depending on their
intra-criterion characteristics. In a true criterion there is no threshold. For a pseudo
criterion there are thresholds that may be one of the following or both: an
indifference threshold and a preference threshold.
The outranking relation S, is applied over all pairs of alternatives of the set of
alternatives, such as a and b. Therefore, aSb means that alternative a outranks
alternative b, which means that a is at least as good as b.
72 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
ELECTRE Methods
(2.5)
C ( a , b) ¦ wj
j:g j ( a )t g j ( b )
where:
wj is the weight for criterion j; the weights are normalized, such that 1.
¦w
j
j
gj(a) and gj(b) is the value of the outcome for criterion j, respectively for
alternatives a and b.
2.4 Multicriteria Decision Methods 73
There are a few different formulations for the discordance index (Roy 1996;
Vincke 1992; Belton and Stewart 2002). D(a,b) may be given by (2.6):
§ g j (b) g j ( a ) ·
D ( a , b) max¨ ¸ , j | g j (b) ! g j (a ); j, c, d . (2.6)
¨ max[ g j (c ) g j (d )] ¸
© ¹
Having obtained these formulations and parameters, this step for building of
the outranking relation can be finalized, by applying (2.7) for all pair of
alternatives. It may happen with a pair of alternatives that aSb and bSa. In this
case there is a circuit and these alternatives are considered indifferent.
The second step of exploiting the outranking relation can now be worked out.
For the ELECTRE I method, the purpose of this step is to obtain the kernel, which
is the subset of alternatives, in which each of its elements is not outranked by any
other in the kernel. If only one alternative is found in the kernel, the choice
problematic reaches its particular case of optimization. Otherwise, the alternatives
in the kernel have been found to be incomparable.
More details on ELECTRE methods may be found in many basic texts on
MCDM/A methods (Roy 1996; Vincke 1992; Belton and Stewart 2002; Figueira
et al. 2005).
PROMETHEE Methods
n
S a, b ¦ w F ( a , b)
j 1
j j (2.8)
where:
wj is the weight for criterion j; the weights are normalized, such that 1.
¦w
j
j
I(a) = ¦S a, b
bA
(2.9)
where, n-1 gives a normalized scale between 0 and 1, since n is the number of
criteria.
The ingoing flow I(a) indicates the disadvantage of the alternative a compared
with all other alternatives b in the set of alternatives A. I(a) is obtained from
(2.10):
I(a) = ¦S b, a .
bA
(2.10)
2.4 Multicriteria Decision Methods 75
Another index for evaluation of the alternatives is the liquid flow I(a), given by
(2.11), which is obtained in a scale of -1 to 1.
I(a)=I(a)-I(a) (2.11)
Now the second step of exploiting the outranking relation, may be concluded
by using these indices on specific procedures for each problematic.
In the PROMETHEE I method two pre-orders are built, based on (2.9) and
(2.10), which indicate the relations of preference (P), indifference (I) and
incomparability (J) between the pairs of alternatives of set A (Brans and Vincke
1985; Belton and Stewart 2002). Therefore, PROMETHEE I outputs a partial pre-
order of the elements of A.
The PROMETHEE II method is based on the liquid flow I(a) from (2.11), in
which each alternative has a score. Therefore, PROMETHEE II outputs a
complete pre-order on the elements of A.
The family of PROMETHEE methods includes other methods: PROMETHEE
III and IV, for a stochastic situation; PROMETHEE V for a portfolio problematic,
as discussed in the following sub-section; and PROMETHEE VI, when the DM
specifies a range for each criterion weight, instead of a precise value of weight.
The PROMETHEE V method (Brans and Mareschal 1992) is applied for selecting
portfolios using a non-compensatory method for evaluating of alternatives in a
model similar to that in (2.3). The only difference is in computing the value of the
portfolio V pr , which is based on the application of PROMETHEE II for scoring
the items ai (projects) as values vi(ai).
There is also a problem of scale with this method, although different from that
with the additive model. In this case PROMETHEE II presents positive and
negative scores for vi(ai) = I(ai) to be applied in (2.3). Therefore, to work in the
maximization model, the negative scores have to be transformed into positive
scores, thereby changing the properties of the ratio scale (Vetschera and de
Almeida 2012).
This transformation has a similar effect, with the possibility of selecting the
wrong portfolio. Contrary to the case of the additive model in (2.3), the ratio scale
cannot be applied in the PROMETHEE V. In order to overcome this problem, an
analysis should be conducted based on the concept of a c-optimal portfolio
(Vetschera and de Almeida 2012; de Almeida and Vetschera 2012).
76 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
There are other approaches and concepts that may be seen either as specific
methods or tools that can be applied in any method, such as those presented above.
Belton and Stewart (2002) consider the latter option for fuzzy sets and rough sets.
A comprehensive view of fuzzy approaches for modeling MCDM/A problems is
given by Pedrycz et al. (2011), while the rough sets approach is briefly described
in the next subsection.
There are a few approaches classified as disaggregation methods, which are
based on holistic (or global) evaluation by the DM, followed by a subsequent step
of inference of the parameters of an aggregation model. Pardalos et al. (1995)
consider these approaches as a fourth group of methods in their classification.
Some of these approaches, such as the UTA method (Jacquet-Lagréze and
Siskos 1982), are related to the single criterion of synthesis methods. However,
inference procedures proposed for the ELECTE TRI method use the same process
of collecting information from the DM on global evaluation for posterior inference
of the parameters of inter-criteria evaluation. The preference learning approach
(Slowinski et al. 2012) uses a similar process.
Rough Sets
Articulation
MCDA methods
of preferences
Global Criterion Method (Osyczka 1984); Goal Programming (Charnes and
Cooper 1961); Goal-Attainment Method (Chen and Liu 1994); Lexicographic
A priori
Method (Rao 1984); Min-Max Optimization (Osyczka 1984); Surrogate Worth
Trade-Off (Haimes et al. 1975).
A posteriori Weighted Sum; İ-constraint Method (Miettinen 1999).
STEP Method (Benayoun et al. 1971); (SEMOPS) Sequential Multiobjective
Progressive
Problem Solving Method (Duckstein et al. 1975)
Table 2.5 First and second generations of multiobjective evolutionary algorithms (MOEAs)
In many decision processes there is more than one DM. In such situations a group
decision model or a negotiation process has to be applied in order to come to a
final solution. Therefore, a brief overview is given of Group Decision and
Negotiation (GDN) methods and processes, particularly of those aspects most
closely related to MCDM/A models. The GDN area covers decision problems
with multiple DMs, over a wide range of topics such as: Conflict Analysis (Fraser
and Hipel 1984; Keith et al. 1993; Kilgour and Keith 2005), web-based
negotiation support systems (Kersten and Noronha 1999), evolutionary systems
design (Shakun M.F 1988), connectedness (Shakun 2010), formal consciousness
(Shakun 2006) and fair division (Brams and Taylor 1996).
As stated by Kilgour and Eden (2010) negotiation and group decision contain
both unity and diversity. Regarding the latter, some of the scholars in the field of
GDN understand that it is appropriate to distinguish between Group Decision
(GD) making and negotiation. Kilgour and Eden (2010) explain that in this view
GD making is related to a decision problem shared by more than one DM, who
must make a choice, for which all DMs will have some responsibility. On the
other hand, a negotiation is seen as a process in which two or more DMs, acting in
an independent way, may either: make a collective choice, or not do so. For the
latter, one (or more) of the DMs may give up taking further part in the decision
process and walk away.
Additionally, it can be considered that a GD process involves an analytical
procedure in order to aggregate the preferences of the individual DMs, which
results in a kind of collective representation of the preferences of the group. With
regard to negotiation, this involves a process of interaction between DMs, in order
to find a collective solution for the problem of their mutual interest.
As to using the analytical procedure in order to aggregate the DMs’
preferences, the process for building models pays great attention to following rules
of rationality, related to a normative perspective. Also, there are some concerns
about dealing with some paradoxes, as shown by the descriptive perspective. As
for the negotiation process, the interaction between people invokes other concerns,
such as the accuracy of their communication process.
These issues show some diversity between GD making and a negotiation
process. However, there are some elements of unity between them. For instance,
most negotiation processes are grounded in analytical results and endeavor to
ensure the rationality and fairness of the collective choice. Also, the building
process for the GD model involves agreements with the group of DMs, regarding
several issues and parameters of the model, especially when the problem also
involves multiple objectives, leading to an integrated MCDM/A and GD model.
Therefore, in order to build GD models, some interaction processes may be
necessary between the DMs. The process for building GD models will depend on
the available time of these DMs, and most of all, on how simultaneously their
80 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
subject, since they can learn something new from other experts. That is why many
studies are focused on searching for consensus regarding the experts’ perceptions
of that particular topic.
Regardless of these differences on these two kinds of aggregation, some
models are built in order to tackle these two issues mutually, since both are
present in many GDN problems.
Many fuzzy approaches are applied to this kind of problem (Ekel et al. 2008;
Pedrycz et al. 2011), and deal with factors such as ambiguity and uncertainties the
experts have as to describing their perception on the variables that are being
evaluated.
There is a particular kind of situation related to experts’ aggregation of
probabilities, which is related to prior probabilities S(T) on the state of nature T.
There are many studies in the literature on Decision Theory (or Decision Analysis)
related to the elicitation of prior probabilities (Raiffa 1968; Berger 1985) and the
aggregation of a group of experts’ prior probabilities (Edwards et al. 2007). At the
end of Chap. 3 there are more details about this topic.
The following subsection gives a brief description of types of group decision
aggregations regarding DMs’ preferences.
Regarding the aggregation of DMs, different actors may play specific roles. For
instance, instead of an analyst, a facilitator or a mediator may act in some
situations. For instance, a facilitator may act so as to intensify the interaction
process between DMs or among other actors in the decision process. With regard
to DMs, the way in which they act and are available for the interaction in the
decision process, for a particular problem, plays an important role when
classifying the types of GD aggregation.
The GD aggregation process consists of reducing the set of individual DMs’
preferences to a collective DMs’ preference. There are some situations in which
one of the actors in the GD process is a supra-DM. This supra-DM makes decision
on final issues, in general, related to global evaluations in the process, such as
evaluating the other DMs’ choices. The supra-DM may have a hierarchical
position above the other DMs in the organization’s structure. Keeney (1976)
considers two types of GD process, with regard to DMs’ interrelationships: the
‘benevolent dictator problem’ and the ‘participatory group problem’. The former
is related to the situation regarding a supra-DM and in the latter, the group acts
jointly in the GD process, with the same power.
Whether or not a supra-DM is present in the process, two kinds of GD
aggregation general procedures may be considered (Kim and Ahn 1999; Leyva-
López and Fernández-González 2003; Dias and Clímaco 2005):
82 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
GD procedure for
aggregating DMs’ GD procedure for aggregating
initial preferences DMs’ individual choices
References
Ackermann F, Eden C (2001) SODA - journey making and mapping in practice. In: Rosenhead J,
Mingers J (eds) Rational Analysis in a Problematic World Revisited, 2nd ed. John Wiley &
Sons Inc., United Kingdom, pp 43–61
Ackoff RL, Sasinieni MW (1968) Fundamentals of operations research. John Wiley & Sons,
New York, p 455
Bana e Costa C, De Corte J-M, Vansnick J-C (2005) On the Mathematical Foundation of
MACBETH. Mult. Criteria Decis. Anal. State Art Surv. SE - 10. Springer New York, pp
409–437
Belton V, Stewart TJ (2002) Multiple Criteria Decision Analysis. Kluwer Academic Publishers
Benayoun R, de Montgolfier J, Tergny J, Laritchev O (1971) Linear programming with multiple
objective functions: Step method (stem). Math Program 1(1):366–375
Berger JO (1985) Statistical decision theory and Bayesian analysis. Springer Science & Business
Media, New York
Bidgoli H (1989) Decision support systems: principles and practice. West Pub. Co.
Borcherding K, Eppel T, von Winterfeldt D (1991) Comparison of Weighting Judgments in
Multiattribute Utility Measurement. Manage Sci 37:1603–1619
Bouyssou D (1986) Some remarks on the notion of compensation in MCDM. Eur J Oper Res
26(1):150–160
Bouyssou D, Marchant T, Pirlot M, Tsoukis A, Vincke P (2006) Evaluation and decision models
with multiple criteria: Stepping stones for the analyst. Springer Science & Business Media
Brams SJ, Taylor AD (1996) Fair Division: from cake-cutting to dispute resolution. Cambridge
University Press, New York
Brans JP, Mareschal B (1992) PROMETHEE V: MCDM Problems with Segmentation
Constraints. INFOR 30(2):85-96
Brans JP, Vincke Ph (1985) A preference ranking organization method: the Promethee method
for multiple criteria decision making, Manage Sci 31:647–656
Brunsson N (2007) The consequences of decision-making. Oxford University Press New York,
NY
Cailloux O, Mayag B, Meyer P, Mousseau V (2013) Operational tools to build a multicriteria
territorial risk scale with multiple stakeholders. Reliab Eng Syst Saf 120:88–97
Charnes A, Cooper WW (1961) Management Models and Industrial Applications of Linear
Programming. John Wiley & Sons
Chen Y-L, Liu C-C (1994) Multiobjective VAr planning using the goal-attainment method. IEE
Proc. - Gener. Transm. Distrib. IET, pp 227–232
Coello CAC, Toscano Pulido G (2001) A Micro-Genetic Algorithm for Multiobjective
Optimization. In: Zitzler E, Deb K, Thiele L, Coello Coello CA, Corne D (eds) First
International Conference on Evolutionary Multi-Criterion Optimization: 126-140. Springer-
Verlag. Lecture Notes in Computer Science No. 1993
Coello CC, Lamont GB, Van Veldhuizen DA (2007) Evolutionary algorithms for solving multi-
objective problems. Springer Science & Business Media
Corne DW, Jerram NR, Knowles JD, Oates MJ (2001) PESA-II: Region based Selection in
Evolutionary Multiobjective Optimization. In: Spector L, Goodman ED, Wu A, Langdon
WB, Voigt HM, Gen M, Sen S, Dorigo M, Pezeshk S, Garzon MH, Burke E (eds)
Proceedings of the Genetic and Evolutionary Computation Conference (GECCO’2001): 283-
290, San Francisco, California, Morgan Kaufmann Publishers
Cox LA Jr (2009) Risk analysis of complex and uncertain systems. Springer Science & Business
Media, New York
Cox LA Jr (2012) Evaluating and Improving Risk Formulas for Allocating Limited Budgets to
Expensive Risk-Reduction Opportunities. Risk Anal 32(7):1244–1252
84 Chapter 2 Multiobjective and Multicriteria Decision Processes and Methods
Daher S, de Almeida A (2012) The Use of Ranking Veto Concept to Mitigate the Compensatory
Effects of Additive Aggregation in Group Decisions on a Water Utility Automation
Investment. Group Decis Negot 21(2):185–204
Danielson M, Ekenberg L, Larsson A, Riabacke M (2014) Weighting under ambiguous
preferences and imprecise differences. Int J Comput Int Sys 7(1):105-112
Davis CB, Olson MH (1985) Management Information Systems: Conceptual Foundations,
Structure and Development. McGraw-Hill
de Almeida A, Vetschera R, de Almeida J (2014) Scaling Issues in Additive Multicriteria
Portfolio Analysis. In: Dargam F, Hernández JE, Zaraté P, et al. (eds) Decis. Support Syst. III -
Impact Decis. Support Syst. Glob. Environ. SE - 12. Springer International Publishing, pp
131–140
de Almeida AT (2013a) Processo de Decisão nas Organizações: Construindo Modelos de
Decisão Multicritério (Decision Process in Organizaions: Building Multicriteria Decision
Models), São Paulo: Editora Atlas
de Almeida AT (2013b) Additive-veto models for choice and ranking multicriteria decision
problems. Asia-Pacific J Oper Res 30(6):1-20
de Almeida AT, Almeida JA, Costa, APCS, ALmeida-Filho AT (2014b) A New Method for
Evaluation of Criteria Weights in Additive Models by Interactive Flexible Elicitation.
Working paper, CDSID
de Almeida AT, Costa, APCS, Almeida JA, Almeida-Filho AT (2014a) A DSS for Resolving
Evaluation of Criteria by Interactive Flexible Elicitation Procedure, In: Dargam F, Hernández
J, Zaraté P, Liu S, Ribeiro R, Delibasic B, Papathanasiou J. Decision Support Systems III -
Impact of Decision Support Systems for Global Environments”. LNBIP 184 (Lecture Notes
in Business Information Processing), Springer. 157–166
de Almeida AT, Ferreira RJP, Cavalcante CAV (2015) A review of multicriteria and multi-
objective models in maintenance and reliability problems. IMA Journal of Management
Mathematics 26(3):249–271
de Almeida AT, Vetschera R (2012) A note on scale transformations in the PROMETHEE V
method. Eur J Oper Res 219:198–200
Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multiobjective genetic
algorithm: NSGA-II. IEEE Trans Evol Comput 6:182–197
Dias LC, Climaco JN (2000). Additive aggregation with variable interdependent parameters: The
VIP analysis software. J Oper Res Soc 51:1070–1082
Dias LC, Clímaco JN (2005) Dealing with imprecise information in group multicriteria
decisions: A methodology and a GDSS architecture. Eur J Oper Res 160(2) 291–307
Duckstein L, Monarchi D, Kisicl CC (1975) Interactive Multi-Objective Decision Making Under
Uncertainty. Theor Decis Pract Hodder Stoughtor, London 128–147.
Eden C (1988) Cognitive mapping. Eur J Oper Res 36(1):1–13
Eden C, Ackermann F (2004) SODA. The Principles. In: Rosenhead J, Mingers J (eds) Rational
Analysis for a Problematic World Revisited. Second Edition, Chichester: John Wiley & Sons
Ltd.
Edwards W, Barron FH (1994) SMARTS and SMARTER: Improved Simple Methods for
Multiattribute Utility Measurement. Organ Behav Hum Decis Process 60(3):306–325
Edwards W, Miles Jr RF, Von Winterfeldt D (2007) Advances in decision analysis: from
foundations to applications. Cambridge University Press
Ehrgott M (2006) Multicriteria optimization. Springer Science & Business Media, Berlin
Erickson M, Mayer A, Horn J (2001) The Niched Pareto Genetic Algorithm 2 Applied to the
Design of Groundwater Remediation Systems. In: Zitzler E, Thiele L, Deb K, et al (eds)
Evol. Multi-Criterion Optim. SE - 48. Springer Berlin Heidelberg, pp 681–695
Figueira J, Greco S, Ehrgott M (eds) (2005) Multiple Criteria Decision Analysis: State of the Art
Surveys. Springer Verlag, Boston, Dordrecht, London
Fishburn PC (1976) Noncompensatory preferences. Synthese 33:393–403
References 85
Kirkwood CW, Sarin RK (1985) Ranking with Partial Information: A Method and an
Application. Oper Res 33:38-48
Knowles JD, Corne DW (2000) Approximating the nondominated front using the Pareto
Archived Evolution Strategy. Evol Comput 8:149–172
Korhonen P (2005) Interactive Methods. Mult. Criteria Decis. Anal. State Art Surv. SE - 16.
Springer New York, pp 641–661
Korhonen P (2009) Multiple objective programming support Multiple Objective Programming
Support. In: Floudas CA, Pardalos PM (eds) Encycl. Optim. SE - 431. Springer US, pp 2503–
2511
Korhonen P, Wallenius J (2010) Interactive Multiple Objective Programming Methods. In:
Zopounidis C, Pardalos PM (eds) Handb. Multicriteria Anal. SE - 9. Springer Berlin
Heidelberg, pp 263–286
Leyva-Lopez JC, Fernandez-Gonzalez E (2003) A new method for group decision support based
on ELECTRE III methodology. Eur J Oper Res 148(1):14-27
Likert R (1932) A technique for the measurement of attitudes. Arch Psychol 22(140):1–55
Miettinen K (1999) Nonlinear multiobjective optimization. Springer Science & Business Media
Morais DC, de Almeida AT (2012) Group Decision Making on Water Resources based on
Analysis of Individual Rankings. Omega 40:42-45
Munda G (2008) Social multi-criteria evaluation for a sustainable economy. Springer, Berlin
Mustajoki J, Hämäläinen RP (2005) Decision Support by Interval SMART/SWING -
Incorporating Imprecision in the SMART and SWING Methods. Decision Sci 36(2):317-339
Nurmi H (2002) Voting Procedures under Uncertainty. Springer Verlag, Berlin-Heidelberg, New
York
Nurmi H (1987) Comparing Voting Systems. Dordrecht: D. Reidel Publishing Company
Osyczka A (1984) Multicriterion optimisation in engineering. Halsted Press
Pardalos PM, Siskos Y, Zopounidis C (eds) (1995) Advances in Multicriteria Analysis. Kluwer
Academic Publishers
Partnoy, F (2012) Wait: The Art and Science of Delay. Perseus Group books
Pawlak Z, Slowinski R (1994) Rough set approach to multiattribute decision-analysis. Eur J
Oper Res 72:443-459
Pedrycz W, Ekel P, Parreiras R (2011) Fuzzy Multicriteria Decision-Making: Models, Methods,
and Applications. John Wiley & Sons, Chichester
Polmerol J-C, Barba-Romero S (2000) Multicriterion Decision in Management: Principles and
Practice. Kluwer
Raiffa H (1968) Decision analysis: introductory lectures on choices under uncertainty. Addison-
Wesley, London
Rao S (1984) Multiobjective optimization in structural design with uncertain parameters and
stochastic processes. AIAA J 22:1670–1678
Rauschmayer F, Kavathatzopoulos I, Kunsch PL, Le Menestrel M (2009) Why good practice of
OR is not enough—Ethical challenges for the OR practitioner. Omega 37(6):1089–1099
Rosenhead J, Mingers J (eds) (2004) Rational Analysis for a Problematic World Revisited.
Second Edition, John Wiley & Sons Ltd.
Roy B (1996) Multicriteria Methodology for Decision Aiding. Springer US
Roy B, SáowiĔski R (2013) Questions guiding the choice of a multicriteria decision aiding
method. EURO J Decis Process 1(1-2):69–97
Roy B, Vanderpooten D (1996) The European school of MCDA: Emergence, basic features and
current works. J Multi-Criteria Decis Anal 5(1):22–38
Saaty, TL (1980) The Analytic Hierarchy Process. McGraw-Hill
Salo A, Hämäläinen RP (2001). Preference ratios in multiattribute evaluation (PRIME) -
elicitation and decision procedures under incomplete information. IEEE Sys Man Cybern
31(6):533-545
Salo A, Punkka A (2005) Rank inclusion in criteria hierarchies. Eur J Oper Res 163(2):338-356
References 87
Salo AA, Hämäläinen RP (1992) Preference assessment by imprecise ratio statements. Oper Res
40:1053-1061
Schaffer JD (1984) Multiple Objective Optimization with Vector Evaluated Genetic Algorithms.
PhD thesis, Vanderbilt University
Shakun MF (1988) Evolutionary Systems Design: Policy Making Under Complexity and Group
Decision Support Systems. Holden-Day, Oakland, CA
Shakun MF (2006) ESD: A Formal Consciousness Model for International Negotiation. Group
Decis Negot 15:491–510
Shakun MF (2010) Doing Right: Connectedness Problem Solving and Negotiation. In: Kilgour
DM, Eden C (Eds.) Handbook of Group Decision and Negotiation, Advances in Group
Decision and Negotiation 4. Springer Science
Simon HA (1955) A Behavioral Model of Rational Choice. Q J Econ 69(1):99–118
Simon HA (1960) The New Science of Management Decision. Harper & Row Publishers, Inc,
New York
Simon, HA (1982) Models of Bounded Rationality. MIT Press
Slack N, Chambers S, Harland C, Harrison A, Johnson R (1995) Operations Management,
Pitman Publishing, London
Slowinski R, Greco S, Matarazzo B (2012) Rough set and rule-based multicriteria decision
aiding. Pesq Oper 32:213-269
Sprague Jr RH, Watson HJ (eds) (1989) Decision Support Systems - Putting Theory into
Practice, Prentice-Hall
Srinivas N, Deb K (1994) Multiobjective Optimization Using Nondominated Sorting in Genetic
Algorithms. Evol Comput 2:221–248
Steuer RE (1986) Multiple Criteria Optimization: Theory, Computation, and Application. Wiley,
New York
Stewart TJ (2005) Dealing with uncertainties in MCDA. In: Figueira J, Greco S, Ehrgott M (eds)
Multiple Criteria Decision Analysis: State of the Art Surveys. Springer Verlag, Boston,
Dordrecht, London, 445-470
Thierauf, RJ (1982) Decision support systems for effective planning and control - A case study
approach. Prentice-Hall, Inc., Englewood Cliffs, New Jersey
Vetschera R, de Almeida AT (2012) A PROMETHEE-based approach to portfolio selection
problems. Comput Oper Res 39(5):1010–1020
Vincke P (1992) Multicriteria Decision-Aid. John Wiley & Sons, New York
Von Neumann J, Morgenstern O (1944) Theory of games and economic behavior. Princeton:
Princeton University Press
Wallenius J (1975) Comparative Evaluation of Some Interactive Approaches to Multicriterion
Optimization. Manage Sci 21(12):1387–1396
Weber M, Borcherding K (1993) Behavioral influences on weight judgments in multiattribute
decision making. Eur J Oper Res 67(1):1–12
Zitzler E, Thiele L (1999) Multiobjective evolutionary algorithms: a comparative case study and
the strength Pareto approach. Evol Comput IEEE Trans 3(4):257–271
Chapter 3
Basic Concepts on Risk Analysis, Reliability
and Maintenance
There are many concepts on risk found in the literature and also different
perceptions to it. However, if a decision is being made and risk is involved, then,
the risk concept should combine consequences and probabilities, incorporating the
DM’s preferences over that, as seen in Chap. 2.
Actually, a ‘decision process’ with no DM’s preference has no decision being
made, as discussed at the end of Chap. 1. Instead of that, that process either: a) has
some preference structure incorporated within the model, at random; b) is just
arbitrary following a previous decision of someone else.
Even so, in most of real cases, the consequences are multidimensional, and
therefore, require an MCDM/A approach for building a decision model. The
following topics are mainly based on the basic RRM literature and do not
incorporate the idea of decision support, as given in Chap. 2. That is, DM’s
preferences are not necessarily considered in the model.
In recent times, undertaking risk studies has become an increasingly complex task,
making it of great importance in different spheres of society. Modern world
facilitates access to information, making people more conscious of decision-
making on risk and its consequences in the social and environmental context. On
the other hand, organizations seek to manage appropriately all risks perceived as
being the most relevant ones in the production of goods or services to ensure that
their final product meets the minimum legal requirements, regulations and
resolutions as well as society’s expectations. However, it is of paramount
importance to emphasize that in the so-called real world, despite organizations
being concerned with identifying and monitoring risks, the restricted availability
of resources is a crucial point, which leads to some risks receiving special
attention with regard to the immediate allocation of resources, while others have
to wait until resources become available.
Although in the literature there are several definitions of the term risk, the basic
concept is associated with uncertainty in an environment and this is related to the
likelihood of an undesirable event occurring and the impact of its consequences.
This is why, according to Theodore and Dupont (2012), risk is defined as a
measure of financial loss or damage to persons, in terms of the likelihood of an
incident occurring and the magnitude of the loss. To Yoe (2012), risk is a measure
of the likelihood and consequences of uncertain future events. It is the chance of
an unwanted result where the lack of information about events that have not yet
occurred is one of the factors inherent in the chance of its happening. Cox (2009)
considers the preferences for consequences.
In the risk context, a change was recently observed by Aven (2012), who states
that traditionally dangerous activities were designed and operated from references
based on codes, standards and hardware requirements. However, what is verified
today is that this trend is more directed towards a functional orientation, where the
focus is associated with what it is sought to achieve. Therefore, the ability to
define risk is the key element in each functional system. Identifying and
categorizing risk are necessary to provide a decision support. The ability to define
what may occur in the future, to evaluate risks and uncertainties and to choose
among alternatives is what guides the decision-making process in the context of
risk.
As seen in Chap. 2, the risk concept in the decision process combines the
consequences with its probabilities, and incorporates the DM’s preferences over
that combination. Even so, in most real cases, the consequences are multi-
dimensional, and therefore, involve an MCDM/A approach, which may involve
tradeoffs as pointed out by Cox (2009), considering dimensions such as: financial,
reliability and health.
Risk Management, Risk Assessment and Risk Analysis are supposed to ensure
proper risk management and control, taking into account aspects such as
3.1 Basic Concepts on Risk Analysis 91
Society deals with risks in everyday life so much so that risk analysis is an
inherent characteristic of human beings. In daily routine activities, risk is always
present e.g. when walking in the street, using public transportation to work, eating
fatty foods, etc. Each person who participates in a hazard/risk analysis gives their
own opinion, memory, attitude and global view of the situation under study.
Moreover, these people are often affected by different types of personal biases
such as their level of education, beliefs, experience, culture, etc. Even experts
come to different conclusions when presented with the same data. The literature
examines the issues that arise by discussing different physical situations and
contexts.
van Leeuwen (2007) supports that perceptions of risk vary among individuals
and the general public, business and other stakeholders, and change over time and
in accordance with the prevailing culture. People continually assess situations and
decide if the risks associated with a particular action can be justified. In some
circumstances, dangerous effects are clearly associated with a particular course of
action. However, in other cases, the impact of each effect can be uncertain and not
immediately obvious.
To Modarres et al. (1999), the perception of risk often differs from the
perception of objective measures, thereby distorting risk management decisions.
Subjective judgments, beliefs and social bias with respect to events with low
probability and high consequence may affect how the results of risk analysis are
understood.
In this context, according to Crowl and Louvar (2001), the general public has
great difficulty with understanding the concept of risk acceptability. The major
problem is related to the involuntary nature of accepting a given degree of risk.
For instance, designers of chemical plants who specify a level of acceptable risk
assume that these risks are satisfactory to those living in the vicinity of the plant.
3.1 Basic Concepts on Risk Analysis 93
type of analysis. For Smith and Simpson (2010), the degree of risk considered as
tolerable depends on a number of aspects such as the degree of control under the
circumstances, the nature of risk analysis (intentional or unintentional), the
number of persons subject to risk, etc. Finally, the concept of intolerable risk
consists of not tolerating a specific risk level, thus not allowing activities to be
developed at this level. Additional comments with regard to these definitions can
be verified in the section dealing with ALARP concept.
To Crowl and Louvar (2001), it is impossible to eliminate any kind of risk
completely. At some point in the design stage, someone needs to determine
whether the risks are acceptable or not. In other words, are the risks under analysis
greater than the daily risks that individuals are subject to in their daily lives?
According to Modarres et al. (1999), risk acceptability is a complex and
controversial issue. However, making use of risk assessment results is a common
way to rank the exposure level of risk, where the risk exposure levels that are
socially acceptable should be defined based on risk acceptance thresholds.
In this context, some risk measures can be verified such as Individual Risk,
Societal Risk, Population Risk and Risk Indices. Each of these measures expresses
the risk, taking into account different aspects and contexts.
According to Smith (2011), Individual Risk refers to the frequency of a fatality
for a hypothetical person with respect to a specific hazard scenario, while the
Societal risk reflects the risk measure for a group of people, taking into account
multiple fatalities. Theodore and Dupont (2012) describe Population Risk as the
risk for the entire population, expressed as a certain number of deaths expressed as
thousands or millions of people potentially exposed to danger. Theodore and
Dupont (2012) also define Risk Indices, describing them as measures represented
by a unique number associated with a facility. Some risk indices are quantitative
while others are semi-quantitative, ranking risks in various categories. Risk
indices can also be quantitative average or benchmarkings based on other risk
measures.
In this context, Crowl and Louvar (2001) add that among these risk measures,
losses and accidents based on statistical data are relevant measures. However, they
should be considered with some caution, given that many of these statistics
represent an average, and do not reflect the occurrence of a specific accident with
potential losses. In contrast, no specific method is capable of measuring all aspects
simultaneously. Some of those commonly used are the OSHA incident rate, the
Fatal Accident Rate (FAR) and the Fatality Rate.
More specifically, according to Tweeddale (2003), FAR is a risk measure used
to assess the risks associated with the employees of an industrial plant. FAR is
defined as the number of fatalities, due to accidents at work, per 100 million hours
worked.
In conclusion, the definition of risk measures is necessary so that reference
values are established in risk studies and used in order that objectives are met
regarding monitoring and controlling risk.
3.1 Basic Concepts on Risk Analysis 95
According to Andrews and Moss (2002), HAZOP is a method that was first used
in the chemical industry, where industrial plants are evaluated with regard to
identifying potential hazards to operators and society. These hazards may arise in
a particular system and may be result of interaction among different systems of the
industrial process.
96 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
The Fault Tree (FT) is a tool widely used in industrial processes within the risk
environment. It can be classified as a qualitative or quantitative tool depending on
the availability of the likelihood values of failure events.
3.1 Basic Concepts on Risk Analysis 97
Top Event
AND
gate
Basic OR
event gate
Basic
Basic
event
event
yes
Scenario 1
yes
no
Scenario 2
Gas release
(Rupture)
yes
Scenario 3
yes
no
no Scenario 4
no
Scenario 5
Fig. 3.2 Illustrative example of event tree applied to the risk analysis of a pipeline
Regarding the events that comprise ET accident sequences, Zio (2007) states
that they are characterized by: intervention (or not) of protection systems that
should come into operation (or not) to mitigate the accident (System Event Tree);
the running (or not) of security functions (Functional Event Tree); and the
occurrence (or not) of physical phenomena (Phenomenological Event Tree).
According to Zio (2007), these event trees types are applied in different
contexts:
x System Event Tree – this is used to identify accident sequences that have
developed within a plant, involving protection and security systems;
x Functional Event Tree – this is an intermediate step when constructing the
System Event Tree. From the ET initial event, safety functions that need to be
established are identified, and are subsequently replaced by the corresponding
protection and security systems;
100 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
Success
Success Scenario 1
Fail
Scenario 2
Initial event
Success
Scenario 3
Success
Fail
Fail
Scenario 4
Fail
Scenario 5
Top Event
AND
gate
Basic Basic
event event
Fig. 3.3 How integrated tools (FTA and ETA) are used to determine failed states
The likelihood of an ETA failure is the same of a top event obtained from FTA,
implemented for each specific failure observed in the ETA. The likelihood of
success is calculated as being the complement of this failure likelihood.
Otherwise, P(Success) = 1 – P(Fail).
3.1 Basic Concepts on Risk Analysis 101
Identification of
Critical Events
Coarse
Consequence
Analysis
Quantitative
Cause Analysis
Detailed
Consequence
Analysis
Risk Calculation
Considering the tools available for hazard identification, such studies are
usually supported by the use of checklists, statistics on failure, a database of
accidents, HAZOP studies and similar risk analysis studies. The experiences
obtained from similar projects are also an important source used to identify
hazards.
After identifying the critical hazards to be considered, it is necessary to identify
the causes of these hazards and which events may lead to an accident scenario
occurring. Identifying the starting point for a potential accident enables the chain
of events that may cause an accident to be established.
During the analysis of the cause analysis (the third step), it is determined which
causes may lead to the initiating events in order to support the assessment of the
probabilities of initiating events. From the cause analysis, it is possible to identify
risk reducing actions that would prevent or interrupt the chain of events that may
cause an accident. In the initial steps of the cause analysis, qualitative techniques
are usually deployed followed by quantitative approaches if there are data
available for quantification. Qualitative approaches are used to identify causes and
conditions for initiating events, thereby establishing the basis for a possible
subsequent quantitative analysis. With regard to the techniques used to identify the
causes, there are: HAZOP, Fault Tree Analysis (FTA), Preliminary Hazard
Analysis (PHA), FMEA and human error analysis techniques, which are also used
in traditional reliability analysis.
Quantitative studies in cause analysis are conducted in order to establish the
probability of the occurrence of initiating events, while using historical statistics to
calculate the frequency of initiating events is one of the most common approaches.
The fourth step in Fig. 3.4 is related to the consequence analysis of accident
scenarios. A consequence analysis considers the existence of barrier functions and
elements to contain hazards and the accident sequences in order to evaluate the
possible function or failure of barriers involved. According to Vinnem (2014), fire
and explosions are two of the main factors evaluated, and both may be assessed by
using the same calculation steps for all scenarios which may involve fire and/or an
explosion. These steps depend on which conditions and sequences are related to
the factors evaluated. Fire and explosions may be a result from a leakage,
punctures or pinhole, any of which may expose a hazardous material which when
associated with a chain of events may result in a Vapor Cloud Explosion (VCE), a
Boiling Liquid Expanding Vapor Explosion (BLEVE), a Flash Fire, a Jet Fire, and
so forth. Thus, the steps of the calculation are used to estimate the amount of
material leaked by considering temperature and pressure conditions in the system
associated with system barriers and mitigation actions. TNO’s Colored Books
present systematic procedures to assist QRA studies, especially consequence
analysis, regarding the estimation of thermal radiation, ignition probabilities,
conditional probabilities for fatalities, damages and other consequences.
Regarding fatalities, probit functions are usually used to calculate the probability
of death due to exposure to toxic substances and / or heat radiation at a given level
of exposure.
104 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
The results from a QRA study arise from the risk calculation, the last step of
Fig. 3.4. These results are usually compared and associated with risk tolerance
level. QRA studies are usually performed until barriers and safety actions are
strong enough to assure that any risk is above reference levels. A QRA study aims
to provide a risk picture, which results from the hazard identification, and a cause
and frequency analysis that are combined to express the risk level associated with
all critical hazards.
Although the terms risk calculation, risk analysis and risk assessment can be
easily misunderstood since they have the same general meaning, there are
differences related to the scope of each term. Risk calculation uses information
from consequence analysis and cause analysis thereby providing a risk level
calculated from frequencies and the magnitudes of consequences. While risk
analysis refers to the entire process described in Fig. 3.4, which includes the risk
calculation, Risk assessment is the entire process of risk analysis when the results
are evaluated regarding risk reference levels, which are defined by considering a
notion of risk tolerance.
To ensure the reliability of the results of QRA studies, there are several factors
that must be considered, such as:
x The technical description of the system (activities, operational phases);
x Purpose and target of risk analysis;
x Activity levels on the installation;
x Operation of safety systems;
x Study assumptions: how these are verified and accepted;
x Data Sources.
Thus, QRA is a systematic development of numerical estimates of the expected
frequency and/or consequence of potential accidents associated with a facility or
operation.
According to Arendt and Lorenzo (2000), there are two main misconceptions
about QRA which are to do with the lack of adequate data on equipment failure
and the cost of conducting QRA, i.e. whether it is cheap or expensive.
Regarding the availability of data, there are industry wide databases that can
provide data for frequency rate estimates and regulation authorities that provide
periodical reports, such as:
x The Guidelines for Process Equipment Reliability Data with Data Tables;
x IEEE Guide to the Collection and Presentation of Electrical, Electronic,
Sensing Component, and Mechanical Equipment Reliability Data for Nuclear
Power Generating Stations (IEEE Std 500);
x The OREDA Offshore Reliability Data Handbook;
x Non-electronic Parts Reliability Data 1991 (NPRD-91) and Failure
Mode/Mechanism Distributions 1991;
x Systems Reliability Service Data Bank;
3.1 Basic Concepts on Risk Analysis 105
3.1.8 ALARP
According to Bedford and Cooke (2001) the ALARP (As Low as Reasonably
Practicable) principle has guided the setting of tolerance risk levels since the
1980s and 1990s, in order to achieve safety goals. The USNRC policy statement
(NRC 1986) and the UK tolerability of risk document (HSE 1987) seek to convert
the principle of setting this ‘as low as reasonably practicable’ (ALARP) into a
numerical definition, which establishes upper levels for risk intolerance and lower
levels at which risks can be considered as tolerable.
Given that criteria for risk acceptance are generally combined with risk
analysis, some industries and countries have regulations that require such criteria
to be defined prior to the risk analysis (Aven 2012).
106 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
Risk Magnitude
ALARP or
Tolerable region
The definition of ALARP regions are based on everyday risks. Thus, risks that
are considered typical and commonly expected may include risks from all causes,
including bad health.
To measure the risk level, the Fatal Accident Rate (FAR) is used with
particular regard to the employees of some hazardous installations who are usually
exposed to higher risks than those working in less hazardous workplaces.
Aven (2012) adds that in practice the value considered to reflect risk is an
estimation of FAR or the probability p of a certain accident event, since the true
value of FAR or p is unknown. Thus using tolerance limits means to compare an
estimated value with acceptable values. This means that using a best-estimate
approach may not produce clear recommendations, and thus standardized models
and input data may be required. Thus, the acceptance level is a function of such
models and input data.
According to Tweeddale (2003) in some cases there is a subjective opinion and
a potential debate about whether ALARP standards are achieved and this can lead
to such issues being questioned in court. Nevertheless, if the hazardous installation
uses the best technology available and can be set up, and it also uses the best oper-
able and maintainable management systems in order to improve safety by keeping
the equipment maintained to high standards, the risk is usually an ALARP one.
There are some criticisms regarding the use of ALARP as a way to justify risk
exposure. In addition, there is a problem with the term regarding ‘acceptable risk’.
This is because it is commonly used by those who generate the risk to excuse the
fact that others will be exposed to it. And this is why some authors call the very
concept of ALARP into question. Tweeddale (2003) remarks that the level of risk
that an individual accepts is particular to that individual. That is, it is not an
108 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
It is important to remark that these everyday risks may change from country to
country; for example, even in the same country, some regions may have a
significantly higher homicide rate than the rest of the country. Therefore, it is
possible that a given risk level could be considered tolerable in an undeveloped
country and intolerable in a developed country. Thus, the values given by
Macdonald (2004), with some examples given in Table 3.2 reflect the reality of a
developed country, and these might be lower than in undeveloped countries.
After assessing the probability of hazardous events, all possible actions must be
deployed in order to achieve a tolerable risk level. In fact, if the risk to life is so
high that is beyond economic concern, the equipment or plant must be considered
safe, otherwise it must be closed.
However, when the tolerable risk level is reached, investments in risk reduction
are only justified by a cost effective evaluation. Thus, when a risk level is
considered tolerable or an ALARP one, any costs to improve safety must be
followed by a compatible benefit. Otherwise, it should not be implemented.
Usually the cost per life saved with a previously established level is considered.
Aven (2008) defines a similar measurement as the implied value of a statistical
life or the implied cost of averting a fatality, by dividing the cost of the safety
improvement by the number that represents the expected reduction in the number
of fatalities. Therefore this ratio can also be considered by quantities other than
lives saved, if for example, an environmental risk is being considered, the
reference may be to tons of oil spilled.
This allows expending resources in order to improve safety by acting where
one can find the greatest benefits while taking the budget allocated to improve
safety into account. These costs vary according to the type of system, complexity
and regulatory standards regarding the activity. Enterprises usually avoid
disclosing data on levels of cost per life saved. According to Smith (2011), this
value is between £500,000 to £4,000,000, while if the risk has potentially multiple
fatalities, then higher amounts may be considered.
Thus, the more that the number of potential fatalities increases, the more risk
averse the analysis becomes, which leads to choosing a higher cost per life saved
level. It is valuable to observe that utility theory provides an axiomatic structure to
evaluate DM’s behavior regarding risks, including risk aversion, as it seen in
Chap. 2.
As examples of how these values are considered, Smith (2011) points out that
for passenger road transportation, there is a voluntary aspect to the exposure and a
small number of casualties per incident, so the value considered for cost per life
saved is approximately £1,000,000. For the transportation of dangerous material,
where the risk is not under an individual´s personal control which means that there
110 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
is an involuntary risk, Smith (2011) presents a cost per life saved of approximately
£2,000,000 to £4,000,000. When considering multiple offshore fatalities, where
there are a large number of fatalities and no personal control by the victims, Smith
(2011) shows that the cost per life saved can be between £5,000,000 to
£15,000,000. Therefore, these values are quite controversial and may change when
they came under scrutiny in the media or are reported as catastrophic accidents,
thereby making the analysis even more risk averse.
Smith (2011) states that the maximum tolerable risk for a single fatality does
not always coincide with the societal risk calculations. Thus, while societal risk
measures the frequency of a fatal event, when considering individual risk, it is the
frequency of individual deaths that is considered. One of the main differences
between estimating individual risk and societal risk is about whether the risk is
voluntary or involuntary. When considering individual risk it is important to
highlight that these individuals are voluntarily exposing themselves to risk, in a
specific place that sets specific conditions for the frequency and risk assessment.
When considering societal risk, what is considered is the involuntary exposure to
risk that may reach random individuals, and it characterizes this concept of an
involuntary risk.
According to Tweeddale (2003) there is no unanimous formal agreement
regarding a specific value that can be considered as denoting a tolerable level of
risk. But in many countries it is typical to consider that an additional risk of 1
chance in a million per year (10-6 per year), due to industrial sources affecting the
person most exposed to these, is a very low risk level compared to everyday risks
that an ordinary person is usually exposed to without questions being raised about
this. Aven (2008) points out that the probability of a fatality for a third person
associated with exposure to risk in an industrial plant is required to be less than
10-5 per year. Therefore, the value that defines whether a risk should be considered
tolerable and therefore accepted by the wider community is the Individual Risk
level. Some of these everyday risks have been exemplified.
When calculating individual risk the focus must be on an event in which one
specific person is seriously injured or killed. Aven (2008) defines individual risk
as the frequency of death for the person or critical group of personnel most at risk
from a given activity due to their location, habits or periods that make them
vulnerable. Thus, individual risk is measured as the annual frequency of an
accident with one or more fatalities over a homogeneous group of people who
voluntarily expose themselves to risk. This is an approximation of the probability
that a random person of a group who conducts a specific voluntary activity will be
killed while he/she is at the industrial facility over the course of the time period
considered, usually a year. This measurement is used to calculate the FAR.
As to the risk to any individual who is involuntarily exposed to some risk,
consideration has to be given to the possibility that more than one person may be
killed due to that risk source. Thus societal risk cannot be measured only by
individual risk, but must include the possibility that there may be 1 to N fatalities.
The more the number of fatalities increases, the more risk averse the analysis is.
3.1 Basic Concepts on Risk Analysis 111
Societal risk is usually represented by F-N curves, which show the frequency of
accident events with at least N fatalities.
Tweeddale (2003) recognizes the controversy regarding attempts to put a value
on a human life, since human life can be considered priceless due to the emotional
values that money cannot compensate for. Nevertheless, there is a need to
establish a limit for the amount that can be spent per life saved, otherwise it is
impossible to decide if a choice can be made on economic grounds between
improving safety in order to keep an industrial plant running or closing the plant.
According to Tweeddale (2003), one absolute limit to be established for cost per
life saved would be obtained by dividing the annual gross national product by the
annual number of births. This value represents the amount that may possibly be
spent in order to extend the life expectancy of each new-born baby, if there are no
other expenses in the community. As there are many other requests for financial
resourcing from the wealth derived from the community, the real value of the limit
to the cost per life saved would be less. Therefore the definition of this value will
depend on particular priorities and other characteristics of the problem, such as
those pointed out in the examples which discussed calculating risk values to do
with the transportation of passengers by road, the transportation of dangerous
materials and substances and multiple offshore fatalities.
Ackermann et al. (2014) present a risk map to engage multiple stakeholders and
build a comprehensive view of risks. The authors use risk map as a dynamic tool
to update information and create knowledge to the decision making process.
Bostrom et al. (2008) presented the foundation for designing and testing
alternative ways to communicate risk and uncertainty for low-probability and
high-consequence events, using the knowledge about the effects of spatial
information, communication of risk, and uncertainty in spatial information and
how these can be tailored effectively for earthquake risk analysis.
Fedra (1998) highlights that technological and environmental risks have an
obvious spatial dimension. Floods, mudslides, and avalanches as much as toxic
spills, explosions, transportation of dangerous goods, or hazardous waste
management are all spatially distributed problems.
Eppler and Aeschimann (2009) present a conceptual framework for risk
visualization in risk management. This framework is based in the answers for the
questions of: ‘why’ (purposes), ‘what’ (contents), ‘for whom’ (target groups),
‘when’ (usage situations), and ‘how’ (formats).
In this context, some applications can be observed in the literature. Brito and
Almeida (2009), Alencar and de Almeida (2010) and Lins and de Almeida (2012)
contextualize the multidimensional risk view in the context of natural and
hydrogen gas pipelines. The multidimensional risk analysis results are presented
by the risk difference between pipeline sections. These risk increments provide to
DMs a different interpretation with regards risk visualization, allowing that the
DM allocates resources according to the risk hierarchy. It also allows the
visualization of the gap size between the risks of two subsequent sections of the
ranking.
Additionally, Garcez and de Almeida (2014) present a multidimensional risk
assessment under an intra-criterion vision in the underground electricity
distribution context. This information view allows the DM to identify the relevant
consequence dimensions for each alternative and thus allocate resources to prevent
and mitigate risk more effectively, prioritizing only those dimensions that impact
the alternative. For example, an alternative that impacts only humans, should not
receive resources that are allocated to the environmental dimension, thus
preventing a misallocation of resources.
Tariq (2013) presents damage curves and maps based on estimated losses and
probabilities of all floods considered. The maps illustrate the flood risk
distribution over the study area, including agricultural land-use zoning and
comparisons over the area before and after crop.
Finally, a specific point that should be mentioned concerns with the application
of Geo Information Technology (GIT), Geo Information Systems (GIS), and
software for visualization of qualitative and quantitative analyses. In this context,
an overview of 3D visualization tools for quantitative analyses could be observed
in Kaufmann and Haring (2014).
As an example of application, Jaedicke et al. (2014) uses a GIS (Geographic
Information Systems) solution to warn avalanches in Norway. Maps are used in
3.2 Basic Concepts on Reliability 115
design projects that ended up with products being produced (Rausand and
Høyland 2004). On the other hand, it should also be noted that how pieces of
equipment are operated and maintained may significantly affect the chances of
their developing faults and failures.
This observation, in fact characterizes the second approach which makes use of
common sense and everyday experience to highlight that the effectiveness of a
functioning system depends not only on its “innate” properties, but also on the
quality of its operations, maintenance, repair, or on any activity that interferes
with the operational performance of the equipment. At one extreme, if all
maintenance actions are limited to emergency repairs only after the system has
suffered a failure, then the operational characteristics of the system are likely to be
very low and the system will not operate in an efficient manner (Scarf et al. 2009).
As a result, the second approach deals with numerous issues, the main concerns of
which are related to the system already in operation and its nature and have regard
to proposing measures that will obtain the best possible operational characteristics.
The importance of this point of view for this book is to do with the ease with
which maintenance activity is seen to be related to its proper purpose. Indeed, the
main objective of maintenance is to anticipate failure and consequently to reduce
of the probability of its occurrence, which in turn, contributes to mitigating
possible consequences associated with failure. Therefore, the way in which
maintenance actions affect reliability are also discussed in this chapter and which
actions can be undertaken to ensure the operating performance of equipment is
good.
Despite the fact that reliability can be tackled over a wide field of study, and may
cover issues not only associated with the project, but also the actions that can be
performed to maximize the performance of the equipment already in place, there
are other narrower views of reliability.
According to Márquez (2007) reliability, as well as risk, are in fact elements
that quantify uncertainties. Thus, as the quantification of uncertainty is not in itself
an end but the means by which it is possible to make better decisions, it can be
said that using risk analysis methods and reliability supports the decision process
under uncertainty.
Viewed from this perspective, reliability would then be a set of methods that
helps in decision making regarding the performance of the system under study.
Commonly reliability is deemed to have three main branches, namely:
The reliability of hardware;
The reliability of software;
Human Reliability.
3.2 Basic Concepts on Reliability 117
This chapter refers to the branch of reliability that is associated with the
operation of components and equipment. On the other hand, the existence of
different branches emphasized the need to study the different aspects involved in
socio-technical systems: man; the machine and software (intangible elements that
are used to operate these machines) (Pham 1999).
Indeed, the fact of there being different approaches to dealing, separately, with
the different agents involved in the operation of production systems indicates that
the procedures for using a particular approach, in fact, depict only part of the real
problem. This kind of reductionism in modeling a decision problem eliminates the
influences that might be present from the other actors and by doing so, this makes
it feasible to find solutions.
Thus, it is important to keep always in mind that the domain of the
consequences of a failure is limited to the perspective that is actually used.
Consequently the awareness that reliability analysis may provide incomplete
information about the actual performance of the system in its completeness warns
of the need to keep an open mind and to look for complementary dimensions of
decision making or other issues that were possibly left out by the reliability
method adopted.
As an example, consider the behavior of a piece of equipment that had a
failure. This is done by seeking to discover how the process of wear and tear due
to how the equipment operated occurred which typically entails undertaking an
analysis using a reliability-driven approach to machinery and equipment, in
isolation, and without considering the influence of other elements (human and
software). Yet, at the same time, as equipment becomes older, due to wear and
tear, there is also a set of circumstances that leads operators to misusing
equipment and this can lead to failure. This is not taken into account. Similarly, in
an automated system, the malfunction of the system can lead to failure. Such
irregular operating regimes may be linked to failures in control programs or other
items of software, neither of which is taken into account under a reliability-driven
approach.
Besides reducing the scope of analysis when adopting a reliability-driven
approach, it is very common to limit looking at how the components of a system
are impacted by changes in other parts of the system as a whole or how each of
them may impact the system.
According to Jorgenson et al. (1967) this type of simplification is a way to
overcome the difficulties imposed by the complexity of large systems. Moreover,
reliability analysis at the component level is consistent with the actions that are
carried out in practice. The most frequent failures occur in a component, so it is
unnecessary to replace the complete system. Furthermore, most scheduled
maintenance activities also require software components, not equipment, to be
replaced.
118 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
R(t ) P (T ! t ) . (3.1)
f (t )
O (t ) . (3.2)
R (t )
1 d
O (t ) R (t ) . (3.3)
R (t ) dt
ª t º
R(t ) exp « O (t )dt » .
³ (3.4)
« 0 »
¬ ¼
Even though every care is taken over a set of items, either in the design phase,
or in the phase when the item is already in use, it is observed that failures still
occur. Such failures are characterized into different types depending on what the
predominant mechanisms were that worked most effectively in bringing them
about.
First, there are the failures that occur quite early in the life of a component. The
most likely cause of this type of failure is that equipment parts were defective due
to their having been improperly manufactured or constructed. It is this which leads
to high rates of early failures of engineering devices. Loss of parts, substandard
materials, components that are out of tolerance, and defects caused during
transportation are among the causes of failures. This is indicative of inefficient
quality control and results in excessive failure rates near the beginning of the life-
cycle of the project (Lewis 1987).
The middle part of the bathtub curve contains the lowest levels of failure rate
and shows little variation, behaving approximately as a constant. It is referred to
as useful life. Failures during this period of time are often ratified as chance
failures i.e., they happen irregularly and unexpectedly. They probably arise due to
unavoidable loads. External loads above equipment design capacity can lead to an
increase in failure rate, for example, due to the equipment material fatigue
(Guedes Soares and Garbatov 1996; Garbatov and Guedes Soares 2001).
The part on the right of the bathtub curve is a region in which the failure rate
increases. During this period, failures due to the aging are prevalent and the
cumulative effects of such matters fatigue and as corrosion tend to be the
dominant causes of these. Wear out failures are symptomatic of components aging
(Lewis 1987; Bazovsky 2004). These failures happen only if the item is not
appropriately maintained. In practice, it is the calculation of when the failure rate
will start to increase rapidly that usually forms the basis for determining not only
when parts should be replaced but also for specifying the design life of the
component.
120 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
O(t)
Each failure mode and the consequent behavior of the failure rate can be
represented by an analytical expression, which is associated with the distribution
of the density of probability over time with respect to faults.
Random failure models are among the most widely used models to describe
reliability phenomena worldwide.
For a device that needs to be free of failure, the magnitude of the effect of early
failures may be limited by controlling product quality and narrowing the
production process, plus a later stage of wear control before its operating life
begins (burn in and debugging). Wear out failures should be limited if there is
careful preventive maintenance with periodic replacement of parts or components
in areas of the production system where the effect of wear is concentrated. Thus,
attention is mainly focused on failures and the chances of preventing, reducing or
completely eliminating consequences.
In order to do so, it is important to model this kind of failure. The lifetime
distribution that describes failures, which occur at random intervals, where the
number of failures is the same for equally long operating periods, is the
exponential distribution. This is given by (3.5) (Bazovsky 2004)
f (t ) Oe Ot . (3.5)
3.2 Basic Concepts on Reliability 121
where O is a constant called the chance failure rate. Its cumulative distribution is
given by (3.6).
F (t ) 1 e Ot . (3.6)
This reliability formula could be used on devices which are not subject to early
failures, and which have not yet suffered from aging. In others words, the time
where this formula is valid is the useful life of the device. This interval of time
varies widely for different devices. One of the most important aspects of this kind
of distribution is the fact that the reliability of a device is approximately the same
for operating times of equal length. Thus, the time t in (3.6) measures the
operating hours in an arbitrarily chosen operating period of a device, regardless of
for how many hours the device has already been in operation before this specific
operating period. During its useful life, the device is always as good as new. This
is because its failure rate remains the same.
From (3.3) it follows (3.8).
O (t ) O 1/T . (3.8)
f t
1 ( )
E (t ) ³ t e T T. (3.9)
T
0
For early failures as well as failures due to the cumulative effect of wear and tear,
also called failures due to age, it is necessary to define the most appropriate
distributions which model the failure time, the context in which time influences
the failure process. Although the log-normal distribution and the standard
distribution are often used to represent the model that demonstrates the effect of
age, the Weibull distribution is the most universally employed. The following
shows some other distributions used to model the behavior of failures due to wear
and tear as well as early failures related to design problems (O’Connor and
Kleyner 2012).
122 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
The Weibull distribution is widely used, can assume a very wide variety of forms
and is therefore very flexible and it can be used for various types of data (Nelson
2004; Jiang et al. 2001).
The Weibull Distribution can have two parameters, and the probability density
function is given in (3.10):
ª § ·E º
t
E 1 « ¨¨ ¸¸ »
E ªt º « ©K ¹ »
¼.
f (t ) e¬ (3.10)
K «¬K »¼
where:
E - the shape parameter
K -the scale parameter
One can observe a very important role related to E:
E = 1, the failure rate is constant, in which the exponential function is a special
case.
E > 1, the failure rate is increasing. In this case, the corresponding failure phase
caused by wear and tear can be modeled using the bathtub curve.
E <1, the failure rate is decreasing. In this case, the early failure phase can be
modeled using the bathtub curve.
Fig. 3.7 displays the graph for this function
2
1.6
1.2
f(t)
0.8
0.4
0
0 0.5 1 1.5 2 2.5
Fig. 3.7 Weibull probability density function f(t) for ȕ=3 (____); ȕ=0.5 (…….); ȕ=1 (_ . _)
0.8
0.6
R(t)
0.4
0.2
0
0 0.5 1 1.5 2 2.5
Fig. 3.8 Reliability for the Weibull Distribution R(t) for ȕ=3 (____); ȕ=0.5 (…….); ȕ=1 (_ . _)
The function concerning the failure rate is a Weibull density which is given by
(3.12).
E 1
E ªt º
O (t ) . (3.12)
K «¬K »¼
1.6
1.2
O(t)
0.8
0.4
0
0 0.5 1 1.5 2 2.5
Fig. 3.9 Failure rate function for the Weibull distribution O(t) for ȕ=3 (___); ȕ=0.5 (….); ȕ=1 (_ . _)
124 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
1 (ln t [ ) 2
f (t ) exp[ ], 0tf (3.13)
Vt 2S 2V 2
0.045
f(t)
0.03
0.015
0
0 0.5 1 1.5 2 2.5
Fig. 3.10 Density function for Log-Normal Distribution for ȝ=3; ı=0.5 (____);ı =1 (_ . _); ı =1.5
(…….)
§ ln t [ · 1
f (t ) I¨ ¸ , 0t f. (3.14)
© V ¹ Vt
§ ln t [ ·
R (t ) 1 )¨ ¸. (3.15)
© V ¹
3.2 Basic Concepts on Reliability 125
0.75
R(t)
0.5
0.25
0
0 0.5 1 1.5 2 2.5
Fig. 3.11 Reliability for Log-Normal distribution for ȝ=3; ı=0.5 (____);ı =1 (_ . _); ı =1.5 (…….)
§ ln t [ ·
I¨ ¸
O (t ) © V ¹ . (3.16)
§ ln t [ ·
Vt Vt)¨ ¸
© V ¹
device, it is possible to identify periods in which the equipment was used outside
the design conditions. Should such times be long, or even if they are short but
frequent, the chances are that the aging process will change. Thus, the distribution
of failure times can be modified.
For the second category of preventive maintenance actions, the main goal is to
control the level of wear and tear that arises from using the equipment, whether or
not this is due to the conditions of use set in the design project being met. Thus, by
replacing a part or component, it is expected that the condition of the device gets
close to that of the original design and consequently the probability of failure is
reduced.
In practice it is common to say that a device that undergoes routine intervention
type 1 is only undergoing corrective actions, when no preventive action type 2 is
performed. This is because a device, even when its design conditions are assured,
it still degrades and ages.
Moreover, the implementation of actions, aimed only to ensure operation
within the design conditions, is really the basic assumption, in order to enable
estimates of reliability to be made, given that the vast majority of reliability
models assume that such conditions are indeed ensured.
Thus, as a result of not taking actions in advance to avoid failure due to wear,
the equipment will be doomed to fail sooner or later. Understanding these issues is
essential for effective maintenance planning.
3.2.7 FMEA
FMEA (Failure Mode and Effects Analysis) emerged in the 40s, and was derived
from standards set for U.S. military systems. It is a qualitative method, used to
identify potential failure modes and their effects and to make recommendations
regarding measures to be taken to mitigate risks which can impact the reliability of
a system.
The FMEA is structured in tabular form, usually on spreadsheets, where
knowledge and experience of those involved are considered as input for a
historical database. Information can be extracted, for example, from drawings,
process specifications, technical manuals, flow and operational procedures. The
aim of applying FMEA is to identify design issues, the critical process and
maintenance components.
Ericsson (2005) argues that a more detailed version of FMEA is known as:
FMECA (Failure Mode, Effects and Criticality Analysis), where three criteria are
usually defined order to calculate the RPN (Risk Priority Number): severity (S),
occurrence (O) and detectability (D). These three criteria define the RPN (Risk
Priority Number) given by (3.17):
Despite many critical questions related to RPN being raised in the literature
(Zammori and Gabbrielli 2012; Yang et al. 2008; Dong 2007; Braglia et al. 2003;
Puente et al. 2002; Braglia 2000; Chang et al. 1999), the RPN value is used in
many studies as a comparison measure for analysis and investigation.
Although many authors emphasize the importance of differentiating between
FMEA and FMECA, Rausand (2011) states that where the frontier between them
lies is rather vague and there is no reason to distinguish between them.
However, some negative aspects of FMEA deserve attention: a significant time
period is required if they are to be applied effectively and they do not take human
factors into account (Stephans 2004); and FMEA is not useful in the process for
identifying combined failures (Nolan 2011). According to Assael and Kakosimos
(2010), each individual failure is considered as an independent event which is not
related to other system failures, except with regard to subsequent effects that may
arise. However, it can be applied in conjunction with other techniques such as
HAZOP (Hazard and Operability Study) when special investigations into complex
systems are made, for example.
On the other hand, in the literature, many of its positive points are highlighted:
FMEA principles are easy to understand (Stephans 2004); FMEA’s description of
failures provides analysts with a basis for making changes to improve a system
(Assael and Kakosimos 2010); it is a useful tool for making analyses and
recording recommendations for design changes (Ericsson 2005).
Taking these aspects into account, it is observed that the wording of a suitable
model definition to address questions relating to the reliability of equipment
and/or system depends on both the context and the problem type being analyzed.
It is important to note that the objectives of reliability studies can affect the
modeling in different ways. Different goals require different approaches and
methods for modeling and analysis. Furthermore, goals can also directly impact
the choice of the computational approach to be used in the analysis (Aven and
Jensen 2013).
Considering the multiple factors and objectives above mentioned, as dealt with
in Chap. 2, this kind of decision making process and analysis combines the
multiple factors or objective and may incorporate the DM’s preferences over those
factors, therefore, an MCDM/A approach shall be applied.
3.2.9 Simulation
Redundant systems are used in different industrial plants so that the systems
continue operating for longer, even if a failure in a system unit occurs when it is in
operational mode. According to Calixto (2013), two redundancy types can be
observed: passive and active redundancy. In passive redundancy, the redundant
equipment (in standby mode) is for most of the time in a passive state. In other
words, these passive devices operate only when the active equipment fails.
Modarres et al. (1999) add that passive redundant systems are also called
standby redundant systems. The units of this system type remain out of operation
until activated by a sensing and switching device. This process continues to be
carried out until all standby units have been brought into operation and failed. In
this last case, the system is considered failed. Calixto (2013) states that systems
can provide active redundancy, in addition to passive redundancy. This occurs
when similar pieces of equipment perform in conjunction the same function in a
system, in an environment where there is a condition that defines production
losses when several pieces of equipment fail. In some cases the charge distribution
effect may occur, where some items of equipment fail and other pieces of
equipment maintain the same level of production in the system, despite their
degrading faster than usual due to this overload. Since in active redundancy the
components operate constantly, it is expected that the mean time between failures
will be lower than in the case of passive redundancy.
According to Modarres et al (1999), a reliability function for a redundant
system with a standby unit is defined by the following mathematical equation:
f
R p (t ) R I (t ) ³f I (t I )dt I R pp (t I ) R II` (t I ) R II (t t I ) . (3.18)
0
where:
fI(t)I = pdf for failure time (unit I);
Rpp (t)I = sensing and switching device reliability;
R`II (tI) = unit II reliability in standby mode operation;
RII(t- tI) = unit II reliability after coming into operation in time tI.
130 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
Calixto (2013) states that the most part of the redundancy increases projects
and maintenance costs, introducing in many types of organizations risk systems,
such as pipelines and tanks.
However, redundancies are often essential systems designed according to the
requirements and specifications of industrial plant systems (Kuo and Zuo 2003;
Tian and Zuo 2006; Kuo and Zhu 2012). There is an extensive literature regarding
redundant systems (Kuo and Prasad 2000).
The maintenance function has some very specific characteristics, which differ
from those of the project function. For example, there is a clearly defined
beginning and end of each project for which this function is responsible.
Maintenance function does not have a defined period during which each item
or equipment of a system will be under the care of maintenance. Maintenance
function seems to be timeless, since its objective is associated with the perform-
ance of a specific system, so whether the system is supposed to be working or not,
there is no time at or during which maintenance activities can be left aside.
The fact that there is a strong demand for maintenance activities does not mean
that there is a right time at which they should take place. Any time can be the time
to do maintenance and these ranges from corrective maintenance that could
happen randomly due to a failure, to preventive maintenance. Even when a system
is not working, it is possible that some maintenance is being undertaken. Indeed,
for some kinds of systems to which accessibility is very difficult, maintenance
activities can only be done when the system stops operation.
Another interesting feature of the maintenance function that further enhances
the first feature of timelessness is the fact that maintenance is used to cope
with and counteract some natural processes, which never cease interfering in
operations. For example, a set of actions has to be frequently undertaken in order
to reduce the consequences of the aging process, the influence of which, most of
the time, is reflected in bringing about changes in failure behavior. Therefore, if
132 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
any device is under the influence of one or more natural processes, such as,
damage, corrosion, and wear and tear, it is not possible to stop doing maintenance
without increasing the risk of serious consequences. It is necessary to conduct
maintenance actions continuously to achieve system’s desirable performance.
Currently, the maintenance function has received the importance that it
deserves. The point to be made about this is that this recognition of its importance
was not given immediate. Maintenance departments spread around the world
faced critical battles until it was recognized that maintenance is a strategic ally in
the struggle to remain competitive. Indeed, even today, there are, right now, some
companies where these battles are still being waged, and where the maintenance
function is viewed only as an unwelcome and burdensome source of costs. In fact,
even although it is no longer acceptable not to acknowledge the importance of
maintenance, since the impact of poor maintenance on a company´s production
targets is quickly evident, there are still a great many companies that only do the
minimum in terms of maintenance, i.e. they correct what has failed.
The approach of maintenance and the level of importance given to maintenance
activities are related to many different issues. Therefore, the way in which the
maintenance is approached may be specific for each company due to distinct
characteristics of the productive system, and the different levels of development of
the maintenance function. Thus, despite the fact that problems related to
maintenance are similar, the particular features of each company´s production
system, and a set of different matters, such as company culture, make the problem
unique for each company.
Maintenance can be defined as the set of activities that aims to ensure the levels of
performance necessary to guarantee the achievement of production targets. This
could be by avoiding failure or by restoring the operating condition when the
failure has already occurred. In the first case, this is only possible by means of
planned maintenance actions; in the second case, corrective actions are addressed
and their purpose is to change the state of failure so as to restore the operational
status of equipment, fast enough in order to ensure losses will not be high.
By a closer examination of the maintenance problems, there is a key event,
around which a number of different actions must be performed in order to safe-
guard the competitive existence of different production systems. A failure is, in
fact, the non-operating condition of a device or a condition of productive disability
and this often reveals itself with different consequences, which can often be
summarized as monetary losses but at other times result in very negative outcomes
that are difficult to convert into objective monetary values, such as the loss of a
human life and serious damage to the environment.
This diversity in the nature of consequences, coupled with the behavior of
equipment failure, in which the uncertainties governing different fault events, are
major complicating aspects when attempting to establish systematic maintenance
actions. This makes it very difficult to adopt standard procedures, with regard
to dealing with failures when they have already occurred, or anticipating such
failures. They demand a more rigorous treatment, with an emphasis on using
mathematical maintenance models (Dekker 1996).
134 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
The definition of maintenance strategies must be aligned with business goals and
therefore the characteristics of the production system (Pinjala et al. 2006). Thus,
although it is possible for there to be variations in the objectives of maintenance
due to the peculiarities of the system, the main and common objectives in various
sectors in which maintenance is conducted can be identified, such as (Corder
1976):
1. To extend the useful life of assets;
2. To ensure satisfactory levels of availability;
3. To ensure operational readiness of systems, and;
4. To safeguard people who use the facilities.
The first three objectives are, in fact, directly associated with the way, whether
this is good or bad, that the maintenance activities are being performed. On the
other hand, the last objective is rather indirect. Actually, maintenance is not in
charge of safety. But, obviously, each time that a failure with dangerous
consequences for humans and for the integrity of the system is avoided as a result
of maintenance activities, these have contributed to making the plant safer.
3.3 Basic Concepts on Maintenance 135
There is no doubt that these four aspects are in fact the most important
objectives of maintenance. But it may be observed in the literature some variations
in these common objectives of maintenance. For example, Dekker (1996) summarizes
maintenance objectives under four headings, namely: ensuring that the system
functions (availability, efficiency and production quality); ensuring the system’s
life (asset management); ensuring safety; and, ensuring human well-being.
Despite there being some slight differences in the main objectives of maintenance,
the simultaneous achievement of these objectives is not a trivial job. In fact, due to
objectives conflicting with each other, it is very common to adopt only one
objective under the supposition that the one chosen is the one that is the most
closely associated with the strategic objective of the business (Rosqvist et al.
2009; Khazraei and Deuse 2011).
The problem with this approach is the fact that by reducing maintenance
objectives to only the main one, the DM’s view of the problems in the maintenance
field is considerably restricted.
This restriction is even more serious if it is observed the contemporary aspects
that invite the maintenance managers to think more broadly about maintenance
problems. Some of these aspects are listed above (Levitt 2003; Newbrough and
Ramond 1967).
8. The widening of the consequence domain, and the diversification of its nature.
It is quite impossible, at a failure event of some kinds of systems, to track all
the agents affected by one failure, or to determine the nature of this affect;
9. The dangers that arise from managerial mistakes with regard to maintenance
activities that are being currently emphasized (Levitin 2000; Wang and Pham
2006);
10.Finally the fact of the business scenario being so competitive that the aim is to
avoid all failures;
All these aspects emphasize the importance of decision-making in the
maintenance context. To achieve the maintenance objective, the decision maker
develops or follows maintenance policies that are the most appropriate for his/ her
objectives and the characteristics of the production system.
The main challenge for the maintenance manager is to structure the maintenance
procedures and activities to be undertaken in such a way that the strategic
objectives associated with them are achieved. This means that the manager has to
plan the maintenance actions with these objectives in mind.
According to Márquez (2007), a maintenance plan is a structured set of tasks
that includes activities, procedures, resources and time required to perform
maintenance tasks. The implementation of maintenance planning in practice leads
to establishing maintenance policies. Maintenance policy is the process of
coordinating maintenance activities with the particular characteristics of each
system, as well as with the goals that the decision-makers wish to reach, which
reflect the company’s strategic objectives.
To define a maintenance policy a mathematical model is associated with it in
order to make sure that the policy achieves best results. The model, by using a
performance function, defines the levels of each action and these should be used to
optimize this function. The most appropriate action can be defined before defining
its frequency. However, there are models in which not only the activity, but also
the frequency, is defined simultaneously (Scarf et al. 2009).
The next section presents a more structured discussion about how mathematical
models contribute to maintenance management, emphasizing the contribution of
the maintenance policy to the maintenance management process.
durable equipment. Inventories provide items for the production process; durable
equipment provides services. The management of durable equipment, however,
imposes two additional problems, choosing appropriate levels of service for the
equipment and keeping up these services.
Years ago, choosing an appropriate level of service was discussed, while
maintenance costs were assumed to be constant or nonexistent. In the 1960s, how-
ever, an entire theory on the maintenance of equipment started to be constructed.
Optimal maintenance policies were proposed and characterized for a wide variety
of situations.
The first studies on maintenance policies treated the problems as being
deterministic (Taylor 1923; Hotelling 1925), i.e., problems in which the result of
each maintenance action is non-random. Some years later, however, different
studies properly faced up to the stochastic aspects of maintenance problems
(Barlow and Proschan 1965; Barlow and Hunter 1961; Barlow and Proschan
1975; Barlow and Hunter 1960; Glasser 1969), for which the consequences of
the maintenance actions would be random. What most motivated developing
maintenance policies emerged largely and importantly from tackling practical
problems of maintaining complex electronic equipment, i.e., aircraft, missiles,
spacecraft, communications equipment, computers, and so on.
The methodology and the theoretical development related to stochastic
maintenance have a striking resemblance to the stochastic theory of inventory
management. Both have their roots in simple deterministic models. Stochastic
inventory theory models usually assume that for a particular item, demand per unit
time and delivery time are random variables. The corresponding stochastic
elements in the theory of maintenance are the time to failure of equipment as well
as repair time.
Indeed, from a broader perspective, Jorgenson et al. (1967) state that no
distinction is made between the principles of management of inventory control
and of durable goods. For durable equipment, the outputs of conservation
activities are services rather than individual items. The level of conservation
activity depends not only on acquiring a spare part of productive assets, but also
various other inputs of materials and services that represent the maintenance
activity. The output of service equipment can be fed into other activities.
Terborgh (1949) refers to the same issue. According to him, the hand of time
lies heavily on the work or the deeds of men. He also notes that it is a fact that
more practical consequences confront the owner of the item with two problems:
the first is to distinguish the speed of death, or, in other words, to say if an asset
that has not been exhausted physically still has a life of economic usefulness,
either generally or for the particular function it performs. The second is to make
the financial provision, in order to be possible to prevent wear of durable goods
over its service life. Jardine (1973) classifies maintenance policies into two
general classes: probabilistic models of maintenance policies and deterministic
models. A striking difference between these classes of problem, besides the
existence of a stochastic process that governs the events processed within the
138 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
policies, is to check whether or not there has been a complete failure on an item.
For example, for complex systems where the probability of its main function
being stopped is very low. However, since with the passage of time there is a
considerable increase in operating costs, it is quite appropriate to use a deter-
ministic approach which consists of observing a cost function and has features
very similar to inventory control models.
Therefore, Jardine (1973) establishes different subgroups in each class, which
consist of models with similar features:
1. A class of Deterministic Models: models for replacing equipment when
operating costs and use increase; models for replacing equipment when
operating costs use, for use with finite time horizons; models for replacing
equipment considered capital investment, taking into account the discounted
net benefit; models for replacing equipment considered capital investment,
taking into account the technological improvement.
2. Class of Probabilistic Models: Age Replacement based models, taking into
account the time of repair and replacement, and finally, the block replacement
model.
Sherif (1982), discusses different maintenance policies in an article that
summarizes several studies on the subject. He also makes the same classification
as Jardine (1973), although he makes different divisions existing within these
subgroups of two broad classes.
McCall (1965) presents a survey of maintenance policies which is in a quite
different format from those of Jardine (1973) and Sherif (1982)
McCall (1965) does not consider any deterministic model and makes an in-
depth analysis of maintenance policies for systems with stochastic failure.
According to McCall (1965), the development of such policies is based on a
variety of mathematical techniques. This foundation, along with a variety of
applications, sometimes obscures the underlying structure common to all policies.
The first author’s purpose is to identify this common structure, and thus clarify the
relationships between the various maintenance policies.
McCall (1965) classifies models into two categories. The first corresponds
to the class of models called (preparedness) preparation or readiness, when
equipment fails stochastically. Its state, in fact, is not known with certainty.
Alternative maintenance actions for such equipment include inspection and
replacement. Preventive maintenance models constitute the second class of
maintenance models. In these models, the machine is subject to stochastic failures,
and machine status is always known with certainty. If the equipment displays an
increasing rate of failures and, moreover, if it is more costly to repair the fault
when the system is in operation than to replace the equipment before it fails, then
it may be advantageous to replace the equipment before it fails. The problem is to
determine a suitable replacement plan (Nakagawa 1984; Nakagawa 1989).Currently
the first class of models has gained considerable development. Investigating the
state of equipment is now supported by a large number of technological tools, in
3.3 Basic Concepts on Maintenance 139
addition to which the research field has increased and diversified in recent years.
This class has been referred to as condition-based maintenance and has been
declared as a new milestone of a new generation of approaches in the practice of
production and maintenance management (Ahmad and Kamaruddin 2012; Wang
2012; Baker and Christer 1994).
In contrast with this, for the second class, the current preventive maintenance
models are almost the same as the earliest ones. Basically, these models deal with
the process of failure, by observing the lapse of time since the last preventive
maintenance activity (Chang 2014). Despite the small evolution of preventive
maintenance models per se, one of the most important contributions, currently, is
the combination of the two distinct strategies, for instance, checking the state and
replacing preventively after some time. These kinds of combination sometimes are
known as hybrid policies, where both actions could be taken following different
rules. For some examples of this policy (Scarf and Cavalcante 2012; Scarf and
Cavalcante 2010).
Another very important class of policies that is very useful for systems
comprising more than one part or multi-component is known as an opportunistic
policy. The main distinguishing feature of these policies is that maintenance actions
for a piece of equipment depend on the state of the rest of the equipment. The con-
nection established between the states of the components allows more favorable
outcomes to occur when compared with those that could arise from individual
policies for each component. By restricting the vision to one device at a time, the
opportunity to observe actions for multiple components simultaneously is lost, as
are the savings that would be made by dealing with opportunities in the most
intelligent way.
Just as in the preventive maintenance policies, in which the most recent major
contributions are associated with the combination of different actions (Drapella and
Kosznik 2002; Jiang and Jardine 2007; Thangaraj and Rizwam 2001), the combi-
nation of actions was also used in order to improve the opportunistic maintenance
policies. The advantages arising from the combined use of the activities inherent
to the two most typically known groups: preventive and corrective maintenance.
This combination was systematically studied by representatives of the RAND
Corporation in the early sixties (Radner and Jorgenson 1963; McCall 1965;
Jorgenson and McCall 1963). As a result, opportunistic maintenance basically
refers to the situation where preventive maintenance is performed arising from
opportunities related to the choice of a date or constraints due to the impossibility
of postponement, given a failure event. In many cases, it is assumed that the
process of generating opportunity is completely independent of the fault (Dekker
1996). On the other hand, it is common to consider the opportunities that coincide
with the time of failure of individual components. Due to economies of scale in
the cost maintenance function, the undesirable event of a fault in a component is
also considered as an opportunity for preventive maintenance of other components.
One must note that in many situations, a combination of preventive and
corrective maintenance repairs is not realistic. The need for corrective
140 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
The ever increasing need for higher productivity, in the face of growing
competition, has demanded of the various sectors of the economy a constant
search for tools that will enable them to acquire competitive advantages.
In order for these organizations to meet these requirements, it is essential that
their production systems are able to operate under normal conditions; in other
words they must be reliable and available. It is the maintenance function that is in
charge of ensuring the normal operation of these systems. To be successful in this
objective, paying due attention to the maintenance structure is the best way to deal
with common problems related to the management of maintenance.
According to Kelly (1983), maintenance planning is a traditional practice,
recommended for the maintenance of machinery, equipment and tools, and should
be conducted by preparing work plans and setting norms and standards for their
conduct. Márquez (2007) states that what is needed for any level of maintenance is
a structured set of tasks that includes activities, procedures, resources and defines
the time required to perform maintenance tasks. These definitions explain the
scope of maintenance planning:
x What should be done?
x When should it be done?
x Which resources should be employed?
The more correct the answers are to these three points, the more efficient the
planning of maintenance resulting from these issues is. In this sense, effective
maintenance planning enables managers to take actions using the correct equip-
ment, at the right time, and with the proper tools. The successful implementation
of maintenance activities is directly related to precise preplanning.
142 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
The answers to these questions usually follow a hierarchical order. So, initially,
it is essential to specify which activity or activities to conduct on each device;
subsequently establish the frequency with which each of the activities should be
undertaken on each of the items of equipment, and finally define the set of
resources that will be used.
Tajiri and Gotoh (1992) and Shirose (1992) state that a definition of TPM
contains the following five points:
1. It aims at getting the most efficient use of equipment.
2. It establishes a total (company-wide) planned maintenance system, (preventive
maintenance, and improvement related maintenance).
3. It motivates the participation of department workers, equipment operators, and
equipment designers.
4. It involves everyone from top management down.
5. It promotes and implements planned maintenance based on autonomous, small
group activities.
In other words, the goal of TPM is to redesign the system of the company, by
seeking to improve the performance of people and equipment. Improving staff
performance is based on training employees (operators and maintenance workers)
so that they can maintain the machines working as per their specifications, and
when an abnormality occurs, the operator himself is able to identify it and solve
the problem, whenever possible.
Improving equipment consists of structural modifications that represent some
kind of benefit to the yield of machine and operator. Another relevant point is to
reduce future maintenance costs when evaluating the purchase of new machines.
Companies want to increase their productivity and reduce losses. TPM is one of
the tools used to eliminate such losses, which can be classified, according to
Shirose (1992), as:
1. Breakdown losses – these can be failures because of stoppage in the operation,
which is caused suddenly, or by deterioration in function, which is a partial
reduction in the capacity and function of the equipment compared to the
original state. This loss is related to the loss of function of the equipment, and
leads to both chronic failures and sporadic faults, resulting in the loss of time
and productivity;
2. Setup and adjustment losses – these happen when one device produces different
products, so it may take excessively long to adjust the equipment so that it is
able to produce another product with the desired quality;
3. Idling and minor stoppage losses – occur when short breaks are not taken into
consideration, but when added together can result in a high loss of time, and
empty operations;
4. Reduced speed losses – occur when the machine operates for any reason at a
slower speed than normal;
5. Quality defects and rework – occur when there are defects that can lead to
disposing of the product and so time and materials for production are lost,
while defects need to be corrected and to fix them an additional amount of
operating and labor time will be necessary which entails sustaining a loss;
3.3 Basic Concepts on Maintenance 145
6. Startup / yield losses (reduced yield between machine startup and stable
production) – correspond to the period in which the performance conditions of
the equipment after it has been triggered do not reach stable production.
So that TPM may develop in organizations, it is necessary that the foundations,
called pillars, are constructed in teams and coordinated by managers or leaders of
each team. The eight pillars of TPM philosophy form a support system that targets
ensuring productive efficiency for the entire organization. Nakajima (1988) lists
the eight pillars of TPM as being:
1. Autonomous Maintenance - places responsibility for routine maintenance, such
as cleaning, lubricating, and inspection, in the hands of operators;
2. Planned Maintenance - schedules maintenance tasks based on predicted and/or
measured failure rates;
3. Quality Maintenance – detects design errors and prevents them from entering
into production processes. It applies root cause analysis to eliminate recurring
sources of quality defects;
4. Focused Improvement - Small groups of employees work together pro-actively
to achieve regular, incremental improvements in the operation of equipment;
5. Early Equipment Management - directs practical knowledge and understanding
of manufacturing equipment gained through TPM towards improving the
design of new equipment;
6. Training and Education – fill in gaps in knowledge which is required to achieve
TPM goals. Training and educational opportunities are given to operators,
maintenance personnel and managers;
7. Safety, Health, Environment – These are about maintaining a safe and healthy
working environment;
8. TPM in Administration - Applying TPM techniques to administrative functions.
The concept of overall equipment effectiveness (OEE) is an important TPM
topic. It is calculated by multiplying the availability of equipment by its perform-
ance efficiency and by its quality rating. OEE gives a useful measure for tracking
the progress and improvements from the TPM program; but it does not give enough
detail to determine why the equipment is better or worse (Mobley et al. 2008).
According to Nakajima (1989), the results from TPM are an increase in the
machine availability index by decreasing the number of breakdowns; a decrease in
the number of failures in the process and thus a decrease in the number of customer
complaints; a reduction in production costs; and a decrease in the number of
workplace accidents. All this is possible by preparing and developing people,
combined with greater integration between man and machine that operates so as to
improve productivity and increase the competitiveness of the entire organization.
There is a trend for many companies to adopt TPM as a tool since they are
interested in the potential success of this methodology. It is also true that many of
the targets are quite challenging, which is why it is important to motivate people
to seek continuous improvement in order to achieve zero losses in the production
146 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
To be at its most effective, RCM needs to be based on certain factors such as:
x The involvement of engineers, operators and maintenance technicians;
x Due importance being given to the study of the consequences of failures that
drive maintenance tasks;
x The scope of the analysis, as this should include safety issues, the environment,
and operation costs;
x Suitable importance given to proactive activities that involve predictive and
preventive tasks;
x Avoiding hidden failures that reduce system reliability;
According to Moubray (1997) there are seven basic questions that should be
used by an RCM program:
1. What are the functions and associated performance standards of the asset in its
present operating context?
2. In what ways does it fail to fulfill its functions?
3. What causes each functional failure?
4. What happens when each failure occurs?
5. In what way does each failure matter?
6. What can be done to predict or prevent each failure?
7. What should be done if a suitable proactive task cannot be found?
In RCM, the four most important terms are: system, subsystem, functional
failure and mode of failure:
x System: This is the plant as a whole or a subdivision thereof which is identified
in the RCM analysis;
x Subsystem: This is a group of items of equipment and/or components which
together perform one or more functions and can be considered as a separate
functional unit within the system;
x Functional failure: Every subsystem performs a certain function. A functional
failure describes how each subsystem failure occurs;
x Failure Mode - Identifies each specific condition related to a specific piece of
equipment which causes loss of function of a subsystem.
RCM provides functional requirements and standards for the desirable
performance of equipment; For each function, functional failures are defined, and
the failure modes and effects of failures analyzed using FMEA (Failure Modes
and Effects Analysis). Consequences of each failure are analyzed for impacts
arising from the effects of the failure modes. Errors fit into one of four categories:
hidden; associated with safety or the environment; operational; and non-operational.
According to Rausand and Vatn (2008), the RCM analysis process can be
carried out over the following 12 steps:
1. Study preparation;
2. System selection and definition;
3. Functional failure analysis (FFA);
148 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
In contrast, Clemen and Winkler (1999) argue that a group of experts can
defend a course of action that is more risky than that by an individual or a group
of experts reached without discussion. This is probably because experts rely on
information provided by others, or there is a sharing of responsibilities among
experts.
In order to choose which procedure (method) should be used to aggregate
experts’ knowledge, it is necessary to consider pragmatic issues such as cost and
acceptance. Cost considerations generally favor using simpler procedures, such as,
the statistical average. However, when there are considerations that affect
acceptability, it is likely that more complex aggregation procedures for interaction
between experts are more favorable, such as face-to-face procedures or using the
Delphi methodology, for example (Fischer 1981).
Clemen and Winkler (1999) list some general guidelines to determine what
approach to aggregating knowledge from experts should be considered:
x What information is provided by experts? Is the probability distribution
complete? It is not, if there is only partial information on some of these
distributions (e.g., means, variances, etc.);
x Who is involved? A single or a group of experts?;
x What degree of modeling should be performed?;
x What type of aggregation rules are to be used?;
x What parameters necessary for the aggregation method? (e.g., setting weights),
and;
x What is the level of complexity of the aggregation process to be adopted?
In the literature, there are two main approaches to aggregating experts’
knowledge (opinions), when it is represented by a probability distribution: the
mathematical approach and behavioral approach.
The mathematical aggregation procedures consist of analytical models that
work on each individual probability distribution so as to produce a combined
probability distribution. Aggregation in the behavioral approach tries to generate
associations between experts by their interacting with each other and reaching
agreement. This can be face-to-face or may involve the exchange of information
without direct contact. This approach considers the quality of individual
information and dependence between these (Garcez et al. 2011).
Instead of probabilities aggregation, there are other approaches related to fuzzy
logic, which may be found in the literature. Ekel et al. (2009) specify two main
approaches to reach a consensus: first, expert’s opinions are combined into a
collective opinion, using weighted aggregation. The disadvantage of this approach
is when there is an expert who has a deep knowledge of the problem and there is a
discrepancy between her/him and other experts. Also, there is a disadvantage
when an expert can be neglected due to reducing the weight to his/her opinions,
and also defining the set of weights may require significant computational effort.
The second approach, described by Ekel et al. (2009), is to maintain the
weights of each expert constant. To reach consensus, the weight given to the
References 155
expert who most disagrees with the rest of the group is reevaluated. A
disadvantage of this approach is that an expert who disagrees may have to change
his/her opinion drastically (perhaps unjustifiably), or this expert may be repeatedly
asked to revise the opinions of his/her initial position, which requires greater
intellectual effort.
Furthermore, decisions typically require the multiple views of different experts,
as a single person may not have sufficient knowledge about the problem and
therefore cannot solve it alone (Ekel et al. 2009; Parreiras et al. 2010).
As to aggregating experts’ knowledge, even though experts can agree what the
relevant variables to be analyzed are, this does not mean that they have a
consensus on the probability distribution. If they do not disagree on any point,
there is no need to consult more than one expert and therefore there would be no
need to make expert aggregation (Clemen and Winkler 1999).
In other words, the members of the group would have a uniform opinion and
therefore their knowledge would be the same as the one that would have been
made, if there had been only one expert. Although such a situation rarely occurs,
in cases where the consequences of taking wrong decisions are potentially very
serious, when experts are selected, it may be valuable to make a preliminary effort
to determine whether they do disagree over the probability distribution of the
variables.
To increase the overall satisfaction level of the solution (collective opinion),
experts should have the chance to influence the consensus by providing
information about their individual knowledge.
References
Ackermann F, Howick S, Quigley J, et al. (2014) Systemic risk elicitation: Using causal maps to
engage stakeholders and build a comprehensive view of risks. Eur J Oper Res 238:290–299
Ahmad R, Kamaruddin S (2012) An overview of time-based and condition-based maintenance in
industrial application. Comput Ind Eng 63(1):135–149
Ale B, Burnap P, Slater D (2015) On the origin of PCDS – (Probability consequence diagrams).
Saf Sci 72:229–239
Alencar MH, de Almeida AT (2010) Assigning priorities to actions in a pipeline transporting
hydrogen based on a multicriteria decision model. Int J Hydrogen Energy 35(8):3610–3619
Al-Kassab J, Ouertani ZM, Schiuma G, Neely A (2014) Information visualization to support
management decisions. Int J Inf Technol Decis Mak 13(2):407–428
Andrews JD, Dunnett SJ (2000) Event-tree analysis using binary decision diagrams. Reliab IEEE
Trans 49(2):230–238
Andrews JD, Moss TR (2002) Reliability and risk assessment. Wiley-Blackwell, Suffolk
Arendt JS, Lorenzo DK (2000) Evaluating process safety in the chemical industry: a user’s guide
to quantitative risk analysis. American Chemistry Council
Assael MJ, Kakosimos KE (2010) Fires, explosions, and toxic gas dispersions: Effects
calculation and risk analysis. CRC Press Taylor & Francis Group, Florida
Aven T (2008) Risk analysis: assessing uncertainties beyond expected values and probabilities.
John Wiley & Sons, Chichester
156 Chapter 3 Basic Concepts on Risk Analysis, Reliability and Maintenance
Aven T (2012) Foundations of Risk Analysis. 2nd ed. John Wiley & Sons, p 198, Chichester
Aven T, Jensen U (2013) Stochastic Models in Reliability. Stochastic Modelling and Applied
Probability. Springer, New York
Aven T, Vinnem JE (2007) Risk Management with applications from the offshore petroleum
industry. Springer Series in Reliability Engineering. Springer-Verlag, London
Baker RD, Christer AH (1994) Review of delay-time OR modelling of engineering aspects of
maintenance. Eur J Oper Res 73(3):407–422
Barlow R, Hunter L (1960) Optimum Preventive Maintenance Policies. Oper Res 8:90–100
Barlow RE, Hunter LC (1961) Reliability Analysis of a One-Unit System. Oper Res 9:200–208
Barlow RE, Proschan F (1965) Mathematical theory of reliability. John Wiley & Sons,
New York
Barlow RE, Proschan F (1975) Statistical theory of reliability and life testing: probability
models. DTIC Document
Bazovsky I (2004) Reliability theory and practice. Dover Publications, Mineola
Bedford T, Cooke R (2001) Probabilistic Risk Analysis: Foundations and Methods. Cambridge
University Press, New York
BeńDaya M (2000) You may need RCM to enhance TPM implementation. J Qual Maint Eng
6(2):82–85
Bérenguer C, Grall A, Dieulle L, Roussignol M (2003) Maintenance policy for a continuously
monitored deteriorating system. Probab Eng Informational Sci 17(2):235–250
Berger JO (1985) Statistical decision theory and Bayesian analysis. Springer Science & Business
Media, New York
Birolini A (2014) Reliability Engineering. Theory and Practice. Springer Berlin Heidelberg
Bostrom A, Anselin L, Farris J (2008) Visualizing Seismic Risk and Uncertainty. Ann. N. Y.
Acad. Sci. Blackwell Publishing Inc, pp 29–40
Braglia M (2000) MAFMA: multíattribute failure mode analysis. Int J Qual Reliab Manag
17(2):1017–1033
Braglia M, Frosolini M, Montanari R (2003) Fuzzy TOPSIS approach for failure mode, effects
and criticality analysis. Qual Reliab Eng Int 19(5):425–443
Brissaud F, Charpentier D, Fouladirad M, et al. (2010) Failure rate evaluation with influencing
factors. J Loss Prev Process Ind 23(2):187–193
Brito AJ, de Almeida AT (2009) Multi-attribute risk assessment for risk ranking of natural gas
pipelines. Reliab Eng Syst Saf 94(2):187–198
BS EN 60706-2 (2010) Maintenance. Maintenance terminology, British Standards Institution
Calixto E (2013) Gas and Oil Reliability Engineering: Modeling and Analysis. Gulf Professional
Publishing, Oxford
Carter ADS (1986) Mechanical reliability. Macmillan London
Chang C, Wei C, Lee Y (1999) Failure mode and effects analysis using fuzzy method and grey
theory. Kybernetes 28(9):1072–1080
Chang C-C (2014) Optimum preventive maintenance policies for systems subject to random
working times, replacement, and minimal repair. Comput Ind Eng 67:185–194
Clemen RT, Winkler RL (1999) Combining Probability Distributions from Experts in Risk
Analysis. Risk Anal 19:187–203
Corder AS (1976) Maintenance management techniques. McGraw-Hill
Cox LA Jr (2009) Risk analysis of complex and uncertain systems. Springer Science & Business
Media, New York
Crowl DA, Louvar JF (2001) Chemical Process Safety: Fundamentals with applications. Prentice
Hall, Boston
Dekker R (1995) On the use of operations research models for maintenance decision making.
Microelectron Reliab 35(9):1321–1331
Dekker R (1996) Applications of maintenance optimization models: a review and analysis.
Reliab Eng Syst Saf 51(3):229–240
References 157
Shirose K (1992) TPM for Workshop Leaders. Productivity Press, New York
Smith DJ (2011) Reliability, Maintainability and Risk. Practical methods for engineers. BH
(Elsevier), Oxford
Smith DJ, Simpson KGL (2010) Safety Critical Systems Handbook. A straightforward guide to
functional safety, IEC 61508 (2010 Edition) and related standards. BH, Oxford
Stephans RA (2004) System safety for the 21st century. The update and revised edition of system
safety 2000. Wiley-Interscience, New Jersey
Sutton I (2010) Process Risk and Reliability Management Operational Integrity Management,
William Andrew – Elsevier, Oxford
Tajiri M, Gotoh F (1992) TPM implementation, a Japanese approach. McGraw-Hill Companies
Tariq M (2013) Risk-based flood zoning employing expected annual damages: the Chenab River
case study. Stoch Environ Res Risk Assess 27:1957–1966
Taylor JS (1923) A Statistical Theory of Depreciation. J Am Stat Assoc 18(144):1010–1023
Terborgh GW (1949) Dynamic equipment policy. McGraw-Hill Book Co
Thangaraj V, Rizwam U (2001) Optimal replacement policies in burn-in process for an
alternative repair model. Int J Inf Manag Sci 12(3):43–56
Theodore L, Dupont RR (2012) Environmental Health and Hazard Risk Assessment. Principles
and Calculations. CRC Press, Boca Raton
Tian Z, Zuo MJ (2006) Redundancy allocation for multi-state systems using physical
programming and genetic algorithms. Reliab Eng Syst Saf 91(9):1049–1056
Tweeddale M (2003) Managing Risk and Reliability of Process Plants. Gulf Professional
Publishing. Burlington
Van Leeuwen CJ, Vermeire TG (2007) Risk Assessment of Chemicals. An introduction.
Springer, Dordrecht
Vinnem J-E (2014) Offshore Risk Assessment. Principles, Modelling and Applications of QRA
Studies Vol. 2, Springer-Verlag, London
Walley P (2002) Reconciling frequentist properties with the likelihood principle. J Stat Plan
Inference 105:35–65
Wang H, Pham H (2006) Reliability and Optimal Maintenance. Springer-Verlag, London
Wang W (2012) An overview of the recent advances in delay-time-based maintenance
modelling. Reliab Eng Syst Saf 106:165–178
Winkler RL (1981) Combining Probability Distributions from Dependent Information Sources.
Manage Sci 27:479–488
Winkler RL, Cummings LL (1972) On the choice of a consensus distribution in Bayesian
analysis. Organ Behav Hum Perform 7:63–76
Winkler RL, Hora SC, Baca RG (1992) The quality of expert judgment elicitations. San Antonio,
TX: Center for Nuclear Waste Regulatory Analyses
Woodman RC (1967) Replacement policies for components that deteriorate. OR 18:267–280.
Yang Z, Bonsall S, Wang J (2008) Fuzzy Rule-Based Bayesian Reasoning Approach for
Prioritization of Failures in FMEA. Reliab IEEE Trans 57:517–528
Yoe C (2012) Principles of Risk Analysis – Decision Making under uncertainty. CRC Press,
Boca Raton
Zammori F, Gabbrielli R (2012) ANP/RPN: a multi criteria evaluation of the Risk Priority
Number. Qual Reliab Eng Int 28:85–104
Zheng X (1995) All opportunity-triggered replacement policy for multiple-unit systems. Reliab
IEEE Trans 44(4):648–652
Zheng X, Fard N (1991) A maintenance policy for repairable systems based on opportunistic
failure-rate tolerance. Reliab IEEE Trans 40(2):237–244
Zio E (1996) On the use of the analytic hierarchy process in the aggregation of expert judgments.
Reliab Eng Syst Saf 53:127–138
Zio E (2007) An Introduction to the basics of Reliability and Risk Analysis. Series in Quality,
Reliability and Engineering Statistics vol 1. World Scientific, Singapore
Chapter 4
Multidimensional Risk Analysis
(in 1978), Derbyshire (in 1980), Herald Free Enterprise (in 1987) and Piper Alpha
(in 1988) many maritime sectors started to seek for the improvement and
application of risk modeling and decision making techniques (Wang 2006). Such
risk evaluations and safety concerns has to be observed not only by companies
involved with the specific context, as maritime transportation and offshore
operations, but also other companies related with the sector, in this case, ship
designers and shipbuilders in order to improve safety (Guedes Soares and Teixeira
2001).
Since the widely accepted notion of risk (see Chap. 3) is also based on
consequences, there is a need to estimate consequences/loss/severity. For many
authors, risk assessment deals with estimating possible losses and is an essential
procedure, whose outcome is the foundation on which the DM justify his/her
decisions. Thus, the consequences of an event can be represented based on some
of these aspects, e.g., the number of fatalities, the number of people injured,
financial loss, damage to property, environmental losses, etc. (Alencar et al. 2014;
Alencar and de Almeida 2010; Brito et al 2010; Luria and Aspinall 2003).
Furthermore, both Individual Risk and Societal Risk concepts only consider the
scale of human loss. Cox (2009) argues that, in the risk context, a rational decision
making seeks to ensure that a risk analysis builds evaluations and comparisons of
proposed risk management actions and interventions, not merely describing the
current situation.
In some studies, under a more conservative view, the risk of human loss is
assessed from the perspective of a result of the occurrence of injuries, and not only
fatalities. Therefore, precisely how people are injured (for example, first or second
degree burns) are considered as a consequence for the calculation of risk (Brito
and de Almeida 2009). In this context, Cox (2009) reinforces that dominated
actions should be eliminated, choosing the best option among non-dominated
alternatives and guaranteeing that those alternatives are not ignored. An evaluation
of the total consequences is necessary, in order to provide an effective risk
management. For each alternative, the overall consequences are calculated taking
into account the summing of all the impacts of proposed alternatives on humans
exposures.
Although, there is a need for a multidimensional risk view, under the relevant
aspects of various ways of analyzing the result of accidents in industrial plants and
modes of transportation in various parts of the world, most of the studies consider
only issues related to a single dimension. Doing so makes a single dimension
approach inadequate or incomplete when the issues involved are complex.
Furthermore, nowadays, analyzing the consequences must satisfy the expectations
of society, the state (public sector) and private companies. The magnitude and
severity of the consequences make it essential to develop a more appropriate and
efficient form of risk management, which provides for positive outcomes. In this
sense, Beaudouin and Munier (2009) present a critic on industrial risk manage-
ment techniques based on procedures derived from health, safety, and environ-
ment within quality management programs, and draw attention that decision
4.1 Justifying the Use of the Multidimensional Risk 163
For most of the contexts, besides risk, the cost-benefit analysis is appropriate
for capturing the society priorities, rather than individual preferences, which are
better captured by MCDM/A methods. The former is related to Societal Risk
concept, whereas the latter is related to Individual Risk concept. This issue is
related to the comparison between the use of cost-benefit analysis and MCDM/A
methods (Almeida-Filho and de Almeida 2010a).
Another important point to be highlighted is that financial losses cannot always
be measured with complete accuracy. This occurs due to firms having to take
account of pressure groups in society who are well informed of possible dangers
that their companies may present to society at large. To counter this, several
companies have a strategy for differentiating themselves from their competitors.
This includes creating an image that the care for the environment, and that their
first priority is to safeguard the safety of their employees, their customers and the
community they form part of. Nevertheless, when any kind of accident occurs that
bring losses to any of the “users” of the system, there is pressure from society not
to consume the products of this company. These results in losses to the company
that are not only brought about by the accident itself (in the monetary perspective)
but also because they lose customers and suppliers; contracts may be broken; their
business image is damaged. These losses cannot be “easily” or completely
(precisely) measured in financial terms.
Hence, traditional approaches to risk analysis do not consider the multiple
dimensional impacts (consequences) that industrial accidents may cause. How-
ever, a multidimensional risk view, in many different contexts is necessary.
Furthermore, nowadays, analyzing the consequences must satisfy the
expectations of society, the state (public sector) and private companies. The
magnitude and severity of the consequences make it essential to develop a more
appropriate and efficient form of risk management, which provides for positive
outcomes. In other words, the results must be at an acceptable level of safety and
also from an economic point of view, the survival of the company is necessarily
called into question in the sense that the cost of taking measures that prevent and
mitigate risks has to be balanced against the likelihood of accidents happening to
people or extensive damage to the environments exposed to them.
As already shown in several studies, an approach to risk assessment that uses
only a single dimension of risk cannot be sufficiently comprehensive to ensure
that the most realistic and efficient assessment of risk is made (Alencar and de
Almeida 2010; Apostolakis and Lemon 2005; Brito and de Almeida 2009; Brito et
al. 2010; Garcez and de Almeida 2014b; Lins and de Almeida 2012; Morgan et al.
2000; Willis et al. 2005). Additionally, for Brito and de Almeida (2009), even if
other effects are not as important as the risks to human beings, they also require
substantial attention from DMs.
Hence, there is a need for a multidimensional assessment, which enables more
consistent decision-making to be made and which takes into account the context of
uncertainty risk. In many decision problems, more than one factor influences the
4.1 Justifying the Use of the Multidimensional Risk 165
DM’s preferences with respect to possible outcomes (Montiel and Bickel 2014;
Bedford and Cooke 2001).
According to Salvi et al. (2005), if environmental assessment and risk
management is given more importance, all stakeholders will take part in the
decision-making process. Probably, this feature results from the development of
society, which has increasing access to global information, and this is combined
with people’s concerns related to the sustainable development of society. Society
has also taken a cautious attitude due to experiences caused by industrial disasters
(e.g., Flixborough, Chernobyl, Bhopal, and more recently Fukushima on 11 March
2011).
The occurrence of these and other disasters have shown that there must be
public consultation with various stakeholders, and that this dialogue should not
occur independently of the risk management process, the main objective of which
is to ensure the long term security of populations. Therefore, the maintenance and
consent of an industrial activity is strongly dependent on society accepting the
risks that the activity generates.
Hence, companies are coming to recognize the need to take the different
opinions and preferences of the various stakeholders into account in the risk
decision-making process. MCDM/A methodologies can be extremely useful to
aggregate these different opinions (criteria, preference, weights) so the most
appropriate decision may be taken both at the national and local level (Roy 1996).
Therefore, according to Yoe (2012), what is observed is that a process of
decision making can be simple or complex depending on a few factors that need to
be considered. When the analysis is of a single dimension of the problem and
there is only one DM, the process is simpler. The same is not true under the risk
management process. It is considered to be a complex process due to there being a
number of aspects, such as, the views of interested and involved parties, the
processes of identification, analysis and risk assessment, and the analysis of
consequences, not merely financial impacts.
The process of risk management involves managers and stakeholders with
different values, priorities and objectives. In this process, consideration is given to
such aspects as tradeoffs between risks, costs, benefits, social values and other
impacts of conflicts of values as a result of many perspectives represented by
stakeholders in the decision making process.
Hobbs and Meier (2000) affirm that MCDM/A methods present many positive
as well as negative aspects. The positive points are:
x Emphasis on learning and understanding by the users;
x Tradeoffs more explicit as to the interests involved;
x Values obtained directly by the stakeholders;
x Reject dominated alternatives.
166 Chapter 4 Multidimensional Risk Analysis
This section presents an MCDM/A procedure for building risk evaluation and
decision models, which is adapted from that of Chap. 2, incorporating a specific
situation, in which the DM’s behavior regarding to risk (prone, neutral, averse)
can be approached via utility theory. According to Cox (2012), the application of
utility functions rather than simple risk formulas – composed by terms such as
exposure, probability and consequence - allows to take into account DM’s risk
attitudes, improving the effectiveness of the decision making process to reduce
risks. This procedure has been applied in several contexts described in next
section (Brito and de Almeida 2009; Brito et al. 2010, Alencar and de Almeida
2010; Lins and de Almeida 2012; Garcez and de Almeida 2014a; Garcez and de
Almeida 2014b).
168 Chapter 4 Multidimensional Risk Analysis
What are the What are the hazard What are the
alternatives? scenarios (state of hazard scenarios
nature)? probabilities?
Risk ranking
Another important aspect that has prompted developing the model is that it uses
Decision Support Systems (DSS) to assist in routing between steps in order to
make the process more dynamic, thus making it possible for the DM to make a
more detailed study of all stages of the steps of the risk analysis. Furthermore, use
of DSS aims to support the decision making process, and takes into account both
technical aspects such as its stochastic nature and the variety of the parameters
which will be entered into the model as well as factors related to the decision
making process on risk analysis (Lopes et al. 2010).
Finally, it is worth mentioning that the steps of the proposed methodology are
not static. In other words, there is a transition between steps which allows the DM
to return to the previous steps to adjust a parameter in order to make the result
more dynamic and realistic. Further details of these aspects will be observed
throughout the text.
170 Chapter 4 Multidimensional Risk Analysis
This is the stage used to define who will be responsible for the decision, since it is
this DM’s preference structure which will be adopted. It is extremely important to
identify the DM correctly because decision making in complex environments
(such as transport systems for hazardous products, electric power systems, nuclear
systems, critical infrastructure, etc.) involves potential severely adverse impacts
on society, the environment, economic losses, etc.
Therefore, it is necessary that the DM is thoroughly familiar with the context of
the risk analysis. For example, he/she must be fully alert to possible accident
scenarios, be fully aware of the consequence dimensions of accidents, and be able
to draw up and implement protective and mitigation measures. In other words, not
only must the DM be knowledgeable about the context in which decisions about
risk may have to be taken but also about the needs of the various stakeholders
involved in the decision making process.
It is worth mentioning that the DM’s preferences should reflect the interests
and goals of the organization (company) and also of the managers who are
responsible for any consequences arising from the decision. In some situations it is
necessary to include the preferences of various DMs. This process is characterized
as a group decision, which may involve three main actors: the company
representative of the system considered, the government representative (regulators)
and the representative of the community in which the system is located.
In this model, it is assumed that there is a single DM who fully meets the
requirements of having the necessary experience, the required level of
responsibility and thorough knowledge of the system. This DM is responsible for
seeing to it that public safety (regulatory body) standards are met, and as DM
assumes appropriate responsibilities to society.
4.2 Multidimensional Risk Evaluation Model 171
Additionally, is worth noting that the information from the risk management
should serve as input that should be passed on to other managers with a view to
guiding them on how to perform their functions more adequately. This applies to
such managers as those in charge of maintenance, health, environment and safety,
or even the production manager. The DM can also be the planning or project
manager, where there is an already established system or new systems are being
implemented. Thus, the proposed model can be applied to systems that are not yet
in operation or those that will be developed. It will determine which alternatives
will require most attention in the project design or project execution stage. This
should then lead to preventive and mitigation measures being drawn up and taken
so as to minimize risks at the project level.
Apart from the DM, another person who has an important role to play in the
decision-making process is the expert. Experts provide technical and theoretical
support to assist the DM with any questions or issues that may influence the
decision-making process. Since this model is intended as a tool that assists risk
management. Some experts with relevant knowledge who perform important
functions in the organization can be included.
On some specific occasions, the DM plays the role of DM and an expert at the
same time, due to his/her having technical knowledge regarding some related to
such matters as likelihood, repair times, failure rates, and the characteristics of the
system. The DM’s preference structure is also incorporated into the problem since
it reflects the preference structure of the company, represented by managers’
decisions. However, this is not necessarily a requirement of the model. The model
allows preference aggregation, when the DM is aided by several specialists. This
occurs when the DM does not have the necessary knowledge about specific
information.
This step consists of defining all the possible scenarios which have resulted from
system/subsystem failure modes. These scenarios describe the set of states of nature
= {11, 12, …, 21, 22, …, jk } related to the failure mode j and the resulting
hazard scenario k.
Hazard scenarios do not define the causes of the failure mode or accidents, but
rather the phenomena or accidents associated with the failure mode, which are
influenced by the type of failure mode and by the existence of other interacting
factors (e.g. there is immediate or delayed ignition and a confined space).
In this context, Crowl and Jo (2007) state that accidents originate from
incidents. An incident can be defined as a loss of control over a material or form
of energy. Many incidents are followed by a series of events which propagate
accidents. This can include fire, explosions and toxic gas leaks. According to the
172 Chapter 4 Multidimensional Risk Analysis
At this stage of the model, the alternatives are defined for the DM. The
multicriteria decision model produces a risk hierarchy related to the company’s
systems or subsystems, and it is these which are the alternatives.
In an alternative, the features must be homogeneous, and take into
consideration both technical and social issues as well as aspects that influence the
probability of a hazard scenario occurring. Expert opinion is important, because it
is the expert who has prior knowledge about the behavior of the system. For
example, for technical issues related to a natural gas pipeline system, extremely
important characteristics include the diameter of the pipe, gas pressure, age of the
pipe, characteristics of the soil, composition of the pipe material, the corrosion
protection used, etc. These factors along the sections (alternatives) impact on the
variation in failure rates and the consequences of accidental releases of natural gas
from the pipeline (Jo and Ahn 2002; Jo and Ahn 2005; Sklavounos and Rigas
2006; Jo and Crowl 2008; Brito and de Almeida 2009; Garcez et al. 2010; Alencar
et al. 2010; Brito et al. 2010)
Regarding environmental dimension, characteristics that could be considered
include the type of the surrounding vegetation, the presence of wildlife exposed to
risk, the importance degree of the environment, environmental impact, etc. As to
the human dimension, characteristics that should be considered include land use,
population density and community type.
Returning to the context of natural gas pipelines, Henselwood and Phillips
(2006) assert that these factors may influence the likelihood of an accidental
ignition of a natural gas leak. As an example, in an industrial region, the ignition,
due to the presence of large numbers of ignition sources, of leaking gas is more
likely than in a rural area, where the population density and the presence of
ignition sources are low. More details about these aspects are given in Brito and
de Almeida (2009), Alencar and de Almeida (2010) and Lins and de Almeida
(2012).
Finally, it is important to emphasize that the uniformity of the characteristics
listed above in each system/subsystem comprises a distinct discrete set (A = {a1,
a2, …, an}), where the final system is the sum of all the subsystems analyzed.
4.2 Multidimensional Risk Evaluation Model 173
At this stage objects that are exposed to impacts due to an accident scenario
having occurred Tjk will be analyzed in a particular alternative i, and in the
different consequence dimensions (C = {c1, c2,…, cr,…, cm}) considered. As
mentioned earlier, these consequence dimensions may consider impacts on human
health, environmental impacts, financial loss, company image losses, operating
loss, etc.
For each hazard scenario and alternative, mathematical models are used and
numerical applications made on several features of the objects in the surroundings
exposed to hazard. Through this mathematical study, possible impacts are
estimated on the different consequence dimensions considered.
However, in the first place, it is necessary to determine what the area or danger
zone (Si) is that results from each scenario and each specific alternative. Having
done so, estimates can be made of the impacts and consequences in the
dimensions considered in a particular alternative. The danger zone, according to
DziubiĔski et al. (2006), is a region where impacts exceed critical limits, causing
injury to persons, property and environment losses.
174 Chapter 4 Multidimensional Risk Analysis
During this stage possible impacts (consequences) or payoffs that arise from
accident scenarios (Tjk), are verified, in a danger zone (Si) which has been defined
in the previous step.
The model consists of a set of multidimensional consequences involving risks.
For each consequence dimension considered, the maximum impacts (losses)
resulting from an accident should be defined.
The first step consists of ensuring that the DM understands the purpose of the
utility function and the consequence space. Therefore, one of the most important
insights the DM can have is the issue that there is no great preference to be
defined, but rather a set of consequences in which the DM demonstrates his/her
preferences. As preferences are the DM’s subjective representations, there is not a
correct choice.
Before engaging with the utility elicitation procedures, it is essential to
familiarize the DM with concepts such as: decision analysis, utility functions, and
lotteries. Details of these concepts can be found in Keeney and Raiffa (1976), Roy
(1996) and Vincke (1992).
Another relevant aspect, according to Keeney and Raiffa (1976), concerns the
Von Neumann-Morgenstern expected utility that can be used to characterize an
individual risk attitude through simple lotteries.
The concept of a simple lottery can be seen in the following example where the
DM maker has, certainly, an amount of money to gamble (e.g. $t.00) and needs to
set the probability value p that makes him indifferent towards two situations:
keeping the money or making the lottery bet. In other words, the DM remains
indifferent between having $t.00 with certainty and risks in a lottery with two
possible outcomes: receiving an X amount with probability p or losing the game
with probability 1 – p. Graphically, this may be represented by Fig. 4.2.
$x
p
$t
1-p $y
When the DM understands the concepts, the structure of the decision problem
and the consequence space are established. To reach a better understanding of this,
an example with three consequence dimensions (c1, c2, c3) will be presented (Brito
and de Almeida 2009; Alencar et al 2010; Garcez et al. 2010; Brito et al 2010),
where c1 represents losses in the human dimension (e.g.: the number of people
exposed to fatality), c2 represents losses in the environmental dimension (e.g.:
a vegetation area exposed to fire) and c3 represents the losses in the financial
dimension (e.g.: the maximum monetary amount disbursed). A graphical
representation of this is given in Fig. 4.3.
176 Chapter 4 Multidimensional Risk Analysis
Fig. 4.3 Graphical representation of the consequences space of the MAU function
If there is any inconsistency in the DM’s answers (the DM must state his/her
highest preference for one of these points: Tc1, Tc2 or Tc3), the DM must be given a
new explanation that will lead him/her to a correct understanding of the limits of
the consequence space and the conceptual basis of utility theory.
According to Keeney and Raiffa (1976), some independence utility
assumptions should be verified after defining the limit values of the utility
functions and checking that the DM understands them correctly.
According to Alencar and de Almeida (2010), an attribute c1 is additively
independent of an attribute c2 if two lotteries are equally preferable for all (c1 and
c2) and for a ‘c1 and c2’ arbitrarily chosen, as presented in Fig. 4.4.
cc cc’
p p
and
According Figueira et al. (2005) when attributes (from the perspective of the
Von Neumann-Morgenstern utility model) and the DM’s preferences are
consistent with the conditions of utility independence, then u(c1, c2, …, cr, , …, cm)
can be decomposed into additive, multiplicative or another well-defined structure
in order to simplify the evaluation of these relations.
The MAUT can be expressed in an additive form, if and only if, cr attributes
are mutually independent in utility and the additive independence between the
attributes is observed. Then:
m
u ¦ k uc
r 1
r r (4.1)
value of c w0
corresponds to the consequence c10 ,c20 and the value c1w is
equivalent to the consequence c11,c12 .
Once the p value is defined, the DM is asked about the value of q in which
he/she is indifferent between the certainty of having consequence c10 ,c20 or
playing the lottery , q, . Having obtained the p and q estimated
c10 , c20 c11 , c12
m
values and the condition ¦ k 1 , the following may be defined: k = p, k = 1 2
r 1 r
(1 – p)q and k3 = (1 – p)(1 – q).
The last step consists of verifying the consistency and the variability of the
results if some parameters are modified. Due to the associated uncertainty related
to the parameters of the model, this phase can capture the impact of the results by
using sensitivity analysis on the model.
Several uncertainties are present in scenarios (Tjk) and estimating hazard zones
(Sn), as shown in the earlier stages of the model. These uncertainties are
undesirable, because it becomes impossible to define deterministically which
multidimensional consequences can occur due to an accident scenario. For this, it
is necessary to estimate the probability distributions of the consequences,
represented by a consequence function P, defined by the probability of obtaining a
consequence p, since a scenario Tjk occurred in alternative ai.
In this step of the model, there is a need to estimate the joint probability
distribution over the possible values in “m” consequence dimensions P(c1,…,m |Tjk ,ai)
for each alternative and hazard scenario adopted.
According to Brito and De Almeida (2009), in some contexts it may be
considered that different consequence dimensions can have small or even
negligible correlations between them. This is because the hazard radius covers
several dozen meters. The combination of these consequence dimensions occur
randomly and independently, depending on the specific characteristics of each
alternative, so that the probabilities P(c1|Tjk ,ai),…,P(cr|Tjk ,ai),…,P(cm|Tjk ,ai) can
be estimated independently.
However, in some risk analysis contexts, the probability distributions of these
consequences are not treated independently, as is the case of risk analysis
regarding petroleum extraction platforms, nuclear power plants, etc. where the
4.2 Multidimensional Risk Evaluation Model 179
danger zones usually extend over a wide area and the size of impact interferes
non-randomly in various consequence dimensions.
In the case of probability distributions independent of consequences it is
possible to define mathematical formulations to model the consequence functions
for each loss independently.
There are several models in the context of natural gas pipelines considering
consequence functions for estimating the human, environmental and financial risk
dimensions (Brito and de Almeida 2009; Garcez et al. 2010; Alencar et al. 2009;
Brito et al. 2010). Similarly, for the context of hydrogen gas pipelines with
required adaptations, same approach is considered for estimating risk dimensions
(Alencar and de Almeida 2010; Lins and de Almeida 2012). A model for risk
evaluation in underground vaults of an electricity distribution system considers the
same decision analysis principles for assessing risk dimensions of human impacts,
financial losses, operating losses and disturbance on the local transit vehicles
(Garcez and de Almeida 2014b).
cr
u c r | T jk , ai (4.2)
It can be considered that the consequences are results of the impact dimension
of a given action, which can be estimated by using a probability distribution
function P(c1,…,m |Tjk ,ai).
Keeney and Raiffa (1976) point out that if an appropriate utility is assigned
to each possible consequence and the expected utility of each alternative is
calculated, what is observed as the best course of action is an alternative with the
highest expected utility. Thus, the consequence utility is the expected value of the
utility:
cr E >u c r @ ³ Pc uc dc
r r r (4.3)
cr
u T jk , ai u Pcr | T jk , ai ³ Pcr | T jk , ai ucr dcr (4.4)
cr
m § § ··
¨ ¨ ¸¸
r ai ¦¦ ¨ S ai T ¨ u c r P c r | T jk , ai dc r ¸ ¸ 1S ai T N
³ (4.5)
rT 1¨ ¨ cr
¸¸
© © ¹¹
Thus, the risk concept based on Decision Theory assesses the consequences (cr)
of the hazard scenarios (Tjk), by combining both uncertainties associated with: (i)
the consequences P(cr|Tjk ,ai); and (ii) the hazard scenarios Sai(Tjk).
Additionally, the risk measure used considers the DM’s preference structure in
the set of expected consequences, through utility functions u(cr), representing the
“desirability” that the DM has about property losses (in this particular case, the
consequences of an accident scenario occurring) and allowing a probabilistic
evaluation of the consequences under uncertainty.
These risk measures comprise a descending risk hierarchy of several of the
alternatives (ai) evaluated. Consequently, the results of this hierarchy serve as
input to the decision-making process and risk management.
Natural gas is a fossil fuel with reserves available in many parts of the world. Its
use has grown over the last 30 or so years due to a number of factors, including,
for example, economic and environmental aspects. The high demand for it in
widely scattered different locations requires a mode of transportation to convey
large amounts of gas from its source to its destination, quickly and safely. Thus,
among the existing modes of transportation, pipelines stand out. Although using
pipelines is considered a safe system, some accidents have occurred over the
years, some of which have had critical consequences.
182 Chapter 4 Multidimensional Risk Analysis
Po1 2 d 5 4
CDR # 1,512 1 4
(4.6)
L
Settled danger areas for each section, and the human, environmental and
financial consequences should be defined. This set of consequences will be
included in the analysis using the model, for which the most pessimistic values in
each consequence dimension will be input.
The proposed model seeks to assess risks considering three risk dimensions in
natural gas pipelines: Human Risks (rh), Financial Risks (rf) and Environmental
Risks (rm). The reasons why it is primarily these dimensions that are considered
are based on values that are normally found in both productive organizations and
in other organizations or institutions involved. These will be translated into
principles of social and environmental responsibility and ethical aspects of human
relationships. These aspects should influence company actions that seek to secure
the financial return aimed at.
4.3 Risk Decision Models 183
As to the human dimension, Brito and de Almeida (2009) assume that human
consequences are estimated by the number of people affected physically due to a
particular accident scenario, and who receive at least second degree burns, and not
necessarily by the number on fatalities.
With regard to the environmental dimension, the area of vegetation affected is
used as a measure for the environmental consequences, taking into account the
extent of environmental impacts caused by this type of accident (Alencar et al.
2010; Garcez et al. 2010; Brito et al. 2010; Alencar et al. 2014).
Finally there is the financial dimension for which disbursements on foregone
income, contractual fines for supply disruptions, fines and other indemnifications
for harm caused to people, environment or organizations and companies are
considered. Additionally there are expenses related to maintenance and operational
actions taken with a view to re-establishing the operational conditions of the
pipeline.
The next step corresponds to eliciting a MAU function U(h, f, m), which it is
considered an additive function. The property of additive independence implies
that there is preferential Independence among the payoff sets. U(h, f, m) can be
expressed by the following (4.7).
I
a Qeff H c (4.8)
4 CDR 2
r xi L
i jk jk , xi (4.9)
184 Chapter 4 Multidimensional Risk Analysis
Knowing that:
L T jk , xi
u P p | T jk , xi (4.10)
In this way, losses associated with each scenario and section are summed in the
three dimensions discussed, multiplied by accident scenario probabilities and
added to the losses associated with a normal scenario (TN), as shown in (4.11).
r xi >
ET L T jk , xi @ ¦¦ LT jk ,x S i T jk 1S i T N
i
(4.11)
j k
Due to the additive independence properties of the MAU function and the
independence in probability of the probability distributions over the consequences,
the risk r(xi) is given by (4.12).
ª º
«k h Ph | T , xi u h dh »
« ³ »
« h »
r xi ¦¦ « f ³ i » i jk i N
« k P f | T , x u f df »S T 1S T (4.12)
j k « f »
« »
« k m Pm | T , xi u m dm»
³
«¬ m »¼
Using the risk values obtained from (4.12), pipeline sections can be ordered in
descending order, thereby obtaining a ranking of pipeline sections that should be
used as input for risk management activities.
The MAUT interval scale allows an incremental comparison between the risk
sections in line with the utility value between the alternatives. Thus, (4.13) and
(4.14) are applied to analyze the relationship between alternatives, showing
respectively the absolute difference between alternatives and the difference ratio
between alternatives. The difference ratio DR is used to interpret the values in
relation to the calculated risks.
DA rb xi rb 1 xi (4.13)
rb xi rb1 xi
DR (4.14)
rb1 xi rb 2 xi
where the index (b) represents the position in the ranking of the section and rb(xi)
represents the risk value related to a specific section. Through the analysis
obtained from the results of these equations, the DM can define which sections
4.3 Risk Decision Models 185
should be included given the resources available, thus representing how much
more a section adds to the risk when compared to another section placed further
down in the ranking provided by the risk model.
Therefore, taking into account all the calculation steps described earlier in this
section, Table 4.1 presents the sections prioritized based on comparisons of the
increments of risk. The values listed in rb(xi) – rb+1(xk) column must be multiplied
by (10-5).
Based on Table 4.1, some interpretations may be made. A descending ranking
of values is applied for risk assessment, where S1 shows the highest value of risk
among the sections evaluated. The highest losses associated with the likely
consequences of accidents are expected for S1. Additionally, it is observed that the
increment in the risk values from S4 to S1 is 1.3098 times greater than that from S7
to S4. In the same way, the increment in the risk values from S9 to S6 is almost 14
times greater than that from S8 to S9.
Typically, energy distribution systems are big and complex. These systems are
considered as being among the main elements of the critical infrastructure. Several
other external systems such as systems of water supply, telecommunications,
traffic, public transport, health, food supply, gas distribution, and others are
dependent on this system. Therefore, several impacts other systems can be caused
by small faults in the power system, and to generate a chain of consequences,
which is why it is a critical part of the infrastructure for society.
Greater initial investment is required by installation of the infrastructure of an
underground system. In generally, it is more complex than overhead systems.
There are some disadvantages by use of the underground systems, such as it incurs
higher costs associated with maintenance; it is also difficult to access underground
networks; to upgrade the system (physical and limited space configuration); and,
to operate and maintain auxiliary ventilation systems, etc.
Though, this system have advantages: the operation of underground systems is
more safer and reliable than overhead systems for the population; more immune
to interference from nature (storms, winds, storms, falling trees, etc.); better
accessibility of disabled people, low visual pollution in the city and presenting less
impact on the occurrence of traffic accidents.
Regardless of being safer than overhead systems, many underground vaults
events have occurred. Hundreds of accidents in vaults occur every year in Ney
York, such as smoke, explosions, fires, etc. (Radeva et al. 2009; Rudin et al. 2010;
Rudin et al. 2011; Rudin et al. 2012).
The low frequency of the occurrence of accident scenarios, its magnitude of
their consequences and the complex environment surrounding the hazard zone
make the risk management becoming even more complex and uncertain (Garcez
and de Almeida 2014a; Garcez and de Almeida 2014b). Also, the large number of
subsystems, with each having particular characteristics, and there is a lack of (or
incomplete) historical data of accidents and its failure modes and past events make
the decision process even more complex.
Hazard scenarios can produce various consequences, for instance, fatalities and
injuries to people, blackout, disruptions to local vehicular traffic, explosions and
fires in nearby locations, impact of the company image, the population being
afraid (on account of the uncertainty of when and where an accident will occurs),
affect the system reliability and safety and other consequences which cannot be in
financial terms (Garcez and de Almeida 2014a; Garcez and de Almeida 2014b).
Hence, these consequences can disturb directly or indirectly the sector of the
society, the public sector and business.
According to Garcez and de Almeida (2014b), assessing the risks comprehensively
and realistically is extremely important. It may generate knowledge that can be
applied to assist a DM to choose and implement preventive and mitigating measures.
4.3 Risk Decision Models 187
§ § § ··
¨ § · · ¸
r §¨Vq ·¸ ¦ ¨ ¦ ¨¨ S T ¨ ³ u c P¨ c T , Vq ¸dc ¸ ¸¸ ¸ 1S T N (4.15)
© ¹ ¨ © ¹ ¸¹ ¸
i ¨© T © © c ¹¹
ri Vq ri 1 Vq (4.16)
As the conclusion, Vq3 is ranked as first underground vault and Vq2 as second.
Furthermore, it is observed that the difference between these risk values
corresponds to approximately 44% of the total range of risk. Therefore, it is
evident that is necessary to allocate more resources as a priority to preventive and
mitigating actions on the first vault.
After the risk of the first alternatives (Vq3 and Vq2) has been attend, there is
another gap between the alternative ranked second Vq2 and the one in third place
Vq5 (14% of the total range of the risk). Again, one prioritizes additional actions to
prevent and mitigate the risk addressed in the first two vaults.
4.3 Risk Decision Models 189
Fig. 4.6 Analysis of the intra-criteria of the risk differences of alternatives Vq and Vq
2 5
This section presents the application (Brito et al. 2010) of a different MCDM/A
method, which is integrated with utility function, the ELECTRE TRI method.
Three main issues should be highlighted, when compared with the two previous
models. First, it is a non-compensatory approach, taking into account a specific
kind of DM’s rationality. Second, the problem consists of a sorting problematic,
since the managerial issues in this application are distinct from the two previous.
Third, it integrates the ELECTRE method with utility theory, in order to
incorporate the DM’s behavior regarding to risk (prone, neutral or averse) into
ELECTRE.
As details given subsequently, this application illustrates the step 6 in the
decision process given in Chap. 2, which involves the identification of DM’s
rationality (compensatory or non-compensatory).
4.3 Risk Decision Models 191
In this context, the profiles b that define the particular risk categories, depend
essentially on the perception that the DM has on different risk levels related to his
system, the availability of resources, the occurrence of previous accidents, society
pressures, as well as being dependent on the number of different strategies, policies,
and measures that the company possesses to deploy among the categories.
The highest risk category contains an alternative with higher probabilities of
occurrence of financial, environmental and human consequences. This category
demands relatively urgent actions that often require changes in some aspects of
the project, and that demand a major financial investment in order to obtain
significant reductions of these risks. Similarly, a lower class of risk presents
sections of pipeline with lower levels of risk, thus allowing a little longer planning
time to find effective solutions and at satisfactory costs (Brito et al. 2010).
Brito et al. (2010) highlight that the manner of determining the reference
profiles b for the each risk categories must be carried out very carefully by the
DM, since the sorting process is fundamentally guided by comparisons with these
profiles.
A procedure to aid DM to infer theses profiles is proposed by Mousseau and
Slowinski (1998). It enables the inference by means of a sample of alternatives
directly sorted by the DM.
The third point highlighted on beginning of this section is the integration
between the ELECTRE TRI method and utility theory, in order to incorporate the
DM’s behavior regarding to risk (prone, neutral, averse). The utility theory
presents an axiomatic approach that can assess the DM’s behavior with regard to
the risk (Keeney and Raiffa 1976) when there are accidents consequences.
Let D be the set of all outcomes in a given accident impact dimension.
Uncertainties are related to the states of nature T, the resulting accidental scenarios
of a pipeline accident, and to its impacts under a given dimension of outcomes.
For dealing with uncertainties on D, it is necessary to use a probabilistic approach,
represented by a probability distribution over the deterministic consequences and
by the elicitation of the utility functions for these consequences (Brito et al. 2010).
This procedure is applied in the intra-criterion assessment process (for each risk
dimension) with the aim of risk evaluation for human, environmental and financial
dimensions posed by each section of pipeline.
As defined in (4.11), the risk is assessed as the expected loss, which is
estimated for each section of pipeline. The loss is given by combining the
probability over the deterministic consequences p in D, named by P(p|T,si), and
the utility function (U(p), where p D ) over these consequences, as shown in
(4.18). It is used the traditional notation for decision analysis (Utility Theory),
where p (from payoff) denotes an element of the set of outcomes D, whereas P
(capital P) refers to a probability (which is a probabilistic payoff).
LT , s i ³ P p | T , s U p dp
i (4.18)
p
4.3 Risk Decision Models 193
Therefore, the expected risk can be calculated for each section of pipeline
under each criterion, applying (4.18) in (4.11), then (4.19) is obtained.
r si ¦
T
S T ³ P p | T , s u p dp
i
p
i (4.19)
P p p
U p e (4.20)
§ P p p ·
r p si ¦ S T ¨ ³ P p | T , s e
i i dp ¸ (1)S i T N (4.21)
© p ¹
T
Parameter rh rm rn
b1 (divides the High Risk from the Medium Risk category) 0.025 0.025 0.05
b2 (divides the Medium Risk from the Low Risk group) 0.013 0.01 0.02
weight 0.60 0.10 0.30
q (indifference threshold) 0.001 0.001 0.005
p (strict preference threshold) 0.005 0.009 0.007
196 Chapter 4 Multidimensional Risk Analysis
By analysis of the DM, the sections in the first category demand higher states
of alert, and thus financial resources would be assigned preferentially to this
category in order to increase measures of physical protection and to intensify the
monitoring of the high risk sections. The Medium Risk category involves pipeline
sections which, although they do not lay claim to such intensive care as those in
the previous class, do demand more thorough planning for preventive measures in
order to avoid neglect in relation to maintaining their safety levels. Finally, as to
the sections assigned to the Low Risk category, the maintenance of routine
inspection actions is planned in order to keep these sections with low risk levels
within the human, environmental and financial dimensions of possible outcomes
(Brito et al. 2010).
The analyst has to explain the meaning of ELECTRE TRI parameters in order
to obtain the proper specification. It was decided not to use a veto threshold for
any risk dimension. With regard to the cutting level, k = 0.65 has been applied.
After applying the sorting model for each individual section of pipeline, the results
in Table 4.5 were obtained.
It was observed that, for this application under study, the results were intensely
influenced, but not completely controlled, by the human risks, given their high
weight value. Among the segments under study, 7 out of the total of 12 sections
were assigned to the Medium Risk category (C2), for which more rigorous
preventive measures should be established within 6 months. Sections s3 and s11
were assigned to the High Risk category (C1), for they present risk levels worse
than or very close to the profile b1 in a more significant proportion of impact
dimensions. Finally, sections s1, s4 and s8 were assigned to the Low Risk category
(C3) because they had more satisfactory performances than those presented by
profile b2.
4.4 Other MCDM/A Applications on Multidimensional Risk 197
The generation of electrical energy can be from various sources. Each energy
source will generate different risks inherent in its own production and supply.
Regős (2012) compared the general risk of the four most important energy chains
(coal, nuclear, gas, hydro). For this, he applied an MCDM/A approach, and chose
severe accidents, terrorism, environmental and health risks, risk of price changes
as risk criteria.
Normally, generation and power supply systems are large and complex systems
which society considers form a critical part of the infrastructure. Typically, several
other systems or subsystems, such as water supply systems, telecommunication,
traffic, health, food supply, etc. are dependent on power supply systems. Thus,
failures in the electricity system can impact other systems and generate a chain of
consequences, which is why it is critical for the infrastructure.
198 Chapter 4 Multidimensional Risk Analysis
According to natural hazard theory, risk appears wherever and whenever assets are
subjected to hazards; it is usually defined as ‘the expected potential loss due to a
particular hazard for a given area and reference period’ and can be mathematically
defined as the combination of hazard and vulnerability (Merad et al. 2004).
For Nefeslioglu et al. (2013), it is fully acceptable for a natural event, such as a
flood or an earthquake to become a natural hazard when people are affected by a
natural hazard. Since the world population is increasing, the need to find habitable
areas has increased considerably, which has led to people having to being caught
up in these natural events more often.
Consequently, Viscusi (2009) states that the occurrence of natural disasters
often generates a cluster of fatalities rather than just a single fatality. Hundreds or
sometimes thousands of people could die from the occurrence of a single event.
Additionally, another important point is that deaths perceived to occur due to
the probability of a natural disaster is a very heterogeneous concept and often, and
its probability is much lower when compared to other risks associated other causes
with fatality. However, the risk management of natural disasters should not just
stick to the issue of people being killed or injured.
4.4 Other MCDM/A Applications on Multidimensional Risk 201
Depending on the type and scale of natural event occurring and its impact on
society, other points may be incorporated in the analysis, such as the issue of
safety, security and public health, population migration, cost estimation, information
sharing, planning public and environmental aspects.
For example, when examining the risk of a flood, a check should be made on
aspects that are part of this context, such as the question of the complexity of the
event, broad spatial scales, intervals of time between events, vulnerability and
social-psychological aspects such as depressions, anxieties and conflicts of
interest, in addition to several conflicting aspects between them.
Additionally, taking into account the study described in Levy (2005) for the
operation and management of reservoirs, there is a complex analysis of the
tradeoffs between protection against flooding (i.e., minimizing the discharge of
reservoirs during the peak periods of flooding) and energy production (meeting
the goal of producing pre-defined levels of energy). On the one hand, flood
protection means that the tank must be maintained at the lowest possible level so
that the reservoir can accommodate the excess water coming from the period of
flooding. On the contrary, the production of energy requires that there be the
largest possible amount of water in the reservoir. In this case, the decision-making
process will directly affect risk management, so what is needed is a more
structured analysis that provides satisfactory results.
According to Nefeslioglu et al. (2013) the evaluation of the interaction between
natural and human events in terms of hazards and risks has become a common
topic for analysis in the last 20 years. Modeling consequences and probabilities is
one of the main tools for assessing the impacts of natural hazards.
Given the uncertainty associated with the environmental context, Parlak et al.
(2012) state that analysis based on multicriteria decision methods provides a
systematic approach to managing the complexities and uncertainties associated
with the occurrence of natural disasters, since multicriteria methods make use of
stochastic approach which help to develop this modeling.
Levy (2005) points out the use of MCDM/A has increased in the last three
decades due to a number of factors, including dissatisfaction with conventional
methods that use only a single criterion, as well as ease of access to software and
algorithms that enable the solution of complex environmental problems to be
found. Thus, in the aforementioned study for the operation and management of
reservoirs, MCDM/A is useful for eliciting and modeling stakeholders’ pre-
ferences and to improve coordination between state agencies, organizations and
the affected population in such a way as to minimize the risks associated with
floods e.g., death of or injury to persons, damage to property and possible environ-
mental impacts.
The need for multicriteria approaches can also be observed in planning the
response to a disaster, where, according to Parlak et al. (2012) such planning
requires the engagement of multiple disciplines such as engineering (infrastructure),
management emergencies, health care, mass communication, water supply and
food logistics. Planning the integration scenario by using multicriteria analysis,
202 Chapter 4 Multidimensional Risk Analysis
landslides at the national level in Cuba, and this information being linked to the
national system, which gives early warning of hurricanes, and warns and
evacuates people from areas prone landslides.
Another context of natural hazard together with man’s intervention arises from
ending mining operations in populated regions. According to Merad et al. (2004),
in the Lorraine region of France many landslides and subsidence have occurred,
which led to the need to develop a specific methodology for risk zoning of the
area. The authors propose a methodology based on a multicriteria decision support
tool (ELECTRE-TRI), with the aim of assigning risk zones in predefined classes
of inhabited regions. This approach enabled the knowledge of experts, multiple
qualitative and quantitative criteria and uncertainties to be considered.
In recent decades, the fight against terrorism has been the focus of constant
analysis worldwide. Security measures have been strengthened and new anti-
terrorism policies are presented to the world by nations periodically to society.
One goal of these policies is to establish the benefits of preventing a terrorist
attack, such as reducing the number of deaths and injuries associated with the
human dimension. More specifically, the risk management of terrorist attacks has
intensified, especially after the terrorist attack on the World Trade Center on
September 11, 2001 in the United States.
Risk management, according to Aven and Renn (2009) seeks to ensure that
adequate measures are established to protect people, the environment and assets
from harmful consequences arising from human activities or natural events. The
extent to which risk reduction measures are justified depends on the balance
between costs and benefits in terms of the security gain. Furthermore, several PRA
models have been applied taking into account aspects such as infrastructure, food
supply chains, population, etc., and considering risk as a product of three
components: threat, vulnerability and consequence (Greenberg et al 2012).
Due to terrorism threats, several models and approaches have been proposed in
order to mitigate the risks associated with such events (Merrick and Leclerc 2014;
Shan and Zhuang 2014; Haphuriwat and Bier 2011; Ezell et al. 2010; Parnell et al.
2010; Ngange et al 2008; Leung et al 2004). Among these models, there are
several studies considering MCDM/A approaches (Akgun et al. 2010; Sri
Bhashyam and Montibeller 2012; Koonce et al. 2008; Patterson and Apostolakis
2007).
In their paper, Akgun et al. (2010) stress that assessing the vulnerability of
critical assets (e.g.: airports, dams, chemical plants, nuclear power plants) to
terrorist attacks is a highly complex strategic activity, requiring a methodology
structured to support the decision-making process in defense planning. Their
approach seeks to define the vulnerability of each critical defense asset against
204 Chapter 4 Multidimensional Risk Analysis
terrorist attacks taking into account multiple criteria. They use SMART in
conjunction with Fuzzy Set Theory and Fuzzy Cognitive Maps in a group decision
environment. Their model seeks to identify hidden vulnerabilities and to define the
roles and most critical (or active) components of each system, and five criteria
were established:
x Deterrence (implemented method of defense, perceived by terrorists as hard to
penetrate);
x Detection (of a terrorist attack);
x Delay (the time during which an element of a physical protection system is
designed to prevent terrorist invasions);
x Response (the time taken to respond to a threat), and;
x Recovery (the time taken to return the areas and people affected to their
existing status prior to the event).
In the second study, Sri Bhashyam and Montibeller (2012) propose a
framework that can be used to infer how the priorities of the terrorists may change
over time and the impact that these changes may have on the choice of a harmful
action. This is done based on a multicriteria model that uses MAUT. The
objectives were visualized in three categories: revenge, reputation and reaction.
The alternatives of the decision problem were established by: strikes by the
terrorists, improvised explosive devices in a public place, explosions of portable
nuclear devices in modes of mass transit, detonating bombs and biological
weapons or dirty bombs that combine explosives and radioactive materials.
Therefore, the aim of modeling terrorists’ priorities is to define the objectives
that terrorists will use to evaluate the attack, providing the best tradeoff between
the operational side of an attack (costs) and benefits (if goals are achieved).
In this context, Papamichail and French (2012) points out that radioactive
accidents have emphasized the requirement to provide support for all emergency
management phases. Several decision support tools are currently being developed
to prevent and mitigate the effects of radioactive accidents. Among these tools,
multicriteria decision techniques stand out.
The literature describes some recent applications of multicriteria decision
methods in the context of nuclear energy. Examples include the following:
x Atmaca and Basar (2012) use an Analytic Network Process (ANP) to evaluate
6 different alternatives of nuclear power plants taking into account criteria such
as technological aspects and sustainability, economic viability, quality of life
and socio-economic impacts.
x Hong et al. (2013) use a multi-criteria decision analysis to assess future
scenarios for generating electricity in Japan electricity which take economic,
environmental and social impacts into consideration. Their study is a response
to the nuclear crisis caused by the Fukushima accident.
x Erol et al. (2014) define the location problem of a nuclear power plant in
Turkey as a multicriteria decision problem using fuzzy logic, and consider
qualitative and quantitative criteria. The primary criteria that they establish are:
proximity to the existing electrical infrastructure; proximity to the transportation
infrastructure; access to large amounts of cooling water. The authors also
consider a number of secondary criteria: population density; geological issues;
atmospheric conditions; cost factors; and risk factors.
x Beaudouin (2015) proposes an MCDM/A model that supports debate about
nuclear power plants safety choices. Therefore, six safety criteria are considered in
combination with cost-effectiveness analyses to point out the best portfolio of
power plant design modifications, satisfying security requirements.
Thus, MCDM/A tools can be used at various stages in the context of nuclear
power production in order to contribute to risk management, thereby making the
decision-making process an important aspect when planning safety measures for
nuclear power plants.
A requirement for the building industry, both with regard to permission to build
and to certification that the finished building meets regulatory requirements, is
necessary an environmental management, where the identification and evaluation
of risk to human and environmental health are the first stages. Topuz et al. (2011)
propose an approach that integrates the assessment of risk to humans with environ-
mental health in industries using hazardous materials, to support environmental
DMs with quantitative and directive results. For this, the methodologies used
206 Chapter 4 Multidimensional Risk Analysis
multicriteria and fuzzy logic to deal with the problems arising from the complexity
of the environment and uncertain data.
Specifically, in the context of bridges, the use of bridges is among the most
important structural elements that reduce traffic problems. Risk management of
bridges serves to determine the best allocation of resources. According to Adey et
al. (2003), these systems are usually evaluated against the structural deterioration
which bridges may suffer from as a result of traffic loads. However, these systems
are affected by various other hazards, such as floods and earthquakes, not only the
traffic load.
The destruction of big bridges are usually important and significant events, and
may result in loss of lives, property and economic losses. According to Shetty et
al. (1997), the consequences of the destruction of bridges can be summarized as:
Human elements that impact the number of deaths and injuries, such as the
high rate of vehicular traffic, the flow of pedestrians that pass over or under the
bridge;
Environmental consequences resulting from spills of hazardous substances, due
to the intersection of transport between road, rail, etc.;
Formation of traffic jams, increasing the volume of traffic at a particular site,
causes overload on other transport routes;
Economic factors, including the cost of taking construction material residuals
away; reconstruction; indemnities payable on the destruction of vehicles; the
environmental catharsis; and legal costs.
According to Wang et al. (2008), the risk assessment of bridges is essentially a
multicriteria problem, which involves multiple assessment criteria such as safety
(safety of the public), functionality (effects on the level of service/availability of
the network for use), sustainability (expenditure and workload) and environment
(effects on the environment, including the (aesthetic) appearance of the
structures). Wang et al. (2008) propose an integrated AHP–DEA methodology to
evaluate risks to or from bridges of hundreds or thousands of bridge structures,
based on which maintenance priorities for the bridge structures can be drawn up.
Environmental risk assessment and decision-making strategies in recent
decades have become increasingly sophisticated, and use intensive and complex
information, including approaches such as expert opinions, cost-benefit analyzes
and evaluation of the toxicological risk. According to Linkov et al. (2006), a tool
that has been used to support environmental decision-making is comparative risk
assessment (CRA), but CRA lacks a structured process to arrive at an alternative
optimal design method. The approach of using multicriteria decision analysis fills
this need by providing methods that give better support to comparing alternatives
and also provides a structure which incorporates input from stakeholders of the
project, the aim of which is to rank alternatives.
In the context of the hazard from forest fires, over past decades, in several
regions, especially in tropical and Mediterranean regions, these fires are due to
several underlying factors, which have received increasing attention because of
4.4 Other MCDM/A Applications on Multidimensional Risk 207
the wide range of ecological, economic, social and political impacts. The more
complex fire models require spatial information, which is done by remote sensing
and GIS (Vadrevu et al. 2010; Arianoutsou et al. 2011).
According to Vadrevu et al. (2010), the integration of MCDM/A methods in
the spatial domain provides a new framework for addressing many environmental
problems, including quantifying “fire hazards”. These authors conducted a study
in a thickly-forested area (Indian region), where most of the stakeholders are the
local people, and their dependence on forest resources is immense.
Moreover, the problem of forest fires in the study area is spatially diverse in
nature and involves both biophysical and socioeconomic parameters, providing an
ideal place to use an MCDM/A methodology. Combining these multiple para-
meters using decision-making methods in a collaborative framework may yield
good results, so, the risk of fires in tropical deciduous forests, in India, was
quantified as a function of topographic, vegetation, climatic, and socioeconomic
attributes in order to evaluate the fire risk in the study area.
Still in the environmental context, the contamination of water resources on land
has been a major environmental concern during the last decades, mainly due to
public health concerns. According to Khadam and Kaluarachchi (2003), traditionally,
environmental decision-making scenarios of subsurface contamination are guided
by means of cost-benefit analysis.
This context, the risk assessment includes quantification of the risk to human
health, as well as evaluating the importance of this risk. When the risk is deter-
mined unacceptable, potential remedial alternatives are identified and decision
analysis is performed to choose the best corrective action. There is a tradeoff
between individual risk and societal risk, the tradeoff between the residual risk
and the cost of reducing this risk, and cost-effectiveness as a justification for
remediation. The authors propose an integrated approach for the management of
contaminated ground water using a multicriteria decision framework to assess the
risk to health and to make an economic analysis.
Another current context to be analyzed is in the newish field of nanotechnology
which is increasingly being embedded in innovations that can benefit humanity
(Siegrist et al. 2007). However, there is a variety of factors involved in managing
the development of nanomaterial, ranging from the technical specifications of the
material to possible adverse effects in humans. Therefore, it is important to assess
the benefits and risks inherent in issues of Environmental Health and Safety (EHS)
related to nanotechnology. According to Linkov et al. (2007), there is currently no
structured approach for making justifiable and transparent decisions with explicit
trade-offs among the many factors.
Linkov et al. (2007) conceptualize the use of the MCDM/A as a powerful
analytical framework and scientifically sound decision tool for assessing and
managing risk when using nanomaterial. They seek a balance between social
benefits and unintended side effects and risks. They also investigate how to gather
multiple lines of evidence to estimate the likely toxicity and risks of nanomaterial,
given limited information on its physical and chemical properties. An essential
208 Chapter 4 Multidimensional Risk Analysis
References
Abella EAC, Van Westen CJ (2007) Generation of a landslide risk index map for Cuba using
spatial multi-criteria evaluation. Landslides 4:311–325
Adey B, Hajdin R, Brühwiler E (2003) Risk-based approach to the determination of optimal
interventions for bridges affected by multiple hazards. Eng Struct 25:903–912
Akgun I, Kandakoglu A, Ozok AF (2010) Fuzzy integrated vulnerability assessment model for
critical facilities in combating the terrorism. Expert Syst Appl 37:3561–3573
Alencar MH, Cavalcante CAV, de Almeida AT, Silva Neto CE (2010) Priorities assignment for
actions in a transport system based on a multicriteria decision model. In: Bris R, Soares CG,
Martorell S (eds) European safety and reliability conference, Prague, September 2009.
Reliability, Risk, and Safety: Theory and Applications, Vol. 1-3. 2009. Taylor and Francis,
London, UK, p 2480
Alencar MH, de Almeida AT (2010) Assigning priorities to actions in a pipeline transporting
hydrogen based on a multicriteria decision model. Int J Hydrogen Energy 35(8):3610–3619
Alencar MH, Krym EM, Marsaro MF, de Almeida AT (2014) Multidimensional risk evaluation
in natural gas pipelines: Environmental aspects observed through a multicriteria decision
model. In: Steenbergen RDJM, VanGelder PHAJM, Miraglia S, Vrouwenvelder ACWMT
(eds) 22nd Annual Conference on European Safety and Reliability (ESREL), Amsterdam,
2013. Safety, Reliability and Risk Analysis: Beyond the Horizon. Taylor & Francis Group,
London, UK, p 758
Almeida-Filho AT de, de Almeida AT (2010a) Cost-effectiveness analysis and multicriteria
approaches: two irreplaceable paradigms for different problems in risk and safety problems.
In: Proceedings of the European Safety and Reliability Annual Conference, Rhodes, 2010.
Reliability, Risk and Safety: Back to the Future, p 2293
Almeida-Filho AT de, de Almeida AT (2010b) Multiple dimension risk evaluation framework.
In: Bris R, Soares CG, Martorell S (eds) European Safety and Reliability Conference
(ESREL), Prague, Czech Republic, 2009. Reliability, Risk and Safety: Theory and
Applications. CRC Press-Taylor & Francis Group, p 1049
Andersson G, Donalek P, Farmer R, et al. (2005) Causes of the 2003 major grid blackouts in
North America and Europe, and recommended means to improve system dynamic
performance. Power Syst IEEE Trans 20:1922–1928
References 209
Apostolakis GE, Lemon DM (2005) A Screening Methodology for the Identification and
Ranking of Infrastructure Vulnerabilities Due to Terrorism. Risk Anal 25:361–376
Arianoutsou M, Koukoulas S, Kazanis D (2011) Evaluating Post-Fire Forest Resilience Using
GIS and Multi-Criteria Analysis: An Example from Cape Sounion National Park, Greece.
Environ Manage 47:384–397
Arnaldos J, Casal J, Montiel H, et al. (1998) Design of a computer tool for the evaluation of the
consequences of accidental natural gas releases in distribution pipes. J Loss Prev Process Ind
11:135–148
Atmaca E, Basar HB (2012) Evaluation of power plants in Turkey using Analytic Network
Process (ANP). Energy 44:555–563
Aven T, Renn O (2009) The Role of Quantitative Risk Assessments for Characterizing Risk and
Uncertainty and Delineating Appropriate Risk Management Options, with Special Emphasis
on Terrorism Risk. Risk Anal 29:587–600
Beaudouin F (2015) Implementing a Multiple Criteria Model to Debate About Nuclear Power
Plants Safety Choices. Gr Decis Negot 1–29
Beaudouin F, Munier B (2009) A revision of industrial risk management: Decisions and
experimental tools in risk business. Risk Decis Anal 1:3–20
Bedford T, Cooke R (2001) Probabilistic Risk Analysis: Foundations and Methods. Cambridge
University Press, New York
Berger JO (1985) Statistical decision theory and Bayesian analysis. Springer Science & Business
Media, New York
Berizzi A (2004) The Italian 2003 blackout. Power Eng Soc Gen Meet 2004 IEEE 1673–1679
Vol. 2
Brito AJ, de Almeida AT (2009) Multi-attribute risk assessment for risk ranking of natural gas
pipelines. Reliab Eng Syst Saf 94(2):187–198
Brito AJ, de Almeida AT, Mota CMM (2010) A multicriteria model for risk sorting of natural
gas pipelines based on ELECTRE TRI integrating Utility Theory. Eur J Oper Res 200:812–
821
Cailloux O, Mayag B, Meyer P, Mousseau V (2013) Operational tools to build a multicriteria
territorial risk scale with multiple stakeholders. Reliab Eng Syst Saf 120:88–97
Catrinu MD, Nordgård DE (2011) Integrating risk analysis and multi-criteria decision support
under uncertainty in electricity distribution system asset management. Reliab Eng Syst Saf
96:663–670
Comes T, Wijngaards N, Hiete M, et al (2011) A Distributed Scenario-Based Decision Support
System for Robust Decision-Making in Complex Situations. Int J Inf Syst Cris Response
Manag 3:17–35
Cox LA Jr (2009) Risk analysis of complex and uncertain systems. Springer Science & Busi-
ness Media, New York
Cox LA Jr (2012) Evaluating and Improving Risk Formulas for Allocating Limited Budgets to
Expensive Risk-Reduction Opportunities. Risk Anal 32(7):1244–1252
Crowl DA, Jo Y-D (2007) The hazards and risks of hydrogen. J Loss Prev Process Ind 20:158–
164
de Almeida AT (2005) Multicriteria Modelling of Repair Contract Based on Utility and
ELECTRE I Method with Dependability and Service Quality Criteria. Ann Oper Res
138:113–126
de Almeida AT (2007) Multicriteria decision model for outsourcing contracts selection based on
utility function and ELECTRE method. Comput Oper Res 34(12):3569–3574
de Almeida AT, Ferreira RJP, Cavalcante CAV (2015) A review of multicriteria and multi-
objective models in maintenance and reliability problems. IMA Journal of Management
Mathematics 26(3):249–271
Dong M, Lou C, Wong C (2008) Adaptive Under-Frequency Load Shedding. Tsinghua Sci
Technol 13:823–828
210 Chapter 4 Multidimensional Risk Analysis
DziubiĔski M, Frątczak M, Markowski AS (2006) Aspects of risk analysis associated with major
failures of fuel pipelines. J Loss Prev Process Ind 19:399–408
Erol ø, Sencer S, Özmen A, Searcy C (2014) Fuzzy MCDM framework for locating a nuclear
power plant in Turkey. Energy Policy 67:186–197
Ezell BC, Bennett SP, Von Winterfeldt D, et al. (2010) Probabilistic Risk Analysis and
Terrorism Risk. Risk Anal 30(4):575–589
Faber MH, Stewart MG (2003) Risk assessment for civil engineering facilities: critical overview
and discussion. Reliab Eng Syst Saf 80:173–184
Figueira J, Greco S, Ehrgott M (eds) (2005) Multiple Criteria Decision Analysis: State of the Art
Surveys. Springer Verlag, Boston, Dordrecht, London
Garcez TV, de Almeida AT (2012) Multiple Dimension Manhole Explosion in an Underground
Electrical Distribution System. In: Proceedings of the 11th International Probabilistic Safety
Assessment and Management Conference and the Annual European Safety and Reliability
Conference 2012. Curran Associates, Inc. Helsinki, Finland, p 4893–4899
Garcez TV, de Almeida AT (2014a) A risk measurement tool for an underground electricity
distribution system considering the consequences and uncertainties of manhole events. Reliab
Eng Syst Saf 124:68–80
Garcez TV, de Almeida AT (2014b) Multidimensional Risk Assessment of Manhole Events as a
Decision Tool for Ranking the Vaults of an Underground Electricity Distribution System.
Power Deliv IEEE Trans 29:624–632
Garcez TV, de Almeida AT (2014c) Multidimensional risk assessment of underground electricity
distribution systems based on MAUT. In: Steenbergen RDJM, VanGelder PHAJM, Miraglia
S, Vrouwenvelder ACWMT (eds) 22nd Annual Conference on European Safety and
Reliability (ESREL), Amsterdam, 2013. Safety, Reliability and Risk Analysis: Beyond the
Horizon. CRC Press-Taylor & Francis Group, p 2009
Garcez TV, de Almeida-Filho AT, de Almeida AT, Alencar MH (2010) Multicriteria risk
analysis application in a distribution gas pipeline system in Sergipe. In: Bris R, Soares CG,
Martorell S (eds) Reliability, risk and safety: theory and applications vols 1-3. European
safety and reliability conference (ESREL 2009), Prague, September 2009. Taylor and
Francis, 1043-1047
Geldermann J, Bertsch V, Treitz M, et al. (2009) Multi-criteria decision support and evaluation
of strategies for nuclear remediation management. Omega 37:238–251
Greenberg M, Haas C, Cox LA Jr, et al. (2012) Ten Most Important Accomplishments in Risk
Analysis, 1980–2010. Risk Anal 32(5):771–781
Guedes Soares CG, Teixeira AP (2001) Risk assessment in maritime transportation. Reliab Eng
Syst Saf 74(3):299–309
Haidar AMA, Mohamed A, Hussain A (2010) Vulnerability control of large scale interconnected
power system using neuro-fuzzy load shedding approach. Expert Syst Appl 37:3171–3176
Haphuriwat N, Bier VM (2011) Trade-offs between target hardening and overarching protection.
Eur J Oper Res 213(1):320–328
Henselwood F, Phillips G (2006) A matrix-based risk assessment approach for addressing linear
hazards such as pipelines. J Loss Prev Process Ind 19:433–441
Hobbs BF, Meier P (2000) Energy Decisions and the Environment. A guide to the use of
multicriteria methods (International Series in Operations Research & Management Science).
Kluwer Academic Publisher, Norwell
Hong S, Bradshaw CJA, Brook BW (2013) Evaluating options for the future energy mix of
Japan after the Fukushima nuclear crisis. Energy Policy 56:418–424
IEEE1584 (2002) IEEE Guide for Performing Arc-Flash Hazard Calculations. IEEE Std 1584-
2002
Jo Y-D, Ahn BJ (2002) Analysis of hazard areas associated with high-pressure natural-gas
pipelines. 2 J Loss Prev Process Ind 15:179–188
Jo Y-D, Ahn BJ (2005) A method of quantitative risk assessment for transmission pipeline
carrying natural gas. J Hazard Mater 123:1–12
References 211
Jo Y-D, Crowl DA (2008) Individual risk analysis of high-pressure natural gas pipelines. J Loss
Prev Process Ind 21:589–595
Jozi A, Pouriyeh A (2011) Health-safety and environmental risk assessment of power plants
using multi criteria decision making method. Chem Ind Chem Eng Q 17:437–449
Karvetski CW, Lambert JH, Keisler JM, Linkov I (2011) Integration of Decision Analysis and
Scenario Planning for Coastal Engineering and Climate Change. Syst Man Cybern Part A
Syst Humans, IEEE Trans 41(1):63–73
Keeney RL, Raiffa H (1976) Decisions with multiple objectives: Preferences and Value Trade-
Offs. Wiley Series in Probability and Mathematical Statistics. Wiley and Sons, New York
Khadam IM, Kaluarachchi JJ (2003) Multi-criteria decision analysis with probabilistic risk
assessment for the management of contaminated ground water. Environ Impact Assess Rev
23:683–721
Koonce AM, Apostolakis GE, Cook BK (2008) Bulk power risk analysis: Ranking infrastructure
elements according to their risk significance. Int J Electr Power Energy Syst 30:169–183
Leung M, Lambert JH, Mosenthal A (2004) A Risk-Based Approach to Setting Priorities in
Protecting Bridges Against Terrorist Attacks. Risk Anal 24(4):963–984
Levy J (2005) Multiple criteria decision making and decision support systems for flood risk
management. Stoch Environ Res Risk Assess 19:438–447
Linares P (2002) Multiple criteria decision making and risk analysis as risk management tools
for power systems planning. Power Syst IEEE Trans 17:895–900
Linkov I, Satterstrom FK, Kiker G, et al. (2006) From comparative risk assessment to multi-
criteria decision analysis and adaptive management: Recent developments and applications.
Environ Int 32:1072–1093
Linkov I, Satterstrom FK, Steevens J, et al. (2007) Multi-criteria decision analysis and
environmental risk assessment for nanomaterials. J Nanoparticle Res 9:543–554
Lins PHC, de Almeida AT (2012) Multidimensional risk analysis of hydrogen pipelines. Int J
Hydrogen Energy 37:13545–13554
Lopes YG, de Almeida AT, Alencar MH, Wolmer Filho LAF, Siqueira GBA (2010) A Decision
Support System to Evaluate Gas Pipeline Risk in Multiple Dimensions. In: Bris R, Soares
CG, Martorell S (eds) European Safety and Reliability Conference (ESREL), Prague, Czech
Republic, 2009. Reliability, Risk and Safety: Theory and Applications. Crc Press-Taylor &
Francis Group, p 1043
Luria P, Aspinall PA (2003) Evaluating a multi-criteria model for hazard assessment in urban
design. The Porto Marghera case study. Environ Impact Assess Rev 23:625–653
Merad MM, Verdel T, Roy B, Kouniali S (2004) Use of multi-criteria decision-aids for risk
zoning and management of large area subjected to mining-induced hazards. Tunn Undergr Sp
Technol 19:125–138
Merrick JRW, Leclerc P (2014) Modeling Adversaries in Counterterrorism Decisions Using
Prospect Theory. Risk Anal. doi: 10.1111/risa.12254
Montiel LV, Bickel JE (2014) A Generalized Sampling Approach for Multilinear Utility
Functions Given Partial Preference Information. Decis Anal 11(3):147–170
Morgan MG, Florig HK, DeKay ML, Fischbeck P (2000) Categorizing Risks for Risk Ranking.
Risk Anal 20:49–58
Mousseau V, Slowinski R (1998) Inferring an ELECTRE TRI Model from Assignment
Examples. J Glob Optim 12:157–174
Nefeslioglu HA, Sezer EA, Gokceoglu C, Ayas Z (2013) A modified analytical hierarchy
process (M-AHP) approach for decision support systems in natural hazard assessments.
Comput Geosci 59:1–8
Nganje W, Bier V, Han H, Zack L (2008) Terrorist Threats to Food: Guidance for Establishing
and Strengthening Prevention and Response Systems. Am J Agric Econ 90(5):1265–1271
Papamichail KN, French S (2013) 25 Years of MCDA in nuclear emergency management. IMA
J. Manag. Math. pp 481–503
212 Chapter 4 Multidimensional Risk Analysis
Parlak AI, Lambert JH, Guterbock TM, Clements JL (2012) Population behavioral scenarios
influencing radiological disaster preparedness and planning. Accid Anal Prev 48:353–62
Parnell GS, Smith CM, Moxley FI (2010) Intelligent Adversary Risk Analysis: A Bioterrorism
Risk Management Model. Risk Anal 30:32–48
Patterson SA, Apostolakis GE (2007) Identification of critical locations across multiple
infrastructures for terrorist actions. Reliab Eng Syst Saf 92:1183–1203
Radeva A, Rudin C, Passonneau R, Isaac D (2009) Report Cards for Manholes: Eliciting Expert
Feedback for a Learning Task. Mach Learn Appl 2009 ICMLA ’09 Int Conf 719–724
Raiffa H (1968) Decision analysis: introductory lectures on choices under uncertainty. Addison-
Wesley, London
RegĘs G (2013) Comparison of power plants’ risks with multi criteria decision models. Cent Eur
J Oper Res 21:845–865
Rezaian S, Jozi SA (2012) Health- Safety and Environmental Risk Assessment of Refineries
Using of Multi Criteria Decision Making Method. APCBEE Procedia 3:235–238
Rogner H-H (2013) World outlook for nuclear power. Energy Strateg Rev 1:291–295
Roy B (1996) Multicriteria Methodology for Decision Aiding. Springer US
Rudin C, Passonneau R, Radeva A, et al. (2010) A process for predicting manhole events in
Manhattan. Mach Learn 80:1–31
Rudin C, Passonneau RJ, Radeva A, et al. (2011) 21st-century data miners meet 19th-century
electrical cables. Computer (Long Beach Calif) 44:103–105
Rudin C, Waltz D, Anderson RN, et al. (2012) Machine Learning for the New York City Power
Grid. Pattern Anal Mach Intell IEEE Trans 34:328–345
Salvi O, Merad M, Rodrigues N (2005) Toward an integrative approach of the industrial risk
management process in France. J Loss Prev Process Ind 18:414–422
Scawthorn C (2008) A Brief History of Seismic Risk Assessment. In: Bostrom A, French S,
Gottlieb S (eds) Risk Assessment, Model. Decis. Support. Springer Berlin Heidelberg, Berlin,
Heidelberg, pp 5–81
Shan X, Zhuang J (2014) Subsidizing to disrupt a terrorism supply chain-a four-player game.
J Oper Res Soc 65(7):1108–1119
Shetty NK, Chubb MS, Halden D (1997) An overall risk-based assessment procedure for
substandard bridges. In: Das PC (ed) Saf. Bridg. Telford, London, Uk, pp 225–235
Siegrist M, Keller C, Kastenholz H, et al. (2007) Laypeople’s and Experts’ Perception of Nano-
technology Hazards. Risk Anal 27:59–69
Sklavounos S, Rigas F (2006) Estimation of safety distances in the vicinity of fuel gas pipelines.
J Loss Prev Process Ind 19:24–31
Sri Bhashyam S, Montibeller G (2012) Modeling State-Dependent Priorities of Malicious
Agents. Decis Anal 9:172–185
Stefanidis S, Stathis D (2013) Assessment of flood hazard based on natural and anthropogenic
factors using analytic hierarchy process (AHP). Nat Hazards 68:569–585
Tamura H, Yamamoto K, Tomiyama S, Hatono I (2000) Modeling and analysis of decision
making problem for mitigating natural disaster risks. Eur J Oper Res 122:461–468
Topuz E, Talinli I, Aydin E (2011) Integration of environmental and human health risk
assessment for industries using hazardous materials: A quantitative multi criteria approach
for environmental decision makers. Environ Int 37:393–403
Tweeddale M (2003) Managing Risk and Reliability of Process Plants. Gulf Professional
Publishing. Burlington
Vadrevu K, Eaturu A, Badarinath KVS (2010) Fire risk evaluation using multicriteria analysis—
a case study. Environ Monit Assess 166:223–239
Van der Vleuten E, Lagendijk V (2010) Transnational infrastructure vulnerability: The historical
shaping of the 2006 European “Blackout.” Energy Policy 38:2042–2052
Vincke P (1992) Multicriteria Decision-Aid. John Wiley & Sons, New York
Viscusi WK (2009) Valuing risks of death from terrorism and natural disasters. J Risk Uncertain
38:191–213
References 213
Walsh BP, Black WZ (2005) Thermodynamic and mechanical analysis of short circuit events in
an underground vault. Power Deliv IEEE Trans 20:2235–2240
Wang J (2006) Maritime Risk Assessment and its Current Status. Qual Reliab Eng Int 22(1):3–19
Wang Y-M, Liu J, Elhag TMS (2008) An integrated AHP–DEA methodology for bridge risk
assessment. Comput Ind Eng 54:513–525
Willis HH, DeKay ML, Fischhoff B, Morgan MG (2005) Aggregate, Disaggregate, and Hybrid
Analyses of Ecological Risk Perceptions. Risk Anal 25:405–428
Yoe C (2012) Principles of Risk Analysis – Decision Making under uncertainty. CRC Press,
Boca Raton
Chapter 5
Preventive Maintenance Decisions
5.1 Introduction
In the face of growing competition, leading to an ever increasing need for higher
productivity, there is a need for methods, tools and technologies that enable the
producing systems to acquire competitive advantages. Preventive maintenance
decisions are quite relevant to the strategic results of any business organization, in
which a producing systems has to make products, may them be goods or services.
The type of product makes a great difference in the way that maintenance in
general (and preventive maintenance in particular) is linked to business results.
For instance, a service producing system has a feature of simultaneousness (Slack
et al. 2010), which means that at the time the system is producing the product, the
customer is being served. In such a context, when a failure in the system occurs,
the maintenance has an immediate impact on the business competitiveness (de
Almeida and Souza 2001). Therefore, preventive maintenance planning becomes a
more strategic decision that is linked to a higher level of the hierarchical organizational
structure. For given decision context, the consequences are characterized by
multiple and less tangible objectives, which may require an MCDM/A support.
These issues are discussed in Chap. 1, presenting the peculiarities of two different
types of systems, service and goods producing systems.
This chapter addresses preventive maintenance planning by the selection of the
preventive maintenance time interval. This kind of decision is applied to a
component or item (or device) and it is not applicable to a system, unless this
system is replaced as whole, using the failure behavior of the system.
The next section presents a classical optimization approach, followed by a
general MCDM/A preventive maintenance model. The last two sub-sections deal
with two different MCDM/A approaches to support the preventive maintenance
time interval.
It is important to note that there are many different mathematical models
related to preventive maintenance (Shafiee and Finkelstein 2015), although similar
process to build MCDM/A models can be applied considering the required
adaptations.
One of the most important problems in the maintenance area is the definition of
the frequency at which preventive maintenance actions should be performed.
In both producing systems, this decision has a great impact.
In a literature review on MCDM/A models in maintenance, around 22 % of
research found is related to preventive maintenance (de Almeida et al. 2015).
Multiobjective optimization in preventive maintenance has been considered since
the late 1970. Inagaki et al. (1978) considered three objectives: mission reliability,
total cost, system weight in a multiobjective nonlinear mixed-integer problem and
proposed a procedure based on interactive optimization and a nonlinear pro-
gramming algorithm, called ICOM (Interactive Coordinatewise Optimization
Method). Hwang et al. (1979) formulated a scheduled-maintenance policy problem
and set three objectives: minimum replacement cost-rate, maximum availability,
and lower-bound on mission reliability. Four multicriteria methods were analyzed:
strictest-selection; lexicographic; Waltz lexicographic, and the sequential multiple-
objective problem-solving technique (SEMOPS). Jiang and Ji (2002) consider
four attributes: cost, availability, reliability, and lifetime, via a multiple attribute
value theory (MAVT).
Before the presentation of a general MCDM/A preventive maintenance model,
a classical optimization approach is presented in the next subsection.
5.2 A General MCDM/A Model for Preventive Maintenance 217
Glasser (1969) presents age replacement and block replacement as two methods of
planning replacement in a program of preventive maintenance. Although these
methods have been previously described by other authors (Barlow and Hunter
1960; Cox 1962), the main contribution of Glasser (1969) is the focus on the
managerial impacts of these methods. These issues have been presented in many
other texts in the literature (Scarf et al. 2005).
According to Glasser (1969), the main problem of preventive maintenance is
associated with the uncertainty about the exact time at which an item will fail.
This uncertainty establishes a difficulty in guarantee the effectiveness of the
replacement, which in some times could happen earlier than failure, in others only
after a failure takes place.
Glasser (1969) structures a two-phase process to model the problem of
replacement planning. The phases consist of: 1) the description of the pattern of
failures of the item over time, in terms of a probability density function f(t); and 2)
the development of an equation that describes the expected cost per time of
following a particular policy of planned replacement.
A general description of this model is given by (5.1).
E ( c(t ))
cr (t ) (5.1)
E ( v (t ))
where:
cr(t) is the cost rate;
E(c(t)) is the expected cost;
E(v(t)) is the expected cycle length.
The final expression of cr(t) depends on the assumptions of the model, which
may be related to the influence of the action on the system.
A great number of papers deal with these different aspects, almost all following
the general structure of Glasser (1969). Although these models have a great potential
to support the maintenance manager, they may not be sufficient to describe the
consequence space of failures. So, the rate cost as a criterion should be considered
together with other criteria. This is described in a general framework that could be
used to address the problem via the MCDM/A approach, given in the next sub-
section.
The assumptions related to the simplest case of the replacement model (Cox
1962) are valid for the MCDM/A models in the subsequent sub-sections. The
assumptions of the simplest replacement age-based model are (Cox 1962):
1. The state of the item is known;
2. The alternatives set is defined as opportunistic intervals, which can be days,
weeks, months, or other period;
218 Chapter 5 Preventive Maintenance Decisions
3. The failure probability density function f(t) of the item is IFR (an increasing
failure rate);
4. The system can only be in one of two states, failed or operational;
5. Replacement prior a failure is worthwhile, there are saving in avoiding a failure
by doing preventive replacements;
6. The item replacement restores the system to the as good as new state;
7. The equipment failure times can be modeled by a known probability density
function f(t);
8. The time necessary to perform a replacement is negligible compared to the time
between failures, so it is not considered into a cycle.
From these assumptions (5.1) becomes (5.2).
ca (1 R(t )) cb R(t )
cr (t ) t
(5.2)
³ xf ( x)dx t( R(t ))
0
where
cr(t) is the cost rate;
R(t) is the reliability function;
f(t) is the density probability function;
ca is the cost replacement after a failure;
cb is the cost replacement before a failure.
As already stated, most of the models restrict the analysis to only the cost rate
criterion cr(t). Therefore, it is important to understand the behavior of this aspect.
To illustrate the different behaviors of the cost rate, a Weibull function is
assumed to f(t). Also, different values for its parameters may be applied to give an
idea of how the cost rate may change with different failure data patterns.
Assuming a Weibull distribution in (5.2), (5.3) can be obtained.
E E
ª § t ·º ª § t ·º
« ¨¨ ¸¸ » « ¨¨ ¸¸ »
K K
ca (1 e «¬ © ¹ »¼ ) cb e «¬ © ¹ »¼
cr (t ) E E
(5.3)
t E 1 ª« §¨ x ·¸ º» ª § t ·º
« ¨¨ ¸¸ »
E ªxº ¨ ¸
¬« © K ¹ ¼» ¬« © K ¹ ¼»
³ x K «¬K »¼
0
e dx te
Some behaviors of the cost rate function (5.3) for different values of parameter
ȕ, of the Weibull density function f(t) are shown in Fig. 5.1. This parameter is
associated with the intensity at which the failure rate function increases.
2.5
2
Cr(t)
1.5
0.5
0
0.5 2.5 4.5 6.5 8.5 10.5 12.5
t (age)
Fig. 5.1 The cost rate function for ca=10, cb=1, Ș =10 and different values of ȕ: ȕ =1 (____); ȕ=2
(…….); ȕ=3 (_ . _); ȕ=4 (_ Ƒ _); ȕ=5 (_ × _)
2.5
2
Cr(t)
1.5
0.5
0
0.5 2.5 4.5 6.5 8.5 10.5 12.5
t (age)
Fig. 5.2 The cost rate function for cb=1, Ș =10 and different values of ca: ca=1(____); ca=3
(…….);ca=10 (_ . _); ca=50 (_ Ƒ _); ca=100 (_ × _)
It is important to note that in some situations the cost rate does not provide
information about the best time to carry out preventive maintenance because the
different alternatives (t ages) have almost the same evaluation in terms of cost
rate. Therefore, for the purpose of making a decision, this aspect does not help the
DM. It should not be considered, for example, in the curve for ca=3 (…….) when
considering t > 4.5.
Alternatives with almost the same evaluation according to one specific criterion
could have very different evaluations in terms of others. That is why it is essential
to make sure that the DM has as broad a view as possible, to make consistent
decisions.
In the next sub-section, other criteria are introduced into the decision problem.
This includes one step from the MCMD/A framework to build the multicriteria
decision problem to support the selection of the preventive maintenance interval.
As stated in Chap. 2, this is one of the most important steps, as the objectives
influence every step in the decision process. In the preventive maintenance
context, the cost is only part of the maintenance objective. As discussed in Chap. 3,
the main objectives of the maintenance function are: to extend the useful life of
assets, to ensure satisfactory levels of availability, to ensure operational readiness
of systems, and to safeguard the people who use the facilities. These objectives are
pursued by the maintenance function as a whole.
It is not necessary to emphasize that for service producing systems, the system
availability is even more important. When failures lead to interruptions of these
systems, they are easily perceived by the customer. Thus, an increase in the
availability may increase the level of user satisfaction. The downtime provides an
indirect measure of this objective. The availability is also related to the reliability,
the capability of the system to work without interruption. The behavior of the
reliability function is shown in Fig. 5.3.
1
0.8
0.6
R(t)
0.4
0.2
0
0 2 4 6 8 10 12
t (age)
Fig. 5.3 Reliability function for, Ș =10 and different values of ȕ: ȕ=1 (____); ȕ=2 (…….); ȕ=3
(_ ___); ȕ=4 (_ Ƒ _); ȕ=5 (_ × _)
As stated in Chap. 2 the state of nature (T) corresponds to aspects that could not be
controlled by the DM and influence the outcome. In fact, they may change
randomly, and consequently, may deeply influence the consequences of the
decision process. The modeling process for this ingredient uses decision theory,
which includes MAUT.
A typical T is the reliability (de Almeida and Souza, 2001), which influences
the outcomes, such as the availability, a usual criterion in the preventive maintenance
decision problem. That is, the reliability is not a consequence, although it may be
considered as such, as a simplification of the model (Cavalcante and de Almeida
2007; de Almeida 2012).
5.2 A General MCDM/A Model for Preventive Maintenance 223
Similar to the set of alternatives, the set of states of nature may be discrete or
continuous and incorporate prior probabilities S(T) on T.
In the maintenance preventive problem, as the data on failure are scarce, the
probability density function that models the time to failure may be completely
unknown or otherwise have undetermined parameters. Thus, if prior probabilities
S(T) are incorporated, a probabilistic modeling task complements the preference
modeling.
Preference Modeling
This step provides information for choosing the MCDM/A method, aligned with
the DM’s preference structure, which may consider, among other factors,
compensatory or non-compensatory rationality. The main question of this factor
considers which of these classes of methods would be more appropriate for a
particular problem. This process could use the model building procedure in Chap.
2. The analysis of the DM’s rationality is essential to ensure that the results from
the MCDM/A model truly reflect the DM’s preferences.
In the next two sections, applications illustrate the use of compensatory and
non-compensatory methods to support the problem of selecting intervals of
preventive maintenance.
Intra-Criterion Evaluation
For a specific problem, this step consists of the elicitation of the value function
vj(x), or utility function uj(x), related to the values of different performances of
outcomes of criterion j, for any j = 1, 2, ..., n.
For a non-compensatory method, an ordinal scale is enough, so the intra-
criterion evaluation is easily quantified. Furthermore, for an outranking method,
the parameters related to indifference, preference and the discordance threshold
may be addressed.
For the compensatory methods, the usual results consist of an overall value for
each alternative that reflects a synthesis of all the criteria for that alternative. This
overall value arises from the aggregation of the utility functions related to each
criterion uj(x). The assessment of the uj(x) relies on the elicitation procedure. The
utility function reflects the preferential structure from the DM for uncertainty
contexts, considering his behavior with respect to risk. The DM could be risk
neutral, averse and prone. Each of these standard behaviors is reflected by a
specific form of the utility function uj(x).
224 Chapter 5 Preventive Maintenance Decisions
Inter-Criteria Evaluation
Elaborating Recommendation
Given the insights and the view achieved by the application of an MCDM/A
method, a complete report should include the essentials of the whole decision
process, as well as, the main aspects that came up during this process. It should
provide any detail that would be requested during the explanation of the results
and the recommended decisions.
Assumptions, simplifications and changes to the original problem should be
explicit and clear, to aid in transmitting an understanding of the results of the
model and their limitations.
One insight of this model is the analysis of the consequence space for the
preventive maintenance decision problem. When this consequence space cannot
be reduced to only one dimension, the classical optimization approach is not
useful. Additionally, when the consequences are multidimensional, the DM’s
preference for each criterion has to be treated very carefully because any mis-
conception or mistake in this process may waste the effort to bring the DM to the
center of the problem.
The context of this problem is an electric power company and it considers the cost
rate and reliability criteria (de Almeida 2012). The underlying model that this
application takes as its base is the model of age based replacement, so all assumptions
and expressions that were presented before are valid in this application.
The set of alternatives corresponds to a discrete set of time intervals. For
instance, months or days may be applied as usual intervals. For a month interval
any alternative is a multiple of 30 days, so that an element of the set of alternatives
could be represented by 30i, where i is any positive integer from 1 to N and N is
the number of alternatives. A quantile of the probability distribution of the time
interval could be used to choose N, although it is not explicit in the model.
It is worth noting that for the age replacement based, whenever a first failure
happens, the timing counting should be restarted. This means that the planning of
the preventive maintenance actions should be performed very carefully, because
the calendar time is not useful to help the manager schedule a particular
preventive maintenance action, as once the schedule is put into action, a failure
can force the schedule to be rearranged. In this way, the use of the base time does
not mean that the action will necessarily happen each 30i days, but rather it means
that 30i is the maximum number of days that a specific item will run until it is
replaced by another. The calendar logic is valid for the block replacement based
policy, in which it is not necessary to keep a register of times to failure.
As already stated the criteria are the cost rate cr(t) and reliability R(t), and their
parameters are presented in Table 5.1.
Weibull ȕ 3
Ș 1200
Costs Replacement Cost Cb 600
Failure Cost Ca 1200
It is possible to build a consequence matrix applying the models for cr(t) and
R(t), as shown in Fig. 5.4.
226 Chapter 5 Preventive Maintenance Decisions
0.8
R(t), cr(t)
0.6
0.4
0.2
0
30 280 530 780 1030
t (age)
Fig. 5.4 The criteria in function of t: R(t) (____) and cr(t) (…….)
For the intra-criteria evaluation, a logistic function was found for the reliability
attribute U(R) and an exponential function for the cost U(cr) attribute. The logistic
utility function for reliability shows that the DM considers the variation at R > 0.9
to be small, and views only changes at R < 0.8 to be important. However, the
higher the cost is, the less the utility is, which reveals the risk averse behavior of
the DM.
With regard to the inter-criteria evaluation, the elicitation process (Keeney and
Raiffa 1976) includes the validation of some axioms about the DM’s preferential
structure. The mutual utility independence between the two attributes was
confirmed, and a multilinear utility function is therefore applied, as given by (5.4).
U ( cr , R ) K1U ( cr ) K 2U ( R ) K 3U ( cr )U ( R ) (5.4)
where:
U(cr) is the utility function for the cost rate criterion;
U(R) is the utility function for the reliability criterion;
U(cr,R) is the multiattribute utility function;
K1, K2 and K3 are the scale constants with K1+K2+K3 = 1.
Following the elicitation procedure, the values obtained for these scale
constants were K1 = 0.35, K2 = 0.45, and K3 = 0.20.
5.4 A Non-Compensatory MCDM/A Model for Preventive Maintenance 227
The results show the highest value for the overall utility function for t = 600 days,
corresponding to 20 months. Applying the classical optimization model, which
considers only the cost rate criterion, the time of the minimum cost rate (t*) is
t=780, which corresponds to 26 months.
The reliability for the time of the minimum cost rate (t*), R(780) is around 0.75.
Therefore, it may be too risky to follow the policy that minimizes the cost rate,
because the probability of a failure of this item may be considered too high to risk
interrupting the supply of electricity. Consequently, the reliability for this item
should not be neglected.
For a large time interval [540, 660] the overall utility function varies over a
range of less than 0.009. In practical terms, this means that the DM has flexibility
regarding the time interval for preventive maintenance without a considerable
decrease in the overall utility value.
Another interesting insight can be recognized when analyzing the change on
the utility form of the cost criterion. This analysis shows that when a linear
function is used to model the cost rate instead of an exponential function, the
overall utility is affected and has its highest value at t*=360 days. In this case, the
DM is not averse to risk to slightly increase the cost, so the smaller time intervals
that were judged unfavorable when using the exponential functions have improved
results using a linear utility function.
The DM may view one of the characteristics of the compensatory method not
suitable. In this approach, alternatives with very poor performance in some criteria
can compensate by good performance in other criteria, and this could happen in
unlimited way. This feature does not apply to the non-compensatory methods.
However, this would not be a reason to change the approach, which should be
based only on the DM’s compensatory rationality. Another way to address this
issue is using the compensatory method with veto (de Almeida 2013).
The criteria considered were cost rate and reliability, as in the previous study. Let
the set of alternatives be A={ti}, where ti= 720 i, for i = 1…12.
The parameters for the criteria are given in Table 5.2.
Weibull ȕ 1.4
Ș 1800
Costs Replacement Cost Cb 300
Failure Cost Ca 1800
3 ( a , b) N
° ¦ Pj ( a, b) w j ½°
j 1
® N ¾ (5.5)
°3 (b, a )
¯ ¦ j
P (b, a ) w j °
1 j ¿
The weight for a criterion j (wj.) has to be established for each criterion based
on the DM’s preferences.
5.4 A Non-Compensatory MCDM/A Model for Preventive Maintenance 229
The scores of the alternatives are based on the outcome and income flows, as
shown in Chap. 2, and recalled in (5.6) and (5.7).
k k
1 1
I (a ) ¦
n 1 x A
3(a, x ) ¦¦
n 1 x A j 1
Pj ( a , x ) w j ¦I
j 1
j (a ) w j (5.6)
k k
1 1
I (a ) ¦
n 1 x A
3 ( x, a ) ¦¦
n 1 x A j 1
Pj ( x, a ) w j ¦I
j 1
j (a ) w j (5.7)
Characteristics R Cm
The next step consists of building the outranking relations, as shown in Table 5.5.
Alternatives T I I
T1 720 0.242142 0.660000
T2 1440 0.307792 0.318536
T3 2160 0.444011 0.022754
T4 2880 0.514990 0.037727
T5 3600 0.505728 0.092916
T6 4320 0.459229 0.164855
T7 5760 0.323753 0.248657
T8 7200 0.180039 0.409705
T9 7920 0.111208 0.533040
T10 8640 0.073333 0.674034
The PROMETHEE I method is applied as given by (5.8), in which PI, II, and RI
correspond to preference, indifference and incomparability, respectively.
230 Chapter 5 Preventive Maintenance Decisions
aS b e aS b
°°
aP I b ®aS b e aI b
°
aI b e aS b
¯°
aI I b aI b e aI b (5.8)
aR I b on the other cases
The best alternatives have been found to be T3 and T4, which correspond to
replacing the components every 2160 or 2880 hours, respectively. These
alternatives are not comparable, indicating that the DM must reflect further when
choosing between them, given that there is not sufficient information or reason,
through the comparison, to have a particular preference for one or the other.
T4 T5 T6 T7 T2 T1 T10
T3 T8 T9
T4 T3 T5 T6 T2 T8 T1 T9 T10
This model takes different assumptions than the previous application, as following
(Cavalcante et al. 2010):
x The time spent on maintenance actions, whether preventive replacement, or
corrective replacement, is non-negligible and known;
x The distribution of the time to failure is known, but its parameters are not.
References 231
With these two basic changes the expressions for R(t) and cr(t) are different.
Despite this increasing in complexity, this model is realistic because it is common
the absence of time to failure data, which makes the parameters of the distribution
of time unknown.
The alternatives that are considered are times multiples of 100; inside the
interval [200, 3000]. These time intervals in units of days.
The data for this problem are given in Table 5.6.
The preference function for the criteria are both type V, as presented in Chap. 2,
which corresponds to the case in which the difference in performance increases as
the difference of evaluation in a criterion increase in a linear relationship. In
addition there are two thresholds: the indifference and preference threshold.
Different values of weight are used to give to the DM more information about
the sensitivity of these values. From the variations applied the first solution
indicated by the PROMETHEE II rank changed from 600 days to 800 days. The
sensitivity analysis provides the DM with more information, indicating the level
of variation that is expected (Cavalcante et al. 2010).
References
Barlow R, Hunter L (1960) Optimum Preventive Maintenance Policies. Oper Res 8:90–100
Cavalcante CAV, Almeida AT de (2007) A multi-criteria decision-aiding model using
PROMETHEE III for preventive maintenance planning under uncertain conditions. J Qual
Maint Eng 13:385–397
Cavalcante CAV, Ferreira RJP, de Almeida AT (2010) A preventive maintenance decision
model based on multicriteria method PROMETHEE II integrated with Bayesian approach.
IMA J Manag Math 21:333–348
Chareonsuk C, Nagarur N, Tabucanon MT (1997) A multicriteria approach to the selection of
preventive maintenance intervals. Int J Prod Econ 49:55–64
Cox DR (1962) Renewal theory, vol 4. Methuen & Co, London
de Almeida AT (2012) Multicriteria Model for Selection of Preventive Maintenance Intervals.
Qual Reliab Eng Int 28:585–593
232 Chapter 5 Preventive Maintenance Decisions
de Almeida AT (2013) Additive-veto models for choice and ranking multicriteria decision
problems. Asia-Pacific J Oper Res 30(6):1-20
de Almeida AT, Ferreira RJP, Cavalcante CAV (2015) A review of multicriteria and multi-
objective models in maintenance and reliability problems. IMA Journal of Management
Mathematics 26(3):249–271
de Almeida AT, Souza FMC (2001) Gestão da Manutenção: na Direção da Competitividade
(Maintenance Management: Toward Competitiveness) Editora Universitária da UFPE, Recife
Glasser GJ (1969) Planned replacement- Some theory and its application (Probability theory
applied to age and block replacement models in preventive maintenance of parts, noting
inspection cost distribution). J Qual Technol 1:110–119.
Hwang CL, Tillman FA, Wei WK, Lie CH (1979) Optimal Scheduled-Maintenance Policy
Based on Multiple-Criteria Decision-Making. Reliab IEEE Trans R-28:394–399
Inagaki T, Inoue K, Akashi H (1978) Interactive Optimization of System Reliability Under
Multiple Objectives. Reliab IEEE Trans R-27:264–267
Jiang R, Ji P (2002) Age replacement policy: a multi-attribute value model. Reliab Eng Syst Saf
76(3): 311-318
Keeney RL, Raiffa H (1976) Decisions with multiple objectives: Preferences and Value Trade-
Offs. Wiley Series in Probability and Mathematical Statistics. Wiley and Sons, New York
Scarf PA, Dwight R, Al-Musrati A (2005) On reliability criteria and the implied cost of failure
for a maintained component. Reliab Eng Syst Saf 89:199–207
Shafiee M, Finkelstein M (2015) An optimal age-based group maintenance policy for multi-unit
degrading systems. Reliab Eng Syst Saf 134:230-238
Slack N, Chambers S, Johnston R (2010) Operations management, 6th ed. Pearson Education,
Harlow
Chapter 6
Decision Making in Condition-Based
Maintenance
Abstract: (G:9>8I>K: B6>CI:C6C8: BD9:A>C< >H 6 IDDA I=6I 86C EGDK>9: B6CN
7:C:;>IH ID I=: B6>CI:C6C8: B6C6<:B:CI 6G:6 ,=>H 8=6EI:G HJ<<:HIH 6 BJAI>
8G>I:G>6 6EEGD68= ;DG BD9:A>C< 8DC9>I>DC 76H:9 B6>CI:C6C8: 9:8>H>DCH /=:G:6H
I=: (G:9>8I>K: B6>CI:C6C8: JH:H 8DC9>I>DC >C;DGB6I>DC D; 6HH:IH I=: EG:K:CI>K:
B6>CI:C6C8: >H I>B:76H:9 DCAN ,=>H 8=6EI:G 9>H8JHH:H 9:8>H>DC B6@>C< >C
8DC9>I>DC76H:9 B6>CI:C6C8: % 6C9 EG:H:CIH HDB: JH:;JA 6EEGD68=:H ;DG
7J>A9>C<BJAI>EA:D7?:8I>K:BD9:AH>CI=>H8DCI:MI !C>I>6AAN6HJBB6GN>H<>K:CD;
I=: ;JC96B:CI6AH D; % 6C9 H:K:G6A 8DC8:EIH D; BDC>IDG>C< 6C9 >CHE:8I>DC
68I>K>I>:H 6H>8 8DC8:EIH D; 9:A6N I>B: 6G: EG:H:CI:9 6C9 9>H8JHH:9 L>I=>C 6C
%%
6EEGD68= ,=:C 6 HIGJ8IJG: D; 6 BJAI>8G>I:G>6 BD9:A ID 9:I:GB>C:
>CHE:8I>DC>CI:GK6AHD;8DC9>I>DCBDC>IDG>C<76H:9DC%JAI>6IIG>7JI:JI>A>INI=:DGN
%-, >H >CIGD9J8:9 ,=: 6HE:8IH D; EG:;:G:C8: BD9:A>C< H86A: 8DCHI6CI
:A>8>I6I>DCJI>A>INI=:DGN6C96%PH7:=6K>DGIDG>H@EGDC:C:JIG6A6K:GH:6G:
>C8AJ9:9>CI=:9:8>H>DCBD9:A C>AAJHIG6I>K::M6BEA:D;6C%%
BD9:A>H
<>K:C >C I=: 8DCI:MI D; 6C :A:8IG>8 EDL:G 9>HIG>7JI>DC HNHI:B ,=JH I=:
BJAI>8G>I:G>6 BD9:A EG:H:CI:9 L=>8= >H 76H:9 DC %-, 6C9 L=>8= =6H 6C
6M>DB6I>8HIGJ8IJG:6>BHID6CHL:GI=:C::9H>9:CI>;>:96C9:C67A:HI=:IG69:D;;
6BDC<I=:8DHIH9DLCI>B:6C9;G:FJ:C8ND;7G:6@9DLCHID7:9:6AIL>I=
6.1 Introduction
&DLA6C 6C9 :6E HI6I:9 I=6I 6C >I:B >H H6>9 ID 7: B6>CI6>C:9 7N
8DC9>I>DC BDC>IDG>C< >; >I >H E:GB>II:9 ID G:B6>C >C H:GK>8: L>I=DJI EG:K:CI>K:
B6>CI:C6C8: JCI>A 6 ;JC8I>DC6A ;6>AJG: D88JGH DC9>I>DC6H:9 %6>CI:C6C8:
%>H8DCH>9:G:967GD69:GHJ7?:8I6C9>C8AJ9:H8DC9>I>DCBDC>IDG>C< %>H
HI6I:97N"6G9>C::I6A
6H7:>C<6B6>CI:C6C8:EGD<G6BI=6IG:8DBB:C9H
B6>CI:C6C8: 68I>DCH 76H:9 DC I=: >C;DGB6I>DC 8DAA:8I:9 JH>C< 8DC9>I>DC
BDC>IDG>C< :H>9:H % 6II:BEIH ID 6KD>9 JCC:8:HH6GN B6>CI:C6C8: I6H@H 7N
I6@>C<B6>CI:C6C8:68I>DCHDCANL=:CI=:G:>H:K>9:C8:D;67CDGB6A7:=6K>DGHD;
6 E=NH>86A 6HH:I ,=>H IDE>8 L6H ;DJC9 >C D; EJ7A>86I>DCH 8DCH>9:G>C<
%%
6EEGD68=:H>C6A>I:G6IJG:G:K>:L9:AB:>96:I6A
/6C< 6C9 6D
>C9>86I: I=6I I=:G: >H 6C >C8G:6H:9 C::9 ;DG BDG:
:;;:8I>K: 6C9 :;;>8>:CI I:8=C>FJ:H I=6I BDC>IDG B68=>C: 8DC9>I>DCH >C G:6A I>B:
9:I:8I I=: >C8:EI>DC 6C9 EGD<G:HH>DC D; 9:;:8IH 6C9 :C67A: ;A:M>7A: B6>CI:C6C8:
H8=:9JA>C<7:;DG:69:;:8IG:HJAIH>CJC:ME:8I:9B68=>C:9DLCI>B: "6G9>C::I6A
HI6I:I=6I9>6<CDHI>8H6C9EGD<CDHI>8H6G:ILD>BEDGI6CI6HE:8IH>C6%
EGD<G6B /=>A:9>6<CDHI>8H9:6AL>I=I=:9:I:8I>DC>HDA6I>DC6C9>9:CI>;>86I>DCD;
9:;:8IH L=:C I=:N D88JG EGD<CDHI>8H 9:6A L>I= I=: EG:9>8I>DC D; 9:;:8I 7:;DG: >I
D88JGH
>6<CDHI>8 I:8=C>FJ:H 6G: IDDAH I=6I =6K: 6C >C8G:6H>C< 6EEA>867>A>IN >C
8DBE6C>:H 9J: ID I=: EDI:CI>6A ;>C6C8>6A G:IJGCH :HE:8>6AAN L=:C 8DBE6G:9 L>I=
I=: EDA>8>:H D; 8DGG:8I>K: 6C9 EG:K:CI>K: B6>CI:C6C8: !C8AJ9:9 6BDC< I=: B6>C
I:8=C>FJ:H6G:EGD8:9JG:H;DGB:6HJG>C<DG6C6ANO>C<.>7G6I>DC8DJHI>8:B>HH>DC
'>A 6C6ANH>H -AIG6HDC>8 ,=:GBD<G6E=N ,:BE:G6IJG: +E::9 (:G;DGB6C8:
DGGDH>DC'JIEJIEDL:G(G:HHJG:A:8IG>8JGG:CI J>9:A>C:HDC8DAA:8I>C<6C9
6C6ANO>C<8DC9>I>DCBDC>IDG>C<96I686C7:;DJC9>C!+'
,=:9>6<CDHI>8HEGD7A:B86C7:9:H8G>7:96H6=NEDI=:H>HI:HIEGD7A:BH>C8:
I=: HI6I: D; I=: HNHI:B >H JC@CDLC C :;;>8>:CI 9>6<CDH>H >H DC: I=6I =6H 6
B>C>B6A:GGDGG6I: H>C=NEDI=:H>HI:HI>C<I=:G:B6N7:ILD INE:HD;:GGDGH>C
9>6<CDH>H !I86CB>HI6@:CAN8A6>B:9I=6I6EGD9J8I>H9:;:8I>K:L=:C>C;68I>I>H
CDI6C9I=>HINE:D;:GGDG>H6AHD86AA:96;6AH:C:<6I>K: 'CI=:DI=:G=6C9>I86C
7:H6>9I=6I6EGD9J8I>HCDI9:;:8I>K:L=:C>C;68II=:G:>H67J<6AHD86AA:9;6AH:
EDH>I>K: :GG69::I6A
EG:H:CI:9ILDBD9:AH>CDG9:GID9:6AL>I=:GGDG>C
I=:>CHE:8I>DCEGD8:HHHJ8=6H;6AH:EDH>I>K:H;6AH:6A6GBH6C9;6AH:C:<6I>K:HD;
EGDI:8I>DCHNHI:BH
%6GI>C 9>HI>C<J>H=:H 7:IL::C =6G9 6C9 HD;I ;6JAIH ,=:H: ;6JAIH 6G:
H=DLC>C>< ,=:HD;I;6JAIA:69HID6EG:9>8I67A:H>IJ6I>DC>IA:C9H>IH:A;ID
8DC9>I>DC BDC>IDG>C< L=>A: I=: =6G9 ;6JAI I6@:H EA68: >CHI6CI6C:DJHAN : HI6I:H
I=6I EG:9>8I>K: B6>CI:C6C8: >CKDAK:H E:G>D9>8 BDC>IDG>C< DC I=: =:6AI= D; I=:
B68=>C: 6C9 H8=:9JA>C< B6>CI:C6C8: DCAN L=:C 6 ;JC8I>DC6A ;6>AJG: >H 9:I:8I:9
,=>H 6AADLH ;DG IG:C9H D; I=: B68=>C: 8DBEDC:CI ID 7: 8DCHIGJ8I:9 6C9 I>B: ID
;6>AJG:ID7::HI>B6I:9
(GD<CDHI>8I:8=C>FJ:H6>BID:HI>B6I:I=:G:H>9J6AA>;:D;ID6E>:8:D;:FJ>E
B:CII6@>C<8DC9>I>DCBDC>IDG>C<>CID8DCH>9:G6I>DC ,=:8DC8:EID;EGD<CDHI>8H
<D:H7:NDC99>6<CDHI>8H >GHII=:EGD7A:B>H9:I:8I:9I=:C69>6<CDH>H>HB69:
%DC>IDG>C<6C9!CHE:8I>DC8I>K>I>:H
67DJII=:;6>AJG:BD9:6C9>IHH:K:G>IN !I>H6AHD>BEDGI6CIIDEG:9>8II=::KDAJI>DC
D; I=: ;6>AJG: >C DG9:G ID :HI>B6I: I=: G:B6>C>C< JH:;JA A>;: D; I=: B68=>C:
:C6N6:I6A
D.6C6C9TG:C<J:G
Hard failure
Good
Condition
Soft failure
Failure
Time t2
t1
Fig. 6.1DC9>I>DC:KDAJI>DC;GDB=6G96C9HD;I;6>AJG:H
,=: 9:K:ADEB:CI D; BDC>IDG>C< 6C9 9>6<CDHI>8 HNHI:BH ;DG 6>G8G6;I 6C9 DI=:G
8DBEA:MHNHI:BH=6HA:9IDI=:G:8D<C>I>DCI=6IEG:9>8I>K:EGD<CDH>H>H9:H>G:96C9
I:8=C>86AAN EDHH>7A: ,=: BDHI JH:9 EGD<CDHI>8H I:8=C>FJ: >H 8DC8:GC:9 L>I=
EG:9>8I>C<=DLBJ8=I>B:>HA:;I7:;DG:6;6>AJG:D88JGH<>K:CI=:8JGG:CIB68=>C:
8DC9>I>DC 6C9 E6HI DE:G6I>DC EGD;>A: JHJ6AAN 86AA:9 >IH G:B6>C>C< JH:;JA A>;:
*-$ "6G9>C: :I 6A
8A6HH>;>:9 G:H:6G8= DC EGD<CDHI>8H >C I=G:: 6G:6H
G:B6>C>C< JH:;JA A>;: EGD<CDHI>8H >C8DGEDG6I>C< B6>CI:C6C8: EDA>8>:H 6C9
8DC9>I>DCBDC>IDG>C<>CI:GK6A
88DG9>C< ID /6C<
8JGG:CI EGD<CDHI>8 6EEGD68=:H 86C 7: 8A6HH>;>:9
>CID I=G:: 76H>8 <GDJEH 6 BD9:A76H:9 6EEGD68= 6 96I69G>K:C 6EEGD68= 6C9 6
=N7G>9 6EEGD68= *:H>9J6A A>;: BD9:AA>C< K>6 HID8=6HI>8 ;>AI:G>C< >H 6 G:A:K6CI
BD9:AA>C<I:8=C>FJ:JH:9>C%/6C<6C9=G>HI:G
"6G9>C::I6A
EGDEDH:9 EGDEDGI>DC6A =6O6G9H BD9:AA>C< ( % >C DG9:G ID >C8DGEDG6I>C<
:MEA6C6IDGN K6G>67A:H >CID 6 BD9:A ;DG :HI>B6I>C< I=: ;6>AJG: G6I: .AD@ :I 6A
JH:9 6 /:>7JAA EGDEDGI>DC6A=6O6G9H BD9:A ID 9:I:GB>C: I=: DEI>B6A
G:EA68:B:CI ;DG 6C >I:B L=>8= >H HJ7?:8I ID K>7G6I>DC BDC>IDG>C< /6C<
DJIA>C:H6H:B>HID8=6HI>8;>AI:G>C<76H:9G:H>9J6AA>;:EG:9>8I>DC6EEGD68=;DGI=:
>I:BHBDC>IDG:9>C% !C<:C:G6AEGD<CDHI>8H>C;DGB6I>DC>H6@:N:A:B:CI>C
BD9:A>C<I=:9:8>H>DCB6@>C<6HE:8ID;%
6GADL6C9(GD8=6CBD9:A:9>CHE:8I>DCEDA>8>:HL=>8=6HHJB:I=6I;6>AJG:
>H9>H8DK:G:9DCAN7N68IJ6A>CHE:8I>DC6C9>C<:C:G6ADCAN6;I:GHDB:I>B:=6H
:A6EH:9H>C8:I=:D88JGG:C8:D;I=:;6>AJG:6C9:K6AJ6I:9H8=:9JA:HD;>CHE:8I>DC
=6EI:G :8>H>DC%6@>C<>CDC9>I>DC6H:9%6>CI:C6C8:
I>B:HL=>8=B>C>B>O:I=:IDI6A:ME:8I:98DHIG:HJAI>C<;GDB7DI=>CHE:8I>DC6C9
;6>AJG: ,=:N6HHJB:I=6I;6>AJG::KDAK:;GDB69:I:G>DG6I>DCEGD8:HH ,=>HEGD8:HH
>H 6HHJB:9 HID8=6HI>8 6C9 I=: 8DC9>I>DC D; I=: HNHI:B >H @CDLC DCAN I=GDJ<=
>CHE:8I>DC
!C >CHE:8I>DC EDA>8>:H CD G:EA68:B:CI DG G:E6>G >H G:8DBB:C9:9 7:;DG:
9:I:8I>DCD;;6>AJG: 68=>CHE:8I>DC=6H68DHII=6I>BEA>:H7:>C<JC;:6H>7A:>CHE:8I
K:GND;I:C DL:K:G6ADC<A6EH:D;I>B:7:IL::C;6>AJG:6C99:I:8I>DC>BEA>:H6
E:C6AIN8DHI ,=:B6>C8=6AA:C<:>HID;>C9I=:7:HI>CHE:8I>DCEDA>8N>CDG9:GID
B>C>B>O::ME:8I:9IDI6A8DHI DG:M6BEA:I=:G:6G:HNHI:BHHJ7?:8IIDEG:8:9:C8:
8DCHIG6>CIH6C96H:FJ:C8:D;>CHE:8I>DCHH=DJA97:9:I:GB>C:9=>J:I6A
6GADL 6C9 (GD8=6C EGDEDH:9 6 BD9:A I=6I B>C>B>O:H :ME:8I:9 8DHI
JCI>A9:I:8I>DCD;;6>AJG:6BD9:AI=6IB>C>B>O:H:ME:8I:98DHI6HHJB>C<G:C:L6A
6I9:I:8I>DCD;;6>AJG:6C96CDEEDGIJC>HI>8G:EA68:B:CIHIG6I:<ND;6H>C<A:E6GI>C
I=:EG:H:C8:D;H:K:G6ABDC>IDG:9E6GIH
:C6N6 :I 6A
ED>CI:9 DJI I=6I ;DG HDB: HNHI:BH 6 8DCI>CJDJH
BDC>IDG>C<D;I=:>GDE:G6I>C<HI6I:H>HCDI:8DCDB>86AAN?JHI>;>67A:6C9>CHE:8I>DCH
6G: JH:;JA >C BDC>IDG>C< I=: 8DC9>I>DC D; I=: HNHI:B 6I EG:9:I:GB>C:9 I>B:H >C
DG9:GIDG:9J8:I=:EGD767>A>IND;>IHB6A;JC8I>DC>C< 'C8:6CDJID;8DCIGDAHI6I:
>H9:I:8I:96G:E6>G>H86GG>:9DJIIDG:HIDG:I=:HNHI:BID>IH>C8DCIGDAHI6I:
,=:G: 6G: HNHI:BH I=6I HNBEIDBH D; ;6>AJG: 6G: CDI 6EE6G:CI 6C9 I=: A:K:A D;
9:<G696I>DCDG9:I:G>DG6I>DC86C7:@CDLCDCANI=GDJ<=>CHE:8I>DCDJA69>G69
6C9G6AA
=:A7>6C9>I#69>
G6AA:I6A
JNC=:I6A
+DB: :M6BEA:H 6G: 6A6GB 6C9 HI6C97N HNHI:BH C >CHE:8I>DC HIG6I:<N :HI67A>H=:H
I=:I>B:6IL=>8=DC:DGBDG:DE:G6I>C<E6G6B:I:GH=6K:ID7:8DCIGDAA:9>CDG9:G
ID9:I:GB>C:>;I=:HNHI:B>H>C6CDE:G6I>C<DG6;6>AJG:HI6I:=:A7>6C9>I#69>
,LD<:C:G6AH>IJ6I>DCHL:G:>9:CI>;>:97N=:A7>6C9>I#69>
>GHIAN
>CHE:8I>DCH8DCH>HIH>BEAN>C6HH:HH>C<>;I=::FJ>EB:CI>HLDG@>C<DG>C6;6>A:9
HI6I: +:8DC9ANI=::FJ>EB:CI8DC9>I>DC86C7:6HH:HH:9I=GDJ<=9>G:8IDG>C9>G:8I
8DCIGDA6C9EG:K:CI>K:68I>DCH86C7:9DC:7:;DG:;6>AJG:D88JGG:C8:>I>H@CDLC
6H%
76H>8 >CHE:8I>DC BD9:A HJ<<:HI:9 7N =:A7> 6C9 >I#69>
6HHJB:9
I=6I:FJ>EB:CI>CHE:8I:96I>CHI6CIHxi. !;6C>CHE:8I>DCG:K:6AHI=6II=::FJ>EB:CI
>H>C6;6>A:9HI6I:6C:L>9:CI>86ADC:>BB:9>6I:ANG:EA68:H>I ,=:H:FJ:C8:D;
>CHE:8I>DC>CHI6CIH>HH=DLC>C><
Failure
0 x1 x2 … xi-1 xi xi+1
Inspection Times
Fig. 6.2,=:H:FJ:C8:D;>CHE:8I>DCI>B:H
:A6N,>B:%D9:AHID+JEEDGI%
=G>HI:G 6C9 /6AA:G >CIGD9J8:9 6C >CHE:8I>DC BD9:A 76H:9 DC ILD HI6<:
;6>AJG: EGD8:HH 86AA:9 :A6N ,>B: , ,=:N 86AA:9 I=: >C>I>6A ED>CI u D; I=:
9:;:8I I=: ;>GHI DEEDGIJC>IN L=:G: I=: EG:H:C8: D; 6 9:;:8I B><=I G:6HDC67AN 7:
:ME:8I:9ID7:G:8D<C>O:97N6C>CHE:8I>DC6C9I=:I>B:hID;6>AJG:;GDBu86AA:9
I=: 9:A6N I>B: D; I=: 9:;:8I 6H H=DLC >C >< :A6N I>B: G:EG:H:CIH 6 I>B:
L>C9DL;DGEG:K:CI>C<6;6>AJG:6;I:G69:;:8ID88JGG:9/6C<
h
u failure
Fig. 6.3:A6N,>B:hD;6;6>AJG:
!C,BD9:AHI=:;6>AJG:EGD8:HH>H6HHJB:9ID7:6CDC=DBD<:C:DJH(D>HHDC
EGD8:HH ,=: DE:G6I>DC6A 8DHI D; 6EEAN>C< 6C >CHE:8I>DC EDA>8N 86C 7: B:6HJG:9
,=:>CHE:8I>DC8DHI9:CDI:97NCiG:EG:H:CIHI=:K6AJ:D;I=:G:HDJG8:HC::9:9ID
E:G;DGB6C>CHE:8I>DCI6H@ ,=:>CHE:8I>DCG:E6>G8DHI9:CDI:97NCr8DCH>HIHD;
I=: 8DHIH C:8:HH6G>AN >C8JGG:9 ID G:E6>G 6 ;6JAI >9:CI>;>:9 >C I=: >CHE:8I>DC
6H>86AANI=:8DHID;67G:6@9DLC>H6E:C6AIN8DHI6C9>H6HHD8>6I:9L>I=I=:8DHI
D; I=: 8DCH:FJ:C8:H 86JH:9 7N 6 ;6>AJG: L=>8= >H 6I 7DIIDB I=: 8DHI G:A6I:9 ID
ADHH D; EGD9J8I>DC ,=: 7G:6@9DLC G:E6>G 8DHI >H 9:CDI:9 7N Cb 6H:9 DC I=:
9:A6N I>B: 8DC8:EI 6C9 69B>II>C< I=6I I=: BDHI >BEDGI6CI 6HHJBEI>DC I=6I >H
7GDJ<=I;GDBI=>H;JC96B:CI6A6EEGD68=>H8DHI688DG9>C<ID=G>HI:G6C9/6AA:G
I=::ME:8I:98DHID;6C>CHE:8I>DCEDA>8N;DG676H>8>CHE:8I>DCBD9:A>H
<>K:C>C
L=:G:
TQI>B:7:IL::C>CHE:8I>DCH
f(h) Q I=:EGD767>A>IN9:CH>IN;JC8I>DCD;9:A6NI>B:
O>HI=:6GG>K6AG6I:D;9:;:8IHE:GJC>II>B:
CbQI=:6K:G6<:7G:6@9DLCG:E6>G8DHI
CrQI=:6K:G6<:G:E6>G8DHI
CiQI=:6K:G6<:>CHE:8I>DC8DHI
dbQI=:6K:G6<:9DLCI>B:IDG:E6>G67G:6@9DLC
=6EI:G :8>H>DC%6@>C<>CDC9>I>DC6H:9%6>CI:C6C8:
dQI=::ME:8I:99JG6I>DCD;I=:>CHE:8I>DCI>B:9,
b(T)QI=:EGD767>A>IND;6;6JAI86JH>C<67G:6@9DLC
,=: >C>I>6A >CHI6CI 6I L=>8= 6 9:;:8I B6N 7: 6HHJB:9 ID ;>GHI 6G>H: L>I=>C I=:
EA6CI>HJC>;DGBAN9>HIG>7JI:9DK:GI>B:H>C8:I=:A6HI>CHE:8I>DC6C9>C9:E:C9:CI
D;h6H<>K:C7N
T
§T h·
bT ³ ¨© ¸ f h dh
T ¹
!C I:GBH D; 6K6>A67>A>IN I=>H 8G>I:G>DC >H G:A6I:9 ID I=: CDCBDC:I6GN 6HE:8IH
6HHD8>6I:9 L>I= I=: ;6>AJG:H ,=>H G:;A:8IH I=: 67>A>IN D; I=: HNHI:B ID E:G;DGB
JC9:GI=:>C;AJ:C8:D;6C>CHE:8I>DCEDA>8N K6>A67>A>IN>H67DJII=:E:G8:CI6<:D;
I>B:I=6II=:HNHI:B>H6K6>A67A: +D>CI:GBHD;6H:GK>8:HNHI:BI=>H6K6>A67>A>IN
8DGG:HEDC9H ID I=: E:G8:CI6<: D; I>B: I=6I I=: H:GK>8: >H EGDK>9:9 ID I=: 8A>:CI
,=JHI=:ADL:GI=:6K6>A67>A>IN>HI=:<G:6I:GI=:8A>:CIPH9>HH6I>H;68I>DC
99>I>DC6AAN 6 8G>I>86A ;68IDG ;DG 6HH:HH>C< I=: E:G;DGB6C8: D; 6C >CHE:8I>DC
EDA>8N>HHNHI:B6K6>A67>A>IN ,=:9DLCI>B:86C7::K6AJ6I:9>CDG9:GIDG:EG:H:CI
I=>H;68IDG 88DG9>C<ID=G>HI:G6C9/6AA:GI=::ME:8I:99DLCI>B:;DG6
76H>8>CHE:8I>DCBD9:A>H<>K:C7N
O T d b bT d
DT
T d
>6<CDHI>8 I:8=C>FJ:H =6K: 7::C 9:K:ADE:9 L=>8= 6G: 67A: ID >9:CI>;N I=: BDHI
>C8>E>:CI9:;:8IH6C9EGD<CDHI>8I:8=C>FJ:H6G:67A:ID:HI>B6I:G:H>9J6AA>;:BDG:
688JG6I:AN I=:H: ILD 6G:6H D; % =6K: 7::C 76H:9 DC B6I=:B6I>86A 6C9
HI6I>HI>86A :HI>B6I>DC I:8=C>FJ:H "6G9>C: :I 6A
EG:H:CI:9 96I6 68FJ>H>I>DC
96I6 EGD8:HH>C< 6C9 B6>CI:C6C8: 9:8>H>DCB6@>C< 6H I=G:: @:N HI:EH D; 6 %
EGD<G6B +JEEDH: 6 H:I D; :FJ>EB:CI D; I=: H6B: INE: =6K: I=: H6B: G:H>9J6A
JH:;JA A>;: :E:C9>C< DC I=: >BE68I D; I=: ;6>AJG: 6 E>:8: D; :FJ>EB:CI 86C 7:
>CHE:8I:9 BDG: ;G:FJ:CIAN DG EG:K:CI>K: B6>CI:C6C8: 68I>K>IN 86C 7: 6CI>8>E6I:9
9J:IDI=:EGD767>A>IND;;6>AJG: ,=:%PHEG:;:G:C8:H;DGI=:E:G;DGB6C8:D;6
B6>CI:C6C8:EDA>8N86C7:BD9:A:99JG>C<I=:9:8>H>DCEGD8:HH /=:CBDG:I=6C
DC: D7?:8I>K: >H 8DCH>9:G:9 6 %%
BD9:A 86C 7: 9:K:ADE:9 ID HJEEDGI 6
% EGD<G6B +DB: :M6BEA:H D; %%
BD9:AH >C % EGD<G6BH 6G:
9>H8JHH:9
6GC:GD
EGDEDH:9 6C :K6AJ6I>DC HNHI:B D; H:II>C< JE 6 EG:9>8I>K:
B6>CI:C6C8: EGD<G6B JH>C< I=: C6ANI>8 >:G6G8=N (GD8:HH ( 6N:H>6C
%%
%D9:ADCDC9>I>DC%DC>IDG>C<
,=:BDHI8DBBDC6HHJBEI>DC>HI=6I6CN;6>AJG:>H9:I:8I:96II=:I>B:D;I=:C:MI
8=:8@ 6C9 6 G:EA68:B:CI >H >BB:9>6I:AN B69: 6GADL 6C9 (GD8=6C
&6@6<6L6
9:G>K6I>K: 6EEGD68= 6C6ANO:H I=: 9:A6N I>B: :A6N I>B: 6
=6EI:G :8>H>DC%6@>C<>CDC9>I>DC6H:9%6>CI:C6C8:
ILDHI:E;6>AJG:EGD8:HH>HI=:I>B:A6EH:;GDBL=:C6HNHI:B9:;:8I8DJA9;>GHI
=6K:7::CCDI>8:9JCI>AI=:I>B:L=:C>IHG:E6>G86CCDADC<:G7:9:A6N:97:86JH:
D;JC688:EI67A:8DCH:FJ:C8:HHJ8=6H6H:G>DJH86I6HIGDE=:L=>8=B><=I6G>H:9J:
ID;6>AJG:=G>HI:G ,=:>BEDGI6C8:D;9:A6NI>B:>CB6>CI:C6C8:B6C6<:
B:CI 6EEA>86I>DCH L6H >CK:HI><6I:9 7N /6C<
:A6N ,>B: BD9:AH =6K:
7::C6EEA>:9>CH:K:G6A8DCI:MIH !C6B6CJ;68IJG>C<>C9JHIGN"DC:H:I6A
:K6AJ6I: 6 HJ7?:8I>K: B:6HJG: D; I=: ;6>AJG: 8DCH:FJ:C8:H 76H:9 DC 9:A6N I>B:
6C6ANH>H >C I:GBH D; 8DHI ID I=: :CK>GDCB:CI >C BDC:I6GN K6AJ: ID I=: 8DBE6CN
6C9 I=: 96B6<>C< :;;:8I ID I=: 8DBE6CN >B6<: :A6N ,>B: BD9:A >H 6AHD
EGDEDH:9ID9:I:GB>C:>CHE:8I>DC>CI:GK6AHDC;>H=>C<K:HH:AH(>AA6N:I6A
!CHE:8I>DC I6H@H 6G: 67A: ID >9:CI>;N >CI:GB:9>6I: HI6I:H 7:;DG: ;6>AJG:
%%
9:8>H>DC BD9:A >C DG9:G ID 6>9 B6>CI:C6C8: EA6CC>C< >C >CHE:8I>DC
BD9:AHL6H9:K:ADE:976H:9DC%-,7N:GG:>G6:I6A
,=>HBD9:AI6@:H
I=:%PHEG:;:G:C8:H>CID688DJCI6HL:AA6HI=:BDHI>BEDGI6CI6HE:8IHL=:C
8DCH>9:G>C<H:II>C<I=:>CHE:8I>DC>CI:GK6AH;DGE:G>D9>88DC9>I>DCBDC>IDG>C<6C9
I=:H:6G:I=:8DHI6C99DLCI>B:6HHD8>6I:9L>I=I=:>CHE:8I>DCEDA>8N
!CHE:8I>DC 86C 7: 9:;>C:9 6H 6 I6H@ D; :M6B>C>C< 6C9 D7H:GK>C< >C DG9:G ID
8A6HH>;N6C>CHE:8I:9>I:B>CI:GBHD;>IH;:6IJG:H6C9EGDE:GI>:H /=:C>CHE:8I>DCH
6G: L:AA 9:;>C:9 6 8DBE6CN 86C B>C>B>O: B6>CI:C6C8: 8DHIH 6C9 >BEGDK: I=:
6K6>A67>A>IN D; HNHI:BH !C <:C:G6A I=: >CI:G:HI >H >C 9>H8DK:G>C< I=: HI6I: D; 6C
6HH:I> : L=:I=:G>I>H9:;:8I>K:DGCDI
%6C6<:GH 6G: >CI:G:HI:9 >C 76A6C8>C< I=: 8DHIH D; I=: >CHE:8I>DC EDA>8N L>I=
H6K>C<H6G>H>C<;GDB>BEGDK>C<I=:E:G;DGB6C8:D;I=:HNHI:B !C;68I9:E:C9>C<
DC I=: EGD9J8I>DC EGD8:HH I=6I >H HJEEDGI:9 7N 6 8DBEA:M HNHI:B 6 K:GN H=DGI
>CI:GGJEI>DC9J:ID6;6>AJG:86C86JH:6HJ7HI6CI>6A;>C6C8>6AADHH
+>C8:I=:BD9:A>H76H:9DCI=:8DC8:EID;9:A6NI>B:I=:8DCHIGJ8I>DCD;I=:
BD9:AH=DJA98DCH>9:GI=:8DBBDC6HHJBEI>DCH9:;>C:9>CI=:9:A6NI>B:6EEGD68=
%-, BD9:A >H EGDEDH:9 L=>8= 8DCH>9:GH I=6I 6IIG>7JI:H D; 8DHI 6C9
6K6>A67>A>IN6G:699>I>K:>C9:E:C9:CI>;6C9DCAN>;I=:ILD6IIG>7JI:JI>A>IN;JC8I>DC
>H699>I>K: DGI=:H:8G>I:G>6I=:699>I>K:;DGBB6N7:LG>II:C6H
L=:G:
kcQH86A:8DCHI6CI;DGI=:8DHI8G>I:G>DC
uc(C(T))Q8DC9>I>DC6AJI>A>IN;JC8I>DC;DGI=:8DHI8G>I:G>DC
kdQH86A:8DCHI6CI;DGI=:9DLCI>B:8G>I:G>DC
ud(D(T))Q8DC9>I>DC6AJI>A>IN;JC8I>DC;DGI=:9DLCI>B:8G>I:G>DC
88DG9>C< ID #::C:N 6C9 *6>;;6 I=: 6HH:HHB:CI EGD8:HH ;DG 6 %-
JC8I>DC8DCH>HIH76H>86AAND;;>K:HI:EH>CIGD9J8>C<I=:I:GB>CDAD<N6C9>9:6
>9:CI>;N>C<G:A:K6CI>C9:E:C9:C8:6HHJBEI>DCH6HH:HH>C<8DC9>I>DC6AJI>A>IN
;JC8I>DCH 6HH:HH>C< H86A>C< 8DCHI6CIH 8=:8@>C< ;DG 8DCH>HI:C8N 6C9
G:>I:G6I>C<
%%
%D9:ADCDC9>I>DC%DC>IDG>C<
,=:;>GHIHI:E8DCH>HIHD;B6@>C<I=:%JC9:GHI6C9I=:B6>CEJGEDH:D;I=:
JI>A>IN;JC8I>DC6C9:HE:8>6AANHDI=6I=:
H=:JC9:GHI6C9HI=:8DCH:FJ:C8:HE68:
DG DJG E6GI>8JA6G 86H: >I >H >BEDGI6CI ID CDI>8: I=6I t I>B: ID >CHE:8I>DC >H
8DCH>9:G:9 6 ;:6H>7A: 6AI:GC6I>K: 6C9 T >H I=: H:I D; 6AA I>B:D;>CHE:8I>DC
6AI:GC6I>K:H !CI=>H86H:I=:H:ID;6AAI>B:HD;>CHE:8I>DC6AI:GC6I>K:H>H9:;>C:9
7NT3 DG:68=>CHE:8I>DCI>B:tI=:G:>H68DCH:FJ:C8:>CI:GBHD;8DHI
C(t) 6C9 9DLCI>B: D(t) DG :M6BEA: I=: ED>CI C(t1) D(t1) 7:ADC<H ID I=:
8DCH:FJ:C8: HE68: C :M6BEA: D; I=: 8DCH:FJ:C8: HE68: D; I=>H EGD7A:B >H
H=DLC>C><
D(T)
D0(T)
D*(T)
,=: H:8DC9 HI:E 8DCH>HIH D; >9:CI>;N>C< HDB: >C9:E:C9:C8: 6HHJBEI>DCH I=6I
6G: IGJ: ;DG I=: % L>I= G:<6G9 ID I=: 8G>I:G>6 I=6I 6G: 7:>C< 8DCH>9:G:9 -I>A>IN
>C9:E:C9:C8: C::9H ID 7: 8=:8@:9 >C DG9:G ID :K6AJ6I: I=: =NEDI=:H>H D; I=:
I=:DGN !; I=>H >C9:E:C9:C8: D88JGH I=:C I=: %- ;JC8I>DC 86C 7: H>BEA:
DI=:GL>H:BDG:8DBEA:M;JC8I>DCH6G:C:8:HH6GNIDG:EG:H:CII=>H;JC8I>DC
'C8: 699>I>K: >C9:E:C9:C8: >H D7H:GK:9 I=: HIG6I:<N D; 9>K>9: 6C9 8DCFJ:G
8DJA97:I=DGDJ<=AN:MEADG:9L=:C6HH:HH>C<I=:%-;JC8I>DC ,=:G:;DG::68=
DC:9>B:CH>DC6A JI>A>IN ;JC8I>DC ;DG :68= G:HE:8I>K: 6IIG>7JI: H=DJA9 7: :A>8>I:9
AI:GC6I>K:AN>CHDB:86H:H6HE:8>;>86C6ANI>8;JC8I>DC8DJA97:JH:9L=:G:>IH
H=6E:<>K:H6C>CI:G:HI>C<9:H8G>EI>DCD;6HE:8>;>8>CHI6C8:D;I=:%PH7:=6K>DG
;DG6<>K:C6IIG>7JI:
,=: I=>G9 HI:E 8DCH>HIH 76H>86AAN D; :A>8>I>C< 8DC9>I>DC6A JI>A>IN ;JC8I>DCH ;DG
:68= 8G>I:G>DC ,=: 8DC9>I>DC6A JI>A>IN ;JC8I>DCH 86C 7: E:G;DGB:9 76H>86AAN 7N
9>G:8I6HH:HHB:CIDG:HI>B6I>DCD;I=:JI>A>IN;JC8I>DC !CHDB:86H:HI=:G:>H6C
6C6ANI>86A:MEG:HH>DCL=>8=>H@CDLCID7:6<DD9BD9:A;DGG:EG:H:CI>C<JI>A>IN
;JC8I>DCHD;HDB:HE:8>;>88G>I:G>6
H ID I=: ;DJGI= HI:E 6HH:HH>C< I=: H86A>C< 8DCHI6CIH 9:E:C9H B6>CAN DC I=:
H:8DC9HI:E !;699>I>K:>C9:E:C9:C8:>H8DC;>GB:9L:86CJH:I=:ADII:GNH=DLC
>C >< ID >9:CI>;N I=: K6AJ: D; I=: 8DCHI6CI >C6AAN I=: %- ;JC8I>DC
u(C(T),D(T))H=DJA97:B6M>B>O:9>CDG9:GID;>C9I=:7:HIDEI>DC;DG>CHE:8I>DC
I>B:>CI:GBHD;8DHI6C99DLCI>B:
=6EI:G :8>H>DC%6@>C<>CDC9>I>DC6H:9%6>CI:C6C8:
(C*(T), D*(T))
p
(C*(T), D0(T)) ~
1-p (C0(T), D0(T))
Fig. 6.5$DII:GNID;>C9I=:H86A>C<8DCHI6CIkc
!C >< I=: %-, 6EEA>86I>DC E=6H: >H HJBB6G>O:9 H=DL>C< E6GI D; I=:
BD9:A>C< EGD8:9JG: L=>8= >H 76H:9 DC I=: <:C:G6A EGD8:9JG: EGDEDH:9 ;DG
7J>A9>C<I=:%%
BD9:A6H<>K:C>C=6E
E:8JA>6G>IN>CI=>HHE:8>;>8
EGD8:9JG:>HI=:JH:D;I=:(6G:ID;GDCI>9:CI>;>86I>DC;DGI=:<:C:G6I>DCD;I=:H:I
D;6AI:GC6I>K:H
DM identification and
parameters estimation for C(T) and D(T)
Maximization of the
Multiattribute utility function u(C(T),D(T))
Fig. 6.6+IGJ8IJG:D;I=:9:8>H>DCBD9:A;DG>CHE:8I>DC>CI:GK6AHD;8DC9>I>DCBDC>IDG>C<
%6>CI:C6C8: B6C6<:B:CI >H 6 7JH>C:HH ;JC8I>DC I=6I 6>BH ID :CHJG: I=:
6K6>A67>A>IND;EGD9J8I>DCG:HDJG8:HID:C67A:I=:DE:G6I>DCD;6CDG<6C>O6I>DC !C
I=:8DCI:MID;:A:8IG>8EDL:G9>HIG>7JI>DC8DBE6C>:H(HI=:6K6>A67>A>IND;
:A:8IG>8 EDL:G >H :HH:CI>6A ;DG HD8>:IN DG I=>H G:6HDC <DK:GCB:CI G:<JA6IDGN
6.6 Building an MCDM/A Model on Condition Monitoring…
6<:C8>:H E:G;DGB 6C :HH:CI>6A GDA: >C 8DCIGDAA>C< I=: FJ6A>IN D; H:GK>8: D; I=:H:
8DBE6C>:H *:<JA6IDGN 6<:C8>:H EGDK>9: ;6KDG67A: 8DC9>I>DCH ;DG I=: :A:8IG>8>IN
B6G@:I ID 9:K:ADE >C 6 76A6C8:9 :CK>GDCB:CI 6BDC<HI 6<:CIH ;DG I=: 7:C:;>I D;
HD8>:IN !CI=>HH:8I>DC686H:HIJ9N>H86GG>:9DJIID:K6AJ6I:6C96EEANI=:E:G;DGB
6C8: D; 6 9:8>H>DC BD9:A 76H:9 DC 96I6 ;GDB 6 G6O>A>6C EDL:G 9>HIG>7JI>DC
8DBE6CN:GG:>G66C99:AB:>96
!CH:K:G6A8DJCIG>:H:A:8IG>8EDL:G9>HIG>7JI>DC8DBE6C>:H6G:>C6C:CK>GDCB:CI
8DCIGDAA:9 7N <DK:GCB:CI G:<JA6IDGN 6<:C8>:H .6G>DJH E:G;DGB6C8: >C9>86IDGH
L:G:9:K:ADE:9L>I=I=:6>BD;:CHJG>C<I=:FJ6A>IND;H:GK>8:L=>8=>HBDC>IDG:9
7N G:<JA6IDGN 6<:C8>:H ,=: EGD;>I67>A>IN D; 8DBE6C>:H >H 9>G:8IAN G:A6I:9 ID HJ8=
<D6AH ,LD G:A:K6CI 9>HIG>7JI>DC G:A>67>A>IN >C9>8:H I=6I B:6HJG: I=: 9JG6I>DC 6C9
;G:FJ:C8ND;I=:6K:G6<:>CI:GGJEI>DCD;6HNHI:B6G:@CDLC6H+NHI:BK:G6<:
!CI:GGJEI>DCG:FJ:C8N!C9:M+!!6C9+NHI:BK:G6<:!CI:GGJEI>DCJG6I>DC
!C9:M+!!ý:E>C
,=:G:;DG:6B:6HJG:I=6I86C7:JH:9ID6HH:HHI=:
CJB7:GD;8JHIDB:GH6;;:8I:97N6CDJI6<:>H9:G>K:9;GDBI=::ME:8I:9CJB7:G
D;;6>AJG:HNf(T)6+!!:FJ>K6A:CI:HI>B6I:;DG6C>CHE:8I>DCEDA>8N6H9:;>C:9
7N/6C<
>C
T
E3 N f T 4 ³ OF t dt
L=:G:
F(t)Q8JBJA6I>K:9>HIG>7JI>DC;JC8I>DC D;I=:9:A6NI>B:
6H:9 DC I=: 9:A6N I>B: 8DC8:EI I=: :FJ>K6A:CI :HI>B6I: I=6I G:EG:H:CIH I=:
SAIDI>HI=:9DLCI>B:D(T)<>K:C>C
d f E3 N f T 4 d s
D T
T ds
L=:G:
6>AJG:L>AA7:G:E6>G:9>BB:9>6I:AN6I6C6K:G6<:8DHIcf6C99DLCI>B:df
An inspection takes place every T time units, costs cs units and requires ds time
units, where ds << T.
,=:8DHID;6C>CHE:8I>DCEDA>8N86C7:9:I:GB>C:97NI=:9:A6NI>B:8DC8:EI
JH>C< =G>HI:G
c f E 3 N f T 4 c s
C T
T ds
/=:C B6@>C< 9:8>H>DCH 67DJI I=: ;G:FJ:C8N D; >CHE:8I>DCH >C 6C :A:8IG>8
:C:G<N HNHI:B H:K:G6A ;68IDGH H=DJA9 7: I6@:C >CID 688DJCI HJ8= 6H I=:
6K6>A67>A>IND;I=:HNHI:B6C9I=:CJB7:GD;>CI:GGJEI>DCH !CI=::A:8IG>86A:C:G<N
H:8IDG 8DCH:FJ:C8:H 6HHD8>6I:9 L>I= ;6>AJG: H=DJA9 7: 6KD>9:9 9J: ID I=: =><=
>BE68IDCI=:8DC8:HH>DCD;I=:H:GK>8:
=6EI:G :8>H>DC%6@>C<>CDC9>I>DC6H:9%6>CI:C6C8:
BJAI>8G>I:G>6 BD9:A 86C 7: JH:9 >C DG9:G ID 9:;>C: HIG6I:<>:H D; >CHE:8I>DC
>CI:GK6AH I=6I L>AA B::I I=G:: D7?:8I>K:H C6B:AN ID B>C>B>O: I=: CJB7:G D;
:ME:G>:C8:H68JHIDB:G=6HD;HJHI6>C:9>CI:GGJEI>DCDK:G6EG:9:;>C:9E:G>D9D;
I>B:I=:A:C<I=D;6C>CI:GGJEI>DC6C9I=:8DHIIDI=:HNHI:B
%-,#::C:N6C9*6>;;6L6H8=DH:CIDBD9:AI=:EGD7A:B ,=:B6>C
G:6HDC;DGI=>H8=D>8:>H76H:9DCI=:6HHJBEI>DCI=6II=:%PHG:6HDC>C<;DGI=>H
EGD7A:B 86C 7: G:EG:H:CI:9 7N I=: 6M>DB6I>8 HIGJ8IJG: D; I=>H I=:DGN !C I=>H
I=:DGN I=: 8DBE:CH6I>DC 7:IL::C I=: 8G>I:G>6 >BEA>:H I=: JH: D; 6 HNCI=:H>H
;JC8I>DC I=: <D6A D; L=>8= >H ID 6<<G:<6I: 6AA 8G>I:G>6 >C DC: 6C6ANI>8 ;JC8I>DC
,=:G:;DG: I=: %PH EG:;:G:C8: HIGJ8IJG: H=DJA9 7: 76H:9 DC I=: CDI>DC D;
8DBE:CH6I>DC
!I>H6HHJB:9I=6II=:G:>H6B6>CI:C6C8:B6C6<:GL=DH:G:HEDCH>7>A>IN>I>HID
9:I:GB>C: I=: >CHE:8I>DC >CI:GK6AH D; 8DC9>I>DC BDC>IDG>C< ,=JH I=: BD9:A
EGDEDH:9 I=GDJ<= %-, >H 9:K:ADE:9 >C DG9:G ID B::I =>H
=:G G:FJ>G:B:CIH HD
6H ID 6HH:HH I=: %- JC8I>DC !C I=>H 86H: I=: H:I D; 6AA I>B:HD;>CHE:8I>DC
6AI:GC6I>K:H >H 9:;>C:9 7N T 3 DG :68= >CHE:8I>DC I>B: I I=:G: >H 6
8DCH:FJ:C8: >C I:GBH D; 8DHI C(t) 9DLCI>B: D(t) 6C9 ME:8I:9 &JB7:G D;
;6>AJG:H Nf(t) DG :M6BEA: I=: ED>CI 3Ct1 Dt1 Nft14 7:ADC<H ID I=: 8DC
H:FJ:C8:HE68:
C :M6BEA: D; I=: 8DCH:FJ:C8: HE68: D; I=>H EGD7A:B >H H=DLC >C ><
,=: % L>H=:H ID B>C>B>O: I=: I=G:: 9>B:CH>DCH D; I=: 8DCH:FJ:C8: HE68:
H>BJAI6C:DJHAN ,=: EGD7A:B 8DCH>HIH D; 8=DDH>C< I=: 7:HI >CHE:8I>DC I>B: 7JI
I=:G:>H68DC;A>8I7:IL::CI=:I=G::9>B:CH>DCHD;I=:8DCH:FJ:C8:
D(T)
C(T)
Nf(T)
Fig. 6.7DCH:FJ:C8:HE68:;DGI=:I=G::8G>I:G>69DLCI>B:8DHI6C9CJB7:GD;;6>AJG:H
6.6 Building an MCDM/A Model on Condition Monitoring…
,=:8DCH:FJ:C8:HE68:9:;>C:9>C>< H=DLH6AAEDHH>7A:8DB7>C6I>DCHD;
I=:I=G::8DCH:FJ:C8:H7JI<:C:G6AANI=>HHE68:EG:H:CIH6ADID;JC;:6H>7A:ED>CIH
HJ8= 6H I=: DEI>B6A ED>CI >C :68= 9>B:CH>DC I=: ED>CI C*(T), D*(T), Nf*(t)). !;
I=>H ED>CI >H ;:6H>7A: I=:C >I >H CDI C:8:HH6GN ID BD9:A I=: EGD7A:B JH>C< 6
BJAI>8G>I:G>66EEGD68=7:86JH:I=>HED>CI9DB>C6I:H6AADI=:G6AI:GC6I>K:H6C9HD>I
H=DJA97:8=DH:C
(G:;:G:CI>6AJI>A>IN6C9699>I>K:>C9:E:C9:C8:H6G:I=G::>BEDGI6CI8DC8:EIHID
7: :MEADG:9 >C I=: JH: D; %-, -I>A>IN >C9:E:C9:C8: >H 6 8DC8:EI >C %-,
:FJ>K6A:CIIDI=6ID;EGD767>A>HI>8>C9:E:C9:C8:>CBJAI>K6G>6I:EGD767>A>INI=:DGN
-I>A>IN>C9:E:C9:C8:8DC9>I>DCH>BEANI=6II=:%-;JC8I>DCBJHI7:D;6HE:8>;>:9
;DGB !C<:C:G6A>C9:E:C9:C8:6HHJBEI>DCH<G:6IANH>BEA>;NI=:6HH:HHB:CID;I=:
DG><>C6AJI>A>IN;JC8I>DC ,=:I=G::6IIG>7JI:H6G:699>I>K:>C9:E:C9:CI>;I=:E6>G:9
EG:;:G:C8: 8DBE6G>HDC D; 6CN ILD ADII:G>:H 9:;>C:9 7N ILD ?D>CI EGD767>A>IN
9>HIG>7JI>DCH 9:E:C9H DCAN DC I=:>G B6G<>C6A EGD767>A>IN 9>HIG>7JI>DCH #::C:N
6C9*6>;;6
6H:9DCI=:699>I>K:JI>A>IN8DC8:EI>I86C7:8DC8AJ9:9I=6II=:6IIG>7JI:H6G:
699>I>K:>C9:E:C9:CI>;6C9DCAN>;I=:I=G::6IIG>7JI:JI>A>IN;JC8I>DC>H699>I>K:
DGI=:H:8G>I:G>6I=:699>I>K:;DGBB6N7:9:;>C:96H
u C T D T Nf T k c uc C T k d ud D T k n un Nf T
L=:G:
kcQH86A:8DCHI6CI;DGI=:8DHI8G>I:G>DC
uc(C(T))Q8DC9>I>DC6AJI>A>IN;JC8I>DC;DGI=:8DHI8G>I:G>DC
kdQH86A:8DCHI6CI;DGI=:9DLCI>B:8G>I:G>DC
ud(D(T))Q8DC9>I>DC6AJI>A>IN;JC8I>DC;DGI=:9DLCI>B:8G>I:G>DC
knQH86A:8DCHI6CI;DGI=:CJB7:GD;;6>AJG:H8G>I:G>DC
unNf,Q8DC9>I>DC6AJI>A>IN;JC8I>DC;DGI=:CJB7:GD;;6>AJG:H8G>I:G>DC
,=>H 6EEA>86I>DC >H 76H:9 DC 6 8DC;>9:CI>6A 86H: HIJ9N >C 6C :A:8IG>8 EDL:G
9>HIG>7JI>DC 8DBE6CN AI=DJ<= I=: ;><JG:H 6C9 DI=:G 6HE:8IH D; I=>H 6EEA>86I>DC
6G:CDII=:G:6A96I6I=:N=6K:7::C6EEGDEG>6I:AN6AI:G:9>CDG9:GIDG:EG:H:CI6
G:6A>HI>86C98DCH>HI:CI8DCI:MI
,=: D7?:8I>K: D; I=: 8DBE6CN >H ID B>C>B>O: I=: 8DHI 9DLCI>B: 6C9 I=:
:ME:8I:9 CJB7:G D; >CI:GGJEI>DCH D; I=: HNHI:B 7GDJ<=I 67DJI 7N 6C >CHE:8I>DC
EDA>8N 6H:9DCI=:BD9:AEGDEDH:9>CI=:EG:K>DJHH:8I>DCI=:;>GHIHI:EIDL6G9H
6EEAN>C<I=:BD9:A>HID>9:CI>;NI=:% B6>CI:C6C8:B6C6<:GG:HEDCH>7A:;DG
B6@>C<HJ8=9:8>H>DCH>CI=:8DBE6CNL6H>9:CI>;>:9 ,=:C:8:HH6GNE6G6B:I:GH
D;I=:BD9:AL:G::HI>B6I:96C96G:>AAJHIG6I:9>C,67A:
=6EI:G :8>H>DC%6@>C<>CDC9>I>DC6H:9%6>CI:C6C8:
Table 6.1(6G6B:I:GHD;I=:BD9:A
(6G6B:I:G .6AJ:
O ;6JAIHE:G6N
h /:>7JAA
ds 96NH
db
96NH
Cb -+
Ci -+
H6G:HJAID;I=:%-;JC8I>DCI=:B6M>BJBJI>A>IND;I=:>CHE:8I>DCI>B:>H
;DG96NH !C>< I=:JI>A>IND;I=:>CHE:8I>DC>CI:GK6AH>HH=DLC
,=>HH:8I>DCEG:H:CI:96%-,IDHJEEDGII=:EA6CC>C<D;6C>CHE:8I>DCEDA>8N
>C 6C :A:8IG>8 EDL:G 9>HIG>7JI>DC 8DBE6CN ,=: CJB7:G D; I>B:H 6 8JHIDB:G
:ME:G>:C8:96HJHI6>C:9>CI:GGJEI>DCDK:G6EG:9:;>C:9E:G>D9D;I>B:I=:A:C<I=
D;>CI:GGJEI>DC6C98DHID;HNHI:B6G:I=G::D7?:8I>K:H8DCH>9:G:9 ,=:8DC8:EID;
9:A6NI>B:L6HJH:9IDBD9:AI=:;6>AJG:EGD8:HH
Fig. 6.8%-;JC8I>DC
References 247
MAUT was chosen to model a DM’s preferences for the cost, SAIDI and
SAIFI criteria in accordance with regulatory laws of this sector and in a suitable
way to deal with the tradeoff of probabilistic consequences. This model was
evaluated and validated by managers from a Brazilian company.
The modeling of predictive maintenance and monitoring is a tool that can
provide many benefits to the area of maintenance management. This chapter
suggests a multicriteria approach for modeling CBM decisions. Thus, the pro-
posed multicriteria model aimed to answer this need based on the MAUT which
has an axiomatic structure and allows to deal with the conflict between the expected
and the cost of an inspection policy downtime.
References
Barlow RE, Proschan F (1965) Mathematical theory of reliability. John Wiley & Sons, New York
Ben-Daya M, Duffuaa S, Raouf A (eds) (2000) Maintenance, Modeling, and Optimization.
Kluwer Academic Publishers, Norwell
Berrade MD, Cavalcante CAV, Scarf PA (2012) Maintenance scheduling of a protection system
subject to imperfect inspection and replacement. Eur J Oper Res 218:716–725
Carnero MC (2006) An evaluation system of the setting up of predictive maintenance
programmes. Reliab Eng Syst Saf 91:945–963
Čepin M (2011) Assessment of Power System Reliability: Methods and Applications. Springer
London
Chelbi A, Ait-Kadi D (2009) Inspection Strategies for Randomly Failing Systems. In: Ben-Daya
M, Duffuaa SO, Raouf A, et al. (eds) Handb. Maint. Manag. Eng. SE - 13. Springer London,
pp 303–335
Chiu SY, Cox LA Jr, Sun X (1999) Optimal sequential inspections of reliability systems subject
to parallel-chain precedence constraints. Discret Appl Math 96 - 97:327–336
Christer AH (1999) Developments in delay time analysis for modelling plant maintenance.
J Oper Res Soc 50:1120–1137
Christer AH, Waller WM (1984) Delay time models of industrial inspection maintenance
problems. J Oper Res Soc 35:401–406
de Almeida AT, Ferreira RJP, Cavalcante CAV (2015) A review of multicriteria and multi-
objective models in maintenance and reliability problems. IMA Journal of Management
Mathematics 26(3):249–271
Do Van P, Bérenguer C (2012) Condition-Based Maintenance with Imperfect Preventive Repairs
for a Deteriorating Production System. Qual Reliab Eng Int 28(6):624–633
Ferreira RJP, de Almeida AT (2014) Multicriteria model of inspection in a power distribution
company. Reliab Maintainab Symp (RAMS), 2014 Annu 1–5
Ferreira RJP, de Almeida AT, Cavalcante CAV (2009) A multi-criteria decision model to
determine inspection intervals of condition monitoring based on delay time analysis. Reliab
Eng Syst Saf 94:905–912
Fouladirad M, Grall A (2014) On-line change detection and condition-based maintenance for
systems with unknown deterioration parameters. IMA J Manag Math 25(2):139–158
Grall A, Bérenguer C, Dieulle L (2002) A condition-based maintenance policy for stochastically
deteriorating systems. Reliab Eng Syst Saf 76(2):167–180
Huynh KT, Castro IT, Barros A, Bérenguer C (2012) Modeling age-based maintenance strategies
with minimal repairs for systems subject to competing failure modes due to degradation and
shocks Eur J Oper Res 218(1):140-151
=6EI:G :8>H>DC%6@>C<>CDC9>I>DC6H:9%6>CI:C6C8:
7.1 Introduction
Ever since management theory took shape, there has been extensive discussion
with regard to downsizing, core competences, business process re-engineering and
other managerial trends, that are deployed into general outsourcing. Such discussion
is also applicable to maintenance. According to Buck-Lew (1992), a company
outsource when it requests the services of an outside party to fulfill a function or
functions in the organization. Decisions on outsourcing are very close related to
contract selection (de Almeida 2001b), contract design and supplier selection
decisions.
The strategy for outsourcing maintenance goes together with the management of
contracts. The relationship between the contractor (the company outsourcing one
or more of its services – client company) and contracted firms (suppliers of such
services) is regulated by a contract in which the parties involved define the rules
of the service to be performed for an agreed length of time.
Service contracts in the area of maintenance typically emphasize the legal
aspects in clauses (terms) that deal with price, forms of readjusting price, payment
terms, quality and warrant provisions of the service to be provided, technical aspects,
transfers of responsibilities to third parties, retention/fines/damages, termination,
period (deadline), exchanges of information (communication channel) and other
important aspects of this relationship.
According to Brito et al. (2010), selecting contracts is a very important stage in
the process of outsourcing maintenance given the current trend towards reducing
costs and increasing competitiveness by focusing on core competences. Many
studies have been carried out on outsourcing and maintenance contracts, most of
which deal with qualitative aspects (Kennedy 1993; de Almeida 2005). Thus,
MCDM/A, plays an important role supporting DMs to deal with multiple and
conflicting criteria, and associated uncertainties in the process for selecting out-
sourcing contracts (Brito et al. 2010).
Wideman (1992) suggests that when companies consider outsourcing, they
need to make prior enquiries about bidding companies at the start of the hiring
process. According to Martin (1997), maintenance contractors are very interested in
252 Chapter 7 Decision on Maintenance Outsourcing
developing new types of contracts that promise to offer them higher profitability
and increased flexibility and lower maintenance costs to them.
From a historical standpoint, the initial practice of the industrial sector was to
hire maintenance services in the form of manpower, i.e., paid for in terms of man-
hours worked. In this type of contract, it is the sole responsibility of the
maintenance service providers to ensure the presence of their staff in the industrial
plants of their customers, and therefore the suppliers is paid for the total number
of hours their staff worked. Main weakness of manpower contract are: less-skilled
personnel; low productivity of services; low quality of services; higher accident
rates; noncompliance with labor legislation.
Although still widely applied, this type of contract does not require the
commitment of outsourced staff to produce good results and, invariably, the
consequences for the industry may be negative in the medium and long term.
Therefore, this type of contract results in a relatively high business risk and
should not be entered into if the reason is the company’s vision is one of global
optimization. This is because although there may be an apparent reduction in the
cost of maintenance, undesired effects on the overall results can be generated.
This type of contract is practically a unilateral relationship. From the
perspective of game theory, one can assume that the policy contract is “win-lose”
in the short term, but in reality, in the medium or long term it can become a policy
of “lose-lose”. Therefore, sometimes this model proves to be bad for both the con-
tractor and the supplier.
Due to the problems discussed above, industry developed a different type of
contract: hiring for specific maintenance jobs or for special maintenance servicing
of specified equipment and machinery. This type of contract occurs in an isolated
form or as part of a hybrid contract (more than one contract type), the latter being
widely used in the industrial sector. Some advantages of this type of contract are:
better-qualified manpower; increased productivity; better quality of work.
The process of outsourcing maintenance activities evolved into hiring a single
supplier or a few suppliers, who are highly specialized and qualified and are made
responsible for the overall maintenance process. At this stage, the relationships
that have been established by the partnership between companies and their sub-
contractors in the maintenance area mature.
In this type of contract, contractors must give support to the activities out-
sourced, and make the staff of the contracted company feel a bond with the
contracting company and as if they were an integral part of a single organism,
which for its best performance needs to keep its basic functions operating in a
healthy way. Achieving this maximum mutual commitment remains the greatest
challenge for obtaining the best results in the process of maintenance outsourcing.
Another type of contract, with emphasis on both the client and the supplier, is
the type of contract that includes tracking results for performance. Typically, this
type of contract involves greater commitment from both sides of the contract, and
formalizes partnerships in the medium and long term (Wideman 1992).
7.2 Selection of Outsourcing Requirements and Contract Parameters 253
The relationships of these contract types to each other and issues of contract
complexity, client-contractor relationship and client maintenance knowledge base
are shown in Fig. 7.1. The three types of contract discussed in Martin (1997) are
extreme cases. However, contractors may develop different (hybrid) contracts for
different sets of production systems, the skills involved and to split the financial
risk between the client and supplier.
The potential impact of maintenance on equipment and systems in terms of
quality, flexibility, cost, availability, and safety is increasingly evident within the
maintenance management system. Therefore, the need for measuring the performance
of maintenance is evident, and this means that maintenance as a function generates
profitability for the firm.
Therefore, what is critical is the process of setting performance indicators,
which will serve as regulatory elements of quality, variable remuneration (depending
on the chosen type of contract) or other indicators.
Fig. 7.1 Relationships of the various contract types and contract complexity, client-contractor
relationship and client’s knowledge of maintenance
example, both parties to the contract must negotiate on the implications of having
ranges for performance indicators. For example, they must agree on from what
point in the range of reduced costs (cost of contract) that a high handling time
(availability of resources for the outsourced activities) and from what point in the
range that the high cost of the contract will result in it taking a short time to
complete the outsourced service.
Within several contract templates, one can identify the type of partnership
agreement based on indicators of availability and of the reliability of the pro-
duction system (using the Mean Time Between Failure – MTBF - and Mean Time
to Repair – MTTR - indicators), where the company to which services have been
outsourced increases its profitability as it improves the availability and reliability
of the client enterprise system (de Almeida and Souza 2001). Therefore, this type
of contract no longer remunerates services (grants bonuses to), but rather solutions
that will improve the levels of availability and reliability of systems.
However, some factors can disturb this type of contract. One that stands out is
the alignment amongst the strategic objectives of the interested parties. In fact,
there is a conflict of interest because the company to which services have been
outsourced also has difficulties of surviving in the competitive market, and it
needs to be competitive. For outsourced companies this type of service is a core
activity, while for contracting companies, it is a means of supporting an end
activity, so the maintenance services performed by subcontractors is the only
source of funds. This conflict is quite evident when contracts are mainly short and
medium term in length.
Therefore, assuming that the aspects of reliability were properly dealt with in
the phase of the designing the production system (i.e., both parties are aware that
the maintenance service will not have the ability to improve system reliability be-
yond that already specified in the design), it would remain for a maintainability
study to be included in maintenance agreements (de Almeida 2002). In this regard,
one has the administrative time (TD), the effective time to repair (TTR), the
availability of spare parts and the level of training of the outsourced teams.
However, all these previous aspects have a cost (C) associated with obtaining the
levels desired.
TD is the time that it takes to notify a maintenance company of a failure and the
time it takes this company to go to the client to deal with it. TD basically consists
of: the time spent in selecting and making the technical staff ready to perform the
service; the time taken to provide tools and the budget necessary to perform the
service; and the commuting time between the service provider company and the
location of the system to be repaired.
Therefore, TD can be a negotiated in a contract because it directly affects the
interruption time (TI) of the client system. As a counterpoint to this, it has to be
remembered that since outsourced firms have many clients, they try to keep the
idle time of their work teams within certain levels in order to meet the demands of
their diverse clients and to satisfy the times of visits agreed to by contract.
256 Chapter 7 Decision on Maintenance Outsourcing
The time taken for the maintenance team from the start of the repair process to
putting the production system back into a normal state of operation is the TTR .
Normally, this time is directly related to the technical skills of the team, team
training, the team’s learning curve, the modularity of the system/equipment, the
availability of repair spare parts and other variables.
Therefore, when the contract is modeled as a function of the TD and TTR, the
maintenance contract must adequately compensate for the cost of the maintenance
structure of the company providing services that ought to be in a state of readiness
to meet sudden demands from the client company. However, keeping a large
contingent of maintenance staff available to the client and a high level of inventory
of spare parts, so as to guarantee an adequate level of system availability, becomes
very and does not fit the competitive market model that both the client and the
supplier find themselves in.
In reality, what is required is that the company providing maintenance services
has a firm commitment to ensuring the availability of the system (hence the need
for the outsourced team to be available at short notice) and not increasing the cost
of service. Thus, the best choice would to make a contract using a decision model
that incorporates the DM’s preference structure represented by the utility function
of the attributes cost C , of the interruption time (TI=TD+TTR) and of the
maintainability of the system as modeled in probabilistic terms.
Thus, the problem faced by the manager or DM becomes how to proceed with a
decision process that allows the various performance indices considered in the
contract to be optimized, since these various indices can conflict with each other.
Brito et al. (2010) state that contracts that present a lower cost might present a less
satisfactory performance concerning criteria related to quality and availability,
which creates a complex frame of trade-offs.
The DM faces several options for maintenance contracts, each implying
different system performances and related costs. de Almeida (2001a) identifies
that the selection of repair contracts is a non-trivial process since the consequences
of a wrong choice may be critical, for instance, in services where availability is
fundamental, as in telecommunications and electric power distribution services.
With regard to selecting contracts, little work has been conducted on exploring
a multi-criteria decision-making approach. de Almeida (2001b) has presented
MCDM/A models based on MAUT for selecting repair contracts, which aggregate
interruption time and related cost through an additive utility function. A different
approach can be found in de Almeida (2002), where the ELECTRE I method has
been combined with utility functions regarding a repair contract problem.
Brito and de Almeida (2007) and Brito et al. (2010) propose a MCDM/A
methodology to support the selection of maintenance contracts in a context where
in-formation is imprecise, when DMs are not able to assign precise values to the
importance parameters of criteria used for contract selection. Utility theory is
combined with the Variable Interdependent Parameters method (VIP) to evaluate
alternatives using an additive value function regarding interruption time, contract
cost and maintenance service supplier’s dependability.
7.3 MCDM/A Maintenance Service Supplier Selection 257
There are some decision models and applications in the literature that consider
MCDM/A techniques which will be discussed in this section.
x There is prior knowledge S(u) about u that can be assessed from experts.
x TD is deterministic and assumes different values according to the service
supplier and contract.
x DM’s preference structure fits MAUT axiomatic requirements to be represented as
an additive utility function U(TI,C), given by (7.2), where kTI and kC are the
respective scale constants:
7.3 MCDM/A Maintenance Service Supplier Selection 259
U C (C ) e A2C (7.4)
k TI u
U (u , a i ) k CU C (C ) (7.8)
A1 u
ª k TI u º
EuU (u , ai ) ³«A u k e C
A2 C
»S i (u )du (7.9)
u
¬« 1 ¼»
Thus, for each distribution of TTR there will be an implied cost for the
respective service supplier contract, which means that the DM is deciding upon
the TTR pdf and its respective cost (C) in order to maximize his/her multi attribute
utility function. The alternatives for this problem are the existing combination of
maintenance service suppliers and its contract. Solving this problem consists
in solving (7.9) for all alternatives, which are all the existing combination of
maintenance service suppliers and its contract. Therefore, kTI and kC represents the
tradeoff between cost and time to repair according to DM’s preferences.
Thus, from (7.1) and (7.10), follows that (7.12) results in (7.13), considering
that this result would be positive if, and only if TD t 0 and TI t TD , thus
TI t TD t 0 , from this result the integer from (7.12) turns into (7.13):
f
f (TI ) ³Ye ue
YTD u ( TI TD )
dTD (7.12)
f
TI
Yu
f (TI ) (e YTI
e uTI
) (7.14)
u Y
Yu
U (u , a ) ³ >k e k U (C )@ (e e )dTI (7.16)
A1TI YTI uTI
i TI C C
TI u Y
262 Chapter 7 Decision on Maintenance Outsourcing
By developing (7.16) into (7.17), and then replacing (7.14) in (7.17), (7.18) is
obtained, and developed into (7.19), (7.20) and (7.21):
Yu
U (u , a ) i ³ e (e e
Y
k TI
A1TI TI uTI
)dTI
u Y TI
(7.17)
Yu
k U (C ) ³ (e Y e )dTI TI uTI
u Y
C C
TI
Yu f f
U (u , a ) i
k TI ®³ e
( A1 Y ) TI
dTI ³ e ( A1 u ) TI
dTI ½¾
u Y ¯ 0 0 ¿ (7.18)
f
k U (C ) ³ F (TI )dTI
C C
0
Yu f f
e ( A1 Y ) TI dTI e ( A1 u ) TI dTI ½
U (u , a i ) k TI ®³ ³0 ¾
u Y ¯0 ¿ (7.19)
k CU C (C )
f f
Yu ° e Y ( A1
e ) TI ( A1 u ) TI ½°
U (u , a ) k ® ¾
u Y °̄ ( A Y ) ( A u) (7.20)
i TI
1 0 1 0 °¿
k U (C )
C C
Yu 1 1 ½
U (u , a i ) k TI ® ¾ k CU C (C ) (7.21)
u Y (
¯ 1 A Y ) ( A1
u) ¿
Yu
U (u , ai ) k TI k e A2 Ci
(7.22)
( A Y )( A u )
C
1 1
Applying (7.22) in (7.5) gives the expression of EuU(u,ai) for the stochastic TD
model as given in (7.23):
ª Yu º
EuU (u , ai ) ³ «k ( A Y )( A u ) k e
TI C
A2 Ci
»S (u )du (7.23)
u
¬ 1 1 ¼
contract. Thus, this decision model uses one-dimension utility functions values as
the performance of alternatives for each criterion.
This decision model was built for a repairable system considering the
implications of each alternative in terms of two aspects: Interruption time (or
response time) and Costs.
Similarly to the decision model previously presented, this decision model
evaluates the benefit of the maintenance service supplier contract in terms of
maintainability and the associated cost of the service.
The maintenance service supplier contract performance in the response time
reflects its specific condition for spare provisioning and repair capability.
The assumptions of the decision model are (de Almeida 2002):
x TI is explained by TD and TTR; where TI=TD+TTR.
x TTR follows an exponential distribution for all service suppliers in all contracts,
given by (7.1).
x There is prior knowledge S(u) about u that can be assessed from experts.
x TD follows an exponential distribution for all service suppliers in all contracts,
given by (7.10).
x TD and TTR are independent random variables.
x DM’s preference structure fits a non compensatory rationality and requires an
MCDM/A approach compatible with outranking relation preferences according
to Chap. 2.
x DM’s preference intra criterion preference structure fits for both attributes an
exponential utility function for representing DM’s one-dimensional preferences
as in the deterministic administrative time model (de Almeida 2001a) and in
the stochastic administrative time model (de Almeida 2001b), these functions
are given by (7.3) and (7.4).
Since this model considers TTR and TD as two independent random variables,
given by (7.1) and (7.10) respectively, TI is given by (7.14).
Despite the fact of this decision model is not considering tradeoffs, the DM
behavior facing subjected to uncertainties is being modeled in the intra criterion
valuation by the utility functions given by (7.3) and (7.4).
Due to the assumptions of this particular decision model, the costs are not
affected by the state of nature, therefore each maintenance service supplier
contract has its particular cost definition not affected by uncertainties, thus the
cost criterion is evaluated directly by (7.4) according to each alternative’s cost (de
Almeida 2001b).
In the other hand, the response time represented by TI cannot be evaluated
directly for each alternative as the cost, due to the interference of state of the
nature uncertainties over its consequences.
For dealing with this situation, de Almeida (2002) considered the parameter Z,
related to TD. Applying the utility function linearity property as in the previous
models, UTI(Z) is given by (7.24).
7.3 MCDM/A Maintenance Service Supplier Selection 265
uY
U (Y )
TI
(7.25)
( A Y )( A u )
1 1
Based in the intra criterion alternatives evaluation given by (7.4) and (7.25), the
ELECTRE I method builds outranking relations based in concordance index
C(a,b) and in a discordance index D(a,b).
The concordance index is given by (7.26), and measures the relative advantage
of each alternative a compared with an alternative b (Vincke 1992).
C ( a, b)
¦ (W 0.5W )
, (7.26)
¦ (W 0.5W W )
ª ( Z Z ak ) º
D ( a, b) max « bk* », (7.27)
¬ (Z k Z k ) ¼
Thus, the model considers (de Almeida 2005) the following assumptions:
x TI is explained only by TTR, following the maintainability approach given by
Goldman and Slattery (1977);
x TTR follows an exponential distribution for all service suppliers in all contracts,
given by (7.1);
x Although there is prior knowledge S(u) about u that can be assessed from
experts, it is assumed that there is an uncertainty about the real value of ui,
withregard to the respective contract i;
x Based on the last assumption, di is defined as the probability that ui t uie for
action ai, therefore uie is the value committed by contract for ui . Therefore, di
is given by (7.28):
di ³ S (u )du
i i
(7.28)
uie
U d (d i ) B3 C 3 ln( A3 d i ) (7.29)
U q (q i ) B4 C 4 ln( A4 q i ) (7.30)
From the assumptions of this decision model costs are also not affected by the
state of nature as in the previous section. Therefore the cost criterion is evaluated
directly by (7.4) according to each alternative’s cost (de Almeida 2005). Same
applies to dependability and service quality, evaluated respectively by (7.29) and
(7.30). From the assumption of the prior knowledge over TTR , the state of nature
must be considered for evaluating the consequences on repair time by considering
the parameter ui instead of TTR. Similarly to the previous, the decision model
proposed by de Almeida (2005) uses the linearity property to obtain UTI(ui) from
(7.31).
Since Pr(TTR | ui ) is given by (7.1), then, applying (7.1) and (7.3) to (7.31),
(7.32) is achieved.
ui
U TI (u i ) (7.32)
( A1 u i )
In some situations, the DM may not feel comfortable about setting precise values
for the decision model parameters, and thus an approach suitable for dealing with
this situation should be used.
The decision models presented by Brito and de Almeida (2007) and Brito et al.
(2010) for selecting maintenance service supplier contracts addressed such a
particular situation, using an approach that enabled a recommendation to be made
based on imprecise statements with regard to the decision maker’s preferences and
this was supported by an elicitation procedure.
Many DMs have difficulties to fix constant values for criteria ‘weights’ that
must represent not only the importance of the criteria but also the compensation
rates between criteria in additive value functions.
There may be several reasons for a decision maker to avoid precise statements.
One of these may be that he is unsure if a parameter should be 0.75 or 0.7. Thus if
a range of values can be used for such parameters, the decision maker can give
more confident statements regarding the decision problem.
Brito and de Almeida (2007) considered three basic criteria: interruption time,
applicant’s dependability and contract cost in this model.
The dependability criterion is used to assess alternatives of contract in relation
to “deadlines” being met. It is a measure related to keeping delivery promises,
which it is represented by the probability of the company selected achieving the
time to repair under a specified probability distribution, as set out in the contract
proposal of the maintenance service supplier, similarly as in the model presented
in last section.
7.3 MCDM/A Maintenance Service Supplier Selection 269
To Brito and de Almeida (2007), these three criteria may be conflictive among
alternatives. Usually, lower interruption times (times to repair) are related to better
resource conditions, better spares provisioning and higher professional skills, and
they often imply higher costs.
Besides, the dependability of the alternative is not directly related to the
proposal conditions associated with interruption time, but it is assessed by
the contracting company taking into consideration other aspects such as the
applicant’s reputation, previous services, the structure of repair facilities, etc.
The approach used in the decision model by Brito et al. (2010) considers utility
functions aggregated by variable interdependent parameters for an additive
function and uses the following criteria for evaluation:
x Mean time to repair (MTTR).
x Service supplier cost.
x Geographical spread of the service supplier network.
x Service supplier reputation.
x Compatibility of company cultures.
The specific problem considered by Brito et al. (2010) was related to power
distribution services, which may also be extended to the telecommunications
context.
The service supplier’s performance on MTTR indicates its structure and
capabilities, thereby reflecting its maintenance staff’s skills, transportation
resources, facilities and spares inventory.
The geographical spread of the service supplier reveals its logistical network
structure, and relates to the number and spread of local branch offices, which
gives flexibility and speed with regard to performing repairs. This is an important
point for companies with widespread local branch offices, and it is directly related
to the speed of service response and flexibility offered to the contracting
organization or its several units
The service supplier’s reputation is another important factor to be considered
since this may avoid bad experiences from past services or even service level
inconsistencies being repeated during the time span of the contract. Evaluating the
service supplier in this respect may be from external sources, such as other
companies that had previous experiences with the service supplier, verifying if
payment of taxes to the government is up-to-date and possession of the due
certifications in quality and/or safety norms.
Cultural compatibility is an issue that has become more and more relevant,
since many organizations are seeking to establish long-term relations by building
strategic partnerships. Allied to such strategic factors, many companies have
added undertaking social and environmental responsibility activities to their
organizational objectives. This includes their seeking sustainability and requiring
this commitment also from their partners and suppliers.
By using variable interdependent parameters, Brito et al. (2010) considered the
range for each parameter, assessing a lower and an upper bound. Another kind of
270 Chapter 7 Decision on Maintenance Outsourcing
imprecise information was the order (ranking) of the parameters. The assessment
of these imprecise statements given by decision maker enabled dominance
relations among the service suppliers to be established, based on the decision
maker’s assessed preferences.
According to Brito et al. (2010), in order to assess the performances of
alternatives of contracts for the first two criteria, since MTTR and contract cost
can be directly represented by values, utility values should be elicited using the
due procedures. However, the last three criteria present a less objective feature; in
this case, each candidate may be evaluated after completion of a questionnaire,
which is constructed so as to obtain all the information required by the contracting
organization in order to assess the candidates on each of the three criteria.
The problem of supplier selection has been studied in many contexts, rather than
the RRM context. Also, studies have been found in a broader way, therefore
MCDM/A and other approaches to supplier evaluation and selection problem have
been widely studied.
There are various decision making approaches proposed in the literature. Ho
et al. (2010) presented a literature review on this topic emphasizing which
approaches were frequently applied, which criteria were most considered, and
investigated inadequacy with regard to the studies of the approaches found in the
literature within their review of articles published in international journals from
2000 to 2008.
Among other approaches widely used for supplier evaluation and selection, Ho
et al. (2010) highlight the use of: Analytic hierarchy process (AHP), Analytic
network process (ANP), Case-based reasoning (CBR), Data envelopment analysis
(DEA), Fuzzy set theory, Genetic algorithm (GA), Mathematical programming,
Simple multi-attribute rating technique (SMART), and hybrid approaches.
Ho et al. (2010) evaluated MCDM/A approaches versus traditional cost based
approaches. The advantages of applying MCDM/A approaches enable con-
sideration to be given to important and relevant factors for the decision process
other than cost.
Another recent literature review on supplier selection was presented by Chai et
al. (2013), considering articles published in journals from 2008 to 2012 that
presented applications of decision making techniques for supplier selection.
From the literature review conducted by Chai et al. (2013) many decision
making approaches have been applied to these problems recently. Chai et al.
(2013) identified twenty six decision making techniques applied for supplier
evaluation and selection, and grouped these techniques into three categories:
MCDM/A, mathematical programming and artificial intelligence techniques.
References 271
References
8.1 Introduction
parts is also negative, as this may well result in a loss of production due to
increased downtime of equipment while awaiting delivery of the spare parts
needed. Therefore, sizing the number of spare parts that optimally need to be held
strongly influences a company´s costs and profitability. Consequently, the
management of this resource is one of the most critical tasks in maintenance
management (de Almeida and Souza 2001). This topic is relevant for many
contexts, may they be related to individual plants, such as refinery (Porras and
Dekker 2008), or to a logistic network (Syntetos et al. 2009).
When comparing with other types of inventory models, such as raw material
for manufacturing processes, sizing and managing spare parts inventory is a far
more complex task, considering that manufacturing inputs are usually is easier to
forecast its demand, especially if comparing its turnover. Production inventories
usually follows market rules, but spare parts are required based on failure rates
and the system reliability design.
Thus, spare parts are sized according to its relative importance according to the
system reliability. A bad decision on spare part sizing may lead to high losses,
compromising company’s profitability as well as its system availability.
According to British Standards 3843-1:1992, Terotechnology is the study that
allows the maintenance of assets in optimal manner by the combination of
management, financial, engineering and other practices applied to physical assets
such as equipment considering its life cycle costs.
The number of spare parts sized must consider time to repair or time of service
disruption. This decision must assure that parts required will be available when
requested. Thus, the spare parts sizing problem has conflicting goals that are
the increase of spare parts available as a contribution to increase equipment
availability by the reduction of service disruption time. As in the other hand, the
goal to reduce the inventory and purchase costs of spare parts.
Considering that manufacturing inventory model are not suitable to manage
spare parts inventory due to the differences in the items demand for both cases,
spare parts inventory considers that the demand for each item follows a stochastic
process, represented by the random variable of equipment failure.
According to Marseguerra et al. (2005), to avoid risks to plant and costly plant
being unavailable due to a shortage of spares parts, the latter are often over-
stocked, thus leading to huge losses due to having invested unnecessarily in an
excessive number of them or to too many of them becoming obsolete.
For Roda et al. (2014), spare parts management plays a relevant role for
equipment-intensive companies. They review the type of criteria applied to spare
parts classification. An important step of such a process is that of classifying spare
parts (criticality) with a view to enabling different items to be properly managed
by taking into account their peculiarities. Many advantages can be achieved by
proper classification, e.g. an organization may align its policy for the stock
management system with the criticality of the need for holding spare parts
(Macchi et al. 2011); demand forecasting may be driven by data collected on the
parts for different classes while improving the performance of equipment and the
8.1 Introduction 275
system overall may concern critical classes, thus making the work of the analyst
easier by allowing him/her to concentrate on tests of inventory control policies
currently in force. Forecasting spare parts demand is an important issue (Boylan
and Syntetos 2010) for building related decision models.
Some studies have addressed the problem of determining the optimal spare
parts inventory, such as by using gradient methods, dynamic programming,
integer programming, mixed integer and nonlinear programming. Unfortunately,
as mentioned by Marseguerra et al. (2005), such optimization techniques typically
entail the use of simplified plants ‘or systems’ models about which predictions
may be of questionable realism and reliability.
In general, spare parts management has at least two main objectives that are
conflicting: to contribute to increased system availability by acquiring and
stocking spare parts; i.e., ensuring the supply of spare parts in the proper amount
to reduce interruption times and; to reduce the cost of buying and stocking spare
parts (de Almeida 2001).
A decision model on provisions for spares assumes that at least one spare item
is held in stock (de Almeida 2001; de Almeida 1996). Normally, when a failure
occurs, in due time, the failed item is replaced by a spare part, which should be
available in the depot. The faulty item is sent out for repair and after being
repaired (good-as-new), it is shipped back to the plant depot where it serves as a
spare. This decision problem uses an MCDM/A model in order to define how
many additional spares should be provided in accordance with what criteria. de
Almeida (2001) applies a Bayesian approach, based on prior probability
distributions. Aronis et al. (2004) also uses prior distributions of the failure rates
to forecast demand.
Some techniques for planning and inventory control were developed for the
context of manufacturing systems (goods producing systems), and were later
extended to the service producing systems. An example of this is Just-In-Time,
which aims to meet the instantaneous demand i.e. only the amount necessary for
the customer at that moment of need. These techniques are suitable for systems
that have predictable demand and are determined by the client. However, in
studies on reliability (in the maintenance context), demand is a probabilistic event,
represented by the number of failures (a random variable). For this reason, the
literature on sizing stocks of spare parts in the maintenance area addresses the
question in very specific ways.
Other noteworthy conditions that make them different from production
inventories are (Kennedy et al. 2002; Macchi et al. 2011): the number of spare
parts in stock is often too large; the sourcing of spare parts is often limited to one
or a few suppliers, causing constraints regarding procurement lead time and the
costs; or in the opposite case of multiple sourcing, the related risk of variations in
the quality of materials supplied can occur; obsolescence may be a problem;
indeed, it is difficult to determine how many units of a spare part to stock for an
obsolescent machine; the high variety in the characteristics of spare parts can
normally be observed (the rates of consumption for some parts are very much
276 Chapter 8 Spare Parts Planning Decisions
higher than for others; some parts are characterized by being cheap to buy, while
others are very expensive; often, procurement lead times vary greatly and may be
lengthy, especially in the case of specific parts or those that have to be placed on
order); and, the management process often lacks information visibility, due to
poor inventory data record-keeping, inefficient or ineffective ordering processes
and inventory management information being hidden in separated “silos”, these
being only some typical reasons for such low visibility.
Duchessi et al. (1988) propose a top-down methodology, which classifies
spares into distinct categories and associates appropriate controls with each
category. This methodology identifies spare parts that do not have to be stocked.
By eliminating these spare parts from the inventory, the manager can reduce costs
and thus improve profits. Thereafter, it identifies critical spare parts that, if not in
stock when needed, result in excessive downtime costs. Moreover, avoidance of
downtime reduces production lead time and improves performance regarding on
time delivery to customers. Finally, it displays a logical framework so that the
need for and stock of spare parts can be matched with formal control policies,
procedures and techniques.
Molenaers et al. (2012) propose a spare part classification method based on the
criticality of an item, using an MCDM/A model. Starting from a multicriteria
analysis, the proposed model converts relevant criteria on such criticality into a
single score which then is considered the level of criticality of the item. This level
is used to rationalize the efficiency of the spare parts inventory policy.
A literature review on MCDM/A approaches in reliability and maintenance
shows work conducted related to spare parts sizing (de Almeida et al. 2015).
This text highlights some approaches to the problem of sizing the need for spare
parts (de Almeida and Souza 2001):
x An approach based on the risk of inventory shortages;
x An approach based on the risk of inventory shortages by using prior knowledge;
x An approach under the cost constraint;
x An approach according to an MCDM/A model.
The system type, whether repairable or not repairable, will influence how to size
the need for spare parts. For non-repairable systems, the desired lifecycle of the
system should be considered as a variable time T (de Almeida 1996; de Almeida
8.2 Some Sizing Approaches for Spare Parts in Repair 277
and Souza 2001). It should be noted, therefore, that the size of the stock is defined
by the difficulty in acquiring spare parts (price, delivery time, availability of more
than one supplier, etc.) and by issues directly to inventory management (available
space, cost storage, etc.).
As to repairable systems, the variable T is equal to the time at which the item
will be repaired, i.e., the system is restored when the defective item is replaced
with a similar one that is already in stock. The number of spare parts in this case is
equal to N s N 1 , since the defective item returns to stock after being repaired
(de Almeida 1996; de Almeida and Souza 2001).
Another issue that will influence spare parts management is related to the
behavior of the failure rate over time. As seen, the number of items available for
spare parts is directly related to the number of failures, which in turn is directly
related to the reliability of the equipment (system). Therefore, the problem is
directly related to the behavior that the variable deemed the number of failures is a
function of time. One should also consider the independence of failures among the
items that make up the system.
Under the analysis of the bathtub curve, spare parts management, in the repair
context, is usually dealt with only in the second life stage that matches the useful
life or the operational phase of the equipment. At this stage, it is assumed that the
failure rate O(t) has a constant behavior as a function of time (the reliability
function is represented by an exponential probability function).
In the first phase of the bathtub curve, in which the predominant faults are
classified as early failures, these are usually covered by the equipment manu-
facturer’s warranty, with no need for the user to direct efforts to solve this
problem, i.e., it is not necessary to have spare parts in stock to cover this period in
the lifecycle of the equipment. In some specific kinds of contract, it is interesting
to analyze the possibility of having spare parts. From the manufacturer point of
view, the sizing decision for this stage has to be made and may follow the model
presented in this section, with proper assumptions.
In the third phase of the bathtub curve, the equipment is at the end of its useful
life. Therefore it might not make much sense to study the problem of dimension-
ing the need for spare parts, in the context of repair, because at this stage the
failure rate is high due to wear and tear. The failure rate O(t) increases with time
so that repair is not sufficient to change the behavior of degenerative equipment,
so the equipment has reached its use limit at this stage. At this stage, what remains
is to consider the policy for preventive maintenance, replacement, reconstruction
or overhaul. If economically feasible, this period may be prolonged as necessary
until the equipment is deemed obsolescent and can then be discarded.
The spare parts sizing in this different context should use the information
collected for the maintenance decision in that particular context. For instance, if a
preventive maintenance model, such as one of those in Chap. 5, is applied, then,
the information from the decision model regarding to the amount of replacements
necessary in the planning time horizon is related to the sizing of the spare parts.
278 Chapter 8 Spare Parts Planning Decisions
Another relevant factor to be considered for sizing the need for spare parts is
technological outdating, which can be a limiting factor in the lifetime of a piece of
equipment (system) and thus, may well shorten its life expectancy. Therefore, when-
ever the equipment becomes technologically outdated before the end of its useful
life, the spare parts for it that are in stock lose their functionality in short periods,
there being an economic loss (obsolete inventory) that need to be written off.
Furthermore, what to do about perishable goods (spare parts subject to
degradation while held in stock) should be made of the Wilson model defined in
Rezg et al. (2008) and Ben-Daya et al. (2009). Gopalakrishnan and Banerji (2013)
point out that perishable spares, with a short shelf life, must be identified, and the
First in First out method must be practiced. Therefore, the optimal sizing of the
total quantity of each spare part has to be determined, and must take into
consideration the objectives of minimizing the cost to the system and wastage
(loss of materials due to deterioration) as investment constraints (Padmanabhan
and Vrat 1990).
Van Volkenburg et al. (2014) develop a model which addresses the effects of
the shelf-life of spare parts (perishable items) on optimizing the stocking of spare
parts because certain conditions exacerbate their deterioration, thereby affecting
the reliability of the system being supported or the spare part being found to be
unserviceable when required. This is especially evident in non-repairable
components that are stored for extended periods.
This approach involves determining a number of spare parts N, for a given value
of risk of stock shortages D within a particular time value T. Thus, cost is
considered in an indirect way, because as the desire is to reduce the risk D, there is
a resultant increase in cost and vice versa. So the cost is obtained at the instant that
defines what level of risk to run (i.e., will be determined at the time of choosing
the value of D).
The risk of stock shortages D means the probability that the number of spare
parts in stock is less than the number of failures x, namely, Px ! N (Probability
of Stockout of the spare parts (PS)). Thus, the Margin of Safety (MOS) is defined
as MOS 1 D , i.e. MOS 1 PS , which is a measure of the probability that
the stock will not fall outside the range considered (de Almeida 1996; de Almeida
and Souza 2001). Therefore,
MOS 1 D Px d N (8.1)
where N is the number of spare parts kept in stock. Notice that the MOS
corresponds to the cumulative probability distribution of the number of failures.
Assuming a Poisson Process, for a system comprising n items:
8.2 Some Sizing Approaches for Spare Parts in Repair 279
N
nOT k e nOT
MOS P x d N ¦
k 0
k!
(8.2)
As O s nO :
N
O S T k e OST
MOS Px d N ¦ k!
(8.3)
k 0
where N is the number of items held in stock; O s rate is the failure system and T
is the time interval.
Finally, there is a procedure for calculating the number of spare parts of N, for
some Drisk of stockout or the MOS, so that, respectively:
Px ! N D (8.4)
or
Px d N t MOS (8.5)
There are practical situations where it is not possible to obtain the values for the
parameters of reliability and/or maintainability of a system. This approach
provides a procedure for sizing the need for spare parts where at least one of these
parameters is not known (de Almeida 1996; de Almeida and Souza 2001).
In such cases, prior knowledge is used (as discussed in Chap. 3) with respect to
the reliability and/or maintainability of the system. Therefore, the prior probability
is applied to obtain the expected values of risk or MOS in order to determine what
the appropriate number of spare parts to be held in stock should be.
For this study, three scenarios are considered:
x Lack of knowledge about the failure rate O ;
x Lack of knowledge about the MTTR (Mean Time to Repair);
x Lack of knowledge about the parameter O and MTTR.
280 Chapter 8 Spare Parts Planning Decisions
In the first case, the absence of O, one obtains the prior probability of O: S O ;
in the second case, one should obtain the prior probability on the MTTR:
S MTTR . I.e., these two functions of prior probabilities are required. Therefore,
to address the problem of sizing the need for spare parts in the absence of data, it
is considered the expected value of MOS data as being derived from previously
defined situations, respectively:
§ N nOT k e nOT ·
EO >MOS @ ³ MOS S O dO ³¦¨ ¸S O dO (8.6)
¨ k! ¸
O O ©k 0 ¹
ª º
EO , MTTR >MOS @ ³ ³ « MOS S O dO »S MTTR dMTTR (8.8)
«
MTTR ¬ O
»¼
In this approach, the attribute of value is treated directly. The cost criterion is seen
as a limiting factor, since for a given cost limit, one tries to minimize the risk of
breakage of stock shortages, i.e., starting from the amount of (monetary) resources
that have been allocated in order to determine the optimal number of inventory
items that should be held (Goldman and Slatery 1977).
The decision process is to determine the threshold value of cost, which depends
on the availability of resources. This determines what the number of spare parts of
N is that minimizes the risk of stock shortages.
Therefore, an expression for calculating the number of spare parts of N needed,
so that:
8.2 Some Sizing Approaches for Spare Parts in Repair 281
CT d C0 (8.9)
where CT is the final total cost and C 0 is the amount of resources available is:
CT N .C (8.10)
where N is the number of spare parts and C is the unit cost of each item. The
procedure consists of finding the value of N, starting from N 0 (no spare parts),
that minimizes the risk of inventory shortages that meets the condition previously
established by the budget constraint.
In more complex situations, such as a modularized system that has equipment
with a number of J different types of modules, where each module has its own Oj
failure rate. The final total cost is obtained by summing the total final costs of
each module:
J
CT ¦ N j .C j (8.11)
j 0
where Cj is the individual cost of the module type j; and, Nj is the number of
modules (items) of type j.
For this approach, a MOS in which N N 1 ,! , N j is considered to represent
the probability that there will be no stock shortages of N, i.e., this is given by the
product operator between the MOS modules:
J
MOS N Px j d N j (8.12)
j 0
Therefore the solution is to maximize MOS N , such that the total cost is less
than or equal to the initial cost imposed as a constraint, CT N d C 0 . For this, one
needs to use non-linear optimization.
frequently faced with the demand of achieving several targets (e.g. low costs, high
revenues, high reliability, low accident risks), some of which may very well be in
conflict with each other. At the same time, several peculiar requirements (e.g. in
spacecraft systems, maximum allowable weight, volume, etc.) should also be
satisfied (Marseguerra et al. 2005).
Unlike a single objective approach, which often implies the poor performance
of other desired objectives, the set identified by a multi-objective approach
provides a spectrum of ‘acceptable’ solutions and attempts made to find a
compromise. This is one of the advantages of working under the multidimensional
aspect (multiobjective).
According to Jajimoggala et al. (2012), a systematic evaluation of the criticality
of spare parts is the key to effective control of spare parts recorded by inventory
systems. There are many factors to be considered which will measure the
criticality of spare parts for maintenance activities, and these evaluation pro-
cedures involve several objectives and it is often necessary to make compromises
among possibly conflicting tangible and intangible factors. In their study,
Jajimoggala et al. (2012) use an MCDM approach to solve this kind of problem,
namely by using a three-phase hybrid model: the first stage involves identifying
the criteria; the second is to prioritize the different criteria using fuzzy ANP; and,
finally in the third phase, the criticality of spare parts is ranked using fuzzy
TOPSIS.
Molenaers et al. (2012) propose a spare parts classification method based on
the criticality of items. Starting from a multicriteria analysis, the proposed model
converts relevant criteria impacting the criticality of an item into a single score
which thus represents the criticality level. This level is used to rationalize the
efficiency of the spare parts inventory policy. They consider the following criteria:
x The criticality of equipment;
x The probability of an item failing;
x Replenishment time;
x The number of potential suppliers;
x The availability of technical specifications, and;
x Maintenance type.
de Almeida (2001) considers two criteria (risk and cost), which are combined
through a multi-attribute utility function in a decision model for provisioning
spares, i.e., spares provisioning can also be modelled by the multicriteria utility
function (by the MAUT method) based on the need for spares and the risk of no
supply.
MAUT has been rarely used for the spares provisioning problem. Several
criteria such as: availability, risk, and cost are used to estimate the volume of
spares needed. Risk is a common criterion used in Mickel and Heim (1990). Other
models optimize a single criterion such as availability or risk subject to costs
(Goldman and Slattery 1977; Barlow et al. 1996).
8.2 Some Sizing Approaches for Spare Parts in Repair 283
The combination of these two attributes will be made through the utility
function (multiattribute function). The decision to be adopted in this approach is to
determine values for the attributes of cost C and risk D in order to maximize
the multi-attribute utility function of consequence U C , D .
By concepts of Decision Theory, one has the space of action a , which
consists of the possible quantities of spare parts N, which is the element on which
the decision maker can act in order to achieve the desired goal, in which case it is
the maximization of the multiattribute utility function U C , D .
Furthermore, the state of nature T , the reliability of the system and the
maintainability of its structure need to be considered. They can be represented by
the parameters of reliability and maintainability, which can be obtained by using a
statistical procedure or the use of experts’ prior knowledge, as previously
mentioned (de Almeida 2001; de Almeida 1996).
The observation data (obtained by an analysis of likelihood) concerning the
reliability and maintainability of the system under study allows some
considerations about the behavior of the state of nature T . The state of nature
has a direct influence on the results of the consequences of the decision made by
the decision maker, but the decision maker does not have any control or influence
over the state of nature.
The consequence space is given by the expected utility value E >u C , D @ . The
function u C , D is obtained using the procedure for eliciting a multi-attribute
utility function (described in Keeney and Raiffa (1976)), which defines the DM’s
preference structure with respect to the values of cost and risk of stock shortages.
Finally, the goal of the approach is to determine the number of N spare parts,
which maximizes u C , D . The mathematical model is given by:
u T , a E p|T ,a u p ³ u p P p | T , a dp ³ uC, D PC, D | T , a dp (8.13)
P p | T , a PC , D | T , a PC | T , a PD | T , a (8.14)
PCi | T , a 1 iff , a ai (8.15)
Hence,
P p | T , a PC | T , a PD | T , a 1 PD | T , a PD | T , a (8.16)
N
nOT k e nOT
° P D | T , a
°
P D 1 MOS | T , a 1 iff, MOS ¦
k 0
k!
® (8.17)
N
° nOT k e nOT
° P D | T , a
¯
P D 1 MOS | T , a 0 iff, MOS z ¦
k 0
k!
max u ai , T S T dT
³T (8.18)
ai
In this model, one considers that the state of nature has two dimensions: one
that matches the reliability of the equipment comprising the system, represented
by the rate of system failures (Os); and the second dimension is the maintainability
8.3 Multiple Spare Parts Sizing 285
Tmax O max
>
max EO ,T >u O , T ; N @
N
@ (8.20)
ªTmax O max º
max « ³ ³ u O , T ; N S O S T dOdT », iff,
N « »
¬ T0 O 0 ¼ (8.21)
N
nOT k e nOT
D 1 ¦
k 0
k!
and C a i Ci
system revenues and the minimization of the total volume of spares. A Monte
Carlo simulation approach was defined to deal with system failure, repair and
replacement stochastic processes. Ilgin and Tunali (2007) propose an approach
using genetic algorithms to optimize preventive maintenance and spares policies
of a manufacturing system operating in the automotive sector, while Lee et al.
(2008) develop a framework that integrates a multi-objective evolutionary
algorithm (MOEA) with a multi-objective computing budget allocation (MOCBA)
method for the multi-objective simulation optimization problem of allocating
spare parts for aircraft.
In general, the maintenance manager is interested in minimizing the total cost
of spare parts and also minimizing the probability of stockout. In this section, a
multi-objective genetic algorithm is proposed to tackle the multi spare parts
problem. Firstly, it is important to point out that this model assumes the ‘fixed’
shape of the failure rate. Finkelstein and Cha (2013) state that this assumption is
well founded for the spare parts setting. This feature of a failure rate makes sense
only for spare parts used in corrective maintenance, rather than parts used in
preventive maintenance for which consumption can be defined by a periodic
replacement strategy. This setting justifies the use of the Poisson distribution for
the computations of the probability of stockout of an item. It is also assumed that
each item has a failure rate and purchase cost. Each item can be classified into two
levels of importance to the system, in order to manage different levels of criticality
of items to the system. There are critical and non-critical items which compete for
the same resources of a limited budget.
A multi-objective model based on NSGA-II is developed to aid the
management of multiple spare parts. The model was tested in an urban passenger
bus transport company.
The mathematical model proposed for the spare parts inventory problem was a
multi-objective optimization model, where the objectives are to optimize the
average of the probability of stockout, and the total cost of the spare parts
purchased, which should be minimized.
The model aims to answer the main question inherent in any process of
inventory management: what is the ideal inventory level for a spare part which can
be obtained at minimum cost and provide maximum availability.
As shown by Kennedy et al. (2002) and Bevilacqua et al. (2008), the Poisson
distribution is the most widely-used mathematical-statistical model in the
literature for optimizing inventories of spare parts, and is premised on modelling
the behavior of demand for the item by a probability distribution, which is widely
used to describe rare random events. The Poisson distribution is represented by
(8.22):
8.3 Multiple Spare Parts Sizing 287
(Ot ) x e Ot
Px (t ) (8.22)
x!
Ni
( O it ) i e Oit
PS i P ( x ! N i ) 1 Px d N 1 MOS 1 ¦ i!
(8.23)
i 0
where PS i is the probability of stockout of the i-th spare part associated with the
quantity N i ; i represents each spare part (critical or non-critical); N i is the
amount in stock of each spare part; Oi is the monthly rate of consumption of the i-
th spare part; and, C i is the unit cost of the i-th spare part.
Hence, the problem consists of minimizing the objective functions (8.24) and
(8.25).
ª n º
¦
min « PS i »
«¬ i 1 »¼
(8.24)
ª n º
¦
min « C i »
«¬ i 1 »¼
(8.25)
The genetic algorithm used to solve the spare parts inventory problem was an
adaptation of the elitist multi-objective genetic algorithm proposed by Deb et al.
(2002), NSGA II. The algorithm is based on sorting the chromosomes based on
non-dominance to find the Pareto front of multi-objective problems, as well as to
maintain the good solutions during the evolutionary process, since it is an elitist
algorithm. In the proposed algorithm, the length of the chromosome is equal to the
number of different spare parts, where each gene represents the amount to be
purchased.
The selection of parents for crossover operation to generate offspring is
random, and, for each set of parents chosen, two descendants are generated using
the genetic operators. The crossover operator is based on random selection of a
position of the chromosome that is the cutoff point. The offspring 1 inherit the
288 Chapter 8 Spare Parts Planning Decisions
genes of parent 1 until the cutoff position and, from there on, they inherit the
genes of parent 2. Additionally, the offspring 2, thus, inherit the genes of the
parent 2 until the cutoff point, and, from that point on, inherit the genes of parent 1.
For the case study, the procedure followed was: (1) to define the components of
the replacement spare parts of the buses to be studied, such that 14 critical items
used were defined in corrective maintenance actions alone, and (2) to define the
parameters to be quantified, which in this case were the consumption rate Ȝ, the
unit price and the initial stock. These parameters are shown in Table 8.1. Items
which cause transport service failure were determined as critical items.
The data were collected from an urban collective public transport company that
has been operating buses for more than 25 years and is regarded as anonymous in
this study. This company has a fleet of 83 buses, the average age of which is 5.69
years, which run 600,000 km per month.
Initially, the algorithm was run for the critical items. 99.9% was established as
the upper limit of the average of probability of stockout, which ensures a high
quality of service obtained by the purchase of critical items. As mentioned in the
previous section, the initial solution of the genetic algorithm was generated as the
result of applying the model based on cost benefit ratio (CB), which is obtained
through the cost ratio by varying the level of service caused by the purchase of
spares. The algorithm based on cost-benefit presented a total of 142 solutions in
the Pareto front. The population size chosen for use in NSGA II was twice the
amount of solutions obtained by the CB model, i.e., 284. The first 142 chromo-
somes of the initial solution are the same chromosomes obtained by the CB
model, and the other half of the chromosomes is random generated, so that the
diversity in the solutions is preserved.
After 250 iterations of the genetic algorithm, a total of 276 solutions on the
Pareto front are obtained. Of this number, only 21 coincide with the solutions
generated by the CB model, which shows that the genetic operators have
diversified the initial solution a lot. If analyzed together, the two models generated
a total of 397 different solutions, of which 363 are non-dominated. A comparative
graph of the solutions of the model based on cost-benefit and NSGA II for critical
items is shown in Fig. 8.1 and Fig 8.2.
0.7
0.6
Probability of Stockout
0.5
0.4
0.3
0.2
0.1
0
0 10000 20000 30000 40000 50000 60000 70000
Total Cost
Fig. 8.1 Total cost versus probability of stockout for critical items (Cost-benefit)
290 Chapter 8 Spare Parts Planning Decisions
0.7
0.6
Probability of Stockout
0.5
0.4
0.3
0.2
0.1
0
0 10000 20000 30000 40000 50000 60000 70000
Total Cost
Fig. 8.2 Total cost versus probability of stockout for critical items (NSGA-II)
In Fig. 8.1 and Fig. 8.2, it can be realized that from an PS (Probability of
Stockout) of 10% ahead, it is soon seen that there is a “saturation” in the curve,
thus reversing the prevailing logic, i.e. there are then high investments for little
return (low reduction in the probability of stockout), which clearly it is not worth
the company’s spending resources on, in this situation.
The option to deal separately with the critical items allows the manager to have
greater flexibility in managing the contingency element of his/her budget, and
certainly yields a better result for inventory management as it allows the logic of
the program, based on the typical problem of a portfolio of assets, in which several
items, within their group of criticality, compete for resources simultaneously, thus
gaining the one that presents the lowest cost-benefit index, which brings a gain to
the operation as a whole.
It can be concluded that the model developed and applied in a real situation
reached its objective, as it allowed important parameters for controlling the
inventory of replacement spare parts to be monitored efficiently, thus contributing
to the management of an urban bus company. It is further understood that this
model can be replicated in any other company which has replacement spare parts
in its inventory and consumes them when carrying out corrective maintenance.
Probability of failure, inspection period, holding cost and obsolescence are crucial
factors in modeling spare parts inventories. In terms of the maintenance policy,
one can argue that condition monitoring may well give a better forecast of the
residual life of the system monitored and can support better decisions about
8.4 Spare Parts for CBM 291
level can be set up. While in indirect monitoring, one can only collect measure-
ments related to the actual condition of the item monitored in a stochastic manner.
Some enhancements to direct monitoring have been made. Rausch and Liao
(2010) develop a model for joint production and spare part inventory based on
CBM, where the condition monitored can be observed directly. Wang et al. (2009)
present the concept of condition-based replacement and spares provisioning
policy, and through the simulation method and the genetic algorithm, the decision
variables were jointly optimized for minimizing the cost rate. Linear and expo-
nential degradation models is evaluated by Elwany and Gebraeel (2008) in order
to support the dynamic decisions of replacement and inventory based on the
physical condition of the equipment. Ferreira et al. (2009) propose a multicriteria
decision model to determine inspection intervals of condition monitoring based on
delay time analysis.
Ferreira and Wang (2012) assume that there are a number of identical compo-
nent items used in a system, which are condition monitored periodically. For
example, there may be many critical and identical bearings installed on a paper
machine and proper maintenance of these bearings should lead to better availability
and lowering the operating costs of the machine as a result of having condition
monitoring information. Having the appropriate volume of spare parts available at
the right time is a relevant issue when managing maintenance activities.
Opportunities for maintenance actions such as condition monitoring and
preventive maintenance times, likewise the order time and arrivals of acquisitions,
are illustrated in Fig. 8.3 in order to represent the main features of the problem.
CM1 CMi t
PM1 PMk t
OT1 OTn t
t
AT1 ATn
Fig. 8.3 Intervals of condition monitoring (CM), preventive maintenance (PM), order time (OT),
order arrival (AT) and lead time (W) of spare parts
8.4 Spare Parts for CBM 293
Immediate replacement
Stock level > 0
Stock level = 0
Replacement
Stock level > 0
at the next CMi
Failure before CMi
Stock level = 0
No Failures before CMik
Stock level > 0
Stock level = 0
Replacement
Stock level > 0
at the next ATm
Failure before ATm
Stock level = 0
No Failures before ATm
Stock level > 0
Stock level = 0
Replacement
Stock level > 0
at the next PMk
Failure before PMk
Stock level = 0
No Failures before PMik
Stock level > 0
Stock level = 0
From the decision tree structure of Fig. 8.4, it is possible to evaluate dynamically
the performance of a given maintenance policy by comparing the results and
analyzing the replacement times. Based on the monitored information, the structure
of the risk of stockout, costs and estimates of the residual life are derived. These
estimates may vary which implies that the need for spare parts may change.
This section addresses a spare part problem by using condition monitoring
information. CBM is a more cost effective maintenance policy than time-based
maintenance since it can avoid premature maintenance or replacement while
making better forecasts of the need for spare parts. In traditional CBM models,
there is a strong assumption that spare parts are always available when needed,
and in several practical situations this is not true.
References
Aronis K-P, Magou I, Dekker R, Tagaras G (2004) Inventory control of spare parts using a
Bayesian approach: A case study. Eur J Oper Res 154:730–739
Barabadi A, Barabady J, Markeset T (2014) Application of reliability models with covariates in
spare part prediction and optimization - A case study. Reliab Eng Syst Saf 123:1–7
Barlow RE, Proschan F (1965) Mathematical theory of reliability. John Wiley & Sons,
New York
Ben-Daya M, Duffuaa SO, Raouf A, et al. (2009) Handbook of Maintenance Management and
Engineering. Springer, London
Bevilacqua M, Ciarapica FE, Giacchetta G (2008) Spare parts inventory control for the
maintenance of productive plants. 2008 IEEE Int. Conf. Ind. Eng. Eng. Manag. IEEM 2008.
IEEE, Singapore, pp 1380–1384
Boylan JE, Syntetos AA (2010) Spare parts management: A review of forecasting research and
extensions. IMA J Manag Math 21:227–237
Chang PL, Chou YC, Huang MG (2005) A (r,r,Q) inventory model for spare parts involving
equipment criticality. Int J Prod Econ 97:66–74
de Almeida AT (1996) Multicriteria for spares provisioning using additive utility function.
In: International Conference on Operational Research for Development, IFORS-ICORD II.
Rio de Janeiro, RJ, pp 1414–1418.
de Almeida AT (2001) Multicriteria decision making on maintenance: Spares and contracts
planning. Eur J Oper Res 129:235–241
de Almeida AT, Ferreira RJP, Cavalcante CAV (2015) A review of multicriteria and multi-
objective models in maintenance and reliability problems. IMA Journal of Management
Mathematics 26(3):249–271
de Almeida AT, Souza FMC (2001) Gestão da Manutenção: na Direção da Competitividade
(Maintenance Management: Toward Competitiveness) Editora Universitária da UFPE. Recife
Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multiobjective genetic
algorithm: NSGA-II. IEEE Trans Evol Comput 6:182–197
Diallo C, Ait-Kadi D, Chelbi A (2008) (s,Q) Spare parts provisioning strategy for periodically
replaced systems. IEEE Trans Reliab 57:134–139
Duchessi P, Tayi GK, Levy JB (1988) A Conceptual Approach for Managing of Spare Parts. Int
J Phys Distrib Logist Manag 18:8–15
Elwany AH, Gebraeel NZ (2008) Sensor-driven prognostic models for equipment replacement
and spare parts inventory. IIE Trans 40:629–639
References 295
Ferreira RJP, de Almeida AT, Cavalcante CAV (2009) A multi-criteria decision model to
determine inspection intervals of condition monitoring based on delay time analysis. Reliab
Eng Syst Saf 94:905–912
Ferreira RJP, Wang W (2012) Spare parts optimisation subject to condition monitoring. In: 11th
International Probabilistic Safety Assessment and Management. Conference and the Annual
European Safety and Reliability Conference, Helsinki, Finland, 25-29 June 2012
Finkelstein M, Cha JH (2013) Stochastic Modeling for Reliability: Shocks, Burn-in and
Heterogeneous populations. Springer, London
Goldman AS, Slattery TB (1977) Maintainability: a major element of system effectiveness.
Robert E. Krieger Publishing Company, New York
Gopalakrishnan P, Banerji AK (2013) Maintenance and Spare Parts Management. PHI Learning,
New Delhi
Ilgin MA, Tunali S (2007) Joint optimization of spare parts inventory and maintenance policies
using genetic algorithms. Int J Adv Manuf Technol 34:594–604
Jajimoggala S, Rao VVSK, Beela S (2012) Spare parts criticality evaluation using hybrid
multiple criteria decision making technique. Int J Inf Decis Sci 4:350
Kabir ABMZ, Al-Olayan AS (1996) A stocking policy for spare part provisioning under age
based preventive replacement. Eur J Oper Res 90:171–181
Keeney RL, Raiffa H (1976) Decisions with multiple objectives: Preferences and Value Trade-
Offs. Wiley Series in Probability and Mathematical Statistics. Wiley and Sons, New York
Kennedy WJ, Wayne Patterson J, Fredendall LD (2002) An overview of recent literature on
spare parts inventories. Int J Prod Econ 76:201–215
Lee LH, Chew EP, Teng S, Chen Y (2008) Multi-objective simulation-based evolutionary
algorithm for an aircraft spare parts allocation problem. Eur J Oper Res 189:476–491
Macchi M, Fumagalli L, Pinto R, Cavalieri S (2011) A Decision Making Framework for
Managing Maintenance Spare Parts in Case of Lumpy Demand: Action Research in the
Avionic Sector. In: Altay N, Litteral LA (eds) Serv. Parts Manag. pp 89–104
Marseguerra M, Zio E, Podofillini L (2005) Multiobjective spare part allocation by means of
genetic algorithms and Monte Carlo simulation. Reliab Eng Syst Saf 87:325–335
Mickel LS, Heim RL (1990) The spares calculator: a visual aid to provisioning. Annu. Proc.
Reliab. Maintainab. Symp. IEEE, Los Angeles, CA, pp 410 – 414
Molenaers A, Baets H, Pintelon L, Waeyenbergh G (2012) Criticality classification of spare
parts: A case study. Int J Prod Econ 140:570–578
Padmanabhan G, Vrat P (1990) Analysis of multi-item inventory systems under resource
constraints: A non-linear goal programming approach. Eng Costs Prod Econ 20:121–127
Porras E, Dekker R (2008) An inventory control system for spare parts at a refinery: An
empirical comparison of different re-order point methods. Eur J Oper Res 184:101–132
Raiffa H (1968) Decision analysis: introductory lectures on choices under uncertainty. Addison-
Wesley, London
Rausch M, Liao H (2010) Joint production and spare part inventory control strategy driven by
condition based maintenance. IEEE Trans Reliab 59:507–516
Rezg N, Dellagi S, Chelbi A (2008) Joint optimal inventory control and preventive maintenance
policy. Int J Prod Res 46:5349–5365
Roda I, Macchi M, Fumagalli L, Viveros P (2014) A review of multi-criteria classification of
spare parts: From literature analysis to industrial evidences. J Manuf Technol Manag 25:528–
549
Syntetos A, Keyes M, Babai M (2009) Demand categorisation in a European spare parts logistics
network. Int J Oper Prod Manag 29:292–316
Van Horenbeek A, Scarf P, Cavalcante CAV, Pintelon L (2013) On the use of predictive
information in a joint maintenance and inventory policy. In: Steenbergen RDJM, VanGelder
PHAJM, Miraglia S, Vrouwenvelder ACWMT (eds) 22nd Annual Conference on European
Safety and Reliability (ESREL), Amsterdam, 2013. Safety, Reliability and Risk Analysis:
Beyond the Horizon. Taylor & Francis Group, London, UK, p 758
296 Chapter 8 Spare Parts Planning Decisions
9.1 Introduction
Barlow and Prochan (1965) presented models involving redundancy and revealed
how to allocate redundancy among the various subsystems under linear constraints
on weight, volume and cost, in order to maximize system reliability. The problem
of maximizing the reliability of a series system subject to one or more constraints
on total cost, weight, volume, has a number of variations depending on whether the
redundancy is parallel or standby. In parallel systems, redundant units operate
simultaneously, and they are subject to failure. In standby redundancy, redundant
units are place on stand-by as spares and used successively for replacement, and they
are not subject to failure while in standby condition. In some active parallel redundant
configurations, it may be required that k out of the n units must be working for the
system to function. The reliability of a k out of n system, with n independent
components in which all the unit reliabilities are equal, is expressed by the
binomial reliability function (O’Connor and Kleyner 2012; Kuo and Zuo 2003).
Even though Barlow and Prochan (1965) did not use the terminology of an
MCDM/A problem, they indicated a class of problems in which no specific set of
constraints is provided. In this case, one may wish to generate a family of non-
dominated allocations in terms of reliability and cost both in parallel as well as
standby redundancy.
The concept of non-dominated solutions is defined by Barlow and Prochan
(1965) as: x 0 ( x10 , x20 ,..., xn0 ) is non-dominated if R(x)>R(x0) implies cj(x)>cj(x0)
for some j, whereas R(x) = R(x0) implies either cj(x)>cj(x0) for some j or
n
cj(x)=cj(x0) for j, where c j ( x) ¦c x ij i .
i 1
This property is the same as the classical definition of the Pareto Front used in
multiobjective formulations presented in Chap. 2. Barlow and Prochan (1965)
stated that if the set consisting of all non-dominated redundancy allocations is
obtained (the complete family of non-dominated redundancy allocations), then the
solution of a redundancy allocation problem with a set of constraints must be a
member of this family. In other words, Barlow and Prochan (1965) realized that
the mono-objective formulation is a particular case of the MCDM/A formulation
and the solution of a mono-objective case is one from the Pareto Front set.
They presented a procedure that is able to generate an incomplete family of
non-dominated allocations in a single cost factor. The procedure is based on the
principle of adding the most reliability obtained per dollar spent in each iteration,
starting with no redundancy in the system. For a multiple cost factor case, a simple
weighted function of reliability is proposed and arbitrarily chosen values of
weights are recommended. They also suggested a procedure to find a complete
family of non-dominated allocations based on the dynamic programming algorithm
of Kettelle Jr (1962).
A literature review of the optimal redundancy allocation models was carried
out by Tillman et al. (1977). They classified early references in the field in terms
of optimization techniques and system configurations. Among the optimization
techniques, no MCDM/A approach was cited.
9.1 Introduction 299
Kuo and Prasad (2000) updated the literature review of Tillman et al. (1977)
and this included identifying if an MCDM/A approach had been analyzed as a
way to help optimize system-reliability. They found twelve papers within this
scope. They stated that an MCDM/A approach was an important but not widely
studied problem in reliability optimization. Although some exact methods can be
used to solve redundancy allocation problems, heuristics used include: ant colony
optimization method; hybrid genetic algorithm and tabu search.
Kuo and Wan (2007) cited multiobjective optimization as a recent topic and
indicated eleven references to this, which have published since 2000 in this field.
They defined four problem structures, namely: 1) The traditional reliability-
redundancy allocation problem; 2) The percentile life optimization problem; 3)
Multi-state system optimization; 4) Multiobjective optimization.
Some kinds of system configuration are defined as shown in Fig. 9.1 and Fig. 9.2.
1 2 N1
IN OUT
1 2 N2
… … …
1 2 NM
Fig. 9.1 Mixed series-parallel system, N components are connected in series, and M such series
connections are connected in parallel to form the system
IN
1 2 3
4 5
OUT
n
x j 1
maximize R( x j ) (1 (1 p )
j 1
j )
n
subject to ¦c x
i 1
ij j d c j , j 1,..., r (9.1)
Multiobjective particle
Khalili-Damghani et al. (2013) X X X
swarm optimization
Fuzzy multiobjective
Garg and Sharma (2013) X X
particle swarm optimization
Decomposition-based
Cao et al. (2013) X X X
approach
Tchebycheff;
Sahoo et al. (2012) X X Lexicographic; Genetic
Algorithms
Safari (2012) X X NSGA-II
Genetic Pareto set
Okafor and Sun (2012) X X
identification algorithm
Khalili-Damghani and Amiri İ-constraint method and
X X X
(2012) data envelopment analysis
Zio and Bazzo (2011a); Zio and Clustering procedure; Level
X X Availability
Bazzo (2011b) Diagrams and MOGA
NSGA-II and data
Li et al. (2009) X X X
envelopment analysis
Multiobjective hierarchical
Kumar et al. (2009) X X genetic algorithm; SPEA2
and NSGA-II
Physical programming;
Tian et al. (2008) X System utility
Genetic algorithms
(continued)
302 Chapter 9 Decision on Redundancy Allocation
In this section, a decision model (de Almeida and Souza 1993) for a standby
system based on the MAUT is presented. This model addresses the waiting time to
call a repair facility when the first piece of equipment of a 2-unit standby system
fails. The first failure implies only a reliability reduction, not system failure, since
the other unit is still operating. This scheme of waiting-time when the first fault
occurs avoids overtime costs in the repair facility. An expert prior knowledge
approach is lead in order to deal with the uncertainty of the parameters failure and
repair rates. Another decision model (de Almeida and Bohoris 1996) extends this
first model, introducing a Gamma distribution to the repair time. The possible
states for a 2-unit redundant standby system are shown in Fig. 9.4.
The problem involves a question about how to select a suitable maintenance
strategy in order to combine system availability and cost preferences. There is an
assumption that the capacity of repair is limited, and instantaneous repair is not
applicable. An MCDM/A approach can solve the conflicting requirements of
system availability and cost through a multi-attribute utility function taking into
account DM’s preferences over these requirements. In this way, MAUT can also
deal with uncertainty of the consequences.
304 Chapter 9 Decision on Redundancy Allocation
e0 - 2 units working
failure repair
e1 - 1 unit working
and 1 unit failed
failure repair
e2 - 2 units failed
Parameters Description
Ȝ Failure rate of the equipment
ȝ Repair rate of the equipment
a An action, element of the action space, representing the maintenance strategy
e0, e1, e2 State of the system when [0, 1, 2] of the units failed
Ta Decision variable representing the repair delay corresponding to a
T0 Time at which the first failure occurs
T1 Time at which the second failure occurs
Time at which the first-failed unit resumes operation, which could be returning the
T2
system to e0
TTR T2 - Ta
ʌ1(Ȝ) Prior knowledge distribution about Ȝ
ʌ2(ȝ) Prior knowledge distribution about ȝ
Ai Scale parameter of ʌi
(continued)
9.2 An MCDM/A Model for a 2-Unit Redundant Standby System 305
Bi Shape parameter of ʌi
Ci Cost for ai
FCi Fixed cost for ai
CRi Repair cost-rate for ai
MCRi Mean CRi
U{TI,C} Multi-attribute utility function for interruption time and cost
The assumptions of the model (de Almeida and Souza 1993) are:
1. The probability distribution of failure of two units are identically distributed;
2. Each unit has two states: good and failed;
3. The system is down when no unit is available for operation;
4. There is one repair facility;
5. Failure rate (Ȝ) is constant and the number of failures follows a Poisson
distribution;
6. A unit repaired becomes as good as new;
7. Repair rate is constant (ȝ);
8. If during a repair of a failed unit, the other unit also fails, the latter unit waits
for repair until the first unit is repaired;
9. There is prior knowledge about Ȝ and ȝ represented as prior probability
distributions over these parameters;
10.Failure and repair states are s-independent;
11.The DM has a structure of preferences over the consequence space (TI,C)
according to the axiomatic preferences of the utility theory;
12.C and TI are s-independent;
13.The objective function is to maximize the multi-attribute utility function
U{TI,C}
The decision model building was based on the context of a telecommunication
system of an electric power company with a 2-unit standby redundant system. The
DM’s preference elicitation over consequences (interruption time and cost)
produces a multi-attribute utility function, which is introduced into the decision
model, according to (9.5). The expected utility of alternatives is given by (9.2).
Om Pm
E( O , P ) {U {(O , P ), ai }} ³O ³P S 1 (O ) S 2 ( P ) U {(O , P ), ai }dOdP (9.2)
0 0
where:
S 1 (O ) ( B1 / A1 ) (O / A1 ) B1 1 exp[(O / A1 ) B1 ] (9.3)
306 Chapter 9 Decision on Redundancy Allocation
S 2 ( P ) ( B2 / A2 ) ( P / A2 ) B2 1 exp[( P / A2 ) B2 ] (9.4)
§ K P · ª § K kt · º
U {(O , P ), ai } K c U {Ci } ¨¨ t ¸¸ «1 ¨¨ ¸¸ exp( O Tai )» . (9.7)
© K kt P ¹ ¬ © O P ¹ ¼
Prior knowledge about the states of nature can be obtained from prior
distributions of Ȝ and ȝ. There are several prior probability elicitation procedures
available in the literature, such as that given by Winkler (1967). The elicitation
procedure applied is based on equal probable intervals. Based on experts on the
equipment and the system maintainability, respectively, these ʌ1(Ȝ) and ʌ2(ȝ) were
obtained according to (9.3) and (9.4), which are illustrated in Fig. 9.5 and Fig. 9.6.
5.00E+04
4.50E+04
4.00E+04
3.50E+04
3.00E+04
pdf of ʌ Ȝ)
2.50E+04
2.00E+04
1.50E+04
1.00E+04
5.00E+03
0.00E+00
0 0.00001 0.00002 0.00003 0.00004 0.00005 0.00006 0.00007 0.00008 0.00009 0.0001
Ȝ
Fig. 9.5 Prior knowledge about Ȝ, ʌ1(Ȝ) with A1 = 18.06·10-6 and B1 = 1.68
9.2 An MCDM/A Model for a 2-Unit Redundant Standby System 307
4.00E+01
3.50E+01
3.00E+01
2.50E+01
pdf of ʌ ȝ)
2.00E+01
1.50E+01
1.00E+01
5.00E+00
0.00E+00
0 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.1
ȝ
Fig. 9.6 Prior knowledge about ȝ, ʌ2(ȝ) with A2 = 0.028 and B2 = 2.57
There are three possible situations in the state e1 (de Almeida and Souza 1993):
x T1 > Ta, and T1 > T2; therefore TI = 0;
x T1 > Ta, and T1 < T2; therefore TI = T2 - T1 > 0;
x Otherwise, there is an emergency, and the repair facility is called immediately,
so that Ta is set equal to T1.
Then, (9.9) represents the interruption time formulation.
References
Barlow RE, Proschan F (1965) Mathematical theory of reliability. John Wiley & Sons,
New York
Cao D, Murat A, Chinnam RB (2013) Efficient exact optimization of multi-objective redundancy
allocation problems in series-parallel systems. Reliab Eng Syst Saf 111:154–163
Chiang C-H, Chen L-H (2007) Availability allocation and multi-objective optimization for
parallel–series systems. Eur J Oper Res 180:1231–1244
Coit DW, Jin T, Wattanapongsakorn N (2004) System optimization with component reliability
estimation uncertainty: a multi-criteria approach. Reliab IEEE Trans 53:369–380
Coit DW, Konak A (2006) Multiple weighted objectives heuristic for the redundancy allocation
problem. IEEE Trans Reliab 55:551–558
de Almeida AT, Bohoris GA (1996) Decision theory in maintenance strategy of standby system
with gamma-distribution repair-time. Reliab IEEE Trans 45:216–219
de Almeida AT, Ferreira RJP, Cavalcante CAV (2015) A review of multicriteria and multi-
objective models in maintenance and reliability problems. IMA Journal of Management
Mathematics 26(3):249–271
de Almeida AT, Souza FMC (1993) Decision theory in maintenance strategy for a 2-unit
redundant standby system. Reliab IEEE Trans 42:401–407
Dhingra AK (1992) Optimal apportionment of reliability and redundancy in series systems under
multiple objectives. Reliab IEEE Trans 41:576–582
Elegbede C, Adjallah K (2003) Availability allocation to repairable systems with genetic
algorithms: a multi-objective formulation. Reliab Eng Syst Saf 82:319–330
Garg H, Sharma SP (2013) Multi-objective reliability-redundancy allocation problem using
particle swarm optimization. Comput Ind Eng 64:247–255
Gen M, Ida K, Tsujimura Y, Kim CE (1993) Large-scale 0–1 fuzzy goal programming and its
application to reliability optimization problem. Comput Ind Eng 24:539–549
Huang H-Z (1997) Fuzzy multi-objective optimization decision-making of reliability of series
system. Microelectron Reliab 37:447–449
Inagaki T, Inoue K, Akashi H (1978) Interactive Optimization of System Reliability Under
Multiple Objectives. Reliab IEEE Trans R-27:264–267
Kettelle Jr JD (1962) Least-Cost Allocations of Reliability Investment. Oper Res 10:249–265
Khalili-Damghani K, Abtahi A-R, Tavana M (2013) A new multi-objective particle swarm
optimization method for solving reliability redundancy allocation problems. Reliab Eng Syst
Saf 111:58–75
Khalili-Damghani K, Amiri M (2012) Solving binary-state multi-objective reliability redundancy
allocation series-parallel problem using efficient epsilon-constraint, multi-start partial bound
enumeration algorithm, and DEA. Reliab Eng Syst Saf 103:35–44
Kumar R, Izui K, Yoshimura M, Nishiwaki S (2009) Multi-objective hierarchical genetic
algorithms for multilevel redundancy allocation optimization. Reliab Eng Syst Saf 94:891–
904
Kuo W, Prasad VR (2000) An annotated overview of system-reliability optimization. Reliab
IEEE Trans 49:176–187
Kuo W, Wan R (2007) Recent Advances in Optimal Reliability Allocation. In: Levitin G (ed)
Comput. Intell. Reliab. Eng. SE - 1. Springer Berlin Heidelberg, pp 1–36
Kuo W, Zhu X (2012) Importance measures in reliability, risk, and optimization: principles and
applications. John Wiley & Sons, New York
Kuo W, Zuo MJ (2003) Optimal reliability modeling: principles and applications. John Wiley &
Sons, New York
Li Z, Liao H, Coit DW (2009) A two-stage approach for multi-objective decision making with
applications to system reliability optimization. Reliab Eng Syst Saf 94:1585–1592
References 309
Liang Y-C, Lo M-H, Chen Y-C (2007) Variable neighbourhood search for redundancy allocation
problems. IMA J Manag Math 18:135–155
Limbourg P, Kochs H-D (2008) Multi-objective optimization of generalized reliability design
problems using feature models - A concept for early design stages. Reliab Eng Syst Saf
93:815–828
Marseguerra M, Zio E, Podofillini L, Coit DW (2005) Optimal design of reliable network
systems in presence of uncertainty. Reliab IEEE Trans 54:243–253
Misra KB, Sharma U (1991) An efficient approach for multiple criteria redundancy optimization
problems. Microelectron Reliab 31:303–321
O’Connor P, Kleyner A (2012) Practical reliability engineering. John Wiley & Sons, Chichester
Okafor EG, Sun Y-C (2012) Multi-objective optimization of a series–parallel system using
GPSIA. Reliab Eng Syst Saf 103:61–71
Paté-Cornell ME, Dillon RL, Guikema SD (2004) On the Limitations of Redundancies in the
Improvement of System Reliability. Risk Anal 24(6):1423–1436
Rao SS, Dhingra AK (1992) Reliability and redundancy apportionment using crisp and fuzzy
multiobjective optimization approaches. Reliab Eng Syst Saf 37:253–261
Safari J (2012) Multi-objective reliability optimization of series-parallel systems with a choice of
redundancy strategies. Reliab Eng Syst Saf 108:10–20
Sahoo L, Bhunia AK, Kapur PK (2012) Genetic algorithm based multi-objective reliability
optimization in interval environment. Comput Ind Eng 62:152–160
Sakawa M (1978) Multiobjective reliability and redundancy optimization of a series-parallel
system by the Surrogate Worth Trade-off method. Microelectron Reliab 17:465–467
Sakawa M (1980) Reliability design of a standby system by a large-scale multiobjective
optimization method. Microelectron Reliab 20:191–204
Sakawa M (1981) Optimal Reliability-Design of a Series-Parallel System by a Large-Scale
Multiobjective Optimization Method. Reliab IEEE Trans R-30:173–174
Salazar D, Rocco CM, Galván BJ (2006) Optimization of constrained multiple-objective
reliability problems using evolutionary algorithms. Reliab Eng Syst Saf 91:1057–1070
Taboada HA, Baheranwala F, Coit DW, Wattanapongsakorn N (2007) Practical solutions for
multi-objective optimization: An application to system reliability design problems. Reliab
Eng Syst Saf 92:314–322
Taboada HA, Espiritu JF, Coit DW (2008) MOMS-GA: A Multi-Objective Multi-State Genetic
Algorithm for System Reliability Optimization Design Problems. Reliab IEEE Trans 57:182–
191
Tian Z, Zuo MJ (2006) Redundancy allocation for multi-state systems using physical
programming and genetic algorithms. Reliab Eng Syst Saf 91:1049–1056
Tian Z, Zuo MJ, Huang H (2008) Reliability-Redundancy Allocation for Multi-State Series-
Parallel Systems. Reliab IEEE Trans 57:303–310
Tillman FA, Hwang C-L, Kuo W (1977) Optimization Techniques for System Reliability with
Redundancy: A Review. Reliab IEEE Trans R-26:148–155
Winkler RL (1967) The Assessment of Prior Distributions in Bayesian Analysis. J Am Stat
Assoc 62:776–800
Zhao J-H, Liu Z, Dao M-T (2007) Reliability optimization using multiobjective ant colony
system approaches. Reliab Eng Syst Saf 92:109–120
Zio E, Bazzo R (2011) Level Diagrams analysis of Pareto Front for multiobjective system
redundancy allocation. Reliab Eng Syst Saf 96:569–580
Zio E, Bazzo R (2011a) A clustering procedure for reducing the number of representative
solutions in the Pareto Front of multiobjective optimization problems. Eur J Oper Res
210:624–634
Chapter 10
Design Selection Decisions
Abstract: The design selection problem in the RRM context considers long-term
performance, and represents higher additional costs if unforeseen features that
should have been included during the project design phase had to be implemented
afterwards. Design decision involves multiple aspects and may be more critical
depending on the kind of item, such as consumer appliances, industrial equipment
or projects that have to consider safety aspects (airplanes or facilities). Reliability
has an essential role for design selection although other aspects have to be
considered such as maintainability and risk depending on the specific design
problem. Therefore, a multidimensional approach is usually required. In this
chapter, all these aspects are discussed in order to illustrate the importance of a
broader perspective when facing design decision problems. The fundamental
requirements are to consider reliability, maintainability and risk aspects so as to
establish features in the design project, including the definition of material,
redundancies, control systems and safety barriers. To illustrate these decisions,
aspects such as reliability (e.g. MTBF), maintainability (e.g. MTTR), safety, cost,
service life, efficiency, are discussed as criteria for these problems. Multi-attribute
utility theory (MAUT) is applied in this chapter to illustrate how reliability,
maintainability and risk aspects are included in an MCDM/A model for design
selection incorporating states of nature. The decision regarding the selection of
which features to include in a design project may be considered as an MCDM/A
portfolio problem. Finally, an introductory view is given of how the redesign
problem arises in the maintenance context with multicriteria approaches.
10.1 Introduction
The term design may have different meanings, such as project, and aesthetical
conception. The main issue in this chapter is related to the former, although the
latter is specifically applied as one of the criteria decision for a car project,
subsequently presented.
Decisions about the design of a product are determinant to its reliability. Errors
in the design process can increase considerably the costs during the product
development cycle. In this way, reliability is highly connected with problems of
engineering design (O’Connor and Kleyner 2012).
In this field, performance capability and cost are two of the most important
factors of the design. While performance capability means the adequacy of a
design to perform required functions, cost means how much these performance
requirements of a design in monetary units are. To deal with a three-way trade-off
among performance capability, cost and reliability, into the design process, the
designer (the DM) have to examine how reliability requirements are set to build
reliability into a design (Lewis 1987). A literature review on MCDM/A approaches
in reliability and maintenance, shows that 16.7% of the work conducted is related
to design (de Almeida et al. 2015). It is important to note that this decision has
different implications depending on the mode and probability of failures.
Reliability requirements can be defined by several ways, such as by the designer,
by the buyer of the product and by government agencies.
Products that are more reliable imply higher capital costs and lower costs for
maintenance and repair. For the classical optimization approach, it is possible to
describe a function that represents the total cost including capital and repair costs.
Thus, an optimal solution can be found when this function is minimized. How-
ever, DM can have a preference for other solutions instead of the optimal that
minimizes the single objective cost function. This occurs in several practical
situations when the DM has some additional considerations about the trade-off
between capital and maintenance costs. The DM’s preferences are illustrated for
three examples: a mobile phone, an automated industrial machine and an aircraft
turbine.
A mobile phone is a consumer appliance in which reliability requirements rely
on consumer expectations. In this case, the reliability of this product can increase
the prices, reduce repair costs, provide elements to offer a longer warranty than a
competitor and boost sales. While some negative consequences are due to
excessive reliability such as lower sales, some inconveniences are due to the
opposite, such as company reputation for poor design. Design decisions in this
context are made by the manufacturer taking into account public’s preferences
about price and reliability through market surveys.
In a design of an automated industrial machine, or other equipment directed to
large organizations, the DM’s preference structure is completely different. In this
case, the trade-off between capital cost and production lost through breakdowns
should be evaluated. Thus, design decisions have to consider aspects of reliability
and maintainability.
Finally, in a design of an aircraft turbine, failure consequences are so severe
that a higher level of reliability is required. It means that the increase of a turbine
cost can be justifiable due to the level of reliability required to this product.
Additionally, aspects of delays and safety in airline maintenance are also relevant
(Sachon and Paté-Cornell 2000). In this case, insurance underwriters and
government agencies are responsible to define reliability specifications and risk
analysis performs an important role in the design.
10.1 Introduction 313
load2
Ȝ(t)
load1
As stated in the last section, the reliability has a central role in system design.
Additionally, the overall performance of a plant is also associated with
maintainability. The availability of a plant depends on the frequency and the
downtime of interruptions. Maintainability is a design characteristic that reflects
the probability of an item be restored in a given period to specified conditions by
maintenances actions that meet some procedures and resources requirements
(Goldman and Slattery 1977).
The fundamental function of equipment depends on its availability or readiness.
Thus, some trade-offs between reliability and maintainability could be made in
order to meet the availability requirements. Different levels of reliability R(t) and
maintainability M(t) can result an availability A(t) level.
As stated by Goldman and Slattery (1977), maintainability is a concept related
to different aspects from the basic physical characteristic of the design to the
strategic level of the maintenance function. Amongst the different issues
comprised on the spectrum of aspects related to maintainability, it is possible to
highlight design requirements, selection of maintenance strategy, logistic
provisioning, and so on.
Despite these different issues, this section focus on the role of that maintain-
ability plays, providing guidelines on the design specification, in order to provide
to the overall equipment some features related with effectiveness and lifetime
support cost.
In a design selection problem, while the designer has to develop the system,
taking into account various aspects, including the maintainability, and different
constraints imposed by budget project and standards; the user has to accept the
final configuration of the design and handle with the challenges related to
effective operational use of this built design. From the difference between user
and designer perspectives, sometimes, feedbacks have to be considered in both
directions: from the operation to design and from the design to operation.
Design development is complex and non-failure-free task. Thus, besides the
problems that could exceed the design phase, some other problems could arise
related to the discordance between designer and user views. Therefore, analysis of
the design process is essential not only as an alternative to handle with the views
discordances, but also to reduce the number of failures that comes from the project
phase. In this process of improvement, the feedbacks are essential. A simple
reason is that some problems come up after the launch of the product. As any item
is subject to failure, maintenance could be a quite frequent activity. In this case,
the maintenance should be done in order to restore the system for the operational
state, as quickly as possible, so that interruptions resulting from failures do not
affect the production targets.
In order to act as quick as possible, maintenance teams have to deal with some
barriers related with design selection issue considering the difficulty associated to
the maintenance activity. In some cases, the maintainability attributes might be too
restrictive resulting in difficulties to reduce maintenance times. For example, the
316 Chapter 10 Design Selection Decisions
In this section is presented an illustrative example of the design selection for a car
according to the MCDM/A approach presented in Chap. 2.
According to the first step discussed in Chap. 2, it is important to make some
remarks and observations to characterize a DM involved in such problem. In this
illustrative example, the DM is the senior engineer responsible for defining which
car project should be selected as the best design from the product development
stage.
At this point is important to notice that all standards and requirements have
already been achieved by each project, thus, the decision regards on evaluating
these alternatives considering the factors or objectives related to this decision,
which recalls the second step pointed in Chap. 2.
As the objectives for this kind of decision, one may list the following
objectives: Maintainability, Reliability, Safety, Cost, Service life, Efficiency,
Aesthetic.
For simplification purposes this illustrative example will not consider the last
three objectives. Focusing on the first four objectives allows to explore more
deeply the RRM perspective of such problem.
Other objectives may be more emphasized during different phases of the
product development, taking the initial list of objectives as a reference or by the
addition of other objectives depending on specific aspects related to the problem
context.
The definition of criteria related to each objective refers to the third step given
in Chap. 2 with regard to establishing criteria.
In order to measure maintainability, the concepts given in previous sections and
in Chap. 3 shall be used, thus as the maintainability concept relates to the time
spent during repairs.
The maintainability function for any probability distribution may be
represented as in (10.1) (Dhillon 1999; Stapelberg 2009), where t represents time
and fr(t) is the pdf of the repair time.
t
m(t ) ³ f (t )dt
r
(10.1)
0
The objective cost measures and evaluates all aspects that may be converted
into a monetary value scale. It is important to emphasize that MCDM/A
approaches provide methodological support to understand and value alternatives
among different objectives scales and provide a global evaluation that includes all
objectives.
Thus, when measuring cost, all factors that may be represented by monetary
scale shall be summed in order to evaluate the respective cost of each car project.
In this example, it shall be assumed that all these factors have been considered for
estimating each car project cost.
The elements of the set of alternatives are the car projects. The problematic
relies on the choice of the best car project according to the corresponding criteria.
The states of nature in this illustrative problem corresponds to factors that are
not under the DM’s control and are subject to uncertainty, influencing the decision
outcomes, such as when a failure may occur or the time for a repair service. Such
uncertainties may be represented by probability distributions that provide the
performance estimates for each car project.
Assuming that the DM preferences fit the axiomatic structure required by
MAUT, the next step regards the intra criterion evaluation. Establishing intra
criterion evaluation consists in the definition of a utility function for each criterion
by assessing its shape and parameters from the DM’s evaluation of the outcomes
in each criterion through lotteries or other elicitation procedure.
For a DM with additive independence condition, the additive utility function is
represented by (10.3):
where ua is the expected utility of a, based on the additive function over the
expected utility of the attributes maintainability (m), reliability (r), safety (s) and
cost (c); km, kr, ks and kc are the respective scale constants.
The scale constants in MAUT are elicited through lotteries as given in Chap. 2.
For this illustrative problem, consider km, kr and ks equal to 0.3 and kc equals to 0.1.
The respective one-dimensional utility functions are um(m), ur(r), us(s) and
uc(c). For maintainability (m) and reliability (r), the random variable repair time
(m) and time to failure (r) are considered consequences, with its respective pdfs
fa(m), fa(r), for each alternative a. Their expected utilities are respectively given as
follows by (10.4) and (10.5).
f
Ea [um ( m )] ³u m ( m ) f a ( m ) dm (10.4)
0
f
Ea [ur ( r )] ³ u (r) f
r a ( r ) dr (10.5)
0
320 Chapter 10 Design Selection Decisions
Table 10.1 Car project alternatives evaluation for a design selection problem
From the results given by Table 10.1 the best car project for such DM would be
the alternative 4, which achieved the highest value in the additive utility function.
By evaluating alternative 4 individual utilities is interesting to observe that it has
the best performance on the maintainability attribute, although presenting one of
the worst values for the attribute cost, only better than alternative 6.
Thus, for a DM with such preferences, the alternative 4 values for
maintainability, reliability and safety are compensating its bad outcome for the
cost attribute. A different DM may have different preferences and make different
tradeoffs, leading to the selection of a different alternative. As pointed in Chap. 2,
the sensitivity analysis is an important step in order to evaluate the robustness of
the preliminary recommendation, allowing to increase the accuracy of the
elicitation process when it is required.
For a DM with a non-compensatory rationale, the selected alternative could be
other due to a different kind of preference structure. Such DM with a different
preference structure, would set different model parameters and establish different
comparisons, such that in many times would be reflected as the selection of a
different alternative. The use of the MCDM/A approach enriches the decision
process by allowing to incorporate these particularities of the DM preferences
with accuracy, by incorporating it in the decision model.
10.3 Risk Evaluation for Design Selection 321
During the design phase, it is possible to improve system reliability and risk
barriers in order to reduce risks and avoid unnecessary costs to adjust and match
the project to the required risk standards. During this phase, accident rates can be
influenced when deciding on what material and components to use in the project.
Many studies consider that risk evaluation follows a trend, which reflects
probabilities separately from consequences. As a result of this trend, many
decisions are taken which consider only probabilities or consequences, without
aggregating these two important factors when evaluating the overall risk.
This may occur due to the difficulties in estimating and/or simulating these
processes to quantify probabilities and the magnitude of the consequences.
However, these two measures can be considered together with human judgment if
utility functions are used, as do Baron and Paté-Cornell (1999), Brito and de
Almeida (2009), Brito et al. (2010), Almeida-Filho and de Almeida (2010) and
Garcez and de Almeida (2014).
The concept of ALARP has been questioned by many authors in the literature.
Melchers and Stewart (1993) shows that each individual can have a different level
that he/she finds acceptable for different types of risk and also that this can change
also from one culture to another.
Aven and Vinnem (2005) presented a different risk analysis regime that is not
based on risk acceptance criteria at all. They argue that a rule based on cost-
effectiveness should do better than pre-defined risk acceptance limits. In some
situations, it is possible to achieve risks below ALARP levels. Thus a methodo-
logy is required that can consider a DM’s tradeoffs among costs and other loss
dimensions such as environmental and potential losses of life.
Aven and Kristensen (2005) presented a discussion on several perspectives of
risk, establishing a common basis for the different perspectives, emphasizing how
important it is to consider all possible consequences associated with their uncertainties.
When considering the context of risk analysis in the literature on the oil and gas
industry, there are two main models for risk management with a specific focus on
risk evaluation and risk reduction that can guide the selection and design decisions
by evaluating risk levels. Besides these models there are other models in the
literature that can be used after start-up at the facility, such as the framework
proposed by Øien (2001) for structuring risk indicators for risk control during
operation.
Khan et al. (2002) present an example of design selection in order to implement
safety measures. Khan et al. (2002) present an offshore oil and gas facility and
design alternatives that may reduce risk.
The approach of inherent safety design was presented initially by Kletz (1985),
and detailed later in Kletz (1998). Khan and Amyotte (2002) presented a study
showing that safety measures should be a concern from the design stage for a
facility in order to reduce costs throughout its life span.
322 Chapter 10 Design Selection Decisions
One of the main models for risk evaluation can be found in ISO/IEC Guide 51:
2014 and another in the NORSOK Standard Z-013. The model described in
ISO/IEC Guide 51: 2014 updates the 1999 version regarding this subject. It can
also be used combined with IEC Guide 73: 2009, which refers to the vocabulary
and meanings regarding risk management, so as to consolidate terminologies.
The NORSOK Standard Z-013 is a standard edited by the NPD (Norwegian
Petroleum Directorate), which is a Norwegian agency in charge of regulating oil
industry activities in the North Sea.
Brandsæter (2002) describes the implementation and uses of risk analysis using
quantitative and qualitative methodologies for the oil and gas offshore industry
with contributions to the EC-JRC International Workshop on “Promotion of
Technical Harmonisation on Risk-Based Decision Making (2000)”, which is
formatted as if it were a response to a set of questions prepared by workshop
organizers, and in which both models mentioned are discussed.
In ISO/IEC Guide 51: 2014, risk evaluation is defined as a wide process of
estimation and analysis. For ISO/IEC Guide 51: 2014, risk analysis terminology is
defined as the systematic use of information to identify hazards and estimate risk, and
risk estimation is defined as a procedure to determine if the risks are tolerable or not.
Thus the ISO/IEC Guide 51: 2014 model is represented by an iterative process
to evaluate and reduce risks that can be applied to qualitative and quantitative risk
evaluations. In ISO/IEC Guide 51: 2014, it is clear that some tolerance criteria
have to be defined, as in ALARP. However, it does not suggest any procedure to
deal with the situation where the limits have already been satisfied (or not) even
how to choose between non-dominated alternatives considering multiple risk
dimensions.
This iterative process considers that each hazard must be considered and must
satisfy a tolerable risk level. According to ISO/IEC Guide 51, it is necessary to
identify each hazardous situation and event by anticipating stages and conditions
for the system, including installation, operation, maintenance, repair and destruction/
disposal. This iterative process considers an entire process of risk assessment.
The ISO/IEC Guide 51: 2014 presents a “three-step method” from the design
phase and additional measures at the use phase. The risk reduction process starts
from the design of the installation, beginning with inherently safe design as a way
to start the risk reduction process.
Additional risk reduction alternatives have to be implemented after the design
stage such as training and procedures that will reduce residual risks after all
protective measures have been deployed.
The NORSOK Standard Z-013 model for the process of assessing risk and
emergency preparedness describes this in a similar way to that of the ISO/IEC,
although NORSOK includes an assessment of preparedness for emergencies in its
process.
10.3 Risk Evaluation for Design Selection 323
The previous version of the process for this standard already gave more
emphasis to a typical quantitative risk analysis methodology when considering the
risk assessment process. It emphasized the importance of the estimation, analysis
and evaluation approach for typical quantitative risk analysis methodologies when
applied to offshore oil and gas structures. It defines risk as a probability or an
expected frequency, and requires a risk acceptance criterion to be defined, which
should consider the probability or frequency of an associated consequence,
thereby establishing a risk index and an acceptable risk limit. For a human
dimension, for example, the risk for an individual can be used and/ or the FAR
(Fatality Accident Rate) which should be compared with acceptable limits by the
adopted standards in order to establish the risk picture.
Throughout the risk analysis process there is no specific framework about how to
aggregate preferences amongst multiple risk dimensions, especially when there is
a decision problem where some of the alternatives have already reached the
acceptable risk levels defined in standards.
Thus, all dimensions are considered in terms of constraints that must be
respected and considered in terms of cost benefit evaluation. Nevertheless, during
the design process, there are opportunities to improve safety and prioritize safety
alternatives.
To address this issue, this section presents a framework with a numerical
application that aggregates preferences by using metrics, which consider
probability, give values to human judgment on consequences and their behavior
with regard to risk (prone, neutral and risk averse). These metrics are provided by
Utility Theory.
Based on the literature, this procedure supports the structuring of decision
problems in the context of evaluating multidimensional risk, based on the frame-
work given in Chap. 2.
The approach given in this section addresses decision problems regarding risk
reduction and safety improvements for the design of hazardous facilities. These
kinds of decision problems can refer to a choice, ranking, sorting or a portfolio
decision problem, and depending on the type of problematic (Roy 1996) a specific
methodology should be used to aggregate the DM’s preferences and doing so
should be amongst its objectives.
While due consideration should be given to the models for risk estimation,
analysis and evaluation from both standards (NORSOK and ISO/IEC), this
MCDM/A procedure can be used as a framework to evaluate risk reducing
alternatives. This can be done even if some of these risks have already achieved
the acceptable risk levels and assuming there are still safety improvements that
324 Chapter 10 Design Selection Decisions
could be implemented. Also decisions can be taken about which risk reduction
measures should be implemented, according to the priority of each action and the
risk involved in different parts of the process.
This kind of decision has two main actors who, in general terms, can be
identified as a DM and an analyst who will give methodological support to the
DM. Both of these actors will exert influence during the decision process, in
which the former, a DM, will influence the decision because of cognitive aspects
and his/her preference structure. The analyst will influence it in such way that he
may bring bias to the process due to his/her own opinion about the subject and/or
because of the use of his/her preferred methodological approaches (Almeida-Filho
and de Almeida 2010).
Figure 10.2 presents the steps to consider an MCDM/A approach in order to
choose or prioritize design alternatives for risk reduction based on the procedure
for building MCDM/A model presented in Chap. 2.
Design
Risk Aggregation
alternatives
Analysis procedure
evaluation
Results Sensitivity
Analysis analysis
It starts with a Decision Situation that represents the phase when a decision
problem has been identified or has appeared and needs to be outlined. This
includes determining the risk level achieved and the risk reduction alternatives to
be considered in design. The step ahead is to define the kind of problematic (Roy
1996; Vincke 1992) that addresses the problem itself.
10.3 Risk Evaluation for Design Selection 325
This MCDM/A framework allows the DM not only to use acceptance levels as
references but also to evaluate them according to his/her preferences and risk
behavior (prone, neutral, averse).
Alternative Action A B C D E
1 No feature implemented 0 0 0 0 0
2 Implement E 0 0 0 0 1
3 Implement D 0 0 0 1 0
4 Implement D and E 0 0 0 1 1
5 Implement C 0 0 1 0 0
6 Implement C and E 0 0 1 0 1
7 Implement C and D 0 0 1 1 0
8 Implement C, D and E 0 0 1 1 1
9 Implement B 0 1 0 0 0
10 Implement B and E 0 1 0 0 1
11 Implement B and D 0 1 0 1 0
12 Implement B, D and E 0 1 0 1 1
13 Implement B and C 0 1 1 0 0
14 Implement B, C and E 0 1 1 0 1
15 Implement B, C and D 0 1 1 1 0
16 Implement B, C, D and E 0 1 1 1 1
17 Implement A 1 0 0 0 0
18 Implement A and E 1 0 0 0 1
19 Implement A and D 1 0 0 1 0
20 Implement A, D and E 1 0 0 1 1
21 Implement A and C 1 0 1 0 0
22 Implement A, C and E 1 0 1 0 1
23 Implement A, C and D 1 0 1 1 0
24 Implement A, C, D and E 1 0 1 1 1
25 Implement A and B 1 1 0 0 0
26 Implement A, B and E 1 1 0 0 1
27 Implement A, B and D 1 1 0 1 0
28 Implement A, B, D and E 1 1 0 1 1
29 Implement A, B and C 1 1 1 0 0
30 Implement A, B, C and E 1 1 1 0 1
31 Implement A, B, C and D 1 1 1 1 0
32 Implement all features 1 1 1 1 1
10.3 Risk Evaluation for Design Selection 329
The consequence evaluation considered the most credible scenarios for these
units (Khan et al. 2002) are summarized in Table 10.3.
Possibility of Spills
Units
The most credible scenario for the first separator is a BLEVE followed by fire;
for the second separator it is VCE followed by fire; for the compressor units it is a
gas release possibly turning into a jet fire; for the Flash Drum unit it is a VCE
followed by fire; and for the Drier unit the most credible scenario is a BLEVE
followed by fire.
These scenarios, which are the most credible ones, have a higher probability in
the present situation and a lower probability after implementing safety features,
for each scenario (Khan et al. 2002). These probabilities are illustrated in Table 10.4.
is used due to the nature of the utility measure, which is based in an interval scale,
so what really matters is the size of the utility difference ratio instead of absolute
differences between them.
Difference Ratio
Rank Alternative
( ui u /u u
i 1 i 1 i 2
)
1 5 0.65
2 7 1.65
3 6 12.07
4 13 0.08
5 8 2.45
6 15 1.70
7 1 0.28
8 21 5.09
9 3 2.25
10 14 0.07
Thus the last column of Table 10.5 shows that the difference between
alternative 5 and 7 is 65.39% bigger than the difference between alternative 7 and
6, and that the difference between 8 and 15 is 245.48% bigger than the difference
between 15 and 1 (implement none of safety features). This measure gives to DMs
a clear idea about the difference of these alternatives considering probabilities,
consequences, their individual preferences and their behavior towards risk.
By conducting a sensitivity analysis, it was possible to observe that the first
positions of the ranking would not change if the values chosen for the scale
constants were changed. It is also interesting to highlight that according to the
results, safety will be improved, whereas the alternative that represents no invest-
ment in safety appears only in 7th position in the ranking.
From the perspective of the maintenance function, redesign is the action that is
done when the status quo is not acceptable. Some reasons for that are: higher
performance requirements due to competitiveness; more conservative standards
related to environment and safety; and more severe degradation not covered by the
initial design.
332 Chapter 10 Design Selection Decisions
References
Alencar MH, Cavalcante CAV, de Almeida AT, Silva Neto CE (2010) Priorities assignment for
actions in a transport system based on a multicriteria decision model. In: Bris R, Soares CG,
Martorell S (eds) European safety and reliability conference, Prague, September 2009.
Reliability, Risk, and Safety: Theory and Applications, Vol. 1-3. 2009. Taylor and Francis,
London, UK, p 2480
Almeida-Filho AT de, de Almeida AT (2010) Multiple dimension risk evaluation framework. In:
Bris R, Soares CG, Martorell S (eds) European safety and reliability conference, Prague,
September 2009. Reliability, Risk, and Safety: Theory and Applications, Vol. 1-3. 2009. Taylor
and Francis, London, UK, p 2480
Aven T, Kristensen V (2005) Perspectives on risk: review and discussion of the basis for
establishing a unified and holistic approach. Reliab Eng Syst Saf 90:1–14
Aven T, Vinnem JE (2005) On the use of risk acceptance criteria in the offshore oil and gas
industry. Reliab Eng Syst Saf 90:15–24
Baron MM, Paté-Cornell ME (1999) Designing risk-management strategies for critical engineering
systems. Eng Manag IEEE Trans 46:87–100
Brandsæter A (2002) Risk assessment in the offshore industry. Saf Sci 40:231–269
References 333
Kletz TA (1998) Process plants: A handbook of inherently safer design. 2nd ed, Taylor &
Francis, Philadelphia
Lewis EE (1987) Introduction to reliability engineering. Wiley, New York
Li W, Zuo MJ (2008) Optimal design of multi-state weighted k-out-of-n systems based on
component design. Reliab Eng Syst Saf 93(11):1673–1681
Lopes YG, de Almeida AT, Alencar MH, Wolmer Filho LAF, Siqueira GBA (2010) A Decision
Support System to Evaluate Gas Pipeline Risk in Multiple Dimensions. In: Bris R, Soares
CG, Martorell S (eds) European Safety and Reliability Conference (ESREL), Prague, Czech
Republic, 2009. Reliability, Risk and Safety: Theory and Applications. CRC Press-Taylor &
Francis Group, p 1043
Martello S, Toth P (1990) Knapsack problems: algorithms and computer implementations. John
Wiley & Sons, Chichester
Melchers RE, Stewart MG. (1993) Probabilistic risk and hazard assessment. Balkema, Rotterdam
Moubray J (1997) Reliability-centered maintenance. Industrial Press Inc., New York
NORSOK (2010) NORSOK Z-013: Risk and emergency preparedness analysis. Rev. 2,
Norwegian Technology Centre
O’Connor P, Kleyner A (2012) Practical reliability engineering. John Wiley & Sons, Chichester
Øien K (2001) A framework for the establishment of organizational risk indicators. Reliab Eng
Syst Saf 74:147–167
Polovko AM, Pierce WH (1968) Fundamentals of reliability theory. Academic press New York
Rathod V, Yadav OP, Rathore A, Jain R (2013) Optimizing reliability-based robust design model
using multi-objective genetic algorithm. Comput Ind Eng 66:301–310
Ren Y, Bechta Dugan J (1998) Design of reliable systems using static and dynamic fault trees.
Reliab IEEE Trans 47:234–244
Roy B (1996) Multicriteria Methodology for Decision Aiding. Springer US
Sachon M, Paté-Cornell E (2000) Delays and safety in airline maintenance. Reliab Eng Syst Saf
67(3):301–309
Sahoo L, Bhunia AK, Kapur PK (2012) Genetic algorithm based multi-objective reliability
optimization in interval environment. Comput Ind Eng 62:152–160
Stapelberg RF (2009) Handbook of reliability, availability, maintainability and safety in engineering
design. Springer-Verlag, London
US MIL-STD-785B (1980) Reliability Program For System and Equipment Development and
Production, US Military Standard
Vanem E, Endresen Ø, Skjong R (2008) Cost-effectiveness criteria for marine oil spill preventive
measures. Reliab Eng Syst Saf 93:1354–1368
Vincke P (1992) Multicriteria Decision-Aid. John Wiley & Sons, New York
Vrijling JK, van Hengel W, Houben RJ (1998) Acceptable risk as a basis for design. Reliab Eng
Syst Saf 59:141–150
Yanasse HH, Soma NY (1987) A new enumeration scheme for the knapsack problem. Discret
Appl Math 18:235–245
Chapter 11
Decisions on Priority Assignment
for Maintenance Planning
11.1 Introduction
Among the decisions regarding maintenance planning, one of the most important
decisions is related to define which kind of maintenance actions are more
appropriated. This decision involves subjective and technical aspects in order to
evaluate the consequences of failures. This chapter presents MCDM/A models
considering the assignment of priorities before establishing a maintenance plan.
A maintenance plan can be defined with different approaches. Selective
maintenance is an example for building a maintenance plan. This approach
includes the specification of each action that should be done for each item in a
multicomponent system, for an interval longer than one cycle, observing the
constraints for optimizing a single objective. Originally, this problem was formulated
considering a fixed time window (Lust et al. 2009).
Two important aspects of the traditional RCM approach are presented in this
subsection in order to provide a better understanding of the MCDM/A model
built: the procedure steps and the evaluation of failure consequences.
From the twelve steps introduced in Chap. 3, Moubray (1997) emphasizes these
following steps:
x Establish the functions of each asset within the operating context considering
the associated desired standards of performance;
x Define failures that may occur in the physical asset;
x Identify the failure modes;
x The fourth step involves checking the effects of failure;
x The fifth step is to verify and analyze the consequences of failure;
x Finally, the last step is to establish maintenance actions that could be verified
by applying two techniques: proactive tasks and default actions.
Additionally, RCM approach classifies the consequences of failure into four
categories (Moubray 1997):
x Hidden failure consequences: when it does not present a direct impact, but can
expose the organization to multiple failures with serious consequences
(including catastrophic);
x Safety and environmental consequences: when it presents safety consequences
considering the possibility of injury or death. The environmental consequences
might mean that an organization has violated a national or international
environmental standard;
x Operational consequences: failures which affects only production;
x Non-operational consequences: failures in this category do not affect either
production or safety, involving only the direct cost of repair.
According to the facilities analyzed, a failure could produce irrelevant con-
sequences or compromise essential systems for the organization or society or
safety. In RCM approach, the consequences are evaluated by verifying the impacts
338 Chapter 11 Decisions on Priority Assignment for Maintenance Planning
Some terms are relevant to build the model such as: the observed context, the
availability of information and its degree of accuracy, the rationality required the
DM’s preference structure and the problematic. An important aspect is the rationality
for the DM in the problem under study that involves a non-compensatory or
compensatory approach. In this sense, the decision model presented in this
subsection sets out to improve the RCM approach by incorporating contributions
from MAUT.
According to de Almeida (2007), MAUT considers that the DM’s preferences
are modeled for computing the MAU function in which the aggregation of uni-
dimensional utility functions must respect MAUT axiomatic structure. Additionally,
Brito and de Almeida (2009) state that MAUT can be applied to aggregate valued
preferences and uncertain consequences related with multiple criteria, providing
results that can be used as input in the process of maintenance management.
The stages of this RCM MCDM/A model are shown in Fig. 11.1. Traditional
RCM steps that remains are: define the functions of assets; identify functional
failures; define failure modes; identify the effects of failure; and, establish
maintenance actions. Therefore, this subsection focuses mainly on the MCDM/A
model built for evaluating the consequences of failures.
For each objective defined, a dimension of consequences is proposed, representing
the objectives of this decision model. Thus, a set of consequence dimensions are
established.
The consequences of failures are evaluated based on five categories defined as
the dimensions of the consequences, in which some of the characteristics differ
from those established by the traditional RCM approach, as follows:
x Human dimension (h): considers the damage with respect to people affected by
the consequences of failures;
x Environmental dimension (e): considers the area affected due to a failure;
x Financial dimension (f): considers the financial losses due to a failure;
x Operational dimension:
– Operational dimension I (o’): considers failures that do not interrupt the
producing system operation;
– Operational dimension II (o’’): considers failures that interrupt the
producing system operation.
11.2 An MCDM/A Model for the RCM Approach 339
Defining the
functions of
physical assets
Establishing
functional
failures
Identifying
failure modes
Establishing
failure effects MAUT quantitative stages (Failure
consequences):
x Establishing consequences
Defining failure dimensions;
consequences x Analyzing the consequences;
x Probabilistic modeling;
x Computing the overall indices
of utility for each failure mode;
x Ranking the alternatives;
Identifying
maintenance
actions
U (T , ai ) ¦ P(c T , a )U (c)
c
i (11.1)
Finally, (11.2) shows the utility function of these consequences for continuous
cases.
U (T , ai ) ³ P(c T , a )U (c)dc
i (11.2)
c
U (h, e, f , o' , o' ' ) k1U (h) k 2U (e) k3U ( f ) k 4U (o' ) k5U (o' ' ) (11.3)
where:
ki is a scale constant that represents the value of the tradeoff;
When the DM’s preferences require limiting tradeoff effects, a model that
considers veto can be incorporated (de Almeida 2013).
The final results are presented by the obtained ranking, established by multi-
attribute utility values found for each failure mode.
The interval scale of utility function allows the incremental value to be
compared to the failure modes Keeney and Raiffa (1976). Applying the interval
scale, it may be affirmed that the difference U(MFx)ȕx – U (MFy)ȕx+1 is M times
greater than the difference U(MFy)ȕx+1 – U (MFz)ȕx+2. This can be seen from the
increment ratio IR of these differences since IR = (U(MFx)ȕx –
U(MFy)ȕx+1)/(U(MFy)ȕx+1) – U(MFz)ȕx +2).
For this illustrative example, 16 failure modes are considered, FMx, x=1,2,…,16.
Each FMx is associated with human (h), environmental (e), financial (f),
operational I (o’) and operational II (o’’) consequence dimensions (Alencar and de
Almeida 2011).
There is a prior probability ʌ(șx) associated to each FMx, as can be observed
from Table 11.1.
11.2 An MCDM/A Model for the RCM Approach 341
The scale constants k1=0.19, k2=0.13, k3=0.27, k4=0.11, k5=0.30, are elicited
from the DM adopting structured protocols (Keeney and Raiffa 1976).
The interval scale of the utility function allows comparison of the differences in
utility among failure modes. These differences are verified in Table 11.2 (fourth
column).
Ranking position (ȕx) Failure Mode FMi U(FMx)ȕx U(FMx)ȕx - U(FMy)ȕx+1 Difference ratio
ȕ01 FM15 0 0.08788 0.51200
ȕ02 FM9 0.08788 0.17164 4.15593
ȕ03 FM1 0.25952 0.0413 0.40249
ȕ04 FM3 0.30082 0.10261 1.46064
ȕ05 FM5 0.40343 0.07025 0.85598
ȕ06 FM7 0.47368 0.08207 0.70308
ȕ07 FM14 0.55575 0.11673 14.70151
ȕ08 FM8 0.67248 0.00794 0.41966
ȕ09 FM11 0.68042 0.01892 1.09745
ȕ10 FM4 0.69934 0.01724 0.19334
(continued)
342 Chapter 11 Decisions on Priority Assignment for Maintenance Planning
The values presented in Table 11.2 provide important information for the DM.
The difference between the values of the utilities associated with the failure modes
FM14 and FM8 is 0.11673; and the difference between the values of the utilities
associated with the failure modes FM8 and FM11 is 0.00794. The ratio among
differences (fifth column) allows the DM to understand the relative difference
among each FM quantified by the utility scale.
This measure allows to state that the relative difference between FM14 and FM8
is approximately 15 times greater than the difference between FM8 and FM11. It is
important to highlight that these values presented in Table 11.2 reflects the DM’s
preferences among four consequence dimensions. From the numerical example
given, it is possible to observe how undesirable is the differences among such
failure modes for DM.
This section presents a model for identifying critical devices, which classifies the
items from an industrial plant into predetermined category of criticality. Using the
general procedure proposed in Chap. 2, an MCDM/A model is introduced. For
344 Chapter 11 Decisions on Priority Assignment for Maintenance Planning
simplification purposes, only some steps of the Chap. 2 procedure are highlighted
in the model presentation.
Maintenance planning requires a thorough understanding of the system and of
the goals, as well as of the consequence dimensions associated with the failure of
the item. Deciding “what to do” may be based on technical, environmental
and financial aspects. Most of the models try to establish a severity index for
representing a measure in different dimensions in order to support DM for
deciding “what to do”. The weakness of these approaches is that DM’s preferences
are not considered when building such measures.
Facing a large set of piece of equipment, the DM seeks to organize this set into
classes of criticalities. This classification helps the DM to specify the most
appropriated set of actions for each class, considering the adequate resources to be
deployed in a more effective way.
For example, in a power distribution network there are several similar items,
however their location in the network, despite its similarity adds the specific
branch characteristics resulting in different criticality levels, which will point for
different maintenance actions or polices depending on this specific item.
Depending on the item which fails, multidimensional consequence will arise. The
specific location of the item can characterize the number of affected customers,
public services, and result in losses for business supplied in the distribution branch
affected. Depending on the kind of failure, safety aspects also may arise. If such
failure occurs in an underground distribution network, for example, and has the
potential to cause explosions in a high density area, such as in large cities (Garcez
and de Almeida 2014a, Garcez and de Almeida 2014b).
Considering the specific problem presented in this section for assigning devices
into priority classes, by means of an MCDM/A sorting model, based on
ELECTRE TRI, using information about the characteristics of device and the
multidimensional consequences associated with its failures.
ELECTRE TRI was designed to describe actions in ordered categories, it
enables a set of alternatives pre-defined and ranked categories based on multiple
criteria to be classified, as illustrated in Fig. 11.2, where each device xi (x = 1 … n)
is classified according to the device’s criticality.
X1 X2 X3 X1 X2 X9 X10 X5
X4 X5 X3 .. X6
X6 .. X7 X8
X7 X8
Description Scale
Catastrophic consequence 5
Major consequence 4
Severe Consequence 3
Minor Consequence 2
Trivial Consequence 1
Description Scale
Loss of more than 20,000 monetary unities 5
Loss of 15,001 a 20,000 monetary unities 4
Loss of 10,001 a 15,000 monetary unities 3
Loss of 5,001 a 10,000 monetary unities 2
Loss of 0 a 5,000 monetary unities 1
346 Chapter 11 Decisions on Priority Assignment for Maintenance Planning
Description Scale
Failed more than 15 times on interval 5
Failed from 12 to 15 times on interval 4
Failed from 8 to 11 times on interval 3
Failed from 4 to 7 times on interval 2
Failed from 0 to 3 times on interval 1
Description Scale
Mean Delay-time of 0 a 10 time unities 5
Mean Delay-time of 11 a 20 time unities 4
Mean Delay-time of 21 a 30 time unities 3
Mean Delay-time of 31 a 40 time unities 2
Mean Delay-time bigger than 40 time unities 1
Description Scale
Almost Impossible detection 5
Difficult detection 4
Moderate detection 3
Easy detection 2
Immediate detection 1
Alternatives\Criterion g1 g2 g3 g4 g5
x1 1 1 2 1 3
x2 4 5 1 3 4
x3 3 2 3 4 2
x4 3 4 3 4 1
x5 5 5 1 5 1
x6 4 3 2 4 3
x7 1 2 5 2 2
(continued)
11.4 Modeling a Problem for Identifying Critical Devices 347
x8 2 3 5 3 5
x9 1 1 3 2 2
x10 2 3 4 5 3
Besides the scale defined for each criterion, the definition of the parameters of
the preference functions in order for the ELECTRE TRI. It was defined that the
values of the limits of preference and indifference are both equal to 0. Disregarding
the veto threshold and setting Ȝ = 0.5.
Related to the intra-criteria evaluation, the weights defined by the DM are
given by Table 11.9.
Table 11.9 The weights of the criteria
Criterion Weight
g1 0.25
g2 0.35
g3 0.18
g4 0.12
g5 0.1
In this study, five categories are considered, ordered according to their degree
of importance concerning the priority of planning and conduct of maintenance
actions. The classes considered are:
x Highly critical device: the occurrence of a fault in any device belonging to this
class will bring serious damage to the organization;
x High priority devices;
x Intermediate priority devices;
x Low-priority devices;
x Extremely low-priority devices: i.e., one can, in a way, neglect the maintenance
of equipment belonging to this class so as to direct more concentrated efforts to
the most critical equipment;
The equivalence classes serve as standards by which the devices will be
classified. The equivalence classes adopted for this study are defined by lower and
upper (“Profiles”) limits, as shown in Table 11.10.
Table 11.10 Classes of equivalence and their lower and upper limits
C3 2.5 3.5
C4 1.5 2.5
C5 - 1.5
Analyzing the results from Table 11.11, only one device was sorted as
extremely critical (x5). This is due to the fact that, if a failure occurs in this device,
there will be catastrophic losses for the company, with regard to matters relating to
the financial, human and environmental dimensions; and besides, this device has a
short delay-time, which deserves careful attention.
The simulation is therefore useful since it enables the manager of maintenance
to drive the maintenance actions in such a way as to focus on the most critical
devices, while it reinforces that equipment considered less important can be neglected.
The impact from the application of an MCDM/A approach on the maintenance
management process may be reflected in the improved operating performance of
the device, due to more efficient maintenance planning for each class of device
having been adopted.
References
12.1 Introduction
Fig. 12.1 World map of publications on the use of MCDM/A in maintenance and reliability
research
Figure 12.2 shows that there has been an exponential trend in citations which
explains the growth of publications on this subject. The ever increasing amount of
publications indicates the relevance of the topic and the perspectives in the area.
14 12 12
11
12
9
10
7
8 6
5 5
6 4 4 4
3
4 2 2 2 2 2 2
1 1 1 1 1 1 1
2
0
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
Year
Fig. 12.2 Number of articles per year on MCDM/A in maintenance and reliability research
Furthermore, the 170 articles considered until 2012 had received 4,306
citations from 1996 to 2013 according to the Scopus database, which represents an
average of 25.33 citations per paper. Fig. 12.3 reflects the impact of this research
area, measured by citations per year since 1996 (de Almeida el al. 2015). For
instance, the articles received 831 citations in 2012.
12.2 Location of Backup Units in an Electric System 353
576
600
500 453
400
315
279
300
200 161
92
100 44 55
8 8 12 15 18 14 23
0
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
Year
Fig. 12.3 Number of citations per year on MCDM/A in maintenance and reliability research
From this perspective, this chapter presents RRM problems that arise in
different particular contexts not mentioned in previous chapters.
Amongst the many specific problems that require an MCDM/A approach, this
chapter illustrates RRM problems that require appropriate modeling in order to
allow a DM to consider multidimensional consequences.
Thus, this chapter covers topics related to the following problems:
x Location of backup units;
x Sequencing of maintenance activities;
x Natural disasters;
x Reliability in power systems;
x Integrated production and maintenance scheduling;
x Maintenance team sizing;
x Reliability acceptance testing.
the place of the service. When the maintenance team is ready to conduct the
service, there is an elapse of time from service startup to completion, and then the
asset returns to operational status.
In many practical applications, is possible to observe that the maintenance time
needed for a repair has been considered one of the most relevant. On the other
hand, when the resources available to perform maintenance are scarce, some other
factors need to be addressed to minimize the elapsed time for corrective
maintenance actions. In addition to this scarcity, the location of some assets may
be geographically dispersed and may have a strong influence on the time needed
to conduct the maintenance actions.
The example given in this section for location of backup units considers the
context of an electric power distribution company. For this kind of companies,
there are several geographically dispersed systems throughout the distribution
network. Given the geographically dispersion, the maintenance function has to
overcome the logistics obstacles to cope with the required performance standards
that require the availability of equipment located in each of the electric power
substations distributed along the network.
The equipment considered includes high-tension power transformers, which are
heavy, expensive assets and have a high useful life. Such equipment costs millions
of dollars and has a useful life around 30 years. Moreover, the lead-time for
ordering such equipment can take several months, besides the time referring the
logistics of installing the equipment.
Although expensive equipment with a low failure rate cannot justify investment
in a high amount of redundancy for electric power substations, the best use of
resources should be evaluated to deal with an emergency, especially when the
impact on system unavailability is high.
In the case of power transformers, it is known that a limited number of backup
units are available to electric power substations for possible replacements due to a
failure. The decision problem is to define the locations of backup transformers to
minimize the overall consequences of a failure and the need for emergency
replacement.
The consequences of these equipment failures can be characterized as in Chap. 1,
when considering service producing systems. In this particular case, the number of
users affected can escalate from thousands to millions, depending on the con-
sequences of the failure and the blackout effects. Such consequences vary
depending on the equipment location, similar to the example given in Chap. 11
regarding the implications to the priority assignment for maintenance planning
decisions.
To prevent and minimize such consequences, it is essential to plan the location
of backup units in order that it can quickly restore the system in case of failure.
Therefore, the location of backup transformers involves many factors, which
directly influence the operation of a power distribution system. Such process
involves objectives beyond costs related to service interruptions prolonged due to
the absence of a backup unit. These factors have a direct influence on system’s
12.2 Location of Backup Units in an Electric System 355
availability and maintainability (de Almeida et al. 2006; Ferreira et al. 2010;
Ferreira and Ferreira 2012).
Thus, the decision model for this problem seeks to ensure that customers are
minimum affected by the inconvenience of service interruption and the losses
associated. Although a similar problem structure is addressed in classical facility
location (Drezner and Hamacher 2004), different objectives are considered.
A failure in each particular location affects multidimensional consequences.
For this particular problem, three dimensions are considered: number of customers,
health services and local economy affected.
Brandeau and Chiu (1989) give an overview of location problems, which have
been previously studied, with the emphasis on models that were developed in the
field of operations research, formulated as optimization problems, such as the
p-median. The p-median is a classical problem in the field of combinatorial
optimization problem. An algorithm with re-optimization procedures for multi-
objective combinatorial optimization problems is proposed by Bornstein et al.
(2012).
The decision model consists of an MCDM/A p-median model based on three
criteria:
x The size of the population (popi);
x Degree of industrialization (indi);
x The extent of health services (hsi).
The p-median model was adapted to minimize three criteria as shown in (12.1).
The distance factor (dij) represents the distance amongst the electric power
substations. The distance factor works as a multiplier weight in relation to pop, ind
and hs.
ns , (12.1)
¦x
j 1
jj nb
xij d x jj ; i, j N
xij {0,1}; i, j N
where:
K1, K2, K3, and K are scale constants related to the respective attributes;
N is a set of electric power substations, N = {1, …,ns};
nb is the number of back-up transformers;
popi is the size of population served by the substation i;
indi is the degree of industrialization served by the substation i;
hsi is the extent of health services served by the substation i;
356 Chapter 12 Other Risk, Reliability and Maintenance Decision Problems
xij is a decision matrix variable, where xij = 1 if the backup transformer of the
substation i is allocated to substation j, and xij = 0, otherwise. And xjj=1 if the
substation j is allocated to store a back-up transformer (a median) and xjj=0,
otherwise.
For the specific application, regarding the inter-criteria evaluation all attributes
are found to be preference independent, except health service and population
criteria. Thus, DM’s preferences over health are affected when varying values of
the size of the population. The model, which represents these conditions, corresponds
to a multi-linear model expressed in the MAU function, given by (12.1).
The objective function represents the utility for substation k, if the backup is
located in substation j. In order to have an indicator for the location of the backup
in substation j, the sums of the utilities of location for all the electric power
substations k should be calculated.
By calculating the maximum utility for each substation, the recommendation
will be to locate a backup transformer in the power substation for which this
transformer will provide the highest maximum utility.
The considered company has to decide which substation should host a backup
transformer from 19 options considered.
As to the multi-attribute p-median model, the results are shown in Fig. 12.4.
The maximum value of the multi-attribute function for nb = 6 is 17.071. The
constant scales represent the preferences of the manager of the current project.
The parameters K1 = 0.2, K2 = 0.5, K3 =0.2, K = 0.1 were obtained using a
structured process suggested by Keeney and Raiffa (1976). The electric power
substations chosen are S = {3, 6, 11, 8, 13, 19}.
1
3 2
8 7
9
6 4
11
13 15 10
5
14
12 16
19
17
18
Fig. 12.4 Example of a solution of the multi-attribute p-median model (S = {3, 6, 11, 8, 13, 19})
12.3 The Sequencing of Maintenance Activities 357
The maintenance database for this power distribution network is updated from
data from inspections scheduled on a calendar that covers the entire power
distribution network over a period of ten years, with periodical activities that take
place at one, two, five and then ten year intervals.
The maintenance strategy adopted is that of immediately restoring the system
when a failure that disrupts the service supply occurs or even if this failure is
reported as not disrupting the service but it does expose the population to risk.
The maintenance culture adopted in this context is similar to that described by
Moubray (1997), who set out three typical states related to an equipment mode:
the normal state, a defect and a failure. Depending on the effect of a failure on the
functioning of the item, the failure is classified either as:
x A potential failure, which is an observable condition, which implies there will
be a functional failure if no preventive action is taken (Moubray 1997);
x A functional failure, which means the inability of an item of equipment to
perform a specific function within desirable operational limits (Moubray 1997).
The focus of this problem on sequencing maintenance activities is to do with
the potential failures, which are identified by inspections included in the calendar
and recorded on the maintenance information system. Thus, these potential
failures are prioritized to avoid a disruption to the service and its consequences for
strategic and operational objectives. The sequence of maintenance services is
based on MCDM/A, which define the order among services to be performed.
In the particular problem addressed, there were about 25 thousand potential
failures identified by inspections. Given the capacity of the maintenance division
workforce and the annual budget set aside for preventive maintenance, only four
thousand potential failures can be tackled per year. The practical meaning is that
potential failures are sequenced taking their priority into account and are corrected
over the year, which leaves a set of potential failures to be reevaluated by
maintenance services in the following year together with other new potential
failures identified in inspections undertaken as per the inspection calendar. This
preventive maintenance budget is usually defined in the sense that the potential
failures identified and pending repairs are still at a tolerable level in the light of
the organizational targets. There are several models regarding the definition of
preventive maintenance time interval (Jiang and Li 2002; Shafiee and Finkelstein
2015), although there are practical situations when the DM has to consider also the
resources available and the production scheduling to define the exact time of
a maintenance repair or replacement. Sect. 12.6 illustrates this problem with
a decision model.
Some of these organizational targets are defined in order to meet regulatory
aspects, such as those defined by ANEEL, which is the Brazilian government
agency responsible for regulating the generation of electrical power, its
transportation and the distribution companies involved. This agency defines
operational and service levels for these companies, and has the power to levy fines
in accordance with the regulatory rules. Another important issue is that the
12.3 The Sequencing of Maintenance Activities 359
1 k
S ( a, b) ¦ w j Pj (a, b) (12.2)
W j1
360 Chapter 12 Other Risk, Reliability and Maintenance Decision Problems
Equation (12.3) represents the net outranking flow, which consists of the
difference between the positive and the negative flow of an alternative a. Based on
the net outranking flow, a complete pre-order is provided that ranks all
alternatives.
1 ªn n º
I (a) «¦ S (a, b) ¦ S (b, a )» (12.3)
n 1 ¬« bbz1a b 1
bza ¼»
After inputting the MCDM/A model parameters into the required fields for the
decision model and performing the PROMETHEE II method, the sequencing of
the maintenance orders is obtained by considering their priority according to the
set of criteria under which all preventive maintenance orders were evaluated.
The DM can also generate reports on preventive maintenance orders which
takes account of the budgetary constraints and of an analysis of the sensitivity
analysis report. The DSS also enables the DM to perform a scenario analysis
supported by graphs so that he/she can compare the effectiveness of each action
while considering costs and the managerial objectives (DEC and FEC), and can
evaluate the cost levels incurred to carry out preventive maintenance orders as
against losses from potential failures (such as, in revenue and because of fines),
these being the consequences if the failure became a functional one.
It is interesting to observe that some prioritized maintenance actions may
not prove to be financially effective. However, they do prevent losses in other
dimensions, such as service quality, which is monitored by the regulatory agency
(ANEEL) or any special clients affected, for example. This illustrates the
importance of considering the MCDM/A nature of such problems; if this is not
done, these factors would not be given appropriate consideration.
Pine (2009) states that the changes in disasters frequency may be the result of
natural climatic variations that occur over a time period or arise from changes of
variables that impact the frequency or severity of environmental change. The
intensification of human activity in hazardous areas such as the construction of
residences without planning permission on hills subject to landslides or on land
known to be occasionally subject to severe flooding are examples. Additionally,
changes to the environment (such as those caused by buildings, technology and
the infrastructure to support human habitation) that lead to the degradation of
natural systems can also increase the severity of the hazard.
Thus, it is of fundamental importance to consider the dynamic nature of risk
and vulnerability. Karimi and Hüllermeier (2007) reinforce this idea by stating
that, due to there being all manner of uncertainty types, evaluating the risk of
losses due to natural disasters is a complex activity, mainly for lack of sufficient
physical knowledge and inadequate statistical data with respect to the origin,
characteristics and consequences of each disaster that has actually taken place.
Bobrowsky (2013) states that risks related to natural hazards and climate
change are not autonomous or externally generated. Therefore, society ought to be
able to react, adapt or respond to them. These risks are the result of the interaction
between society and the natural or built environment. Consequently, risk
management requires a better understanding of this relationship and the factors
influencing it.
Unlike the aspects considered in traditional risk management, in natural
disaster environments, some additional concepts are important so that DMs can
evaluate the situation more adequately. Hence there is a need to consider aspects
such as vulnerability and resilience, besides the concept of risk. The concepts of
risk, vulnerability and resilience are important in studies on natural disasters, and
are used as an approach to understand the dynamics of natural disasters (Paul
2011).
According to Field et al. (2012), vulnerability is the result of different conditions
and processes, which must include considering historical, social, political, cultural,
institutional and environmental matters and natural resources. Resilience is
defined in the context of natural disasters as a means of promoting sustainable
livelihoods, which enables individuals or systems to be able to cope with an
extreme event without using all the available resources (Paul 2011). Resilient
systems tend to reduce physical damage, thereby providing time for the environ-
ment to recover after an extreme event has occurred. Therefore, it reflects the
interest in improving the capacity of human and physical systems to respond to
natural events that occur.
According to Field et al. (2012), the risk of a disaster can be understood as the
possibility of adverse effects in the future arising from interaction between social
and environmental processes, and a combination of physical hazards and
vulnerabilities in the exposed elements. The simultaneous consideration of risk,
vulnerabilities and dynamic changes in the different phases of crises and disasters
produces a complex scenario out of which the degree of risk and vulnerability that
364 Chapter 12 Other Risk, Reliability and Maintenance Decision Problems
this contains needs to be identified and assessed, as should the measures that need
to be taken to mitigate risk and to adapt strategies. An understanding of extreme
events and disasters is a prerequisite for drawing up adaptation strategies related to
climate change and reducing risk in disaster risk management.
Natural disasters manifest themselves independent of the pre-existing states of
economic, social and physical environmental. Therefore, infrastructure, services
and organizations are prone to being affected by an event triggered by a natural
phenomenon (such as an earthquake or a flood) or a technical event like an
explosion or gas leakage (De-Leon 2006; Guikema 2009). Thus, it is observed that
a disaster is preceded by at least two aspects: the possibility that an initiated event
occurs, usually termed a danger from this potential state, and; a pre-existing
vulnerability. In other words, there is a predisposition for people, processes, infra-
structure, services, organizations or systems to be affected, damaged or destroyed
when an event occurs.
Considering the coupled with vulnerability and danger as a prerequisite of there
being a risk of a disaster, the exposure can be considered as another prerequisite.
The exposure is understood as the number of people and/or other elements at risk
that may be affected by a particular event (Thywissen 2006). Among other
definitions, risk must be understood as a function of hazard, vulnerability,
exposure and resilience.
Another issue discussed in studies related to natural disasters is consequence
analysis. The occurrence of an event or combination of two or more events can
cause different impacts in different dimensions.
The potential effects of climate change on natural hazards are an input to the
formulation of strategies to adapt risk management practices using knowledge
developed about the risks associated with people and with economic impacts
(Zischg et al. 2013).
Extreme natural events can induce higher losses especially when they occur in
vulnerable and/or areas that are densely populated (Huttenlau and Stotter 2011).
Risk analysis in natural hazards is used for estimating the consequences in order to
provide information to the public and DM. It is observed in this type of analysis
that different perceptions of the concept of risk are seen to have different goals
and to take different approaches, as given in Chap. 3, there are different
approaches when considering individual or social risks. Therefore, depending on
the approach given to the risk evaluation, different perspectives may be considered
and depending on how the consequences are evaluated an MCDM/A approach is a
more suitable manner to address the problem of evaluating multidimensional
consequences. Considering the complexity of consequence evaluation activity,
MCDM/A approaches allow to include different types of losses.
There are some aspects that may be considered for measuring the effects of
natural disasters, amongst which are economic and social disruption and
environmental impacts. Social disruption can, for example, include the number of
people made homeless or the incidence of crime such as the number of homicides,
arrests, the extent of civil disorder, including riots and street-fighting (Pine 2009).
12.4 Natural Disasters 365
Girgin and Krausmann (2013) highlight that Natech risks is likely to be more
frequent in the future due to industrial growth, changes in the patterns of
occurrence (due to climate change) of natural disasters and the fact that society is
becoming increasingly vulnerable, the more interconnected it becomes –
something that is happening with each passing day.
In conclusion, organizations and countries should regard the search for more
effective risk management as a fundamental goal. Adequate control of risk and
monitoring should be done in the best way as possible, thereby minimizing the
occurrence of catastrophic consequences.
To illustrate how an MCDM/A model may address such aspects, the following
section structures an MCDM/A model for multidimensional risk evaluation
considering flooding as an example of natural disaster.
There are several natural hazards related to climate changes and global warming.
In this section, an illustrative example is presented of using MCDM/A to evaluate
risk considering multidimensional consequences specifically for one of the most
frequent natural disasters, namely flooding.
Therefore, an MCDM/A model is described (Priori Jr. et al. 2015) focusing on
specific aspects, including the occurrence of different events/scenarios, the choice
of criteria, different methods and the distinct rationality required from the DM`s
preference structure.
Some steps of the general procedure for building MCDM/A models proposed
in Chap. 2 are mentioned throughout the presentation of the model.
This risk evaluation considers urban areas located in coastal regions at sea
level, or even below sea level, for example, in the Netherlands.
In underdeveloped countries, there are poor communities that are more exposed
than others to flooding due to the lack of infrastructure. This vulnerability puts
people at risk from landslides as a consequence of rainfall even without flooding
affecting the safety of such communities.
A probabilistic background is necessary for this type of evaluation. Thus, a risk
hierarchy can be built, based on Utility Theory, for the most critical areas by
assigning priorities to risks with a view to reducing or mitigating them in order to
allocate the available resources better and to a level above the local safety
standards applied. Therefore, when considering step 2 of the general procedure for
building MCDM/A models, the overall objective is to assign priorities to risk so as
to guide how resources will be allocated (Lins and de Almeida 2012). According
to step 3, human, environmental, financial and infrastructure are considered
consequence dimensions. The hierarchical structure of these dimensions and
attributes is shown in Fig. 12.6.
12.4 Natural Disasters 367
Natural hazards
(Consequences)
Communication
Impact area
No fatalities
Transport
Buildings
Drainage
Fatalities
Energy
Cost
Step 5 deals with identifying the state of nature. The DM has to consider a set
of controllable and uncontrollable inputs that impacts the problem analyzed.
Additionally, the DM makes decisions that take account of states of nature and
probabilities for several consequences. Therefore, utility functions are defined so
as to represent the DM’s preferences for different consequences (Cox 2009).
Elements of decision theory are used to evaluate the consequences. The
consequences are represented by c and the set of alternatives by A. The state of
nature ș, represents the uncertainty related to the problem, measured by the
magnitude of the rainfall. The state of nature is represented by a continuous set
denoted by real numbers regarding the rate of rainfall in a given region per hour
(mm/h), in a determined rainfall event. Lognormal or Gamma probability density
functions could be applied to represent the mm/h for a determined rainfall event in
a specific location (Cho et al. 2004).
As to considering consequences of a rainfall, a probabilistic approach can be
introduced to incorporate the associated uncertainties in A, considering a
probability distribution over consequences given the state of nature. By eliciting
utility functions for these consequences, DM’s preferences are represented in the
model. The prior probability ʌ(ș) is introduced as the probability of each state of
nature. Therefore, the expected utility E[U(ș,ai)] is used to represent the risk
associated with each given alternative (Berger 1985).
The utility is calculated by combining the probability of the consequences c in
A, the consequence function P(c|ș,ai). The effects of the tide on run-off rainwater
can be included in this probabilistic mechanism. Therefore, the expected utility
E[U(ș, ai)] of these consequences is represented by (12.4).
According to Berger (1985), the loss function is defined as the negative of the
utility function L(ș, ai) = - E[U (ș, ai)]. Thus, the losses are computed for each
criterion (L(h); L(f); L(e); L(s)) considering the urban area analyzed and the state of
nature. Moreover, the risk to an urban area is defined by (12.5).
where kh, ke, kf, ks are scale constants for the human, environmental, financial and
infrastructure dimensions respectively.
From the hierarchical structure of the attributes, the infrastructure dimension s
and the human dimension h present specific attributes. These attributes are
considered in the MAU function as given by (12.7).
where:
kh, ke, kf, ks are a scale constants that represent the value of the tradeoff
(dimensions);
kh1, kh2, ks1,…, ks5 are a scale constants that represent the value of the tradeoff
(specific attributes).
Step 9 consists of applying an algorithm in the decision model in order to
evaluate the set of alternatives. In this model, the interval scale of the utility
function is applied to provide additional information based on comparing the
utility values and ratios of the increments of utility between alternatives.
Finally, step 10 of the procedure for resolving problems and building MCDM/A
models presented in Chap. 2 consolidates step 9 by conducting a sensitivity
analysis to verify the robustness of the model, incorporating the data and
parameters analyzed.
The demand for electric power has increased and all current forecasts indicate
even higher growth due to improvements in the quality of life, besides population
growth. As mentioned in Sect. 12.4, when discussing the effects and trends of
changes in the climate, consequences in generating and consuming electric power
can also be observed. The effects of climate changes can limit power generation
capacity, when considering renewable sources, and in terms of environmental
constraints for generation when considering other sources such as coal and oil. On
the other hand, the demand for electric power rises due to severe weather such as
colder winters or higher temperatures in summer. Therefore, the impact of power
outages becomes more and more critical due to the great dependence of modern
society on power always functioning.
370 Chapter 12 Other Risk, Reliability and Maintenance Decision Problems
Thus, since there are many social and economic issues that are rising in
importance, there are many MCDM/A problems regarding decisions on the
reliability of power systems.
From this perspective, due to the threat of irreversible damages to the
environment, there is a need to consider various forms of energy production in
order to satisfy the aspects related to demand, environment and cost (Jebaraj and
Iniyan 2006).
Besides environmental and social changes, electric power systems have
evolved and have been restructured according to the country’s needs, and these
depend on each country’s specific energy policies. Therefore, there is a need to
emerge from the unique view of provision of power at “minimum cost” to a
broader perspective that allows consideration of multiple aspects, and this may
consider the different interests of the actors involved in dealing with planning
energy systems (Diakoulaki et al. 2005).
For power systems predominantly based on hydropower generation, such as
Canada and Brazil, additional complexity is introduced when depending on this
power source. Such complexity concerns planning the power system due to the
dynamics of river flows and precipitation patterns in such energy producing
systems.
As to step 5 of the general procedure for building MCDM/A models given in
Chap. 2, the state of nature T reflects the potential energy generation stored in
water reservoirs over time. Depending on the context, different levels of uncertainty
are found. For example, while in Canada river flows are more predictable due to
the relation with the accumulated volume of snow layers; in Brazil the prediction
of potential energy generation stored in water reservoirs is more difficult since
there is no such relation (Albuquerque et al. 2009).
Considering the importance of potential energy generation forecasts, the
operational planning of these power systems is much more complex than in power
systems higher percentages of coal, oil and nuclear generation sources. The
operational planning of generating power in such systems, which are not
predominantly dependent on hydropower, is not highly associated with matching
temporal to spatial data (Diniz and Maceira 2008).
Therefore, many MCDM/A problems arise in order to assure the supply
reliability. Another aspect that must be considered is how the electric power
system is designed, as this may result in different kinds of constraints and
consequences over the system.
The objective of the system operator is to assure that the demand will be met
with minimum cost and the maximum reliability of system supply. Thus the
planning of such systems must be formulated with an MCDM/A paradigm that
takes into account besides cost and the reliability of system supply, other aspects
related both to the quality of service (such as voltage, power-frequency and
harmonics) and to environmental impacts.
Thus, a reliability decision problem in power systems includes several
objectives and a variety of constraints which reflect the physical system (Pinto
12.6 Integrated Production and Maintenance Scheduling 371
not work in practice. Indeed, it is not possible to ensure long lasting results for the
production perspective, since production scheduling does not last for long due to
failures. Nevertheless, a very common hypothesis considers that equipment is
always available during the scheduling period, even though the probability of
failure in intensively used production systems has a significant value (Allaoui
et al. 2008).
Thus, it is interesting to include the activity of preventive maintenance in a
production schedule in an integrated form (Ángel-Bello et al. 2011). According to
Allaoui et al. (2008), in the literature, there are two particularly prominent
approaches with regard to the problem of integrating production and preventive
maintenance. For the first kind, the optimum maintenance schedule in the
production system can be determined. The second approach comprises optimizing
the scheduling of production by considering a preventive maintenance plan. By
doing so, the maintenance schedule decision could be drawn up in advance of the
production schedule. The problem with this approach is that the dynamic nature of
the problem is overlooked.
Despite there being some interesting papers dealing simultaneously with a
maintenance and a production schedule, most of them consider only one decision
criterion (Alardhi et al. 2007; Benmansour et al. 2011; Ji et al. 2007; Sortrakul and
Cassady 2007; Su and Tsai 2010). In fact, since these integrated models derive
from the original problem of the production schedule, some of these papers still
consider only the original objectives such as total weighted expected tardiness.
Therefore, maintenance features that influence joint scheduling are dealt with as
secondary aspects, mostly as elements of constraint.
It is worth stating that maintenance aspects are completely different from the
common criteria used to define the production schedule. Instead of simply having
some rules based on the strategy of client satisfaction, such as, expected tardiness
and makespan; maintenance aspects are related to the performance of the
equipment, such as availability, probability of finishing by the end of the schedule,
the total cost, considering preventive maintenance and interruption; and so on.
Thus, it is not difficult to realize that the operational and maintenance aspects are
complementary.
An integrated decision model that takes into account two objectives to be
optimized simultaneously: minimizing total weighted expected tardiness and
minimizing expected maintenance costs is developed. The conflicts of the
maintenance function and production are dealt with under the approach of
MCDM/A. Some results give evidence that on applying NSGA-II (Deb et al.
2002), satisfactory solutions can be found for the integrated scheduling problem.
It is assumed that there are a number of jobs to be scheduled in a single
machine in a production system. Each job has a fixed processing time, due date
and importance weight. In addition to production scheduling, it is assumed that
this machine may be unavailable due to preventive maintenance or repairs that are
needed due to failures. These features imply a conflict between the production and
maintenance objectives. Whereas the production objective may be related to
12.6 Integrated Production and Maintenance Scheduling 373
n i
minimize F1( xij , y i ) ¦
i 1
w[i ] ( ¦T
k 0
[ i ,k ]S [i ,k ] )
(12.8)
n i
minimize F2 ( xij , y i ) ¦¦ cm
i 1 k 0
[i ,k ]S [i ,k ]
i
cm[i ,k ] cb ¦y
l 1
[l ] ca k k 0,1,..., i; i 1,..., n (12.9)
i i
c[i ,k ] tp ¦
l 1
y[l ] ¦p
l 1
[l ] kt r k 0,1,..., i; i 1,..., n (12.10)
where:
n - Total number of jobs to be scheduled;
p[i] - Processing time for the i-th job performed;
d[i] - Due date for the i-th job performed;
w[i] - Weight for the i-th job performed;
374 Chapter 12 Other Risk, Reliability and Maintenance Decision Problems
205
200
195
F2 - Maintenance Cost
190
185
180
175
170
165
1350 1400 1450 1500 1550 1600 1650 1700
F1 - Tardiness
Fig. 12.7 Result of NSGA-II in the integrated production and maintenance scheduling
nij items. Each MTi has team size si. Each maintenance team has similar
characteristics with respect to repair time. A generic representation of the
maintenance system is shown in Fig. 12.8.
Eq11 n11
MT1 s1
...
...
Eq1q n1q
...
Determining the minimum and maximum number of personnel smin and smax
Computing the expected value of the loss function for the interval [smin, smax]
Establishing the trade-off between the system cost and the waiting time
Fig. 12.9 Structure of decision model with use of prior knowledge, ș = [Ȝ, ȝ]
The objective of the decision model is to define the team size while dealing
with the tradeoff between service level and cost. Thus, the problem is solved by
maximizing the MAU function given by (12.12).
§f f ·
U ( si ) k1U 1 ¨
¨ ³³ L(O , P , s ) S (O )dO S ( P )dP ¸ k 2U 2 c p s
¸
(12.12)
© f f ¹
The expected queue length, L(Ȝ, ȝ, s), is given by (12.13) and the maximum
number of servers, smax , is given by (12.14).
s
§O·
¨¨ ¸¸ O P
©P¹ 1 §O·
L (O , P , s ) ¨¨ ¸¸ (12.13)
s 1!( s P O ) 2 s 1
1 §O·
n
1 §O·
s
§ sP · ©P ¹
¦ n! ¨¨© P ¸¸¹
n 0
¨ ¸
s! ¨© P ¸¹
¨¨ ¸¸
©sP O ¹
1 1 4O
smax 1 (12.14)
2 2 P 'U critical
378 Chapter 12 Other Risk, Reliability and Maintenance Decision Problems
where:
Ȝ - failure rate;
ȝ - repair rate;
s - number of servers;
cp - cost of personnel
ʌ(Ȝ) - prior probability on failure rate;
ʌ(ȝ) - prior probability on repair rate;
ǻȡcritical - critical value of difference in terms of factor utilization.
Considering an illustrative example with ʌ(Ȝ) and ʌ(ȝ) given by Weibull
probability distributions with ȕ=11.3 and Ș=0.05154 for ʌ(Ȝ); and ȕ=10.7 and
Ș=0.01153 for ʌ(ȝ). In addition, an additive function with scale constants k1 = 0.10
and k2 = 0.90 for the MAU function given by (12.12).
The minimum number of servers smin to satisfy the stationary condition of 99%
is found by considering the Ȝmax and ȝmin from the inverse Weibull function.
Therefore, smin = 8.
Based on (12.14), the maximum number of servers is defined as smax = 23.
Thus, for this scenario, when maximizing the MAU function, the best maintenance
team size would be s*=10. This result is illustrated in Fig. 12.10.
0,98
0,96
MAU function
0,94
0,92
0,9
0,88
0,86
8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23
The result observed from the sensitivity analysis indicates that the analyst
should give more attention to the elicitation process that obtained the scale
parameters from the experts. As these are the most sensitive parameters, this
elicitation process has to be more accurate than other parameters not so sensitive.
min Rb (T1 )
(12.15)
s.t. Rb (T 0 ) d D
This formulation corresponds to testing the null hypothesis H0: that the
equipment has a lower failure rate than or equal to a specified against the
alternative hypothesis H1. This means that H0: TO Ȝ0 and H1: TO > Ȝ0.
There are two errors involved in the hypothesis test that should be minimized.
The probability Į of rejecting the null hypothesis when it is true, Rb(ș0), and the
probability ȕ of incorrectly accepting the null hypothesis, Rb (ș1). These errors are
known as error Type I and error Type II, respectively.
12.8 Bayesian Reliability Acceptance Test Based on MCDM/A 381
° i 1 b ½°
min ® ¦ ³ S (T )[ L(T , a0 ) L(T , a1 )] e T e x dT ¾ (12.16)
i
°̄ x 0 x! a °¿
where:
ș - state of nature;
ʌ(ș) - prior probability distribution;
L(ș,ai) - Loss function.
The interval [a, b] corresponds to the given range of TO in a prior distribution.
The solution that minimizes (12.16) represents the maximum number of failures
that there are for not rejecting the null hypothesis. If during the phase of operation
trials, there occurs a greater number of failures than the solution of (12.16), the
null hypothesis must be rejected.
This procedure using a Bayesian approach can support a DM to find the
maximum number of failures that would be acceptable, considering the objectives
and knowledge available.
Thus, when considering reliability acceptance for equipment there are con-
flicting objectives as discussed previously regarding the system’s reliability and
the project being delivered on time.
Therefore, in some situations a DM can be concerned with evaluating of the
tradeoff between the error Type I dimension and the delay in delivering the project
due to rejecting the equipment.
For example, if a DM is interested in the evaluating whether the reliability
defined in the contract is consistent with the actual reliability of the purchased
items, he/she may consider a tolerance level. He/she does so in order to include an
upper limit greater than the equipment’s nominal reliability value, which is usually
the one set out on the contract, for the null hypothesis. Thus, this tolerance level
represents how much the DM is willing to tradeoff in terms of reliability in order
to succeed in delivering the project on time. Therefore, this means that the DM
may decide to accept the null hypothesis since the reliability is lower than the one
defined by contract, but respects the tolerance level accepted by the DM in order
to deliver the project on time.
382 Chapter 12 Other Risk, Reliability and Maintenance Decision Problems
This decision may involve a substantial delay in project execution time from
several months to years due to the specificities of the items that may be rejected.
For these kinds of items, the lead time involved with the purchasing process and
delivery is sufficiently long to compromise delivering the project on time. This is
because the manufacturer usually adopts a make-to-order strategy for this kind of
equipment which means that in most cases, it is not available in the short term i.e.
immediately after ordering the equipment.
Thus, for dealing with a decision that has such relevant and strategic objectives,
the MCDM/A approach provides techniques and methods for modeling the DM’s
preferences in order to give a recommendation on the reliability acceptance test
considering the broader aspects involved in the problem. Therefore, the general
procedure defined for building MCDM/A models in Chap. 2 gives the directions
to build a suitable decision model for a problem with these features.
This section is divided into two topics based on the generation of MOEAs
(Multiobjective Evolutionary Algorithms).
The first papers using multiobjective formulation to find Pareto solutions were
published in the 1980s. Dewispelare (1984) formulated a non-linear multiple
objective problem with regard to a pre-production decision on an airborne tactical
missile where the reliability, survivability, combat effectiveness, cost and flight
area were considered as objective functions. Feasible space was explored for all
non-dominated solutions obtained by a constrained optimization technique.
Although non-dominated solutions should be found, a scalar scoring function is
recommended when the DM is not able to make a choice due to the incomplete
ordering of the Pareto solutions set.
Soltani and Corotis (1988) constructed a trade-off curve of a design for
structural systems as a result of using multiobjective linear programming obtained
and a constrained optimization technique to formulate objective functions of cost
of failure versus initial cost.
Fu and Frangopol (1990) found Pareto optimal solutions in a multiobjective
formulation of structural systems considering three objectives: weight, system
reliability and redundancy. They used the H-constraint method to find Pareto
solutions.
12.9 Some Multiobjective Optimization Models on Reliability and Maintenance 383
Misra and Sharma (1991), Dhingra (1992) and Rao and Dhingra (1992) used
MOEAs for redundancy allocation, as discussed in Chap. 9.
With the development of the Second Generation of MOEAs close to 2000, several
studies have developed with a view to evaluating the effectiveness of these
techniques and more often from this point on, in the field of maintenance and
reliability, such as: NSGA-II, SPEA2 and other approaches.
The use of the second generation of MOEAs in Reliability and Maintenance
problems has become one of the most common approaches across the Pareto-front
approaches. Some cases from the literature are highlighted in Table 12.2.
Table 12.2 Some Pareto-front approaches used in Reliability and Maintenance problems
Min-max concept; Exact algorithm; Misra and Sharma (1991), Certa et al. (2011), Chou and Le
PSO; GPSIA (2011)
Ramirez-Rosado and Bernal-Agustin (2001), Marseguerra
et al. (2002), Marseguerra et al. (2004), Kumar et al.
(2006), Kumar et al. (2008), Cadini et al. (2010), Moradi
et al. (2011), Wang and Hoang (2011), Chiang (2012),
MOEA; MOGA; NSGA-II
Torres-Echeverria et al. (2012), Zio et al. (2012), Gjorgiev
et al. (2013), Jin et al. (2013), Li et al. (2013), Lins et al.
(2013), Rathod et al. (2013), Trivedi et al. (2013), Zidan
et al. (2013)
References
Aghezzaf E-H, Najid NM (2008) Integrated production planning and preventive maintenance in
deteriorating production systems. Inf Sci (Ny) 178:3382–3392
Alardhi M, Hannam RG, Labib AW (2007) Preventive maintenance scheduling for
multícogeneration plants with production constraints. J Qual Maint Eng 13:276–292
Albuquerque LL de, Almeida AT de, Cavalcante CAV (2009) Aplicabilidade da programação
matemática multiobjetivo no planejamento da expansão de longo prazo da geração no Brasil
(Multiobjective mathematical programming applicability in long-term expansion planning of
generation in Brazil). Pesqui Operacional 29 :153–177
Allaoui H, Lamouri S, Artiba A, Aghezzaf E (2008) Simultaneously scheduling n jobs and the
preventive maintenance on the two-machine flow shop to minimize the makespan. Int J Prod
Econ 112:161–167
Almeida-Filho AT de, Ferreira RP, de Almeida AT (2013) A DSS Based on Multiple Criteria
Decision Making for Maintenance Planning in an Electrical Power Distributor. In: Purshouse
R, Fleming P, Fonseca C, et al. (eds) Evol. Multi-Criterion Optim. SE - 58. Springer Berlin
Heidelberg, pp 787–795
ANEEL (2012) Agência Nacional de Energia Elétrica, Brazil. Qualidade do serviço (Quality of
service). Available at http://www.aneel.gov.br/area.cfm?idArea=79&idPerfil=2. Accessed 05
Mai 2012
Ángel-Bello F, Álvarez A, Pacheco J, Martínez I (2011) A heuristic approach for a scheduling
problem with periodic maintenance and sequence-dependent setup times. Comput Math with
Appl 61:797–808
Azaron A, Perkgoz C, Katagiri H, et al. (2009) Multi-objective reliability optimization for
dissimilar-unit cold-standby systems using a genetic algorithm. Comput Oper Res 36:1562–
1571
Barakat S, Bani-Hani K, Taha MQ (2004) Multi-objective reliability-based optimization of pre-
stressed concrete beams. Struct Saf 26:311–342
Batista FRS, Geber de Melo AC, Teixeira JP, Baidya TKN (2011) The Carbon Market
Incremental Payoff in Renewable Electricity Generation Projects in Brazil: A Real Options
Approach. Power Syst IEEE Trans 26:1241–1251
Benmansour R, Allaoui H, Artiba A, et al. (2011) Simulatiońbased approach to joint
production and preventive maintenance scheduling on a failuréprone machine. J Qual
Maint Eng 17:254–267
References 387
Berger J (1985) Statistical Decision Theory and Bayesian Analysis (Springer Series in Statistics).
Springer, New York
Bobrowsky PT (ed) (2013) Encyclopedia of Natural Hazards. Springer, Dordrecht
Bornstein CT, Maculan N, Pascoal M, Pinto LL (2012) Multiobjective combinatorial
optimization problems with a cost and several bottleneck objective functions: An algorithm
with reoptimization. Computers and Operations Research, 39 (9): 1969-1976
Brandeau ML, Chiu SS (1989) An Overview of Representative Problems in Location Research.
Manage Sci 35:645–674
Brans JP, Mareschal B (1984) PROMETHEE: a new family of outranking methods in
multicriteria analysis. Operational Research 84. Brans JP (eds). Amsterdam: North-Holland,
pp. 408–421
Brans JP, Mareschal B (2002) Prométhée-Gaia: une méthodologie d’aide à la décision en
présence de critères multiples. Éditions de l’Université de Bruxelles
Cadini F, Zio E, Petrescu CA (2010) Optimal expansion of an existing electrical power
transmission network by multi-objective genetic algorithms. Reliab Eng Syst Saf 95:173–181
Cassady CR, Kutanoglu E (2003) Minimizing Job Tardiness Using Integrated Preventive
Maintenance Planning and Production Scheduling. IIE Trans 35:503–513
Certa A, Galante G, Lupo T, Passannanti G (2011) Determination of Pareto frontier in multi-
objective maintenance optimization. Reliab Eng Syst Saf 96:861–867
Chiang T-A (2012) Multi-objective decision-making methodology to create an optimal design
chain partner combination. Comput Ind Eng 63:875–889
Cho H-K, Bowman KP, North GR (2004) A Comparison of Gamma and Lognormal
Distributions for Characterizing Satellite Rain Rates from the Tropical Rainfall Measuring
Mission. J Appl Meteorol 43:1586–1597
Chou J-S, Le T-S (2011) Reliability-based performance simulation for optimized pavement
maintenance. Reliab Eng Syst Saf 96:1402–1410
Cox LA Jr (2009) Risk analysis of complex and uncertain systems. Springer Science & Business
Media
de Almeida AT, Cavalcante CAV, Ferreira RJP, et al. (2006) Location of Back-up Transformers.
Eng. Manag. Conf. 2006 IEEE Int. IEEE, Salvador, Bahia, pp 300–302
de Almeida AT, Ferreira RJP, Cavalcante CAV (2015) A review of multicriteria and multi-
objective models in maintenance and reliability problems. IMA Journal of Management
Mathematics 26(3):249–271
de Almeida AT, Souza FMC (1986) Bayes-Like Decisions in Reliability Engineering.
In: Proceedings of International Conference on Information Processing and Management of
Uncertainty in Knowledge-Based Systems, Paris, 30 June-4 July, 87-90
de Almeida AT, Souza FMC (2001) Gestão da Manutenção: na Direção da Competitividade
(Maintenance Management: Toward Competitiveness) Editora Universitária da UFPE. Recife
De León JCV (2006) Vulnerability: A conceptual and methodological review. ‘Studies of the
University: Research, Counsel, Education. No. 4/2006, Institute for Environment and Human
Security (UNU-EHS), Bonn, Available at https://www.ehs.unu.edu/file/get/8337.pdf.
Accessed 16 Jan 2014
Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multiobjective genetic
algorithm: NSGA-II. IEEE Trans Evol Comput 6:182–197
DeWispelare AR (1984) A computer based application of non-linear multiple objective
optimization. Comput Ind Eng 8:143–152
Dhingra AK (1992) Optimal apportionment of reliability and redundancy in series systems under
multiple objectives. IEEE Trans Reliab 41:576–582
Diakoulaki D, Antunes CH, Gomes Martins A (2005) MCDA and Energy Planning. Mult.
Criteria Decis. Anal. State Art Surv. SE - 21. Springer New York, pp 859–890
Diniz AL, Maceira MEP (2008) A Four-Dimensional Model of Hydro Generation for the Short-
Term Hydrothermal Dispatch Problem Considering Head and Spillage Effects. Power Syst
IEEE Trans 23:1298–1308
388 Chapter 12 Other Risk, Reliability and Maintenance Decision Problems
Drezner Z, Hamacher HW (2004) Facility location: applications and theory. Springer Science &
Business Media, New York
Farag A, Al-Baiyat S, Cheng TC (1995) Economic load dispatch multiobjective optimization
procedures using linear programming techniques. Power Syst IEEE Trans 10:731–738
Ferreira RJP, de Almeida AT, Ferreira HL (2010) Multi-attribute p-median model for location of
back-up transformers. Brazilian J Oper Prod Manag 7(2):09–28
Ferreira RJP, Ferreira HL (2012) Decision support system for location of back-up transformers
based on a multi-attribute p-median model. Syst. Man, Cybern. (SMC), 2012 IEEE Int. Conf.
IEEE, Seoul, pp 629–631
Field CB, Barros V et al. (eds) (2012) IPCC – Managing the risks of extreme events and
disasters to advance climate change adaptation. A Special Report of Working Groups I and II
of the Intergovernmental Panel on Climate Change. Cambridge University Press, New York
Fu G, Frangopol DM (1990) Balancing weight, system reliability and redundancy in a multi-
objective optimization framework. Struct Saf 7:165–175
Girgin S, Krausmann E (2013) RAPID-N: Rapid natech risk assessment and mapping
framework. J Loss Prev Process Ind 26(6):949–960
Gjorgiev B, Kanþev D, ýepin M (2013) A new model for optimal generation scheduling of
power system considering generation units availability. Int J Electr Power Energy Syst
47:129–139
Guikema SD (2009) Natural disaster risk analysis for critical infrastructure systems:
An approach based on statistical learning theory. Reliab Eng Syst Saf 94:855–860
Hansson K, Danielson M, Ekenberg L, Buurman J (2013) Multiple Criteria Decision Making for
Flood Risk Management. In: Amendola A, Ermolieva T, Linnerooth-Bayer J, Mechler R
(eds) Integr. Catastr. Risk Model. SE - 4. Springer Netherlands, pp 53–72
Hillier FS (1963) Economic Models for Industrial Waiting Line Problems. Manage Sci 10:119–
130
Huttenlau M, Stötter J (2011) The structural vulnerability in the framework of natural hazard risk
analyses and the exemplary application for storm loss modelling in Tyrol (Austria). Nat
Hazards 58:705–729
IEEE (2012) Std 1366-2012: IEEE Guide for Electric Power Distribution Reliability Indices.
IEEE, New York
Ighravwe DE, Oke SA (2014) A non-zero integer non-linear programming model for
maintenance workforce sizing. Int J Prod Econ 150:204–214
Jebaraj S, Iniyan S (2006) A review of energy models. Renew Sustain Energy Rev 10:281–311
Jha AK, Miner TW, Stanton-Geddes Z (2013) Building Urban Resilience: Principles, Tools, and
Practice. 1–180. © World Bank. http://elibrary.worldbank.org/doi/book/10.1596/978-0-8213-
8865-5
Ji M, He Y, Cheng TCE (2007) Single-machine scheduling with periodic maintenance to
minimize makespan. Comput Oper Res 34:1764–1770
Jiang R, Ji P (2002) Age replacement policy: a multi-attribute value model. Reliab Eng Syst Saf
76(3):311–318
Jin T, Tian Y, Zhang CW, Coit DW (2013) Multicriteria Planning for Distributed Wind
Generation Under Strategic Maintenance. Power Deliv IEEE Trans 28:357–367
Karimi I, Hüllermeier E (2007) Risk assessment system of natural hazards: A new approach
based on fuzzy probability. Fuzzy Sets Syst 158:987–999
Keeney RL, Raiffa H (1976) Decisions with multiple objectives: Preferences and Value Trade-
Offs. Wiley Series in Probability and Mathematical Statistics. Wiley and Sons, New York
Keller EA, DeVecchio DE (2012) Earth’s Processes as Hazards, Disasters, and Catastrophes.
Natural Hazards. Pearson Prentice Hall, New Jersey
Konak A, Kulturel-Konak S, Levitin G (2012) Multi-objective optimization of linear multi-state
multiple sliding window system. Reliab Eng Syst Saf 98(1):24–34
Krausmann E, Cozzani V, Salzano E, Renni E (2011) Industrial accidents triggered by natural
hazards: an emerging risk issue. Nat Hazards Earth Syst Sci 11:921–929
References 389
Krausmann E, Cruz A (2013) Impact of the 11 March 2011, Great East Japan earthquake and
tsunami on the chemical industry. Nat Hazards 67:811–828
Kumar Y, Das B, Sharma J (2006) Service restoration in distribution system using non-
dominated sorting genetic algorithm. Electr Power Syst Res 76:768–777
Kumar Y, Das B, Sharma J (2008) Multiobjective, multiconstraint service restoration of electric
power distribution system with priority customers. Power Deliv IEEE Trans 23:261–270
Li F, van Gelder PHAJM, Ranasinghe R, et al. (2014) Probabilistic modelling of extreme storms
along the Dutch coast. Coast Eng 86:1–13
Li YF, Sansavini G, Zio E (2013) Non-dominated sorting binary differential evolution for the
multi-objective optimization of cascading failures protection in complex networks. Reliab
Eng Syst Saf 111:195–205
Lin C, Madu CN, Chien TW, Kuei C-H (1994) Queueing Models for Optimizing System
Availability of a Flexible Manufacturing System. J Oper Res Soc 45(10):1141–1155
Linnekamp F, Koedam A, Baud ISA (2011) Household vulnerability to climate change:
Examining perceptions of households of flood risks in Georgetown and Paramaribo. Habitat
Int 35:447–456
Lins ID, Rêgo LC, Moura M das C, Droguett EL (2013) Selection of security system design via
games of imperfect information and multi-objective genetic algorithm. Reliab Eng Syst Saf
112:59–66
Lins PHC, de Almeida AT (2012) Multidimensional risk analysis of hydrogen pipelines. Int J
Hydrogen Energy 37:13545–13554
Mareschal B, De Smet Y, Nemery P (2008) Rank reversal in the PROMETHEE II method: Some
new results. Ind Eng Eng Manag 2008 IEEM 2008 IEEE Int Conf 959–963
Marseguerra M, Zio E, Podofillini L (2002) Condition-based maintenance optimization by
means of genetic algorithms and Monte Carlo simulation. Reliab Eng Syst Saf 77:151–165
Marseguerra M, Zio E, Podofillini L (2004) A multiobjective genetic algorithm approach to the
optimization of the technical specifications of a nuclear safety system. Reliab Eng Syst Saf
84:87–99
Misra KB, Sharma U (1991) An efficient approach for multiple criteria redundancy optimization
problems. Microelectron Reliab 31:303–321
Moghaddam KS (2013) Multi-objective preventive maintenance and replacement scheduling in a
manufacturing system using goal programming. Int J Prod Econ 146:704–716
Moradi E, Fatemi Ghomi SMT, Zandieh M (2011) Bi-objective optimization research on
integrated fixed time interval preventive maintenance and production for scheduling flexible
job-shop problem. Expert Syst Appl 38:7169–7178
Moubray J (1997) Reliability-centered maintenance. Industrial Press Inc., New York
Paul, BK (2011) Environmental hazards and disasters: contexts, perspectives and management.
Wiley-Blackwell, Chichester
Pelling M (2003) Vulnerability of cities. Earthscan Publications Ltd., London
Pine JC (2009) Natural hazards analysis: reducing the impact of disasters. CRC Press, Taylor &
Francis Group, Florida
Pinedo ML (2012) Scheduling: theory, algorithms, and systems. Fourth Edition. Springer,
New York.
Pinto RJ, Borges CLT, Maceira MEP (2013) An Efficient Parallel Algorithm for Large Scale
Hydrothermal System Operation Planning. Power Syst IEEE Trans 28:4888–4896
Priori Jr L, Alencar MH, de Almeida AT (2015) Adaptations to possible climate change impacts:
problem structuring based on VFT methodology. CDSID Working report
Ramirez-Rosado IJ, Bernal-Agustin JL (2001) Reliability and costs optimization for distribution
networks expansion using an evolutionary algorithm. Power Syst IEEE Trans 16:111–118
Rao SS, Dhingra AK (1992) Reliability and redundancy apportionment using crisp and fuzzy
multiobjective optimization approaches. Reliab Eng Syst Saf 37:253–261
Rathod V, Yadav OP, Rathore A, Jain R (2013) Optimizing reliability-based robust design model
using multi-objective genetic algorithm. Comput Ind Eng 66:301–310
390 Chapter 12 Other Risk, Reliability and Maintenance Decision Problems
Shafiee M, Finkelstein M (2015) An optimal age-based group maintenance policy for multi-unit
degrading systems. Reliab Eng Syst Saf 134:230–238
Solecki W, Leichenko R, O’Brien K (2011) Climate change adaptation strategies and disaster
risk reduction in cities: connections, contentions, and synergies. Curr Opin Environ Sustain
3:135–141
Soltani M, Corotis RB (1988) Failure cost design of structural systems. Struct Saf 5:239–252
Sortrakul N, Cassady CR (2007) Genetic algorithms for total weighted expected tardiness
integrated preventive maintenance planning and production scheduling for a single machine.
J Qual Maint Eng 13:49–61
Su L, Tsai H (2010) Flexible preventive maintenance planning for two parallel machines
problem to minimize makespan. J Qual Maint Eng 16:288–302
Thywissen K (2006) Components of Risk: A Comparative Glossary. Institute for Environment
and Human Security (UNU-EHS), Bonn
Torres-Echeverría AC, Martorell S, Thompson HA (2012) Multi-objective optimization of
design and testing of safety instrumented systems with MooN voting architectures using a
genetic algorithm. Reliab Eng Syst Saf 106:45–60
Trivedi A, Srinivasan D, Sharma D, Singh C (2013) Evolutionary Multi-Objective Day-Ahead
Thermal Generation Scheduling in Uncertain Environment. Power Syst IEEE Trans 28:1345–
1354
Vari A, Linnerooth-Bayer J, Ferencz Z (2003) Stakeholder Views on Flood Risk Management in
Hungary’s Upper Tisza Basin. Risk Anal 23:585–600
Wang Y, Pham H (2011) A Multi-Objective Optimization of Imperfect Preventive Maintenance
Policy for Dependent Competing Risk Systems with Hidden Failure. Reliab IEEE Trans
60:770–781
Wong KP, Fan B, Chang CS, Liew AC (1995) Multi-objective generation dispatch using
bi-criterion global optimisation. Power Syst IEEE Trans 10:1813–1819
Yokoyama R, Bae SH, Morita T, Sasaki H (1988) Multiobjective optimal generation dispatch
based on probability security criteria. Power Syst IEEE Trans 3:317–324
Zidan A, Shaaban MF, El-Saadany EF (2013) Long-term multi-objective distribution network
planning by DG allocation and feeders’ reconfiguration. Electr Power Syst Res 105:95–104
Zio E, Golea LR, Rocco S. CM (2012) Identifying groups of critical edges in a realistic electrical
network by multi-objective genetic algorithms. Reliab Eng Syst Saf 99:172–177
Zischg A, Schober S, Sereinig N, et al. (2013) Monitoring the temporal development of natural
hazard risks as a basis indicator for climate change adaptation. Nat Hazards 67:1045–1058
Index
B D
Bathtub curve, 119–120, 277 Danger zone, 173–174, 178–179, 182, 329
Bayesian approach, 65, 151, 173, 275, 339, Debugging, 120
379 Decision
Blackout, 186, 200, 354 analysis, 2, 7, 81, 149, 175, 179, 192,
Block-based replacement, 291 200, 202, 205–206
Burn in, 120 matrix, 12, 16, 35, 38, 49
method, 57, 201
model, 1–21, 26, 53, 167–169, 172,
C 181–197, 258, 263, 266, 268,
Catastrophic losses, 174, 337, 365 275, 318, 338, 357
Choice of MCDM/A method, 14, 28, 33, perspectives, 14
38, 51, 63, 164, 166, 325, 366 theory, 7, 35, 168, 179, 181, 183, 187,
Choice of method, 325, 359 258, 260, 283
Climate changes, 199, 202, 204, 366 Decision maker’s
Compensatory, 16–18, 37, 63, 174, 181, preference, 3, 4, 9–10, 20–21, 31, 36,
190–191, 223, 258, 263, 266, 338 51, 59, 89–90, 92–93, 128, 163,
Condition-based maintenance, 139, 234, 166–168, 170–171, 174, 181,
290 187, 256, 258–265, 267–268,
Condition monitoring, 234, 239, 242, 270, 283
290–291 rationality, 16–18, 37, 51, 181, 190–191,
Consequences, 31, 117 263, 267, 325, 338, 366
analysis, 5, 8, 12, 33, 89, 112, 161–162, risk behavior, 7, 21, 39, 154, 161,
168–169, 173–182, 192, 259, 167–168, 172, 174, 181, 190,
267, 283 192–193, 264, 277, 283–284, 286
P
Parallel-series system, 298 R
Parallel system, 298 Rainfall, 366
Perception of risk, 92, 111, 161, 192, 208 RCM. See Reliability Centered
Periodic replacement, 217 Maintenance (RCM)
Perishable items, 278 Redesign, 144, 331, 332
Petrochemical, 142 Redundancy allocation, 297, 313
Planning replacement, 217 Redundant systems, 129–130, 303
P-median, 355 Reliability, 18, 115–117, 126–130, 243,
Portfolio of actions, 285, 290 255, 276, 283, 284, 300, 313, 336,
Portfolio problem, 9, 34, 63, 285, 326 380
Power Reliability acceptance test, 379
generation, 369 Reliability Centered Maintenance (RCM),
outages, 369 146, 336–341, 358, 360
system, 170, 186, 198, 200, 370 Repairable system, 130, 264, 277, 285
transformers, 354 Repair contract selection, 249, 251, 256
Predictive maintenance, 233 Replacement, 130, 216–217, 277, 286–288,
Preference 290–294
modeling, 10, 14, 24, 36, 340, 368 Residual useful life, 234, 290
structure, 3, 9–12, 20, 36, 58, 66, 68, Resilience, 363
170, 171, 181, 188, 198, 244, Response time, 257, 263–266
256, 258, 261, 264, 283, 338 Risk
Preference Ranking Organization Method acceptability, 92–94, 106, 323, 326
for Enrichment Evaluation analysis, 90–95, 101, 112, 114, 161,
(PROMETHEE), 15, 16, 73, 271, 164, 167, 169–170, 173, 178,
359 197, 203, 321
Prescriptive perspective, 14, 68 assessment, 55, 91, 94, 101, 111, 162,
Preventive maintenance, 8, 18, 126, 215, 164, 182, 199
238, 277, 286, 291, 292, 358, 371 categories, 106, 191–193, 195
Index 395