Next Article in Journal
Empirical Evaluation of Unoptimized Sorting Algorithms on 8-Bit AVR Arduino Microcontrollers
Previous Article in Journal
Deep Learning-Based Multi-Lead ECG Reconstruction from Lead I with Metadata Integration and Uncertainty Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Integrated Risk-Informed Multicriteria Approach for Determining Optimal Inspection Periods for Protective Sensors

by
Ricardo J. G. Mateus
*,
Rui Assis
,
Pedro Carmona Marques
,
Alexandre D. B. Martins
,
João C. Antunes Rodrigues
and
Francisco Silva Pinto
RCM2+ Research Centre for Asset Management and Systems Engineering, Lusófona University, Campo Grande, 376, 1749-024 Lisboa, Portugal
*
Author to whom correspondence should be addressed.
Sensors 2026, 26(1), 213; https://doi.org/10.3390/s26010213 (registering DOI)
Submission received: 28 October 2025 / Revised: 14 December 2025 / Accepted: 26 December 2025 / Published: 29 December 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

Equipment failure is the leading cause of industrial operational disruption, with unplanned downtime accounting for up to 11% of manufacturing revenue, highlighting the need for effective proactive maintenance strategies, such as protective sensors that can detect potential failures in critical equipment before a functional failure occurs. However, sensors are also subject to hidden failures themselves, requiring periodic failure-finding inspections. This study proposes a novel integrated multimethodological approach combining discrete event simulation, Monte Carlo, optimization, risk analysis, and multicriteria decision analysis methods to determine the optimal inspection period for protective sensors subject to hidden failures. Unlike traditional single-objective models, this approach evaluates alternative inspection periods based on their risk-informed overall values, considering multiple conflicting key performance indicators, such as maintenance costs and equipment availability. The optimal inspection period is then selected considering uncertainties and the intertemporal, intra-criterion, and inter-criteria preferences of the organization. The approach is demonstrated through a case study at the leading Portuguese electric utility, replacing previous empirical inspection standards that did not consider economic costs and uncertainties, supported by an open, transparent, auditable, and user-friendly decision support system implemented in Microsoft Excel using only built-in functions and modeled based on the principles of probability management. The results identified an optimal inspection period of 90 h, representing a risk-informed compromise distinct from the 120 h interval suggested by cost minimization alone, highlighting the importance of integrating organizational preferences into the decision process. A sensitivity analysis confirmed the robustness of this solution, maintaining validity even as the organizational weight for equipment availability ranged between 35% and 82%. The case study shows that the proposed approach enables the identification of inspection intervals that lead to quantitatively better maintenance cost and availability outcomes compared to empirical inspection standards.

1. Introduction

Unplanned downtime costs are estimated at 11% of revenue in the world’s 500 largest manufacturing companies and continue to increase [1]. Unplanned, reactive, or run-to-failure corrective maintenance is no longer viable for managing critical equipment. Instead, systematic preventive maintenance, after a pre-specified operating period, has become the norm. However, in critical environments, these policies have been increasingly replaced by proactive strategies, such as predictive or condition-based maintenance, capable of detecting potential failures before they degrade into functional failures [2].
Protective sensors are often used to continuously gather real-time data on equipment conditions, indicating that a functional failure is imminent or occurring. These online sensors can detect potential failures before escalation, enabling timely interventions that prevent costly functional failures. For instance, it is often possible to take advantage of the degradation of an equipment’s condition to infer that it has started its failure development period (P–F interval), that is, the time interval until a functional failure occurs [2]. Prompt action can then be taken, either manually or automatically (e.g., via an actuator), to shut down the critical equipment, repair it, and mitigate more severe consequences that could result from a functional failure.
However, protective sensors are not immune to failures, making it essential to ensure their proper functioning. Although some failures are evident, many remain hidden and undetected until they cause severe equipment consequences, leading to high reactivation costs and long downtimes. Hidden failures can account for up to half of all failure modes [2], and their late detection is a major source of downtime and reactive maintenance costs across many industrial sectors, further reinforcing the need to optimize inspection intervals.
In this two-unit sensor–equipment system, two scenarios can arise when the equipment fails:
  • The monitoring sensor is well-functioning, detecting, and issuing an early warning of a potential failure in the protected critical equipment promptly, allowing for a prompt equipment shutdown and repair, with minor equipment repair costs and downtime.
  • The monitoring sensor is either ineffective (due to a hidden failure) or non-operational (because it has failed and is still under repair), and a multiple failure occurs in the system, leading to significant consequences in terms of equipment reactivation costs and downtime.
To reduce the probability that the protective sensor is not effectively monitoring and protecting the critical equipment, the former must be inspected periodically to check for the presence of hidden failures [3]. Inspection intervals (failure finding intervals) may be either constant (periodic) or different (sequential or non-periodic). The common practice in the industry is to implement a calendar-based inspection policy with periodic and scheduled inspections, at fixed, equal, and calendar intervals due to its applicability and feasibility.
Naturally, periodic offline inspections incur their own costs. Hence, the trade-off between the consequences of failure-finding inspections in the protective sensor and the consequences of a functional failure in the equipment must be assessed. It is expected that less frequent sensor inspections (and hence longer inspection periods) will decrease the probability of detecting a failure in the equipment and consequently:
  • On the one hand, the expected sensor inspection costs and equipment availability will decrease.
  • On the other hand, the expected equipment reactivation costs and downtime are likely to increase.
Therefore, an optimal interval between periodic inspections is expected to exist, corresponding to the best compromise between the multiple conflicting organizational Key Performance Indicators (KPIs) (e.g., minimizing total maintenance costs and maximizing equipment availability) regarding the equipment’s functioning. This compromise can be solved by eliciting the specific preferences of the organization regarding those trade-offs, that is, how much it is willing to lose in one KPI (e.g., total costs) to gain in another (e.g., equipment availability).
Unlike previous studies that optimize inspection intervals using a single method (e.g., simulation or risk analysis alone), this work uniquely integrates simulation, optimization, and multicriteria decision analysis into a unified, transparent framework applicable across diverse industrial contexts.
Therefore, the main objective of this article is to present an integrated risk-informed multicriteria value optimization and simulation approach to determine the optimal inspection period for a protective sensor, subject to hidden failures, while monitoring the condition of a critical system. The novelty of this work does not lie in the development of new theoretical algorithms but rather in the coherent integration of established methods (i.e., discrete event simulation, Monte Carlo methods, risk analysis, multicriteria value measurement, and optimization) within a transparent and auditable decision framework. This integrated approach allows organizations to evaluate inspection programs under uncertainty, incorporating their preferences across multiple conflicting KPIs. This aim is proposed to be achieved by addressing the following research questions:
  • How to represent and model uncertainties in the occurrence of failures in the equipment and sensors.
  • How to predict the consequences of the occurrence of system failures on multiple KPIs, such as total maintenance costs and equipment availability, considering distinct sensor inspection periods, costs, times, and other parameters of the system and the organization.
  • How to model and compare the value and risk of distinct sensor inspection periods, considering their uncertainty and consequences on multiple conflicting organizational KPIs.
  • How to find the optimal inspection period considering the previous issues in an integrated and robust approach.
  • How to develop a decision support system that can be easily understood, audited, and used in practice by the organization.
The specific objectives of this article are to answer these research questions through an integrated and generic multimethodological approach and to demonstrate its effectiveness through its application to a real-world case study supported by an open, transparent, auditable, and user-friendly decision support system (DSS).
The remainder of this paper is organized as follows: Section 2 presents a description of the specific proactive maintenance problem tackled by the article, highlighting its multicriteria and stochastic nature, introducing the most important challenges that underpin its analysis and resolution We included a short review of the state of the art on similar published approaches to address the research problem and build the case for the need and contribution of the methodological framework. Section 3 describes the proposed methodology. It starts with a formal definition of the problem, presenting the assumptions and building blocks of the methodology, and detailing the methods, variables, parameters, and equations used in each step of the framework. Section 4 illustrates the application of the approach through a case study supported by a DSS. Section 5 discusses the main contributions to the research literature and practice and identifies its strengths and limitations. Section 6 concludes the paper and suggests potential avenues for future research.

2. State of the Art and Research Gap

Industrial systems degrade over time, leading to system failures that result in increased costs, unexpected machine unavailability, and quality and safety issues. While maintenance has historically been regarded as a reactive necessity, it is now recognized as a critical activity of asset management aimed at improving organizational efficiency, reliability, and profitability [4].
Within the context of Industry 4.0 and 5.0, traditional maintenance approaches—such as reactive and preventive maintenance—are progressively being replaced by more proactive strategies based on condition monitoring and predictive analytics, which aim to reduce maintenance costs and improve equipment availability and reliability [5,6].
A proactive maintenance system typically comprises two main types of components: (i) critical entities that perform the system’s fundamental function and (ii) auxiliary entities that provide information, redundancy, and/or protection [7,8]. Protective components serve to mitigate the consequences of the protected critical equipment failure, which would otherwise be significantly more severe [2].
Proactive maintenance relies heavily on sensor data to monitor equipment condition, enabling early anomaly detection, performance optimization, increased system reliability, and the dynamic scheduling of maintenance interventions before failures occur. Predictive maintenance combines historical sensor data with predictive models based on statistical, forecasting, or artificial intelligence methods to estimate the likelihood and timing of future failures [9,10,11,12].
Reliability Centered Maintenance (RCM) constitutes a structured and widely adopted proactive framework that prioritizes critical assets and enables the development of customized maintenance plans, which reduce downtime and operational disruptions, extend equipment lifespan, and promote more efficient, reliable, sustainable, and cost-effective operations, ultimately improving product quality and customer satisfaction. RCM’s plans track equipment performance through a combination of condition-based monitoring of real-time sensor data and predictive analytics to anticipate failures and schedule preventive maintenance interventions accordingly before failures escalate into major consequences.
Nevertheless, most maintenance decision studies assume that the monitoring sensor is perfect, that is, its information signal is accurate to the actual state of the equipment, overlooking the compounded effect of sensor unreliability due to the presence of hidden failures. Faulty sensors that produce erroneous signals lead to incorrect maintenance decisions [13]. Research on the maintenance of a critical system with an unreliable protective system is limited [14]. Hence, more research should be conducted to study the effects of imperfect condition monitoring on optimal dynamic maintenance decisions, namely, new models and methodologies that determine the optimal inspection maintenance schedule based on predicted failures of the sensor and equipment [15].
A numerical method for determining the optimal inspection calendar of critical equipment that fails according to a Weibull probability distribution was proposed, considering the minimization of total maintenance costs and the time interval between a potential failure and a functional failure (the so-called P–F interval, failure development period, or warning period) [16]. The Weibull distribution is widely used in reliability analysis because of its versatility in describing either failure (event) times or the lifetime of systems or their components subject to random failures or degradation phenomena (e.g., wear-out, corrosion, erosion, and fatigue).
Despite the advances in RCM, the current inspection maintenance optimization best practice still often relies on simplified closed-form formulas for the problem of finding the optimal inspection scheduling for two-unit systems [17]. For instance, to determine the optimal inspection period T (failure-finding interval) for a single independent protective sensor, the RCM standard guide standard suggests a simple formula, expressed in Equation (1), based only on the Mean Time Between Failures (MTBF) of the equipment, the sensor, and the (multiple-failure) system [2]:
T = 2   ×   M T B F s e n s o r   ×   M T B F e q u i p m e n t M T B F s y s t e m .
While formulas provide an intuitive approach to certain types of problems, they are often not suitable for defining the optimal inspection period for two-unit systems because of the computational complexities, dependencies, nonlinear relationships, uncertainties, and dynamic nature of such systems. Alternative methods, such as numerical, simulation, optimization, and multicriteria decision analysis methods, offer greater flexibility and can better accommodate the intricate relationships and real-world complexities inherent in these systems.
Additionally, most existing solutions focus only on a single optimization objective, limiting their applicability in real-world industrial practice [17,18]. For instance, Equation (1) does not account for related maintenance costs or other relevant organizational objectives. However, in maintenance optimization, there are often multiple and typically conflicting objectives, including maintenance cost minimization and equipment availability maximization, among others [6]. Although there are articles that use multicriteria decision analysis methodologies to select the best inspection period [19,20], we are not aware of any that specifically addresses the research problem of this article.
Ref. [18] conducted a systematic literature review on industrial maintenance decision-making to identify the efforts required for future work development and new integrated methodologies, concluding that the following difficulties and challenges are still present in current decision-making models and methodologies:
  • Lack of qualitative/quantitative integration between the correct treatment data (e.g., costs, probability of failures, equipment condition, and risk analysis) and the preferences of the decision maker regarding multiple objectives.
  • Definition of meaningful KPIs and creation of methods for transforming information into actionable knowledge for proactive maintenance decisions.
  • The integration of methodologies in which decision-making modeling is continuously fed using reliable results from a specific step into the models at another stage.
Despite these developments, existing studies still present important limitations. Many rely on single-objective optimization, typically cost minimization or availability maximization, and therefore fail to represent the conflicting nature of real industrial KPIs. Others assume perfect monitoring sensors, disregarding the impact of hidden failures on maintenance decisions. Furthermore, analytical closed-form formulas, while convenient, oversimplify system interactions and cannot adequately capture uncertainty or dynamic behavior. In contrast, the present work combines risk-informed modeling, discrete-event simulation, optimization techniques, and multicriteria decision analysis into a unified and transparent framework that explicitly accounts for sensor unreliability and multiple organizational objectives.
This literature review clearly highlights the need for developing risk-informed optimization-simulation models that support multicriteria decision-making for determining the optimal inspection scheduling of protective sensors monitoring critical equipment, accounting for their failure uncertainties and interdependence dynamics.

3. Integrated Methodological Approach

The target maintenance system, S , consists of two units: (a) an equipment, e (or one of its components), providing a critical function to the organization, which can fail and needs protection; (b) a sensor, s , providing a continuous online protective function to equipment, e , triggering an alarm when abnormal conditions are detected and, therefore, also subject to its own failures.
The proposed maintenance policy involves a periodic inspection of the protective sensor. Inspections occur at fixed and predetermined intervals T , resulting in a calendar schedule where each inspection i is performed at time t i .
Two system-level failure scenarios can occur when the protected equipment fails:
  • Normal detected failure: The sensor is operational and promptly identifies a potential failure f e in the equipment, which is immediately halted and restored within a repair time t r e and at unit cost c r e .
  • Multiple (undetected) failures in the system S : The sensor is ineffective or unavailable, and an equipment failure results in a significantly longer reactivation period, t r S and higher unit cost, c r S .
The following modeling assumptions apply:
  • Failure times of equipment and sensors are statistically independent.
  • Failure times follow a three-parameter Weibull probability distribution.
  • Sensor and equipment inspections are assumed to be perfect, revealing their true operational states. This assumption simplifies the initial model. Incorporating inspection errors (e.g., false negatives or false positives) constitutes a logical extension and would generally lead to shorter optimal inspection intervals to maintain the desired system’s reliability and equipment availability.
  • Inspection duration is assumed to be negligible. In practical applications, sensor inspections may require a brief equipment shutdown. The model can be easily refined to account for this by adding a fixed downtime period to each inspection without affecting the overall simulation logic or methodology.
  • Each sensor inspection incurs a fixed cost.
  • Repairs begin immediately upon failure detection.
  • Each unit is restored to its initial (as-new) condition after each repair or reactivation.
  • Repair and reactivation durations and costs are fixed constants.
  • A constant annual discount rate is applied.
The optimal inspection period T * is determined through a four-step methodology (see Figure 1):
  • Random Failure Time Generation: Generate failure times t f e and t f s according to Weibull distributions to model the uncertainty in the occurrence of equipment and sensor failures.
  • Discrete-Event Simulation: Execute a run until a multiple failure occurs and estimate consequences on KPIs (e.g., total maintenance costs, equipment availability) as a function of T and other relevant input parameters.
  • Monte Carlo Simulation: Repeat the random sampling procedure (in Step 1) to generate the distributions of KPI consequences (from Step 2) for each T .
  • Inspection Period Optimization: Select the T * that maximizes the risk-adjusted overall value using a risk-informed multicriteria decision analysis methodology.
The candidate inspection periods are evaluated over a discrete grid defined by a step size. In practice, the step size should be chosen small enough to ensure that the difference in overall value between adjacent T periods is negligible in the face of the uncertainty of the simulation results A practical guideline is to set the step size to no more than 5–10% of the expected inspection interval, and to perform iterative refinement around the candidate optimum to confirm the robustness of the solution.
The variables, parameters, equations, and methods used in this methodological sequence are detailed next.

3.1. Random Failure Times Generation for the Sensor and Equipment

Uncertainty in the time to failure of the equipment and sensor s is modeled using a three-parameter Weibull distribution (Figure 2).
The probability density function f(t) of each unit is represented by Equation (2):
f t = α β · t t 0 β α 1 · e x p t t 0 β α with   t t 0 ,   α > 0 ,   and   β > 0
where t represents the time of failure, e x p represents the base of natural logarithms (approximately 2.7183), t 0 is a location (or threshold) parameter that defines the lowest value that t can take and allowing to offset the distribution by the repair time t r of the unit that it re-enters into service after a failure (and restoring its initial capability as new), β is a positive scale (or spread) parameter defining the time t at which 63.2% of the units have failed and representing the characteristic life of the unit, and α is a positive shape (or slope) parameter reflecting the failure degradation mechanism of the unit such that it is one if the failure rate is random and either greater or less than one if that rate increases (e.g., early or rapid wear-out) or decreases (e.g., infant mortality) over time respectively. Parameters α and β should be determined by fitting empirical data from experiments, expert analysis, or theoretical values from the respective suppliers or from reliability data sources (e.g., OREDA database [21]).
The cumulative distribution function of the Weibull distribution F t can be derived by integrating the density function f t between t r (i.e., t 0 ) and t resulting in Equation (3):
F t = 1 e x p t t r β α .
By inverting Equation (3) and replacing F t by a uniform continuous random variable r n d between 0 and 1 results in Equation (4), which can be used to generate the random failure time t according to Equation (3) by randomly sampling values from r n d :
t = t r + β · ln 1 1 r n d 1 α .
If the unit has already survived up to time t , the respective probability of a failure can be determined by the conditional probability of failure F t t | t , given by Equation (5), which computes the probability that the unit will fail in time period t given that it has already survived time t (see Figure 3):
F t t t = F t F t 1 F t .
The inverse of Equation (5) results in Equation (6), where F t is computed by Equation (3). Hence, rather than Equation (4), Equation (6) must be used for generating the first random failure time t for a unit that has already accumulated some service time t at the start of the simulation:
t = β · ln 1 1 F t + r n d · 1 F t 1 α t .
After the first failure and subsequent repair, the unit is assumed to restore its initial condition as new. The next random failure time t is afterwards determined using Equation (4).
The repair time t r after a failure event is determined differently for the sensor and the equipment.
The sensor can have at most one failure within each inspection period, since it can only be detected at the time t i of the next inspection i . Time t i can be determined by the multiplication of T by the rounded-up integer of the division of the sensor failure time t f s by T . The sensor is then started repairing at time t i for a repair time t r s until it has been repaired at time t r f s , as expressed by Equation (7):
t r f s = t f s T · T + t r s .
The equipment can have multiple failures within each inspection period, since the sensor can promptly detect any equipment failure. The equipment is then immediately started repairing at the equipment failure time t f e for a repair time t r e until it has been repaired at time t r f e , as expressed by Equation (8):
t r f e = t f e + t r e .

3.2. Discrete-Event Simulation Model

A discrete-event simulation replicates the system until a multiple failure occurs, allowing the estimation of consequences for each T using the previously generated random failure times, and other relevant input parameters.

3.2.1. Determining the Occurrence of a Multiple Failure in the System

A multiple failure occurs at time t f S when an equipment failure occurs within an interval in which the sensor is unavailable, that is, between a sensor failure time t f s and the time t r f s until it is repaired. Figure 4 presents the procedure to find t f S along with the number of equipment failures n f e i and sensor failures n f s i occurring within each period t i = [ t i 1 ; t i   ] between each inspection.

3.2.2. Determining Total Maintenance Costs Until the Occurrence of a Multiple Failure

Total maintenance costs are computed based on the sum of the inspection costs, sensor repair costs, equipment repair costs, and system reactivation costs incurred during time horizon H = t f S + t r S , that is, up to the time when the equipment is reactivated ( t r S ) following a multiple failure ( t f S ).
The organizational time preference to incur costs later than sooner is modeled by discounting them with a discount rate d , allowing the conversion of any future cost F t into a commensurate amount representing its equivalent present cost P as expressed in Equation (9):
P = F t 1 + d t   .
Alternative time series of equivalent present cash outflows can then be meaningfully added up and compared. Discount rates should be set by the organization based on the opportunity costs associated with both its capital cost and the investment risk premium, that is, the average risk-adjusted return on investment that the organization can obtain from alternative investments.
Organizations typically offer discount rates annually. Maintenance events usually occur in shorter time periods though. Therefore, the annual discount rate could be converted into an equivalent rate d expressed in a more convenient time unit (e.g., minutes, hours, weeks) using Equation (10), where n u is the number of time units u (e.g., hours) within one year (e.g., 365 × 24 = 8760 h):
d = 1 + d 1 n u 1 .
Total sensor inspection present costs, T P i , incurred until a multiple failure occurs (in each simulation run) are determined by Equation (11):
T P i = i I c i 1 + d t i
where I is the number of inspections i , c i is the sensor inspection unit cost, and t i is the time of each inspection i . Total sensor repair present costs T P r s are determined by Equation (12):
T P r s = i I c r s · n f s i 1 + d t i + c r s 1 + d t f S
where c r s is the sensor repair unit cost, n f s i is the number of sensor failures (either one or null) within the respective t i , t i is the time of inspection i , and t f S is the system failure time. Note that after the last inspection I there is always an extra sensor repair cost at the time t f S since the sensor failure is detected only then. Total equipment repair present costs T P r e are determined using Equation (13):
T P r e = i I f e n f e i c r e 1 + d t f e , i + f e i = I + 1 n f e i 1 c r e 1 + d t f e , i
where c r e is the equipment repair unit cost, n f e i is the number of equipment failures within the respective t i , and t f e , i is the time of failure f e within the respective t i . Note that after the last inspection I there might be additional equipment repair costs for the first equipment failures f e before the system fails. Finally, total equipment (or system) reactivation present costs, T P r S , are determined using Equation (14):
T P r S = c r S 1 + d t f S
where c r S is the equipment (or system) reactivation unit cost. Since the previous total present costs, T P , are determined based on different time horizons for each simulation run, it is necessary to convert them into a series of (present) costs U C made at equal (uniform) time intervals to compare them. The equivalent uniform cost can be computed by the product of T P by two factors: a uniform time interval t u (e.g., 100 h) and a conversion factor f c as expressed in Equation (15):
f c = d 1 + d H 1 + d H 1 .
Uniform system total costs U C for each simulation run can then be computed as expressed in Equation (16):
U C = f c × t u × T P i + T P r s + T P r e + T P r S .

3.2.3. Determining Other KPIs Until the Occurrence of a Multiple Failure

Additional KPIs can be computed as required. A common KPI in maintenance is the equipment availability A e , that is, the proportion of time the equipment can perform its intended function, as expressed in Equation (17):
A e = t o t a l   t i m e u n a v a i l a b l e   t i m e t o t a l   t i m e = 1 N f e 1 Δ t r e + Δ t r S t f S + Δ t r S .
Other common metrics such as Mean Time To Failure (MTTF), Mean Time Between Failures (MTBF), or Mean Time to Repair (MTTR) may be used depending on the organization’s objective, but equipment availability is retained as the primary reliability KPI in this study.

3.3. Generating Distributions of Consequences on KPIs for Each Δ T Through a Monte Carlo Method

Uncertainty in the equipment and sensor failure behavior is propagated using Monte Carlo simulation. For each Δ T , the discrete-event simulation is repeated R runs, generating a distribution of KPI consequences.
For each KPI, expected values and two-tailed confidence interval for a confidence level 1 α are computed using Equation (18):
C I = x ¯ ± z 1 α 2 × s R
where s is the sample standard deviation and z 1 α 2 is the z-score from the standard normal distribution corresponding to a cumulative probability of 1 α 2 . The precision, stability, and reliability of the results increase with the number of runs set for the Monte Carlo simulation at the cost of higher computation demands in terms of memory and running time. The choice of the right trade-off between precision and performance time is context-dependent and should be determined by the organization based on the specific application.

3.4. Determining the Optimal Inspection Period Δ T * Through a Risk-Informed Multicriteria Value Measurement Method

Organizations often face multiple conflicting performance objectives, including costs, availability or reliability, product quality or throughput, customer service, safety, and environment. Without loss of generality to further objectives, the optimal inspection period Δ T * will be determined in this case study for only two of the most used conflicting objectives in maintenance problems: minimizing total maintenance costs (Equation (16)) and maximizing equipment availability (Equation (17)). Maintenance cost and equipment availability were selected because they are among the most widely adopted and operationally critical KPIs in industrial maintenance. The framework remains fully extensible and can incorporate additional objectives.
Conflicting objectives involve a trade-off: improving the performance on one KPI can only be achieved by compromising the performance on the other, and vice versa. However, organizations must still choose the best Δ T * between conflicting objectives. Multicriteria value measurement methods (MCVM) provide a solution to this conundrum by modeling the intra-criterion and inter-criterion preferences of the organization regarding multiple objectives.
Intra-criterion preferences on each KPI can be defined by value functions v K P I representing how much the organization values marginal consequences on the respective KPI, and inter-criteria preferences by weights w K P I , representing how much the organization is willing to lose on one KPI to increase on another.
The optimal inspection period, Δ T * , is then the one that maximizes the overall value V Δ T of each potential alternative computed by the additive multicriteria value model expressed in Equation (19):
Δ T * = arg max Δ t V Δ T = arg max Δ t K P I w K P I × v K P I Δ T   a n d   K P I w K P I = 1 .
The MCVM literature describes various well-established techniques for constructing a value function for continuous KPIs based on the elicitation of preferences from the decision-maker, namely, the bisection method [22], direct rating [23], or MACBETH [24]. The bisection method provides a simple and sound protocol for assessing either linear or exponential functions on each KPI: (a) it starts by scoring the worst (or a bad) consequence K P I with a value score v K P I = 0 , and the best (or a good) consequence K P I + with a value score v K P I + = 100 ; (b) it then asks the decision-maker to identify the intermediate consequence K P I that makes them indifferent between either improving K P I to K P I or improving K P I to K P I + ; and (c) it finally fits a function that passes through the points K P I , 0 ,   K P I , 50 ,   a n d K P I + , 100 . If the resulting function is linear, its closed-form expression corresponds to Equation (20). Otherwise, it can be fitted through the exponential function respecting the delta property, expressed in Equation (21):
v K P I = 100 × K P I K P I K P I + K P I
v K P I = 100 × 1 e x p K P I K P I ρ 1 e x p K P I + K P I ρ .
The MCVM literature also describes sound procedures for assessing the weights associated with the KPIs based on the elicitation of preferences from the decision-maker, namely the trade-off procedure [25], swing weighting [23], and MACBETH [26]. The trade-off procedure provides a practical and sound questioning protocol for determining the weights associated with the KPIs in this setting because decision-makers usually find it easier to assess non-monetary KPIs against a monetary KPI, such as the total maintenance costs U C . This procedure comprises the following: (a) for each K P I \ { U C } ask the decision-maker to identify the total maintenance cost U C K P I that makes him (or her) indifferent between a fictitious alternative A defined by the consequences U C and K P I + and a fictitious alternative B defined by U C K P I and K P I ; (b) compute the weights w K P I = v U C K P I K P I v U C K P I where v U C U C = 100 .
A notable special case occurs when v U C is assessed to be linear, where weights w K P I can be more simply determined by: (a) asking the decision-maker to assess how much U C K P I is he (or she) willing to pay for improving each KPI from K P I to K P I + ; (b) computing the weight w K P I = U C K P I K P I U C K P I . Additionally, in this special case, steps (b) and (c) of the previously described bisection method can be rather performed by: (b) asking the decision-maker to assess how much U C K P I is he (or she) willing to pay for improving each KPI from K P I to a given intermediate consequence K P I ; (c) fitting a function that passes through the points K P I , 0 ,   K P I , 100 × U C K P I U C K P I + ,   a n d K P I + , 100 .
Decision alternatives comprise a set of evenly distributed inspection periods Δ T . Uncertainty in sensor and equipment failures propagates through the discrete-event simulation model and the additive multicriteria value model, resulting in an uncertain overall value V Δ T for each Δ T . This uncertainty can be represented as a risk profile encapsulating the range of possible overall values V Δ T and their associated probabilities. If the cumulative risk profile of one alternative Δ T lies entirely to the right of another, the former stochastically dominates the latter [27]. When a single Δ T stochastically dominates all others, it is deemed optimal. Otherwise, in contexts where organizations face many similar decisions, a risk-neutral preference is generally the most appropriate risk attitude. Accordingly, the optimal inspection period Δ T * should correspond to the Δ T with the highest expected value, V ¯ Δ T , as defined by the additive multicriteria value model in Equation (19).
Any potential optimal decision must always be stress-tested against the range of significant uncertainties identified during the construction of the respective decision support model, namely, the parameters of the additive multicriteria value model. The sensitivity analysis can inform the organization on the robustness of the preliminary decision to select a given optimal inspection period against changes in the KPIs’ weights.

4. Case Study

A case study of the methodology described in Section 3 is presented next to demonstrate and validate its application in a real-world decision context. The case study will be illustrated using specific inputs but is generalizable to any other scenarios. Although the approach has been implemented in Portuguese organizations through decision support systems specifically developed for user interaction (see Figure 5), the article proceeds by presenting a prototype Excel application, which can be downloaded from the Supplementary File, for reproducibility, reuse, clarity, transparency, and the reader’s convenience. The proposed methodological approach is independent of the software platform: all steps (random failure generation, discrete event simulation, Monte Carlo analysis, multicriteria methods, risk analysis) can be implemented in alternative environments (e.g., Python 3.0, R 4.2, MATLAB R2022b). The Excel-based DSS should therefore be interpreted as a pedagogical and technology transfer tool rather than a methodological constraint. Additionally, the use of probability management principles ensures that all stochastic elements remain fully traceable and reproducible.

4.1. Setting Input Parameters and Reference Units

Figure 6 presents the inputs interface (sheet “Inputs”) where the user can set all the input parameters, including their reference units. Symbols used in the notation of Section 3 are depicted within parentheses. All the remaining sheets are fed in from the values filled in this interface. Time and monetary units can be changed. The general data section also allows setting the annual discount rate d of the organization and a uniform reference period t u of the convenience. The optimization and simulation sections allow setting a particular inspection period Δ T (e.g., 280 h) and simulation run r (e.g., one) to verify the respective outcomes in the other interfaces (sheets). The optimization section also allows defining a set of evenly spaced inspection periods from a minimum of 30 to a maximum of 300 and a step of 10. The initial step should be small enough so that differences in the overall value V(∆T) between adjacent intervals are negligible compared to the result uncertainty. An iterative refinement with a finer step around a candidate optimum is recommended to confirm robustness. The total number of runs R and a particular confidence level ( 1 α ) associated with the desired margin of error of the expected values can be set in the simulation section. The equipment and sensor sections allow filling in the respective parameters t , α , and β of the Weibull distribution representing their random failure time, repair time periods Δ t r e and Δ t r s , and repair unit costs c r e and c r s . The unit cost c i of each inspection can be set in the sensor section, while the equipment’s reactivation unit cost, c r S , and reactivation time, Δ t r S can be set in the equipment section.

4.2. Generation of Random Failure Times for the Sensor and Equipment

Figure 7 and Figure 8 show the generated random failure times for the sensor and equipment. The top panels use data from the “Input” sheet (Figure 6) to compute these values for each simulation run and inspection period. The bottom panels present a set of pre-generated two hundred uniform random real numbers between 0 and 1 using the Excel function RAND() and then frozen and stored for each simulation run.
These streams are called Stochastic Information Packets (SIPs) by the discipline of probability management [28], allowing auditing, transparency, and understanding of the computations carried out in each simulation run by the user.
Figure 7 illustrates how t f s and t r f s are computed based on these data for r = 1 and Δ T = 280 h. The first random failure time is computed through Equation (6) with F t being determined by Equation (3). Repair times are determined by Equation (7). The remaining random failure times are computed through Equation (4).
Figure 8 illustrates how t f e and t r f e are computed for r = 1 and Δ T = 280 h. The first random failure time is computed through Equation (6), with F t being determined by Equation (3). Repair times are determined by Equation (8). The remaining random failure times are computed through Equation (4).

4.3. Determining the System’s Multiple Failure Time and Consequences on the KPIs

The output interface for the discrete-event simulation model and resulting consequences on the KPIs is presented in Figure 9 (sheet “System”). The bottom panel of this interface presents the results of the discrete-event simulation model described in Section 3.2.1 until a multiple failure occurs in the system at time t f S . Figure 9 illustrates the computation of t f S for r = 1 and Δ T = 280 h. After 12 equipment failures ( N f e ) and four sensor failures ( N f s ), a system failure occurs at time t f S = 5502 h, after the 19th and before the 20th inspection is conducted. During this period, the equipment fails a first time at t f e = 5322 h, but since the sensor is operational, the equipment is promptly repaired in Δ t r e = 1 h and is back working at t r f e = 5323 h; later on, the equipment fails again at time t f e = 5502 h, but since at this time the sensor is not operational (from t f s = 5374 to t r f s = 5608 h), a multiple failure occurs at t f S = 5502 h. The last four columns from the bottom panel in Figure 9 show the total present costs associated with sensor inspections ( T P i ), sensor repairs ( T P r s ), equipment repairs ( T P r e ), and equipment reactivation ( T P r S ) computed as described in Section 3.2.2 using Equations (10)–(14). The top panel in Figure 9 presents the relevant inputs, and the resulting consequences on the two organizational KPIs. Total maintenance costs are measured through the respective uniform total maintenance costs U C , determined using Equations (15) and (16) described in the last paragraphs of Section 3.2.2, considering a uniform reference period t u equal to 100 h. Equipment availability A e is computed using Equation (17) described in Section 3.2.3.

4.4. Estimating the Distribution of Consequences of Distinct Inspection Periods on KPIs Through Monte Carlo Simulation

A Monte Carlo simulation method is called from the “Simulations” sheet for each KPI: total maintenance costs, U C , and equipment availability A e . The number R of runs and the set of inspections periods Δ T to simulate are defined in the “Inputs” sheet and automatically updated in this sheet. The simulation is conducted by taking advantage of the Excel “Data Table” feature as suggested by the discipline of probability management. Its advantages over running a VBA procedure are two-fold: it is much faster, and it allows recording and auditing each result. Figure 10 and Figure 11 illustrate excerpts of the results on the two KPIs for R = 5000 and for inspection periods Δ T ranging from 30 to 300 h in increments of 10 h.
Figure 12 and Figure 13 show the expected values and 95% confidence intervals of maintenance costs ( U C ) and equipment availability ( A e ), respectively, as a function of the inspection periods Δ T for R = 5000. On the one hand, it can be concluded from Figure 12 that the optimal inspection period from the point of view of minimizing maintenance costs is Δ T * = 120 h. On the other hand, from the point of view of maximizing equipment availability, Figure 13 clearly shows that the smaller the Δ T , the better the equipment availability.

4.5. Selecting the Optimal Inspection Period That Maximizes the Risk-Adjusted Overall Value to the Organization

Since it is not possible to define a Δ T that is optimal for both KPIs, the organization must choose the optimal inspection period Δ T * based on its intra-criterion and inter-criterion preferences. Since both KPIs are preferential and independent from the perspective of the organization, the additive multicriteria value model expressed in Equation (19) can be applied to evaluate the overall value V Δ T of each potential Δ T .
The value function v U C for maintenance costs ( U C ) was assessed as linear by the organization, using the bisection method and defined as v U C = 100 U C / 60 , based on Equation (20), with U C = 6000 EUR/100 h and U C + = 0 EUR/100 h, considering the range of results obtained for U C in Section 4.4 (Figure 10).
Taking advantage of v U C being linear, the value function v A e for equipment availability ( A e ) and the weights w U C and w A e for both KPIs were assessed by asking the organization’s decision-maker two questions: (a) How much is he (or she) willing to pay in EUR/100 h for improving equipment availability from 0% to 100%? (b) How much is he (or she) willing to pay in EUR/100 h for improving equipment availability from 0% to 50%? Based on his (or her) response of 9000 EUR/100 h to question (a), KPI’s weights were determined as w U C = 6000/(6000 + 9000) = 40% and w A e = 9000/(6000 + 9000) = 60% by applying the special case of the tradeoff procedure described in Section 3.4. Likewise, based on a response of 7550 EUR/100 h to question (b), the value function for equipment availability was assessed as non-linear and defined exponentially as v A e = 100 × 1 e x p A e / 0.3 / 1 e x p 1 / 0.3 , with A e = 0%, A e + = 100%, and ρ = −0.3, determined by fitting Equation (21) to the points (0%, 0), (50%, 83.88), and (100%, 100) as described in Section 3.4 for the special case of the bisection method.
The overall values V Δ T for each inspection period Δ T , ranging from 30 to 300 h in increments of 10 h, were computed from five thousand simulation runs. These values were derived by combining the respective consequences on U C (Figure 10) and A e (Figure 11) and then aggregating them using the additive model defined in Equation (19) with the specified weights and value functions. Figure 14 illustrates the respective cumulative risk profiles, as described in Section 3.4, for a subset of representative Δ T , showing that selecting the optimal inspection period based on (first order) stochastic dominance is not possible.
The optimal inspection period Δ T * should be selected based on the best expected value V ¯ Δ T . Figure 15 shows the expected values and 95% confidence intervals of the overall value V Δ T as a function of the inspection periods Δ T , computed from R = 5000 runs, indicating that the optimal inspection period is Δ T * = 90 h.
A sensitivity analysis on the weights w K P I allowed to verify their impact on the selection of the optimal inspection period Δ T * from the set of inspection periods Δ T ranging from 30 to 300 h in increments of 10 h. It was found that the selection of Δ T * = 90 h is robust to changes in w U C between 18% and 65%, and, consequently, to changes in w A e between 35% and 82%. These variations correspond to an interval between 3300 and 27,000 EUR/100 h (rather than 9000) regarding the amount that the organization is willing to pay for improving equipment availability from 0% to 100%, suggesting that the decision is indeed robust. Figure 16 plots the efficient frontier connecting the partial values v U C and v A e of all evaluated inspection periods Δ T along with the isoline representing the optimal overall value V Δ T * = 90 = 94.12 , with w U C = 40% and w A e = 60%. Dashed lines represent the two iso-value lines for the weights that change the optimal inspection period Δ T * from 90 h to: (a) 80 h, when w U C = 18% and w A e = 82% and (b) 120 h, when w U C = 65% and w A e = 35%.

5. Discussion

This study presents a novel, integrated, and generic multimethodological approach for determining the optimal inspection period for protective sensors monitoring critical equipment. It considers multiple KPIs and the uncertainty associated with hidden failures in the system and the organization’s preferences. A case study illustrating the application of the proposed approach to a real-world maintenance problem via a DSS demonstrates its practical applicability and contributes to the limited literature on proactive maintenance modeling and optimization applied in practice [15]. The following sections summarize the main findings, implications, and limitations.

5.1. Findings

The modeling approach and DSS were developed to answer the five key research questions, identified in the Introduction, and commonly encountered in real-world applications of this maintenance problem. The results demonstrate that the proposed solution effectively addresses each specific research question as follows:
  • The use of three-parameter Weibull probability density functions, Monte Carlo simulation, and the principles of the probability management discipline enable an effective and auditable representation of the uncertainties related to failures in the equipment and sensors.
  • Discrete-event simulation modeling allows for accurate prediction of the consequences of these failures on multiple KPIs, considering varying sensor inspection periods, costs, time periods, and other input parameters.
  • Risk analysis and value-based multicriteria decision analysis methods facilitate risk-informed assessment of the overall value of different sensor inspection periods considering both their uncertain consequences and the preferences of the organization on multiple conflicting organizational KPIs, namely by representing the intertemporal, intra-criterion, and inter-criteria preferences of the organization through discount rate, value functions, and weights of the KPIs, respectively.
  • Risk-informed multicriteria value optimization methods, namely stochastic dominance and/or expected value maximization, provide a sound criterion for determining the optimal inspection period, while sensitivity analysis allows assessing the robustness of the optimal decision against the uncertainty of the organization’s preferences.
  • Finally, organizational buy-in and the successful implementation of maintenance decisions recommended by a decision support model require not only the scientific rigor and integration of the underlying approach, but also key success factors related to the DSS, such as its transparency, openness, understandability, user-friendliness, and auditability. These qualities were ensured by developing a prototype in Microsoft Excel, utilizing only built-in functions and taking advantage of the principles of probability management.
This research addresses a significant gap in the maintenance literature regarding the need for integrated methodologies that combine quantitative and qualitative data, uncertainty and risk analysis, and organizational preferences across multiple KPIs.
The proposed approach is generic, enabling it to accommodate different parameters and integrate additional assumptions, such as incorporating new KPIs, using alternative probability distributions for modeling failure uncertainty, or applying other MCVM methods for determining intra-criterion and/or inter-criteria preferences.

5.2. Implications

The proposed framework contributes to the state-of-the-art by providing a practical and robust approach for optimizing inspection scheduling for protective sensors monitoring critical equipment subject to uncertain hidden failures. In contrast to the traditional maintenance models, which often rely on simplified assumptions and deterministic approaches, this framework explicitly accounts for the uncertainty associated with sensor and equipment failures as well as organizational preferences, aligning with recommendations for more risk-informed multicriteria decision-making in industrial maintenance.
Another key contribution to practice is the development and provision of an open-access DSS in Excel, increasing transparency, comprehensibility, ease of use, and making it valuable both practical decision-making and educational purposes.

5.3. Limitations

The main limitations of the results include the following. The case study focused on a single protection sensor in a specific utility context, which limits generalization. The computational time required to run thousands of simulations can be significant, although using the Table Data feature from Excel instead of VBA macros reduces execution from hours to minutes. Another limitation regards the comparison of discrete inspection periods only, which might miss a better inspection period compared to the continuous case. One solution to this limitation comprises running the same problem again with a set of less evenly spaced inspection periods around the previous optimal inspection period. For instance, in the presented case study, running the DSS again with the optimization input parameters set with a minimum of 80 h, a maximum of 100 h, and a step of 1 h. Minor limitations, which can be easily addressed by model refinement, include modeling sensor inspection time if the sensor is not available to perform its protecting function while the inspection is performed; and probabilistically modeling the duration and/or costs of other maintenance tasks, namely regarding inspection, repair, and reactivation. Modeling assumptions such as perfect inspections, independent failure processes, and Weibull-distributed failure times may not fully capture real system complexities. Imperfect inspections or correlated failures could alter optimal intervals and risk profiles, and alternative distributions may be more suitable in some systems. These factors do not invalidate the framework but should be considered when generalizing the results.

6. Conclusions

This research addresses the problem of determining the optimal inspection period for protective sensors subject to hidden failures. The proposed approach combines simulation, optimization, and multicriteria decision analysis to support this complex decision.
The key finding is that a robust optimal inspection period can be identified by jointly modeling the randomness of failures and the organization’s preferences and risk attitude. In the case study, the period of 90 h represented the best compromise between cost and availability. This result provides an objective and auditable basis for maintenance policies, with a direct impact on reducing unplanned downtime and increasing operational reliability, which are central goals of proactive maintenance.
The proposed approach addresses the underlying research questions required to tackle this maintenance problem by: (a) modeling failure uncertainties using probability density functions, Monte Carlo simulation, and probability management: (b) predicting failure consequences on KPIs through discrete-event simulation; (c) comparing the overall value of alternative inspection periods using multicriteria value measurement and risk analysis methods; (d) determining the optimal inspection period considering uncertainties and preferences by stochastic dominance, expected value maximization, and sensitivity analysis; and (e) developing a user-friendly decision support system for practical implementation.
The developed framework and decision support system empower organizations to make risk-informed multicriteria decisions regarding sensor inspection schedules, leading to optimized industrial maintenance. This translates to reduced downtime, lower maintenance costs, and improved overall equipment effectiveness. The proposed integrated approach represents an advancement over traditional methods that often rely on simplified assumptions and single-criterion optimization. The methodology is readily applicable across various industries reliant on critical equipment with protective sensors, such as power generation, manufacturing, and transportation. By adopting this approach, organizations can enhance their proactive maintenance practices and improve operational efficiency.
Although the case study demonstrates the feasibility and usefulness of the proposed framework, additional empirical validation in real operational environments remains an important direction for future research. Implementing the optimized inspection schedule in practice and monitoring the resulting performance over time would provide evidence of real-world benefits and support the continuous improvement of the model.
Future research should explore the integration of real-time condition monitoring data into the proposed approach, allowing for even more dynamic and adaptive inspection schedules. Further investigation into the scalability of the framework for modeling complex networks with multiple protective sensors interconnected with critical equipment can also offer valuable insights. Two sensors can be used either as a protective function, or one can replace the other while it is being repaired, providing redundancy but with additional costs. Additionally, exploring the relevance of modeling non-neutral risk attitudes on the optimal inspection period could be investigated. We encourage researchers and practitioners to adopt and adapt this framework to their specific contexts, contributing to the advancement of proactive maintenance practices and improved asset management.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s26010213/s1, Prototype Excel application mentioned in the Case Study.

Author Contributions

Conceptualization, R.J.G.M. and R.A.; methodology, R.J.G.M. and R.A.; software, R.J.G.M. and R.A.; validation, all authors; writing—original draft preparation, R.J.G.M. and R.A.; writing—review and editing, P.C.M., A.D.B.M., J.C.A.R. and F.S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article and Supplementary Material. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DSSDecision support system
KPIKey performance indicator
MCVMMulticriteria value measurement methods
MTBFMean time between failures
MTTFMean time to failure
MTTRMean time to repair
RCMReliability centered maintenance
VBAVisual basic for applications

References

  1. SIEMENS AG. The True Cost of Downtime 2024; Siemens AG: Erlangen, Germany. Available online: https://assets.new.siemens.com/siemens/assets/api/uuid:1b43afb5-2d07-47f7-9eb7-893fe7d0bc59/TCOD-2024_original.pdf (accessed on 25 December 2025).
  2. SAE JA1012; A Guide to the Reliability-Centered Maintenance (RCM) Standard. SAE International: Warrendale, PA, USA, 2002.
  3. Zhang, Y.; Ouyang, L.; Meng, X.; Zhu, X. Condition-Based Maintenance Considering Imperfect Inspection for a Multi-State System Subject to Competing and Hidden Failures. Comput. Ind. Eng. 2024, 188, 109856. [Google Scholar] [CrossRef]
  4. Benhanifia, A.; Ben Cheikh, Z.; Oliveira, P.M.; Valente, A.; Lima, J. Systematic Review of Predictive Maintenance Practices in the Manufacturing Sector. Intell. Syst. Appl. 2025, 26, 200501. [Google Scholar] [CrossRef]
  5. Ahmed Murtaza, A.; Saher, A.; Hamza Zafar, M.; Kumayl Raza Moosavi, S.; Faisal Aftab, M.; Sanfilippo, F. Paradigm Shift for Predictive Maintenance and Condition Monitoring from Industry 4.0 to Industry 5.0: A Systematic Review, Challenges and Case Study. Results Eng. 2024, 24, 102935. [Google Scholar] [CrossRef]
  6. Zhu, T.; Ran, Y.; Zhou, X.; Wen, Y. A Survey of Predictive Maintenance: Systems, Purposes and Approaches. arXiv 2019, arXiv:1912.07383. [Google Scholar]
  7. Shen, J.; Hu, J.; Ma, Y. Two Preventive Replacement Strategies for Systems with Protective Auxiliary Parts Subject to Degradation and Economic Dependence. Reliab. Eng. Syst. Saf. 2020, 204, 107144. [Google Scholar] [CrossRef]
  8. Fu, S.; Avdelidis, N.P. Prognostic and Health Management of Critical Aircraft Systems and Components: An Overview. Sensors 2023, 23, 8124. [Google Scholar] [CrossRef] [PubMed]
  9. Rodrigues, J.A.; Farinha, J.T.; Cardoso, A.M.; Mendes, M.; Mateus, R. Prediction of Sensor Values in Paper Pulp Industry Using Neural Networks. In Proceedings of IncoME-VI and TEPEN 2021; Zhang, H., Feng, G., Wang, H., Gu, F., Sinha, J.K., Eds.; Mechanisms and Machine Science; Springer International Publishing: Cham, Switzerland, 2023; Volume 117, pp. 281–291. ISBN 978-3-030-99074-9. [Google Scholar]
  10. Zonta, T.; da Costa, C.A.; Righi, R.d.R.; de Lima, M.J.; da Trindade, E.S.; Li, G.P. Predictive maintenance in the Industry 4.0: A systematic literature review. Comput. Ind. Eng. 2020, 150, 106889. [Google Scholar] [CrossRef]
  11. Garcia, J.; Rios-Colque, L.; Peña, A.; Rojas, L. Condition Monitoring and Predictive Maintenance in Industrial Equipment: An NLP Assisted Review of Signal Processing, Hybrid Models, and Implementation Challenges. Appl. Sci. 2025, 15, 5465. [Google Scholar] [CrossRef]
  12. Mateus, B.C.; Farinha, J.T.; Mendes, M. Fault Detection and Prediction for Power Transformers Using Fuzzy Logic and Neural Networks. Energies 2024, 17, 296. [Google Scholar] [CrossRef]
  13. Roux, M.; Fang, Y.-P.; Barros, A. Impact of Imperfect Monitoring on the Optimal Condition-Based Maintenance Policy of a Single-Item System. In Proceedings of the Book of Extended Abstracts for the 32nd European Safety and Reliability Conference, Dublin, Ireland, 28 August–1 September 2022; Research Publishing Services: Singapore, 2022; pp. 658–664. [Google Scholar]
  14. Zhao, X.; Guo, B.; Chen, Y. A Condition-Based Inspection-Maintenance Policy for Critical Systems with an Unreliable Monitor System. Reliab. Eng. Syst. Saf. 2024, 242, 109710. [Google Scholar] [CrossRef]
  15. De Jonge, B.; Scarf, P.A. A Review on Maintenance Optimization. Eur. J. Oper. Res. 2020, 285, 805–824. [Google Scholar] [CrossRef]
  16. Assis, R.; Marques, P.C. A Dynamic Methodology for Setting Up Inspection Time Intervals in Conditional Preventive Maintenance. Appl. Sci. 2021, 11, 8715. [Google Scholar] [CrossRef]
  17. Werbińska-Wojciechowska, S. Inspection Models for Technical Systems. In Technical System Maintenance; Springer Series in Reliability Engineering; Springer International Publishing: Cham, Switzerland, 2019; pp. 101–159. ISBN 978-3-030-10787-1. [Google Scholar]
  18. Ruschel, E.; Santos, E.A.P.; Loures, E.D.F.R. Industrial Maintenance Decision-Making: A Systematic Literature Review. J. Manuf. Syst. 2017, 45, 180–194. [Google Scholar] [CrossRef]
  19. Park, M.; Lim, J.-H.; Kim, D.-K.; Kim, J.-W.; Park, D.H.; Park, B.-N.; Park, J.H.; Song, M.-O.; Sung, S.-I.; Jung, K.M. A Study to Determine Optimal Inspection Intervals Using a Multi-Criteria Decision Model. J. Appl. Reliab. 2021, 21, 282–293. [Google Scholar] [CrossRef]
  20. Zhao, X. Multi-Criteria Decision Model for Imperfect Maintenance Using Multi-Attribute Utility Theory. Int. J. Perform. Eng. 2018, 14, 3014. [Google Scholar] [CrossRef]
  21. Veritas, D.N. OREDA Offshore Reliability Data Handbook, 4th ed.; OREDA Participants: Høvik, Norway, 2002; ISBN 978-82-14-02705-1. [Google Scholar]
  22. Goodwin, P.; Wright, G. Decision Analysis for Management Judgment, 5th ed.; Wiley: Hoboken, NJ, USA, 2014; ISBN 978-1-118-74073-6. [Google Scholar]
  23. von Winterfeldt, D.; Edwards, W. Decision Analysis and Behavioral Research; Cambridge University Press: Cambridge, MA, USA, 1993; ISBN 978-0-521-27304-6. [Google Scholar]
  24. Bana, E.; Costa, C. MACBETH—An Interactive Path towards the Construction of Cardinal Value Functions. Int. Trans. Oper. Res. 1994, 1, 489–500. [Google Scholar] [CrossRef]
  25. Keeney, R.L.; Raiffa, H. Decisions with Multiple Objectives: Preferences and Value Trade-Offs, 1st ed.; Cambridge University Press: Cambridge, MA, USA, 1993; ISBN 978-0-521-43883-4. [Google Scholar]
  26. Bana e Costa, C.A.; Vansnick, J.-C. Applications of the MACBETH Approach in the Framework of an Additive Aggregation Model. J. Multi-Criteria Decis. Anal. 1997, 6, 107–114. [Google Scholar] [CrossRef]
  27. Clemen, R.T.; Reilly, T. Making Hard Decisions with Decision Tools; South-Western/Cengage Learning: Mason, OH, USA, 2014; ISBN 978-0-538-79757-3. [Google Scholar]
  28. Probability Management. Available online: https://www.probabilitymanagement.org/ (accessed on 16 December 2024).
Figure 1. Diagram logic of the methodology to select the optimal T * .
Figure 1. Diagram logic of the methodology to select the optimal T * .
Sensors 26 00213 g001
Figure 2. Probability density f(t) and cumulative distribution F(t) functions of the Weibull distribution.
Figure 2. Probability density f(t) and cumulative distribution F(t) functions of the Weibull distribution.
Sensors 26 00213 g002
Figure 3. Conditional probability of failure F t t | t in time t t having already survived until time t .
Figure 3. Conditional probability of failure F t t | t in time t t having already survived until time t .
Sensors 26 00213 g003
Figure 4. Discrete-event simulation model to find the system multiple failure time.
Figure 4. Discrete-event simulation model to find the system multiple failure time.
Sensors 26 00213 g004
Figure 5. DSS was implemented in a Portuguese electric utility organization.
Figure 5. DSS was implemented in a Portuguese electric utility organization.
Sensors 26 00213 g005
Figure 6. User inputs interface (sheet “Inputs”).
Figure 6. User inputs interface (sheet “Inputs”).
Sensors 26 00213 g006
Figure 7. Sensor random failure times interface (sheet “Sensor”).
Figure 7. Sensor random failure times interface (sheet “Sensor”).
Sensors 26 00213 g007
Figure 8. Equipment random failure interface (sheet “Equipment”).
Figure 8. Equipment random failure interface (sheet “Equipment”).
Sensors 26 00213 g008
Figure 9. System (multiple) failure and KPIs interface (sheet “System”).
Figure 9. System (multiple) failure and KPIs interface (sheet “System”).
Sensors 26 00213 g009
Figure 10. Total maintenance costs U C across 5000 runs for inspection periods Δ T = 30 to 300 h.
Figure 10. Total maintenance costs U C across 5000 runs for inspection periods Δ T = 30 to 300 h.
Sensors 26 00213 g010
Figure 11. Equipment availability A e across 5000 runs for inspection periods Δ T = 30 to 300 h.
Figure 11. Equipment availability A e across 5000 runs for inspection periods Δ T = 30 to 300 h.
Sensors 26 00213 g011
Figure 12. Expected values (thick central line) and 95% confidence intervals (thin outer lines) of maintenance costs ( U C ) for inspection periods Δ T = 30 to 300 h computed from 5000 simulation runs.
Figure 12. Expected values (thick central line) and 95% confidence intervals (thin outer lines) of maintenance costs ( U C ) for inspection periods Δ T = 30 to 300 h computed from 5000 simulation runs.
Sensors 26 00213 g012
Figure 13. Expected values (thick central line) and 95% confidence intervals (thin outer lines) of equipment availability ( A e ) for inspection periods Δ T = 30 to 300 h computed from 5000 simulation runs.
Figure 13. Expected values (thick central line) and 95% confidence intervals (thin outer lines) of equipment availability ( A e ) for inspection periods Δ T = 30 to 300 h computed from 5000 simulation runs.
Sensors 26 00213 g013
Figure 14. Cumulative risk profiles for the overall value V Δ T of inspection periods Δ T = 30, 50, 70, 90, 120, and 300 h (The x-axis was truncated before 75 to enhance clarity).
Figure 14. Cumulative risk profiles for the overall value V Δ T of inspection periods Δ T = 30, 50, 70, 90, 120, and 300 h (The x-axis was truncated before 75 to enhance clarity).
Sensors 26 00213 g014
Figure 15. Expected values (thick central line) and 95% confidence intervals (thin outer lines) of overall value V Δ T for inspection periods Δ T = 30 to 300 h from 5000 simulation runs.
Figure 15. Expected values (thick central line) and 95% confidence intervals (thin outer lines) of overall value V Δ T for inspection periods Δ T = 30 to 300 h from 5000 simulation runs.
Sensors 26 00213 g015
Figure 16. KPI value space. The dotted curve represents the values v U C and v A e of inspection periods Δ T = 30 to 300 h. The solid line is the optimal V Δ T * = 90 . The dashed lines represent V Δ T * = 80 and V Δ T * = 120 .
Figure 16. KPI value space. The dotted curve represents the values v U C and v A e of inspection periods Δ T = 30 to 300 h. The solid line is the optimal V Δ T * = 90 . The dashed lines represent V Δ T * = 80 and V Δ T * = 120 .
Sensors 26 00213 g016
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mateus, R.J.G.; Assis, R.; Carmona Marques, P.; Martins, A.D.B.; Rodrigues, J.C.A.; Pinto, F.S. An Integrated Risk-Informed Multicriteria Approach for Determining Optimal Inspection Periods for Protective Sensors. Sensors 2026, 26, 213. https://doi.org/10.3390/s26010213

AMA Style

Mateus RJG, Assis R, Carmona Marques P, Martins ADB, Rodrigues JCA, Pinto FS. An Integrated Risk-Informed Multicriteria Approach for Determining Optimal Inspection Periods for Protective Sensors. Sensors. 2026; 26(1):213. https://doi.org/10.3390/s26010213

Chicago/Turabian Style

Mateus, Ricardo J. G., Rui Assis, Pedro Carmona Marques, Alexandre D. B. Martins, João C. Antunes Rodrigues, and Francisco Silva Pinto. 2026. "An Integrated Risk-Informed Multicriteria Approach for Determining Optimal Inspection Periods for Protective Sensors" Sensors 26, no. 1: 213. https://doi.org/10.3390/s26010213

APA Style

Mateus, R. J. G., Assis, R., Carmona Marques, P., Martins, A. D. B., Rodrigues, J. C. A., & Pinto, F. S. (2026). An Integrated Risk-Informed Multicriteria Approach for Determining Optimal Inspection Periods for Protective Sensors. Sensors, 26(1), 213. https://doi.org/10.3390/s26010213

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop