Accendo Reliability

Your Reliability Engineering Professional Development Site

  • Home
  • About
    • Contributors
    • About Us
    • Colophon
    • Survey
  • Reliability.fm
    • Speaking Of Reliability
    • Rooted in Reliability: The Plant Performance Podcast
    • Quality during Design
    • CMMSradio
    • Way of the Quality Warrior
    • Critical Talks
    • Asset Performance
    • Dare to Know
    • Maintenance Disrupted
    • Metal Conversations
    • The Leadership Connection
    • Practical Reliability Podcast
    • Reliability Hero
    • Reliability Matters
    • Reliability it Matters
    • Maintenance Mavericks Podcast
    • Women in Maintenance
    • Accendo Reliability Webinar Series
  • Articles
    • CRE Preparation Notes
    • NoMTBF
    • on Leadership & Career
      • Advanced Engineering Culture
      • ASQR&R
      • Engineering Leadership
      • Managing in the 2000s
      • Product Development and Process Improvement
    • on Maintenance Reliability
      • Aasan Asset Management
      • AI & Predictive Maintenance
      • Asset Management in the Mining Industry
      • CMMS and Maintenance Management
      • CMMS and Reliability
      • Conscious Asset
      • EAM & CMMS
      • Everyday RCM
      • History of Maintenance Management
      • Life Cycle Asset Management
      • Maintenance and Reliability
      • Maintenance Management
      • Plant Maintenance
      • Process Plant Reliability Engineering
      • RCM Blitz®
      • ReliabilityXperience
      • Rob’s Reliability Project
      • The Intelligent Transformer Blog
      • The People Side of Maintenance
      • The Reliability Mindset
    • on Product Reliability
      • Accelerated Reliability
      • Achieving the Benefits of Reliability
      • Apex Ridge
      • Breaking Bad for Reliability
      • Field Reliability Data Analysis
      • Metals Engineering and Product Reliability
      • Musings on Reliability and Maintenance Topics
      • Product Validation
      • Reliability by Design
      • Reliability Competence
      • Reliability Engineering Insights
      • Reliability in Emerging Technology
      • Reliability Knowledge
    • on Risk & Safety
      • CERM® Risk Insights
      • Equipment Risk and Reliability in Downhole Applications
      • Operational Risk Process Safety
    • on Systems Thinking
      • The RCA
      • Communicating with FINESSE
    • on Tools & Techniques
      • Big Data & Analytics
      • Experimental Design for NPD
      • Innovative Thinking in Reliability and Durability
      • Inside and Beyond HALT
      • Inside FMEA
      • Institute of Quality & Reliability
      • Integral Concepts
      • Learning from Failures
      • Progress in Field Reliability?
      • R for Engineering
      • Reliability Engineering Using Python
      • Reliability Reflections
      • Statistical Methods for Failure-Time Data
      • Testing 1 2 3
      • The Hardware Product Develoment Lifecycle
      • The Manufacturing Academy
  • eBooks
  • Resources
    • Special Offers
    • Accendo Authors
    • FMEA Resources
    • Glossary
    • Feed Forward Publications
    • Openings
    • Books
    • Webinar Sources
    • Journals
    • Higher Education
    • Podcasts
  • Courses
    • Your Courses
    • 14 Ways to Acquire Reliability Engineering Knowledge
    • Live Courses
      • Introduction to Reliability Engineering & Accelerated Testings Course Landing Page
      • Advanced Accelerated Testing Course Landing Page
    • Integral Concepts Courses
      • Reliability Analysis Methods Course Landing Page
      • Applied Reliability Analysis Course Landing Page
      • Statistics, Hypothesis Testing, & Regression Modeling Course Landing Page
      • Measurement System Assessment Course Landing Page
      • SPC & Process Capability Course Landing Page
      • Design of Experiments Course Landing Page
    • The Manufacturing Academy Courses
      • An Introduction to Reliability Engineering
      • Reliability Engineering Statistics
      • An Introduction to Quality Engineering
      • Quality Engineering Statistics
      • FMEA in Practice
      • Process Capability Analysis course
      • Root Cause Analysis and the 8D Corrective Action Process course
      • Return on Investment online course
    • Industrial Metallurgist Courses
    • FMEA courses Powered by The Luminous Group
      • FMEA Introduction
      • AIAG & VDA FMEA Methodology
    • Barringer Process Reliability Introduction
      • Barringer Process Reliability Introduction Course Landing Page
    • Fault Tree Analysis (FTA)
    • Foundations of RCM online course
    • Reliability Engineering for Heavy Industry
    • How to be an Online Student
    • Quondam Courses
  • Webinars
    • Upcoming Live Events
    • Accendo Reliability Webinar Series
  • Calendar
    • Call for Papers Listing
    • Upcoming Webinars
    • Webinar Calendar
  • Login
    • Member Home
Home » Articles » on Product Reliability » Beyond the Numbers » Integrating Human Factors with Traditional Reliability Techniques

by Chris Weir Leave a Comment

Integrating Human Factors with Traditional Reliability Techniques

Integrating Human Factors with Traditional Reliability Techniques

In Part 1 of Beyond the Numbers, I reflected on why Human Factors matter in reliability engineering and how the human element of the system can be overlooked by traditional hardware-focussed approaches.
This article explores how Human Factors principles can be integrated into traditional reliability analyses such as Reliability Block Diagrams (RBDs), Fault Tree Analysis (FTA) and Failure Mode, Effects and Criticality Analysis (FMECA) and without reinventing the wheel or introducing additional complexity.

The short answer is that we are already doing much of this implicitly. Applying Human Factors principles makes those assumptions explicit and therefore visible, challengeable and open to improvement.

Human Performance Assumptions

Reliability practitioners should already be comfortable in dealing with uncertainty. We routinely estimate failure rates from incomplete data, apply confidence limits and accept that models are approximations of reality rather than perfect representations.

Yet when it comes to human performance, we often default to a silent assumption that the human will do the right thing, at the right time, every time. This assumption is rarely stated, rarely challenged and rarely justified.

In practice however, many of the systems we model rely on people to:

  • Perform system configuration or start-up actions
  • Diagnose faults and select recovery actions
  • Restore redundancy following failures
  • Execute maintenance tasks correctly and completely

If those actions are not performed, or not performed in the way that was envisaged, the modelled reliability may never be realised in operation.

Reliability Block Diagrams: Hardware Modelling to Functional Success

RBDs are widely used to represent how system functions depend on combinations of components. They are often interpreted as hardware models, but in reality they model functional success paths and many of those paths include human action, examples include:

  • Manual switching between redundant paths
  • Operator-initiated reconfiguration
  • Fault acknowledgement and reset actions
  • Human-dependent restoration of failed items

In many models, these actions are assumed to be instantaneous and perfectly reliable.

Making human dependency visible

A practical first step is to represent human-dependent actions explicitly:

  • Introduce blocks representing critical human tasks
  • Assign task success probabilities
  • Explore sensitivity rather than precision

The intent is not to define precise human error probabilities, but to understand how dependent system success is on human performance.

For example, a system may appear highly resilient due to redundancy, but if both paths rely on the same human action under time pressure, the functional reliability may be far more fragile. The RBD below first presented in Part 1, illustrates this example:

Taking this approach leads to better design questions:

  • Can this critical dependency be removed or designed out?
  • Can a manual task be automated or simplified?
  • Are there conditions where the failure of this action becomes unacceptable?

Fault Tree Analysis: Human Errors as Basic Events

FTA represents combinations of events that lead to an undesired outcome. There is no technical reason why human actions cannot be represented as basic events, yet in practice they are often excluded or oversimplified.

The following Fault Tree re-expresses the previous RBD in failure logic terms, treating human actions in the same way as technical failures. For practitioners, the value lies less in the precise numbers and more in the visibility this provides into where human actions meaningfully influence system risk and where changes to design, procedures, or training may have the greatest impact.

Practical integration

Human actions can be modelled as:

  • Failure to respond to an alarm
  • Incorrect diagnosis leading to wrong action
  • Failure to complete a required recovery step

Where data is limited (which it often is), conservative probability ranges can be used, or qualitative sensitivity studies performed.

The value lies not in the exact number, but in understanding:

  • Whether human performance dominates the risk
  • How sensitive the outcome is to task difficulty or context
  • Where design or procedural changes would have the greatest benefit

FTA becomes a powerful tool for challenging optimistic assumptions about human performance, particularly in time-critical or high-stress scenarios.

Failure Mode, Effects and Criticality Analysis: Integrating Human Actions

FMECA is a familiar and trusted technique for understanding how systems can fail and for prioritising mitigation. Traditionally, failure modes are associated with hardware or software items. However, people also fail in identifiable and predictable ways and there is no technical reason why human actions cannot be represented within the same analytical framework.

Applying FMECA thinking to human activity does not require a new method. It requires recognising that human performance is shaped by task design, interfaces, procedures and operating context and that these factors can be analysed just as systematically as technical causes.

Types of human failure

Human failure modes can be grouped in a simple and practical way, based on whether the action was intended or unintended and whether the resulting behaviour reflects a deliberate choice, a planning or decision-making error, or a breakdown in execution or memory, as illustrated in the diagram below:

Crucially, these failure modes are not attributes of individuals. They arise from the way tasks are designed, information is presented, procedures are written and work is carried out under real operating conditions.

Integrating human failure modes into FMECA

When human actions are treated as potential failure modes within a FMECA, the intent is not to list every possible error, but to understand how human performance can influence system outcomes and where mitigation effort is most effective. Rather than conducting a separate analysis, human failure modes can be:

  • Included alongside technical failure modes, rather than treated separately
  • Linked directly to specific tasks, interfaces, or decisions
  • Assessed using the same severity, occurrence, and detectability logic familiar to practitioners

This shifts the focus away from assigning “operator error” as a cause and towards identifying the Performance Influencing Factors (PIFs) that shape human performance, such as workload, interface design, time pressure, or procedural ambiguity.

In turn, this allows mitigation actions to focus on design changes, procedure and interface improvements, training or task restructuring, rather than relying solely on behavioural controls or additional rules.

The table below provides an illustrative example of how different types of human failure can be represented within a FMECA, linking typical examples with relevant PIFs and potential mitigation approaches:

Key takeaway

Human failure modes can be analysed, prioritised and mitigated using the same reliability thinking as hardware failures. The difference lies not in the method, but in recognising that people operate within systems and it is those systems that most strongly shape reliability outcomes.

Avoiding False Precision

A common concern when discussing human reliability is the fear of false accuracy and assigning spurious numbers to complex human behaviour.

This concern is valid – but it is not a reason to ignore the issue entirely.

Reliability models already contain uncertainty:

  • Failure rates are often based on sparse data
  • Environmental and usage assumptions are idealised
  • Human performance is frequently assumed to be perfect by default

Replacing an unstated assumption of perfection with a transparent, bounded assumption of fallibility is an improvement, not a regression. The aim is not to predict human behaviour precisely but to design systems that are robust to human variability.

Incremental Integration, Not Reinvention

One of the most important lessons from my own learning is that integrating human reliability does not require a wholesale change to established practice.

It can start incrementally:

  • Identifying where models depend on human action
  • Making those dependencies visible
  • Challenging whether assumptions are realistic
  • Using sensitivity and scenario thinking rather than precision

Traditional reliability methods remain essential. Integrating human reliability simply extends their relevance to the socio-technical systems we are actually designing, operating and maintaining.

Closing Reflection

Reliability models are not just mathematical interpretations they are expressions of how we believe systems will behave in the real world. When we ignore the human contribution, we risk creating models that are technically sound but operationally fragile.

By integrating human reliability principles into RBDs, FTA and FMECA, we move towards models that better reflect reality, not because they are more complex, but because they are more explicit about system vulnerabilities

Reliability engineering does not stop at understanding how systems fail. Through processes such as Reliability Centred Maintenance (RCM), it directly shapes how maintenance is derived and how tasks are ultimately performed.

In the next part of this series, I will move from reliability-focussed tools to the maintenance domain, exploring how Human Factors principles can be integrated into RCM and Maintenance Task Analysis (MTA) and why realistic assumptions about human performance are just as important when defining maintenance as they are when predicting failure.

I’d be interested to hear from others working in reliability, maintainability and supportability engineering across defence and related sectors:

  • Where do your models already depend on human performance, and how is that dependency made visible or communicated?

… and for Human Factors practitioners:

  • Where do you see opportunities to influence traditional reliability analysis more directly and proportionately?

I’d be genuinely interested to hear others’ perspectives…

 

Filed Under: Articles, Beyond the Numbers, on Product Reliability

About Chris Weir

Chris is expanding his focus into Human Factors and Human Reliability Analysis. In his article series Beyond the Numbers: Human Reliability in Practice, he shares what he’s learning along the way, from practical techniques to real-world examples, showing how human-centred thinking can be woven into traditional reliability engineering tools and processes.

« Top Reliability Engineers Make Risk Management Simple for Every Frontline Worker

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Beyond the Numbers article series logo Photo of Chris WeirArticles by Chris Weir
in the Beyond the Numbers: Human Reliability in Practice article series

Recent Posts

  • Integrating Human Factors with Traditional Reliability Techniques
  • Top Reliability Engineers Make Risk Management Simple for Every Frontline Worker
  • Yet Another Confused MTBF Definition
  • Reliability Test can be Done in Parallel to Design Validation?
  • Safety Culture and Artificial Intelligence

Join Accendo

Receive information and updates about articles and many other resources offered by Accendo Reliability by becoming a member.

It’s free and only takes a minute.

Join Today

© 2026 FMS Reliability · Privacy Policy · Terms of Service · Cookies Policy

Book the Course with John
  Ask a question or send along a comment. Please login to view and use the contact form.
This site uses cookies to give you a better experience, analyze site traffic, and gain insight to products or offers that may interest you. By continuing, you consent to the use of cookies. Learn how we use cookies, how they work, and how to set your browser preferences by reading our Cookies Policy.