<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	xmlns:media="http://search.yahoo.com/mrss/" >

<channel>
	<title>SIL Safe</title>
	<atom:link href="https://silsafe.net/feed/" rel="self" type="application/rss+xml" />
	<link>https://silsafe.net</link>
	<description>Safer Communities, Resilient Operations.</description>
	<lastBuildDate>Tue, 07 Apr 2026 15:42:03 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://silsafe.net/wp-content/uploads/2025/07/cropped-SIL-Safe-Logo-master-white-favicon-square-150x150.png</url>
	<title>SIL Safe</title>
	<link>https://silsafe.net</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Hazard and Risk Assessment (H&#038;RA): The Foundation of Functional Safety</title>
		<link>https://silsafe.net/hazard-and-risk-assessment-hra/</link>
					<comments>https://silsafe.net/hazard-and-risk-assessment-hra/#respond</comments>
		
		<dc:creator><![CDATA[mamerten]]></dc:creator>
		<pubDate>Sat, 04 Apr 2026 21:07:09 +0000</pubDate>
				<category><![CDATA[Beginner]]></category>
		<guid isPermaLink="false">https://silsafe.net/?p=6213</guid>

					<description><![CDATA[A hazard and risk assessment (H&#038;RA) is the foundation of every IEC 61511 safety case. This guide covers what it is, how it works, the methodologies used to conduct one, and what PHA means in the context of OSHA PSM and EPA RMP — written for process safety engineers and functional safety practitioners.]]></description>
										<content:encoded><![CDATA[
<p>If there is one activity in the functional safety life-cycle that sets the tone for everything that follows, it is the Hazard &amp; Risk Assessment (H&amp;RA). Get it right and you have a solid technical foundation for your Safety Instrumented System (SIS). Get it wrong — or skip it — and every decision downstream is built on sand.</p>



<h2 class="wp-block-heading">What Is a Hazard &amp; Risk Assessment (H&amp;RA)?</h2>



<p>A <strong>hazard</strong> is a physical situation with the potential to cause harm. A <strong>risk</strong> is what you get when you combine two things: the <em>probability</em> that the hazard leads to a harmful event, and the <em>severity</em> of the consequences. Risk is never just one of those dimensions. A high-severity outcome with negligible probability may be entirely tolerable. A low-severity outcome that happens constantly may not be. Both dimensions must be assessed together — always.</p>



<p>The H&amp;RA is the structured process of identifying hazards, evaluating the associated risks, and determining whether those risks are tolerable. It is the foundation of the <strong>functional safety life-cycle</strong> — the end-to-end engineering process defined in IEC 61511 that governs how Safety Instrumented Systems are designed, implemented, operated, and maintained. Without the H&amp;RA, there is no technical basis for any of what follows.</p>



<p><strong>A note on terminology</strong> — you will encounter several names and acronyms for this activity depending on the standard or industry context:</p>



<ul class="wp-block-list">
<li><strong>H&amp;RA</strong> (Hazard and Risk Assessment) — the term used in IEC 61511</li>



<li><strong>HRA</strong> and <strong>HARA</strong> — also widely used within the functional safety community; same activity, different shorthand</li>



<li><strong>PHA</strong> (Process Hazard Analysis) — the equivalent term under the OSHA PSM regulation (29 CFR 1910.119) and the EPA RMP regulation; same concept, different regulatory language</li>



<li>Other safety disciplines — machinery safety, for example — may use different terms again</li>
</ul>



<p>The variation is mostly regulatory and organizational preference. The underlying activity is the same.</p>



<h2 class="wp-block-heading">When a Hazard and Risk Assessment Fits in the IEC 61511 Safety Life-Cycle</h2>



<p>A hazard and risk assessment must be conducted early in the safety life-cycle — specifically under Clause 8 — before SIS design begins.</p>



<p>The accepted practice is to conduct the H&amp;RA when the <strong>P&amp;IDs (Piping and Instrumentation Diagrams) are at Rev 0</strong> — the point at which the process design is sufficiently defined to support a meaningful hazard study, but not so advanced that changes identified during the study become costly or impractical to implement. Too early and the hazard picture is incomplete; too late and the window to influence design has closed.</p>



<p>A poorly timed H&amp;RA compounds every problem that follows.</p>



<h2 class="wp-block-heading">Key Steps in Conducting an H&amp;RA</h2>



<p>The H&amp;RA is not a single task — it is a structured sequence of activities. The order matters.</p>



<p><strong>Step 1 – Determine tolerable risk.</strong> Before you can assess whether any risk is acceptable, you need to define what &#8220;acceptable&#8221; means for your organization. This benchmark must be established first. It governs every risk judgment that follows. (See the next section for details.)</p>



<p><strong>Step 2 – Define the scope and boundaries.</strong> What process units, equipment, and operating modes are included in this H&amp;RA? Scope creep and scope gaps are both problems. A clearly documented boundary prevents both.</p>



<p><strong>Step 3 – Identify hazards.</strong> What physical situations exist that have the potential to cause harm? This is where the structured identification methodology — HAZOP, What-If, or similar — is applied.</p>



<p><strong>Step 4 – Identify hazardous events and demand scenarios.</strong> A hazard becomes a hazardous <em>event</em> when a specific initiating cause triggers it — the conditions under which a safety function would be called upon to act.</p>



<p><strong>Step 5 – Assess consequences.</strong> For each hazardous event, what is the worst credible outcome? Consequences are typically assessed in terms of harm to people, environmental impact, and asset damage.</p>



<p><strong>Step 6 – Assess likelihood/frequency.</strong> How often is the hazardous event expected to occur, without any protection layers in place? This is the <em>unmitigated</em> or <em>inherent</em> demand rate.</p>



<p><strong>Step 7 – Identify the risk gap.</strong> With both consequence severity and likelihood established, the assessed risk can be compared against the tolerable risk criteria and the calibrated risk matrix. Where the assessed risk exceeds tolerable risk, a gap exists. That gap is the direct trigger for a SIS.</p>



<h2 class="wp-block-heading">Determining Tolerable Risk — The Foundation of Step 1</h2>



<p>Under IEC 61511, tolerable risk is the level of risk accepted in a given context based on the current values of society. It is <em>not</em> zero risk — every industrial process carries some inherent risk, and the goal of IEC 61511 is to ensure that risk is <em>identified</em>, <em>evaluated</em>, and <em>reduced to a tolerable level</em>, not eliminated entirely. The instinct to demand &#8220;no risk&#8221; is understandable but neither achievable nor the intent of the standard.</p>



<p>Tolerable risk criteria must be documented — a specific IEC 61511 requirement and a common gap at smaller PSM and RMP facilities. Without documented criteria, every risk judgment becomes too subjective and the study loses its technical defensibility.</p>



<p>A <strong>calibrated risk matrix</strong> is the standard tool for establishing and communicating tolerable risk criteria. It is a specific type of risk matrix in which the axis boundaries are anchored to actual numerical frequency and consequence values — not vague qualitative descriptors like &#8220;frequent&#8221; or &#8220;catastrophic.&#8221; Calibration reduces subjectivity as much as possible, improves consistency across the study, and makes the risk criteria defensible under regulatory or third-party scrutiny. Larger organizations and multi-site facilities often maintain multiple calibrated risk matrices — one for each consequence category or tailored to specific process units. The PEAR Model, discussed further in the Related Topics section, provides a useful framework for structuring those consequence categories.</p>



<h2 class="wp-block-heading">The Risk Gap and the Case for a SIS</h2>



<p>The risk gap is the difference between the <em>unmitigated risk</em> of a hazardous event and the <em>tolerable risk</em> threshold. When a gap exists — when the assessed risk exceeds what is tolerable — a protection layer is required.</p>



<p>Protection layers can take many forms:</p>



<ul class="wp-block-list">
<li>Inherently safer design</li>



<li>Basic process control</li>



<li>Physical relief devices</li>



<li>Operator response</li>



<li>Safety Instrumented Functions (SIFs)</li>
</ul>



<p>When a hazard cannot be reduced to a tolerable level through other means, it requires a <strong>Safety Instrumented Function (SIF)</strong> — a specific automated safety action that brings the process to a safe state in response to a defined demand. A process will typically have multiple hazards that each require their own SIF. All of those SIFs together constitute the <strong>Safety Instrumented System (SIS)</strong> — the SIS does not exist independently of the SIFs that define it; it is the sum of them.</p>



<p>This is why the hazard and risk assessment is the primary input to SIS design. Without it, there is no basis for knowing which SIFs are needed or what they must do. The H&amp;RA outputs flow directly into other life-cycle documents, such as the Safety Requirements Specification (SRS).</p>



<h2 class="wp-block-heading">The Link Between H&amp;RA and SIL Determination</h2>



<p>Identifying that a SIF is needed is only the first step. The H&amp;RA provides the information that a SIL allocation is required — the size and nature of the risk gap determines how much risk reduction each SIF must deliver.</p>



<p><strong>Safety Integrity Level (SIL)</strong> is a discrete measure of the required risk reduction performance of a SIF. IEC 61511 defines three SIL levels for the process sector — SIL 1, SIL 2, and SIL 3 — each representing an order-of-magnitude increase in performance. The required SIL is determined by the size of the risk gap: the larger the gap, the higher the SIL required to close it.</p>



<p><strong>Layer of Protection Analysis (LOPA)</strong> is the most widely used SIL determination methodology in the process industry. LOPA is <em>not</em> an H&amp;RA methodology — it is a SIL allocation tool that operates later in the life-cycle, using the hazardous event data and demand rates produced by the H&amp;RA as its primary inputs.</p>



<h2 class="wp-block-heading">H&amp;RA Methodologies</h2>



<p>IEC 61511 does not prescribe a specific H&amp;RA methodology. The right choice depends on process complexity, the stage of design, available data, and the level of rigor required. What the standard does require is that the methodology is appropriate and applied systematically.</p>



<p>All H&amp;RA methodologies fall into one of three umbrella categories:</p>



<p><strong>Qualitative</strong> — judgment-based and descriptive. No numerical failure frequencies or consequence magnitudes are required. The output is a structured list of hazards, causes, consequences, and existing safeguards. Qualitative methods are the most widely used in the process industry for H&amp;RA purposes.</p>



<p><strong>Semi-quantitative</strong> — structured scoring or ranking. Numbers are used to characterize risk, but a full probabilistic analysis is not performed. The output provides more consistency and defensibility than a purely qualitative approach.</p>



<p><strong>Quantitative</strong> — numerical frequency and consequence analysis with full probabilistic treatment. The output is a calculated risk value that can be compared directly against a numerical tolerable risk criterion.</p>



<h3 class="wp-block-heading">Qualitative Methods</h3>



<p><strong>HAZID (Hazard Identification Study)</strong> is an upstream screening tool used to identify major hazards early in a project, before detailed design is available. It is typically applied at the conceptual or FEED stage and is less structured than a HAZOP. The primary reference standard is ISO 17776.</p>



<p><strong>HAZOP (Hazard and Operability Study)</strong> is the dominant H&amp;RA methodology in the process industry. It uses a systematic, guide-word-driven approach applied by a multi-disciplinary team to identify deviations from design intent and evaluate their causes and consequences. HAZOP produces a structured, auditable record that serves as the primary H&amp;RA documentation. The governing standard is IEC 61882.</p>



<p><strong>What-If Analysis</strong> is a less formal brainstorming technique useful for simpler processes, preliminary reviews, or as a supplement to a more structured study.</p>



<p><strong>FMEA (Failure Mode and Effects Analysis)</strong> examines individual components to identify failure modes and their system-level effects. It is a bottom-up analysis most commonly used in equipment-focused assessments.</p>



<h3 class="wp-block-heading">Semi-Quantitative Methods</h3>



<p><strong>Risk Graph</strong> is a structured method that uses defined parameters — consequence severity, occupancy, probability of avoiding harm, and demand rate — to assign a SIL target. It is widely used as an initial SIL targeting tool where a full LOPA is not warranted.</p>



<h3 class="wp-block-heading">Quantitative Methods</h3>



<p><strong>Event Tree Analysis (ETA)</strong> models the sequences of events following an initiating cause, branching at each point where a safety barrier succeeds or fails. It is used to quantify outcome frequencies and is commonly paired with fault tree analysis in broader QRA work.</p>



<p><strong>Quantitative Risk Analysis (QRA)</strong> is a formal methodology combining frequency analysis and consequence modeling to produce a quantified risk picture — typically expressed as individual risk contours or F-N curves. QRA draws on ETA, fault tree analysis, and dispersion analysis as inputs and is the most rigorous and resource-intensive approach.</p>



<p><strong>FMEDA (Failure Mode, Effects and Diagnostic Analysis)</strong> is primarily a <strong>manufacturer and device certification tool</strong> conducted against IEC 61508. It produces quantitative failure rate data — including dangerous undetected failure rate and diagnostic coverage — that process facilities consume from equipment supplier safety manuals rather than generating themselves.</p>



<h2 class="wp-block-heading">Related Topics and Tools</h2>



<p><strong>Bow-Tie Analysis</strong> places the top event at the center, with threat pathways on the left and consequence pathways on the right. Barriers appear on both sides. Bow-tie diagrams are effective for communicating H&amp;RA outputs to management, regulators, and operating teams.</p>



<p><strong>Fault Tree Analysis (FTA)</strong> is a top-down tool that works backward from an undesired top event to identify the combinations of failures that could cause it. FTA can be used qualitatively as a logic map or quantitatively with failure rate data. It is often paired with ETA in QRA studies.</p>



<p><strong>Dispersion Analysis</strong> models the physical spread of a hazardous material release — gas cloud, toxic plume, or flammable vapor — to quantify the geographic extent and frequency of harm. It is a quantitative calculation that supports consequence assessment and is an essential input to QRA at facilities handling hazardous materials.</p>



<p><strong>The PEAR Model</strong> structures consequence assessment around four dimensions: <em>People</em>, <em>Environment</em>, <em>Assets</em>, and <em>Reputation</em>. It is particularly relevant where tolerable risk criteria extend beyond personnel safety — and in practice, many facilities maintain a separate calibrated risk matrix for each PEAR dimension.</p>



<p><strong>Hazardous Area Classification</strong> is a discipline that should be completed prior to the H&amp;RA. The already-established hazardous areas (or zones) can serve as an independent protection layer and may be credited as such during the risk assessment process.</p>



<p><strong>Security Risk Assessment (SRA)</strong> is a parallel requirement under IEC 61511 Clause 8 — the same clause that governs the H&amp;RA. Where the H&amp;RA addresses process hazards, the SRA addresses intentional threats and cybersecurity vulnerabilities to the SIS. Both are required to meet full Clause 8 obligations.</p>



<h2 class="wp-block-heading">Common Mistakes and Pitfalls</h2>



<p><strong>Timing it wrong.</strong> Too early and the hazard picture is incomplete; too late and design changes to address identified hazards may already be impractical. The H&amp;RA should be conducted when the P&amp;IDs are at Rev 0.</p>



<p><strong>No documented tolerable risk criteria.</strong> Without a defined, documented calibrated risk matrix, every risk judgment in the study is a personal opinion — and the study cannot be defended or verified by a third party.</p>



<p><strong>Poorly defined hazardous events.</strong> Vague events produce incorrect consequence and frequency assessments, which produce incorrect risk gap calculations. Each hazardous event needs a defined initiating cause, affected equipment, and demand scenario.</p>



<p><strong>Ignoring human factors and demand rates.</strong> Human error is a significant source of SIF demand. Underestimating it produces an overly optimistic risk picture and may result in incorrect SIL targets.</p>



<p><strong>Treating the H&amp;RA as a one-time exercise.</strong> The H&amp;RA is a living document. When the process changes, it must be reviewed and updated.</p>



<p><strong>Confusing inherent risk with residual risk.</strong> Inherent risk is assessed <em>before</em> protection layers. Residual risk is what remains <em>after</em>. Conflating the two leads to incorrect safeguard crediting and inaccurate risk gap calculations.</p>



<h2 class="wp-block-heading">Keeping Your Hazard and Risk Assessment Current (Management of Change)</h2>



<p>IEC 61511 is explicit: the H&amp;RA must be reviewed whenever changes occur that could affect its validity. Triggers include process modifications, near misses, new or modified equipment, regulatory updates, and periodic revalidation obligations under the safety life-cycle. A formal Management of Change (MOC) process that flags these triggers is the most reliable way to keep the H&amp;RA aligned with the actual process.</p>



<h2 class="wp-block-heading">Frequently Asked Questions</h2>



<p><strong>Q1: I keep hearing different terms — PHA, HARA, HRA, H&amp;RA. Are these all the same thing?</strong></p>



<p>H&amp;RA, HRA, and HARA all refer to the same IEC 61511 activity. Process Hazard Analysis (PHA) is the equivalent term under the OSHA PSM and EPA RMP regulations — different regulatory language, same concept. Other safety disciplines, such as machinery safety, may use different terms again — it can be a bit confusing.</p>



<p><strong>Q2: I&#8217;ve heard that LOPA and HAZOP are often done together in the same study. Is that correct, and if so, how does that work?</strong></p>



<p>Yes — many organizations run HAZOP and LOPA back-to-back in the same workshop series for efficiency, completing the HAZOP for each node before immediately running the LOPA on the identified scenarios. They remain technically distinct activities: HAZOP is the H&amp;RA, LOPA is SIL determination — combining the workshops is a logistical choice, not a technical one.</p>



<p><strong>Q3: Is HAZOP the standard way to conduct an H&amp;RA in the process industry?</strong></p>



<p>In practice, yes — HAZOP is the de facto H&amp;RA methodology for most process facilities operating under IEC 61511, ISA 84, PSM, and RMP. That said, IEC 61511 does not mandate it; other methodologies such as What-If or QRA are appropriate depending on the project stage and complexity.</p>



<p><strong>Q4: Does IEC 61511 require a specific H&amp;RA methodology?</strong></p>



<p>No — IEC 61511 requires that an appropriate methodology is applied systematically, but does not mandate a specific one. HAZOP is the most common choice in the process sector, but What-If and other approaches are all valid depending on the context.</p>



<p><strong>Q5: What is the difference between inherent risk and residual risk?</strong></p>



<p>Inherent risk is the risk before any protection layers are applied. Residual risk is what remains after all protection layers — including the SIS — have been credited.</p>



<p><strong>Q6: Who should be involved in conducting an H&amp;RA?</strong></p>



<p>An H&amp;RA is a multi-disciplinary team activity requiring input from process, operations, instrumentation, and safety disciplines at a minimum — these can be multi-day affairs, painful, particularly when LOPA is run back-to-back with the HAZOP. IEC 61511 requires that at least one team member holds functional safety competence; this role is often referred to in practice as the Team Leader or HAZOP Leader.</p>



<p><strong>Q7: How often does an H&amp;RA need to be revalidated?</strong></p>



<p>IEC 61511 requires H&amp;RA review whenever changes occur that could affect its validity — process modifications, near misses, new equipment, or regulatory changes. Periodic revalidation is also required as part of the broader safety life-cycle review obligations.</p>



<h2 class="wp-block-heading">Further Reading</h2>



<p><strong>Internal — related articles on this site:</strong></p>



<ul class="wp-block-list">
<li><a href="https://silsafe.net/functional-safety-for-the-process-industry/" data-type="post" data-id="6100">What is Functional Safety?</a></li>



<li><a href="https://silsafe.net/proof-test-coverage-cpt-to-improve-pfdavg/" data-type="post" data-id="2784">Proof Testing and Management of Change</a></li>



<li><a href="https://silsafe.net/glossary/stated-risk/" data-type="glossary" data-id="898">Stated Risk</a> vs. <a href="https://silsafe.net/glossary/revealed-risk/" data-type="glossary" data-id="814">Revealed Risk</a></li>



<li>SIL Safe Glossary &#8211; <a href="https://silsafe.net/glossary-cat/hazard-and-risk-assessment/" data-type="page" data-id="3787">H&amp;RA Terms</a></li>



<li>HAZOP — A Deep Dive &#8211; <em>coming soon</em></li>



<li>Layer of Protection Analysis (LOPA) and SIL Determination &#8211; <em>coming soon</em></li>
</ul>



<p><strong>External authoritative references:</strong></p>



<ul class="wp-block-list">
<li><a href="https://www.isa.org/" target="_blank" rel="noopener">ISA — ISA 84 / IEC 61511 resources</a></li>



<li><a href="https://www.iec.ch/" target="_blank" rel="noopener">IEC — IEC 61511 standard</a></li>



<li><a href="https://www.osha.gov/" target="_blank" rel="noopener">OSHA — Process Safety Management regulation (29 CFR 1910.119)</a></li>



<li><a href="https://www.epa.gov/" target="_blank" rel="noopener">EPA — Risk Management Program (RMP)</a></li>
</ul>



<h2 class="wp-block-heading">Conclusion</h2>



<p>A hazard and risk assessment is not a compliance checkbox. It is the technical foundation on which every downstream SIS decision is built — from SIF definition to SIL determination to verification. A well-conducted H&amp;RA, performed at the appropriate time, gives you a defensible, documented basis for your safety case. A poor one — or one conducted too early or too late — leaves your entire SIS design without a credible technical basis.</p>



<p>For facilities operating under IEC 61511, PSM, or RMP, the message is straightforward: invest in getting the H&amp;RA right, time it correctly, and make sure it is conducted by people who understand both the process and the standard.</p>



<p>Functional safety is complex, and the stakes are high. If you have questions about your SIS design, SIL verification, or where to start with IEC 61511, the team at SIL Safe is here to help. Reach out to us today.</p>



<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "I keep hearing different terms — PHA, HARA, HRA, H&RA. Are these all the same thing?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "H&RA, HRA, and HARA all refer to the same IEC 61511 activity. Process Hazard Analysis (PHA) is the equivalent term under the OSHA PSM and EPA RMP regulations — different regulatory language, same concept. Other safety disciplines, such as machinery safety, may use different terms again — it can be a bit confusing."
      }
    },
    {
      "@type": "Question",
      "name": "I've heard that LOPA and HAZOP are often done together in the same study. Is that correct, and if so, how does that work?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Yes — many organizations run HAZOP and LOPA back-to-back in the same workshop series for efficiency, completing the HAZOP for each node before immediately running the LOPA on the identified scenarios. They remain technically distinct activities: HAZOP is the H&RA, LOPA is SIL determination — combining the workshops is a logistical choice, not a technical one."
      }
    },
    {
      "@type": "Question",
      "name": "Is HAZOP the standard way to conduct an H&RA in the process industry?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "In practice, yes — HAZOP is the de facto H&RA methodology for most process facilities operating under IEC 61511, ISA 84, PSM, and RMP. That said, IEC 61511 does not mandate it; other methodologies such as What-If or QRA are appropriate depending on the project stage and complexity."
      }
    },
    {
      "@type": "Question",
      "name": "Does IEC 61511 require a specific H&RA methodology?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "No — IEC 61511 requires that an appropriate methodology is applied systematically, but does not mandate a specific one. HAZOP is the most common choice in the process sector, but What-If and other approaches are all valid depending on the context."
      }
    },
    {
      "@type": "Question",
      "name": "What is the difference between inherent risk and residual risk?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Inherent risk is the risk before any protection layers are applied. Residual risk is what remains after all protection layers — including the SIS — have been credited."
      }
    },
    {
      "@type": "Question",
      "name": "Who should be involved in conducting an H&RA?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "An H&RA is a multi-disciplinary team activity requiring input from process, operations, instrumentation, and safety disciplines at a minimum — these can be multi-day affairs, painful, particularly when LOPA is run back-to-back with the HAZOP. IEC 61511 requires that at least one team member holds functional safety competence; this role is often referred to in practice as the Team Leader or HAZOP Leader."
      }
    },
    {
      "@type": "Question",
      "name": "How often does an H&RA need to be revalidated?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "IEC 61511 requires H&RA review whenever changes occur that could affect its validity — process modifications, near misses, new equipment, or regulatory changes. Periodic revalidation is also required as part of the broader safety life-cycle review obligations."
      }
    }
  ]
}
</script>
]]></content:encoded>
					
					<wfw:commentRss>https://silsafe.net/hazard-and-risk-assessment-hra/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Functional Safety for the Process Industry: 10 Core Concepts Every Engineer Should Know</title>
		<link>https://silsafe.net/functional-safety-for-the-process-industry/</link>
					<comments>https://silsafe.net/functional-safety-for-the-process-industry/#respond</comments>
		
		<dc:creator><![CDATA[mamerten]]></dc:creator>
		<pubDate>Wed, 25 Mar 2026 01:12:24 +0000</pubDate>
				<category><![CDATA[Beginner]]></category>
		<guid isPermaLink="false">https://silsafe.net/?p=6100</guid>

					<description><![CDATA[Functional safety in the process industry is more than calculations—it's a lifecycle approach to reducing risk. This guide explains IEC 61511, SIS, SIL, LOPA, and how safety systems actually work in practice, helping engineers move from basic understanding to real-world application.]]></description>
										<content:encoded><![CDATA[
<h2 class="wp-block-heading">Functional Safety at the Most Basic Level</h2>



<p>Functional safety is the engineering discipline that is about making sure that when something goes wrong, the systems intended to protect people and the environment work when they are needed.</p>



<p>A simple illustration helps.</p>



<p>Over roughly 20 years, I have had about six driver’s side window regulators fail. Annoying—but not dangerous.&nbsp; If the window is used about three times per week, that’s roughly 3,120 operations. Six failures over that period corresponds to a probability of failure of about 1.92E-3 per demand.&nbsp; Is this acceptable?&nbsp; It has been annoying for me, but never once was a safety issue.&nbsp; It also was not a cost risk for the manufacturer since they all failed post-warranty.&nbsp; Now, imagine the regulator is a life safety device, that failure rate would be too high.&nbsp; Functional safety is the program that takes that failure rate and lowers it multiple orders of magnitude, to perhaps 1.92E-5.&nbsp; One cannot do that by building a “better” regulator.&nbsp; You can get there, but it would be via a layer of controls throughout the lifecycle of the regulator.</p>



<p>Driving failure probabilities down by orders of magnitude is the core driver of functional safety. It is not achieved by a single “better” device. It requires coordinated engineering: design choices, architecture, testing, diagnostics, maintenance, and governance across the life of the system.&nbsp;&nbsp;In the process industry, this structured approach is most commonly defined and implemented through IEC 61511, which provides the framework for how functional safety is applied across the lifecycle of a facility.</p>



<p>Functional safety is implemented through active protection functions that span the full loop—sensor, logic solver, and final element. When a hazardous condition is detected, this chain must act correctly to move the process to a safe state, whether that means closing a valve, stopping a motor, or isolating energy.</p>



<p>Just as important, that performance has to hold over time. Functional safety is not about working once—it is about continuing to meet required performance over years of operation, testing, maintenance, and change.</p>



<h2 class="wp-block-heading">Why Functional Safety is Applied in the Process Industry</h2>



<p>Consider a reactor that begins to over‑pressurize.</p>



<p>If nothing intervenes, the vessel could rupture, leading to injury, environmental release, or significant damage. Operators may not respond in time. Standard control systems can help, but systems designed for normal control are typically not sufficient for safety-critical action when consequences are severe.</p>



<p>A system that fails 1% of the time may be acceptable for control. It is often unacceptable for protection.</p>



<p>Functional safety via IEC 61511 is applied to reduce the probability that the protection fails when it is needed, thus making the process move to a safe state under abnormal conditions.</p>



<h2 class="wp-block-heading">What Is Functional Safety?</h2>



<p>Functional safety is an engineering discipline that ensures safety‑significant functions perform correctly when required. It is not a single device, calculation, or tool. It is a structured approach to designing, implementing, and managing systems so that their performance is sufficiently reliable for the hazards they control.</p>



<p>The core concept to understand is risk.&nbsp; A process will have hazards (such as a tank rupture) and that hazard has an associated level of risk.&nbsp; Risk is always the combination of probability and severity.</p>



<p>Functional safety contributes to risk reduction by lowering the probability that a hazardous outcome occurs and, in some designs, by limiting its severity.</p>



<h2 class="wp-block-heading">How Functional Safety Reduces Risk</h2>



<p>Every facility has hazards. But before the hazards are evaluated, the facility needs to determine what risks are acceptable.&nbsp; This is an odd concept for people new to process safety because one has to decide what is an acceptable amount of death or injury.&nbsp; This decision making is documented in a calibrated risk matrix.&nbsp; The next task is to understand those hazards, determine what risks they hold, are those risks tolerable, and reduce risk where they exceed the tolerable level.</p>



<p>Functional safety provides this method of reducing this risk.</p>



<p>The central metric is the probability of failure on demand (PFDavg). If a safety function has a PFDavg of 0.01, it will fail about 1% of the time when demanded and succeed about 99% of the time. This corresponds to a risk reduction factor (RRF) of 100.&nbsp; RRF = 1/PFDavg</p>



<p>In practical terms, this safety function reduces the probability of a hazardous outcome by a factor of 100. This quantified reduction is the mechanism by which functional safety works.</p>



<h2 class="wp-block-heading">Standards Governing Functional Safety</h2>



<p>Functional safety is applied across multiple industries, each with standards built on the same underlying principles. Examples include ISO 26262 for automotive systems, IEC 62061 for machinery, and EN 50126/50128/50129 for rail.</p>



<p>At the foundation is IEC 61508, which defines the general framework for functional safety of electrical, electronic, and programmable electronic systems.</p>



<p>IEC 61511 applies these principles to the process industry. It is not a separate concept; it is a sector-specific implementation of the IEC 61508 framework tailored to process facilities.</p>



<p>Note that in my experience as an engineer, almost all codes and standards are applicable to a certain area.&nbsp; Such as a country or perhaps the European Union.&nbsp; But functional safety is truly one of the few global standards.&nbsp; In almost all counties, if they implement a process safety approach, it will be functional safety.&nbsp; This is one of the reasons SIL Safe exists.</p>



<h2 class="wp-block-heading">Hazard and Risk Assessment (H&amp;RA)</h2>



<p>The functional safety lifecycle begins with understanding hazards in your process.&nbsp; This is called the&nbsp;hazard and risk assessment (H&amp;RA) which identifies scenarios that could lead to harm, estimates their risk (probability and severity), and determines whether risk is tolerable.&nbsp; Under IEC 61511, the H&amp;RA is the starting point of the safety lifecycle, and it directly drives the identification of Safety Instrumented Functions (SIFs) and their required performance.</p>



<p>Typical activities in an H&amp;RA include identifying credible scenarios, estimating probability and severity, and deciding where additional protection is required.</p>



<p>Common methods in the process industry include HAZOP and risk matrices. The outcome of this work will establish whether SIFs are needed.</p>



<p>When this occurs in the design process is important.&nbsp; The H&amp;RA cannot be too early or too late in the process.&nbsp; If too early, much of the hazards would likely change, not exist, or only be a partial list.&nbsp; If done too late, you are asking for challenges as changes may need to be modified in existing equipment.&nbsp; Think cutting pipe to add a SIF.&nbsp; That is never fun.</p>



<p>See this full article for<a href="https://silsafe.net/hazard-and-risk-assessment-hra/" data-type="post" data-id="6213"> deeper dive into an H&amp;RA</a>.</p>



<h2 class="wp-block-heading">Independent Protection Layers (IPLs)</h2>



<p>Facilities rely on multiple layers of protection rather than a single safeguard.&nbsp; Meaning safeguards should be put in place BEFORE a SIF is needed.&nbsp;If this happens, then perhaps a SIF would not be needed or a SIF at a lower SIL level.&nbsp;</p>



<p>These layers can include the basic process control system (BPCS), relief devices, operator response, and safety instrumented systems. For layers to count independently, they must not fail for the same reasons.&nbsp; This collection of independent items that can protect a SIF are called Independent Protection Layers (IPL).</p>



<p>The key question is whether the existing layers reduce risk to tolerable levels. If they do not, additional protection is required.</p>



<p>This evaluation is often formalized through a Layer of Protection Analysis (LOPA), which builds directly on H&amp;RA results.&nbsp; This is one of the most common methods used within IEC 61511 programs to determine the required SIL for each SIF.&nbsp;&nbsp;This asks&#8230;</p>



<ul class="wp-block-list">
<li>Is the risk associated with each hazard tolerable?&nbsp; This would assess the risk for each hazard against the calibrated risk matrix.</li>
</ul>



<ul class="wp-block-list">
<li>If not, are there IPLs present or can they be added?&nbsp; This could be extra pressure relief valves, other instruments in the BPCS.&nbsp;</li>



<li>Does each hazard scenario need a SIF?&nbsp; This compares the risk with the IPLs, against the tolerable risk.&nbsp; If the risk is not tolerable, a SIF is added to mitigate the risk of that hazard.</li>



<li>How much risk must each SIF reduce?&nbsp; Generally, this is thought of in orders of magnitude via a calibrated risk matrix.&nbsp; Each box the hazard moves is considered one order of magnitude.</li>



<li>What is the SIL level of each SIF?&nbsp; This is how many orders of magnitude the risk must be reduced to be tolerable.</li>
</ul>



<h2 class="wp-block-heading">Safety Instrumented Systems (SIS)</h2>



<p>When existing protection is insufficient, a Safety Instrumented System (SIS) is the system that comes into play.&nbsp; The SIS is the core system involved in IEC 61511, although the standard also has lifecycle requirements.&nbsp; Meaning functional safety is not just the SIS.</p>



<p>A SIS is an independent system designed to detect hazardous conditions and move the process to a safe state. It operates separately from the basic process control system (BPCS), which manages normal operation. A typical SIS consists of sensors, a logic solver, and final elements such as shutdown valves or motor isolation devices.&nbsp;&nbsp;</p>



<p>A Safety Instrumented Function (SIF) is a single protection loop within the SIS. A facility may have one SIF or many, depending on the number of scenarios requiring protection.&nbsp; Again, it is the LOPA that dictates where the SIFs are and the SIL needed per SIF.</p>



<p>For example, a simple SIF could be a pressure switch sensing high pressure and turning off a pump with a contactor via the controls in a PLC.&nbsp; Then another SIF can sense a high level and open a dump valve to prevent an overflow.&nbsp; It goes to the same PLC as the first SIF.&nbsp; This would have two SIFs, both using the same PLC and together they would be the SIS.</p>



<h2 class="wp-block-heading">Safety Integrity Levels (SIL)</h2>



<p>Each SIF must meet the required level of reliability, expressed as a Safety Integrity Level (SIL).</p>



<p>SIL is always determined after the H&amp;RA and during the allocation of safety layers, typically determined during the LOPA (which itself is typically done directly after the H&amp;RA, often in the same long meeting). In the process industry, most applications are SIL 1 or SIL 2, with occasional SIL 3. SIL 4 is rarely used.</p>



<p>SIL is often associated with equipment ratings, which is technically correct. But more importantly, it defines the required performance of the safety function across design, implementation, and maintenance. For example, higher SIL levels will need more redundancy and would have different levels of controls and different rules about software.&nbsp; This can be complicated and best discussed elsewhere.</p>



<p>Below is the standard relationship between SIL, Risk Reduction Factor (RRF), and PFDavg for low demand mode:</p>



<figure class="wp-block-table"><table class="has-fixed-layout"><thead><tr><td><strong>SIL</strong></td><td><strong>RRF (Risk Reduction Factor)</strong></td><td><strong>PFDavg Range</strong></td><td><strong>Interpretation</strong></td></tr></thead><tbody><tr><td>&nbsp;1</td><td>10 to 100</td><td>1E-1 to 1E-2</td><td>Reduces risk by 1–2 orders of magnitude</td></tr><tr><td>2</td><td>100 to 1,000</td><td>1E-2 to 1E-3</td><td>Reduces risk by 2–3 orders of magnitude</td></tr><tr><td>3</td><td>1,000 to 10,000</td><td>1E-3 to 1E-4</td><td>Reduces risk by 3–4 orders of magnitude</td></tr><tr><td>4</td><td>10,000 to 100,000</td><td>1E-4 to 1E-5</td><td>Rare in process industry; extreme risk reduction</td></tr></tbody></table></figure>



<h2 class="wp-block-heading">Other Design Considerations in Functional Safety</h2>



<p>Designing a SIF involves more than selecting components.&nbsp; There are many other concepts that need to be considered, understood, and decided during the design phase.</p>



<ul class="wp-block-list">
<li>hardware fault tolerance (HFT) &#8211; Reliability requirements</li>



<li>voting architecture &#8211; Such as 1oo1 voting, 2oo3, and others</li>



<li>proof test interval (TI) &#8211; Duration between proof tests.  See <a href="https://silsafe.net/proof-testing-of-sifs/" data-type="post" data-id="33">this in-depth article</a>.</li>



<li>proof test coverage (Cpt) &#8211; How good proof testing is at detecting failures.  See this <a href="https://silsafe.net/proof-test-coverage-cpt-to-improve-pfdavg/" data-type="post" data-id="2784">deeper dive on Cpt</a>.</li>



<li>mean time to restore (MTTR) &#8211; Facility&#8217;s time to repair a problem, relates to available spare parts.  Read more about <a href="https://silsafe.net/mean-time-to-restore-mttr/" data-type="post" data-id="1129">MTTR</a>.</li>



<li>spurious trip rate (STR) &#8211; The balance between safety and facility uptime</li>



<li>diagnostic coverage (DC) &#8211; How good the on-board diagnostics are at detecting failures</li>



<li>common cause failure (beta factor) &#8211; Redundant systems may jointly fail due to a common design flaw.  Lean more on <a href="https://silsafe.net/how-to-apply-beta-factor-for-common-cause-failure/" data-type="post" data-id="3628">CCF in this article</a>.</li>
</ul>



<p>These parameters all interact. There is a push and pull between the terms and even departments in a facility.&nbsp; Decisions about testing, architecture, spare parts, and maintenance can materially change achieved performance via PFDavg. Each topic is substantial and better explored individually.&nbsp;&nbsp;</p>



<p>These concepts receive significant attention in functional safety engineering and certification exams.&nbsp; These will also take significant effort to decide and work through during the detailed design phase of the SIS.</p>



<h2 class="wp-block-heading">Verifying Safety Instrumented Function Performance (PFDavg)</h2>



<p>Once a SIF is designed, IEC 61511 requires that it be verified against its SIL requirement, per SIF.&nbsp;&nbsp;PFDavg calculations quantify whether the design meets the target. They account for failure rates of the SIF components, architecture, testing intervals, coverage, and repair assumptions.</p>



<p>The process of doing the calculations is complicated.&nbsp; There are simple equations and complex equations.&nbsp; Think of it as a specific PFDavg calculation stems from a series of decisions that are made by engineering.&nbsp; For example, if the SIF is tested only when the unit is shutdown versus bypassed &#8211; that impacts PFDavg in different ways.&nbsp; Generally, practitioners use software or Excel to facilitate.</p>



<p>There is a related approach called Markov Analysis which we will not get into here as it is a more advanced approach.</p>



<p>While central, PFDavg is only one step in a broader discipline. It verifies performance; it does not define the entire program.&nbsp; Inexperienced users may think PFDavg is all functional safety is about.&nbsp; But that is an over-simplification.</p>



<p>The simplest PFDavg equation is shown below.&nbsp; This is for a 1oo1 architecture.&nbsp;</p>



<figure class="wp-block-image size-full"><img decoding="async" width="212" height="58" src="https://silsafe.net/wp-content/uploads/2025/04/pfdavg-basic-equation.png" alt="PFDavg basic equation" class="wp-image-188"/></figure>



<ul class="wp-block-list">
<li>TI &#8211; proof test interval</li>



<li>λ<sub>DU</sub> &#8211; the dangerous undetected failure rate.</li>
</ul>



<h2 class="wp-block-heading">The Functional Safety Life-Cycle</h2>



<p>Functional safety is governed by a structured life-cycle, and this is one of the most important concepts to understand. The structure of this lifecycle is defined in IEC 61511, and following it is what distinguishes a complete functional safety program from isolated design efforts.&nbsp; Many engineers initially approach functional safety as a design activity, but that is only one portion of the overall process.</p>



<p>The life-cycle defines how safety is managed from the earliest concept of a facility through to its eventual decommissioning. It ensures that safety functions are not only designed correctly, but also installed, operated, maintained, and periodically assessed in a consistent and auditable manner.</p>



<p>Typical phases include:</p>



<ul class="wp-block-list">
<li>Hazard and risk assessment (H&amp;RA) such as a HAZOP</li>



<li>Allocation of safety functions (such as via a LOPA)</li>



<li>Design and engineering &#8211; detailed design of the SIFs and SIS</li>



<li>Verification and validation</li>



<li>Operation and maintenance</li>



<li>Functional safety assessment (FSA)</li>



<li>Decommissioning</li>
</ul>



<p>Each of these phases has specific deliverables and expectations. For example, the design phase may define the architecture of a SIF, but the operation and maintenance phase ensures that proof testing is performed and that failures are addressed correctly.</p>



<p>The key point is continuity. Functional safety is not a one-time effort. A system that is properly designed but poorly maintained will not achieve its required performance over time.</p>



<p>An all-too-common scenario is that a compliant SIS is installed in a facility.&nbsp; The company is then bought by a firm in an adjacent industry that has never implemented, nor understands, IEC 61511.&nbsp; Over time, budgets can be cut or key people can navigate to other departments, companies, or retire.&nbsp; In this possible scenario various things could happen that could impact the SIS design.&nbsp; For example, proof tests being done at the wrong intervals or not in accordance with the requirements.&nbsp; Components could be replaced with non-SIL qualified versions, or the competency and training program could become weak.&nbsp; All of this is why the functional safety lifecycle is so important.</p>


<div class="wp-block-image">
<figure class="aligncenter size-full is-resized"><a href="https://silsafe.net/wp-content/uploads/2025/05/SIS-safety-lifecycle.webp"><img fetchpriority="high" decoding="async" width="757" height="981" src="https://silsafe.net/wp-content/uploads/2025/05/SIS-safety-lifecycle.webp" alt="SIS Safety Lifecycle. The very important diagram from IEC 61511 figure 7 which overlays the entire Functional Safety Process" class="wp-image-760" style="width:354px;height:auto" srcset="https://silsafe.net/wp-content/uploads/2025/05/SIS-safety-lifecycle.webp 757w, https://silsafe.net/wp-content/uploads/2025/05/SIS-safety-lifecycle-231x300.webp 231w" sizes="(max-width: 757px) 100vw, 757px" /></a></figure>
</div>


<h2 class="wp-block-heading">Regulatory Context</h2>



<p>In the United States, regulatory frameworks such as OSHA’s Process Safety Management (PSM) and the EPA’s Risk Management Program (RMP) are the core two regulations.&nbsp; These get triggered if a certain weight of various materials is on site (called threshold quantity).&nbsp; These regulations do not prescribe exact methods for managing risk. Instead, they require that facilities follow sound engineering practices.&nbsp; IEC 61511 is not listed specifically as applicable in the Code of Federal Regulations (CFR), but both OSHA and EPA have stated in writing that using that standard is a sufficient and preferred way to meet the regulations.&nbsp; In other words, it is considered a RAGAGEP (recognized and generally acceptable good engineering practice).</p>



<p>Therefore, IEC 61511 is often used to demonstrate that a facility is meeting those regulations. It provides a structured and well-understood approach that regulators, auditors, and engineers recognize.&nbsp; Other countries have similar laws and regulations requiring the standard.</p>



<p>In practice, this means that even though IEC 61511 is not mandatory by law, it is frequently treated as if it were, because it defines what &#8220;good&#8221; looks like for functional safety in the process industry.</p>



<p>In addition to RMP and PSM, other things could trigger using IEC 61511.&nbsp; This could be contracts between parties or insurance requirements.&nbsp; At times, projects for various reasons will not invoke IEC 61511 in full, but may require certain instruments or automated valve maintain a SIL certification.&nbsp; These SIL only requirements have become more typical as SIL rated components become more common.&nbsp; SIL Safe welcomes this change in the industry.</p>



<h2 class="wp-block-heading">Common Misconceptions About Functional Safety</h2>



<p>Functional safety via IEC 61511 is often misunderstood, particularly by those who are new to the discipline or who have only been exposed to portions of it.</p>



<p>One common misconception is that functional safety is primarily about calculations, especially PFDavg. While calculations are important, they are only one part of the overall process. Without proper hazard assessment, design, testing, and maintenance, calculations alone do not ensure safety.</p>



<p>Another misconception is that functional safety is only relevant during design. In reality, long-term performance depends heavily on proof testing, maintenance practices, and how changes are managed over time.</p>



<p>A third misconception is that higher SIL automatically means a better system. In practice, SIL is determined by the risk of the hazard along with the IPLs that exist. A higher SIL requirement often indicates a more severe hazard rather than a superior design choice.</p>



<p>Understanding these misconceptions is important because they often lead to incomplete or ineffective implementations of functional safety programs.</p>



<h2 class="wp-block-heading">Why Functional Safety Programs Matter — and When Expertise Is Needed</h2>



<p>Functional safety programs provide a structured way to manage risk across a facility. At a high level, they help ensure that hazards are identified, risks are evaluated, and appropriate protections are implemented and maintained over time.</p>



<p>Effective programs reduce the probability and severity of major accidents, support regulatory compliance, and provide confidence that safety systems will perform when required.</p>



<p>In practice, many organizations require additional expertise at certain points, such as:</p>



<ul class="wp-block-list">
<li>Implementing IEC 61511 for the first time</li>



<li>Performing SIL determination or verification</li>



<li>Preparing for and conducting functional safety assessments</li>



<li>Modifying or upgrading existing SIS implementations</li>
</ul>



<p>These situations often involve complex decisions, tradeoffs, and documentation requirements that benefit from experienced practitioners.</p>



<h2 class="wp-block-heading">Q&amp;A Section</h2>



<ol start="1" class="wp-block-list">
<li>What is functional safety in simple terms?</li>
</ol>



<p>Functional safety is the part of overall facility safety that depends on safety functions operating correctly when required. It focuses on ensuring that protection systems perform reliably enough to reduce risk to tolerable levels.&nbsp; It has the ability to take a typical failure rate of a safety system of perhaps 0.01 (1%) down multiple orders of magnitude.&nbsp; It does this through a layer of requirements throughout the lifecycle of the system.</p>



<p>Functional Safety is applied to various industries.&nbsp; SIL Safe focuses on its application to the process industry via IEC 61511.</p>



<ol start="2" class="wp-block-list">
<li>What is the difference between a SIF and a SIS?</li>
</ol>



<p>A Safety Instrumented System (SIS) is the overall system that performs safety functions. A Safety Instrumented Function (SIF) is a single protection loop within that system, typically consisting of a sensor, logic solver (think a PLC), and final element (like a valve or contactor).</p>



<ol start="3" class="wp-block-list">
<li>If I have hazardous scenarios, why can&#8217;t I just add an extra instrument?</li>
</ol>



<p>Adding an extra instrument and connecting that to your BPCS does not necessarily reduce risk enough. Risk reduction must be quantified, and the resulting protection must meet the required performance level. Without it, the hazard may still exceed tolerable risk.</p>



<p>For example, at times a SIF will have to have an architecture of 2oo3 (meaning three instruments at one point).&nbsp; One would not know that it was needed unless the process was followed and a PFDavg was calculated</p>



<ol start="4" class="wp-block-list">
<li>I&#8217;ve worked on projects where SIL 2 instruments were specified, but the facility was not doing functional safety in its entirety. What is happening there?</li>
</ol>



<p>Some projects contractually require certain instruments or &#8220;safety instruments&#8221; to be SIL rated (for example SIL 2 transmitters or valves). This does not mean that the facility is implementing the full functional safety lifecycle. In many cases, companies attempt a compromise where equipment meets SIL capability requirements even if the broader IEC 61511 functional safety program is not fully implemented.</p>



<ol start="5" class="wp-block-list">
<li>What standards govern functional safety?</li>
</ol>



<p>The foundational standard is IEC 61508, which defines the general framework for functional safety of electrical, electronic, and programmable electronic systems. For the process industry specifically, IEC 61511 defines how those principles are applied to Safety Instrumented Systems.</p>



<ol start="6" class="wp-block-list">
<li>What about machinery safety?</li>
</ol>



<p>Machinery safety is important, of course.&nbsp; But it is distinct.&nbsp; Functional safety for the process industry is focused on reducing the risk (probability and severity) of a major accident.&nbsp; Machinery safety focuses on the user of the machine.</p>



<p>However, &#8230;. as SIL ratings are more common, what is happening is machinery safety risk assessment now will often require a SIL rated instrument.&nbsp;&nbsp;SIL Safe fully supports this excellent use of SIL instruments.&nbsp; However, this should not be construed as functional safety.</p>



<h2 class="wp-block-heading">Conclusion</h2>



<p>Functional safety via IEC 61511 combines hazard analysis, engineered protection systems, and structured life-cycle management to reduce the probability and severity of hazardous events.</p>



<p>It is not a single calculation or device, but a coordinated engineering approach that spans the entire life of a facility.</p>



<p>For engineers working in the process industry, understanding these concepts is essential to designing and operating safe systems.</p>



<h2 class="wp-block-heading">Call to Action</h2>



<p>If your facility is implementing or improving a functional safety program, expert guidance can make the process significantly more effective.</p>



<p>Contact SIL Safe to discuss consulting services for IEC 61511 programs, SIS design, and functional safety assessments.</p>



<h2 class="wp-block-heading">Additional Resources:</h2>



<ul class="wp-block-list">
<li><a href="https://webstore.iec.ch/en/publication/24241" target="_blank" rel="noopener">IEC 61511 Official Standard</a></li>



<li><a href="https://www.isa.org" target="_blank" rel="noopener">International Society of Automation (ISA)</a></li>



<li><a href="https://www.hse.gov.uk" target="_blank" rel="noopener">UK Health and Safety Executive (HSE)</a></li>



<li><a href="https://www.aiche.org/ccps" target="_blank" rel="noopener">CCPS Guidelines</a></li>
</ul>



<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "What is functional safety in simple terms?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Functional safety is the part of overall facility safety that depends on safety functions operating correctly when required. It focuses on ensuring that protection systems perform reliably enough to reduce risk to tolerable levels. It has the ability to take a typical failure rate of a safety system of perhaps 0.01 (1%) down multiple orders of magnitude. It does this through a layer of requirements throughout the lifecycle of the system. Functional safety is applied to various industries. SIL Safe focuses on its application to the process industry via IEC 61511."
      }
    },
    {
      "@type": "Question",
      "name": "What is the difference between a SIF and a SIS?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "A Safety Instrumented System (SIS) is the overall system that performs safety functions. A Safety Instrumented Function (SIF) is a single protection loop within that system, typically consisting of a sensor, logic solver (think a PLC), and final element (like a valve or contactor)."
      }
    },
    {
      "@type": "Question",
      "name": "If I have hazardous scenarios, why can't I just add an extra instrument?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Adding an extra instrument and connecting that to your BPCS does not necessarily reduce risk enough. Risk reduction must be quantified, and the resulting protection must meet the required performance level. Without it, the hazard may still exceed tolerable risk. For example, at times a SIF will have to have an architecture of 2oo3, meaning three instruments at one point. One would not know that it was needed unless the process was followed and a PFDavg was calculated."
      }
    },
    {
      "@type": "Question",
      "name": "I've worked on projects where SIL 2 instruments were specified, but the facility was not doing functional safety in its entirety. What is happening there?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Some projects contractually require certain instruments or safety instruments to be SIL rated, for example SIL 2 transmitters or valves. This does not mean that the facility is implementing the full functional safety lifecycle. In many cases, companies attempt a compromise where equipment meets SIL capability requirements even if the broader IEC 61511 functional safety program is not fully implemented."
      }
    },
    {
      "@type": "Question",
      "name": "What standards govern functional safety?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "The foundational standard is IEC 61508, which defines the general framework for functional safety of electrical, electronic, and programmable electronic systems. For the process industry specifically, IEC 61511 defines how those principles are applied to Safety Instrumented Systems."
      }
    },
    {
      "@type": "Question",
      "name": "What about machinery safety?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Machinery safety is important, of course, but it is distinct. Functional safety for the process industry is focused on reducing the risk, meaning probability and severity, of a major accident. Machinery safety focuses on the user of the machine. However, as SIL ratings are more common, machinery safety risk assessment will often require a SIL rated instrument. SIL Safe fully supports this excellent use of SIL instruments. However, this should not be construed as functional safety."
      }
    }
  ]
}
</script>



<p></p>
]]></content:encoded>
					
					<wfw:commentRss>https://silsafe.net/functional-safety-for-the-process-industry/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Failure Rates in Functional Safety: A Practical Guide for Working Engineers</title>
		<link>https://silsafe.net/failure-rates-in-functional-safety/</link>
					<comments>https://silsafe.net/failure-rates-in-functional-safety/#respond</comments>
		
		<dc:creator><![CDATA[mamerten]]></dc:creator>
		<pubDate>Sun, 07 Dec 2025 23:07:26 +0000</pubDate>
				<category><![CDATA[Beginner]]></category>
		<guid isPermaLink="false">https://silsafe.net/?p=4221</guid>

					<description><![CDATA[Failure rates are central to SIL verification, device selection, diagnostics, and proof testing. This guide explains λDU, λDD, λSU, and λSD, where the values come from, and how functional safety engineers use them correctly under IEC 61511.]]></description>
										<content:encoded><![CDATA[
<h2 class="wp-block-heading" id="introduction-why-failure-rates-matter-in-functional-safety">Introduction: Why Failure Rates Matter in Functional Safety</h2>



<p>Failure rates sit at the center of how we evaluate, verify, and maintain safety instrumented systems (SIS) under IEC 61511. They show up in SIL verification, PFDavg and STR calculations, equipment selection, and proof test strategy. If you understand the failure‑rate categories and how to obtain them correctly, you can avoid many of the mistakes that derail SIL verification or misrepresent SIS performance.</p>



<p>This article explains the four failure‑rate categories, where the values come from, how to interpret them, and how a functional safety engineer uses them in practice.</p>



<h2 class="wp-block-heading" id="the-big-picture-what-we-mean-by-a-failure-rate">The Big Picture: What We Mean by a &#8220;Failure Rate&#8221;</h2>



<p>In functional safety, <em>failure rate</em> is represented by <strong>λ (lambda)</strong>, typically shown in units of 1/hour or in FITs (failures per 1E9 hours). It represents the frequency of <em>random hardware failures</em>—the only failures that can be mathematically modeled.</p>



<h3 class="wp-block-heading" id="random-vs-systematic-failures">Random vs. Systematic Failures</h3>



<p>Random hardware failures are the only failures that can be described by a rate. Systematic failures absolutely matter, but because they arise from design or process weaknesses, they cannot be represented by λ. You must manage them through quality processes and functional safety management—not statistics.</p>



<h3 class="wp-block-heading" id="constant-failure-rate-assumption">Constant Failure Rate Assumption</h3>



<p>IEC 61511 modeling assumes a <em>constant</em> failure rate. Real‑world devices follow a classic bathtub curve: higher failures early in life (infant mortality), a long flat useful‑life period, and then increasing failures late in life. Failure‑rate data used for SIL verification assumes you are in that useful‑life region.</p>



<h3 class="wp-block-heading" id="statistical-nature-of-λ-values">Statistical Nature of λ Values</h3>



<p>Certification bodies and data handbooks treat λ values as statistical estimates with confidence bounds. The SIL certificate condenses this into a single number, but it should be remembered that every published λ carries uncertainty.&nbsp; However, most of that is a bit behind the scenes for functional safety engineers.</p>



<h2 class="wp-block-heading" id="the-four-failure-rate-categories-in-functional-safety">The Four Failure Rate Categories in Functional Safety</h2>



<p>Failure modes in functional safety fall into four buckets based on whether the failure is safe or dangerous, and whether it is detected or undetected by diagnostics:</p>



<ul class="wp-block-list">
<li><strong>λSD – Safe Detected</strong></li>



<li><strong>λSU – Safe Undetected</strong></li>



<li><strong>λDD – Dangerous Detected</strong></li>



<li><strong>λDU – Dangerous Undetected</strong></li>
</ul>



<p>These categories determine whether the failure affects safety, reliability, or uptime—and how it appears in SIL and STR calculations.&nbsp; Note that &#8220;detected&#8221; means detected by diagnostics, not by proof tests.</p>



<h2 class="wp-block-heading" id="how-the-four-failure-rates-feed-functional-safety-calculations">How the Four Failure Rates Feed Functional Safety Calculations</h2>



<p><strong>λDU</strong> is the largest driver of safety risk. It represents failures that prevent the SIF from acting and are <em>not</em> discovered by diagnostics. This value is always used in <strong>PFDavg</strong>.&nbsp;&nbsp;<strong>λDD</strong> may be used in PFDavg if the detected failure notifies only (does not trip the SIF).</p>



<p><strong>λSU</strong> always contributes to <strong>spurious trip rate (STR).</strong>&nbsp;&nbsp;<strong>λDD and λSD</strong> contribute to <strong>STR</strong>&nbsp;if the control logic forces a safe‑state action when diagnostics detect a failure.&nbsp; See <a href="https://silsafe.net/spurious-trip-rate-explained/" data-type="post" data-id="3315">this article on STR </a>for more background.  </p>



<p><strong>λSU and λSD</strong> influence reliability and uptime but do not affect PFDavg.</p>



<p>Finally, proof tests exist specifically to reveal <em>undetected dangerous failures</em>—the λDU portion. Some engineers misunderstand this and assume proof tests simply “detect failures,” but in functional safety terms, proof tests are how you manage the DU accumulation.&nbsp; See this article on <a href="https://silsafe.net/proof-testing-of-sifs/" data-type="post" data-id="33">proof testing</a>.</p>



<h2 class="wp-block-heading" id="where-failure-rates-come-from">Where Failure Rates Come From</h2>



<h3 class="wp-block-heading" id="where-the-functional-safety-engineer-actually-gets-failure-rates">Where the Functional Safety Engineer Actually Gets Failure Rates</h3>



<p>In real SIS design work, most failure‑rate data comes from <strong>certified products</strong>, where a certification body (CB) has already performed a detailed IEC 61508 assessment. The FS engineer reads the <strong>SIL certificate</strong> and the <strong>Safety Manual</strong>, which provide the extracted λDU, λDD, λSU, and λSD values. These two documents are the authoritative sources for day‑to‑day engineering. The underlying FMEDA exists, but it is not normally reviewed or needed by the practitioner.</p>



<p>When a certified device is not available, several alternate data routes exist. Each route has tradeoffs and requires engineering judgment:</p>



<ul class="wp-block-list">
<li><strong>Manufacturer‑supplied reliability data</strong> – useful when transparent and well‑supported, but assumptions must be confirmed.</li>



<li><strong>Validated site or company datasets</strong> – often the most realistic if maintenance and failure tracking are strong.</li>



<li><strong>User‑generated field data</strong> – applicable for legacy equipment with a long operating history.</li>



<li><strong>Industry sources such as OREDA</strong> – helpful when carefully matched to device type, service, and environment.</li>
</ul>



<p>These alternatives are less typical in functional safety practice, but they require more scrutiny than certified data.</p>



<h3 class="wp-block-heading" id="how-failure-rates-are-determined-typical-scenario">How Failure Rates Are Determined (Typical Scenario)</h3>



<p>For certified devices, IEC 61508 defines the process for establishing failure rates. Behind the scenes, the CB reviews or performs:</p>



<ul class="wp-block-list">
<li>FMEDA (failure‑mode analysis and diagnostic modeling)</li>



<li>Test campaigns and empirical validation</li>



<li>Diagnostic behavior evaluation</li>



<li>Environmental and installation assumption checks</li>
</ul>



<p>The FS engineer does <strong>not</strong> redo this work. Instead, their responsibility is to:</p>



<ul class="wp-block-list">
<li>Use the published λ values correctly</li>



<li>Ensure the application matches the assumptions in the Safety Manual</li>



<li>Integrate diagnostics the way the certification expects</li>
</ul>



<p>This is where many real‑world errors occur—not because the values are wrong, but because the application does not match the assumptions behind them.</p>



<h2 class="wp-block-heading" id="broader-reliability-concepts">Broader Reliability Concepts</h2>



<p>Failure rates are not standalone constants; they are shaped by reliability principles that sit behind the λ numbers. A functional safety engineer must understand these broader ideas to avoid misusing published data.</p>



<p><strong>Systematic failures are not described by λ values.</strong> Random hardware failures can be modeled with rates; systematic failures cannot. They arise from design issues, configuration errors, software defects, or procedure gaps. They must be controlled through functional safety management (such as what IEC 61508 does), not reliability math.</p>



<p><strong>Failure‑rate uncertainty is always present.</strong> λ values are statistical estimates derived from limited testing, modeling, or field data. Certification bodies select a representative value for the SIL certificate, but there is natural variability behind every λ. The published number is not a perfect truth—it is a useful engineering approximation.</p>



<p><strong>Application and environment can change the true failure rate.</strong> A device used in corrosive service, high vibration, or aggressive cycling may experience a higher effective λ than the certified value. Likewise, poor installation, improper mounting, or low‑quality air supply (for valves) can shift failure behavior. The published λ applies only when the Safety Manual conditions are met.</p>



<p><strong>The Safety Manual controls the validity of the data.</strong> A λ value is only valid if the equipment is installed, wired, maintained, and operated according to the Safety Manual. If diagnostics are not used, if limits are exceeded, or if maintenance intervals differ from expectations, the certified failure rates no longer describe the real system.</p>



<h2 class="wp-block-heading" id="automated-valve-assemblies-and-how-their-failure-rates-combine">Automated Valve Assemblies and How Their Failure Rates Combine</h2>



<p>Automated valves used as final elements are not single devices—they are assemblies made of several components, each with their own failure behavior. A functional safety engineer must gather λ values for <strong>each</strong> sub‑component and understand how they combine to represent the full final element.</p>



<p>Typical valve‑assembly components include the valve body, actuator, solenoid, positioner, and any boosters or air relays. Each component contributes its own λDU, λDD, λSU, and λSD. Because the SIF fails if <strong>any one</strong> of these components cannot perform its intended action, the failure rates are combined using <strong>Boolean OR logic</strong>:</p>



<p><strong>λ_total ≈ λ₁ + λ₂ + λ₃ + …</strong></p>



<p>In practice, most λDU comes from mechanical components such as the actuator and valve body. These parts typically lack meaningful diagnostics, so λDU remains the dominant contributor for the final element. Electronic components—like smart positioners—may reduce λDD by improving diagnostics, but they seldom reduce λDU in a significant way.</p>



<p>Some manufacturers have begun certifying complete valve assemblies under IEC 61508. When available, this simplifies the engineer’s task: the assembly‑level λ values are already validated and consolidated under a single device boundary. See this example from Emerson: <a href="https://www.emerson.com/en-us/automation/valves/controlvalves/digital-isolation-solutions" target="_blank" rel="noopener">https://www.emerson.com/en-us/automation/valves/controlvalves/digital-isolation-solutions</a>. We at SIL Safe expect and hope this trend continues.</p>



<h2 class="wp-block-heading" id="practical-examples">Practical Examples</h2>



<h3 class="wp-block-heading" id="sensor-example-how-the-engineer-obtains-λ-values">Sensor Example: How the Engineer Obtains λ Values</h3>



<p>A functional safety engineer begins by locating the λDU, λDD, λSU, and λSD values published in the device’s SIL certificate or Safety Manual. These documents reflect the certification body’s IEC 61508 assessment and define how the device behaves under expected diagnostic, installation, and environmental conditions. The engineer then confirms that the plant’s SIS logic and wiring actually use the diagnostic features assumed in the certification. Once these steps are complete, the engineer has the correct λ values that will be applied in later PFDavg or STR calculations.</p>



<h3 class="wp-block-heading" id="final-element-example-how-the-engineer-obtains-λ-values">Final Element Example: How the Engineer Obtains λ Values</h3>



<p>For an automated valve assembly, the process is more involved because a final element is made of multiple components that must all function correctly. The engineer identifies each sub-component—such as the valve body, actuator, solenoid, positioner, and boosters—and retrieves λ values from each component’s SIL certificate or Safety Manual. The installation and diagnostic assumptions must match the application for the values to be valid. Because a failure in any single sub-component prevents the valve from performing its safety function, the engineer combines the λ values using OR logic to produce the total assembly failure rate. This assembled λ dataset will be used in downstream PFDavg and STR calculations.</p>



<h2 class="wp-block-heading" id="common-mistakes-engineers-make-with-failure-rates">Common Mistakes Engineers Make with Failure Rates</h2>



<p>Even experienced engineers can misapply failure‑rate data if the context behind the numbers is not fully understood. A few issues show up repeatedly in real SIS design and verification work.</p>



<p><strong>Misinterpreting λDU vs. λDD.</strong> These two values behave very differently. λDU always goes into PFDavg because it represents failures that diagnostics cannot find. λDD may or may not impact STR or PFDavg depending on how diagnostics are integrated. Treating DD like DU—or assuming DD never matters—produces incorrect verification results.</p>



<p><strong>Using generic values without validating assumptions.</strong> Generic data tables, old spreadsheets, or handbook values can be misleading if the assumptions behind them do not match your application. Certified values come with defined conditions; generic values usually do not.</p>



<p><strong>Ignoring diagnostics.</strong> Sometimes diagnostics exist on the device but do not make it into the SIS logic or maintenance workflow. If a diagnostic bit is unwired, unmapped, filtered out, or simply ignored in operations, detected dangerous failures behave like undetected failures. In this case, λDD effectively becomes λDU.</p>



<p><strong>Treating λ values as universal constants.</strong> A λ value from a certificate is not automatically valid everywhere. Installation, environment, cycling, mounting, and maintenance determine whether the published λ truly reflects your plant’s conditions. Failure rates must be applied with engineering judgment, not copied blindly.</p>



<h2 class="wp-block-heading" id="when-diagnostics-exist-but-are-not-used">When Diagnostics Exist but Are Not Used</h2>



<p>Diagnostics only add value when the SIS actually acts on them. A device may have excellent internal diagnostics, but if they are not used by the controls, the failure behaves as if it were <em>undetected</em>. In this situation, λDD is effectively added to λDU for purposes of SIL verification because the SIF remains impaired until someone actively responds.&nbsp; It could really impact PFDavg.</p>



<p>This scenario is more common in brownfield facilities, older installations, poorly integrated SIS/BPCS architectures, or sites where diagnostics alarm but no work process exists to ensure timely repair. The lesson for the engineer is simple: <strong>diagnostics only help if the entire chain—from device to logic to maintenance—uses them correctly.</strong></p>



<h2 class="wp-block-heading" id="where-failure-rates-influence-sis-design-decisions">Where Failure Rates Influence SIS Design Decisions</h2>



<p>Failure‑rate data influences several real‑world engineering choices throughout the SIS life‑cycle. Understanding how λ values behave helps the engineer select architectures, manage proof‑test strategies, and apply diagnostics intentionally—not blindly.</p>



<p><strong>Architecture selection.</strong> If λDU is high, additional redundancy may be required to achieve the target SIL. Failure‑rate data helps determine whether 1oo1, 1oo2, or 2oo3 architectures are appropriate for the SIF.</p>



<p><strong>Choosing the proof‑test interval (TI).</strong> Proof tests exist to reveal λDU—the part diagnostics cannot see. A higher λDU or lower proof‑test coverage (Cpt) typically requires a shorter TI. Failure‑rate data directly shapes the proof‑test strategy.&nbsp; See <a href="https://silsafe.net/proof-test-coverage-cpt-to-improve-pfdavg/" data-type="post" data-id="2784">this other article about CPT.</a></p>



<p><strong>Partial‑stroke testing for final elements.</strong> For valves that dominate λDU, partial‑stroke testing may reduce the exposure time of dangerous failures. This decision depends on understanding which failure modes are found by diagnostics versus proof tests.</p>



<p><strong>Diagnostic selection and integration.</strong> λDD and λSD only help if diagnostics are wired, mapped, and acted on. Understanding the diagnostic coverage of a device and the assumptions in the Safety Manual helps engineers design logic and maintenance workflows that truly reduce risk.</p>



<h2 class="wp-block-heading" id="summary-a-practical-way-to-think-about-failure-rates">Summary: A Practical Way to Think About Failure Rates</h2>



<p>Failure rates are the foundation of how we model and manage random hardware failures in functional safety. λDU represents the portion of failures that silently erode the ability of a SIF to act when needed. λDD and λSD describe failures that diagnostics can reveal, informing how often a SIF may trip unnecessarily and how reliably it stays available. λSU supports reliability but does not influence risk directly.</p>



<p>These four failure‑rate categories show up throughout the IEC 61511 safety life‑cycle: in equipment selection, architectural decisions, proof‑test strategy, diagnostic design, and SIL verification. If the failure‑rate assumptions in the SIL certificate and Safety Manual are respected—and if diagnostics are used correctly—then λ values become powerful tools for designing and maintaining a dependable SIS.</p>



<h2 class="wp-block-heading" id="more-help">More Help</h2>



<p>For more help applying failure‑rate data correctly—or for third‑party SIS verification—reach out through SIL Safe’s contact page.</p>



<ul class="wp-block-list">
<li>Explore the SIL Safe <a href="https://silsafe.net/functional-safety-glossary/" data-type="page" data-id="391">glossary </a>for clear explanations of related terms.</li>



<li>See <a href="https://www.isa.org/intech-home/2020/may-june/departments/updated-guidelines-for-using-isa-iec-61511-funct" target="_blank" rel="noopener">ISA&#8217;s main guidelines</a></li>



<li>Fully <a href="https://www.emerson.com/en-us/automation/valves/controlvalves/digital-isolation-solutions" target="_blank" rel="noopener">certified automated valves</a></li>



<li>Deeper blog article on <a href="https://silsafe.net/proof-testing-of-sifs/" data-type="post" data-id="33">proof testing</a></li>
</ul>



<h2 class="wp-block-heading" id="q-a-section">Q&amp;A Section</h2>



<ul class="wp-block-list">
<li><strong>What’s the practical difference between λDU and λDD?</strong><br>λDU drives PFDavg because the failure is both dangerous and <em>undetected</em>. λDD is dangerous but <em>detected</em>, so it typically results in an alarm or forced-safe trip and may contribute to STR depending on configuration.</li>



<li><strong>Do failure rates change over time in my facility?</strong><br>Yes. The published λ values assume controlled conditions; however, real‑world factors like environment, cycling, installation quality, and maintenance can shift the true failure rate up or down.&nbsp;However, we assume constant for modeling purposes.</li>



<li><strong>Why do certified products help for accurate λ values?</strong><br>Certified devices provide validated λ values and clearly defined DU/DD/SU/SD splits under IEC 61508. This reduces interpretation errors and ensures the assumptions behind the numbers are understood and controlled.</li>



<li><strong>Is it okay to use generic failure rates?</strong><br>Only if you confirm that they match your device and application. Generic values may not reflect your environment, proof-test strategy, or diagnostic coverage.</li>



<li><strong>What if the device doesn’t have a SIL certificate?</strong><br>You can still use it, but you must rely on credible manufacturer reliability data, validated site history, or reputable sources such as OREDA. These represent the alternate data routes when a certified data path is not available. Assumptions must match the application, and justification must be documented.</li>



<li><strong>Why do final elements almost always have higher λDU than sensors?</strong><br>Most λDU in a final element comes from mechanical components like the actuator and valve body, which lack strong diagnostics. Sensors generally have better diagnostic coverage and fewer mechanical wear points, so their λDU values are typically much lower.</li>



<li><strong>I understand why λDU impacts PFDavg, but why would λDD impact PFDavg?</strong><br>It depends on how the controls and diagnostics are set up. In many low-demand configurations, only λDU enters PFDavg. But if a dangerous detected failure leaves the SIF unable to perform its function—and the system does not act on or repair that diagnostic—then that portion of λDD effectively behaves like λDU and may need to be included in the PFDavg analysis.</li>



<li><strong>How are λ values used differently in high-demand or continuous modes?</strong><br>Failure rate is highly relevant but used differently.&nbsp; In low-demand mode (most common), we use λDU to calculate <strong>PFDavg</strong>.&nbsp;&nbsp;In continuous or high-demand modes, <strong>PFDavg is not used</strong>. Instead, SIL is based on <strong>PFH</strong> — the rate of dangerous failure per hour. In these cases, λDU is used to calculate PFH through a different series of equations than PFDavg.&nbsp;</li>
</ul>



<!-- FAQPage Schema (JSON-LD) — Verbatim conversion -->
<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "What’s the practical difference between λDU and λDD?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "λDU drives PFDavg because the failure is both dangerous and undetected. λDD is dangerous but detected, so it typically results in an alarm or forced-safe trip and may contribute to STR depending on configuration."
      }
    },
    {
      "@type": "Question",
      "name": "Do failure rates change over time in my facility?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Yes. The published λ values assume controlled conditions; however, real-world factors like environment, cycling, installation quality, and maintenance can shift the true failure rate up or down. However, we assume constant for modeling purposes."
      }
    },
    {
      "@type": "Question",
      "name": "Why do certified products help for accurate λ values?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Certified devices provide validated λ values and clearly defined DU/DD/SU/SD splits under IEC 61508. This reduces interpretation errors and ensures the assumptions behind the numbers are understood and controlled."
      }
    },
    {
      "@type": "Question",
      "name": "Is it okay to use generic failure rates?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Only if you confirm that they match your device and application. Generic values may not reflect your environment, proof-test strategy, or diagnostic coverage."
      }
    },
    {
      "@type": "Question",
      "name": "What if the device doesn’t have a SIL certificate?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "You can still use it, but you must rely on credible manufacturer reliability data, validated site history, or reputable sources such as OREDA. These represent the alternate data routes when a certified data path is not available. Assumptions must match the application, and justification must be documented."
      }
    },
    {
      "@type": "Question",
      "name": "Why do final elements almost always have higher λDU than sensors?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Most λDU in a final element comes from mechanical components like the actuator and valve body, which lack strong diagnostics. Sensors generally have better diagnostic coverage and fewer mechanical wear points, so their λDU values are typically much lower."
      }
    },
    {
      "@type": "Question",
      "name": "I understand why λDU impacts PFDavg, but why would λDD impact PFDavg?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "It depends on how the controls and diagnostics are set up. In many low-demand configurations, only λDU enters PFDavg. But if a dangerous detected failure leaves the SIF unable to perform its function—and the system does not act on or repair that diagnostic—then that portion of λDD effectively behaves like λDU and may need to be included in the PFDavg analysis."
      }
    },
    {
      "@type": "Question",
      "name": "How are λ values used differently in high-demand or continuous modes?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Failure rate is highly relevant but used differently. In low-demand mode (most common), we use λDU to calculate PFDavg. In continuous or high-demand modes, PFDavg is not used. Instead, SIL is based on PFH—the rate of dangerous failure per hour. In these cases, λDU is used to calculate PFH through a different series of equations than PFDavg."
      }
    }
  ]
}
</script>
]]></content:encoded>
					
					<wfw:commentRss>https://silsafe.net/failure-rates-in-functional-safety/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>How to Apply the Beta Factor: A Practical Guide to Common Cause Failures in SIL Verification</title>
		<link>https://silsafe.net/how-to-apply-beta-factor-for-common-cause-failure/</link>
					<comments>https://silsafe.net/how-to-apply-beta-factor-for-common-cause-failure/#respond</comments>
		
		<dc:creator><![CDATA[mamerten]]></dc:creator>
		<pubDate>Sun, 19 Oct 2025 16:44:46 +0000</pubDate>
				<category><![CDATA[Math Related]]></category>
		<category><![CDATA[PFDavg]]></category>
		<guid isPermaLink="false">https://silsafe.net/?p=3628</guid>

					<description><![CDATA[Learn how to apply the beta factor (β) in SIL verification and understand its link to common cause failures. This guide walks through where β values come from, how they affect PFDavg, and why small changes can shift your SIL rating.]]></description>
										<content:encoded><![CDATA[
<p>When engineers perform SIL verification, most of the attention goes toward failure rates, proof test intervals, or diagnostics. But one input that carries enormous influence is <strong>beta factor (β)</strong> — the number that represents how much of your redundancy can be trusted to behave independently. Misunderstanding it can make even a perfect SIL calculation look better on paper than it really is in the field. This article walks through what β is, how it links to <strong>common cause failure (CCF)</strong>, and how to apply it correctly with practical math and examples.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">What Is It? (CCF vs. Beta Factor and Why It Matters)</h2>



<p><strong>Common cause failure (CCF)</strong>&nbsp;is a concept of multiple things failing for the same core reason. It is a concept in various industries and practices. In a Safety Instrumented System (SIS) it means multiple channels in redundant architectures fail together because of a shared cause. For example:</p>



<ul class="wp-block-list">
<li><strong>Sensors:</strong> Two pressure transmitters mounted side‑by‑side exposed to the same vibration or plugged impulse lines.</li>



<li><strong>Logic solver:</strong> Both processor cards affected by the same unknown software bug.</li>



<li><strong>Final element:</strong> Two solenoids sharing the same instrument air header that fail simultaneously.</li>
</ul>



<p>The <strong>beta factor (β)</strong> is the fraction of all dangerous failures that occur from a CCF. It is the mathematical representation of the CCF concept. It converts the qualitative idea of “common cause” into a quantifiable term used in equations. Ignoring β will almost always make your <strong>average probability of failure on demand (PFDavg)</strong> appear lower than it really is — giving a false sense of security.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">Where to Get Beta Factor Values</h2>



<p>There are two primary sources:</p>



<ol class="wp-block-list">
<li><strong>SIL certificates or FMEDA reports</strong> from manufacturers. These often include β values based on test data and modeling assumptions.</li>



<li><strong>IEC 61508‑6 Annex D</strong> – the formal method for determining β using dependent failure analysis. This process is complex and beyond the scope of this article, but it is the standard reference.</li>
</ol>



<p>Other useful references include <strong>ISA TR84.00.02</strong>, <strong>OREDA</strong>, and <strong>CCPS</strong> reliability publications. Some companies maintain internal databases based on field performance.</p>



<p>A practical approach:</p>



<ul class="wp-block-list">
<li>Start with the β from the SIL certificate.</li>



<li>Review Annex D factors to see if your installation justifies adjustment.</li>



<li>Document and justify the final value in your <strong>Safety Requirements Specification (SRS)</strong>.</li>
</ul>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">How Architecture Affects β</h2>



<p>The β‑factor is not constant; it changes with architecture. This is a common point of confusion within Functional Safety. Think of redundancy as a team of <strong>security guards protecting a building</strong>:</p>



<ul class="wp-block-list">
<li>In a <strong>1oo2</strong> architecture, either guard can respond to stop a robbery. If both guards eat the same lunch and get food poisoning, that’s a CCF — a shared vulnerability.</li>



<li>In a <strong>1oo4</strong> architecture, four guards must all get sick for the robbery to succeed. The common cause event would have to be much stronger or more universal, or &#8220;more common,&#8221; to take them all down. Think of a really widespread case of food poisoning. Therefore, the effective β is lower.</li>
</ul>



<p>This creates a feedback loop: changing architecture alters the PFDavg equation, but it also justifies a new β — which again changes the PFDavg. It is a bit odd, but it is correct. IEC 61508‑6 Annex D Table D.5 gives architectural correction factors (roughly 0.3 – 1.74) that help account for this effect. Most β values published in SIL certificates are assumed for 1oo2 configurations.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">NooN Architectures and Why β Is Not Present</h2>



<p>For <strong>NooN</strong> architectures (e.g., 2oo2 or 3oo3), β is normally not present in the mathematics. Any single channel failure will cause the function to fail because all channels must work. The total λDU (dangerous undetected failure rate) already includes both independent and common‑cause contributions. Applying a separate β term would double‑count the effect.</p>



<p>Returning to the guard analogy: if three guards must all respond correctly and any one failure causes system failure, it doesn’t matter if their failures were independent or shared — the result is the same. For this reason, modeling tools and standards do not include β for NooN designs.</p>



<p>Note that this can be thought of as &#8220;β is set to 0 for NooN&#8221; but that is not how we at SIL Safe think of this. Yes, you can use the MooN PFDavg equations, set&nbsp;β = 0, and get the same results. But as we discussed above,&nbsp;β is accounted for in λDU.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">Typical Values and Influencing Factors</h2>



<figure class="wp-block-table"><table class="has-fixed-layout"><thead><tr><th>Subsystem</th><th>Typical β Range</th></tr></thead><tbody><tr><td>Sensors</td><td>0.02 – 0.10</td></tr><tr><td>Logic Solvers</td><td>0.01 – 0.05</td></tr><tr><td>Final Elements</td><td>0.05 – 0.15</td></tr></tbody></table></figure>



<p>Factors that increase β:</p>



<ul class="wp-block-list">
<li>Common environment (same enclosure, same power, same impulse lines)</li>



<li>Identical design and firmware</li>



<li>Shared maintenance procedures or simultaneous testing</li>
</ul>



<p>Factors that decrease β:</p>



<ul class="wp-block-list">
<li>Physical separation and shielding</li>



<li>Vendor or technology diversity</li>



<li>Independent power and utilities</li>



<li>Separate calibration schedules</li>



<li>Staggered proof testing</li>
</ul>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">PFDavg Equation Discussion</h2>



<p>In a <strong>1oo2</strong> system, the total PFDavg is made up of independent and common‑cause terms. Using a simplified version of the equation:</p>



<figure class="wp-block-image size-full is-resized"><img decoding="async" width="560" height="78" src="https://silsafe.net/wp-content/uploads/2025/10/PFDavg-1oo2-basic.webp" alt="Equation showing how the beta factor (β) influences PFDavg in a 1oo2 safety architecture by combining independent and common cause failure probabilities." class="wp-image-3626" style="aspect-ratio:7.180925666199158;width:438px;height:auto" srcset="https://silsafe.net/wp-content/uploads/2025/10/PFDavg-1oo2-basic.webp 560w, https://silsafe.net/wp-content/uploads/2025/10/PFDavg-1oo2-basic-300x42.webp 300w" sizes="(max-width: 560px) 100vw, 560px" /></figure>



<p>Where:</p>



<ul class="wp-block-list">
<li>λ<sub>DU</sub> = dangerous undetected failure rate (per hour)</li>



<li>TI = proof test interval (hours)</li>



<li>β = fraction of failures due to common causes</li>
</ul>



<p>The first term (squared) represents two independent latent failures. The second (linear) term represents the probability that both channels fail from a shared cause. Because the independent term is squared, it becomes much smaller — making β a powerful driver in redundant designs.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">PFDavg Calculation Examples (1oo2)</h2>



<h3 class="wp-block-heading">Case A – Good β Factor (Target: SIL 2)</h3>



<p><strong>Given:</strong> β = 0.03, λ<sub>DU</sub> = 2E-6/hr, TI = 8,760 hr (1 year)</p>



<ol class="wp-block-list">
<li>Independent term:<br>(1/3) × [(1 − 0.03) × 2E-6 × 8,760]² = 9.63E-5</li>



<li>Common‑cause term:<br>(1/2) × 0.03 × 2E-6 × 8,760 = 2.63E-4</li>



<li><strong>Total PFDavg </strong>= 3.59E-4</li>



<li><strong>RRF = 1 / PFDavg </strong>= 2,785 → SIL 2</li>
</ol>



<p><strong>STR (spurious trip rate):</strong> assume λ<sub>SP</sub> = 1E-5/hr per channel. For 1oo2 (trip on any channel):<br>STR ≈ 2 × λ<sub>SP</sub> = 2E-5/hr → 0.175 trips per year ≈ a trip every 5.7 years.</p>



<h3 class="wp-block-heading">Case B – Poor β Factor (Target: SIL 1)</h3>



<p><strong>Given:</strong> β = 0.15, λ<sub>DU</sub> = 2E-6/hr, TI = 8,760 hr (same assumptions)</p>



<ol class="wp-block-list">
<li>Independent term:<br>(1/3) × [(1 − 0.15) × 2E-6 × 8,760]² = 7.4E-5</li>



<li>Common‑cause term:<br>(1/2) × 0.15 × 2E-6 × 8,760 = 1.31E-3</li>



<li><strong>Total PFDavg </strong>= 1.38E-3</li>



<li><strong>RRF = 1 / PFDavg </strong>= 725 → SIL 1</li>
</ol>



<p><strong>STR (same λ<sub>SP</sub>):</strong> 2E-5/hr → 0.175 trips/year or one trip every 5.7 years (same as Case A).</p>



<p><strong>Comparison:</strong> Raising β from 0.03 to 0.15 increased total PFDavg by ≈ 4×, dropping the SIL from 2 to 1.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">Impact of Common Cause Portions of PFDavg</h2>



<p>Because the independent term is squared, it diminishes as reliability increases, while the common‑cause term remains linear. The result: CCF often dominates total PFDavg in redundant architectures.</p>



<ul class="wp-block-list">
<li><strong>Case A (β = 0.03)</strong>: PFDavg_ind = 9.63E-5, PFDavg = 2.63E-4 → CCF ≈ 73% of total.</li>



<li><strong>Case B (β = 0.15)</strong>: PFDavg_ind = 7.4E-5, PFDavg = 1.31E-3 → CCF ≈ 95% of total.</li>
</ul>



<p>Even at modest β values, most of the total unavailability stems from common causes — which is why reducing β through independence and diversity is far more effective than upgrading the architecture. The same pattern holds true for higher architectures like 2oo3.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">Common Mistakes</h2>



<ul class="wp-block-list">
<li>Applying the same β to every subsystem without justification.</li>



<li>Assuming diagnostics lower β (they don’t; they reduce λ<sub>DU</sub>).</li>



<li>Forgetting to justify β selection in the SRS.</li>



<li>Focusing only on adding redundancy instead of lowering β.</li>
</ul>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">Summary</h2>



<p>The β‑factor is what connects real‑world dependencies to your math. A small change in β can shift a design from SIL 2 to SIL 1 — even when every other parameter is identical. Real independence, not just redundancy, drives reliability.</p>



<p>Document your β assumptions, revisit them throughout the SIS life‑cycle, and when in doubt, verify them against IEC 61508‑6 Annex D or manufacturer data.</p>



<p><em>If you’d like a second set of eyes on your β assumptions or SIL verification model, contact <strong>SIL Safe</strong> for a practical review or training session.</em></p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">Q&amp;A</h2>



<p><strong>1. Does beta factor (β) apply only to 1oo2 architectures?</strong><br>No. The concept applies to any redundant system where channels can fail together, though it’s most visible in 1oo2.</p>



<p><strong>2. Can β ever be zero?</strong><br>Not realistically. There’s always some shared factor — environment, maintenance, or design — that introduces correlation.&nbsp; For NooN scenarios,&nbsp;<strong>β</strong> does not appear in the mathematics, so some say&nbsp; β is zero, but it is still included in the&nbsp;λ<sub>DU</sub> term.</p>



<p><strong>3. What does β actually represent?</strong><br>It’s the percentage of dangerous failures that are shared between channels rather than truly independent.</p>



<p><strong>4. How often should β be revisited?</strong><br>Any time design, environment, or procedures change — typically during periodic review or FSA Stage 4.&nbsp; It is common for&nbsp;β to change over the safety lifecycle.</p>



<p><strong>5. Can one simply apply the β in the SIL certificate?</strong><br>Use it as a starting point. Then compare against Annex D’s checklist and justify any adjustment.</p>



<p><strong>6. What portion can common cause contribute to total PFDavg?</strong><br>Often the majority. For 1oo2 with β ≈ 0.1–0.2, common cause failures can make up 70–95 % of total PFDavg — the same trend appears in higher architectures.</p>



<p><strong>7. I&#8217;ve heard that for NooN architectures, β is set to 0, but I thought that is not possible?</strong><br>β is involved in NooN, but it is already included in the&nbsp;λ<sub>DU</sub> term. Meaning we don&#8217;t want to double count. The NooN equations will not have any&nbsp;β terms. Thus, it can be thought of as they were set to zero, but that is not technically correct.</p>



<h2 class="wp-block-heading">Further Reading</h2>



<p>Here are additional resources</p>



<ol class="wp-block-list">
<li><a href="https://silsafe.net/functional-safety-glossary/" data-type="page" data-id="391">SIL Safe Glossary </a>&#8211; This is SIL Safe&#8217;s glossary of 100+ functional safety terms.</li>



<li>Guidance on quantification of common cause failure&nbsp; <a href="https://ez.analog.com/ez-blogs/b/engineerzone-spotlight/posts/how-to-quantify-common-cause-failures" target="_blank" rel="noopener">https://ez.analog.com/ez-blogs/b/engineerzone-spotlight/posts/how-to-quantify-common-cause-failures</a></li>



<li>Nuclear Regulatory Commission (NRC) Guidance &#8211; The US nuclear industry does not follow IEC 61511-1, but they care a LOT about common cause failure.&nbsp; Here is their guidance.&nbsp;<a href="https://www.nrc.gov/docs/ml0729/ml072970404.pdf" target="_blank" rel="noopener">https://www.nrc.gov/docs/ml0729/ml072970404.pdf</a></li>



<li>More NRC Guidance &#8211;&nbsp;<a href="https://www.nrc.gov/reading-rm/doc-collections/nuregs/staff/sr2225/index" target="_blank" rel="noopener">https://www.nrc.gov/reading-rm/doc-collections/nuregs/staff/sr2225/index</a></li>
</ol>



<!-- FAQPage Schema (JSON-LD) — Verbatim conversion -->
<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "Does beta factor (β) apply only to 1oo2 architectures?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "No. The concept applies to any redundant system where channels can fail together, though it’s most visible in 1oo2."
      }
    },
    {
      "@type": "Question",
      "name": "Can β ever be zero?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Not realistically. There’s always some shared factor — environment, maintenance, or design — that introduces correlation. For NooN scenarios, β does not appear in the mathematics, so some say β is zero, but it is still included in the λDU term."
      }
    },
    {
      "@type": "Question",
      "name": "What does β actually represent?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "It’s the percentage of dangerous failures that are shared between channels rather than truly independent."
      }
    },
    {
      "@type": "Question",
      "name": "How often should β be revisited?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Any time design, environment, or procedures change — typically during periodic review or FSA Stage 4. It is common for β to change over the safety lifecycle."
      }
    },
    {
      "@type": "Question",
      "name": "Can one simply apply the β in the SIL certificate?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Use it as a starting point. Then compare against Annex D’s checklist and justify any adjustment."
      }
    },
    {
      "@type": "Question",
      "name": "What portion can common cause contribute to total PFDavg?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Often the majority. For 1oo2 with β ≈ 0.1–0.2, common cause failures can make up 70–95 % of total PFDavg — the same trend appears in higher architectures."
      }
    },
    {
      "@type": "Question",
      "name": "I’ve heard that for NooN architectures, β is set to 0, but I thought that is not possible?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "β is involved in NooN, but it is already included in the λDU term. Meaning we don't want to double count. The NooN equations will not have any β terms. Thus, it can be thought of as they were set to zero, but that is not technically correct."
      }
    }
  ]
}
</script>

]]></content:encoded>
					
					<wfw:commentRss>https://silsafe.net/how-to-apply-beta-factor-for-common-cause-failure/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Spurious Trip Rate Explained: 6 Facts Every Functional Safety Engineer Should Know</title>
		<link>https://silsafe.net/spurious-trip-rate-explained/</link>
					<comments>https://silsafe.net/spurious-trip-rate-explained/#respond</comments>
		
		<dc:creator><![CDATA[mamerten]]></dc:creator>
		<pubDate>Wed, 03 Sep 2025 02:24:51 +0000</pubDate>
				<category><![CDATA[Math Related]]></category>
		<category><![CDATA[Beginner]]></category>
		<guid isPermaLink="false">https://silsafe.net/?p=3315</guid>

					<description><![CDATA[Spurious Trip Rate (STR) is often overlooked in functional safety. Learn what STR is, how it’s calculated, and why balancing STR with PFDavg matters.]]></description>
										<content:encoded><![CDATA[
<h2 class="wp-block-heading">Introduction</h2>



<p>When engineers think about <strong>functional safety</strong> and IEC 61511, they often focus on <em>probability of failure on demand (PFDavg)</em>. But another metric—<strong>spurious trip rate (STR)</strong>—directly affects plant uptime, productivity, and costs. Unlike PFDavg, STR doesn’t measure safety performance. Instead, it reflects how often a Safety Instrumented Function (SIF) causes unwanted shutdowns. In this article, we’ll break down STR, introduce its inverse (MTTFsp), show how it’s calculated, and explain why balancing STR with PFDavg is essential for effective SIF design.</p>



<h2 class="wp-block-heading">What Is Spurious Trip Rate?</h2>



<p>Spurious trip rate (STR) is the <strong>frequency of unwanted or false trips</strong> of a Safety Instrumented Function (SIF).</p>



<p>These aren’t dangerous failures—they’re <em>safe failures</em> that cause a SIF to shut down the process when no actual demand exists. STR is usually expressed in failures per hour (1/hr) and then converted to years between trips for practical interpretation.</p>



<p>To clarify:</p>



<ul class="wp-block-list">
<li><strong>Dangerous failures</strong> → drive safety risk and are addressed through PFDavg.</li>



<li><strong>Safe failures (that force shutdowns)</strong> → drive nuisance trips and are captured in STR.</li>
</ul>



<h2 class="wp-block-heading">Why STR Matters in Functional Safety</h2>



<p>Unnecessary shutdowns don’t just reduce production—they can create safety hazards of their own. Restarting equipment introduces operational risk, and frequent nuisance trips erode operators’ trust in SIS performance. IEC 61511 requires that designers consider both <strong>safety integrity (via PFDavg)</strong> and <strong>availability (via STR)</strong>, even though the standard does not prescribe numeric STR limits.</p>



<p>A SIF with an overly high STR might meet its SIL target but still be unacceptable to operations because of lost uptime. Spurious trips matters because it links functional safety design to real-world economics.</p>



<p>Alternatively, some operations are so automated that recovering from a SIF trip is only a minor problem. It just depends.</p>



<h2 class="wp-block-heading">STR and Its Inverse: MTTFsp</h2>



<p>STR is expressed in trips per unit time—often failures per hour. But that can be difficult to interpret. So, engineers often use its inverse: <strong>MTTFsp (Mean Time to Spurious Trip)</strong> .</p>



<ul class="wp-block-list">
<li><strong>MTTFsp = 1 / STR</strong></li>



<li>Example: STR = 0.01 trips/year → MTTFsp = 100 years</li>
</ul>



<p>Both terms are useful:</p>



<ul class="wp-block-list">
<li><strong>STR</strong> highlights the frequency of nuisance trips.</li>



<li><strong>MTTFsp</strong> highlights how long, on average, a SIF runs without a false trip.</li>
</ul>



<p>At SIL Safe, we always calculate both terms in verification reports. This makes it easier for both engineers and managers to understand the balance between reliability and uptime.</p>



<h2 class="wp-block-heading">How STR Is Calculated</h2>



<p>The first step is defining the <strong>λsp (spurious failure rate)</strong>. This aggregates all relevant failure modes in the SIF path that can drive a false trip:</p>



<ul class="wp-block-list">
<li>If a detected failure is configured to <strong>vote-to-trip</strong>, it contributes to λsp.</li>



<li>If a detected failure is configured to <strong>notify-only</strong>, it does not contribute.</li>



<li>If there are no diagnostics, then the diagnostic terms are already 0.</li>
</ul>



<p>Thus, the voting philosophy directly affects which categories—safe detected (SD), safe undetected (SU), dangerous detected (DD)—feed into λsp. Dangerous undetected (DU) never contributes to STR.</p>



<ul class="wp-block-list">
<li><strong>1oo1 SIF:</strong> STR = λsp (per unit time).</li>



<li><strong>1ooN SIF:</strong> STR increases with more elements in parallel (anyone can trip the system).
<ul class="wp-block-list">
<li>STR=N*λsp</li>
</ul>
</li>



<li><strong>MooN SIF (M &gt; 1):</strong> STR decreases because multiple channels must trip simultaneously (e.g., 2oo3 cuts nuisance trip probability). This is done through the often-confusing MooN equation, which is beyond the scope here.</li>
</ul>



<p>Contrast: PFDavg calculations focus on dangerous undetected failures (DU), while STR calculations focus on safe or detected failures that cause spurious trips.</p>



<h2 class="wp-block-heading">Example Calculations</h2>



<h3 class="wp-block-heading">Case A – 1oo1, No Diagnostics</h3>



<ul class="wp-block-list">
<li><strong>Assumptions:</strong> Single-channel SIF, no diagnostics.</li>



<li><strong>Failure rates per channel:</strong>
<ul class="wp-block-list">
<li>λ<sub>DU</sub> = 1E-6 /hr (listed for completeness; not in calculation)</li>



<li>λ<sub>DD</sub> = 0 /hr</li>



<li>λ<sub>SU</sub> = 2E-6 /hr</li>



<li>λ<sub>SD</sub> = 0 /hr</li>
</ul>
</li>



<li><strong>λsp definition:</strong>&nbsp;As there are no diagnostics, DD and SD are not applicable.&nbsp; DU is never applicable.&nbsp; Thus, λsp = λ<sub>SU</sub> = 2E-6 /hr</li>



<li><strong>Results:</strong><br>STR = <strong>2E-6 /hr</strong><br>MTTFsp = 1 / (λ<sub>sp</sub> × 8,760) = <strong>≈ 57.1 years</strong> (≈ 1 trip every 57 years per channel)</li>
</ul>



<h3 class="wp-block-heading">Case B – 1oo2 Sensors, Diagnostics with Vote-to-Trip</h3>



<ul class="wp-block-list">
<li><strong>Assumptions:</strong> Dual-channel 1oo2 sensor architecture; online diagnostics present; detected faults (SD and DD) vote-to-trip. Only the sensor portion is considered.</li>



<li><strong>Failure rates per channel:</strong>
<ul class="wp-block-list">
<li>λ<sub>DU</sub> = 1.0E-6 /hr (listed for completeness; not in calculation)</li>



<li>λ<sub>DD</sub> = 2.0E-7 /hr</li>



<li>λ<sub>SU</sub> = 4.0E-7 /hr</li>



<li>λ<sub>SD</sub> = 8.0E-7 /hr</li>
</ul>
</li>



<li><strong>λsp per channel:</strong> λ<sub>sp</sub> = λ<sub>SD</sub> + λ<sub>SU</sub> + λ<sub>DD</sub> = 1.4E-6 /hr</li>



<li><strong>System results:</strong><br>STR = 2 × 1.4E-6 = <strong>2.8E-6 /hr</strong><br>MTTFsp = 1 / (STR × 8,760) = <strong>≈ 40.8 years</strong> (≈ 1 trip every 41 years for sensor portion)</li>
</ul>



<p><strong>Key takeaway:</strong> architecture, diagnostics, and voting logic can drastically change nuisance trip rates. Presenting both STR and MTTFsp highlights the trade-offs clearly.</p>



<h2 class="wp-block-heading">Who Decides What STR Is Acceptable?</h2>



<p>Many plants rely on guidance from firms like SIL Safe to set realistic STR expectations.&nbsp; Unlike SIL targets, there is <strong>no universal STR requirement</strong>. Acceptability is decided by plant management and operations during safety design reviews. It’s typically documented in the <strong>Safety Requirements Specification (SRS)</strong>.</p>



<ul class="wp-block-list">
<li><strong>Typical ranges:</strong> MTTFsp of 10–100 years per channel is often considered reasonable in industry.</li>



<li><strong>Higher MTTFsp expectations:</strong> High-availability processes (e.g., refineries, offshore platforms).</li>



<li><strong>Lower MTTFsp expectations:</strong> Non-critical utilities or batch processes.</li>



<li><strong>Unrealistic targets:</strong> “Once every million years” is neither achievable nor useful.&nbsp; However, this does happen as new people are introduced to functional safety.</li>
</ul>



<p>The point: STR goals are practical business decisions, not dictated by IEC 61511.&nbsp; The standard IEC 61511-1 requires the team to consider it against safety and PFDavg.</p>



<h2 class="wp-block-heading">Balancing Safety Integrity and Availability</h2>



<p>The art of SIF design is balancing <strong>safety integrity (low PFDavg)</strong> with <strong>availability (low STR)</strong>:</p>



<ul class="wp-block-list">
<li>Use redundancy wisely (2oo3 voting can cut STR).&nbsp; This is the primary reason 2oo3 is so common.</li>



<li>Apply diagnostics carefully (vote-to-trip vs notify-only matters).</li>



<li>Base calculations on realistic failure data, not overly optimistic assumptions.</li>
</ul>



<p>A SIS with perfect safety but terrible availability—or vice versa—fails its mission.</p>



<h2 class="wp-block-heading">Conclusion</h2>



<p>Spurious Trip Rate doesn’t affect whether a plant is safe, but it absolutely affects whether it runs. That’s why STR is as important as PFDavg in practice. Engineers must present both STR and MTTFsp to give operators a realistic picture of system performance. The best designs find the balance: safe enough <strong>and</strong> reliable enough.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">Q&amp;A Section</h2>



<p><strong>1. Is a low STR always better?</strong><br>Yes for uptime, but not if it compromises safety. STR must be balanced against PFDavg.</p>



<p><strong>2. Does IEC 61511 require specific STR values?</strong><br>No. It requires that availability and spurious trips be considered, but it does not prescribe limits.</p>



<p><strong>3. Can proof testing increase STR?</strong><br>Yes. Poorly planned tests (e.g., cycling valves without bypass) can cause nuisance trips. This temporarily raises STR and lowers MTTFsp.</p>



<p><strong>4. Which devices dominate STR?</strong><br>Final elements (valves) often contribute the most to spurious trips because of high safe failure rates.</p>



<p><strong>5. How does redundancy reduce STR?</strong><br>Voting architectures (like 2oo3) allow one sensor to fail without tripping the system, lowering STR.</p>



<p><strong>6. Where is acceptable STR documented?</strong><br>Usually in the Safety Requirements Specification (SRS) or reliability design basis documents.</p>



<p><strong>7. Can STR goals cause conflict between engineers and operations?</strong><br>Absolutely. Operations may push for fewer nuisance trips, while safety engineers emphasize conservatism. Finding balance is key.&nbsp; The stakeholders need to discuss it together.</p>



<h2 class="wp-block-heading">Learn More</h2>



<ul class="wp-block-list">
<li>SIL Safe has a <a href="https://silsafe.net/functional-safety-glossary/">full glossary</a> and a much shorter entry for <a href="https://silsafe.net/glossary/spurious-trip-rate/" data-type="glossary" data-id="895">spurious trip rate</a>.  </li>



<li><a href="https://webstore.iec.ch/en/publication/61289" target="_blank" rel="noopener">IEC 61511-1 page</a></li>
</ul>



<!-- FAQPage Schema (JSON-LD) — Verbatim conversion -->
<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "Is a low STR always better?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Yes for uptime, but not if it compromises safety. STR must be balanced against PFDavg."
      }
    },
    {
      "@type": "Question",
      "name": "Does IEC 61511 require specific STR values?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "No. It requires that availability and spurious trips be considered, but it does not prescribe limits."
      }
    },
    {
      "@type": "Question",
      "name": "Can proof testing increase STR?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Yes. Poorly planned tests (e.g., cycling valves without bypass) can cause nuisance trips. This temporarily raises STR and lowers MTTFsp."
      }
    },
    {
      "@type": "Question",
      "name": "Which devices dominate STR?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Final elements (valves) often contribute the most to spurious trips because of high safe failure rates."
      }
    },
    {
      "@type": "Question",
      "name": "How does redundancy reduce STR?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Voting architectures (like 2oo3) allow one sensor to fail without tripping the system, lowering STR."
      }
    },
    {
      "@type": "Question",
      "name": "Where is acceptable STR documented?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Usually in the Safety Requirements Specification (SRS) or reliability design basis documents."
      }
    },
    {
      "@type": "Question",
      "name": "Can STR goals cause conflict between engineers and operations?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Absolutely. Operations may push for fewer nuisance trips, while safety engineers emphasize conservatism. Finding balance is key, and stakeholders need to discuss it together."
      }
    }
  ]
}
</script>
]]></content:encoded>
					
					<wfw:commentRss>https://silsafe.net/spurious-trip-rate-explained/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>How to Use Proof Test Coverage (Cpt) to Improve PFDavg</title>
		<link>https://silsafe.net/proof-test-coverage-cpt-to-improve-pfdavg/</link>
					<comments>https://silsafe.net/proof-test-coverage-cpt-to-improve-pfdavg/#respond</comments>
		
		<dc:creator><![CDATA[mamerten]]></dc:creator>
		<pubDate>Sun, 03 Aug 2025 22:15:25 +0000</pubDate>
				<category><![CDATA[Beginner]]></category>
		<category><![CDATA[Math Related]]></category>
		<guid isPermaLink="false">https://silsafe.net/?p=2784</guid>

					<description><![CDATA[Proof test coverage (Cpt) plays a major role in determining whether a SIF meets its target SIL. This article explains what Cpt is, how it affects PFDavg, and how even a small change in coverage can shift your system from SIL 1 to SIL 2—or vice versa. Includes examples, tips, and FAQs.]]></description>
										<content:encoded><![CDATA[
<p>When designing or verifying a Safety Instrumented Function (SIF), it’s common to hear terms like PFDavg, SIL level, and test interval. But one factor that’s often misunderstood — or just overlooked — is <strong>proof test coverage (Cpt)</strong>. This is a critical element that directly impacts how effectively your testing finds dangerous failures.</p>



<p>If your facility is working toward compliance with <strong>IEC 61511-1</strong>, understanding how Cpt works — and how to apply it — can make the difference between an overly optimistic SIL claim and a realistic, defensible safety case.</p>



<p>Let’s walk through what Cpt is, how it affects your calculations, and how to apply it the right way.</p>



<h2 class="wp-block-heading">What Is Proof Test Coverage (Cpt)?</h2>



<p><strong>Proof test coverage (Cpt)</strong> is the fraction of dangerous undetected (DU) failures that your proof test is capable of finding.</p>



<ul class="wp-block-list">
<li>A Cpt of 1.0 (or 100%) means your test detects <strong>all</strong> dangerous undetected failures.</li>



<li>A Cpt of 0.7 (or 70%) means your test only finds <strong>70%</strong> of those failures.</li>
</ul>



<p>This is important because any dangerous failures that your test doesn’t catch will accumulate over time, increasing the <strong>average probability of failure on demand (PFDavg)</strong>.</p>



<p>Cpt is often used alongside another key term: <strong>proof test interval (TI)</strong> — how often you do the testing. But the test interval doesn’t matter much if your test isn’t catching what matters.</p>



<p>Also worth noting: Cpt is <em>not</em> the same thing as <strong>diagnostic coverage (DC)</strong> — though they both relate to detecting failures, they’re measured differently and come from different sources.</p>



<p>Note that proof tests can and do capture beyond DU failures. It can also catch safe failures (SU, SD). But the main <span style="text-decoration: underline;">purpose of</span> proof test is to find DU failures and that is the <span style="text-decoration: underline;">only failure</span> Cpt is associated with.</p>



<h2 class="wp-block-heading">How Cpt Affects PFDavg</h2>



<p>The most common form of the PFDavg equation used in training looks like this:</p>



<figure class="wp-block-image size-full is-resized"><img loading="lazy" decoding="async" width="212" height="58" src="https://silsafe.net/wp-content/uploads/2025/04/pfdavg-basic-equation.png" alt="PFDavg basic equation" class="wp-image-188" style="aspect-ratio:3.6554492127199754;width:160px;height:auto"/></figure>



<p>But this assumes you catch <em>all</em> dangerous undetected (DU) failures — which is rarely true. A more accurate form includes <strong>Cpt</strong>:</p>



<figure class="wp-block-image is-resized"><img loading="lazy" decoding="async" width="516" height="83" src="https://silsafe.net/wp-content/uploads/2025/08/PVDavg-with-Cpt.webp" alt="Mathematical equation for calculating PFDavg using proof test coverage (Cpt), test interval (TI), and SIS lifetime (LT) for SIL verification." class="wp-image-2781" style="aspect-ratio:6.218258624735977;width:340px;height:auto" srcset="https://silsafe.net/wp-content/uploads/2025/08/PVDavg-with-Cpt.webp 516w, https://silsafe.net/wp-content/uploads/2025/08/PVDavg-with-Cpt-300x48.webp 300w" sizes="auto, (max-width: 516px) 100vw, 516px" /></figure>



<p>Where:</p>



<ul class="wp-block-list">
<li><strong>λ<sub>DU</sub></strong> is the dangerous undetected failure rate</li>



<li><strong>TI</strong> is the proof test interval</li>



<li><strong>LT</strong> is the SIS lifetime (e.g., 15 or 20 years)</li>
</ul>



<p>The two terms in the equation represent different contributions to PFDavg, as explained below:</p>



<ul class="wp-block-list">
<li>The first term is the contribution <strong>between tests</strong>.</li>



<li>The second term is the contribution of failures that remain hidden even during proof tests that apply for the lifetime.  </li>
</ul>



<p>In many training or spreadsheet tools, the second term is omitted if the <strong>lifetime is similar to the test interval</strong>. But when the lifetime is significantly longer (e.g., TI = 1 year, LT = 15 years), <strong>ignoring it underestimates risk</strong>.</p>



<p>Let’s make a quick comparison:</p>



<p><strong>Example:</strong></p>



<ul class="wp-block-list">
<li>λ<sub>DU</sub> = 2E-6 per hour</li>



<li>TI = 1 year (8,760 hours)</li>



<li>LT = 15 years (131,400 hours)</li>



<li>Case A: Cpt = 0.55</li>



<li>Case B: Cpt = 0.95</li>
</ul>



<p><strong>Case A: PFDavg ≈ (2E-6 × 0.55 × 8760)/2 + (2E-6 × 0.45 × 131400)/2</strong> = <strong>1.04E-2 → RRF ≈ 96</strong> (SIL 1)</p>



<p><strong>Case B: PFDavg ≈ (2E-6 × 0.95 × 8760)/2 + (2E-6 × 0.05 × 131400)/2</strong> = <strong>2.06E-3 → RRF ≈ 485</strong> (SIL 2)</p>



<p>👉 This is the difference between a <strong>SIL 1 system</strong> and a <strong>SIL 2 system</strong> — driven entirely by proof test coverage.</p>



<p>Even though both cases used the same failure rate, test interval, and SIS lifetime, the lower test coverage in Case A pulled the risk performance down an entire SIL level. This is a powerful reminder that <strong>increasing test frequency is not enough</strong> if the test itself isn’t catching the right failure modes.</p>



<h2 class="wp-block-heading">What’s a Realistic Cpt?</h2>



<p>You’ll often see vendors or safety books quote generic Cpt ranges. Here’s a quick cheat sheet to get you started:</p>



<figure class="wp-block-table"><table class="has-fixed-layout"><thead><tr><th>Component</th><th>Typical Cpt</th><th>Notes</th></tr></thead><tbody><tr><td>Pressure Transmitter</td><td>85–95%</td><td>Depends on how it&#8217;s tested</td></tr><tr><td>Logic Solver</td><td>95–99%</td><td>High diagnostic coverage helps</td></tr><tr><td>Final Element (valve)</td><td>50–95%</td><td>Greatly depends on stroke testing</td></tr></tbody></table></figure>



<p>Cpt is influenced by:</p>



<ul class="wp-block-list">
<li><strong>Test method</strong> (partial stroke, full stroke, leak test, etc.)</li>



<li><strong>Equipment design</strong> (some valves are inherently testable)</li>



<li><strong>Human factors</strong> (procedures, training, consistency)</li>
</ul>



<h2 class="wp-block-heading">How to Determine Proof Test Coverage (Cpt)</h2>



<h3 class="wp-block-heading">If Using IEC 61508 Certified Equipment</h3>



<p>If you’re using components that are <strong>certified per IEC 61508</strong>, your job is easier. Look at the <strong>SIL certificate</strong> or <strong>safety manual</strong>. Most will include Cpt values based on an FMEDA (Failure Modes, Effects, and Diagnostic Analysis).</p>



<ul class="wp-block-list">
<li>Example: A final element might claim 65% for partial-stroke testing and 90% for full-stroke testing.</li>



<li>You need to match <strong>your test procedure</strong> to what was assumed in the FMEDA.</li>
</ul>



<p>This is especially important with valves. Partial stroke tests (PVST) might not catch failure modes that a full test (FVST) would — and the difference in Cpt can be dramatic.</p>



<h3 class="wp-block-heading">If Using Non-61508 Equipment (Route 2H or 2S)</h3>



<p>If your hardware isn’t certified, you’ll need to get data and use the <strong>proven in use</strong>&nbsp;method (this takes the <strong>route 2H or 2S</strong> approach, routes are confusing and will be discussed elsewhere).:</p>



<ul class="wp-block-list">
<li>Use industry databases like <strong>OREDA</strong></li>



<li>Refer to books like <em>Safety Instrumented System Verification</em> by Goble</li>



<li>Review ISA technical reports and peer-reviewed FMEDAs</li>



<li>Document your <strong>engineering judgment</strong> and <strong>conservatism</strong></li>
</ul>



<p>Example: You might assign a Cpt of 70% to a test routine that checks for mechanical failure in a solenoid but can’t detect seat leakage. Be transparent about assumptions — auditors and assessors will ask.</p>



<h2 class="wp-block-heading">Common Proof Test Coverage Misunderstandings</h2>



<ul class="wp-block-list">
<li><strong>Cpt ≠ diagnostic coverage</strong>: Diagnostic coverage comes from built-in self-checks. Cpt is about your <strong>manual or automatic testing</strong> procedures.</li>



<li><strong>You can’t just assume 100%</strong>: Even a full-stroke test may not catch all dangerous failures, especially in actuators and valve internals.</li>



<li><strong>Test frequency doesn’t override poor Cpt</strong>: Doing a weak test more often doesn’t give the same benefit as a strong test less frequently.</li>



<li>Copying vendor Cpt while using a much weaker plant test; assuming partial‑stroke test Cpt equals full functional Cpt.</li>



<li>Not coordinating the type of Cpt with the plant possibly needing to be shutdown. For example, a Functional Safety Engineers establishes a high Cpt assuming a FVST every six months, which would require a plant shutdown in their situation. But the facility cannot be shutdown every six months. This is a classic example of not including all stakeholders in decisions.</li>
</ul>



<h2 class="wp-block-heading">Practical Tips for Beginners</h2>



<ul class="wp-block-list">
<li>If possible, <strong>use certified equipment</strong> — it saves work and improves defensibility.</li>



<li>For valves and final elements, be clear with your operations team: A test that’s easy to perform (like PVST) often has lower coverage.</li>



<li>Document exactly <strong>what your test does and doesn’t detect</strong>.</li>



<li>For new designs, <strong>select devices that are easier to proof test</strong>.</li>
</ul>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">Related Reading and Resources</h2>



<p><strong>Internal Links:</strong></p>



<ul class="wp-block-list">
<li>Glossary: <a href="https://silsafe.net/glossary/proof-test-coverage-cpt/" data-type="glossary" data-id="794">Cpt</a>, <a href="https://silsafe.net/glossary/proof-test/" data-type="glossary" data-id="788">Proof Test</a>, <a href="https://silsafe.net/glossary/proof-test-interval/" data-type="glossary" data-id="790">Proof Test Interval</a></li>



<li>Blog: Intro course on <a href="https://silsafe.net/functional-safety-for-the-process-industry/" data-type="post" data-id="6100">Functional Safety for the Process Industry</a></li>



<li>Blog: <a href="https://silsafe.net/proof-testing-of-sifs/" data-type="post" data-id="33">Understanding Proof Testing</a></li>



<li>Blog: Deep dive on <a href="https://silsafe.net/failure-rates-in-functional-safety/" data-type="post" data-id="4221">failure rate</a></li>
</ul>



<p><strong>External Resources:</strong></p>



<ul class="wp-block-list">
<li><a href="https://oreda.com/" target="_blank" rel="noopener">OREDA Failure Database</a></li>



<li>Exida&#8217;s Safety Equipment Reliability Handbook (<a href="https://www.exida.com/books/safety-equipment-reliability-handbook-4th-edition" target="_blank" rel="noopener">SERH</a>) &#8211; this does have proof test coverage</li>
</ul>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<h2 class="wp-block-heading">Q&amp;A</h2>



<p><strong>1. How do I figure out what Cpt to use in my SIFs?</strong><br>Start with the equipment documentation. If certified, use the FMEDA. If not, use judgment, external databases, and document everything.</p>



<p><strong>2. Can I assume 100% Cpt if I fully test a valve via a FVST?</strong><br>Not quite. While FVST gets close, it might miss failure modes like sticking during partial actuation or internal bypass.</p>



<p><strong>3. How is Cpt different from diagnostic coverage?</strong><br>It measures what your manually performed or manually-initiated <strong>test</strong> can catch. Diagnostic coverage measures what the device’s <strong>self-checks</strong> can catch.</p>



<p><strong>4. Does increasing test frequency help more than increasing Cpt?</strong><br>They both help — but increasing Cpt often gives more impact with fewer operational interruptions.</p>



<p><strong>5. What’s the best way to improve Cpt without changing the system?</strong><br>Upgrade your test method. Add leak testing, position feedback, or combine manual and automated routines.</p>



<p><strong>6:&nbsp;Does Cpt apply to safe failures as well?</strong><br>No. Cpt is defined only on dangerous undetected failure modes. Proof tests may reveal safe failures, but those affect spurious trip rate and availability, not the Cpt value used in PFDavg.</p>



<hr class="wp-block-separator has-alpha-channel-opacity"/>



<p>Want to go deeper? Check out our full article on <a href="https://silsafe.net/proof-testing-of-sifs/" data-type="post" data-id="33">Proof Testing</a> here or visit the <a href="https://silsafe.net/functional-safety-glossary/" data-type="page" data-id="391">glossary </a>for more functional safety terms.</p>



<!-- FAQPage Schema (JSON-LD) — Verbatim conversion -->
<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "How do I figure out what Cpt to use in my SIFs?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Start with the equipment documentation. If certified, use the FMEDA. If not, use judgment, external databases, and document everything."
      }
    },
    {
      "@type": "Question",
      "name": "Can I assume 100% Cpt if I fully test a valve via a FVST?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Not quite. While FVST gets close, it might miss failure modes like sticking during partial actuation or internal bypass."
      }
    },
    {
      "@type": "Question",
      "name": "How is Cpt different from diagnostic coverage?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "It measures what your manually performed or manually initiated test can catch. Diagnostic coverage measures what the device’s self-checks can catch."
      }
    },
    {
      "@type": "Question",
      "name": "Does increasing test frequency help more than increasing Cpt?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "They both help — but increasing Cpt often gives more impact with fewer operational interruptions."
      }
    },
    {
      "@type": "Question",
      "name": "What’s the best way to improve Cpt without changing the system?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Upgrade your test method. Add leak testing, position feedback, or combine manual and automated routines."
      }
    },
    {
      "@type": "Question",
      "name": "Does Cpt apply to safe failures as well?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "No. Cpt is defined only on dangerous undetected failure modes. Proof tests may reveal safe failures, but those affect spurious trip rate and availability, not the Cpt value used in PFDavg."
      }
    }
  ]
}
</script>
]]></content:encoded>
					
					<wfw:commentRss>https://silsafe.net/proof-test-coverage-cpt-to-improve-pfdavg/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>How MTTR Affects Your SIL Calculations: A Beginner’s Guide to Mean Time to Restore</title>
		<link>https://silsafe.net/mean-time-to-restore-mttr/</link>
					<comments>https://silsafe.net/mean-time-to-restore-mttr/#respond</comments>
		
		<dc:creator><![CDATA[mamerten]]></dc:creator>
		<pubDate>Fri, 09 May 2025 00:55:19 +0000</pubDate>
				<category><![CDATA[Math Related]]></category>
		<category><![CDATA[Advanced]]></category>
		<guid isPermaLink="false">https://silsafe.net/?p=1129</guid>

					<description><![CDATA[Learn how MTTR (Mean Time to Restore) fits into PFDavg and STR in IEC 61511 compliance. A clear, beginner-friendly guide for functional safety.]]></description>
										<content:encoded><![CDATA[
<p>When functional safety calculations are discussed, engineers often focus on things like hardware fault rates, test intervals, or architectural constraints. One input that is often misunderstood — or misused — is MTTR: <strong>Mean Time to Restore</strong>. While it sounds simple, this term carries specific meaning in the context of IEC 61511 and can significantly impact SIL verification and spurious trip rate (STR) evaluations.</p>



<p>This article is a beginner’s guide to what MTTR actually is, when it should (and should not) be used in probability of failure on demand (PFDavg) calculations, and how getting it wrong could lead to overconfident or unrealistic safety claims.</p>



<h2 class="wp-block-heading">What Is MTTR (Mean Time to Restore)?</h2>



<p>Per IEC 61511, <strong>MTTR</strong> refers to the time it takes to restore a safety instrumented function (SIF) to its proper operating state after a failure has occurred.&nbsp; Importantly — and often overlooked — this is <strong>different from</strong>&nbsp;mean repair time (MRT).</p>



<p><strong>This includes the following four components:</strong></p>



<ul class="wp-block-list">
<li><strong>(a)</strong> Time to detect the failure &#8211; meaning from the point of failure to when the diagnostic alert is triggered</li>



<li><strong>(b) </strong>Time spent before starting the repair.  Such as paperwork and ordering parts.  </li>



<li><strong>(c)</strong> Time to do the repair itself.</li>



<li><strong>(d)</strong> Time to perform the restoration</li>
</ul>



<p>So:</p>


<div class="wp-block-image">
<figure class="aligncenter size-full is-resized"><img loading="lazy" decoding="async" width="531" height="80" src="https://silsafe.net/wp-content/uploads/2025/06/MTTR-equation.webp" alt="Equation showing MTTR equals a plus b plus c plus d, representing Mean Time to Repair" class="wp-image-2485" style="width:229px;height:auto" srcset="https://silsafe.net/wp-content/uploads/2025/06/MTTR-equation.webp 531w, https://silsafe.net/wp-content/uploads/2025/06/MTTR-equation-300x45.webp 300w" sizes="auto, (max-width: 531px) 100vw, 531px" /></figure>
</div>


<p>In contrast, <strong>MRT (Mean Repair Time)</strong> is:</p>


<div class="wp-block-image">
<figure class="aligncenter size-full is-resized"><img loading="lazy" decoding="async" width="432" height="73" src="https://silsafe.net/wp-content/uploads/2025/06/MRT-equation.webp" alt="Equation showing MRT equals b plus c plus d, representing Mean Restoration Time" class="wp-image-2484" style="width:168px;height:auto" srcset="https://silsafe.net/wp-content/uploads/2025/06/MRT-equation.webp 432w, https://silsafe.net/wp-content/uploads/2025/06/MRT-equation-300x51.webp 300w" sizes="auto, (max-width: 432px) 100vw, 432px" /></figure>
</div>


<p>This distinction matters. MTTR as defined in IEC 61511 considers the full window of vulnerability — from initial failure to full recovery.&nbsp; Both MTTR and MRT are used in PFDavg calculations.</p>



<p>This is important because in safety calculations, you&#8217;re modeling how long a function is unavailable and how much risk accumulates during that time.</p>



<p><strong>Typical Ranges:</strong></p>



<ul class="wp-block-list">
<li><strong>Low-end </strong>(e.g., redundant sensor swap with on-site spares): 2–4 hours</li>



<li><strong>High-end </strong>(e.g., submerged valve with specialist access): weeks or even months</li>
</ul>



<h2 class="wp-block-heading">When MTTR Is (and Isn’t) Used in PFDavg</h2>



<p>Mean time to restore is NOT always used in PFDavg.&nbsp;</p>



<p><strong>Rule of thumb:</strong> It only affects PFDavg when a diagnostic failure is <strong>detected</strong>&nbsp;and that failure is reported only.&nbsp; Meaning, it doesn’t cause an immediate trip of the SIF, nor does it vote to trip.&nbsp; In essence it means the SIF is still online but <strong>functionally unavailable</strong>&nbsp;and has effectively already &#8220;failed.&#8221;</p>



<p><strong>Not used in PFDavg:</strong></p>



<p>The following are examples when PFDavg would not include MTTR components</p>



<ul class="wp-block-list">
<li>Example &#8211; A 2oo3 instrument SIF where one instrument has a DD error.&nbsp; That instrument has voted TRUE in the 2oo3 logic.&nbsp; Note this is a common scenario and is a typical way SIFs are designed.</li>



<li>Example &#8211; a 2oo3 instrument where one instrument has a DD error.&nbsp; The system is designed to trip the ENTIRE SIF (this is a rare scenario but could happen)</li>
</ul>



<p><strong>Incorrect usage example:</strong> Applying MTTR to a demand-mode final element that fails silently and has no on-board diagnostics.&nbsp; This underestimates PFDavg.</p>



<h2 class="wp-block-heading">MTTR Impact on PFDavg</h2>



<p>This concept can influence PFDavg calculations in some hard-to-understand ways.&nbsp; The first way is the classic method of DD failure which has the most impact.&nbsp; The second is included as it is very adjacent to MTTR, but is actually another term called mean repair time (MRT).</p>



<h3 class="wp-block-heading">1. DD Failures that are Reported only</h3>



<p>In this case, a component within the SIF experiences a <strong>dangerous detected failure (DD)</strong> that is reported but not acted upon with a trip nor vote-to-trip. The SIF remains active, but risk is accumulating because the failure has not yet been remedied. It is assumed the SIF is non-operational.</p>



<figure class="wp-block-image size-full is-resized"><img loading="lazy" decoding="async" width="512" height="71" src="https://silsafe.net/wp-content/uploads/2025/06/PFDavg-DD.webp" alt="Equation showing PFDavg equals lambda DD times MTTR" class="wp-image-2487" style="width:211px;height:auto" srcset="https://silsafe.net/wp-content/uploads/2025/06/PFDavg-DD.webp 512w, https://silsafe.net/wp-content/uploads/2025/06/PFDavg-DD-300x42.webp 300w" sizes="auto, (max-width: 512px) 100vw, 512px" /></figure>



<p><strong>Where:</strong></p>



<ul class="wp-block-list">
<li>λ<sub>DD</sub> = dangerous detected failure rate</li>



<li>MTTR = mean time to restore</li>
</ul>



<h3 class="wp-block-heading">2. Very Closely Related &#8211; SIF Bypass Periods (e.g., Proof Testing or Maintenance)</h3>



<p>When the entire SIF is placed in bypass — such as during <strong>proof testing</strong> or <strong>temporary overrides</strong> — MRT (not MTTR) contributes. The logic is that if a failure is discovered during testing, the SIF must be restored, so MRT becomes relevant.</p>


<div class="wp-block-image">
<figure class="aligncenter size-full is-resized"><img loading="lazy" decoding="async" width="673" height="130" src="https://silsafe.net/wp-content/uploads/2025/06/PFDavg-bypass-term.webp" alt="Equation showing PFDavg equals PTD over TI plus lambda DU times MRT" class="wp-image-2486" style="width:292px;height:auto" srcset="https://silsafe.net/wp-content/uploads/2025/06/PFDavg-bypass-term.webp 673w, https://silsafe.net/wp-content/uploads/2025/06/PFDavg-bypass-term-300x58.webp 300w" sizes="auto, (max-width: 673px) 100vw, 673px" /></figure>
</div>


<p><strong>Where:</strong></p>



<ul class="wp-block-list">
<li>PTD = proof test duration</li>



<li>TI = test interval</li>



<li>λ<sub>DU</sub> = dangerous undetected failure rate</li>



<li>MRT = mean repair time</li>
</ul>



<p>It can be confusing that λ<sub>DU</sub> is included as the failure was &#8220;detected&#8221; during proof testing.&nbsp; Reminder, the &#8220;detected&#8221; versus &#8220;undetected&#8221; applies to diagnostics.&nbsp; If a proof test finds a seized solenoid valve, that would fall under&nbsp;λ<sub>DU</sub>.&nbsp; It can also be confusing that MTTR and MRT can equal each other in times of fast diagnostics.</p>



<h2 class="wp-block-heading">How Much Does MTTR Contribute to PFDavg?</h2>



<p>Now that we&#8217;ve seen where Mean Time to Restore enters the PFDavg equations, the next logical question is: <em>how much of the total PFDavg is typically due to this value?</em></p>



<p>The answer depends on the architecture and diagnostic design of the SIF, but in many cases the MTTR-related terms can be a <strong>major portion</strong> of the overall PFDavg.</p>



<h3 class="wp-block-heading">Example Breakdown:</h3>



<p>Suppose a SIF has the following contributors:</p>



<ul class="wp-block-list">
<li>Dangerous undetected term (TI/2 component): <strong>6.00E-3</strong></li>



<li>Diagnostic-only (λ<sub>DD</sub> × MTTR): <strong>2.00E-3</strong></li>



<li>Bypass and restoration term: <strong>1.00E-3</strong></li>
</ul>



<p><strong>Total PFDavg = 9.00E-3</strong></p>



<p>In this example, <strong>MTTR-based contributions account for 3.00E-3</strong>, or <strong>33%</strong> of the total PFDavg.</p>



<h3 class="wp-block-heading">Sensitivity to MTTR Changes</h3>



<p>If the time the SIF is unavailable increases from 8 hours to 72 hours or beyond — as can easily happen with inaccessible equipment — that 33% could grow to <strong>50% or more</strong> of the total risk profile.</p>



<p>This reinforces a key takeaway:</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p><strong>It isn&#8217;t just a footnote — it&#8217;s a meaningful design and reliability driver.</strong></p>
</blockquote>



<p>When conducting SIL verification, always explore how much of your calculated PFDavg comes from MTTR-related sources. If it’s significant, you may want to revisit response logistics, spares availability, or automated recovery strategies.&nbsp; Designing a SIS is a balance of many things.</p>



<h2 class="wp-block-heading">Impact on Spurious Trip Rate (STR)</h2>



<p>A longer MTTR increases the <strong>exposure window</strong> for nuisance or spurious trips. Consider a logic solver that falsely detects a dangerous condition. If the system cannot be restored quickly, operations may suffer unnecessary downtime or escalation.</p>



<p>Designs that balance fault tolerance, diagnostic alerts, and <strong>realistic restore times</strong> can minimize the impact of STR on both safety and availability.&nbsp; Different facilities will have completely different tolerances of STR.&nbsp; Perhaps one per year is acceptable, perhaps not.</p>



<h2 class="wp-block-heading">What IEC 61511 Says</h2>



<p>IEC 61511 emphasizes the need to use <strong>realistic</strong> — not idealized — values in SIL verification.</p>



<blockquote class="wp-block-quote is-layout-flow wp-block-quote-is-layout-flow">
<p>“The MTTR shall take into account all delays including diagnosis, personnel response, spares availability, and repair.” <em>(paraphrased for clarity)</em></p>
</blockquote>



<p><strong>Be cautious of defaulting to 8 hours.</strong> Unless supported by site history, vendor specs, or service agreements, this assumption could invalidate a SIL claim.</p>



<h2 class="wp-block-heading">Common Mistakes </h2>



<ul class="wp-block-list">
<li>Confusing MTTR with MRT</li>



<li><strong>Using overly optimistic values</strong> — e.g., assuming parts, tools, and staff are instantly available</li>



<li><strong>Ignoring logistics</strong> — A poorly managed work order system may introduce delays of several days</li>



<li><strong>Assuming the same value for all elements</strong> — A sensor may be quick to restore, but a final element might require complex isolation and drainage</li>
</ul>



<h2 class="wp-block-heading">Best Practices for MTTR in SIL Verification</h2>



<ul class="wp-block-list">
<li><strong>Document separately </strong>for each device type (sensor, logic solver, final element)</li>



<li><strong>Source your data</strong>: field data, OEM specs, or reliability databases (OREDA, etc.)</li>



<li><strong>Challenge vendor-provided </strong>values if they seem unrealistic</li>



<li><strong>Responsibility</strong>: SIS designers typically own the MTTR assumption, but input from operations and maintenance is essential</li>



<li><strong>Record all assumptions</strong> in the Safety Requirements Specification (SRS) and in SIL verification reports</li>
</ul>



<h2 class="wp-block-heading">Q&amp;A</h2>



<p><strong>1. When does mean time to restore impact PFDavg?</strong><br>Only when a component of a SIF has detected a failure and reported it only.&nbsp; Thus, the SIF is unavailable but not tripped.</p>



<p><strong>2. How does MTTR differ from MRT?</strong><br>The former includes detection time (a); MRT starts with diagnosis (b). MTTR is longer and more conservative.</p>



<p><strong>3. Do these values affect high-demand SIFs?</strong><br>Generally no — high-demand systems use different metrics like PFH (probability of failure per hour).</p>



<p><strong>4. Should MTTR differ by component?</strong><br>Yes. Sensors, logic solvers, and final elements can have drastically different restore times.</p>



<p><strong>5. Can MTTR be impacted by facility bureaucracy or procurement delays?</strong><br>Absolutely. Delays in approvals, permits, or spare part procurement can dramatically increase the time a SIF is unavailable.&nbsp; These should be discussed and accounted for in determining these values.&nbsp; It could be that the architecture needs to change if the facility is unwilling to have working spares.</p>



<h2 class="wp-block-heading">Conclusion</h2>



<p>Mean time to restore might seem like a small parameter, but it has outsized influence on how we evaluate and justify risk reduction in a functional safety system. Whether you&#8217;re calculating PFDavg or trying to keep STR under control, using accurate, documented, and conservative MTTR values helps ensure your SIL claims are realistic — and defensible.</p>



<p>Audit your assumptions. Align with IEC 61511. And don&#8217;t let this &#8220;minor&#8221; input undermine your entire safety case.</p>



<p>Limble has a <a href="https://limble.com/learn/metrics/failure-metrics/" target="_blank" rel="noopener">great article</a> discussing MTTR and comparing it with other similar concepts.</p>



<!-- FAQPage Schema (JSON-LD) — paste into a WordPress “Custom HTML” block -->
<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "When does mean time to restore impact PFDavg?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Only when a component of a SIF has detected a failure and reported it. In this case, the SIF is unavailable but not tripped."
      }
    },
    {
      "@type": "Question",
      "name": "How does MTTR differ from MRT?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "MTTR includes detection time (a), while MRT starts with diagnosis (b). MTTR is therefore longer and more conservative."
      }
    },
    {
      "@type": "Question",
      "name": "Do these values affect high-demand SIFs?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Generally no — high-demand systems use different metrics such as PFH (probability of failure per hour)."
      }
    },
    {
      "@type": "Question",
      "name": "Should MTTR differ by component?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Yes. Sensors, logic solvers, and final elements can have drastically different restore times."
      }
    },
    {
      "@type": "Question",
      "name": "Can MTTR be impacted by facility bureaucracy or procurement delays?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Absolutely. Delays in approvals, permits, or spare part procurement can dramatically increase the time a SIF is unavailable. These should be discussed and accounted for when determining MTTR. In some cases, the architecture may need to change if the facility is unwilling to maintain working spares."
      }
    }
  ]
}
</script>

]]></content:encoded>
					
					<wfw:commentRss>https://silsafe.net/mean-time-to-restore-mttr/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>PFDavg Explained: 6 Essentials for Getting Started with SIL Calculations</title>
		<link>https://silsafe.net/pfdavg-explained/</link>
					<comments>https://silsafe.net/pfdavg-explained/#respond</comments>
		
		<dc:creator><![CDATA[mamerten]]></dc:creator>
		<pubDate>Sat, 26 Apr 2025 19:50:21 +0000</pubDate>
				<category><![CDATA[Beginner]]></category>
		<category><![CDATA[Math Related]]></category>
		<category><![CDATA[PFDavg]]></category>
		<category><![CDATA[proof testing]]></category>
		<category><![CDATA[SIFs]]></category>
		<guid isPermaLink="false">https://silsafe.net/?p=184</guid>

					<description><![CDATA[This beginner-friendly guide explains what PFDavg (Probability of Failure on Demand) means in functional safety, how it's calculated, and why it matters under IEC 61511. With clear examples, key definitions, and a breakdown of the safety loop, this article is your starting point for designing effective Safety Instrumented Functions (SIFs).]]></description>
										<content:encoded><![CDATA[
<p>Understanding <strong>PFDavg</strong>—short for <em>Probability of Failure on Demand, average</em>—is foundational if you&#8217;re new to functional safety and the IEC 61511 framework. Whether you&#8217;re supporting a Safety Instrumented System (SIS) in oil &amp; gas, biogas, chemicals, or any other process industry, developing a solid understanding of this metric helps you design safer systems and comply with regulatory expectations.</p>



<p>This guide walks you through six essential concepts about PFDavg that every engineer should understand when evaluating or designing Safety Instrumented Functions (SIFs).</p>



<h2 class="wp-block-heading">1. What Is PFDavg and Why It Matters</h2>



<p>A Safety Instrumented Function (SIF) is a specific set of equipment—usually a sensor, logic solver, and final element—designed to take a process to a safe state when a defined hazardous condition is detected.</p>



<p>PFDavg is the average likelihood that a Safety Instrumented Function (SIF) will fail <em>when</em> it&#8217;s needed. It quantifies the chance that your safeguard won&#8217;t respond in a dangerous situation—basically, the &#8220;gap&#8221; in protection.</p>



<p>Most SIS applications operate in <strong>low-demand mode</strong>, meaning they’re called upon infrequently (e.g., fewer than once per year). For these systems, PFDavg is the go-to metric for quantifying performance.</p>



<p>Why does it matter? Because PFDavg is directly tied to your Safety Integrity Level (SIL) target. If your value is too high, your SIF doesn’t meet the required SIL—and that means your process risk isn&#8217;t sufficiently reduced.</p>



<p>External resource: <a href="https://www.isa.org/standards-and-publications/isa-standards/isa-standards-committees/isa84" target="_blank" rel="noopener">ISA Functional Safety</a></p>



<h2 class="wp-block-heading">2. How PFDavg Determines Safety Integrity Level (SIL)</h2>



<p>The <strong>Safety Integrity Level (SIL)</strong> is a measure of how much risk reduction a SIF provides. It’s defined by using thresholds.&nbsp; The following table summarizes the P ranges and associated risk reduction factor (RRF) for each SIL level:</p>



<figure class="wp-block-table"><table class="has-fixed-layout"><thead><tr><th>SIL Level</th><th>PFDavg Range</th><th>RRF Range</th></tr></thead><tbody><tr><td>SIL 1</td><td>≥1.0E-2 to &lt;1.0E-1</td><td>10 to &lt;100</td></tr><tr><td>SIL 2</td><td>≥1.0E-3 to &lt;1.0E-2</td><td>100 to &lt;1,000</td></tr><tr><td>SIL 3</td><td>≥1.0E-4 to &lt;1.0E-3</td><td>1,000 to &lt;10,000</td></tr><tr><td>SIL 4</td><td>≥1.0E-5 to &lt;1.0E-4</td><td>10,000 to &lt;100,000</td></tr></tbody></table></figure>



<p>In the process industry, you’ll typically see SIL 1 to SIL 3 used. SIL 4 is rare and generally reserved for nuclear or aerospace applications.</p>



<p>External reference: <a href="https://www.iec.ch/functionalsafety/" target="_blank" rel="noopener">IEC Functional Safety</a></p>



<h2 class="wp-block-heading">3. The Core PFDavg Equation</h2>



<p>For low-demand systems, a simplified PFDavg formula looks like this:</p>



<figure class="wp-block-image size-full"><img loading="lazy" decoding="async" width="212" height="58" src="https://silsafe.net/wp-content/uploads/2025/04/pfdavg-basic-equation.png" alt="PFDavg basic equation" class="wp-image-188"/></figure>



<p>Where:</p>



<ul class="wp-block-list">
<li>λdu: Dangerous undetected failure rate (failures per hour)</li>



<li>TI: Proof test interval (hours)</li>
</ul>



<p>This equation assumes perfect testing, no redundancy, and no diagnostics. It&#8217;s a good starting point but real-world systems require more advanced modeling.</p>



<p>More complete equations may also include:</p>



<ul class="wp-block-list">
<li><strong>Proof Test Coverage (Cpt)</strong></li>



<li><strong>Mean Time to Restore (MTTR)</strong></li>



<li><strong>Common Cause Failure, beta factor (β)</strong></li>
</ul>



<p>These concepts—<em>Cpt</em>, <em>MTTR</em>, and <em>β</em>—will be covered in future posts.</p>



<h2 class="wp-block-heading">4. Key Terms You Need to Know</h2>



<h3 class="wp-block-heading">Proof Test Interval (TI)</h3>



<p>How often the system is tested to reveal hidden failures. A longer TI increases PFDavg because failures remain latent for longer.</p>



<h3 class="wp-block-heading">Mean Time to Restore (MTTR)</h3>



<p>The average time it takes to restore the system to a working state once a failure is discovered. This influences overall system unavailability. <em>(Future article topic)</em></p>



<h3 class="wp-block-heading">Proof Test Coverage (Cpt)</h3>



<p>Represents the fraction of dangerous undetected failures that are revealed by a proof test. The higher the Cpt, the more effective the test, and the lower your average probability of failure on demand. This is especially critical in systems that lack built-in diagnostics. <em>(Future article topic)</em></p>



<h3 class="wp-block-heading">Common Cause Failure</h3>



<p>Common cause failure occurs when two or more components that are supposed to provide redundancy fail simultaneously due to a shared cause. These causes might include shared environmental factors (like temperature or humidity), a common power supply, or even human error.&nbsp; In SIL verification calculations, the beta factor (β) represents the portion of failure that cannot be assumed to be independent. Properly accounting for β is critical in ensuring that your redundant architecture isn&#8217;t giving a false sense of risk reduction. <em>(Future article topic)</em></p>



<h2 class="wp-block-heading">5. A Simple Example</h2>



<p>Let’s say you have a single-channel SIF configured in a 1oo1 architecture, using a sensor that has a dangerous undetected failure rate of 1E-6 per hour. The system is proof tested every year (8760 hours).</p>



<figure class="wp-block-image size-full"><img loading="lazy" decoding="async" width="314" height="42" src="https://silsafe.net/wp-content/uploads/2025/04/pfdavg-calculation-example.png" alt="Example calculation of PFDavg (PFD Average)" class="wp-image-190" srcset="https://silsafe.net/wp-content/uploads/2025/04/pfdavg-calculation-example.png 314w, https://silsafe.net/wp-content/uploads/2025/04/pfdavg-calculation-example-300x40.png 300w" sizes="auto, (max-width: 314px) 100vw, 314px" /></figure>



<p>This value falls within the SIL 2 range. This corresponds to a Risk Reduction Factor (RRF) of 228.</p>



<p>Now, if you want to increase this to SIL 3, you will need to&#8230;</p>



<ul class="wp-block-list">
<li>Reduce the failure rate</li>



<li>Reduce the TI</li>



<li>Add redundancy by adjusting the architecture (which would use another equation)</li>



<li>Increase diagnostic coverage</li>
</ul>



<p>Note this formula is only valid for 1oo1 and a simple approach of doing these calculations.</p>



<h2 class="wp-block-heading">6. How Design Choices Affect PFDavg</h2>



<p>Designers can manipulate several variables to drive the value down:</p>



<ul class="wp-block-list">
<li><strong>Shorter TI</strong>: More frequent testing catches failures earlier.</li>



<li><strong>Redundant architecture</strong>: 1oo2 or 2oo3 configurations use different equations.</li>



<li><strong>Improved diagnostics</strong>: Increases Cpt, reducing undetected failure exposure.</li>



<li><strong>Better components</strong>: Lower intrinsic failure rates (λD).</li>
</ul>



<p>Each choice comes with trade-offs in cost, complexity, and operational downtime. The goal is to achieve <em>just enough</em> safety—not overdesign.</p>



<h2 class="wp-block-heading">In Summary</h2>



<p>PFDavg is a key performance metric in functional safety for low-demand SIFs. It links directly to SIL and guides the engineering design and validation process. Understanding the basics of how it’s calculated—and what affects it—helps ensure you’re designing systems that meet the risk reduction targets set by IEC 61511.</p>



<p>Future articles will dive deeper into Cpt, MTTR, and β, each of which play roles in real-world SIL verification.</p>



<h2 class="wp-block-heading">Quick Q&amp;A</h2>



<p><strong>What is the difference between PFDavg and PFH?</strong><br>PFDavg applies to low-demand systems; PFH (Probability of Failure per Hour) applies to high/continuous-demand systems.</p>



<p><strong>Can PFDavg be too low?</strong><br>Yes. Overdesign increases cost and complexity without necessarily increasing safety proportionally.</p>



<p><strong>Does IEC 61511 require PFDavg calculations?</strong><br>Yes—for each SIF, PFDavg (or PFH) must be calculated to demonstrate compliance with the required SIL.</p>



<p><strong>Is there a standard way to calculate it?</strong><br>Yes and no. IEC 61508 and IEC 61511 provide methods and allow both simplified and detailed probabilistic approaches. But there are different levels of equations the engineer can choose to get into.</p>



<p><strong>What determines what the PFDavg needs to be?</strong><br>The required value is dictated by your hazard and risk assessment (H&amp;RA)—typically performed through a Layer of Protection Analysis (LOPA). This is what tells you how much risk reduction is needed, which in turn defines your SIL target.</p>



<p><strong>What role does proof testing play?</strong><br>It reveals hidden failures. The longer the proof test interval (TI), the higher the PFDavg.</p>



<p><strong>Does the math formula change with different voting logic (e.g., 1oo2, 2oo3)?</strong><br>Absolutely. The basic formula you&#8217;ve seen assumes a 1oo1 architecture. Different configurations use different math, and these differences can significantly change the resulting calculation.</p>



<!-- FAQPage Schema (JSON-LD) — VERBATIM, EXACT REPLICATION -->
<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "What is the difference between PFDavg and PFH?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "PFDavg applies to low-demand systems; PFH (Probability of Failure per Hour) applies to high/continuous-demand systems."
      }
    },
    {
      "@type": "Question",
      "name": "Can PFDavg be too low?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Yes. Overdesign increases cost and complexity without necessarily increasing safety proportionally."
      }
    },
    {
      "@type": "Question",
      "name": "Does IEC 61511 require PFDavg calculations?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Yes—for each SIF, PFDavg (or PFH) must be calculated to demonstrate compliance with the required SIL."
      }
    },
    {
      "@type": "Question",
      "name": "Is there a standard way to calculate it?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Yes and no. IEC 61508 and IEC 61511 provide methods and allow both simplified and detailed probabilistic approaches. But there are different levels of equations the engineer can choose to get into."
      }
    },
    {
      "@type": "Question",
      "name": "What determines what the PFDavg needs to be?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "The required value is dictated by your hazard and risk assessment (H&RA)—typically performed through a Layer of Protection Analysis (LOPA). This is what tells you how much risk reduction is needed, which in turn defines your SIL target."
      }
    },
    {
      "@type": "Question",
      "name": "What role does proof testing play?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "It reveals hidden failures. The longer the proof test interval (TI), the higher the PFDavg."
      }
    },
    {
      "@type": "Question",
      "name": "Does the math formula change with different voting logic (e.g., 1oo2, 2oo3)?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Absolutely. The basic formula you’ve seen assumes a 1oo1 architecture. Different configurations use different math, and these differences can significantly change the resulting calculation."
      }
    }
  ]
}
</script>

]]></content:encoded>
					
					<wfw:commentRss>https://silsafe.net/pfdavg-explained/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Proof Testing of SIFs: Understanding Its 3 Purposes and Its Importance</title>
		<link>https://silsafe.net/proof-testing-of-sifs/</link>
					<comments>https://silsafe.net/proof-testing-of-sifs/#respond</comments>
		
		<dc:creator><![CDATA[mamerten]]></dc:creator>
		<pubDate>Sat, 15 Mar 2025 19:02:46 +0000</pubDate>
				<category><![CDATA[Beginner]]></category>
		<category><![CDATA[Math Related]]></category>
		<category><![CDATA[PFDavg]]></category>
		<category><![CDATA[proof testing]]></category>
		<guid isPermaLink="false">https://silsafe.net/?p=33</guid>

					<description><![CDATA[Proof testing ensures your Safety Instrumented Functions (SIFs) work when needed by uncovering hidden faults. Learn why proof tests are crucial for lowering PFDavg and maintaining Safety Integrity Level (SIL) targets under IEC 61511.]]></description>
										<content:encoded><![CDATA[
<p>If you&#8217;re new to functional safety, you&#8217;ve likely encountered terms like proof testing, Safety Instrumented Functions (SIF), and Probability of Failure on Demand (PFDavg). Understanding these concepts is crucial for ensuring the reliability of safety systems in process industries. In this article, we&#8217;ll clearly explain <strong>proof testing</strong>, why it&#8217;s essential, how it impacts the PFDavg calculation, and its relationship with Safety Integrity Level (SIL).</p>



<h2 class="wp-block-heading">What Exactly is Proof Testing?</h2>



<p>Proof testing refers to manually performed or manually initiated periodic tests conducted on Safety Instrumented Systems (SIS) to reveal hidden or dormant faults. The main goal is ensuring each SIF—which consists of sensors, logic solvers, and final elements—can reliably activate when needed. These tests simulate conditions that activate the safety system, confirming functionality and identifying issues that could compromise performance during an actual emergency.</p>



<p>IEC 61511, the international standard guiding functional safety in the process industry, explicitly mandates proof tests to maintain compliance and assure safety integrity.</p>



<p>Proof testing is not just a regulatory checkbox—it&#8217;s a central activity in any process safety lifecycle. When done properly, it enhances equipment reliability, informs maintenance strategies, and helps avoid unnecessary spurious trips or overlooked latent failures. These tests are especially important in facilities with aging infrastructure or complex safety systems, where regular inspections can be challenging.</p>



<h2 class="wp-block-heading">Why is it Essential?</h2>



<p>Regular proof testing helps:</p>



<ul class="wp-block-list">
<li><strong>Ensure Reliability:</strong> Detects hidden faults, increasing confidence that safety systems respond correctly when required.</li>



<li><strong>Achieve and Maintain SIL Ratings:</strong> Directly influences SIL by lowering PFDavg values, ensuring a system meets its designated safety objectives.</li>
</ul>



<h2 class="wp-block-heading">Connecting to PFDavg</h2>



<p>PFDavg is a metric used to gauge the likelihood that a safety system will fail to respond correctly upon demand. Frequent and thorough testing significantly reduces the PFDavg by detecting hidden failures and correcting them promptly.</p>



<h3 class="wp-block-heading">Understanding PFDavg Through a Simple Calculation:</h3>



<p>A simplified formula for calculating PFDavg is:</p>



<figure class="wp-block-image size-full is-resized"><img loading="lazy" decoding="async" width="212" height="58" src="https://silsafe.net/wp-content/uploads/2025/04/pfdavg-basic-equation.png" alt="PFDavg basic equation showing impact of proof testing interval (TI)" class="wp-image-188" style="width:216px;height:auto"/></figure>



<ul class="wp-block-list">
<li>λ<sub>DU</sub> is the dangerous undetected failure rate, often in 1/hour</li>



<li><em>TI</em> s the test interval between proof tests, often in years</li>
</ul>



<p>Shorter test intervals result in a lower PFDavg, directly supporting higher SIL ratings.</p>



<h3 class="wp-block-heading">Practical Example:</h3>



<p>Imagine a SIF component that has 0.002 failures per year. With annual testing, your calculation would be:</p>



<figure class="wp-block-image"><img loading="lazy" decoding="async" width="223" height="41" src="https://silsafe.net/wp-content/uploads/2025/03/pfdavg-with-proof-test.png" alt="PFDavg example calculation showing proof test" class="wp-image-181"/></figure>



<p>This PFDavg of 0.001 complies with SIL 2 requirements (ranging from 0.001 to 0.01). Extending test intervals increases PFDavg, reducing safety system reliability and potentially affecting the SIL rating.</p>



<h2 class="wp-block-heading">Types of Tests and Operational Impact</h2>



<p>Different types of tests have distinct impacts on plant operations:</p>



<h3 class="wp-block-heading">Full Functional Test</h3>



<p>A complete system test, involving sensors, logic solvers, and actuators.</p>



<ul class="wp-block-list">
<li><strong>Operational Impact:</strong> Usually requires full shutdown.</li>



<li><strong>Advantages:</strong> Highest accuracy, confirms full loop functionality.</li>



<li><strong>Drawbacks:</strong> Production downtime and increased costs.</li>
</ul>



<h3 class="wp-block-heading">Partial-Stroke Test</h3>



<p>Typically used for valves, it partially moves the valve to test performance without shutting down the process.</p>



<ul class="wp-block-list">
<li><strong>Operational Impact:</strong> Minimal interruption, production continues.</li>



<li><strong>Advantages:</strong> Frequent monitoring, limited disruption.</li>



<li><strong>Drawbacks:</strong> Less comprehensive, may miss certain faults.</li>
</ul>



<h3 class="wp-block-heading">Diagnostic Testing &#8211; is Great &#8211; but is not Proof Testing</h3>



<p>Automated tests performed by components to detect faults continuously or at short intervals is a common feature in Functional Safety.  However, these are NOT considered proof tests.  Remember, proof test are manually performed or manually initiated tests.  </p>



<p>Automatic diagnostic testing is an amazing way to help ensure safety, but is a distinct concept and has distinct math terms from proof testing. Some notes on diagnostics are&#8230;</p>



<ul class="wp-block-list">
<li><strong>Operational Impact:</strong> No direct operational impact.</li>



<li><strong>Advantages:</strong> Early fault detection, reduced manual testing frequency.</li>



<li><strong>Drawbacks:</strong> Can miss faults undetectable by automated methods.</li>
</ul>



<h2 class="wp-block-heading">Balancing Test Frequency and Practicality</h2>



<h3 class="wp-block-heading">Typical Test Intervals (TI) for SIFs</h3>



<p>The appropriate Test Interval (TI) for a Safety Instrumented Function (SIF) depends on many factors, including the required Safety Integrity Level (SIL), failure rate of components, Proof Test Coverage (Cpt), and the risk reduction targets.</p>



<p>In general practice:</p>



<ul class="wp-block-list">
<li><strong>SIL 1 SIFs</strong> often have TIs ranging from 1 to 5 years</li>



<li><strong>SIL 2 SIFs</strong> typically fall within a 1 to 2 year TI</li>



<li><strong>SIL 3 SIFs</strong> usually require TIs of 6 months to 1 year or even more frequent</li>
</ul>



<p>These ranges are not hard-rules and must be validated by PFDavg calculations, which incorporate actual device data and proof test protocol effectiveness. For high-demand or continuous mode operations, the calculation basis and interval definitions change, often requiring more sophisticated modeling.</p>



<p>Determining testing frequency involves balancing safety, reliability, and operational practicality. More frequent testing reduces PFDavg but increases operational costs and downtime, while less frequent testing can increase risks due to undetected faults. The key is aligning testing frequency with safety goals and operational realities.</p>



<h3 class="wp-block-heading">Proof Test Coverage (Cpt)</h3>



<p>A related topic, Proof Test Coverage (Cpt), refers to the effectiveness of proof tests in detecting hidden faults. Or more specifically, dangerous undetected (DU) failures. While this article focuses on the broader concept of proof testing, it&#8217;s worth noting that Cpt plays a significant role in how PFDavg is calculated. In essence, if a test is only capable of detecting a fraction (Cpt ) of possible dangerous undetected failures, that limited effectiveness must be factored into the safety equations. This complicates the math and emphasizes the importance of having well-defined proof test protocols that clearly state what is and isn&#8217;t being tested.</p>



<p>To learn more about Cpt, see <a href="https://silsafe.net/proof-test-coverage-cpt-to-improve-pfdavg/" data-type="post" data-id="2784">this article</a> to get into more details.</p>



<h3 class="wp-block-heading">Balancing Testing with Operational Requirements</h3>



<p>While proof testing is essential for maintaining the integrity of SIFs, it often competes with operational demands. Some tests, especially full functional tests, require shutting down part or all of a process—a decision that carries significant production and cost implications. In other cases, testing may require bypassing safety functions, temporarily reducing the facility’s protection layers.</p>



<p>Therefore, organizations must carefully weigh the benefits of frequent and thorough testing against the impact on production schedules, safety availability, and maintenance workload. This balance often involves coordination across operations, safety, and engineering teams to align test intervals with planned outages, maintenance windows, and risk tolerance levels.</p>



<h2 class="wp-block-heading">Common Testing Pitfalls</h2>



<p>While not exhaustive, here are some common pitfalls organizations encounter:</p>



<ul class="wp-block-list">
<li>Poor planning leading to unnecessary operational disruptions.</li>



<li>Overreliance on partial-stroke and diagnostic testing without periodic full functional tests.</li>
</ul>



<p>Avoiding these issues enhances both safety and efficiency.</p>



<h2 class="wp-block-heading">Wrap-Up</h2>



<p>Proof testing of SIFs is fundamental in maintaining the reliability and effectiveness of safety systems. Regular and thorough tests directly influence PFDavg calculations, help maintain SIL ratings, and ensure compliance with IEC 61511. By clearly understanding and implementing these practices, your organization significantly enhances safety, reduces risks, and sustains operational efficiency.</p>



<h2 class="wp-block-heading">Quick Q&amp;A:</h2>



<p><strong>Q1:</strong> What is the primary purpose of a proof test protocol?<br><strong>A:</strong> To detect hidden failures within the SIF, ensuring reliability and readiness. Or more specifically, dangerous undetected (DU) failures.</p>



<p><strong>Q2:</strong> How does testing frequency influence functional safety?<br><strong>A:</strong> More frequent testing reduces the PFDavg, directly improving the SIL rating. </p>



<p><strong>Q3:</strong> Can partial-stroke testing replace full functional testing?<br><strong>A:</strong> No, partial-stroke testing complements full tests but can&#8217;t entirely replace them, as it&#8217;s not comprehensive enough to detect all types of faults.</p>



<p><strong>Q4:</strong> Is diagnostic testing (done automatically by newer components) a type of proof testing.<br><strong>A:</strong> No, diagnostic testing and proof testing are different concepts in the functional safety ecosystem. Proof testing is always operator-performed or operator-initiated. Both concepts are in PFDavg calculations and the math works different ways.</p>



<p><strong>Q5:</strong>&nbsp;If my SIF already uses devices with high diagnostic coverage, do I still need frequent proof testing?<br><strong>A:</strong>&nbsp;Yes. But perhaps with very strong diagnostics you can use a longer TI and still achieve the needed PFDavg.</p>



<h2 class="wp-block-heading">Ready to Learn More?</h2>



<p>Stay informed and up-to-date on functional safety and industry best practices by subscribing to our newsletter. You&#8217;ll receive the latest insights directly in your inbox to help improve your safety management practices.</p>



<h2 class="wp-block-heading">Additional Resources:</h2>



<ul class="wp-block-list">
<li><a href="https://webstore.iec.ch/en/publication/24241" target="_blank" rel="noopener">IEC 61511 Official Standard</a></li>



<li><a href="https://www.isa.org" target="_blank" rel="noopener">International Society of Automation (ISA)</a></li>



<li><a href="https://www.hse.gov.uk" target="_blank" rel="noopener">UK Health and Safety Executive (HSE)</a></li>



<li><a href="https://www.aiche.org/ccps" target="_blank" rel="noopener">CCPS Guidelines</a></li>



<li>Internal <a href="https://silsafe.net/proof-test-coverage-cpt-to-improve-pfdavg/" data-type="post" data-id="2784">blog post</a> on Cpt</li>



<li>See our <a href="https://silsafe.net/functional-safety-glossary/" data-type="page" data-id="391">glossary </a>for loads more terms</li>



<li>Internal <a href="https://silsafe.net/failure-rates-in-functional-safety/" data-type="post" data-id="4221">article </a>on failure rate</li>



<li>Internal <a href="https://silsafe.net/functional-safety-for-the-process-industry/" data-type="post" data-id="6100">article </a>on the basics of functional safety</li>
</ul>



<!-- FAQPage Schema (JSON-LD) — VERBATIM, EXACT REPLICATION -->
<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "Q1: What is the primary purpose of a proof test protocol?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "A: To detect hidden failures within the SIF, ensuring reliability and readiness. Or more specifically, dangerous undetected (DU) failures."
      }
    },
    {
      "@type": "Question",
      "name": "Q2: How does testing frequency influence functional safety?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "A: More frequent testing reduces the PFDavg, directly improving the SIL rating."
      }
    },
    {
      "@type": "Question",
      "name": "Q3: Can partial-stroke testing replace full functional testing?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "A: No, partial-stroke testing complements full tests but can't entirely replace them, as it's not comprehensive enough to detect all types of faults."
      }
    },
    {
      "@type": "Question",
      "name": "Q4: Is diagnostic testing (done automatically by newer components) a type of proof testing.",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "A: No, diagnostic testing and proof testing are different concepts in the functional safety ecosystem. Proof testing is always operator-performed or operator-initiated. Both concepts are in PFDavg calculations and the math works different ways."
      }
    },
    {
      "@type": "Question",
      "name": "Q5: If my SIF already uses devices with high diagnostic coverage, do I still need frequent proof testing?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "A: Yes. But perhaps with very strong diagnostics you can use a longer TI and still achieve the needed PFDavg."
      }
    }
  ]
}
</script>
]]></content:encoded>
					
					<wfw:commentRss>https://silsafe.net/proof-testing-of-sifs/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>

<!--
Performance optimized by W3 Total Cache. Learn more: https://www.boldgrid.com/w3-total-cache/?utm_source=w3tc&utm_medium=footer_comment&utm_campaign=free_plugin

Page Caching using Disk: Enhanced 
Minified using Disk

Served from: silsafe.net @ 2026-04-24 10:27:20 by W3 Total Cache
-->