PSC Myths vs. Reality

The 5 most common PSC myths that are not supported by PSC Data

Click a myth to reveal what the data actually says.
7
Primary PSC KPIs (DPI/DER) for CY 2025

MYTH #1

Detained Ships Are Problematic

THE REALITY

Detained Ships Are Just Unprepared

MYTH #2

MoU Matters

THE REALITY

Ship-Port Tier (SPOT) Matters

MYTH #3

PSCI Priority Matters

THE REALITY

PSCI Risk Level (PRL) & Ship- Port Tier (SPOT) Matters

MYTH #4

A Good Checklist Is Enough

THE REALITY

Effective PSC Checklist Must Be Ship & Port Adaptive, Prioritized

MYTH #5

I Know My Fleet Performance

THE REALITY

PSC Performance Monitoring Requires Port Type-Specific KPI Values & Benchmark

MYTH #1

Myth vs Reality: Detained Ships Are Not Problematic – They Are Unprepared

  • The Myth

“Detained ships are problematic, old, or poorly managed. We don’t have ‘problem ships’ in our fleet. That won’t happen to us – we run a tight ship, we are a blue chip company.”

This is one of the most persistent myths in shipping. Detentions are often seen as a red flag for systemic failure – poor maintenance, weak management, or incompetent crew.

The underlying belief is: “That won’t happen to us – we run a tight ship.”

But the data tells a different story.

  • Normalcy Bias - “It Won’t Happen to Us”

This bias explains the belief behind Myth #1 that detentions mainly happen to “problem ships.” In reality, PSC detentions are statistically normal across the global fleet. Operators underestimate their own exposure and assume their vessels are less likely to be targeted, creating dangerous complacency in preparation.

  • The Reality

REALITY #1

Detentions are widespread and non-discriminatory

Across Port State Control regimes globally, approximately 7,000 detentions occur every 36 months, involving around 5,700 unique vessels.

This is not a marginal phenomenon. With more than 1 in 7 ocean-going commercial ships projected to face detention at least once in the coming 36 months, detentions are statistically normal – not an exception.

This includes ships of all types, ages, and management standards.

Ships will face detention at least once in 36 months

REALITY #2

Most detained ships have no previous detention record

More than 80% of ships detained within a 36-month window have only one detention in that period.

They are not repeat offenders. They are vessels with otherwise clean records – simply caught unprepared at the wrong time.

The idea of the “problem ship” is largely a myth.

The truth is: one-off detentions are the rule, not the exception.

%

Ships detained within a 36-month window have only one detention

REALITY #3

SPOT analysis proves no Manager is immune

RISK4SEA’s SPOT Matrix (Ship-Port Tier) reveals that detentions occur across all management profiles, including Top-Tier managers.

Data shows that even highly ranked operators – with low historic DPI or DER – occasionally get caught by a focused PSCI.

This is not about operator “quality” – it’s about port-specific preparation failures.

SPOT analysis, similarly to Pareto principle, proves that we may identify well in advance the 15% of the PSCIs that create the 85% of the detentions.

%

of the PSCIs create the 85% of the detentions

REALITY #4

PRL research exposes the unpredictability of PSCI exposure

The POCRA Risk Level (PRL) research proves that detentions often occur when vessels are not even “eligible” under formal PSCI targeting logic.

  • 1 in 5 inspections happens outside the official open window.

  • 1 in 4 detentions hits ships with closed PRL windows.

That means a vessel can be off the radar – but still detained. Eligibility offers no protection if unprepared.

Inspections happens outside the official open window

  • At A Glance

Detentions have been recorded on Last 36 months

%

of Ships detained have NO other detention within the last 36 months

Ocean-Going Ships (45k in total) will be detained at least ONCE the next 36-months

Conclusion

The majority of ships detained are NOT substandard but UNPREPARED

  • The Takeaway

Reality: Detained ships are not problematic – they are unprepared.

They are hit not because they are poorly managed, but because they walked into a port with:

  • The wrong preparation,

  • At the wrong time,

  • Facing the wrong focus.

This is why data-driven risk intelligence, not generic compliance, is the only sustainable strategy:

  • Port-specific readiness (via SPARC & SPOT insights)

  • Predictive targeting risk (via PRL levels)

  • Continuous benchmarking

  • Focused crew preparation

At RISK4SEA, we help managers shift from defensive thinking to proactive risk reduction. Because detention is not a sign of failure – it’s a sign of being caught unready. And that can be fixed.

MYTH #2

Myth vs Reality: In PSC, MoU Does not Matter – Port Matters

  • The Myth

“We’re prepared. We check the MoU site and know the requirements. It’s all standardized. If I know the MoU rules, I’m covered anywhere in that MoU region.”

Many operators take comfort in monitoring their PSC MoU website. The logic is: “If I know the MoU rules, I’m covered anywhere in that MoU region.”

But that confidence is misplaced.

  • Representativeness Bias - “Wrong Map Thinking”

This bias drives Myth #2 by encouraging operators to rely on the MoU label as a shortcut for risk. The assumption that ports within the same MoU behave similarly ignores well-documented port-to-port variability. As a result, managers navigate PSC risk using an oversimplified regional map instead of port-specific intelligence.

  • The Reality

REALITY #1

There is no true MoU-wide standardization

Consider the scale and diversity:

  • Paris MoU covers 1,508 PSC stations across 25 nations.

  • Tokyo MoU spans 728 PSC stations in 19 nations.

While these MoUs coordinate principles, they do not enforce common standards of training, checklists, inspection practices, or interpretations.

What does that mean for you?
A ship inspected in one Paris MoU country faces a very different inspection approach in another. There is no guarantee of consistency even within a single MoU.

PSC stations across 25 nations in Paris MoU

REALITY #2

Even single-nation regimes show regional differences

Take the US Coast Guard (USCG) – often seen as the most unified PSC regime:

  • 143 PSC stations in one country.

  • One national oversight authority.

  • One official checklist.

Yet despite that centralization, inspection patterns differ noticeably between:

  • Atlantic Coast

  • Pacific Coast

  • US Gulf

Even with identical rules and oversight, local practice varies. It’s unrealistic to expect consistency across dozens of nations in an MoU.

PSC stations in one country for USCG

REALITY #3

Even the same port, same inspectorate, is not uniform

Our analysis of ports with 100+ PSC inspections in 36 months reveals:

  • Across different ship types (Bulkers, Tankers, Containers), only 30–50% of findings are truly common.

  • The remaining findings are highly specific, reflecting inspector focus, local risk priorities, or vessel preparation gaps.

Even in the same port, predictability is limited.

%

of findings are truly common across different ship types

  • At A Glance

SPOT is the plot of Ship & Port Performance Tiers (from A to E) before the PSCI. It represents more than 230k PSCIs in the last 36 months of the global oceangoing fleet.

The Pareto principle revealed: The SPOT Zones marked red represent the 29% of PSCIs responsible for the 73% of the Detentions. This is actionable intelligence on where to focus attention and divert training & resources.

  • The Takeaway

Reality: In PSC, MoU Does not Matter – Port & Ship combination Matters.

  • The port matters.

  • The local authority and its training matter.

  • The individual inspector matters.

Operators who rely solely on MoU-level guidance risk being blindsided by local practice.


What should managers do?

  • Shift from “MoU-level” preparation to port-level intelligence.

  • Use data analytics to predict local risk.

  • Train crews on flexibility and readiness for local variations.

At RISK4SEA, we provide detailed port-level insights to help managers prepare where it actually matters. Because in PSC, MoU does NOT matter – Port matters.

MYTH #3

Myth vs Reality: PSCI severity is more important than PSCI priority/probability

  • The Myth

“PSCI Open Window and Ship Priority matter. I’ll just check the MoU site to see if my ship is eligible for inspection. If my inspection window is closed, I’m safe. I won’t be inspected.”

This is a widely held belief among ship managers and operators. The assumption is simple: “If my inspection window is closed or my priority is low, I’m safe. I won’t be inspected.”

But that assumption can be dangerously misleading.

  • Oversimplification Bias - “Single-Number Comfort”

This bias sits behind Myth #3, where PSCI priority is treated as the main indicator of inspection exposure. PSC risk is multidimensional and highly context dependent, but decision-makers gravitate toward one easy metric. The result is false comfort and missed risk signals that only emerge through combined PRL and SPOT analysis.

  • The Reality

REALITY #1

A large share of inspections happen outside the official “window”

Our research shows:

  • 20% of PSC Inspections happen when the ship’s official PSCI window is closed.

  • 25% of all detentions occur in these same supposedly “ineligible” situations.

In other words, 1 in 5 inspections and 1 in 4 detentions defy the standard eligibility logic. Relying only on “priority” and “window” status offers a false sense of security.

%

of PSC Inspections happen when official PSCI window is closed

REALITY #2

You have limited control over PSCI probability

Even if you carefully track the MoU site and your inspection window, you can’t control:

  • Random selection processes.

  • Information sharing between MoUs.

  • Intelligence-led targeting.

  • Special campaigns.

  • Port State Control Officer (PSCO) discretion.

And once a PSCO steps onboard, you can’t tell them to leave. Your “priority” level becomes irrelevant in that moment.

OUT OF

YOUR CONTROL

Random selection processes

Information sharing between MoU

Intelligence-led targeting

Special campaigns

PSCO discretion

REALITY #3

In risk management, severity always trumps probability

This is Risk Assessment 101:

  • Probability tells you how likely something is to happen.

  • Severity tells you how bad it is if it does happen.

Even if your inspection probability is low, the severity of a detention is high. Fines, delays, reputation damage, off-hire costs – the consequences can be enormous.

That’s why effective risk management focuses on severity.

RISK ASSESSMENT 101

Probability

Tells you how likely something is to happen

Severity

Tells you how bad it is if it does happen

REALITY #4

Data proves this popular myth is inaccurate

RISK4SEA’s ongoing research on PSCI Risk Levels (see: www.risk4sea.com/PRL) reveals trends that directly contradict this myth:

  • 1 out of 5 PSCIs occur with closed windows.

  • 1 out of 4 detentions happen outside official eligibility.

Many inspections are driven by intelligence or other unpredictable factors, not MoU-listed windows or priorities.

PSCIs occur with closed windows

  • At A Glance

The PRL Matrix is the result of the plot of actual data of more than 230k PSCIs for the last 36 months and the results of the POCRA providing an assessment of the PSCI Probability and PSCI Severity generating a unique Risk Level.

The PSC Risk Levels (PRL) represent Risk Conditions of escalating severity (e.g. PRL 2 represents a higher PSC risk compared to PRL1 and so on, with PRL 9 being the highest Risk that anyone may face.

  • The Takeaway

Reality: PSCI priority is an illusion – Ship & Port Tier Combination (SPOT) matters.

  • Recognize that PSCI Severity “eats” Probability for breakfast.

  • Focus on always-on readiness.

  • Build a robust inspection-prepared culture.

  • Use data to understand real risk, not just MoU-listed priorities.

At RISK4SEA, our mission is to replace guesswork and myths with evidence-based insights that help managers reduce real operational risk. Because in PSC, it’s not about if you’ll be inspected – it’s about how ready you are when it happens.

MYTH #4

Myth vs Reality: A Good PSC Checklist Is Not Enough Unless It Is Ship & Port Specific

  • The Myth

“I have a good static PSC checklist – it’s from Class, Flag, colleagues, and past experience over the last 10 years. I’ve even added a Top 20 list of detainable items. That should cover it.”

This is one of the most persistent illusions in PSC preparation. It gives crews and superintendents the comfort of structure and simplicity. But unfortunately, it also gives them false confidence.

A checklist built on global statistics or generic advice will not reflect the real risks faced in specific ports, especially when ship types, regimes, and regional enforcement patterns vary dramatically.

  • Compliance Bias - “Checklist Comfort”

This bias underpins Myth #4 and reflects the industry’s box-ticking culture. Many operators equate completing a generic checklist with being PSC-ready, primarily to satisfy perceived authority expectations. In practice, static compliance does not control dynamic port risk, and true readiness requires adaptive, risk-based preparation.

  • The Reality

REALITY #1

Static checklists lack critical port intelligence

Generic Topxx checklists are often seen as “safe.” But when RISK4SEA tested these against real PSCI data across high-activity ports, the result was striking:

The global Top 20 strategy missed more than half (50%) of actual detainable deficiency codes found during inspections.

  • In over 1,090 ports (86% of ports), the checklist missed more than half of the real detainable items

  • In over 120K PSCIs (91% of PSCIs), the checklist missed more than half of the real detainable items

  • In over 6K Detentions (94% of Detentions), the checklist missed more than half of the real detainable items

This means crews may diligently tick all boxes – yet remain exposed to the exact deficiencies that get ships detained in that specific port.

%

or more

of actual detainable deficiency codes were missed with a global Top 20 strategy for the 94% of Detentions

REALITY #2

Static checklists prioritize comfort over detention avoidance

Why are these static checklists still used? Because they’re easy and do not take office or ship staff out of their comfort zone.

  • They’re manageable, short, and familiar.

  • Crews can complete them, managers can review them, and the forms look clean.

But this efficiency is misleading: “The checklist is easy because it ignores the majority of items that actually lead to detention in that specific port.”

Checklist completion becomes a bureaucratic comfort – not a risk control strategy.

REALITY #3

Real life bites – checklists must be dynamic and risk-weighted

If your checklist were truly effective, detentions would vanish. But they don’t. Why?

Because real life – and real PSCI focus – varies per port, ship type, and campaign.

A RISK4SEA study confirms that each port had an average of 14 and a median (PSCI weighted) of 23 unique detainable codes, while many ports had more than 30. Detention risk is fragmented, dynamic, and local.

A static list may have worked yesterday. It will not work tomorrow.

unique detainable codes in the median port

  • At A Glance

A short or fixed global checklist may reduce effort, but it does not reduce risk.

According to the white paper:

  • Global Top 20 checklists fail to align with real port-level enforcement.

  • Compliance with such checklists gives a false sense of readiness.

  • PSC inspections remain unpredictable unless preparation adapts to local patterns.

  • The Takeaway

Reality: Effective PSC Checklist must be dynamic, port-specific, prioritized. It should include port-specific detainable items, by ship type, prioritized by occurrence. Comfort vs compliance is a plan to fail.

To align with real-world detention risk:

  • Start with port-specific detainable history, not global statistics.

  • Filter by ship type, as detention profiles vary widely across fleets.

  • Order by frequency, not alphabetically or arbitrarily.

Crew focus must narrow only when data justifies it. In high-variance ports, the checklist must expand – even if this adds workload. Reducing workload must not be mistaken for reducing risk.

Ship managers need to move away from legacy comfort checklists. Detentions are not avoided by familiarity – they are avoided by relevance, risk-focus, and real-time intelligence.

At RISK4SEA, we provide the tools to do exactly that. Because in PSC, a “good” checklist isn’t enough unless it knows the ship, and the port.

MYTH #5

Myth vs Reality: To Monitor PSC Performance, You Need the Port KPIs to Do It

  • The Myth

“I know how my fleet performs as I monitor my own fleet deficiencies, detentions, and trends over time and I have the full picture, I know where my PSC pain points are.”

Many managers assume that because they see their fleet’s own inspection records, they know how well they are performing. They look at their ships’ deficiencies, detentions, and perhaps trends over time-and think they have the full picture.

But that assumption misses a fundamental requirement for real performance management: benchmarking.

  • Aggregation Bias - “Average Comfort”

This bias explains Myth #5, where fleet-level KPIs create a misleading sense of control. High-level averages often mask port-specific risk pockets and vessel-level exposure. Managers believe they understand performance because headline metrics look acceptable, while localized PSC vulnerability remains unaddressed.

  • The Reality

REALITY #1

Benchmarking is about deviation from a standard, not absolute values

The basics of benchmarking require that you compare your performance against a meaningful reference point.

In the PSC world, that means measuring your ship’s or fleet’s Key Performance Indicators (KPIs):

  • Deficiency Per Inspection (DPI) to assess deficiencies vs benchmark

  • Detention Rate (DER) to assess detentions vs benchmark

  • SMS Deficiency Rate (SDR) to assess ISM & repeat deficiencies vs benchmark

  • Predicted Deficiency Index (PDI) to assess how predictable the deficiencies were

  • Human Factors & Crewing DPI to assess relevant topic findings vs benchmark

  • Regulations & Operational Compliance DPI to assess relevant topic findings vs benchmark

  • Ship Structure & Equipment DPI to assess relevant topic findings vs benchmark


Against what?

  • The Key Performance Indicator (KPI) for the same ship type (Bulkers, Tankers, Containers, etc.)

  • In the specific port where inspections occurred

  • Over a meaningful period (e.g. the past 36 months)

Without this context, raw numbers are meaningless.

BASIC PSC KPIs

DPI
DER
SDR
PDI
HFC DPI
ROC DPI
SSE DPI

Deficiency Per Inspection

to assess deficiencies vs benchmark

Detention Rate

to assess detentions vs benchmark

SMS Deficiency Rate

to assess ISM & repeat deficiencies vs benchmark

Predicted Deficiency Index

to assess how predictable the deficiencies were

Human Factors & Crewing DPI

to assess relevant topic findings vs benchmark

Regulations & Operational Compliance DPI

to assess relevant topic findings vs benchmark

Ship Structure & Equipment DPI

to assess relevant topic findings vs benchmark

REALITY #2

Each port has its own PSC profile

Ports are not interchangeable. Inspection cultures, local risk focus, inspector training, and campaigns vary widely.

For example:

  • A DPI of 0.5 in a port with a bulk carrier average of 0.3 is underperformance.

  • The same DPI of 0.5 in a port where the average is 1.2 is excellent.

Without knowing the port-specific benchmark, you can’t tell the difference.

DPI in a port with a bulk carrier average of 0.3 is underperformance

REALITY #3

Ship type matters within each port

Even within the same port, PSC statistics differ by ship type:

  • Bulk Carriers may be targeted differently than Tankers.

  • Containers may face entirely different deficiency profiles.

A proper benchmark must match ship type within port. Generalized numbers won’t cut it.

DID YOU KNOW?

Bulk Carriers
Containers

Bulk Carriers

may be targeted differently than Tankers

Containers

may face entirely different deficiency profiles

REALITY #4

You need both the average and the volume of inspections

To even create a reliable benchmark, you need:

  • The average KPI  for each port and ship type.

  • The number of PSC inspections (PSCIs) for that ship type in that port.

Why? Because inspection volume gives the benchmark its weight.

  • A port with 500 bulk carrier inspections has a robust average.

  • A port with 5 inspections does not.

RELIABLE BENCHMARK

The average KPI

for each port and ship type

The number of PSCIs

for that ship type in that port

  • At A Glance

Missing Port Averages Means Missing True Performance Insight

If you don’t know the Port KPI averages for each PSCI your ship undergoes, you’re missing the most basic context:

  • Is your ship better or worse than the port average?

  • Are you improving relative to where you’re actually being inspected?

  • Where are your real pain points?

Without these Port KPIs, you’re managing blind.

KPI Calculation Worked Example

DPI Value (Ship, Manager, Port, etc.)
Example: Deficiencies=20, PSCIs=5, DPI Value=4

Formula to be used when BEST value is either Zero or the closest to zero, to provide SCORE for 0% to 100% with 50%=Benchmark

DPI Benchmark (PSCI History)
Avg #DEFs/#PSCIs of PSCI History L36M
Example: Deficiencies=240, PSCIs=100, DPI Benchmark=2.4

SCORE from 0% to 100% indicating:

  • 0% = Worst

  • 50% = Average (KPI Value=KPI Benchmark)

  • 100% = Best (KPI Value=Ø & KPI Benchmark >0)

  • The Takeaway

Reality: PSC performance monitoring requires Port-level, ship-type-specific benchmarking.

  • Don’t just track your own DPI and DER in isolation.

  • Compare them against Port KPIs for the same ship type.

  • Use weighted averages to see where your real risks lie.

At RISK4SEA, we specialize in giving ship managers access to exactly this level of intelligence-so you can stop guessing and start managing PSC performance with real data.

Because in PSC, knowing your own numbers isn’t enough. You need the Port KPIs to understand them.