Safety and Information Flow

After reading Sidney Dekker’s book The Field Guide to Understanding ‘Human Error’, here are some of my emerging reflections and connections. As a consequence of me forming my thinking, I’ll be developing this series over the next few weeks.

A consistent theme throughout the series is the concept of information flow and the safety culture within organisations. Information flow not only impacts physically safety in many industries; it is also an indicator of organisational effectiveness regardless of their industry. As Ron Westrum states, “By examining the culture of information flow, we can get an idea of how well people in the organization are cooperating, and also, how effective their work is likely to be in providing a safe operation.”

Book summary

Based on my notes from the book, here is a summary (aided by Google Gemini)

The “Field Guide” distinguishes between the “Old View” and the “New View” of human error. In the Old View, human errors are the primary cause of problems, and the focus is on controlling people’s behavior through attitudes, posters, and sanctions.

In the New View, human errors are seen as a symptom or consequence of deeper organizational issues, and the focus is on understanding why people behaved in a certain way and improving the conditions in which they work.

The Field Guide provides examples of practices that indicate an organization adheres to the Old View, such as investigations and disciplinary actions being handled by the same department or group, and behavior modification programs being implemented to address human error. These practices tend to suppress the symptoms of human error rather than addressing the root causes.

Information Flow

In this article of the ‘human error’ series, I relate Ron Westrum’s concept of information flow to ‘human error’.

In Ron Westrum’s paper on The Study of information flow, he writes

The important features of good information flow are relevance, timeliness, and clarity. Generative environments are more likely to provide information with these characteristics, since they encourage a “level playing field” and respect for the needs of the information recipient. By contrast, pathological environments, caused by a leader’s desire to see him/herself succeed, often create a “political” environment for information that interferes with good flow.

Westrum, The study of information flow, Safety Science 67 (2014) 58–63

As Westrum describes in his paper, generative environments are those that focus on mission, where everything is subordinated to performance aided by a free flow of information. Pathological environments are characterised by a fear and threat, where individuals hoard information for political reasons.

Between those environments, he identified bureaucratic organisations which tend to retain information within departments for protection.

These cultural styles determine how organisations are likely to respond to ‘human errors’, such as mishaps and mistakes, and to unwelcome surprises such as project-halting issues. In his paper, Westrum sets out ways in which organisations may respond to anomalous information:

  1. the organization might “shoot the messenger.”
  2. even if the “messenger” was not executed, his or her information might be isolated.
  3. even if the message got out, it could still be “put in context” through a “public relations” strategy.
  4. maybe more serious action could be forestalled if one only fixed the immediately presenting event.
  5. the organization might react through a “global fix” which would also look for other examples of the same thing.
  6. the organization might engage in profound inquiry, to fix not only the presenting event, but also its underlying causes.

High performance organisation achieve success with global fixes and inquiry where the ‘bearer of bad news’ is welcomed and supported by their leaders and peers.

Practitioner’s insights

There are means to reasonably approximate where along Westrum continuum an organisation may be. An organisation’s position along the continuum will likely reveal why intractable problems may be nigh on impossible to address.

In my own work, I’ve helped a financial services organisation monitor their position by conducting regular surveys. For the department in focus, individuals at all-levels were asked to rate anonymously each of the following kinds of statements:

  • Cross-functional collaboration is encouraged and rewarded
  • Failures are treated primarily as opportunities to improve the system
  • Information is actively sought
  • Messengers are not punished when they deliver news of failures or other bad news
  • New ideas are welcomed
  • Responsibilities are shared

This particular set of statements are from the DORA Research.

To understand the state of information flow within an organisation, practitioners should consider employing the same approach.

Connections with other concepts

Over the next few weeks I’ll be adding to this series of posts which makes further connections between aspects of The Field Guide and the following concepts:

Safety: Stressors, Resilience & Drift

After reading Sidney Dekker’s book The Field Guide to Understanding ‘Human Error’, here are some of my emerging reflections and connections. As a consequence of me forming my thinking, I’ll be developing this series over the next few weeks.

A consistent theme throughout the series is the concept of information flow and the safety culture within organisations. Information flow not only impacts physically safety in many industries; it is also an indicator of organisational effectiveness regardless of their industry. As Ron Westrum states, “By examining the culture of information flow, we can get an idea of how well people in the organization are cooperating, and also, how effective their work is likely to be in providing a safe operation.”

Book summary

Based on my notes from the book, here is a summary (aided by Google Gemini)

The “Field Guide” distinguishes between the “Old View” and the “New View” of human error. In the Old View, human errors are the primary cause of problems, and the focus is on controlling people’s behavior through attitudes, posters, and sanctions.

In the New View, human errors are seen as a symptom or consequence of deeper organizational issues, and the focus is on understanding why people behaved in a certain way and improving the conditions in which they work.

The Field Guide provides examples of practices that indicate an organization adheres to the Old View, such as investigations and disciplinary actions being handled by the same department or group, and behavior modification programs being implemented to address human error. These practices tend to suppress the symptoms of human error rather than addressing the root causes.

Using stressors to gain safety resilience

Here’s my considerations relating Nassim Taleb’s thinking on stressors and Dekker’s work.

In his Antifragile book, Taleb writes

The crux of complex systems, those with interacting parts, is that they convey information to these component parts through stressors, or thanks to these stressors: [you] get information about the environment not through your logical apparatus, your intelligence and ability to reason, but through stress.

Nassim Taleb, Anti-Fragile

I believe Taleb’s statement relate to what Dekker writes about Erik Hollnagel’s concept of safety resilience

[Resilience is] the ability of a system to adjust its functioning before, during or after changes and disturbances, so that it can sustain required operations under both expected and unexpected conditions. This needs to translate into using the variability and diversity of people’s behavior to be able to respond to a range of circumstances, both known and new.

Sidney Dekker, The Field Guide to Understanding ‘Human Error’

To me, incidents such as mistakes and mishaps are stressors which individuals should feel comfortable to share. A condition for sharing are individuals knowing they can do so without fear of retribution or intolerance.

This sharing and the subsequent adjustments to their environment is ideally done in near real-time, and in a manner satisfactory to those closest to the problem; and less through formal procedures.

This ability to safely share stressors for continuous and appropriate improvement creates resilience. It creates resilience against stock-events since it flushes out and addresses weakness that are typically hidden in plain sight. It allows people, processes and services to continuously adapt in a contextual manner.

On the other hand, formal procedures often result in delay and are further hampered through the use of erroneous targets (e.g. targeting zero incidents). Those targets may satisfy the needs of a somewhat separate function, but create perverse incentives for those making decisions in the frontline.

A typical consequence are frontline colleagues striving for flexible practices despite the purported support of formal procedures. Often frontline colleagues establish workarounds and ‘inside knowledge’ which become normalised. Sentiment often extends towards cynicism of the internal ‘powers that be’ and a collective feeling of ‘us-vs-them’.

Drift

The mismatch between formal procedures and practices grows over time. It’s an almost universal phenomenon which Dekker describes as Drift. For many in the frontline, it’s a necessary departure from the original ideal, on which counter-productive targets, expectations and officialdom is set. The over-design of procedures takes agency away from the very people who typically know best and instead favours individuals who are rule-followers.

Consider your average corporation, government department or large nonprofit, and we will sadly find this corrosive phenomenon to be endemic. Often entire departments and functions exist to create and prop-up counter-productive ill-fitting procedures.

Mistakes and Mishaps

Hoffman’s Pyramid of Advantage and Catastrophe
Illustrated in the book Unlearn, by Barry O’Reilly

Above I use the words ‘mistakes’ and ‘mishap’ with specific meaning. Ed Hoffman’s states a ‘mistake’ is when something doesn’t go to plan; a ‘mishap’ occurs when the initiative or operation doesn’t fail, but an element of it is wrong.

‘Mistakes’ and ‘mishaps’ are stressors that indicate something is not performing as it should.

It should be down to the frontline individuals to recognise, share and collectively address such incidents. It should be down to their managers to ensure these individuals feel safe to share such incidents, and ensure that the learning is accounted for, regardless of how deep it impacts the organisation.

This sharing and learning can be described as information flow through an organisation (more on this later in the series). A clear upside is the organisation’s ability to adjust itself in a timely and necessary manner.

Practitioner’s insights

The success of open, safe and frank enquiry which leads to lasting change will depend on the organisation’s culture. So, when considering the likelihood of an organisation addressing issues and their underlying causes, practitioners should consider the wider cultural environment.

This experience is backed up by the work of Ron Westrum, which I introduce in the next article. In that article I provide practitioners with a practical recommendation to understand the cultural environment and how it relates to the concept of information flow.

Connections with other concepts

Over the next few weeks I’ll be adding to this series of posts which makes further connections between aspects of The Field Guide and the following concepts:

Thanks

Many thanks to Anne Gambles, Zsolt Berend, Nawel Lengliz and Carlo Volpi for giving feedback on this article.

Book Review: Accelerate by John Kotter

Accelerate is a book by John Kotter, first published in 2014

I’ve been aware of John Kotter’s book Accelerate: Building Strategic Agility for a Faster-Moving World as foundational text in the Scaled Agile Framework (SAFe) canon. As someone who operates within organisations adopting the framework, I was curious to understand the significance of the text, and try to appreciate why it is so compelling to help kick start the SAFe movement.

John Kotter is respected thought-leader, author and academic in the field of leadership, strategy and consulting. I had previously read the fictional story Our Iceberg is Melting – another of his books that talks of change and success in a fast changing world. Since this book is written as a fable, I was hoping that Accelerate would not only offer models and practices, but also citable research and thorough case studies to underpin Kotter’s thinking.

The “dual operating system”

Eight accelerators for the network.

In Accelerate, Kotter does well in describing how organisations can apparently adapt to rapid change by adopting a ‘dual operating system’, where the traditional hierarchical co-exists with a more nimble and entrepreneurial network of individuals. These intrinsically motivated individuals are driven by a sense of urgency, are supported by a guiding coalition, are catalysed by short-term success, and other such accelerators that would be familiar to readers of this post.

The dual operating system. Traditional hierarchy on the left, integrated with the entrepreneurial network on the right.

I find the apparent successful marriage of the traditional hierarchical with this network-like structure to be an intriguing, but huge claim. A claim that disappointingly isn’t backed up by citable research or thorough case studies.

For example, Kotter says those operating on the network side need to find time in addition to the time needed to fulfil their duties in the hierarchical side. To explain this, the book is strong on rhetoric, but weak on providing something like detailed first-hand accounts, or day-in-the-life stories. I was hoping to understand how individuals had navigated the trials and tribulations to overcome the cultural status quo, protectionism and suspicion. Without this, I find Kotter claim a tall order, even with the presence of his accelerators.

Extraordinary claims require extraordinary evidence

Knowing that Kotter is an academic, I’m surprise by the lack of evidence-based research from which his models, concepts and practices should emerge. The phrase “Extraordinary claims require extraordinary evidence” comes to mind. I found extraordinary evidence is lacking in Accelerate.

The evidence you’re given are mostly brief paragraphs describing what anonymous individuals and organisations have done to adopt the ‘dual operating system’. This stands in stark contrast to other leadership and strategy books, such as Alan Lefley’s and Roger Martin’s Play to Win. Throughout Play to Win the authors lean upon their lessons transforming Proctor and Gamble. I acknowledge this is only one organisation, but at least the case study is authentic, comprehensive and underpins their advice.

Maybe Kotter has other books and research which provide the rigour I’m seeking (let me know if there are). Until I see it, I am sadly unconvinced about the feasibility of what Kotter advocates in Accelerate.

Leader’s role in enabling organisational competitiveness

How well is your organisation responding to rapid change and uncertainty?

In responding to change, as a leader, how are you engaging individuals across your organisation?  

How are you enabling organisational learning in order to enable organisation competitiveness?  

These are crucial questions that will be discussed in this talk. 

We’ll discuss the principles for nurturing and promoting continuous transformation. A transformation that is respectful of legacy, recognises established success and is accommodating to each individual’s growth journey. 

Supported by the speaker’s experience, the talk will help leaders use the following approaches to foster exploration and emergence – prerequisites for organisation competitiveness:

  1. Leaders supporting individuals to test assumptions
  2. Leaders acting as grand conductors
  3. Leaders creating a safe-to-learn environment
  4. Leaders forming and protecting an adaptive space for innovators and operators

Come along to this talk if you are a leader facing the challenge and opportunity of inclusive organisational change.

This is a TEKsystems Global Services talk created and delivered by Dean Latchana

Slides

Previous Dates

Social symbolism and attached meaning of shared objects

I know, I keep banging on about the importance of co-location, and on what we’re losing when working remotely.

Here’s another lose I’d like to share: “Deep social problem-solving with the aid of physical material”

Let’s relate this to research that reveals tool-use as an extension of our body.

“When we do something as apparently simple as picking up a screwdriver, our brain automatically adjusts what it considers body to include the tool. We can literally feel things with the end of the screwdriver.”

12 Rules for Life (Jordan B Peterson)

I speculate that tool-use (including physical material) doesn’t only extend our physical perception of self. We also attach symbolism and meaning. Think of children personifying and creating stories with a crude rag doll.

When objects are used socially, such as during a collaboration workshop, the symbolism and attached meaning have unspoken shared significance.

Does the digital and remote world in which we now operate offer the same potential as the placemaking of a shared physical environment? Does it equate to the emergent symbolism, attached meaning and collective perception that arises when collaborating with physical material?


Related reading

Photo by Kaboompics .com from Pexels