Category: Equipment/Equifactor®

Equipment Root Cause Tip: Raise your hand if you have never reset a circuit breaker…….

August 7th, 2013 by


What are the risks of setting a circuit breaker without knowing why it opened?


I just saw this local news article about a father teaching his daughter about the circuit breaker panel in their house after a ceiling fan stopped working. End result….. House on fire.  Read more here.


With eighteen years in aviation and having worked on the  C-141 Aircraft, this incident brought to mind the wrong pump replaced and resetting the circuit breaker during testing explosion. Read more here.

There are additional ways to gain equipment troubleshooting experience without starting a fire. The easiest way is to attend one of our upcoming Equifactor® Course coming up in your local area. See the schedule here:

Root Cause Analysis Tip: Equipment Troubleshooting … What’s Your Approach?

June 26th, 2013 by

Before you can start the analysis of the root causes of equipment problems, you need to troubleshoot the equipment failure to find out what happened.

Some companies do this by using “smart people”.

Others hire outside tech reps to handle their most difficult problems.

I’ve even seen a company that keeps a retired employee on contract to provide expert guidance.

But the majority have NO APPROACH AT ALL. Well at least nothing more than blindly replacing parts until something works.

TapRooT® users have a resource that others don’t. An expert in a box.

What? Well … we’ve taken the expertise of Heinz Bloch (and some others) and built it into our TapRooT® Software.

It’s like having an expert equipment troubleshooter in a box.

Just follow the troubleshooting guidance and learn much more about what is causing equipment problems.

What’s even better is that the expert in a box can also be used proactively before a failure occurs to help keep equipment failures from happening.

Where can you learn about Equifactor® and TapRooT®?

At one of the 3-Day TapRooT® / Equifactor® Equipment Troubleshooting and Root Cause Analysis Courses that we hold around the world.

CLICK HERE for more information about our courses and course locations.

Detailed Up-Front Machinery Quality Assessments (MQA) — a Key Prerequisite to Reliable Major Machinery

March 27th, 2013 by

Heinz Bloch explained how Best-of-Class companies use MQA on critically important compressors, drivers, and essential process pumps at the 2013 Global TapRooT® Summit. The MQA effort consists of structured and well-defined reviews of vendor experience, drawings, calculations, and other documentation. For highest possible value, MQA is often carried out before purchase orders are issued. Mr. Bloch described three distinct phases of successful MQA tasks.

8437Images-Heinz-Bloch-150-1Heinz Bloch is a graduate of New Jersey Institute of Technology (BSME, MSME, Cum Laude). After his retirement from Exxon Chemical Central Engineering (in Baytown, Texas), he worked as a consulting engineer and author of 17 books. He is the equipment/reliability editor of Houston-based Hydrocarbon Processing and has published over 460 papers and articles on reliability improvement subjects. For several decades, he has advised industry on maintenance cost reduction and reliability improvement issues and has taught over 500 equipment uptime improvement courses on all six continents.

Click on the icon below to view the presentation and learn the three phases of successful MQA tasks:

Bloch.Heinz.Detailed Up-Front-1

Monday Accident & Lessons Learned: A Published Use of TapRooT® Following an Equipment Failure

September 10th, 2012 by

This root cause report was prepared for Fermilab Research Alliance (FRA) on September 14, 2007 following the “Large Hadron Collider Magnet System Failure”.

1)  On November 25, 2006 a heat exchanger internal to one of the Fermilab supplied magnets collapsed in a pressure test

2)  On March 27, 2007 structural supports internal to one of the Fermilab supplied magnets failed in a pressure test.

Here is the link to the Incident PDF:

Here at System improvements, Inc. and in our TapRooT® Root Cause Analysis Courses that we teach, we encourage our process be used for multiple business processes. In this Root Cause Report, the areas below were investigated using our root cause process as one of the investigation tools:

• Project Management

• Agreements

• Specifications

• Design

• Procurement & Construction

• Acceptance & Testing

• Delivery

• Commissioning & Startup

Read the report and see what they determined and also how they integrated TapRooT® into the actual report. Let me know what you think.

Root Cause Network Newsletter – September 2012

August 30th, 2012 by

Click the link below to enjoy our September 2012 issue of the Root Cause Network Newsletter. In it, you’ll learn:

– 4 Ways your root cause analysis will improve performance,

– What we’re doing to help you prevent fatalities,

– Whether more safety regulations mean fewer incidents,

– How to stop picking the low-hanging fruit of performance improvement, &

– How finding root causes saves lives.

Click here to read our September 2012 Newsletter.

Damage to Motiva’s New Crude Unit Seems Like an Excellent Opportunity for Advanced Root Cause Analysis

July 17th, 2012 by

Corrosion because a valved leaked caustic into a relatively new crude unit that was off line for some quick repairs will cause the unit to be down for perhaps a year. Here’s the story:

Human error? Equipment failure? Bad operating or maintenance practices? Unexpected corrosion? Non of these are root causes. To find the root causes you need a systematic process like Equifactor® and TapRooT® to troubleshoot equipment problems and dig down to the real, fixable root causes of the problems. 

For more information about TapRooT®, see:

Time for Fireworks Root Cause Analysis?

July 6th, 2012 by

The Wall Street Journal reported that the was a fireworks malfunction in San Diego on the 4th. It seems all the fireworks went up at once. Here’s a video of what it looked like:

The WSJ says that Garden State Fireworks Inc. issued a statement blaming the mishap on a technical malfunction.

Time for fireworks root cause analysis?

One more idea … was the theme “Fast & Furious Six” or “Shock and Awe”?

Here’s an interview with the owner of Garden State Fireworks (the company that did the display):

Good news … no one was injured.

Might have been the biggest finally ever! (And the only one that was at the start of the show!)

UK Rail Accident Investigation Branch Reports on a Rail Accident Equipment Failure

July 3rd, 2012 by

Here’s the Summary from the UK RAIB:

Detachment of a Cardan Shaft at Durham Station – 10 April 2011


On 10 April 2011, at around 12:30 hrs, a cardan shaft fell from an empty class 142 passenger train travelling through Durham station at 75 mph (120 km/h). The train ran for a distance of approximately 2 miles (3.2 km) before being stopped. A member of the public standing on a platform suffered a minor injury from ballast thrown up as the cardan shaft fell onto the track; the train suffered damage, including loss of diesel fuel.

The immediate cause of the detachment was the complete fracture of a final drive input shaft. The input shaft fractured because a seized input bearing generated a large amount of frictional heat between the shaft and bearing. The input shaft was locally heated to a temperature at which its strength was reduced so that it could no longer carry its normal loading.

The RAIB established that the seizure of the bearing was due to the setup of the bearings during overhaul which resulted in a lack of end float in the bearings when in operation. The final drive failure was not detected by the checks which were in place to identify the onset of such failures. The detached cardan shaft was not retained by its safety loops.

The RAIB has made six recommendations to Northern Rail and owners of class 14x vehicles. Two recommendations relate to reviewing the end float and alignment requirements for the class 14x final drives and ensuring that any changes to the setup of safety critical components are validated. One recommendation covers the detection of impending final drive failures. The fourth recommendation relates to the final drive post-overhaul testing and the fifth covers the provision of key design information to overhaul and maintenance contractors. The final recommendation relates to the completion of the review work associated with the events in the immediate aftermath of the accident.

For the complete report, see:

Monday Accident and Lessons Learned: When High Reliability Systems Fail

June 25th, 2012 by

What if you had a system with two regular power supplies, two back-up power supplies (diesels), and a battery back up with a separate diesel to keep it charged?

Wow!  This should be highly reliable right?

Read about how this system failed here:


Now here’s the question …

What did they miss in their “root cause analysis”?

I think they had great troubleshooting.

They even had actions to address generic problems.

But I don’t think they found the root causes of the “cloud failure” incident.

What do you think? Leave your comments here…

Equipment Troubleshooting: Taming Those Misbehaving Motors

May 18th, 2012 by

The points in the May 2012 edition of Maintenance Technology’s article, “Taming Those Misbehaving Motors,” (Thomas H. Bishop, P.E.) can help the most seasoned maintenance pros become even better equipment troubleshooters.

“Given the number of motors in most plants, it’s not surprising that they sometimes misbehave.  While maintenance professionals are typically well equipped to tame the unruly motors that come their way, they’re occasionally puzzled by the following three behaviors” (Learn 3 behaviors and troubleshooting tips).

Need to find the root causes of equipment and machinery problems at your plant? Learn more about our Equifactor® Equipment Troubleshooting & Root Cause Failure Analysis course  (view info).


Root Cause Analysis Tip: Best Practice Sharing #2 – TapRooT® Summit

April 25th, 2012 by

In today’s Root Cause Analysis Tip, Phil Goodman shares his TapRooT® best practice at our 2012 Global TapRooT® Summit.

Today is Part 2 of 12. Click here for Part 1.

Next week, hear Jeff Cooper of Boart Longyear share his TapRooT® best practice.

Time for Equifactor®? Maybe Past Time!

January 25th, 2012 by


Here the text that came with the picture ,,, don’t know if it is true …

Here are some photos of what happens when bearings overheat
in the transmissions of these monster windmills.

To date no gear oil  has been invented to withstand the pressures produced within these transmissions.

Most recently, the government gave Dow-Corning a big  grant to work on it.

Previously, many others had tried and failed.

As they age there will be many  more bearing failures.



Hard to believe that every wind turbine will fail due to inadequate gear lubrication.

I had heard that many wind turbines are not getting proper maintenance.

Wonder what Equifactor® has to say about this?

Investigation of Fatal Elevator Accident in New York Continues – Maintenance Work May Be the "Cause"

January 24th, 2012 by

The New York Times reported that Robert LiMandri, the Commissioner of the Buildings Department in New York City, said:

We know that there was work being done right before the unfortunate event, and we do believe that is a contributing cause, or the cause.

He also said:

We know for sure that those events directly before this unfortunate accident clearly are part of our investigation.

Suzanne Hart was killed while when the elevator suddenly shot upwards as she boarded.

The story also says that the about 60,000 elevators in New York produced 53 accident in the previous year.

Great Human Factors: Wrong Tools, Bad Access by Design, Per “Ingenuity” or All of the Above?

January 19th, 2012 by

As an ex-aircraft mechanic and a “sometimes gotta work on my own car” mechanic, I have in the past borrowed or made some of the tools pictured below. The questions remain:

Wrong Tool?

Bad Access by Design?

Mechanic’s Ingenuity?

Or a little bit of them all?

Finally, ever have one of your modified tools bite you back?  Share your stories in the comment section.



Oil Cooler Line Wrench #2 009 (Medium)

Drinking Water Emergency at Point Hope Caused by Pump Impeller Problems

December 27th, 2011 by

How can bad equipment reliability cause a crisis? Imagine losing the water supply at your house or business for an extended period.

It seems that all five impellers on their five pumps failed due to corrosion on pumps at the Point Hope, CA, water plant.

The previous impellers lasted lasted 67 years without failure but the new pumps at a new plant commissioned in 2005 only made it until 2011. The first impeller inspection wasn’t even scheduled until 20012.

For complete details, see these stories:–upkeep-issues-ruled-out-as-port-hope-water-emergency-cause

And if you want to learn more about troubleshooting pump problems, attend the TapRooT®/Equifactor® Equipment Troubleshooting and Root Cause Analysis Course. CLICK HERE to see the public course schedule for 2012.

Monday Accident & Lessons Learned: Make Sure You Remove the Grounding Strap Before You Energize the Switchgear!

November 21st, 2011 by

Pictures sent to me by a TapRooT® User of an unfortunate accident …

Screen Shot 2011-11-04 At 6.24.47 Pm

Screen Shot 2011-11-04 At 6.25.46 Pm

Screen Shot 2011-11-04 At 6.26.19 Pm

Screen Shot 2011-11-04 At 6.27.00 Pm

Screen Shot 2011-11-04 At 6.27.35 Pm

Screen Shot 2011-11-04 At 6.28.14 Pm

Screen Shot 2011-11-04 At 6.28.58 Pm

Monday Accident & Lessons Learned: Bad Maintenance Practices Lead to Failed Train Wheel Set and Derailment

October 31st, 2011 by

Do your maintenance folks “make it work”?

Screen Shot 2011-10-06 At 3.20.41 Pm

Looks like “just make it work” was a cause of this accident.

See the accident report from the UK Rail Accident Investigation Branch:

Blackberry Outage – Is a Three Day Outage on a High Reliability Business Application OK?

October 13th, 2011 by

Many people count on their Blackberries to run their business. They get concerned about even a one hour outage. But the most recent outage has been going on for three days.

Here’s a quote from a recent Forbes story about the unexpected outage:

In a Wednesday afternoon conference call for reporters, RIM’s Chief Technology Officer for software, David Yach, said the company is working “around the clock” to fix the service issues. Though RIM says it is still investigating the root cause of the problem, Yach expressed certainty that the global outage stemmed from the failure of a single “core switch” in Europe and was not the result of a network breach or hack. Since RIM provides back-end service support for all BlackBerrys, the company operates multiple nodes and switches around the world for routing data.

This failure caused a backlog that overwhelmed the system.

Does this sound like reliability issues you face?

Could they have avoided this issues with some proactive application of root cause analysis?

We’ll watch what comes out in future press reports.

Lightning NOT the Root Cause of Amazon Data Center Outage

August 17th, 2011 by

The Inquirer published this article:

Lightning did not cause Amazon datacentre outage

Interesting to see the root cause analysis of a computer reliability problem being discussed.

First, we could argue if “lightning” could be a root cause. But let’s save that argument for some other time.

But what I found interesting in this article was that they were eliminating a potential cause and then going on to look further.

Looks like it is a power supply reliability root cause analysis. The first step in this process is evidence collection and troubleshooting of the “cause” of the failure.

Since they don’t know the reason that the transformer exploded, finding a root cause is going to be difficult.

It would be interesting to see the process used in this engineering analysis that is in the start of the evidence collection and evaluation process that contributes to the root cause analysis.

Next, the article goes on to discuss problems with the load transferring to backup diesel generators. This would be a second causal factor that needs to be analyzed (troubleshooting and root cause analysis).

The approach for corrective action was mentioned in the article:

– more redundancy and more isolation to its PLCs, in order to prevent failures from spreading,
– a new “environmentally friendly” backup PLC
– improved load balancing
– drastically shorter recovery times

All this will be accomplished “… as soon as possible.”

Of course these corrective actions aren’t very specific (they would not meet the SMARTER criteria in TapRooT®) but they are just a list out of an article. Perhaps the company corrective actions are more detailed.

Also, it is interesting to see additional safeguards being suggested before the failure of the current safeguards are understood.

For cloud computer users, let’s hope a successful root cause analysis with effective corrective action is completed so that future outages can be minimized.

Connect with Us

Filter News

Search News


Angie ComerAngie Comer


Anne RobertsAnne Roberts


Barb CarrBarb Carr

Editorial Director

Chris ValleeChris Vallee

Human Factors

Dan VerlindeDan Verlinde

VP, Software

Dave JanneyDave Janney

Safety & Quality

Garrett BoydGarrett Boyd

Technical Support

Ken ReedKen Reed

VP, Equifactor®

Linda UngerLinda Unger


Mark ParadiesMark Paradies

Creator of TapRooT®

Michelle WishounMichelle Wishoun

Licensing Paralegal

Per OhstromPer Ohstrom

VP, Sales

Shaun BakerShaun Baker

Technical Support

Steve RaycraftSteve Raycraft

Technical Support

Wayne BrownWayne Brown

Technical Support

Success Stories

We started using TapRooT® in the mid 1990s after one of our supervisors wanted to instill a more formal process to our investigations…


In this short period of time interesting problems were uncovered and difficult issues were…

Contact Us