Here we can see how the P-F interval informs our decision on frequency of condition monitoring tasks. It shows a deteriorating resistance to failure over time, and a range of check points throughout that time period.
This is a good description for just about any failure that condition monitoring could be useful for detecting.
As an example, I will look at the failure of a roller bearing due to metal fatigue, using the points on the curve as a guide.
Metal fatigue is the effect of taking a paper clip and folding it back on itself several times. What happens? The metal within the paperclip weakens, becomes fatigued, and then breaks altogether. What we don’t see are the range of microscopic effects and events that lead up to the paperclip breaking.
It is the same with a bearing. As we have reviewed previously, over time the metal within the races, balls and other elements weakens. (We will avoid getting into the discussion about why it would weaken for now) While it remains sub-surface, that is within the walls of the race for example, we often do not even know that it is developing, nor are we able to do anything to detect it.
However, once it actually breaks the surface, say at point B, then we can start to detect it via changes in vibration levels. As the resistance to failure deteriorates even further, point C, then we may be able to detect it via different means. Changes in heat, noise, amperage draw, and equipment performance are all examples of differing means to predict failure.
The time remaining between when we actually predict the warning signs of failure, and when it actually fails functionally (no longer able to do what we require of it), is the P-F interval.
So, the frequency of any task that sets out to detect the warning signs of failure needs to be less than the P-F interval. This is a fundamental issue and one that is often misunderstood. Working through the criteria for applicability and effectiveness are what guides us as to whether or not to apply a task. It has nothing to do with whether we find anything or not.
Our decision logic (applicability and effectiveness) have already told us that it is wise for us to apply the task, because if we don’t make the efforts to predict it, then we will be faced with undesirable consequences of failure.
If we take it upon ourselves to lengthen the frequency, or worse, to remove the task altogether, merely because it has not detected anything yet, then we are setting ourselves up for an unpredicted failure. For example, if we extend the frequency to six months or greater, then we will detect the warning signs of failure based solely on good fortune. If we remove the task altogether, we are destined for the exact equipment failure that this task was designed to protect us from.
I hope this is of use in either defending the condition monitoring tasks you already have in place, or in helping you to determine the frequency of on-condition tasks that you are currently working with. There are a whole range of additional factors that should also be considered such as the accuracy of the task, the severity of the consequences etcetera, but this should help as a basic guide.