Jump to content

Availability (system)

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Arjayay (talk | contribs) at 19:25, 26 January 2021 (a > an +). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Availability is the probability that a system will work as required when required during the period of a mission. The mission could be the 18-hour span of an aircraft flight. The mission period could also be the 3 to 15-month span of a military deployment. Availability includes non-operational periods associated with reliability, maintenance, and logistics.

This is measured in terms of nines. Five-9's (99.999%) means less than 5 minutes when the system is not operating correctly over the span of one year.

Availability is only meaningful for supportable systems. As an example, availability of 99.9% means nothing after the only known source stops manufacturing a critical replacement part.

Definition

There are two kinds of availability.

  • Operational
  • Predicted

Operational availability is presumed to be the same as predicted availability until after operational metrics become available.

Availability

Operational availability is based on observations after at least one system has been built. This usually begins with the brassboard system that is used to complete system development, and continues with the first of kind used for live fire test and evaluation (LFTE). Organizations responsible for maintenance use this to evaluate the effectiveness of the maintenance philosophy.

Predicted availability is based on a model of the system before it is built.

Downtime is the total of all of the different contributions that compromise operation. For modeling, these are different aspects of the model, such as human-system interface for MTTR and reliability modeling for MTBF. For observation, these reflect the different areas of the organization, such as maintenance personnel and documentation for MTTR, and manufacturers and shippers for MLDT.

MTB

Mean Time Between (MTB) depends upon the maintenance philosophy.

If a system is designed with automatic fault bypass, then MTB is the anticipated lifespan of the system if these features recover all possible failure modes (infinity for all practical purposes). Such systems will continue with noticeable interruption when these conditions are satisfied unless there is an open request. This is called active feedback, which requires maintenance to prevent mission failure. Active response is required for systems that can be maintained, such as satellites.

If a system has no redundancy, then MTB is in return of failure rate, .

Systems with spare parts that are energized but that lack automatic fault bypass are to accept actually results because human action is required to restore operation after every failure. This depends upon Condition-based maintenance and Planned Maintenance System support.

MTTR

Mean Time To Recover (MTTR) is the length of time required to restore operation to specification.

This includes three values.

  • Mean Time To Discover
  • Mean Time To Isolate
  • Mean Time To Repair

Mean Time To Discover is the length of time that transpires between when a failure occurs and the system users become aware of the failure. There are two maintenance philosophies associated with Mean Time To Discover.

CBM works like your car where an oil indicator tells you when oil pressure is too low and a temperature indicator tells you when engine temperature is too high. There is zero time to discover a failure where an indicator is placed in front of a system operator.

PMS is required for silent failures that lack CBM. PMS works is periodic maintenance, like when you perform diagnostic tests on your car every 90 days (or 3,000 miles). A failure may occur any time during the 90 days, such as a broken light, but you will not become aware until you perform diagnostic test.

Mean Time To Discover is statistical when PMS is the dominant maintenance philosophy. For example, if a fault is discovered during PMS diagnostic procedure that is run every 10 days, the average fault duration will be 5 days. This creates a dependency between availability performance and labor costs. There is no such dependency associated with CBM.

Mean Time To Isolate is the average length of time required to identify a setting that needs to be adjusted or a component that needs to be replaced. This is dependent on documentation, training, and technical support. This tends to be less on systems that have CBM because users can begin with the list of items connected to the indicator used to notify users about the fault. This also tends to be less on fully documented systems.

Mean Time To Repair is the average length of time to restore operation. For mission critical systems, this is generally estimated by dividing time required to replace all parts by the number of replaceable parts.

MLDT

Mean Logistics Delay Time is the average time required to obtain replacement parts from the manufacturer and transport those parts to the work site.

MAMDT

Mean Active Maintenance Down Time is associated with Planned Maintenance System support philosophy. This is average amount of time while the system is not 100% operational because of diagnostic testing that requires down time.

For example, an automobile that requires 1 day of maintenance every 90 days has a Mean Active Maintenance Down Time of just over 1%.

This is separate from the type of down time associated with repair activities.

Nines

Availability expectations are described in terms of nines.

The following table shows the anticipated down-time for different availabilities for a mission time of one year. This is the typical time-span used with commercial systems.

90% 99% 99.9% 99.99% 99.999% 99.9999%
40 days 4 days 9 hours 50 minutes 5 minutes 30 seconds

Supportability

Systems that require maintenance are said to be supportable if they satisfy the following criteria.

  • Adequate labor for PMS diagnostics or fault indicators for failure modes that lack PMS
  • Access to manufacturing sources for replacement parts
  • Sufficient resources to sustain access to replacement parts and maintenance labor (money, lodging, ...)
  • Training and maintenance documentation suitable to restore operation
  • Transportation services to deliver replacement parts

Systems that lack any of these requirements are said to be unsupportable.

Mission Failure

Mission failure is the result of trying to use a system in its normal mode when it is not working.

Apart from human error, mission failure results from the following causes.

  • Lack of instrumentation for the fault in the form of an indicator suitable for CBM
  • Failure after PMS diagnostic testing has been completed but before operation is attempted
  • Lack of diagnostic PMS testing for any failure mode that lacks CBM

See also

References

Public Domain This article incorporates public domain material from Federal Standard 1037C. General Services Administration. Archived from the original on 2022-01-22. (in support of MIL-STD-188).