Not to be confused with power factor.

In electrical engineering the load factor is defined as the average load divided by the peak load in a specified time period.[1]

${\displaystyle f_{Load}={\frac {\text{Average load}}{\text{Maximum load in given time period}}}}$

Typical example of a large commercial electrical bill: kW Demand = 436 kW; kWh Use = 57,200 kWh; Number of days in billing cycle = 30d

Load Factor [%] = {(57,200 kWh / (30 days X 24 hours per day X 436 kW)} X 100% = 18.22%

It can be derived from the load profile of the specific device or system of devices. Its value is always less than one because maximum demand is always higher than average demand, since facilities likely never operate at full capacity for the duration of an entire 24-hour day. A high load factor means power usage is relatively constant. Low load factor shows that occasionally a high demand is set. To service that peak, capacity is sitting idle for long periods, thereby imposing higher costs on the system. Electrical rates are designed so that customers with high load factor are charged less overall per kWh. This process along with others is called load balancing or peak shaving.

The load factor is closely related to and often confused with the demand factor.

${\displaystyle f_{Demand}={\frac {\text{Maximum load in given time period}}{\text{Maximum possible load}}}}$

The major difference to note is that the denominator in the demand factor is fixed depending on the system. Because of this, the demand factor cannot be derived from the load profile but needs the addition of the full load of the system in question.