Improve Forecast Accuracy, Eliminate Excess Inventory, & Maximize Service Levels
In this video, Dr. Thomas Willemain, co-Founder and SVP Research, talks about improving forecast accuracy by measuring forecast error. We begin by overviewing the various types of error metrics: scale-dependent error, percentage error, relative error, and scale-free error metrics. While some error is inevitable, there are ways to reduce it, and forecast metrics are necessary aids for monitoring and improving forecast accuracy. Then we will explain the special problem of intermittent demand and divide-by-zero problems. Tom concludes by explaining how to assess forecasts of multiple items and how it often makes sense to use weighted averages, weighting items differently by volume or revenue.
Four general types of error metrics
1. Scale-dependent error
2. Percentage error
3. Relative error
4 .Scale-free error
Remark: Scale-dependent metrics are expressed in the units of the forecasted variable. The other three are expresses as percentages.
1. Scale-dependent error metrics
- Mean Absolute Error (MAE) aka Mean Absolute Deviation (MAD)
- Median Absolute Error (MdAE)
- Root Mean Square Error (RMSE)
- These metrics express the error in the original units of the data.
- Ex: units, cases, barrels, kilograms, dollars, liters, etc.
- Since forecasts can be too high or too low, the signs of the errors will be either positive or negative, allowing for unwanted cancellations.
- Ex: You don’t want errors of +50 and -50 to cancel and show “no error”.
- To deal with the cancellation problem, these metrics take away negative signs by either squaring or using absolute value.
2. Percentage error metric
- Mean Absolute Percentage Error (MAPE)
- This metric expresses the size of the error as a percentage of the actual value of the forecasted variable.
- The advantage of this approach is that it immediately makes clear whether the error is a big deal or not.
- Ex: Suppose the MAE is 100 units. Is a typical error of 100 units horrible? ok? great?
- The answer depends on the size of the variable being forecasted. If the actual value is 100, then a MAE = 100 is as big as the thing being forecasted. But if the actual value is 10,000, then a MAE = 100 shows great accuracy, since the MAPE is only 1% of the actual.
3. Relative error metric
- Median Relative Absolute Error (MdRAE)
- Relative to what? To a benchmark forecast.
- What benchmark? Usually, the “naïve” forecast.
- What is the naïve forecast? Next forecast value = last actual value.
- Why use the naïve forecast? Because if you can’t beat that, you are in tough shape.
4. Scale-Free error metric
- Median Relative Scaled Error (MdRSE)
- This metric expresses the absolute forecast error as a percentage of the natural level of randomness (volatility) in the data.
- The volatility is measured by the average size of the change in the forecasted variable from one time period to the next.
- (This is the same as the error made by the naïve forecast.)
- How does this metric differ from the MdRAE above?
- They do both use the naïve forecast, but this metric uses errors in forecasting the demand history, while the MdRAE uses errors in forecasting future values.
- This matters because there are usually many more history values than there are forecasts.
- In turn, that matters because this metric would “blow up” if all the data were zero, which is less likely when using the demand history.
The special problem of intermittent demand
- “Intermittent” demand has many zero demands mixed in with random non-zero demands.
- MAPE gets ruined when errors are divided by zero.
- MdRAE can also get ruined.
- MdSAE is less likely to get ruined.
Recap and remarks
- Forecast metrics are necessary aids for monitoring and improving forecast accuracy.
- There are two major classes of metrics: absolute and relative.
- Absolute measures (MAE, MdAE, RMSE) are natural choices when assessing forecasts of one item.
- Relative measures (MAPE, MdRAE, MdSAE) are useful when comparing accuracy across items or between alternative forecasts of the same item or assessing accuracy relative to the natural variability of an item.
- Intermittent demand presents divide-by-zero problems which favor MdSAE over MAPE.
- When assessing forecasts of multiple items, it often makes sense to use weighted averages, weighting items differently by volume or revenue.
RECENT POSTS
Managing Spare Parts Inventory: Best Practices
In this blog, we’ll explore several effective strategies for managing spare parts inventory, emphasizing the importance of optimizing stock levels, maintaining service levels, and using smart tools to aid in decision-making. Managing spare parts inventory is a critical component for businesses that depend on equipment uptime and service reliability. Unlike regular inventory items, spare parts often have unpredictable demand patterns, making them more challenging to manage effectively. An efficient spare parts inventory management system helps prevent stockouts that can lead to operational downtime and costly delays while also avoiding overstocking that unnecessarily ties up capital and increases holding costs.
12 Causes of Overstocking and Practical Solutions
Managing inventory effectively is critical for maintaining a healthy balance sheet and ensuring that resources are optimally allocated. Here is an in-depth exploration of the main causes of overstocking, their implications, and possible solutions.
FAQ: Mastering Smart IP&O for Better Inventory Management.
Effective supply chain and inventory management are essential for achieving operational efficiency and customer satisfaction. This blog provides clear and concise answers to some basic and other common questions from our Smart IP&O customers, offering practical insights to overcome typical challenges and enhance your inventory management practices. Focusing on these key areas, we help you transform complex inventory issues into strategic, manageable actions that reduce costs and improve overall performance with Smart IP&O.