Direct to the Brain of the Boss – Inventory Analytics and Reporting

I’ll start with a confession: I’m an algorithm guy. My heart lives in the “engine room” of our software, where lightning-fast calculations zip back and forth across the AWS cloud, generating demand and supply scenarios used to guide important decisions about demand forecasting and inventory management.

But I recognize that the target of all that beautiful, furious calculation is the brain of the boss, the person responsible for making sure that customer demand is satisfied in the most efficient and profitable way. So, this blog is about Smart Operational Analytics (SOA), which creates reports for management. Or, as they are called in the military, sit-reps.

All the calculations guided by the planners using our software ultimately get distilled into the SOA reports for management. The reports focus on five areas: inventory analysis, inventory performance, inventory trending, supplier performance, and demand anomalies.

Inventory Analysis

These reports keep tabs on current inventory levels and identify areas that need improvement. The focus is on current inventory counts and their status (on hand, in transit, in quarantine), inventory turns, and excesses vs shortages.

Inventory Performance

These reports track Key Performance Indicators (KPIs) such as Fill Rates, Service Levels, and inventory Costs. The analytic calculations elsewhere in the software guide you toward achieving your KPI targets by calculating Key Performance Predictions (KPPs) based on recommended settings for, e.g., reorder points and order quantities. But sometimes surprises occur, or operating policies are not executed as recommended, so there will always be some slippage between KPPs and KPIs.

Inventory Trending

Knowing where things stand today is important, but seeing where things are trending is also valuable. These reports reveal trends in item demand, stockout events, average days on hand, average time to ship, and more.

Supplier Performance

Your company cannot perform at its best if your suppliers are dragging you down. These reports monitor supplier performance in terms of the accuracy and promptness of filling replenishment orders. Where you have multiple suppliers for the same item, they let you compare them.

Demand Anomalies

Your entire inventory system is demand driven, and all inventory control parameters are computed after modeling item demand. So if something odd is happening on the demand side, you must be vigilant and prepare to recalculate things like mins and maxes for items that are starting to act in odd ways.

Summary

The end point for all the massive calculations in our software is the dashboard showing management what’s going on, what’s next, and where to focus attention. Smart Inventory Analytics is the part of our software ecosystem aimed at your company’s C-Suite.

 Smart Reporting Studio Inventory Management Supply Software

Figure 1: Some sample reports in graphical form

 

Fifteen questions that reveal how forecasts are computed in your company

In a recent LinkedIn post, I detailed four questions that, when answered, will reveal how forecasts are being used in your business.  In this article, we’ve listed questions you can ask that will reveal how forecasts are created.

1. When we ask users how they create forecasts, their answer will often be “we use history.” This obviously isn’t enough information, as there are different types of demand history that require different forecasting methods. If you are using historical data, then make sure to find out if you are using an averaging model, a trending model, a seasonal model, or something else to forecast.

2. Once you know the model used, ask about the parameter values of those models. The forecast output of an “average” will differ, sometimes significantly, depending on the number of periods you are averaging.  So, find out whether you are using an average of the last 3 months, 6 months, 12 months, etc.

3. If you are using trending models, ask how the model weights are set. For example, in a trending model, such as double exponential smoothing, the forecasts will differ significantly depending on how the calculations weight recent data compared to older data (higher weights put more emphasis on the recent data).

4. If you are using seasonal models, the forecast results are going to be impacted by the “level” and “trending weights” used. You should also determine whether seasonal periods are forecasted with multiplicative or additive seasonality.  (Additive seasonality says, e.g., “Add 100 units for July”, whereas multiplicative seasonality says “Multiply by 1.25 for July.”) Finally, you may not be using these types of methods at all.  Some practitioners will use a forecast method that simply averages prior periods (i.e., next June will be forecasted based on the average of the prior three Junes).

5. How do you go about choosing one model over another? Does the choice of technique depend on the type of demand data or when new demand data are available? Is this process automated? Or if a planner chooses a trend model subjectively, will that item continue to be forecasted with that model until the planner changes it again?

6. Are your forecasts “fully automatic,” so that trend and/or seasonality are detected automatically? Or are your forecasts dependent on item classifications that must be maintained by users? The latter requires more time and attention from planners to define what behavior constitutes trend, seasonality, etc.

7. What are the item classification rules used? For example, an item may be considered a trending item if demand increases by more than 5% period-over-period. An item may be considered seasonal if 70% or more of the annual demand occurs in four or fewer periods. Such rules are user-defined and often require overly broad assumptions. Sometimes they are configured when a system was originally implemented but never revised even as conditions change. It’s important to make sure any classification rules are understood and, if necessary, updated.

8. Does the forecast regenerate automatically when new data are available, or do you have to manually regenerate the forecasts?

9. Do you check for any change in forecast from one period to the next before deciding whether to use the new forecast? Or do you default to the new forecast?

10. How are forecast overrides that were made in prior planning cycles treated when a new forecast is created? Are they reused or replaced?

11. How do you incorporate forecasts made by your sales team or by your customers? Do these forecasts replace the baseline forecast, or do you use these inputs to make planner overrides to the baseline forecast?

12. Under what circumstances would you ignore the baseline forecast and use exactly what sales or customers are telling you?

13. If you rely on customer forecasts, what do you do about customers who don’t provide forecasts?

14. How do you document the effectiveness of your forecasting approach?  Most companies only measure the accuracy of the final forecast that is submitted to the ERP system, if they measure anything. But they don’t assess alternative predictions that might have been used. It is important to compare what you are doing to benchmarks. For example, do the methods you are using outperform a naïve forecast (i.e., “tomorrow equals today,” which requires no thought), or what you saw last year, or the average of the last 12 months.  Benchmarking your baseline forecast insures you are squeezing as much accuracy as possible out of the data.

15. Do you measure whether overrides from sales, customers, and planners are making the forecast better or worse? This is just as important as measuring whether your statistical approaches are outperforming the naïve method.  If you don’t know whether overrides are helping or hurting, the business can’t get better at forecasting – you need to know which steps are adding value so that you can do more of those and get even better. If you aren’t documenting forecast accuracy and conducting “forecast value add” analysis, then you aren’t able to properly assess whether the forecasts being produced are the best you could make.  You’ll miss opportunities to improve the process, increase accuracy, and educate the business on what type of forecast error is to be expected.

 

 

How to Handle Statistical Forecasts of Zero

A statistical forecast of zero can cause lots of confusion for forecasters, especially when the historical demand is non-zero.  Sure, it’s obvious that demand is trending downward, but should it trend to zero?  When the older demand is much greater than the more recent demand and the more recent demand is very low volume (i.e., 1,2,3 units demanded), the answer is, statistically speaking, yes.  However, this might not jive with the planner’s business knowledge and expected minimum level of demand.  So, what should a forecaster do to correct this? Here are three suggestions:

 

  1. Limit the historical data fed to the model. In a down trending situation, the older data is often much greater than the recent data.   When the older much higher volume demand is ignored, the down trend won’t be nearly as significant.  You’ll still forecast a down trend, but results are more likely to be line with business expectations.
  1. Try trend dampening. Smart Demand Planner has a feature called “trend hedging” that enables users to define how a trend should phase out over time. The higher the percentage trend hedge (0-100%), the more pronounced the trend dampening.  This means that a forecasted trend will not continue through the whole forecast horizon.  This means the demand forecast will start to flatten before it hits zero on a downtrend.
  1. Change the forecast model. Switch from a trending method like Double Exponential Smoothing or Linear Moving Average to a non-trending method such as Single Exponential Smoothing or Simple Moving Average. You won’t forecast a downtrend, but at least your forecast won’t be zero and thus more likely to be accepted by the business.