The Smart Forecaster

Pursuing best practices in demand planning, forecasting and inventory optimization

A readable, well-organized textbook could be invaluable to “help corporate forecasters-in-training understand the basics of time series forecasting,” as Tom Willemain notes in the conclusion to this review, originally published in Foresight: The International Journal of Applied Forecasting. Principally written for an academic audience, the review also serves inexperienced demand planning professionals by pointing them to an in-depth resource.

This neat little book aims to “introduce the reader to quantitative forecasting of time series in a practical, hands-on fashion.” For a certain kind of reader, it will doubtless succeed, and do so in a stylish way.

The author, Dr. Galit Shmueli, is the SRITNE Chaired Professor of Data Analytics and Associate Professor of Statistics and Information Systems at the Indian School of Business, Hyderabad. She has authored or coauthored several other books on applied statistics and business analytics.

The book is meant to be a text for a “mini-semester” course for graduate or upper-level undergraduate students. I think it would be a stretch to believe there is enough technical material here to serve as the basis for a graduate course, but I could see it working well for undergraduates in industrial engineering or management who have had a prior statistics course (and therefore will indeed be able to “recall that a 95% prediction interval for normally distributed errors is…”).

There are end-of-chapter exercises of appropriate size and even setups for three real-world semester projects, so instructors could use the book as envisioned by the author. The book illustrates its points using XLMiner, an Excel add-in, and students can use the free demo version for almost all the exercises. Text datasets are available from the book’s web site, which also provides a free time series analysis “dashboard” application. The author notes that other software can be used in place of XLMiner and mentions Minitab, JMP, and Rob Hyndman’s forecast library in R.

While reading this book, I was delighted by its clarity. Having spent time recently correcting the technical prose of two otherwise good graduate students, I found the writing in this book to be a refreshing contrast, making technical concepts understandable.

Another virtue of this book is its selection of topics. The technical ones are reasonably standard (smoothing methods, regression using polynomial trends, and dummy variables) but also range a bit toward the more exotic (logistic regression, neural nets, a bit of ARIMA). More impressive is the inclusion of what might be called “meta-topics” relevant to forecasting: performance assessment, an overview of alternative technical approaches, and one on the forecasting process, from definition of goals to ways to gear reports differently for managerial and technical audiences. This is the kind of forecasting wisdom we find in Chris Chatfield’s book (2004), though presented rather less tartly and with less mathematical exposition. I typically recommend Chatfield’s introductory book for more technical readers interested in getting into time series; I would recommend Shmueli’s book for a more general audience.

No review is complete without quibbles. Here are a few—too few to undo my very positive view of this impressive little book:

• The text makes a good case for “well formatted and easily readable” charts (p. 179). But I found many of the screen shots to be poorly printed and difficult to see. The book is otherwise so visually pleasing that these defects seem very out of character. It uses luxurious amounts of white space and whimsical marginal art to great effect, producing a very “light” feel that must surely help comprehension.

• The author claims (p. 115) that smoothing methods (e.g., moving averages, exponential smoothing) cannot be fully automated because “the user must specify smoothing constants.” Of course, this is not so, since there are several software packages that do this, and the text later contradicts itself on this point on page 127.

• The otherwise good discussion of autocorrelation misleads when it claims (p. 88) that negative lag-1 autocorrelation means that “high values are immediately followed by low values and vice versa.” Well, usually, but not always.

When I finished reading this book, I realized immediately that there is another target audience outside the classroom. My company often conducts training sessions on the use of our software, and these include some general background on forecasting methods and processes. If we could excise the material on XLMiner, and even if we couldn’t, this text would make a wonderful “leave behind” to help corporate forecasters-in-training understand the basics of time series forecasting. The book is so well written, well organized and well designed that it might even be read. We can certainly use it to help our new programmers understand the applications they are developing. And this book might even serve as guilty reading for a graduate student who wants to really “get” what’s going on in Box, Jenkins and Reinsel (2008).

Thomas Willemain, PhD, co-founded Smart Software and currently serves as Senior Vice President for Research. Dr. Willemain also serves as Professor Emeritus of Industrial and Systems Engineering at Rensselaer Polytechnic Institute and as a member of the research staff at the Center for Computing Sciences, Institute for Defense Analyses.

Leave a Comment

Related Posts

Infrequent Updates to Inventory Planning Parameters Costs Time, Money, and Hurts Service

Infrequent Updates to Inventory Planning Parameters Costs Time, Money, and Hurts Service

Inventory planning parameters such as safety stock levels, reorder points, Min/Max settings, lead times, order quantities, and DDMRP buffers directly impact inventory spending and ability to meet customer demand. Ensuring that these inputs are optimized regularly will dramatically improve customer service levels and will reduce the amount of unnecessary inventory spending.

Want to Optimize Inventory? Follow These 4 Steps

Want to Optimize Inventory? Follow These 4 Steps

Service Level Driven Planning (SLDP) is an approach to inventory planning based on exposing the tradeoffs between SKU availability and inventory cost that are at the root of all wise inventory decisions. When organizations understand these tradeoffs, they can make better decisions and have greater variability into the risk of stockouts. SLDP unfolds in four steps: Benchmark, Collaborate, Plan, and Track.

Riding the Tradeoff Curve

Riding the Tradeoff Curve

In the supply chain planning world, the most fundamental decision is how to balance item availability against the cost of maintaining that availability (service levels and fill rates). At one extreme, you can grossly overstock and never run out until you go broke and have to close up shop from sinking all your cash into inventory that doesn’t sell.

Recent Posts

  • Warehouse supervisor with a smartphone.Top 3 Most Common Inventory Control Policies
    To make the right decision, you’ll need to know how demand forecasting supports inventory management, choice of which policy to use, and calculation of the inputs that drive these policies.The process of ordering replenishment stock is sufficiently expensive and cumbersome that you also want to minimize the number of purchase orders you must generate. […]
  • Office staff are analyzing The Right Forecast Accuracy Metric for Inventory PlanningThe Right Forecast Accuracy Metric for Inventory Planning
    Testing software solutions via a series of empirical competition can be an attractive option. In the case of forecasting and demand planning, a traditional “hold out” test is a good way to assess monthly or weekly forecast accuracy, but it is minimally useful if you have a different objective: Optimizing inventory. […]