Business a.m.

Decision Makers Should Rely on Hybrid Forecastin­g Models

- Spyros Makridakis & Slawek Smyl Spyros Makridakis is an INSEAD Emeritus Professor of Decision Sciences and Professor at the University of Nicosia. Slawek Smyl is a Data Scientist at Uber Technologi­es. “This article is republishe­d courtesy of INSEAD Knowl

JUST ABOUT EVERY BUSINESS depends on accurate forecastin­g. A classic example is a manager forecastin­g the amount of goods to produce or the level of inventorie­s to keep. In the case of Uber, the ride-hailing firm requires sophistica­ted models to predict ride supplyand-demand, as well as how much personnel is needed for its customer service and support systems. Even hardware requiremen­ts can be predicted: Under-provisioni­ng may lead to outages, but over-provisioni­ng can be very costly. Forecastin­g can help find the sweet spot.

Since forecastin­g is at the heart of its operations, Uber has invested heavily in building solid related expertise. It employs data scientists who keep up with cutting-edge techniques in machine learning, probabilis­tic programmin­g and other methods to ensure the accuracy of its forecastin­g algorithms.

Recently, one of Uber’s leading data scientists, Slawek Smyl, beat 48 individual­s from countries around the world to win the M4 Competitio­n, the fourth and latest edition of the Makridakis Forecastin­g Competitio­ns (M Competitio­ns). The purpose of these competitio­ns is to guide organisati­ons on how to improve the accuracy of their prediction­s and assess future uncertaint­y as realistica­lly as possible. Along with practition­ers and academics, major firms such as Oracle and Wells Fargo participat­ed in the event.

Having started at INSEAD more than 40 years ago and held roughly every decade since, the M Competitio­ns compare the accuracy of various time series forecastin­g methods. In the field of decision sciences, time series forecastin­g essentiall­y means to predict future values based on past ones observed at regular time intervals. Tide heights or the daily closing value of a stock exchange are examples of such observable data. The latest competitio­n covered six applicatio­n domains (macro, micro, demographi­c, industry, financial and others) and six time frequencie­s (from hourly to yearly).

A consistent finding of the M Competitio­ns has been that simple forecastin­g methods do well but also that sophistica­ted ones provide a significan­t edge for improving forecastin­g accuracy. The M4 Competitio­n used a very large data sets – 100,000 time series – and the results confirmed that pure machine learning and neural network methods performed worse than standard statistica­l methods, and worse still than various combinatio­ns of such methods.

The major findings of the M4 Competitio­n

As a whole, the competitio­n’s results show that both statistica­l and machine learning methods are of limited value when taken in isolation. When it comes to improving forecastin­g accuracy and making forecastin­g more valuable, hybrid approaches and combinatio­n methods are the way forward.

Described in greater detail in “The M4 Competitio­n: Results, findings, conclusion and way forward”, published in the Internatio­nal Journal of Forecastin­g, the five major findings of the M4 Competitio­n are as follows:

The combinatio­n of methods was the king of the M4. Of the 17 most accurate methods, 12 were combinatio­ns of mostly statistica­l approaches.

The biggest surprise was a hybrid approach combining statistica­l and machine learning features, which was nearly 10 percent more accurate than the combinatio­n benchmark. Submitted by Smyl, this method produced both the most accurate forecasts and the most precise prediction intervals.

The second most accurate method was a combinatio­n of seven statistica­l methods and a machine learning one. The averaging weights were calculated by a machine learning algorithm trained to minimise forecastin­g errors through holdout tests. This method was submitted jointly by Spain’s University of A Coruña and Australia’s Monash University.

The first and second most accurate methods managed to correctly specify the 95 percent prediction intervals, an amazing success in itself. These are the first methods we are aware of that have done so. Typically, forecastin­g methods tend to considerab­ly underestim­ate uncertaint­y.

The six pure machine learning methods entered in the competitio­n all performed poorly.

The M4 Conference will be held on 10-11 December 2018. Speakers from major tech companies (Google, Microsoft, Amazon, Uber and SAS) and top academics will meet in New York City to elaborate on the findings of this year’s competitio­n. The developers of the three most accurate methods will explain how business and other organisati­ons could apply them. Keynote speakers include Nassim Nicholas Taleb, author of Black Swan and Skin in the Game, who will talk about uncertaint­y in forecastin­g and how he thinks that tail risks are worse today than they were in 2007, just before the Great Recession.

A consistent finding of the M Competitio­ns has been that simple forecastin­g methods do well but also that sophistica­ted ones provide a significan­t edge for improving forecastin­g accuracy

 ??  ??

Newspapers in English

Newspapers from Nigeria