Time Series for Dummies – The 3 Step Process
Time series forecasting is an easy to use, lowcost solution that can provide powerful insights. This post will walk through introduction to three fundamental steps of building a quality model.
By Chris St. Jeor & Sean Ankenbruck, Zencos
After a satisfying meal of Chinese takeout, you absentmindedly crack open the complimentary fortune cookie. Glancing at the fortune inside, you read, “A dream you have will come true.” Scoffing, you toss the small piece of paper and pop the cookie in your mouth. Being the intelligent, wellreasoned person you are, you know the fortune is insignificant—no one can predict the future. However, that thought may be incomplete. There is a way to predict the future with great accuracy: time series modeling.
Time series modeling may not be able to tell you when you will meet the love of your life or whether you should wear the blue or the red tie to work, but it is very good at using historical data to identify existing patterns and use them to predict what will happen in the future. Unlike most advanced analytics solutions, time series modeling is a lowcost solution that can provide powerful insights.
This post will walk through the three fundamental steps of building a quality time series model: making data stationary, selecting the right model, and evaluating model accuracy. The examples in this post use historical page views data for a major automotive marketing company.
Step 1: Making Data Stationary
Time series involves the use of data that are indexed by equally spaced increments of time (minutes, hours, days, weeks, etc.). Due to the discrete nature of time series data, many time series data sets have a seasonal and/or trend element built into the data. The first step in time series modeling is to account for existing seasons (a recurring pattern over a fixed period of time) and/or trends (upward or downward movement in the data). Accounting for these embedded patterns is what we call making the data stationary. Examples of trending and seasonal data can be seen if figures 1 and 2 below.
Figure 1: Example of Upward Trending Data
Figure 2: Example of Seasonal Data
What is Stationarity?
As we previously mentioned, the first step in time series modeling is to remove the effects of the trend or season that exist within the data to make it stationary. We keep throwing around the term stationarity, but what exactly does it mean?
A stationary series is one where the mean of the series is no longer a function of time. With trending data, as time increase the mean of the series either increases or decreases with time (think of the steady increase in housing prices over time). For seasonal data, the mean of the series fluctuates in accordance with the season (think of the increase and decrease in temperature every 24 hours).
How do we achieve stationarity?
There are two methods that can be applied to achieve stationarity, difference the data or linear regression. To take a difference, you calculate the difference between consecutive observations. To use linear regression, you include binary indicator variables for your seasonal component in the model. Before we decide which of these methods to apply, let’s explore our data. We plotted the historical daily page views using SAS Visual Analytics.
Figure 3: Time Series Plot of Raw Page Views
The initial pattern seems to repeat itself every seven days indicating a weekly season. The prolonged increase in the number of page views over time indicates that there is a slightly upward trend. With a general idea of the data we then applied a statistical test of stationarity, the Augmented DickeyFuller (ADF) test. The ADF test is a unitroot test of stationarity. We won’t get into the details here, but a unitroot indicates if the series is nonstationary so we use this test to determine the appropriate method to handle the trend or season (differencing or regression). Based on the ADF test for the data above we removed the sevenday season by regressing on dummy variables for day of the week and removed the trend by differencing the data. The resulting stationary data can be seen in the figure below.
Figure 4: Stationary Data After Removing Season and Trend
Step 2: Building Your Time Series Model
Now that the data is stationary, the second step in time series modeling is to establish a base level forecast. We should also note that most base level forecasts do not require the first step of making your data stationary. This is only required for more advanced models such as ARIMA modeling which we will discuss momentarily.
Establish a Base Level Forecast
There are several types of time series models. To build a model that can accurately forecast future page views (or whatever you are interested in forecasting), it is necessary to decide on the type of model that is appropriate for your data.
The simplest option is to assume that future values of y (the variable you are interested in forecasting) are equal to the most current value of y. This is considered the most basic, or “naïve model”, where the most recent observation is the most likely outcome for tomorrow.
A second type of model is the average model. In this model, all observations in the data set are given equal weight. Future forecasts of y are calculated as the average of the observed data. The forecast generated could be quite accurate if the data is level, but would provide a very poor forecast if the data is trending or has a seasonal component. The forecasted values for the page views data using the average model can be seen below.
Figure 5: Average (Mean) Model Forecast
If the data has either a seasonal or trend element, then a better option for a base level model is to implement an exponential smoothing model (ESM). ESMs strike a happy medium between the naïve and average models mentioned above, where the most recent observation is given the greatest weight and the weight of all previous observations decrease exponentially into the past. ESMs also allow for a seasonal and/or trending component to be incorporated into the model. The following table provides an example of an initial weight of 0.7 decreasing exponentially at a rate of 0.3.
Observation  Weight (α = 0.7) 
Yt (current observation)  0.7 
Yt1  0.21 
Yt2  0.063 
Yt3  0.0189 
Yt4  0.00567 
Table 1: Example of the exponentially decreasing effect of past observations of Y.
There are various types of ESMs that can be implemented in time series forecasting. The ideal model to use will depend on the type of data you have. The table below provides a quick guide as to what type of ESM to use depending on the combination of trend and season in the data.
Type of Exponential Smoothing Model  Trend  Season 
Simple ESM  
Linear ESM  X  
Seasonal ESM  X  
Winters ESM  X  X 
Table 2: Model Selection Table
Because of the strong sevenday season and upward trend in the data, we selected an additive winters ESM as the new base level model. The forecast generated does a decent job of continuing the slight upward trend and captures the sevenday season. However, there is still more pattern in the data that can be removed.
Figure 6: Additive Winters ESM Forecast
ARIMA Modeling
After identifying the model that best accounts for the trend and season in the data, you ultimately have enough information to generate a decent forecast, as we see in Figure 2 above. However, these models are still limited in that they do not account for the correlation that the variable of interest has with itself over previous periods of time. We refer to this correlation as autocorrelation, which is commonly found in time series data. If the data has autocorrelation, as ours does, then there may be additional modeling that can be done to further improve upon the baseline forecast.
To capture the effects of autocorrelation in a time series model, it is necessary to implement an Autoregressive Integrated Moving Average (or ARIMA) model. ARIMA models include parameters to account for season and trend (like using dummy variables for days of the week and differencing), but also allow for the inclusion of autoregressive and/or moving average terms to deal with the autocorrelation imbedded in the data. By using the appropriate ARIMA model, we can further increase the accuracy of the page views forecast as seen in Figure 3 below.
Figure 7: Seasonal ARIMA Model Forecast
Step 3: Evaluating Model Accuracy
While you can see the improved accuracy of each of the models presented, visually identifying which model has the best accuracy is not always reliable. Calculating the MAPE (Mean Absolute Percent Error) is a quick and easy way to compare the overall forecast accuracy of a proposed model – the lower the MAPE the better the forecast accuracy. Comparing the MAPE of each of the models previously discussed, it is easy to see that the seasonal ARIMA model provides the best forecast accuracy. Note that there are several other types of comparison statistics that can be used for model comparison.
Model Error Validation  
Model  MAPE 
Naïve  9.0 
Mean  10.2 
Winters ESM  6.9 
Seasonal ARIMA  4.4 
Table 3: Model Error Rate Comparison
Summary
In summary, the trick to building a powerful time series forecasting model is to remove as much of the noise (trend, season, and autocorrelation) as possible so that the only remaining movement unaccounted for in the data is pure randomness. For our data we found that a seasonal ARIMA model with regression variables for day of the week provided the most accurate forecast. The ARIMA model forecast was more accurate when compared to the naïve, average, and ESM models mentioned above.
While no time series model will be able to help you in your love life, there are many types of time series models at your disposal to help predict anything from page views to energy sales. The key to accurately predicting your variable of interest is to first, understand your data, and second, apply the model that best meets the needs of your data.
Chris St. Jeor, a consultant for the Zencos Data Science team, helps businesses find actionable insights in their data. While working as an Economist for the Idaho Department of Labor, Chris produced labor forecasts and economic impact analyses for technical stakeholders. Chris has a BS in Economics and an MS in Analytics from the Institute for Advanced Analytics at NC State University. Chris’s enthusiasm for machine learning and predictive modeling is only rivaled by his enthusiasm for his family. Whether hiking trails or exploring a new museum, he is constantly on the move with his wife and two kids.
Sean Ankenbruck brings a diverse background to the Zencos team with experience in government, investment banking, wholesale energy trading and web application development. He has a passion for code and enjoys the opportunity to use his skills to solve complex problems. When not working on an analytical project, Sean spends his time contributing to the Zencos blog where he uses data to tell an interesting story. A North Carolina native, Sean received a BS in Business and a MS in Analytics from NC State University.
Related:
 Time Series Analysis: A Primer
 Time Series Analysis with Generalized Additive Models
 Automated Feature Engineering for Time Series Data
Top Stories Past 30 Days

