Stat701 Fall 1999
Time series: models and forecasts.
References:
Chris Chatfield. The Analysis of Time Series.Chapman and Hall.
Todays class.
- Introduction.
- Descriptive techniques.
- Series with trend - linear filters
- Autocorrelation
- The correlogram
- Processes
- A purely random process
- A random walk
- Moving average process (MA)
- Autoregressive process (AR)
- Mixed ARMA models
- Types of series
- Economic series: interest rate
- Physical series: temperature
- Marketing series: sales
- Demographic: population series
- Process control: yield
- Objectives in time series analysis
- Description
- Explanation
- Prediction
- Control
- Overall plan:
descriptive techniques,
probability models,
fitting the models,
forecasting procedures
- Analysis in the time domain not the frequency domain
(see stat 711 for this).
Types of variation: extracting the main properties in a series.
- Seasonal effect - measure and/or remove
- Other cyclic changes
- Trend - long term change in the mean level
- Other irregular fluctuations - the residuals,
can they be explained via models such as moving average or
autoregressive.
- Stationary time series - no systematic change in the mean or
the variance and no strictly periodic variations. Most theory is
concerned with stationary time series - so need this assumption to use
theory.
- The time plot: plotting the series against time - picks up important
features such as trend, seasonality, outliers and discontinuities
Analyzing series that contain a trend
- Model it:
- Filter it
- Turn one time series into another by a linear
operation
where is a set of weights.
- To smooth out fluctuations and estimate the local mean set
- the moving average. Often symmetric with s = q and . The simple moving average: a symmetric smoothing filter with
for .
- Can also choose weights to fit a local polynomial - very close to
splines.
- Exponential smoothing: , where .
- What sort of filter? To see the trend need to remove local fluctuations, the high frequency variation - therefore need a low-pass filter.
- To see residuals remove low frequency variation - want a high-pass
filter.
- A very familiar filter , differencing. For
non-seasonal data this may be enough to obtain near stationarity.
Series with seasonal variation
A simple additive model:
.
To estimate the seasonal effect for a particular period, find the average for
that period minus the corresponding yearly average.
To remove a seasonal effect with monthly data use the filter
To remove a seasonal effect with quarterly data use the filter
Autocorrelation
Measures of the correlation at different distances apart in time.
Lag 1 autocorrelation measures the correlation between observations 1 time
unit apart.
Lag k autocorrelation:
The correlogram: a plot of against k.
Interpretation of the correlogram:
If time series is completely random then for large N, .
In fact is approx N(0,1/N), so estimates should typically
be within
Short term correlation. Stationary series often have a fairly large ,
a few coefficients greater than 0 but getting successively smaller, and
for large k, is about 0.
Alternating series: Successive observations alternating on different
sides of the overall mean, the correlogram alternates too.
Non-stationary series with a trend: comes down slowly. In fact
is only meaningful for stationary series - detrend first.
Seasonal fluctuations: if time series has a seasonal fluctuation then
correlogram has fluctuation at the same frequency.
Richard Waterman