forecasting and time series 4-6

Lakukan tugas rumah & ujian kamu dengan baik sekarang menggunakan Quizwiz!

stationary

the general linear process is (stationary/not stationary)

nonstationary

With a _ series, the ACF typically does not die off quickly as the lag increases.

larger

for _ sample sizes, the approximate sampling distribution of rk cane found, when the data come from ARMA-type models

no die when go back in time

for the IMA(1,1) process the weights on the et's

AR(p)

for this model the autocorrelations don't cut off at a certain lag; they die off gradually toward zero

jagged

for φ near −1, the overall graph of the process will appear _

AIC

=−2logL+2k, where L here is the maximized likelihood function and k = p + q + 1 for a model with an intercept term and k = p + q for a model without an intercept.

zero weaker

AR(1) process, Since |φ| < 1, the autocorrelation gets closer to _ or _ as the number of lags increases.

recent

AR(1), the value of the process is associated with very _ values

ARIMA

Auto Regressive Integrated Moving Average

stationary invertible

For ARMA(p,q) models, we restrict attention to those models which are both _ and _

smooth

For φ near 1, the overall graph of the process will appear _

integrated moving average

If the ARIMA process has no autoregressive terms

strongly positive

IMA (1,1) as time goes on correlation between values of the process will be _ _ for small lags

larger

IMA (1,1) as time goess on var(Yt) gets

autoregressive integrated process

If the ARIMA process has no moving average terms

yule-walker equations

If the general AR process is stationary, these can be formed

normally distributed

If the model of AR(p) is correct, then the sample partial autocorrelations for lags greater than p are approximately _ _ with means 0 and variances 1/n.

PACF

If we use an AR model for these data, we should examine the _ to determine the order of the AR model.

conditional bivariate

If {Yt} is a normally distributed time series, the PACF can be defined as the correlation coefficient of a _ _ normal distribution: φkk = corr(Yt,Yt−k|Yt−1,...,Yt−k+1)

general linear process

Let {Yt} be our observed time series and let {et} be a white noise process (consisting of iid zero-mean r.v.'s). {Yt} is a general linear process if it can be represented by: Yt = et + ψ1et−1 + ψ2et−2 + · · · where et , et−1, . . . are white noise.

nonunique

MA(1) process is _, we get the same autocorrelation function if we replace θ by 1/θ.

1. choosing appropriate values for p,d,q 2. estimating the parameters 3. checking model adequacy

Model specification for ARIMA(p,d,q) models involves:

noninvertible

Overdifferencing will create a _ model, which leads to problems with interpretability and parameter estimation.

do not have evidence against

Sample PACF, for any lag k > p, if the sample partial autocorrelation φˆkk is within 2 standard errors of zero (between −2/√n and 2/√n), then this indicates that we _____________ the AR(p) model.

invertible

So the MA(1) model is _ if and only if |θ| < 1

exceed

The AR(p) process is stationary if and only if the solutions of the AR characterstic equation _ 1 in absolute value

p q

The EACF table for an ARMA(p, q) process should theoretically have a triangular pattern of zeroes with the top-left zero occurring in the _-th row and _-th column (with the row and column labels both starting from 0).

correlation

The PACF at lag k is denoted φkk and is defined as the _ between Yt and Yt−k after removing the effect of the variables in between: Yt−1, . . . , Yt−k+1.

The null hypothesis is that the series is nonstationary, but can be made stationary by differencing.

The _ _ of the dickey-fuller unit root test is that the series is nonstationary, but can be made stationary by differencing.

ARIMA(p,d,q)

The _ class of models as a broad class can describe many real time series

Dickey-Fuller Unit Root

The __________ test is a formal hypothesis test for whether the time series is "difference nonstationary."

zero

The autocorrelations ρk are _ for k > q and are quite flexible, depending on θ1, θ2, . . . , θq, for earlier lags when k ≤ q.

should not

The fact that the sample autocorrelations do not cut off after a certain lag tells us that we _______ use an MA model for this time series.

theta1 theta2

The formulas for the lag-k autocorrelation, ρk, and the variance γ0 = var(Yt) of an AR(2) process are complicated and depend on _ and _

invertible

We can solve the nonuniqueness problem of MA processes by restricting attention only to _ MA models.

ARIMA

We have seen that many real time series exhibit nonstationary behavior. For these _ models are better

exponential growth

When AR with |φ| > 1, we get an _ _ model in which the weights on past disturbance terms blow up (rather than dying out) as we go further into the past.

autoregressive moving average process of order p and q

a time series that has both autoregressive and moving average components

sample autocorrelation function

a useful tool to check whether the lag correlations that we see in a data set match what we would expect under a specific model

rk

are estimates of patterns of the ρk for ARMA models

simple sampling distribution

because of its form and the fact that it is a function of possibly correlated values, rk does not have a _ _ _

positive

box-cox transformation assumes the data values are all

arma(p,q)

can be found numerically solving a series of equations that depend on φ1, . . . , φp or θ1,...,θq.

partial autocorrelation function

can be used to determine the order p of an AR(p) model

first differences

d = 1

0

general linear process expected value

present and past white noise terms

general linear process is a weighted linear combination of

linearly

if Yt is changing exponentially, the logged series will change

constant

if box-cox transformation data values are not all positive add a _ to make them positive

white noise

if the original data follow a random walk process, then taking first differences would produce a stationary _ _ model.

zero

in the AR(p) process, φkk = _ for all k > p.

MA(q)

in this model we know the autocorrelations should be zero for lags beyond q

IMA(d,q)

integrated moving average

extended autocorrelation function

is one method proposed to assess the orders of a ARMA(p, q) model.

Box Jenkins Method

process of iteratively proposing, checking, adjusting, and re-checking the model for fitting time series

autoregressive processes

process that takes the form of a regression of Yt on itself, or more accurately on past values of the process

non stationary

process with a mean function that varies over time

disturbance term

so the value of the process at time t is a linear combination of past values of the process plus some independent _ _

moving average

special case of the general linear process

model adequacy

strategy for checking _ _ is to see whether the rk's fall within 2 standard error of their expected values (the pk's)

stationary

taking the natural log and then taking first difference should produce a series that is

MA(q)

the PACF for a _ process behaves similarly as the ACF for an AR process of the same order

shapes

the autocorrelation function can have a wide variety of _, depending on the values of φ1 and φ2

increases

the autocorrelation ρk dies out toward 0 as the lag k _

decays increases

the partial autocorrelation of an MA(1) process never equals zero exactly, but it _ to zero quickly as k _.

exponentially decaying

we often assume that weights in the general linear process are

reciprocal transformation

when box-cox lambda = -1

square root

when box-cox lambda = 1/2

regularity condition on the coefficients

when the number of terms in the general linear process is actually infinite we need some

BIC

−2logL+klog(n),


Set pelajaran terkait

Informative Essay Vocabulary Terms

View Set

BIO141: MasteringBio - Chapter 5

View Set

Chapter 11 video quizzes and reading quiz

View Set

NAEMT All Hazards Disaster Response - BIF

View Set

English 2 Honors 1st semester final

View Set

Lab #11: Spinal Cord, Spinal Nerves and Cranial Nerves

View Set

Testing & Evaluation - Chapter 9

View Set