Chapter 2 - Time series Flashcards

(55 cards)

1
Q

in general, what can we say that simple tiem series models attempt to do?

A

capture teh linear relationship between r_t and all prior information.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

why is correlation/covariance linear

A

we take two random variables. Then subtract their mean, which gives us only the “differnece” in both cases (movement, relative, not absolute). We can write this as : E[AB], where A and B is the mean-adjusted variables. Since A and B are random variables, they are associated with a probability distribution. When multiplied together, we get a joint distribution. When taking expectation, we average over the distribution to find the most likely outcome (the weighted joint average). but since A and B hold relative movements, the expectation gives us the “expected joint movement”. We can consider this as the average movement in X relative to average movement in Y, and if they move opposite of each other, we get something negative as the covriance. This is inherently linear, because we compare movemnt in one variable as a function of another linear variable. I suppose that if we did E[AB^2] instead, so that B is squared, we’d model something non linear.

REMEMBER: The whole concept is to explore correlation. Movement in one variable is associated with movement in some other variable. Therefor,e it makes sense to consider locksteps.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

recall the variance property of weakly stationary time series

A

constant variance.

this means that var(r_t) == var(r_{t-1}) etc

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

elaborate on very basic hypothesis testing of lag correlation

A

We use lag-l sample correlation estimator. Under some conditions, lag-l sample correlation is a consistent estimator to the real lag-l correlation (autocorrelation).

For instance, we require the secodn moment of the return series to be bounded. Meaning, bounded variance.
we also require r_t to be iid.
Under these assumptions, lag-l sample correlation is asymptotically normal with mean 0 and variance 1/T.

Then we’d use the t-ratio test to figure out if the coefficient is statistically significant or not.

However, Bartlett is better because it assumes weakly stationary and bounded second moment, wheras the previous one require iid sequence. So the different variances test for different things.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

there is sometihng key to rememver that simplify the lag-l sample correlation estimator?

A

var(r_t) == var(r_{t-1})

This means that we can simply use var(r_t) in the denominator instead of both lags.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

briefly mention null and alternaive hypothesis when using Bartlett for testing of individual lag autocorrelations

A

null: p=0
h1: p!=0

If the time series is weakly stationary and bounded second moment, we expect the sample lag-l autocorrelation estimator to asynmptotically follow normal distirbution with mean 0 and variance as given by bartletts formula. Based on THIS,. we test whether the value observed was extreme enough to say that the correlation is most definitely not 0 or not.

when using the t-ratio, we operate with a standard normal vairable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is portmantaeu

A

sammentrekning, flere i en

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does Q statistic and Ljung Box do?

A

Test multiple lags at once. Chi squared

Q-statistic is simple. Sums together “m” squares of lag-l autocorrelations for differnet lags. Since they are assumed normal given the estimator (sample, aysymptotic etc) we know that the sum of squares gives a chi squared variable with degrees of freedom equal to the number of squared entries: m.

Q is bad in small samples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

if we have estimates for p_1, p_2, … what do we have?

A

sample autocorrelation function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

elaborate on white noise

A

A tiem series is defined as white noise if it is iid with finite mean and variance.

Gaussian noise has mean 0 and variance sigma^2.in

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

in practice, what qualifies as white noise?

A

if all sample ACF’s are 0, we have white noise.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Assume we use 1/T (iid + finite second moment finite/bounded) to apply a t-ratio to test for ACF. We get significant result and reject null. What can we say in regards to Bartlett+

IMPORTANT CARD

A

The thing is that they are used to test for different things. 1/T is used to test for whether the tiem series is iid or not (white noise, basically). Bartlett is an assumption (weak stationarity) that we can use. We need to test whether this assumption is valid or not by testing for stationarity. if stationary, we can use bartlett for variance to test for significant ACF lags.

if the series is iid, the ACF should have variance 1/T. If the estimator gives high values, it would indicate that the series is NOT iid. it would not be iid because there is a clear structure where ACF show signs of linear correlation.

If the series is weakly stationary, the variance is Bartlett. therefore, a large value from ACF relative to Bartlett would idnicate that the series is autocorrelated. We have to assume weak stationarity to use Bartlett.

IT IS WORHT ADDING that if ACF is insignificant at 1/T, it is also insig at Bartlett. Therefore, 1/T is a nice quick-test.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

define a linear time series

A

A tiem series is linear if it can be written as:

r_t = mu + ∑w_i a_{t-i}

where a_t is a sequence of iid random variables with mean zero and a well defined distribution (white noise).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

give the simple AR(1) model

A

r_t = ø_0 + ø_1 r_{t-1} + a_t

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

given the simple AR model, what is its conditional mean and variance?

A

E[r_t | r_{t-1}] = ø_0 + ø_1 r_{t-1}

Var[r_t |r_{t-1}] = Var(a_t) = sigma_a^2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

for a simple AR(1) model, give the requirements for stationarity

A

We need constant mean ,constant variance, and constant lag-l autocovariance.

Start with constant mean.

E[r_t] = mu

E[r_t] = ø_0 + ø_1 E[r_{t-1}]

if the mean is constant, we have :

mu(1-ø_1) = ø_0

mu = ø_0 / (1 - ø_1)

This entails that ø_1 cannot be 0, and we know that if ø_0 is 0 then the mean mu is 0 as well.

VARIANCE:

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

what is the stationarity requirement for AR models

A

unit roots larger than 1 in absolute value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

what is characteristic root?

A

1/x, where x

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

briedly itnroduce how we can determine the order of AR model

A

1) PACF

2) Information criteria

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

how do we use AIC and BIC?

A

Select the one with the smallest value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

when forecasting, how do we do it?

A

We forecase from some origin and up to some point.

We need to represent forecasting as conditional expectation of r_{t+1}. This basically removes the error term / shocks term at time t+1, as it has zero mean.

The variance of the error is sigma_a^2.

Basides from this, we simply use the model in a deterministic way.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

build the ACF for MA

A

WE start wit hthe MA model, and subtract the mean. This is equivalent to making it zero-mean. Since MA models have mean equal to the constant term, this is done by subtracting this constant from both sides.

The new time series is adjusted.

We take r_t and r_{t-l}, multiply them together and take expectation. This is the covariance computation. Normally, we’d include subtracting the mean here, for both variables, but since we have established a time series with zero mean, we dont need to.

E[r_t r_{t-l}].

then we need to change both into their “model” form.

Taking the expectation of this will yield -w_i sigma_a^2 for lag 1, 0 for all above.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

stationairty condition for MA models

A

THey are always stationary because they are linear combinations of white noise sequence for which the first two moments are time invariant.

24
Q

elaborate on estimating the parameters in the MA models

A

MLE is the way to go.

There are 2 ways of performing MLE:
1) Conditional maximum liklihood
2) Exact MLE

Conditional:
It is typical to assume that the very first hsocks are 0. Then, we can start at r_0 with shocks at earlier times than 0, and 0 itself, being 0. We then get shocks according to:

a_t = r_t - c_0

so, a_1 = r_1 - c_0.

a_2 = r_2 - c_0 + a_1 ø_1

EXACT:
Treats earlier shocks as parameters to be estimated as well. More expensive computationally.

Under large sample sizes, these two methods bhave similarily.

25
what is the most obvious characteristics of MA models in regards to forecasting?
They forecast the unconditional mean when after x steps when they are of order x. This is a result of their finite memory.
26
give general form ARMA
27
elaborate on properties of ARMA models
their unconditonal mean is the same as for AR. Its ACF looks like AR, while its PACF looks like MA. AS a result, both are decaying. The stationarity condition is the same as for AR, because the MA part is always stationary.
28
what is a unit root nonstationary time series?
One that has a root on the unit circle
29
name an exmaple of unit root nonstationary time sreries
random walk
30
define random walk using time series
r_t = r_{t-1} + a_t p_0 is a real value denoting the starting point, and {a_t} is a white noise series.
31
is random walk stationary+
no
32
recall the notaiton used for forecasting
X steps ahead from origin point h: p_h(X), typically with a prediction notaiton on the "p" as well
33
when forecasting with random walk, what can we say about its mean?
always the same as the origin point. doesnt matter what the forecasting horizon it, the expceted value is the current value.
34
what is interesting about the constant term in the various models
it has differenrt interpretation. This highlights some key differences in the various models. For instance, for MA models, the constant is the mean. For AR, the constant is related to the mean, but it also based on an interplay with the other constants. For regression models, constant is the intercept. For random walk with drift, the constant represent the slope of the trend line.
35
what happens if we have ARMA; and force it to have a unit root?
becomes unit root non stationary.
36
define an ARIMA process
a process is ARIMA(p,1,q) if it is non stationary, but its first-differenced version is stationary.
37
elaborate on ADF
test for whether a time series contains a unit root. it is primarily based on AR terms. It has a unit root if the coefficeint of the first AR term is 1. therefore, we test for this. if the value shows that this s highly unlikely, the series is likely stationary. This is sort of based on trying to prove whether a time series follow a random walk. it is random walk, potentially with drift, if the coefficeint of the first AR term is 1. The ADF test is actually testing whetehr the time series has exactly one unit root or not. Therefore, it answers the question "time series will become stationary with first differencing".
38
in this chapter, what relationship are we trying to capture?
Linear relationships. this is primarily linear time series analysis
39
why doesnt these models contain exogenuous variables?
they are univeriate. A univariate time series model contains only its own variable (and shocks=
40
why d owe rarely work with strict stationarity?
difficult to measure empirically, and is perhaps not that common. Also, it so happens that weak stationarity is all we need to for simple models
41
when is weak stationarity equal to strict?
if the first two moments of the asset return series is finite, and r_t is normally distributed, then they are the same.
42
if two random variabels are normally distributed, when is their correlation 0?
when indenpendently distributed
43
consider the sample covariance estimator. do we include weighting of probs?
no. This is because they are implicitly given in the sample. by weighting uniformly, we assume that the sample represent the correct distribution already.
44
elaborate on portmanteau testing in regards to ACF
these are testing multiple acfs at once. we have 2: Q-statistic Ljung Box. They are based on the same idea, but Ljung-Box is better in finite samples because it adjusts for some bias. BOht are build on the idea that if one assume that the sample lag-l autocorrelations (ACFs) are normally distributed, or at least asymptotically normal, then the sum of their squares tend to a chi squared variable. These are testing under the assumption that the ACF's are iid. and certain moment conditions, which I assume makes sure that it gets its asymptotic properties. so if the final varialbe appear to be likely chi squared with m degrees of freedom (m lags), multiplied by T, then we know that the individual ACfs are likely 0.
45
why do we care so much about the ACF?
the ACF is literally perfect because the models we are interested in now have the goal of modeling linear relationships in a univariate time series. Since correlation models degree of linear association between variables, this is all we need.
45
can white noise have mean differnet from 0?
yes. if it is zero, it is called gaussian white noise
46
what can we say about the ACFs of white noise series?
All 0
47
what is the conditional mean of an AR model? what about conditiona lvariance?
we use prior information, which means that we neglect the a_t term. for an AR(1) model, we'd get: E[r_t | r_{t-1}] = ø_0 + ø_1 r_{t-1} for the conditional variance, we're looking at: Var[r_t | prior] = var(ø_0 + ø_1 r_{t-1} + a_t) since r_{t-1} is now given ,it has 0 variance. We'd get: == var(a_t) = sigma_a^2 This indicates that the conditional variance is equal to the variance of the shock. Makes sense. Everything else is deterministic.
48
do we need stationarity conditions for AR models?
Yes, because they are not necessarily stationariy by default. We must enforce certain requirments to make it stationariyu.
49
elaborate on finding the conditions/requirements for stationary time series in regards to the mean
we need the mean to be constant, mu. Therefore, we utilize the fact that the expected value of r_t must be teh same as r_{t-1}. We get this setup: E[r_t] = E[ø_0 + ø_1 r_{t-1} + a_t] = ø_0 + ø_1 E[r_{t-1}] => E[r_t] = ø_0 + ø_1E[r_t] => mu = ø_0 + ø_1 mu => mu = ø_0 / (1 - ø_1) The same procedure is used for higher orders of the AR model. This gives a constraint, as we must enforce the denominator to not be 0.
50
elaborate on the role of covariance in AR models and stationarity
there is no additional condition that must be enforced. Rather, by taking care of mean and variance, we're good. This means, unit root fuckery.
50
elaborate on stationarity conditioon for variance regarding AR models
variance must be constant. This means that var[r_t] = var[r_{t-1}] r_t = ø_0 + ø_1 r_{t-1} + a_t var(r_t) = var(ø_0 + ø_1r_{t-1} + a_t) var(r_t) = ø_1^2 var(r_{t-1}) + var(a_t) using the fact that two are the same: var(r_t) (1-ø_1^2) = sigma_a^2 var(r_t) = sigma_a^2 / (1 - ø_1^2) As a result, we'd need 1-ø_1^2 to not be 0. 1-ø_1^2 = 0 ø_1^2 = 1 ø_1 = sqrt(1) so, for plus minus 1, ø_1 takes a value that is going to make the time series by non-stationary. NB: This is not the entire story! Variance cannot be negative. Therefore, we also have the follwong constraint: ø_1^2 < 1 so, if ø_1^2 >= 1, then it is non-stationary.
51
derive the moment equation for AR (2) model
start by finding the mean. Simple procedure, utilize the condition that mean is constant. WE get: mu = ø_0 / (1 - ø_1 - ø_2) then we use ø_0 = u(1 - ø_1 - ø_2) Subsitute this into the formula for AR(2) as ø_0 r_t = u (1-ø_1-ø_2) + ø_1r_{t-1} + ø_2r_{t-2} subtract u from both sides r_t - u = u - u - uø_1 -uø_2 + ø_1r_{t-1} + ø_2r_{t-2} r_t - u = ø_1 [r_{t-1} - u] + ø_2 [r_{t-2} - u] + a_t Now we have a formula for r_t - u. Then we multiply it by (r_{t-l} - u), and then take expectation of it. (r_t - u)(r_{t-l}-u) = ø_1 [r_{t-1} - u] (r_{t-l}-u) + ø_2 [r_{t-2} - u](r_{t-l}-u) + a_t(r_{t-l} - u) taking expectation: = ø_1 E[[r_{t-1} - u] (r_{t-l}-u)] + ø_2 E[[r_{t-2} - u](r_{t-l}-u)] + 0 gamma_l = ø_1 gamma_(l-1) + ø_2 gamma_(l-2) Then, we divide by gamma_0 to get: p_l = ø_1 p_(l-1) + ø_2 p_(l-2) for lag 1, we get: p_1 = ø_1 p_0 + ø_2 p_(-1) We know that p_(-1)=p_(1) in weakly stationary series: p_1 = ø_1 + ø_2 p_1 p_1 (1-ø_2) = ø_1 p_1 = ø_1 / (1 - ø_2) In GENERAL, we would have: p_l = ø_1 p_(l-1) + ø_2 p_(l-2) using backshift operatoe: p_l = (ø_1B + ø_2B^2)p_l ==> (1 - ø_1B - ø_2B^2)p_l = 0 We need the roots to be all solutions to be greater than 1 in absolute value. This will make sure that the effects of higher lags converge to 0.
52
assume we have fitted an AR model. How do we check to see if it is adequeate?
The residual series should behave as white noise. We can use ljung box to check this. One can also use R^2
53