Time series Flashcards
(54 cards)
define a time series {r_t}
A sequence of random variables over time
what does the models of AR, MA, and ARMA sort of “do”?
They attempt to make use of whatever information they have prior to some point, and the goal is to make a prediction by utilizing the linear relationship between the prior information and the prediction point.
what is the ultimate focus on linear time series analysos?
the aspect of correlation. In time series, it is referred to as autocorrelation.
What is the foundation of time series analysis?
A concept called stationarity.
Define stationarity
The book differ between strict and weak stationarity.
Strict stationarity: A time series {r_t} is said to be strictly stationariy if the joint distribution (r_t1, r_t2, …, r_tk) and the joint distribution that has the same variables, but shifted by t, are the same. Same distribution, same mean, same variance all identical.
Weak stationarity: A time series {r_t} is said to be weakly stationary if both the mean of r_t and the covariance of r_t and r_{t-l} is time invariant. This means that E[r_t] = mu, and cov(r_t, r_{t-l}) = gamm_l, which depend only on l.
Make no mistake: This notaiton entail that each r_t in the time series {r_t} has the same mean. this makes sure that the trend is flat around this mean value.
Also take time to consider the covariance requirement. It say that the covaraince between r_t, which is just a random variable in the time series, and its lag-l corresponding random variable, is the same regardless of what r_t is. all that matters, is the lag-l. Thus, the covariance between each consecutive random variable in the time series must be the same, and the same goes for each second consecutive etc and so on. This is basically saying that the same relationship among all variables must remain true through the entire time series. Note that it is still very plausible that the covariance is close to 0 for many of the cases, which is totally fine. This is ultimately a part of what we want to figure out with time series analysis.
From the requirement of cov() = constant, we also get that the variance of each random variable in the time series must be the same.
Note as well what the cov() requirement does to seasonality. It enforce that there is no seasonality because of for instance how lag-1 covariance being constant for all consecutive random variables means that we cant get those seasonal swings.
what is lag-l autocovariance
Facny word for the correlation between two random variables separated by a lag of “l” in a time series.
what are we actually considering with {r_t}+
I believe a “true” time series with infinite random variables. I believe this is the case because of how the book describe samples of a time series as {r_t}^T_{t=0}
how do we denote the lag-l autocorrelation of r_t?
p_l
What is portmanteau and Ljung-box used for?
testing to see whether multiple differnet lag-l autocorrelations are 0 or nto
define white noise
White noise is a time series consisting of iid random variables with finite mean and variance.
It is typical to use white noise with mean zero and variance equal to sigma^2, along with normal distribution. Then it is called gaussian white noise.
what can we say about the ACF’s of white noise series+
all are zero because of how they are iid.
what is actually being defined here?
Here we define what it would entail for the UNDERLYING PROCESS that governs the time series data generation to be linear. It is not related to a linear model.
it simply say that IF the time series is inherently linear, its data generation process will follow that formula.
This means that if we assume that some time series is linear, we basically assume that its data generation process can be described using that formula.
On the specifics, the assumption is that a linear time series follow a sum of weights and white noise. This indicate that the next data point is a function of all the previous movements multiplied by some weight.
Also, do not neglect the mu. This is the mean. The white noise series is more related to representing some kind of relationship regarding “what happens next if the shock was this”. we know that the mean is mu on aggregated metrics, but our goal is to relate the shocks to future movement in order to make our predicitons. We’re essentially looking to analyze how the time series react to certain shocks.
The more interesting part is how past values are not included. We’re only using the mean and how the time series react to shocks.
AR models depend on r_(t-i) and not just the white noise. Can it then be defined as linear?
Yes, because we can rewrite the model (using the expected value for mu etc) to get:
r_t - mu = ø1 (r_{t-1} - mu) + a_t
and if we repeat the substitutions, we get that the result is linear, where the weights are powered increasingly. Since the ø’s must be less than 1 in absolute value, this sum will diminish.
in the book, they do this for all models. They say that “for a weakly stationary AR(1) model….”. what exactly are they referring to? What is weakly stationary?
The model in its abstract form represent a process of data generation. The idea is that if the underlying process that we wish to describe with a model is stationary, we need our model to inhibit the same stationary features. Because of this, it is important to us that we know our model do this.
Stationarity is a baseline for time series because without stationarity in the data process that generate the time series in the first place, there is nothing to predict. THere will be no way to use a time series model to forecast. Because of this, we place great emphasis on the fact that our time sereis are at least weakly stationary.
And due to the fact that our time series are weakly stationary, we need a model that capture this. And the model should capture it with the same values as well, but this enters into the world of estimation.
what is meant by “autocovariance”
Autocovariance determine how some random variable relate to its previous values. Covariance of the same random variable.
what is the interpretation of autocovariance/covariance?
Nothing really. This is why we use correlation
express corrlation using only gamma
correlation is defined as covariance divided by the product of standard deviation. Since standard deviation is the same regardless of which random variable we look at in the time series, we know that this is just the variance. thus we get “covariance/variance” and for some specific lag we simply take the covariance (auto) and divide on the variance. gamma_l / gamma_0.
what is the autocorrelation function?
the graph we get from plottign the correlation values for each lag.
what is the key regarding white noise series
the covariance (autocovariance) between different variables in the white noise series is always 0 unless we take the covariance with itself, which is the variance, which is sigma^2.
what does th autocorrelation function of white noise look like?
A single peak of 1 at the lag 0, all 0 otherwise
what is the distribution of a white noise autocorrelation function?
Note that this applies only for the sample autocorrelation function of a gaussian white noise series.
Normal with mean 0 and variance 1/T.
It will approximate zero variance for larger time series.
This means that while we should expect a value of 0 for hte autocorrelation, it will swing/fluctuate with a variance of 1/T.
We use this in hypothesis testing to see if something is basically white noise or not. If we use a t-ratio test, we’re cehcking the hypothesis that the mean is indeed 0. If it turns our that the observed values are CRITICALLY EXTREME for the t-ratio statistic, we know that the observed autocorrelation is likely not a result of a white noise series, because it indicates that there is significant autocorrelation in our sample for the specific lag we’re checking (autocorrelation is specified on a par-lag basis).
how can we easily test if something is white noise? Say we have an autocorrelation funciton, and we want to establish bands indicating signifance?
Easiest ir pobably to set up confidence interval like this:
+- 1.96 x 1/sqrt(T)
this gives us the limits for 95% confidence.
for MA models, we typically neglect the constant mu term. Why?
It considerably eases computaitons, and at no loss of generality. This is because we can obtain a zero-mean series by simply subtracting the mean, shifting all values/observations.
how does a q’th order MA model look like+