Slide coverage Flashcards
(33 cards)
there is a CLRM assumption that can replace two others. elaborate
E[u_t | X_t] = 0
This can replace E[u_t]=0 AND cov(x_t, u_t)=0
how to ensure that assumption 1 holds
Include a constant term
what is endogeneity? relate to exogeneity
Exogeneity refers to the assumption that exogeneuous variables are not correlated with the error terms. In other words,: E[u | X] = 0
Endogeneity is a condition where exogenity is violated.
why would endogeneity be present?
This is the same as asking for reasons why the explanatory variables and error terms might be correlated.
If we are missing an important variable, both the constant term and the explanatory variables can try to explain this movement without actually being related. Unless if completely uncorrelated. in that case, only the constant will be affected.
Measurement error is also a source.
Endogeneity is present. What do we do?
The obvious one is to account for missing variables.
if we have panel data, this can help leverage common effects.
IV estimation.
Heckman Correction
elaborate on instrumental variables
This is a method for endogeneity.
First we split the X into two parts.
1) the variables not correlated with error term
2) the varibles that are correlated with the error term
Then we extract the part of X that is NOT correlated. We use this to get a consistent estimate of the variation in Y.
We want to use instrumental variables. These are correlated with the endogenuous variables of X, but not the error terms.
We find some other variable, call it Z. we want this to be correlated with the endogenuous variables, but not the error terms.
We would have used this Z directly, but we want to understand the effect that X (endo) has on Y. This is why we go to greater lengths to include it.
We use Z in a spearate regression that will predict values for the endogenuous variable(s). these are then used.
This is referred to as two-stage as well. first regress on the IV, then regress on the real shit.
It is worth noting that good IV’s are usually difficult to find.
elaborate on the possible forecasting methods
We have 3)
1) Scaled
2) Iterative
3) Direct
Direct skips directly tpo the forecasting horizon, but iterative goes point by point. This leads to different results.
Scaled predicts 1 step ahead and then scales it with the number of steps we actually want. Does not perform well.
what is realized volatility?
Realized volatility entails getting a on-the-spot measure of the volatility by utilizing high frequency returns like 5min intervasl, compute the corresponding volatility and scale to daily. This gives us a daily volatility estimate that represent the current level sort of .
what is MIDAS
Mixed Data Sampling
what does MIDAS do
it is a regression technique that allows for data sampled at different frequencies to be used in the same model.
Specifically, we sample the depdnent variable at a lower frequency than the independent ones.
the price at which electricity is sold is determined by the ?
merit order principle
In europe, it is day ahead markets
what is merit order?
a way of ordering available sources of energy.
These are ordered from lowest to highest marginal cost to determine which ones should be produced first to supply a given demand at minimum cost for the system.w
give the merit order ordering
solar, wind, nuclear, water, some sort of burning, coal, oil
why water not at 0?
Because it is really bad if we exhaust the resource
does solar actors and offshore wind actors earn the same?
No, because hte produce electricity at different times of the day. Wind ismostly at night. solar mostly at day. there is also different demand at these times which result in differnet clearing price.
elaborate on interpretation of the constant term
we need to be careful with this. It is very common that the constant intercept the y-axis at a point where we basically have no data points. In these chases, it doesnt really makes much sense to interepret any base level.
what can we do if we weant to test relationships that are more intricate?
We need to use testng under maximum likelihood.
this includes the 3 boys:
1) Wald
2) LR
3) LM
what is the basis for which the ML testing works?
checking if the likelihood value drops when we enforce a restriction
what can we do about heteroskedasticity?
If we happen to know the shape, we can do somethin
say the hteroskedasticity depends on: var(u_t) = sigma^2 z_t^2
What can we do?
divide the entire regression on z_t^2. this gives us error terms that are constant in variance.
elaborate on Durbin Watson
Tests a single lag for autocorrelation.
DW =ish 2(1-p^pred)
Since correlation is between -1 and 1, DW is between 0 and 4.
Values close to 2 represent no correlation.
There are also cirtical values that we need to find, and these will be somewhere between 0 and 2 and between 2 and 4.
There are actually 2 sets of critical values: lower and upper.
We need to be beyond the lower to say that there is autocorrelation
what is the error-in-variables problem?
Measurement error in the explanatory variables. this is a serious problem and will cause violation of assumption.
what is 2SLS?
two stage least squares.
Refers to first running a regression to predict some endogenuous variable by using some instrumental variable. Then when this is done, we replace the endogenuous, and run a new actual regression.
elaborate on Heckman correction
Heckman correction is a technique used in regards to fixing sample bias. It works best when there is a systematic sample bias.