Model Selection Flashcards
(36 cards)
What is the traditional strategy?
No model selection. You run one model, and assume every continuous variable has a linear relationship. Include all interactions
Problems with traditional strategy?
No multicollinearity assessed, overfiting data and not all relationships are linear.
Historical strategy
Based on past models that ran just fine. One advantage is that with this, you are able to compare data from past models because you include same predictors and same variables
Problems with historical designs
You may ignore variables that were not initially considered. Design may be insufficient eventually
The exploratory approach
Here a wide range of models are run and then usually the one that worked is the one reported.
Problems with the exploratory strategy?
A lot of experimenter degrees of freedom. Results could be too flexible and could lead to replication problems. Use a more conservative model selection approach (AIC and BIC are always safe bets)
Don’t do this in calling analyses
Call exploratory analysis confirmatory
Theoretically driven approach
Only a limited number of models are run here, 3 tops has to be theoretically driven. Is very systematic and highly localized. Less prone to overfitting
Problems with theoretically driven?
Can miss the analysis of variables already collected. Is okay to include both models just report everything. But for confirming that you should run an additional experiment and run it as confirmatory.
What is the mixed strategy?
Report which decisions models were exploratory, theoretically driven, etc. in its appropriate sections. Can segment the analysis into different parts, lessens the overfitting issue
Disadvantages of mixed strategy?
Researcher degrees of freedom
P- focused strategy
Choose the model that has your key variables significant.
Interocular Test
You look at the results, and it hits you between the eyes. Plot your data,a which says a lot. Statistics are more like an afterthought. Worry if the graph shows no differences and the analysis is telling you otherwise
What happens with the R square
The proportion of variance accounted for. To calculate this you have the mean somewhere in the formula. Mean is a measure of central tendency in a NORMAL DISTRIBUTION. Which is why in models where this is not met, R square is not a good model fit indicator
RMSE
Is the square root of the variance squared or the ss formula. The squared errors are averaged. In other words this is a measure of the average distance any point is from the mean. But this does not by any means deal with complexity
Mallow CP
In JMP, when you try to do a stepwise regression. Another inde,x butit is not common
The AIC and BIC
The Akaike and Bayesian Information criteria. One penalizes for complexity. BIC favors simpler models over complex models
Likelihood Ratios
When you are comparing one model to the other we talk about how much likely one model is to fit the data versus another model. It gives a ratio because… is 10 times more likely, is 35 times more likely.
AIC / BIC / Likelihood ratios
Since AIC and BIC are based on Log likelihoods, you can include them in the model fit inspection, or you can transform AICs and BICs to the log likelihood scale to make them more interpretable
Cross validation
Uses two types of data. Sample data to develop a model. Test the dataset to see if that model also applies to more data. This is an empirical way to determine if the model is going to perform well out of that sample. If it does, it cross-validates.
LOO
Leave-one-out cross-validation. It holds one line out of your data out and after the model is done and created, see how well it does with that one out, and repeats for every single row in your dataset
Bootstrapping
Sampling with replacement, and treat the distribution as the population. It gives you an empirically estimated distribution Some people use it to make up for a small sample size.
When do you use bootstrapping
When you have bad, very bad distributional properties that there is no statistical test to determine the differences. This happens in something called interportal ranges.
Where could you use this?
In multilevel repeated meassures designs when you have imbalances but slight ones like in each condition some values were missing but does not apply if the whole level is missing