week 7 - Parameter Estimation Flashcards
(64 cards)
Name three popular model fitting frameworks
Least-squared estimation (LSE)
Maximum-likelihood estimation (MLE)
Bayesian estimation (BE)
How does the LSE fit a model?
by minimising the (squared) discrepancies between predictions and observations
How does the MLE fit a model?
finds the parameter values that give the highest likelihood of the observed data
How does Bayesian estimation fit a model?
by combining prior knowledge with the observed data to derive a range of likely parameter values
What is a simple way in which to fit a model via LSE?
by fitting a linear regression (straight line fit)
For LSE, what is the objective of parameter estimation? How is this achieved?
find the parameter values that minimise the discrepancy function via optimisation algorithms
What does RMSD stand for?
What is it?
-Root Mean Square Deviation
-measure in stats which calculated the discrepancy between predicted and observed values
If you have a high RMSD what does this tell you about model fit?
model fit is not good and discrepancy between predicted and observed values is large
What are some popular optimisation algorithms?
Nelder-Mead Simplex
Simulated annealing
What is the error surface?
What can you find on the error surface?Why is that good?
-all the possible RMSD/error values that can be calculated
-place where the RMSD/error is at a minimum so you have found your model fit
What is the ‘drunken triangle’?
Which optimisation algorithm is this?
-triangles moves in a way so that its tumbling down error surface to find (optimal) minimum error
-Nelder-Mead Simplex
What are the four strategies for deciding where to move the drunken triangle/simplex?
What are they in order of instruction?
reflection, expansion, contraction, shrinking
- reflection
- reflection success -> expansion
- reflection fail -> contraction
- contraction fail -> shrinking
What does ‘reflection’ involve in parameter estimation using the Nelder-Mead Simplex?
removing the point with largest discrepancy (from error minimum) and flip it to opposite side
How does expansion move the simplex/drunk triangle? What step must it come after?
After successful REFLECTION, extend the flipped point out to take a larger step down (thus closer to error minimum)
What strategy must you apply if reflection fails in the drunken triangle/simplex?
reflection fails -> contraction: move the worst fitting point more toward the centre (error minimum)
What must you do if contraction fails in the drunken triangle/simplex?
contraction fail -> shrinking: shrink reduce the triangle/simplex by half in the direction of the error minimum
In the Nelder-Mead Simplex Algorithm, what are the starting values calculated from?
plausible values are collected from data, experiments you did or from the simulations done AND discrepancy calculated from them
What are the two steps in the Nelder-Mead Simplex?
- compute discrepancy for starting values
- tumble down the error surface until you reach error minimum
For LSE, what is the issue with finding the minimum error?
How do you try to fix this?
-you can sometimes end up at a local minimum and think it is the global minimum
-Bootstrapping: by repeating process of LSE many times with different starting values. Then you look at the variability between the model parameter estimates (how much error there is for each set of starting values)
What is bootstrapping?
What is using the variability calculated in bootstrapping like?
-provides an indicator of variability around the model parameter estimates by repeatedly sampling from the model or the data.
-like using confidence intervals
What are the drawbacks of using LSE to fit model (which even bootstrapping can’t relieve)?
-RMSD doesn’t tell you about the goodness of fit -> only about if the discrepancy is small or not
-can’t statistically compare to other models -> can’t tell whether difference is meaningful or due to chance
-parameter estimates don’t have any inherent statistical properties -> dont come with confidence intervals unless you do boot strapping
What does it mean when a drawback of LSE includes ‘not being able to statistically compare models’?
can’t compare two models which have different fits (different parameters and parameter values) and see which is better -> can’t tell whether difference is meaningful or due to chance
What mathematical function does LSE use to calculate the discrepancy?
RMSD
root mean SQUARE deviation
How is the logic in MLE kind of the opposite of LSE?
LSE the discrepancy between predicted and observed data values whereas MLE finds parameter values that give highest likelihood of the observed data