Regression Models Flashcards
(34 cards)
What is supervised learning?
Learning a function that maps an input and some parameters to a predicted output (label), which is then evaluated based on a ground truth value
What is an observation in supervised learning?
One row of data in the dataset
What is a feature in supervised learning?
One column of data in the dataset
What are hyperparameters?
Parameters that we define before running the model, and are not learned directly from the data
Models are typically trained in two phases: […] and […]
Training and evaluation (/prediction)
The training phase involves…
Teaching the models what predictors fall into what category, learning parameters that define the relationships between the features
The prediction/evaluation phase involves…
Gaining new observations and feeding them into our trained model to create a prediction
An update rule is defined which is calculated using the value from a […].
Loss function
What is a loss function?
A quantitative measure of error in predicted values of a model
Problems with quantitative responses are usually […] problems, while those with qualitative responses are usually […] problems.
Regression, classification
Provide an example of a regression model used for qualitative responses.
Logistic regression, since it estimates the probability of a choice
What are regression models?
Regression models are used to model any continuous target or outcome (i.e. loss, revenue)
How does a linear regression model create predictions?
Defining a straight mathematical line that attempts to go through each point, with which we can substitute x and y to find y’.
In a linear regression model, epsilon represents…
A hyperparameter chosen at implementation that guides the model to learn the data
In a linear regression model, b0 and b1 represent…
Parameters that represent the y-intercept and slope of the line respectively, calculated via an equation
What is Mean Squared Error (MSE), and how can we use it for a linear regression model?
MSE is the sum of squared errors divided by the number of data points, which we can use to measure the error of our model and adjust coefficients accordingly
Error metrics are useful for training a model because…
We can focus on easier-to-read parameters rather than the predictions themselves to generate model/data insights
What is a black-box model?
A black-box is a type of learning model that does not explain its process, creating a sort of ‘black-box’ between input and output
The best step-by-step practice for modelling involves…
Establishing the ideal cost function, developing multiple models with different hyperparameters, and comparing the results according to our loss function (establishing, developing, comparing)
We can calculate the parameters of our linear regression model by…
Solving the linear system of equations formed by our data points for epsilon, calculating the sum of these equations, then using the derivative to solve for bn.
The area above the mean of a linear regression line is called…
Explained variation
The area below the mean of a linear regression line is called…
Unexplained variation
Total variation is calculated as the sum of…
Explained and unexplained variation
How can we add more curvature to our linear regression model?
By using a polynomial regression instead