Exam 8 TIA Flashcards

(35 cards)

1
Q

Considerations for which risk characteristics to use

A
  1. Relationship of Risk Characteristics & Expected Outcomes
  2. Causality
  3. Objectivity
  4. Practicality
  5. Applicable Law
  6. Industry Practices
  7. Business Practices
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Considerations in establishing risk classes

A
  1. Intended Use
  2. Actuarial Considerations - Homogeneity, Credibility, and practicality
  3. Other Considerations - Law, industry, and business practices
  4. Reasonableness of results
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Advantages of multiplicative rating plans

A

Simple and practical, guarantees positive premiums, intuitive impact of risk characteristics.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Choices for Severity distributions

A

Gamma and Inverse Gaussian are common. Gamma is most used; Inverse Gaussian suits more skewed distributions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Choices for Frequency distributions

A

Poisson (possibly overdispersed with φ ≠ 1), or Negative Binomial (Poisson with Gamma mixing, uses κ).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Degrees of freedom

A

Number of parameters to be estimated.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

GLM outputs for each predicted coefficient

A
  1. Standard error
  2. p-value: an estimated probability that the absolute value of a particular β is at least that different from 0 by pure chance
  3. Confidence interval
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Impact of more observations and φ=dispersion parameter on p-values

A

p-values (and standard errors and confidence intervals) will
be smaller with larger datasets that have more observations.
They will also be smaller with smaller values of φ.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Define multicollinearity and give a way to detect it

A

Linear dependency among 3+ predictors. Detected by VIF ≥ 10.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Define aliasing and how GLM software deals with it

A

Perfect linear dependency. Software drops one variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

2 limitations of GLMs

A
  1. GLMs give full credibility: The estimated coefficients are
    not credibility-weighted to recognize low volumes of data
    or high volatility. This concern can be partially addressed
    by looking at p-values or standard errors.
  2. GLMs assume that the randomness of outcomes are
    uncorrelated: Two examples of violations of this are:
    * Using a dataset with several renewals of the same policy,
    since the same insured over different renewals is likely to
    have correlated outcomes.
    * When the data can be affected by weather, the same
    weather events are likely to cause similar outcomes to
    risks in the same areas
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Steps of model-building process

A

Components of model-building process
1. Setting goals and objectives
2. Communication with key stakeholders
3. Collecting and processing the data
4. Conducting exploratory data analysis
5. Specifying the form of the model
6. Evaluating the model output
7. Validating the model
8. Translating the model results into a product
9. Maintaining and rebuilding the model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Considerations in merging policy and claim data

A

Match claims to vehicles/coverages, address timing differences, ensure unique keys, consider aggregation level (PY vs CY).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Considerations in Modifying the Data

A
  • Check for duplicate records and remove them
  • Check categorical field values against documentation (i.e.,
    are there code values not in the documentation, and are
    these new codes or errors?)
  • Check reasonability of numerical fields (e.g., negative
    premiums, significant outliers)
  • Decide how to handle errors and missing values (e.g., how
    much time to investigate, anything systematic about these
    records such as a specific location, maybe discard these
    records or replace the bad values with average values or an
    error flag)
  • Convert continuous variables into categorical (called
    binning)? Group levels in categorical variables? Combine
    or separate variables?
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Other possible data adjustments before modeling

A
  • Capping large losses
  • Removing cats or giving them less weight
  • Developing losses
  • On-leveling premiums for LR models
  • Trending exposures and losses
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Purpose of using a separate dataset for testing

A

Avoid overfitting by testing model on different data to measure true predictive performance.

17
Q

List 3 Model Testing Strategies

A
  1. Train/test split
  2. Train/validate/test
  3. Cross-validation (e.g., k-fold)
18
Q

Steps for k-fold cross-validation

A
  1. Split data into k folds
  2. Train on k−1 folds, test on 1
  3. Repeat
19
Q

Combine frequency and severity into pure premium

A

Multiply relativities (if both log-linked), or add linear predictors.

20
Q

2 disadvantages of modeling freq/sev separately

A

Takes more time, requires detailed data.

21
Q

Advantages of modeling freq/sev separately

A
  • Gaining more insight and intuition about the impact of each
    predictor variable.
  • Each of frequency and severity separately is more stable
    (e.g., a variable that only impacts frequency will look less
    significant in a pure premium model).
  • Pure premium modeling can lead to overfitting if a
    predictor variable only impacts frequency or severity but
    not both, since the randomness of the other component may
    be considered a signal effect.
  • The Tweedie distribution in a pure premium model
    assumes both frequency and severity move in the same
    direction, but this may not be true
22
Q

Steps to combine separate models by peril

A
  1. Model each peril
  2. Aggregate expected losses
  3. Build model on all-peril loss costs
23
Q

Criteria for variable inclusion

A

Statistical significance, data cost, legal/applicability, system constraints.

24
Q

Transformations after partial residual graph

A
  • Binning the variable: i.e., turning it into a categorical
    variable with separate “bins”. Downsides include that this
    increases the degrees of freedom of the model, it can result
    in inconsistent and/or impractical patterns, and variation
    within bins is ignored.
  • Adding polynomial terms: i.e., x2j , x3j , etc. Drawback is loss
    of interpretability without a graph.
  • Add piecewise linear functions: Add hinge functions
    max(0,xj − c) at each break point c. Drawback is break
    points must be manually chosen.
25
2 diagnostic test measures
Log-likelihood and Scaled Deviance (Adding more variables to a model always increases llmodel and reduces D∗ since there is more freedom to fit the data)
26
Conditions for comparing deviances
Same dataset and same distribution
27
3 ways to assess model stability
* The influence of an individual record on the model can be measured using the Cook’s distance, which can be calculated by most GLM software. Records with the highest Cook’s distance should be given additional scrutiny as to whether they should be included in the dataset or not. * Cross-validation can be used to assess model stability by comparing in-sample parameter estimates across different model runs. * Bootstrapping can be used to create new datasets with the same number of records by randomly sampling with replacement from the original dataset. The model can then be refit on many different datasets and we can get statistics like the mean and variance for each parameter estimate
28
2 reasons not to use refinement for selection
Models may be proprietary, selection may be business-driven.
29
3 cautions for actual vs. predicted plots
* Use holdout data (to prevent overfit) * It can help to aggregate data before plotting if the dataset is very large (e.g., into 100 buckets based on percentiles of predicted values) * Taking the log of all values before graphing prevents large values from skewing the picture
30
3 criteria to choose model from quantile plot
1. Predictive accuracy: Difference between actual and predicted in each quantile. 2. Monotonicity: The actual pure premium should consistently increase across quantiles. 3. Vertical distance of actual loss cost between first and last quantiles: This indicates how well the model distinguishes between the best and worst risks.
31
Sensitivity, specificity, false positive rate
Sensitivity = True positives / Total event occurrences Specificity = True negatives / Total event non-occurrences False positive rate = 1 - Specificit
32
Why coverage related variables should be first priced outside of GLMs
Coverage related variables (such as deductibles or limits) in GLMs can give counterintuitive results, such as indicating a lower rate for more coverage. This could be due to correlations with other variables outside of the model, including possible selection effects (e.g., insureds self-selecting to higher limits since they know they are higher risk, underwriters forcing high risk insureds to have higher deductibles). Charging rates for coverage options that reflect anything other than pure loss elimination could lead to changes in insured behavior, which means the indicated rates based on past experience would no longer be expected to be appropriate for new policies. As such, rates for coverage options should be estimated outside of the GLM first and included in the GLM as offset terms.Avoid counterintuitive results due to selection/correlation; include as offset.
33
Why price territories outside GLM
Territories are challenging in GLMs since there may be a very large number of territories, and aggregating them into a smaller number of groups may cause you to lose important information. Techniques like spatial smoothing can be used to price territories, and then territorial rates can be included in the GLM with the offset terms. However, the territory model should also be offset for the rest of the classification plan, so the process should be iterative until each model converges to an acceptable degree.
34
Why ensemble models can offer improved predictions
Different models will over-predict and under-predict for different segments of the book, but using an average of multiple models helps balance these predictions out for those segments. However, this really only works when the model errors are as uncorrelated as possible, which generally happens when models are built separately by different people with little or no sharing of information.
35
Problem and options for GLMs with highly correlated variables
This can result in an unstable model with erratic coefficients that have high standard errors. Two options for dealing with very high correlation include: 1. Removing all highly correlated variables except one. This eliminates the high correlation in the model, but it also potentially loses some unique information contained in the eliminated variables. 2. Use dimensionality-reduction techniques such as principal components analysis or factor analysis to create a new subset of variables from the correlated variables, and use this subset of variables in the GLM. The downside is the additional time required to do this extra analysis