L7: Time Series Forecating Flashcards

1
Q

What is the goal of time series forecating?

A

Predict future state of values based on historical values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is unique about time series forecasting vs. regression or classification?
A) Time dependency and sequential data rather than individual draws
B) Feature engineering: use lag values (past values) as features and time-based features
C) Different evaluation metrics: mean absolute error, root mean squared error etc.
D) Time-based cross validation or walk-forward validation
E) Interdependence between observations
F) Different data partitioning: test and training “periods” rather than “data”
G) Captures seasonality and trends

A

All options are true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

The longer the horizon, the larger the uncertainty of the forecast, leading to the benefit of forecast updating

TRUE/ FALSE

A

TRUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Time series ANALYSIS is descriptive as it describes the patterns in historical data.

Time series FORECASTING is predictive, as it predicts future states of values

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does residual error mean in times series forecasting?

A

Residual error: difference between forecast value vs. actual value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does “TREND” in times series mean?

A

Whether the series has an overall increasing/ decreasing trend

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does “SEASONALITY” in times series mean?

A

Whether any spikes repat itself at fixed intervals (seasonality) or non-fixed interval (cycle)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does “RANDOM NOISE” in times series mean?

A

Any form of observation left after accounting for trend and seasonality

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is global trend vs. local trends?

A

Local trend: Represents shorter-term fluctuations or patterns within data

Global trend: represents the overall long-term direction of data - captures the broad movement/ pattern

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Exponential Smoothing means that observations in the past are weighted depending on how far back they are in time (closer in time observations are assigned higher weights).

TRUE/FALSE

A

TRUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

The degree of weighting depends on the domain: whether it is fast-changing or fairly constant over time

TRUE/FALSE

A

TRUE

The degree of decay of weights back in time depends on how important newer data is

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Data partitioning in time series forecast: e.g., fixed partitioning implies that you use all past data as training dataset to forecast the future

TRUE/FALSE

A

TRUE

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are the common data pre-processing considerations in TSF? Select all correct
A) Missing values: some models, e.g., ARIMA cannot handle missing values
B) Unequally spaced series: equal intervals between datapoints is necessary, if not, imputation is required (e.g., setting N.A. to 0)
C) Choice of time span: older data is less informative, but important to have sufficient dataset size

A

ALL ARE CORRECT:

A) Missing values: some models, e.g., ARIMA cannot handle missing values
B) Unequally spaced series: equal intervals between datapoints is necessary, if not, imputation is required (e.g., setting N.A. to 0)
C) Choice of time span: older data is less informative, but important to have sufficient dataset size

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Sliding window is one partitioning technique in TSF. What does this mean?
A) Older data is dropped due to the assumption that newer data is more representative - smaller training set
B) Longer training periods are created as more data is revealed with to be incorporated into the model

A

A) Older data is dropped due to the assumption that newer data is more representative - smaller training set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Expanding window is one partitioning technique in TSF. What does this mean?
A) Older data is dropped due to the assumption that newer data is more representative - smaller training set
B) Longer training periods are created as more data is revealed with to be incorporated into the model

A

B) Longer training periods are created as more data is revealed with to be incorporated into the model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

The automatically generated Naive benchmark in TSF represents what value?

A

The latest observed value

17
Q

What does sNaive stand for?
It gives us the last seasonal value (i.e. we predict something in December, it will give us the last observation in December last year

A

Seasonal naive

18
Q

In TSF, you want to extract all information from the data, leaving behind only white noise.
To check if this is accomplished, you can use which measure?
A) autocorrelation function
B) autocorrection function
C) autocausation function

A

A) autocorrelation function
The higher the ACF, the more data is left on the table-

If most information has been extracted in the model, the ACF is very low, indicating low interdependency