Week 10 - Kalman Filter, Recurrent Neural Networks Flashcards

(20 cards)

1
Q

Define Kalman Filter

A

It is similar to HMMs, but it is for Kalman filter is for continuous hidden state (๐’๐’•)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does each hidden state (๐’๐’•) depend on?

A

It depends on the one before it in time ๐‘๐‘กโˆ’1, for all ๐‘ก = 1, 2, โ€ฆ , ๐‘‡

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the observations depends on in Kalman filter?

A

It only depend on the associated hidden state (๐’๐’•)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Kalman filter often used for?

A

It is often used for target tracking or motion smoothing with noisy observations applications in automated airplane or ship guidance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the linear- Gaussian recurrence relation?

A

LGRR is the simplest model. The form is:

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the main focus of reinforcement learning (RL) in the context of decision-making?

A

RL focuses on finding optimal policies to maximize rewards by making decisions based on observed effects of those choices, rather than predicting sequences from observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How can dynamic programming be applied in reinforcement learning problems?

A

Dynamic programming is used in reinforcement learning when the problem is modeled as a Markov Decision Process (MDP), a special kind of sequential graphical model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What components define the progression in a Markov Decision Process (MDP)?

A

An MDP consists of a sequence of input actions Zt that lead to observed states St where each state has an associated deterministic reward R(St).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How are rewards and utility defined in a Markov Decision Process?

A

Each state St has a deterministic reward R(St). The utility U of a sequence of states (S1,โ€ฆ,St) is also a deterministic function of these states

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What limitation do standard deep networks have when dealing with sequential data?

A

Standard deep networks do not take the ordering of data into account.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How do Recurrent Neural Networks (RNNs) differ from standard deep networks?

A

It treats the hidden layer from the previous stage as an additional input to the current stage, enabling them to process sequential data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What role does the hidden state Zt-1 play in RNNs?

A

The value of the previous hidden state Zt-1 is included in computing the current hidden state Zt, allowing the network to retain memory of past inputs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How do RNNS handle the order of inputs?

A

RNNS explicitly incorporate order dependence into their structure, allowing them to model temporal or sequential relationships in data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What an example of RNN?

A

Stock prediction

17
Q

What is the goal of using an RNN for stock prediction?

A

The goal is to predict the next dayโ€™s stock price given the prices from the previous N days (e.g N = 3)

18
Q

What are the inputs and key parameters in an RNN for stock prediction?

A

Input are daily stock prices (plus a bias). The parameters include weight w0,w1, recurrent weight u1, and output weight v0,v1, which are shared across time steps.

19
Q

How are the hidden and output values computed in a stock prediction RNN?

A

Each hidden state ๐‘ง๐‘ก = ๐‘“(๐‘ค0 + ๐‘ค1๐‘ฅ๐‘ก + ๐‘ข1๐‘ง๐‘กโˆ’1), and the output y = ๐‘“(v0 + v1Zt, using the same weights across time steps

20
Q

What is a major challenge when training RNNs and what are the solutions?

A

RNNs suffer from the vanishing or exploding gradient problem, Solution include LSTM networks, and Tranformers