Week 10 - Kalman Filter, Recurrent Neural Networks Flashcards
(20 cards)
Define Kalman Filter
It is similar to HMMs, but it is for Kalman filter is for continuous hidden state (๐๐)
What does each hidden state (๐๐) depend on?
It depends on the one before it in time ๐๐กโ1, for all ๐ก = 1, 2, โฆ , ๐
What are the observations depends on in Kalman filter?
It only depend on the associated hidden state (๐๐)
What is Kalman filter often used for?
It is often used for target tracking or motion smoothing with noisy observations applications in automated airplane or ship guidance
What is the linear- Gaussian recurrence relation?
LGRR is the simplest model. The form is:
What is the main focus of reinforcement learning (RL) in the context of decision-making?
RL focuses on finding optimal policies to maximize rewards by making decisions based on observed effects of those choices, rather than predicting sequences from observed data.
How can dynamic programming be applied in reinforcement learning problems?
Dynamic programming is used in reinforcement learning when the problem is modeled as a Markov Decision Process (MDP), a special kind of sequential graphical model.
What components define the progression in a Markov Decision Process (MDP)?
An MDP consists of a sequence of input actions Zt that lead to observed states St where each state has an associated deterministic reward R(St).
How are rewards and utility defined in a Markov Decision Process?
Each state St has a deterministic reward R(St). The utility U of a sequence of states (S1,โฆ,St) is also a deterministic function of these states
What limitation do standard deep networks have when dealing with sequential data?
Standard deep networks do not take the ordering of data into account.
How do Recurrent Neural Networks (RNNs) differ from standard deep networks?
It treats the hidden layer from the previous stage as an additional input to the current stage, enabling them to process sequential data
What role does the hidden state Zt-1 play in RNNs?
The value of the previous hidden state Zt-1 is included in computing the current hidden state Zt, allowing the network to retain memory of past inputs.
How do RNNS handle the order of inputs?
RNNS explicitly incorporate order dependence into their structure, allowing them to model temporal or sequential relationships in data.
What an example of RNN?
Stock prediction
What is the goal of using an RNN for stock prediction?
The goal is to predict the next dayโs stock price given the prices from the previous N days (e.g N = 3)
What are the inputs and key parameters in an RNN for stock prediction?
Input are daily stock prices (plus a bias). The parameters include weight w0,w1, recurrent weight u1, and output weight v0,v1, which are shared across time steps.
How are the hidden and output values computed in a stock prediction RNN?
Each hidden state ๐ง๐ก = ๐(๐ค0 + ๐ค1๐ฅ๐ก + ๐ข1๐ง๐กโ1), and the output y = ๐(v0 + v1Zt, using the same weights across time steps
What is a major challenge when training RNNs and what are the solutions?
RNNs suffer from the vanishing or exploding gradient problem, Solution include LSTM networks, and Tranformers