00 Summaries Flashcards

1
Q

Reinforcement learning

A

Reinforcement learning: How an agent can become proficient in an unknown environment, given only its percepts and occasional rewards

Utilities can be learned using:

  • Direct utility estimation
  • Adaptive dynamic programming (ADP)
  • Temporal-difference (TD)

Three overall agent designs:

  • Model-based design using a model and a utility function
  • Model-free design, using an action-utility function Q
    • Q-unfctions can be learning by ADP or TD
  • Reflex design, using a policy π

Value of information is relevant for active learning

Approximate functions are required for large state spaces

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Natural Language Processing

A

Natural Language Understatning
- Require empirical investigation of actual human behavior

Formal language theory
Phrase structure grammars
(Lexicalized) Probabilistic Context-free grammar (PCFG) formalism
Treebank
Augmented grammar for semantic interpretation
Parsing
Machine translation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Learning from Examples

A

Learning from examples: Inductive learning of functions from examples. Inductive learning involves finding a hypothesis that agrees well with the examples.

Ockham’s razor (choose the simples consistent hypothesis)
Supervised learning
Classification (discrete-valued function)
Regression (continuous function)
Decision trees (with information-gain heuristic)
Perceptron
- Trained by simple weight update rule
Neural networks
- Represent complex nonlinear functions with a network of linear-treshold units.
- Multilayer feed-forward neural networks can represent any function
- Back-propagation algorithm implements a gradient descent in paramater space to minimize the output error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Making Simple Decisions

A

Decision theory (what should an agent do) = Probability theory (what should an agent believe) + Utility theory (what does an agent want)

MEU
Decision networks (extension of Bayesian networks)
Value of information: Expected improvement in utility compared with making a decision without the information
Decision-theoretic expert systems

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Probabilistic Reasoning

A

Probabilistic Reasoning: Bayesian networks

- a representation for uncertain knowledge and conditional independence relationships

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Quantifying Uncertainty

A

Uncertainty arise because of laziness and ignorance
Probabilities express the agent’s inability to reach a definite decision regarding the truth of a sentence
Decision tehory
Prior probabilities and conditional probabilities
Full joint porobability distribution
Bayes’ rule
Conditional independence
Naive Bayes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Inference in FOL

A
  • Inference using rules (UI, EI) to propositionalize the problem
  • Unification
    Generalized Modus Ponens (applied by forward- and backward-chaining)
    Resolution provides a complete proof system
How well did you know this?
1
Not at all
2
3
4
5
Perfectly