Information Theory Flashcards

1
Q

log(A) + log(B)

A

log(AB)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Entropy for N equiprobable events

A

Entropy = E[-log p(x)] = - sum_{over classes} p(x_c) log p (x_c)

For equiprobable events, p(x_c) = 1/C, so
entropy = log(C)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Is entropy ever negative?

A

No

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

When is entropy additive?

A

For independent events

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Interpretation of -log p_i

A

of Bits required to represent ith symbol efficiently

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Shannon bit

A

0 or 1, but not both simultaneously (as opposed to qubit)
1 bit = amount of info gained
if we have apriori info about which event is more probable, then amount of info gain < 1 bit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Shanon Coding

A

Minimize the bits required to represent message
Done by assigning shorter codes to symbols with higher probabilities
Mean Length = sum(p_x log_k (p_x)), where k = arity of code e.g., 2 for binary and p_x is probability for symbol x
Coding must be efficient but also decodable (unique symbols)
Here, log_k(p_x) is the bits of information and p_x is the proba of x happening

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Convert base of log

A

log_c(a) = log_b(a)/log_b(c)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Interpretation of cross-entropy

A
  • sum_x p_x log(q_x) where p is the true distribution and q is the predicted distribution
    Avg bits required to transmit p if q is used instead
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Interpretation of KL divergence

A

KLD(P||Q) = H(P, Q) - H(P)
i.e., cross-entropy - entropy (not symmetric)
i.e., additional bits required to transmit p since we are using the predicted distribution q instead of true distribution p

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Is KLD a distance measure? Why/Why not?

A

No, since it is not symmetric

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Mutual Information

A

I(X, Y) = H(X) + H(Y) - H(X, Y)
I(X, Y) = H(Y) - H(Y|X)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly