Probabilistic algorithms Flashcards

1
Q

heuristics

A

t

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

approximation algorithm

A

t

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

randomized algorithm

A

t

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

stochastic

A

describes something that was randomly determined

randomly determined; having a random probability distribution or pattern that may be analyzed statistically but may not be predicted precisely.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Probability density function

A

a probability density function (PDF), or density of a continuous random variable, is a function, whose value at any given point in the sample space (the set of possible values taken by the random variable) can be interpreted as providing a relative likelihood that the value of the random variable would equal that sample

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

pivot

A

t

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Probabilistic Algorithm paradigm

A

for each input make a random choice and hope that it is good.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Non-Deterministic execution of algorithm

A

for the same input different sequences of
operations are possible

vs. for the same input the same sequence of
operations is executed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Probabilistic Algorithm

A

A Probabilistic (or Randomized) Algorithm is an algorithm which employs a degree of randomness as part of its logic

  • > During execution, it takes random choices depending on those random numbers
  • > The behavior (output) can vary if the algorithm is run multiple times on the same input

=> probabilistic algorithms are often simpler and faster than their deterministic counterparts
=> however, in the worst execution case, the algorithm may be very slow

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

positive vs negative probability

A

??

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

probabilistic algorithm

A

+ gain execution time

- loose correctness of answer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Las Vegas algorithms

A

a randomized algorithm that

  • always gives correct/optimal results
  • the execution time is variable
  • example: RandQSORT
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Monte Carlo algorithms

A

a randomized algorithm that

  • the running time is deterministic
  • the output may be incorrect with a certain probability
  • example: Check matrices equality
  • for Decision problem (Yes/No): one- or two-sided error
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

one-sided error

A

Monte Carlo algorithm: Yes or No always correct

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

two-sided error

A

Monte Carlo algorithm: a non-zero probability to err for both outputs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

E[X]

A

Expected value of discrete X: E[X]

The expected value of random variable X is often written as E(X)

sl. 15 lecture_01

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Random number generation

A
  • the kernel of Monte Carlo simulation
  • the heart of many standard statistical methods (bootstrap, Bayesian analysis)
  • cryptography
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

bisection method

A

d

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

sucessive quadratic estimation method

A

d

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Powell’s algorithm

A

d

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

inflection point

A

An inflection point is a point on a curve at which the sign of the curvature (i.e., the concavity) changes. Inflection points may be stationary points, but are not local maxima or local minima.

http://mathworld.wolfram.com/InflectionPoint.html

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

first order condition

A

The first order condition means that the local maximum or local minimum of some continuous function within some range (not including the end points) must occur where the derivative of that function is equal to 0.

  • > At the highest and lowest points of a curve, the tangent to the curve at such points is horizontal.
  • > The slope of the curve is zero.

http://users.etown.edu/p/pauls/ec309/lectures/lec04_unconst.html

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

second order condition

A

SOC for a local max is that f’‘(x) < 0

SOC for a local min is that f’‘(x) > 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

loss function

A

a function that maps an event or values of one or more variables onto a real number intuitively representing some “cost” associated with the event.

An optimization problem seeks to minimize a loss function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Hessian matrix?

A

useful in characterizing the shape of L and to

distinguish, among the zeros of gradient g(d), the minimum points, the maximum points and the inflexion points

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

why an inverse matrix?

A

we can multiply by an Inverse, which achieves the same thing as “dividing by a number”

remember: with Matrices the order of multiplication matters. AB is almost never equal to BA.
https: //www.mathsisfun.com/algebra/matrix-inverse.html

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Jacobian matrix

A

is the matrix of all first-order partial derivatives of a vector-valued function. When the matrix is a square matrix, both the matrix and its determinant are referred to as the Jacobian in literature

The Jacobian matrix is important because if the function f is differentiable at a point x (this is a slightly stronger condition than merely requiring that all partial derivatives exist there), then the Jacobian matrix defines a linear map ℝ^n → ℝ^m, which is the best (pointwise) linear approximation of the function f near the point x. This linear map is thus the generalization of the usual notion of derivative, and is called the derivative or the differential of f at x.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Differenzierbarkeit / differentiable

A

Als Differenzierbarkeit bezeichnet man in der Mathematik die Eigenschaft einer Funktion, sich lokal um einen Punkt in eindeutiger Weise linear approximieren zu lassen.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

TSP

A

find a Hamiltonian cycle with minimal path length in G

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

TSP extension

A
  • Asymmetric TSP
  • optimum configuration for the ATSP
  • time dependent TSP
  • Vehicle Routing Problem
  • ## probabilistic VRP
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

Branch & Bound

A

an algorithm design paradigm for discrete and combinatorial optimization problems, as well as mathematical optimization. A branch-and-bound algorithm consists of a systematic enumeration of candidate solutions by means of state space search: the set of candidate solutions is thought of as forming a rooted tree with the full set at the root. The algorithm explores branches of this tree, which represent subsets of the solution set. Before enumerating the candidate solutions of a branch, the branch is checked against upper and lower estimated bounds on the optimal solution, and is discarded if it cannot produce a better solution than the best one found so far by the algorithm.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

Branch & Bound TSP

A

in B & B we try to reduce the search space:

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

linear programming

A

LP, also called linear optimization

best outcome of a mathematical model (provided constraints)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

convex set

A

a convex set is a subset of an affine space that is closed under convex combinations.

in a Euclidean space, a convex region is a region where, for every pair of points within the region, every point on the straight line segment that joins the pair of points is also within the region

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

Convex polytope

A

a special case of a polytope, having the additional property that it is also a convex set of points in the n-dimensional space R^n

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

Steepest Descent algorithm

A

tbd

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

Newton-Raphson

A

tbd

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

Branch & Bound

A

tbd

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

Localized Random Search

A

tbd

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

Direct Random Search

A

tbd

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

convex

A

convex describes a surface that curves outward, or is thicker in the middle than on the edges.

https://writingexplained.org/concave-vs-convex-difference

42
Q

concave

A

Concave is an adjective that describes a surface that curves inward, or is thinner in the middle than on the edges.

43
Q

hypercube

A

a hypercube is an n-dimensional analogue of a square (n = 2) and a cube (n = 3)

44
Q

multivariate

A

involving two or more variable quantities.

45
Q

multivariate normal distribution

A

is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions

One definition is that a random vector is said to be k-variate normally distributed if every linear combination of its k components has a univariate normal distribution

46
Q

Enhanced Localized Random Search

A

tbd

47
Q

Simple Random (“Blind”) Search

A

tbd

48
Q

Direct methods for stochastic search

A

enlist…

49
Q

Pattern search

A

describe…

50
Q

Random search with noise-free measurements

A

describe…

51
Q

slack variable

A

a slack variable is a variable that is added to an inequality constraint to transform it into an equality

Introducing a slack variable replaces an inequality constraint with an equality constraint and a non negativity constraint

By introducing the slack variable y >= 0, the inequality Ax <= b can be converted to the equation Ax + y = b

52
Q

n-simplex

A

is a generalization of the notion of a triangle to arbitrary dimensions

53
Q

Basic Pattern Search Algorithm

A

t

54
Q

Compass search

A

t

55
Q

Nonlinear Simplex (Nelder-Mead) Algorithm

A

t

  • derivative-free optimization algorithm -> derivatives are unavailable or expensive to compute
  • > derivative-free method
  • > recommended for solving optimization problems with noisy objective functions
56
Q

convex hull

A

a polygon which is the smallest perimeter fence enclosing a set of N points

57
Q

Constraints

A

limitation of the allowed values of the degree of
freedom of the system

e.g. boundaries (localisation constraints)
e.g. deadline for event occurrence (temporal
constraints)

hard & soft constraints

58
Q

penalty function

A

a solution which violates a constraint is not rejected, but “punished”

59
Q

heuristics

A

is any approach to problem solving, learning, or discovery that employs a practical method not guaranteed to be optimal or perfect, but sufficient for the immediate goals.

60
Q

Heuristic optimization algorithms

A

Two natural ways to design an heuristic algorithm:

  • Construction heuristics
  • Markovian improvement heuristics
61
Q

configuration

A

t

62
Q

construction heuristics vs. improvement heuristics

A

t

63
Q

Insertion heuristic

A

the new node is inserted somewhere inside the sequence (not appended!)

  • How to choose the new node?
  • Where is the optimum place to insert it?
64
Q

Construction heuristics for TSP

A

e.g. Nearest Neighbor Heuristic

65
Q

Best insertion heuristic

A

you insert the next randomly chosen node between those two nodes where the increase in the total length is minimum
until all nodes are visited.

66
Q

Nearest Neighbor Heuristic

A

choose a random node. then choose a random node which is closest to the first node. Then choose a random node which is closest to the second node and so on… (the new node is appended!)

67
Q

Best-Best insertion / cheapest insertion heuristic

A

t

68
Q

Worst-best insertion / farthest insertion algorithm

A

t

69
Q

nearest insertion heuristic

A

t

70
Q

Shortest edge heuristic

A

t

71
Q

Ruin & Recreate

A

t

72
Q

Ruin & Recreate for TSP

A

t

73
Q

energy landscape

A

The values of loss function + the search space form the energy landscape of the system

74
Q

improvement heuristics

A

t

75
Q

Markov improvement heuristic

A

t

76
Q

local search approach

A

?

77
Q

Swap move algorithm TSP

A

t

78
Q

Translation move algorithm

A

t

79
Q

Inversion move algorithm

A

t

80
Q

Improvement heuristics for TSP

A

Swap move algorithm TSP
Translation move algorithm
Inversion move algorithm

81
Q

Greedy algorithm for TSP

A

t

82
Q

Simulated Annealing

A

t

83
Q

TA is considered as a “first approximation” of SA

A

much faster than SA (no need to generate delta uniformly on (0,1))

84
Q

i.i.d.

A

independent and identically distributed

a sequence or other collection of random variables is independent and identically distributed if each random variable has the same probability distribution as the others and all are mutually independent

85
Q

Noise:

A

independent, identically distributed (i.i.d.) measurement

errors

86
Q

Steepest Descent with Noisy/Noise-Free Gradient

A

_09

87
Q

Random Search Algorithms with Noisy Loss Function

A

Basic implementation of random search assumes perfect (noise-free)
values of L

88
Q

Statistical and Engineering Convergence Conditions

A

_09

89
Q

Stochastic Approximation

A

_09 ???

90
Q

stochastic gradient

A

_09 sl. 31

91
Q

Stochastic Gradient Measurement and Algorithm (SGA=

A

_09

92
Q

Gradient-Free Algorithms

A

_09

93
Q

Finite-Difference Algorithm

A

_09

94
Q

FDSA algorithm

A

_09

95
Q

gain value?

A

_09

96
Q

Simultaneous Perturbation Stochastic Approximation

Algorithm (SPSA)

A

_09

97
Q

gain sequence

A

the sequence of a_k values used in the calculation of d^_k+1, as a factor to Y_k(d_k)

98
Q

d*

A

d* e D* -> a global optimal solution
D* is the solution space, the set of all solutions of an optimization problem

d^ e D (problem space) is a solution candidate

99
Q

The loss function L()

The gradient g(d)

A

The gradient g(d) of L(d) is the vector of first order derivatives

100
Q

SP

A

Simultaneous Perturbation