Escolar Documentos
Profissional Documentos
Cultura Documentos
January 8, 2018
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 1 / 60
Contents
2 Fundamentals of Optimization
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 2 / 60
Basic Theory of Sets and Functions
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 2 / 60
Sets and Functions
Topics
Sets
Sequences
Mappings & Functions
Continuous Functions
Infimum and Supremum of Functions
Maxima and Minima of Functions
Differentiable Functions
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 3 / 60
Sets
Introduction
Sets
A set A is a collection of any kind which are called the elements or points of A.
The set of real numbers is denoted by R.
If a is an element of set A, we write as a ∈ A, if a is not an element of A, it is
represented as, a ∈/A
Eg: The set of all non-negative real numbers,
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 4 / 60
Sets
Introduction
Sets
A set A is a collection of any kind which are called the elements or points of A.
The set of real numbers is denoted by R.
If a is an element of set A, we write as a ∈ A, if a is not an element of A, it is
represented as, a ∈/A
Eg: The set of all non-negative real numbers,
A = {x|x ∈ R, x ≥ 0}.
If A and B are two sets and all elements of A are elements of B, we write A ⊂ B,
that is, A is contained in B or that B contains A and write B ⊃ A.
Then, A is called a subset of B and B is called the super-set of A.
If A ⊂ B and B ⊂ A, then A = B.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 4 / 60
Sets
Operations on Sets
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 5 / 60
Sets
Operations on Sets
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 6 / 60
Sets
Operations on Sets
A lower bound ᾱ is the greatest lower bound or infimum of S, (i.e. inf {x|x ∈ S}) if
no number greater than ᾱ is a lower bound of S.
An upper bound β̄ is the least upper bound or supremum of S, (i.e. sup{x|x ∈ S}),
if no number smaller than β̄ is an upper bound of S.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 7 / 60
Sets
Topological Properties of Rn
Topological Properties of Rn
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 8 / 60
Sets
Topological Properties of Rn
A point is said to be interior point of the set S ⊂ Rn , if there exists an > 0 such
that N (X) ⊂ S.
The set of all interior points of S is called the interior of S and is denoted by int.S.
Obviously, int.S ⊂ S.
S is called open if S = intS, that is if every point of S is an interior point.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 9 / 60
Sets
Topological Properties of Rn
Boundary Points
Bounded Sets
A set S ⊂ Rn is said to be bounded if there exists a real number α > 0 such that
kXk < α, for each X ∈ S.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 10 / 60
Sets
Sequences
Sequences
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 11 / 60
Sets
Sequences
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 12 / 60
Sets
Sequences
Limit of Sequence
lim Xn = X̄
n→∞
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 13 / 60
Sets
Sequences
Caushy Sequence
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 14 / 60
Sets
Mappings & Functions
Mappings
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 15 / 60
Sets
Mappings & Functions
If the mapping Γ from a set X into a set Y is such that the image set Γ(x) always
consists of a single element, then, Γ is called single valued mapping of single
valued function.
Multi-valued Functions
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 16 / 60
Sets
Mappings & Functions
Continuous Functions
A function for which sufficiently small changes in the input result in arbitrarily small
changes in the output.
Let S be a non-empty set in Rn . A function f : S → R is said to be continuous at
X̄ ∈ S, if for any given > 0, there exists a δ > 0 such that,
X ∈ S, kX − X̄k < δ ⇒ |f (X) − f (X̄)| < .
Equivalently,
.
A vector valued function is said to be continuous at X̄, if each of its components is
continuous at X̄.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 17 / 60
Sets
Mappings & Functions
Discontinuous Functions
Many functions, however, will have isolated points where they are not connected.
These problem points are called discontinuities.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 18 / 60
Sets
Infimum and Supremum of Functions
Bounded Functions
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 19 / 60
Sets
Infimum and Supremum of Functions
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 20 / 60
Sets
Maxima and Minima of Functions
Minima of Functions
Let f be a real single valued function defined on the set S. If there is point X0 ∈ S
such that f (X0 ) ≤ f (X), ∀X ∈ S is called the minimum of S and is written as,
Maxima of Functions
Let f be a real single valued function defined on the set S. If there is point X ∗ ∈ S
such that f (X ∗ ) ≥ f (X), ∀X ∈ S is called the minimum of S and is written as,
f (X ∗ ) = max f (X)
X∈S
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 22 / 60
Sets
Differentiable Functions
Differentiable Functions
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 23 / 60
Matrices and Determinants
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 24 / 60
Fundamentals of Optimization
Fundamentals of Optimization
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 24 / 60
Fundamentals of Optimization
Topics
Fundamentals of Optimization
Ref: Igor Griva, Stephen G. Nash, Ariela Sofer, "Linear and Nonlinear Optimization",
SIAM, Second Edn., 2009.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 25 / 60
Fundamentals of Optimization
Feasibility and Optimality
Feasibility
where,
{gi } are given functions that define the constraints in the model
E is an index set for the equality constraints
I is an index set for the inequality constraints
Any set of equations and inequalities can be rearranged in this form.
Examples
g1 (x) = 3x21 + 2x2 − 3x3 + 9 = 0
g2 (x) = −sinx1 + cosx2 ≥ 0
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 26 / 60
Fundamentals of Optimization
Feasibility and Optimality
Feasibility
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 27 / 60
Fundamentals of Optimization
Feasibility and Optimality
Feasibility
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 28 / 60
Fundamentals of Optimization
Feasibility and Optimality
Feasibility
At the feasible point xa = (0, 0, 2)T , the first two inequality constraints x1 ≥ 0 and
x2 ≥ 0 are active, while the third is inactive.
At the point xb = (3, 0, 1)T only the second inequality is active.
While at the interior point xc = (1, 1, 1)T none of the inequalities are active.
The boundary of the feasible region is indicated by bold lines.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 29 / 60
Fundamentals of Optimization
Feasibility and Optimality
Optimality
It may seem surprising that there is any question about what is meant by a "solu-
tion" to an optimization problem.
The confusion arises because there are a variety of conditions associated with an
optimal point and each of these conditions gives rise to a slightly different notion of
a "solution."
Let us consider the n-dimensional problem;
min f (x)
x∈S
.
There is no fundamental difference between minimization and maximization prob-
lems. We can maximize f by solving
min −f (x)
x∈S
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 30 / 60
Fundamentals of Optimization
Feasibility and Optimality
Optimality
f (x∗ ) ≤ f (x), ∀x ∈ S
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 31 / 60
Fundamentals of Optimization
Feasibility and Optimality
Optimality
If we cannot find the global solution, then at the least we would like to find a point
that is better than its surrounding points.
More precisely, we would like to find a local minimizer of f in S, a point satisfying,
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 32 / 60
Fundamentals of Optimization
Convexity
Convexity
There is one important case where global solutions can be found, the case where
the objective function is a convex function and the feasible region is a convex set.
A set S is convex if, for any elements x and y of S,
αx + (1 − α)y ∈ S, ∀, 0 ≤ α ≤ 1
In other words, if x and y are in S, then the line segment connecting x and y is also
in S.
More generally, every set defined by a system of linear constraints is a convex set.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 33 / 60
Fundamentals of Optimization
Convexity
Convexity
∀, 0 ≤ α ≤ 1 and ∀, x, y ∈ S.
This definition says that the line segment connecting the points (x, f (x)) and (y, f (y))
lies on or above the graph of the function
Intuitively, the graph of the function is bowl shaped.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 34 / 60
Fundamentals of Optimization
Convexity
Convexity
∀, 0 ≤ α ≤ 1 and ∀, x, y ∈ S.
We say that a function is strictly convex if,
min f (x)
x∈S
min f (x)
subject to gi (x) ≥ 0, i = 1, ..., m,
is a convex optimization problem if f is convex and the functions {gi } are concave.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 35 / 60
Fundamentals of Optimization
Convexity
Convexity
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 36 / 60
Fundamentals of Optimization
Convexity
Convexity
k
X
y= α i xi
i=1
where,
k
X
αi = 1
i=1
and αi ≥ 0, i = 1, ..., k
There will normally be many ways in which y can be expressed as a convex com-
bination of {xi }.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 37 / 60
Fundamentals of Optimization
Convexity
f ”(x) ≥ 0, ∀x ∈ S
Examples
f (x) = x4 → Convex function
f (x) = sinx → neither convex nor concave
In the multidimensional case, the Hessian matrix of second derivatives must be
positive semidefinite; that is, at every point x ∈ S,
y T ∇2 f (x)y ≥ 0, ∀y
In the multidimensional case, if the Hessian matrix ∇2 f (x) is positive definite for
all x ∈ S, then the function is strictly convex on S.
Alternatively, it would have been possible to show that the eigenvalues of the Hes-
sian matrix were all greater than or equal to zero.
Eg: f (x1, x2) = 4x21 + 12x1 x2 + 9x22
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 38 / 60
Fundamentals of Optimization
Convexity
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 39 / 60
Fundamentals of Optimization
Convexity
Function f with one continuous derivative is convex over a convex set S if and only
if it satisfies,
f (y) ≥ f (x) + ∇f (x)T (y − x), ∀x, y ∈ S
Proof is self study
This property states that the function is on or above any of its tangents.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 40 / 60
Fundamentals of Optimization
The General Optimization Algorithm
min f (x)
then, the optimality test will often be based on the condition f 0 (x) = 0
If f 0 (xk ) 6= 0, then xk is not optimal, and the sign and value of f 0 (xk ) indicates
whether f is increasing or decreasing at the point xk , as well as how rapidly f is
changing.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 41 / 60
Fundamentals of Optimization
The General Optimization Algorithm
In this algorithm, pk is a search direction that we hope points in the general direction
of the solution, or that "improves" our solution in some sense.
The scalar αk is a step length that determines the point xk+1 .
Once the search direction pk has been computed, the step length αk is found by
solving some auxiliary one-dimensional problem
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 42 / 60
Fundamentals of Optimization
The General Optimization Algorithm
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 43 / 60
Fundamentals of Optimization
The General Optimization Algorithm
For an unconstrained problem of the form, we will typically require that the search
direction pk be a descent direction for the function f at the point xk .
This means that for "small" steps taken along pk the function value is guaranteed
to decrease for some :
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 44 / 60
Fundamentals of Optimization
The General Optimization Algorithm
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 45 / 60
Fundamentals of Optimization
The General Optimization Algorithm
min
x∈S
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 46 / 60
Fundamentals of Optimization
Rates of Convergence
Rates of Convergence
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 47 / 60
Fundamentals of Optimization
Rates of Convergence
Rates of Convergence
For many optimization methods, the number of operations or steps required to find
an exact solution will be infinite.
The rate of convergence describes how quickly the estimates of the solution ap-
proach the exact solution.
Let us assume that we have a sequence of points xk converging to a solution x∗ .
We define the sequence of errors to be,
We say that the sequence {xk } converges to x∗ with rate r and rate constant C if
kek+1 k
lim =C
k→∞ kek kr
where, C < ∞
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 48 / 60
Fundamentals of Optimization
Rates of Convergence
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 49 / 60
Fundamentals of Optimization
Rates of Convergence
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 50 / 60
Fundamentals of Optimization
Rates of Convergence
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 51 / 60
Fundamentals of Optimization
Taylor Series
Taylor Series
The Taylor series is a tool for approximating a function f near a specified point x0 .
It has many uses:
It allows you to estimate the value of the function near the given point (when the function
is difficult to evaluate directly).
The derivatives and integral of the approximation can be used to estimate the derivatives
and integral of the original function.
It is used to derive many algorithms for finding zeroes of functions, for minimizing func-
tions, etc.
Let x0 be a specified point. Then the nth order Taylor series approximation is,
1 2 pn (n)
f (x0 + p) ≈ f (x0 ) + pf 0 (x0 ) + p f ”(x0 ) + ... + f (x0 )
2 n!
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 52 / 60
Fundamentals of Optimization
Taylor Series
Taylor Series
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 53 / 60
Fundamentals of Optimization
Taylor Series
Taylor Series
The first two terms of the Taylor series give us the formula for the tangent line for
the function f at the point x0 .
y = f (x0 ) + (x − x0 )f 0 (x0 ).
where p = x − x0 .
The first three terms of the Taylor series give a quadratic approximation to the
function f at the point x0 .
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 54 / 60
Fundamentals of Optimization
Taylor Series
Taylor Series
For n variables,
1 T 2
f (x0 + p) ≈ f (x0 ) + pT ∇f (x0 ) + p ∇ f (x0 ) + ...
2
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 55 / 60
Fundamentals of Optimization
Taylor Series
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 56 / 60
Fundamentals of Optimization
Newton’s Method for Nonlinear Equations
f (x) = 0
for p to obtain
f (xk )
p=−
f 0 (xk )
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 57 / 60
Fundamentals of Optimization
Newton’s Method for Nonlinear Equations
The new estimate of the solution is then xk+1 = xk +p or xk+1 = xk −f (xk )/f 0 (xk ).
This is the formula for Newton’s method
Newton’s method corresponds to approximating the function f by its tangent line at
the point xk .
The point where the tangent line crosses the x-axis (i.e., a zero of the tangent line)
is taken as the new estimate of the solution.
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 58 / 60
Fundamentals of Optimization
Newton’s Method for Nonlinear Equations
Theorem (Convergence of Newton’s Method). Assume that the function f (x) has
two continuous derivatives. Let x∗ be a zero of f with f 0 (x∗ ) 6= 0. If |x0 − x∗ | is
sufficiently small, then the sequence defined by.
f ”(x∗ )
C=| |
2f 0 (x∗ )
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 59 / 60
Fundamentals of Optimization
Newton’s Method for Nonlinear Equations
Much of the discussion in the one-dimensional case can be transferred with only
minor changes to the n-dimensional case.
Suppose that we are solving f (x) = 0, which represents,
f1 (x1 , ..., xn ) = 0
f2 (x1 , ..., xn ) = 0
...
...
fn (x1 , ..., xn ) = 0
Dr. Sudhish N. George NIT Calicut EC6422, Linear and Nonlinear Optimization January 8, 2018 60 / 60