hw5: done 1.1, 2, 3

This commit is contained in:
Claudio Maggioni 2021-06-03 15:23:37 +02:00
parent 259751dc3e
commit 96d5e805a0
4 changed files with 126 additions and 13 deletions

View file

@ -16,6 +16,54 @@ header-includes:
--- ---
\maketitle \maketitle
# Excecise 1
## Exercise 1.1
### The Simplex method
The simplex method solves constrained minimization problems with a linear
cost function and linearly-defined equality and inequality constraints. The main
approach used by the simplex method is to consider only basic feasible points
along the feasible region polytope and to iteratively navigate between them
hopping through neighbours and trying to find the point that minimizes the cost
function.
Although the Simplex method is relatively efficient for most in-practice
applications, it has exponential complexity, since it has been proven that
a carefully crafted $n$-dimensional problem can have up to $2^n$ polytope
vertices, thus making the method inefficient for complex problems.
### Interior-point method
The interior point method aims to have a better worst-case complexity than the
simplex method but still retain an in-practice acceptable performance. Instead
of performing many inexpensive iterations walking along the polytope boundary,
the interior point takes Newton-like steps travelling along "interior" points in
the feasible region (hence the name of the method), thus reaching the
constrained minimizer in fewer iterations. Additionally, the interior-point
method is easier to be implemented in a parallelized fashion.
### Penalty method
The penalty method allows for a linear constrained minimization problem with
equality constraints to be converted in an unconstrained minimization problem,
and to allow the use of conventional unconstrained minimization algorithms to
solve the problem. Namely, the penalty method builds a new uncostrained
objective function with is the summation of:
- The original objective function;
- An additional term for each constraint, which is positive when the current
point $x$ violates that constraint and zero otherwise.
With some fine tuning of the coefficients for these new "penalty" terms, it is
possible to build an equivalend unconstrained minimization problem whose
minimizer is also constrained minimizer for the original problem.
## Exercise 1.2
## Exercise 1.3
# Exercise 2 # Exercise 2
## Exercise 2.1 ## Exercise 2.1
@ -55,7 +103,7 @@ index set $1, 2, \ldots, n$ such that:
indices in $\beta$ are linearly independent from each other. indices in $\beta$ are linearly independent from each other.
The geometric interpretation of basic feasible points is that all of them The geometric interpretation of basic feasible points is that all of them
are verticies of the polytope that bounds the feasible region. We will use this are vertices of the polytope that bounds the feasible region. We will use this
proven property to manually solve the constrained minimization problem presented proven property to manually solve the constrained minimization problem presented
in this section by aiding us with the graphical plot of the feasible region in in this section by aiding us with the graphical plot of the feasible region in
figure \ref{fig:a}. figure \ref{fig:a}.
@ -85,7 +133,7 @@ Figure 1 taken from the book.-->
Since the geometrical interpretation of the definition of basic feasible point Since the geometrical interpretation of the definition of basic feasible point
states that these point are non-other than the vertices of the feasible region, states that these point are non-other than the vertices of the feasible region,
we first look at the plot above and to these points (i.e. the verticies of the we first look at the plot above and to these points (i.e. the vertices of the
bright green non-trasparent region). Then, we look which constraint boundaries cross these bright green non-trasparent region). Then, we look which constraint boundaries cross these
edges, and we formulate an algebraic expression to find these points. In edges, and we formulate an algebraic expression to find these points. In
clockwise order, we have: clockwise order, we have:
@ -130,7 +178,8 @@ x^*_3 = \frac{1}{13} \cdot \begin{bmatrix}3\\24\end{bmatrix} \;\;\; f(x^*_3) = 4
\cdot \frac{3}{13} + 3 \cdot \frac{24}{13} = \frac{84}{13}$$$$ \cdot \frac{3}{13} + 3 \cdot \frac{24}{13} = \frac{84}{13}$$$$
x^*_4 = \frac12 \cdot \begin{bmatrix}3\\2\end{bmatrix} \;\;\; f(x^*_4) = 4 \cdot x^*_4 = \frac12 \cdot \begin{bmatrix}3\\2\end{bmatrix} \;\;\; f(x^*_4) = 4 \cdot
\frac32 + 3 \cdot 1 = 9$$$$ \frac32 + 3 \cdot 1 = 9$$$$
x^*_5 = \begin{bmatrix}2\\0\end{bmatrix} \;\;\; f(x^*_5) = 4 \cdot 2 + 1 \cdot 0 = 8$$ x^*_5 = \begin{bmatrix}2\\0\end{bmatrix} \;\;\; f(x^*_5) = 4 \cdot 2 + 1 \cdot 0
= 8$$
Therefore, $x^* = x^*_1 = \begin{bmatrix}0 & 0\end{bmatrix}^T$ is the global Therefore, $x^* = x^*_1 = \begin{bmatrix}0 & 0\end{bmatrix}^T$ is the global
constrained minimizer. constrained minimizer.
@ -138,6 +187,73 @@ constrained minimizer.
# Exercise 3 # Exercise 3
## Exercise 3.1 ## Exercise 3.1
<!--
I consider the given problem, which is exactly the same as one of the problems
of the previous assignment (Homework 4):
$$\min_{x} f(x) = 3x^2_1 + 2x_1x_2 + x_1x_3 +
2.5x^2_2 + 2x_2x_3 + 2x^2_3 - 8x_1 - 3x_2 - 3x_3
$$$$\text{ subject to } x_1 + x_3 = 3 \;\;\; x_2 + x_3 = 0$$
defining $x$ as $(x_1,\,x_2,\,x_3)^T$, that can be written in the form of a
quadratic minimization problem:
$$\min_{x} f(x) = \dfrac{1}{2} \langle x,\, Gx\rangle + \langle x,\, c\rangle \\
\text{ subject to } Ax = b$$
Where $G\in \mathbb{R}^{n\times n}$ is a symmetric positive definite matrix,
$x$, $c \in \mathbb{R}^n$. The equality constraints are defined in terms of the
matrix $A\in \mathbb{R}^{m\times n}$, with $m \leq n$ and vector $b \in
\mathbb{R}^m$. Here, matrix $A$ has full rank.
-->
Yes, the problem can be solved with _Uzzawa_'s method since the problem can be
reformulated as a saddle point system. The KKT conditions of the problem can be
reformulated as a matrix-vector to vector equality in the following way:
$$\begin{bmatrix}G & -A^T\\A & 0 \end{bmatrix} \begin{bmatrix}
x^*\\\lambda^* \end{bmatrix} = \begin{bmatrix} -c\\b \end{bmatrix}.$$
If we then express the minimizer $x^*$ in terms of $x$, an approximation of it,
and $p$, a search step (i.e. $x^* = x + p$), we obtain the following system.
$$\begin{bmatrix}
G & A^T\\
A & 0
\end{bmatrix}
\begin{bmatrix}
-p\\
\lambda^*
\end{bmatrix} =
\begin{bmatrix}
g\\
h
\end{bmatrix}$$
This is the system the _Uzzawa_ method will solve. Therefore, we need to check
if the matrix:
$$K = \begin{bmatrix}G & A^T \\ A& 0\end{bmatrix} = \begin{bmatrix}
6 & 2 & 1 & 1 & 0 \\
2 & 5 & 2 & 0 & 1 \\
1 & 2 & 4 & 1 & 1 \\
1 & 0 & 1 & 0 & 0 \\
0 & 1 & 1 & 0 & 0 \\
\end{bmatrix}\text{ recalling the computed values of }A\text{ and }G\text{ from the
previous assignment}$$
Has non-zero positive and negative eigenvalues. We compute the eigenvalues of this
matrix with MATLAB, and we find:
$$\begin{bmatrix}
-0.4818\\
-0.2685\\
2.6378\\
4.3462\\
8.7663\end{bmatrix}$$
Therefore, the system is indeed a saddle point system and it can be solved with
_Uzzawa_'s method.
## Exercise 3.2 ## Exercise 3.2

View file

@ -58,21 +58,18 @@ legend('2x1 + 3x2 <= 6', '-3x1 + 2x2 <= 3', '2x2 <= 5', ...
'2x1 + x2 <= 4', 'x1 > 0 and x2 > 0', 'feasible region'); '2x1 + x2 <= 4', 'x1 > 0 and x2 > 0', 'feasible region');
hold off hold off
%% gsppn %% Exercise 3.1
for i=1:5
obj = 4 * px(i) + 3 * py(i);
fprintf("x1=%g x2=%g y=%g\n", px(i), py(i), obj);
end
%% Exercise 3.2
G = [6 2 1; 2 5 2; 1 2 4]; G = [6 2 1; 2 5 2; 1 2 4];
c = [-8; -3; -3]; c = [-8; -3; -3];
A = [1 0 1; 0 1 1]; A = [1 0 1; 0 1 1];
b = [3; 0]; b = [3; 0];
K = [G A'; A zeros(2)];
eig(K)
%% Exercise 3.2
[x, lambda] = uzawa(G, c, A, b, [0;0;0], [0;0], 1e-8, 100); [x, lambda] = uzawa(G, c, A, b, [0;0;0], [0;0], 1e-8, 100);
display(x); display(x);
display(lambda); display(lambda);