Midterm for CS 4220
You may use any resources (course notes, books, papers) except consulting other people, providing attribution for any good ideas you might get. Solutions that involve computing an explicit inverse (or pseudoinverse) without directions to do so will not get full credit.
1. Efficient operations (4 points)
Rewrite each of the following codes to have the desired complexity.
p1d_ref (generic function with 1 method)
2. A bit of basis (4 points)
The Chebyshev polynomials are $T_0(x) = 1$, $T_1(x) = x$, and then
$$T_{j+1}(x) = 2x T_j(x) - T_{j-1}(x).$$
The Legendre polynomials are $P_0(x) = 1$, $P_1(x) = x$, and then
$$(n+1) P_{n+1}(x) = (2n+1) x P_n(x) - n P_{n-1}(x).$$
The Legendre polynomials satisfy
$$\int_{-1}^1 P_i(x) P_j(x) \, dx = \begin{cases} \frac{2}{2j+1}, & i = j \\ 0, & \mbox{otherwise} \end{cases}$$
The function poly_xform
below computes a matrix
$$A = \begin{bmatrix} a_{00} & a_{01} & \ldots & a_{0d} \\ 0 & a_{11} & \ldots & a_{1d} \\ 0 & 0 & \ddots & \vdots \\ 0 & 0 & \ldots & a_{dd} \end{bmatrix}$$
such that
$$T_j(x) = \sum_{i = 0}^j P_i(x) A_{ij}.$$
If $T(x)$ and $P(x)$ are row vectors of Chebyshev and Legendre polynomials, we can write this relation concisely as $T(x) = P(x) A$.
poly_xform (generic function with 1 method)
(1 point) Suppose $p(x) = \sum_{j=0}^d T_j(x) c_j = T(x) c$ and $A$ is given by
poly_xform
. How can we compute $w$ such that $p(x) = \sum_{j=0}^d P_j(x) w_j = P(x) w$?(1 point) Argue that $\int_{-1}^1 p(x)^2 \, dx = \sum_{j=0}^d 2 w_j^2 / (2j+1)$.
(2 points) It is possible to write $\int_{-1}^1 p(x)^2 \, dx$ as $c^T M c$ for some symmetric positive definite matrix $M$. How could you compute $M$ using the tools above, and what is $\int_{-1}^1 p(x)^2 \, dx$ for the case given below?
1.26607
1.13032
0.271495
0.0443368
0.00547424
0.000542926
4.49773e-5
3.19844e-6
1.99212e-7
1.10368e-8
5.5059e-10
3. Closest point (4 points)
A certain optimization algorithm involves repeatedly solving linear systems of the form
$$\begin{bmatrix} A & c \\ c^T & 0 \end{bmatrix} \begin{bmatrix} u \\ v \end{bmatrix} = \begin{bmatrix} f \\ g \end{bmatrix}$$
where $c, f, g \in \mathbb{R}^n$ change at each step, but the symmetric positive definite matrix $A \in \mathbb{R}^{n \times n}$ remains fixed. Complete the following function to solve the above linear system in $O(n^2)$ time given a precomputed Cholesky factor $R$ of $A$. For full credit, do it in three solves with $R$ or $R^T$ (and $O(n)$ additional work).
Scoring:
(1 point) Correct derivation
(2 point) Working $O(n^2)$ code
(1 point) Done in three solves
solve_p2 (generic function with 1 method)
8.86134
1.0
4. Silly structure (4 points)
Let $A \in \mathbb{R}^{n \times n}$ be upper triangular except for the last row.
(2 points) Write an $O(n^2)$ function to compute an unpivoted LU decomposition of $A$
(2 points) Write an $O(n^2)$ function to compute $R$ in a QR factorization of $A$.
house (generic function with 1 method)
p4lu (generic function with 1 method)
p4qr (generic function with 1 method)
0.405139
0.0122195
0.0721597
5. Fun with floating point (4 points)
Consider the linear system
$$\begin{bmatrix} 1 & 1-c \\ 1-c & 1 \end{bmatrix} \begin{bmatrix} x \\ y \end{bmatrix} = \begin{bmatrix} f - d \\ f + d \end{bmatrix}$$
We note that inverse of the matrix can be written as
$$\begin{bmatrix} 1 & 1-c \\ 1-c & 1 \end{bmatrix}^{-1} = \frac{1}{1-(1-c)^2} \begin{bmatrix} 1 & -(1-c) \\ -(1-c) & 1 \end{bmatrix}$$
where $c, d$ are small positive real numbers of similar magnitude (say around $\epsilon_{\mathrm{mach}}$) and $f$ is around 1.
We initially consider a straightforward solver for this system using Julia's backslash.
p5v1 (generic function with 1 method)
(1 point) Explain the exception the happens with
p5v1(1e-17, 3e-17, 1.5)
(but not withp5v1(1e-16, 3e-16, 1.5)
).(1 point) Explain why
p5v1(1e-16, 3e-16, 1.5)
is inaccurate.(2 point) Write
p5solve(c, d, f)
to get high relative accuracy in both $x$ and $y$. What is the output ofp5solve(1e-16, 3e-16, 1.5)
? Why do you believe your modified approach?
Hint: It is worth sanity checking your algebra by also comparing to p5v1
for more benign values of $c$ and $d$ (e.g. $c = 0.1$ and $d = 0.3$). Also, this is a place where it is fine to start from an explicit inverse (though calling inv
on the matrix directly works no better than using backslash).
6. Norm! (4 points)
Suppose $M \in \mathbb{R}^{n \times n}$ is an invertible matrix.
(2 point) Argue that $\|v\|_* = \|Mv\|_\infty$ is a vector norm.
(2 point) Write a short Julia code to compute the associated operator norm.
Note: In Julia, opnorm(A, Inf)
computes the infinity operator norm of a matrix $A$; norm(A, Inf)
computes the vector infinity norm (i.e.~$\max_{i,j} |a_{ij}|$).
7. Rank one approximation (4 points)
(1 point) How would you find $x$ to minimize $\|bx-a\|_2^2$ when $a, b \in \mathbb{R}^m$ are given vectors?
(1 point) Suppose $A \in \mathbb{R}^{m \times n}$ and $b \in \mathbb{R}^m$ are given. How would you find $u \in \mathbb{R}^n$ to minimize $\|A-bu^T\|_F^2$?
(1 point) Suppose $A \in \mathbb{R}^{m \times n}$ and $b \in \mathbb{R}^m, c \in \mathbb{R}^n$ are given. Show that $\langle A, bc^T \rangle_F = b^T A c$. (Recall that the Frobenius inner product is $\langle X, Y \rangle_F = \sum_{i,j} y_{ij} x_{ij} = \operatorname{tr}(Y^T X)$).
(1 point) Suppose $A \in \mathbb{R}^{m \times n}$ and $b \in \mathbb{R}^m$ and $c \in \mathbb{R^n}$ are given. How would you find $\gamma \in \mathbb{R}$ to minimize $\|A-\gamma bc^T\|_F^2$?
8. Your turn (2 points)
Choose your favorite two, or answer all three!
What is one thing you think is going particularly well in the class?
What is one thing you think could be improved?
What is one something you find particularly interesting right now, from this class or otherwise?