[Linear Algebra] Norm
Norm
Norm of vector
Norm is an operator that gives elements a kind of ‘length’ or ‘size’. Mathematically, norm is defined as
$\mathbf{Definition.}$ Norm
An operator $|| \cdot || : \mathbb{C}^m \to \mathbb{R}$ is a norm if it satisfies:
1) $|| \mathbf{x} || \geq 0; || \mathbf{x} || = 0 \leftrightarrow \mathbf{x} = \mathbf{0}$
2) $|| \mathbf{x} + \mathbf{y} || \leq || \mathbf{x} || + || \mathbf{y} ||$.
3) $|| \alpha \mathbf{x} || = | \alpha | || \mathbf{x} ||$.
Thus, if any operator statisfies the above 3 axioms, we can call it norm. In practice, various kinds of norm are used.
$L_p$ norm
$L_p$ norm is defined as
The most commonly used $L_p$ norm are $L_1$, $L_2$, and $L_\infty$ norm.
$p = 1$
$L_1$ norm is also called Manhattan norm or Taxicab norm:
$p = 2$
When $p=2$, we call it $L_2$ norm or Euclidean norm:
$p = \infty$
Also, $p$ may have an infinite value. When $p = \infty$, we call it $L_\infty$ norm or supremum norm:
Weighted $p$-norm
With a variation of $L_p$ norm, weighted $\mathbf{p}$-norm is defined as follows:
where $W = \text{diag}(\mathbf{w})$.
Norm of matrix
The previous definition of norm is an operator that receives the vector as input and outputs a positive number. However, can we extend it as an operator of the matrix?
(Induced) Matrix Norm
Using the norm of the vector, one can define the norm of the matrix as follows:
$\mathbf{Definition.}$ (Induced) Matrix Norm
Let $A \in \mathbb{C}^{m \times n}$ be given. A matrix norm induced by $|| \cdot ||$, $|| A ||$, is the smallest $C \geq 0$ such that $\forall \mathbf{x} \in \mathbb{C}^n$, $|| A \mathbf{x} || \leq C || \mathbf{x} ||$, i.e.,
$\mathbf{Remark.}$ We may interpret $|| A ||$ as the maximum factor by which $A$ can stretch $\mathbf{x}$, so it’s also called spectral norm.
$L_p$ matrix norm
We may use $L_p$ norm for matrix norm.
$p = 1$
$L_1$ norm of matrix is defined as
$p = 2$
$L_2$ norm of matrix is defined as
$\mathbf{Remark.}$ If $A$ is a single row vector, $|| A ||_2$ corresponds to the vector norm:
and bound is tight when $\mathbf{x} = \mathbf{a}$.
$p=\infty$
$L_\infty$ norm of matrix is defined as
Here are some useful facts to compute the $L_p$ matrix norm:
$\mathbf{Thm\ 1.}$ $L_1$ matrix norm equals to maximum absolute column sum
$\mathbf{Proof.}$
Let $\mathbf{a}_1, \cdots, \mathbf{a}_n$ be column vectors of a matrix $A$.
Note that the bound is tight. Done.
There is an special cases where $L_2$ matrix norm can be easily computed.
$\mathbf{Thm\ 2.}$ $L_2$ norm of outer product
$\mathbf{Proof.}$
Since the equality is achieved when $\mathbf{x} = \mathbf{v}$, done.
$\mathbf{Thm\ 3.}$ $L_2$ norm of diagonal matrix $D \in \mathbb{C}^{m \times m}$
$\mathbf{Proof.}$
Let $D = \text{diag}(d_1, \cdots, d_m) \in \mathbb{C}^{m \times m}$. Then,
Since the equality is tight, done.
$\mathbf{Thm\ 4.}$ $L_\infty$ matrix norm equals to maximum absolute row sum:
$\mathbf{Proof.}$
Let $\mathbf{a}_1^\star, \cdots, \mathbf{a}_m^\star$ be row vectors of $A$. Consider $|| \mathbf{x} || = 1$.
Since the equality is achievable, done.
General Matrix Norm
Indeed, matrix norms do not need to be induced by vector norms; it just suffices to satisfy 3 conditions of norm. The famous example of this case is Frobenius norm.
Frobenius norm
Frobenius norm, or Hilbert-Schmidt norm is defined as:
Significant Properties of Norm
Cauchy-Schwartz & Hölder Inequality
$\mathbf{Thm\ 5.}$ Hölder Inequality
For $\frac{1}{p} + \frac{1}{q} = 1 \; (1 \leq p, q \leq \infty)$, the following inequality holds:
\[\begin{align*} | \left< \mathbf{x}, \mathbf{y} \right> | \leq \| \mathbf{x} \|_p \cdot \| \mathbf{y} \|_q \end{align*}\]and it is known as Hölder Inequality. In special case of $p = q = 2$, we call it Cauchy-Schwartz Inequality.
Bounding $|| AB ||$
$\mathbf{Thm\ 6.}$ Let any $A \in \mathbb{C}^{l \times m}, B \in \mathbb{C}^{m \times n}$. Then,
$\mathbf{Proof.}$
Proof is quite simple:
For any vector (with appropriate dimension) $\mathbf{x}$,
Note that it is not an equality, for example, $| | A^n | | \neq | | A | |^n (n \geq 2)$ generally.
$\mathbf{Remark.}$ We can bound Frobenius norm, too:
Unitary matrix preserves norm
$\mathbf{Thm\ 7.}$ For any $A \in \mathbb{C}^{m \times n}$ and unitary $Q \in \mathbb{C}^{m \times m}$,
$\mathbf{Proof.}$
Note that it remains valid if $Q$ is generalized to a rectangular matrix with orthonormal columns.
Reference
[1] Lloyd N. Trefethen & David Bau III, Numerical Linear Algebra, 1997.
Leave a comment