Definition: Suppose that u = (u1, …, un) and v = (v1, …, vn) are both vectors in ℝn. Then the dot product of u and v is u ⋅ v = u1v1 + … + unvn.
Theorem: Let u, v, w be in ℝn. Then the dot product has the following properties:
Definition: Let x be a vector in ℝn, then the norm of x is given by $\|x\|=\sqrt{x\cdot x}$. Note that ∥cx∥ = |c|∥x∥.
For two vectors u and v, the distance between u and v is given by ∥u − v∥.
Definition: Let u and v be vectors in ℝn are orthogonal if u ⋅ v = 0.
Theorem: (Pythagorean Theorem) Suppose that u and v are in ℝn. Then ∥u + v∥2 = ∥u∥2 + ∥v∥2 if and only if u ⋅ v = 0.
Theorem: (Triangle Inequality) If u and v are in ℝn, then ∥u + v∥ ≤ ∥u∥ + ∥v∥.
Definition: Let S be a subspace of ℝn. A vector u is orthogonal to S if it is orthogonal to every vector in S. The set of all vectors orthogonal to S is called the orthogonal complement of S and is denoted S⊥.
The orthogonal complement to a subspace is also a subspace.
Theorem: Let B = {v1, …, vn} be a basis for a subspace S. Then u ∈ S⊥ (u is orthogonal to S) if and only if u is orthogonal to each vi.
Example: Let s1 = (1, 0, − 1) and s2 = (1, 1, 1) and S be the span of s1 and s2. Is u = ( − 1, 1, 1) ∈ S⊥? What is a basis for S⊥?
Definition: A set of vectors V in ℝn form an orthogonal set the vectors are pairwise orthogonal. This means that if vi and vj are distinct vectors in V, then vi ⋅ vj = 0.
Example:
Theorem: An orthogonal set of nonzero vectors is linearly independent.
Definition: A basis that is orthogonal as a set is called an orthogonal basis. A basis that is orthogonal as a set and is comprised of vectors of norm 1 is called an orthonormal basis.
Theorem: Let S be a subspace with orthogonal basis {v1, …, vk}. Then any vector s ∈ S can be written as a linear combination v = c1v1 + … + ckvk with ci = vi ⋅ s/∥vi∥2.
Example: (THIS IS A BAD EXAMPLE. TURNS OUT NOT TO BE ORTHOGONAL.) Let v1 = ( − 2, 1, 1), v2 = (1, − 1, − 3), v3 = (4, 7, − 1). Write (3, − 1, 5) as a linear combination of vi.
For finite dimensional spaces, we have that (S⊥)⊥ = S. Use this to show that every subspace is the null space of some matrix.