8.1 Dot Products and Orthogonal Sets

Definition: Suppose that u = (u1, …, un) and v = (v1, …, vn) are both vectors in n. Then the dot product of u and v is u ⋅ v = u1v1 + … + unvn.

Theorem: Let u, v, w be in n. Then the dot product has the following properties:

Definition: Let x be a vector in n, then the norm of x is given by $\|x\|=\sqrt{x\cdot x}$. Note that cx∥ = |c|∥x.

For two vectors u and v, the distance between u and v is given by u − v.

Definition: Let u and v be vectors in n are orthogonal if u ⋅ v = 0.

Theorem: (Pythagorean Theorem) Suppose that u and v are in n. Then u + v2 = ∥u2 + ∥v2 if and only if u ⋅ v = 0.

Theorem: (Triangle Inequality) If u and v are in n, then u + v∥ ≤ ∥u∥ + ∥v.

Definition: Let S be a subspace of n. A vector u is orthogonal to S if it is orthogonal to every vector in S. The set of all vectors orthogonal to S is called the orthogonal complement of S and is denoted S.

The orthogonal complement to a subspace is also a subspace.

Theorem: Let B = {v1, …, vn} be a basis for a subspace S. Then u ∈ S (u is orthogonal to S) if and only if u is orthogonal to each vi.

Example: Let s1 = (1, 0,  − 1) and s2 = (1, 1, 1) and S be the span of s1 and s2. Is u = ( − 1, 1, 1) ∈ S? What is a basis for S?

Definition: A set of vectors V in n form an orthogonal set the vectors are pairwise orthogonal. This means that if vi and vj are distinct vectors in V, then vi ⋅ vj = 0.

Example:

Theorem: An orthogonal set of nonzero vectors is linearly independent.

Definition: A basis that is orthogonal as a set is called an orthogonal basis. A basis that is orthogonal as a set and is comprised of vectors of norm 1 is called an orthonormal basis.

Theorem: Let S be a subspace with orthogonal basis {v1, …, vk}. Then any vector s ∈ S can be written as a linear combination v = c1v1 + … + ckvk with ci = vi ⋅ s/∥vi2.

Example: (THIS IS A BAD EXAMPLE. TURNS OUT NOT TO BE ORTHOGONAL.) Let v1 = ( − 2, 1, 1), v2 = (1,  − 1,  − 3), v3 = (4, 7,  − 1). Write (3,  − 1, 5) as a linear combination of vi.

For finite dimensional spaces, we have that (S) = S. Use this to show that every subspace is the null space of some matrix.