Compare commits
No commits in common. "6f693b9da50f4e87be299eaee36a0ff690d68744" and "5128e83d2e8fffc48d49eb8ab648e8daca14dcf9" have entirely different histories.
6f693b9da5
...
5128e83d2e
3 changed files with 2 additions and 158 deletions
Binary file not shown.
|
@ -1,7 +1,7 @@
|
||||||
#import "@preview/unequivocal-ams:0.1.1": ams-article, theorem, proof
|
#import "@preview/unequivocal-ams:0.1.1": ams-article, theorem, proof
|
||||||
|
|
||||||
#show: ams-article.with(
|
#show: ams-article.with(
|
||||||
title: [A Digression on Abstract Linear Algebra],
|
title: [Week 2],
|
||||||
authors: (
|
authors: (
|
||||||
(
|
(
|
||||||
name: "Youwen Wu",
|
name: "Youwen Wu",
|
||||||
|
@ -10,156 +10,6 @@
|
||||||
url: "https://youwen.dev",
|
url: "https://youwen.dev",
|
||||||
),
|
),
|
||||||
),
|
),
|
||||||
bibliography: bibliography("refs.bib"),
|
|
||||||
)
|
)
|
||||||
|
|
||||||
= Introduction
|
$ integral^oo_(-oo) e^(-x^2) dif x = sqrt(pi) $
|
||||||
|
|
||||||
Many introductory linear algebra classes focus on _application_. In general,
|
|
||||||
this is a red herring and is engineer-speak for "we will teach you how to
|
|
||||||
crunch numbers with no regard for conceptual understanding."
|
|
||||||
|
|
||||||
If you are a math major (or math-adjacent, such as Computer Science), this
|
|
||||||
class is essentially useless for you. You will learn how to perform trivial
|
|
||||||
numerical operations such as the _matrix multiplication_, _matrix-vector
|
|
||||||
multiplication_, _row reduction_, and other trite tasks better suited for
|
|
||||||
computers.
|
|
||||||
|
|
||||||
If you are taking this course, you might as well learn linear algebra properly.
|
|
||||||
Otherwise, you will have to re-learn it later on, anyways. Completing a math
|
|
||||||
course without gaining a theoretical appreciation for the topics at hand is an
|
|
||||||
unequivocal waste of time. I have prepared this brief crash course designed to
|
|
||||||
fill in the theoretical gaps left by this class.
|
|
||||||
|
|
||||||
= Basic Notions
|
|
||||||
|
|
||||||
== Vector spaces
|
|
||||||
|
|
||||||
Before we can understand vectors, we need to first discuss _vector spaces_. Thus
|
|
||||||
far, you have likely encountered vectors primarily in physics classes,
|
|
||||||
generally in the two-dimensional plane. You may conceptualize them as arrows in
|
|
||||||
space. For vectors of size $>3$, a hand waving argument is made that they are
|
|
||||||
essentially just arrows in higher dimensional spaces.
|
|
||||||
|
|
||||||
It is helpful to take a step back from this primitive geometric understanding
|
|
||||||
of the vector. Let us build up a rigorous idea of vectors from first
|
|
||||||
principles.
|
|
||||||
|
|
||||||
=== Vector axioms
|
|
||||||
|
|
||||||
The so-called _axioms_ of a _vector space_ (which we'll call the vector space
|
|
||||||
$V$) are as follows:
|
|
||||||
|
|
||||||
#enum[
|
|
||||||
Commutativity: $u + v = v + u, " " forall u,v in V$
|
|
||||||
][
|
|
||||||
Associativity: $(u + v) + w = u + (v + w), " " forall u,v,w in V$
|
|
||||||
][
|
|
||||||
Zero vector: $exists$ a special vector, denoted $0$, such that $v + 0 = v, " " forall v in V$
|
|
||||||
][
|
|
||||||
Additive inverse: $forall v in V, " " exists w in V "such that" v + w = 0$. Such an additive inverse is generally denoted $-v$
|
|
||||||
][
|
|
||||||
Multiplicative identity: $1 v = v, " " forall v in V$
|
|
||||||
][
|
|
||||||
Multiplicative associativity: $(alpha beta) v = alpha (beta v) " " forall v in V, "scalars" alpha, beta$
|
|
||||||
][
|
|
||||||
Distributive property for vectors: $alpha (u + v) = alpha u + alpha v " " forall u,v in V, "scalars" alpha$
|
|
||||||
][
|
|
||||||
Distributive property for scalars: $(alpha + beta) v = alpha v + beta v " " forall v in V, " scalars" alpha, beta$
|
|
||||||
]
|
|
||||||
|
|
||||||
It is easy to show that the zero vector $0$ and the additive inverse $-v$ are
|
|
||||||
_unique_. We leave the proof of this fact as an exercise.
|
|
||||||
|
|
||||||
These may seem difficult to memorize, but they are essentially the same
|
|
||||||
familiar algebraic properties of numbers you know from high school. The
|
|
||||||
important thing to remember is which operations are valid for what objects. For
|
|
||||||
example, you cannot add a vector and scalar, as it does not make sense.
|
|
||||||
|
|
||||||
_Remark_. For those of you versed in computer science, you may recognize this
|
|
||||||
as essentially saying that you must ensure your operations are _type-safe_.
|
|
||||||
Adding a vector and scalar is not just false, it is an _invalid question_
|
|
||||||
entirely because vectors and scalars and different types of mathematical
|
|
||||||
objects. See #cite(<chen2024digression>, form: "prose") for more.
|
|
||||||
|
|
||||||
=== Vectors big and small
|
|
||||||
|
|
||||||
In order to begin your descent into what mathematicians colloquially recognize
|
|
||||||
as _abstract vapid nonsense_, let's discuss which fields constitute a vector space. We
|
|
||||||
have the familiar space where all scalars are real numbers, or $RR$. We
|
|
||||||
generally discuss 2-D or 3-D vectors, corresponding to vectors of length 2 or
|
|
||||||
3; in our case, $RR^2$ and $RR^3$.
|
|
||||||
|
|
||||||
However, vectors in $RR$ can really be of any length. Discard your primitive
|
|
||||||
conception of vectors as arrows in space. Vectors are simply arbitrary length
|
|
||||||
lists of numbers (for the computer science folk: think C++ `std::vector`).
|
|
||||||
|
|
||||||
_Example_. $ vec(1,2,3,4,5,6,7,8,9) $
|
|
||||||
|
|
||||||
Moreover, vectors need not be in $RR$ at all. Recall that a vector space need
|
|
||||||
only satisfy the aforementioned _axioms of a vector space_.
|
|
||||||
|
|
||||||
_Example_. The vector space $CC$ is similar to $RR$, except it includes complex
|
|
||||||
numbers. All complex vector spaces are real vector spaces (as you can simply
|
|
||||||
restrict them to only use the real numbers), but not the other way around.
|
|
||||||
|
|
||||||
In general, we can have a vector space where the scalars are in an arbitrary
|
|
||||||
field $FF$, as long as the axioms are satisfied.
|
|
||||||
|
|
||||||
_Example_. The vector space of all polynomials of degree 3, or $PP^3$. It is
|
|
||||||
not yet clear what this vector may look like. We shall return to this example
|
|
||||||
once we discuss _basis_.
|
|
||||||
|
|
||||||
== Vector addition. Multiplication
|
|
||||||
|
|
||||||
Vector addition, represented by $+$, and multiplication, represented by the
|
|
||||||
$dot$ (dot) operator, can be done entrywise.
|
|
||||||
|
|
||||||
_Example._
|
|
||||||
|
|
||||||
$
|
|
||||||
vec(1,2,3) + vec(4,5,6) = vec(1 + 4, 2 + 5, 3 + 6) = vec(5,7,9)
|
|
||||||
$
|
|
||||||
$
|
|
||||||
vec(1,2,3) dot vec(4,5,6) = vec(1 dot 4, 2 dot 5, 3 dot 6) = vec(4,10,18)
|
|
||||||
$
|
|
||||||
|
|
||||||
This is simple enough to understand. Again, the difficulty is simply ensuring
|
|
||||||
that you always perform operations with the correct _types_. For example, once
|
|
||||||
we introduce matrices, it doesn't make sense to multiply or add vectors and
|
|
||||||
matrices in this fashion.
|
|
||||||
|
|
||||||
== Vector-scalar multiplication
|
|
||||||
|
|
||||||
Multiplying a vector by a scalar simply results in each entry of the vector
|
|
||||||
being multiplied by the scalar.
|
|
||||||
|
|
||||||
_Example_.
|
|
||||||
|
|
||||||
$ beta vec(a, b, c) = vec(beta dot a, beta dot b, beta dot c) $
|
|
||||||
|
|
||||||
== Matrices
|
|
||||||
|
|
||||||
Before discussing any properties of matrices, let's simply reiterate what we
|
|
||||||
learned in class about their notation. We say a matrix with rows of length $m$,
|
|
||||||
and columns of size $n$ (in less precise terms, a matrix with length $m$ and
|
|
||||||
height $n$) is a $m times n$ matrix.
|
|
||||||
|
|
||||||
Given a matrix
|
|
||||||
|
|
||||||
$ A = mat(1,2,3;4,5,6;7,8,9) $
|
|
||||||
|
|
||||||
we refer to the entry in row $j$ and column $k$ as $A_(j,k)$ .
|
|
||||||
|
|
||||||
=== Matrix transpose
|
|
||||||
|
|
||||||
A formalism that is useful later on is called the _transpose_, and we obtain it
|
|
||||||
from a matrix $A$ by switching all the rows and columns. More precisely, each
|
|
||||||
row becomes a column instead. We use the notation $A^T$ to represent the
|
|
||||||
transpose of $A$.
|
|
||||||
|
|
||||||
$
|
|
||||||
mat(1,2,3;4,5,6)^T = mat(1,4;2,5;3,6)
|
|
||||||
$
|
|
||||||
|
|
||||||
Formally, we can say $(A_(j,k))^T = A_(k,j)$.
|
|
||||||
|
|
|
@ -1,6 +0,0 @@
|
||||||
@misc{chen2024digression,
|
|
||||||
author = {Evan Chen},
|
|
||||||
title = {Digression on Type Safety},
|
|
||||||
year = {2024},
|
|
||||||
howpublished = {\url{https://web.evanchen.cc/upload/1802/tsafe-1802.pdf}},
|
|
||||||
}
|
|
Loading…
Reference in a new issue