documents: add selected solutions

This commit is contained in:
Youwen Wu 2024-10-24 19:26:36 -07:00
parent b28b8455a9
commit 866b86c69a
Signed by: youwen5
GPG key ID: 865658ED1FE61EC3
5 changed files with 418 additions and 31 deletions

View file

@ -0,0 +1,98 @@
#import "@preview/unequivocal-ams:0.1.1": ams-article, theorem, proof
#show: ams-article.with(title: "Exam Solutions")
Q6.
The general clockwise rotation matrix in $RR^2$ is
$
mat(cos theta, sin theta; -sin theta, cos theta)
$
We have $theta = (3pi)/2$, and
$ cos((3pi) / 2) = 0, sin((3pi) / 2) = -1 $
So our particular rotation matrix is
$
T = mat(0, -1; 1, 0)
$
Clearly, the linear transformation that reflects a vector across the vertical
axis changes the first standard basis vector, $vec(1, 0)$, to $vec(-1, 0)$.
This corresponds to the linear transformation (matrix)
$
S = mat(-1, 0; 0, 1)
$
Matrix composition, $compose$, is an equivalent notion to matrix multiplication. Therefore, we have the two compositions.
+ #[
$T compose S$, the linear transformation corresponding to a reflection followed by rotation:
$
T compose S &= mat(0, -1; 1, 0) mat(-1, 0; 0, 1) \
&#[using _column by coordinate_ rule] \
&= mat((-1 vec(0,1) + 0 vec(-1,0)), (0 vec(0,1) + 1 vec(-1,0))) \
&= mat(0, -1,;-1, 0)
$
]
+ #[
$S compose T$, the linear transformation corresponding to rotation, followed by reflection. Since matrix composition is generally not commutative, we obtain a different matrix.
$
S compose T &= mat(-1,0; 0,1) mat(0,-1; 1,0) \
&#[using _column by coordinate_ rule] \
&= mat((0 vec(-1, 0) + 1 vec(0, 1)), (-1 vec(-1, 0) + 0 vec(0, -1))) \
&= mat(0, 1; 1, 0)
$
]
#pagebreak()
Q7.
#enum(
numbering: "7.1",
[
The matrix $A$ corresponds to a linear transformation
$T: RR^4 |-> RR^3$. $A$ has 3 rows and 4 columns, so its matrix-vector
multiplication is only defined when with vectors in $RR^4$. Accordingly, it
will output a vector in $RR^3$.
So, $p = 4$.
],
[
See above explanation. $q = 3$.
],
[
To find all vectors $arrow(x) in RR^4$ whose image under $T$ is $arrow(b)$, we seek all solutions $arrow(x) = (x_1, x_2, x_3, x_4, x_5)^T$ to the equation
$ T arrow(x) = arrow(b) $
We can do this using our usual row reduction methods.
$
mat(augment: #(-1), -2,3,7,-11,-3;1,0,-2,1,3; 1,-1,-3,4,2) \
mat(augment: #(-1), 1,-1,-3,4,2; 0, 1, 1, -3, 1; 0, 0, 0, 0, 0)
$
We now have the augmented matrix in echelon form. So, $x_4$ and $x_3$ are free. Then, let $s,t in RR$ be free variables
$
x_1 &= 3 - t + 2s \
x_2 &= 1 + 3t - s \
x_3 &= s \
x_4 &= t
$
So, all vectors $arrow(x)$ are of the form
$
arrow(x) = vec(3+2s - t, 1 - s + 3t, s, t)
$
],
)

View file

@ -0,0 +1,34 @@
{
typstPackagesCache,
typixLib,
cleanTypstSource,
...
}:
let
src = cleanTypstSource ./.;
commonArgs = {
typstSource = "main.typ";
fontPaths = [
# Add paths to fonts here
# "${pkgs.roboto}/share/fonts/truetype"
];
virtualPaths = [
# Add paths that must be locally accessible to typst here
# {
# dest = "icons";
# src = "${inputs.font-awesome}/svgs/regular";
# }
];
XDG_CACHE_HOME = typstPackagesCache;
};
in
typixLib.buildTypstProject (
commonArgs
// {
inherit src;
}
)

View file

@ -0,0 +1,9 @@
@article{netwok2020,
title={At-scale impact of the {Net Wok}: A culinarically holistic investigation of distributed dumplings},
author={Astley, Rick and Morris, Linda},
journal={Armenian Journal of Proceedings},
volume={61},
pages={192--219},
year=2020,
publisher={Automatic Publishing Inc.}
}

View file

@ -15,21 +15,22 @@
= Introduction
Many introductory linear algebra classes focus on _application_. In general,
this is a red herring and is engineer-speak for "we will teach you how to
crunch numbers with no regard for conceptual understanding."
Many introductory linear algebra classes focus on _application_. They teach you
how to perform trivial numerical operations such as the _matrix
multiplication_, _matrix-vector multiplication_, _row reduction_, and other
trite tasks better suited for computers.
This class is essentially useless. Linear algebra is really a much deeper
subject, when viewed through the lens of _linear maps_ and _vector spaces_. In
particular, taking an abstract point-free approach allows the freedom to prove
theorems that generalize to linear algebra on arbitrary vector spaces, and
indeed, even infinite vector spaces.
If you are a math major (or math-adjacent, such as Computer Science), this
class is essentially useless for you. You will learn how to perform trivial
numerical operations such as the _matrix multiplication_, _matrix-vector
multiplication_, _row reduction_, and other trite tasks better suited for
computers.
If you are taking this course, you might as well learn linear algebra properly.
Otherwise, you will have to re-learn it later on, anyways. Completing a math
course without gaining a theoretical appreciation for the topics at hand is an
unequivocal waste of time. I have prepared this brief crash course designed to
fill in the theoretical gaps left by this class.
course without gaining a theoretical appreciation for the topics at hand is a
complete and utter waste of time.
= Basic Notions
@ -78,37 +79,42 @@ example, you cannot add a vector and scalar, as it does not make sense.
_Remark_. For those of you versed in computer science, you may recognize this
as essentially saying that you must ensure your operations are _type-safe_.
Adding a vector and scalar is not just false, it is an _invalid question_
entirely because vectors and scalars and different types of mathematical
objects. See #cite(<chen2024digression>, form: "prose") for more.
Adding a vector and scalar is not just "wrong" in the same sense that $1 + 1 =
3$ is wrong, it is an _invalid question_ entirely because vectors and scalars
and different types of mathematical objects. See #cite(<chen2024digression>,
form: "prose") for more.
=== Vectors big and small
In order to begin your descent into what mathematicians colloquially recognize
as _abstract vapid nonsense_, let's discuss which fields constitute a vector space. We
have the familiar space where all scalars are real numbers, or $RR$. We
generally discuss 2-D or 3-D vectors, corresponding to vectors of length 2 or
as _abstract vapid nonsense_, let's discuss which fields constitute a vector
space. We have the familiar field of $RR$ where all scalars are real numbers,
with corresponding vector spaces $RR^n$, where $n$ is the length of the vector.
We generally discuss 2D or 3D vectors, corresponding to vectors of length 2 or
3; in our case, $RR^2$ and $RR^3$.
However, vectors in $RR$ can really be of any length. Discard your primitive
conception of vectors as arrows in space. Vectors are simply arbitrary length
lists of numbers (for the computer science folk: think C++ `std::vector`).
However, vectors in $RR^n$ can really be of any length. Vectors can be viewed
as arbitrary length lists of numbers (for the computer science folk: think C++
`std::vector`).
_Example_. $ vec(1,2,3,4,5,6,7,8,9) $
_Example_. $ vec(1,2,3,4,5,6,7,8,9) in RR^9 $
Moreover, vectors need not be in $RR$ at all. Recall that a vector space need
only satisfy the aforementioned _axioms of a vector space_.
Keep in mind that vectors need not be in $RR^n$ at all. Recall that a vector
space need only satisfy the aforementioned _axioms of a vector space_.
_Example_. The vector space $CC$ is similar to $RR$, except it includes complex
numbers. All complex vector spaces are real vector spaces (as you can simply
restrict them to only use the real numbers), but not the other way around.
_Example_. The vector space $CC^n$ is similar to $RR^n$, except it includes
complex numbers. All complex vector spaces are real vector spaces (as you can
simply restrict them to only use the real numbers), but not the other way
around.
From now on, let us refer to vector spaces $RR^n$ and $CC^n$ as $FF^n$.
In general, we can have a vector space where the scalars are in an arbitrary
field $FF$, as long as the axioms are satisfied.
field, as long as the axioms are satisfied.
_Example_. The vector space of all polynomials of degree 3, or $PP^3$. It is
not yet clear what this vector may look like. We shall return to this example
once we discuss _basis_.
_Example_. The vector space of all polynomials of at most degree 3, or $PP^3$.
It is not yet clear what this vector may look like. We shall return to this
example once we discuss _basis_.
== Vector addition. Multiplication
@ -138,6 +144,129 @@ _Example_.
$ beta vec(a, b, c) = vec(beta dot a, beta dot b, beta dot c) $
== Linear combinations
Given vector spaces $V$ and $W$ and vectors $v in V$ and $w in W$, $v + w$ is
the _linear combination_ of $v$ and $w$.
=== Spanning systems
We say that a set of vectors $v_1, v_2, ..., v_n in V$ _span_ $V$ if the linear
combination of the vectors can represent any arbitrary vector $v in V$.
Precisely, given scalars $alpha_1, alpha_2, ..., alpha_n$,
$ alpha_1 v_1 + alpha_2 v_2 + ... + alpha_n v_n = v, forall v in V $
Note that any scalar $alpha_k$ could be 0. Therefore, it is possible for a
subset of a spanning system to also be a spanning system. The proof of this
fact is left as an exercise.
=== Intuition for linear independence and dependence
We say that $v$ and $w$ are linearly independent if $v$ cannot be represented by the scaling of $w$, and $w$ cannot be represented by the scaling of $v$. Otherwise, they are _linearly dependent_.
You may intuitively visualize linear dependence in the 2D plane as two vectors
both pointing in the same direction. Clearly, scaling one vector will allow us
to reach the other vector. Linear independence is therefore two vectors
pointing in different directions.
Of course, this definition applies to vectors in any $FF^n$.
=== Formal definition of linear dependence and independence
Let us formally define linear independence for arbitrary vectors in $FF^n$. Given a set of vectors
$ v_1, v_2, ..., v_n in V $
we say they are linearly independent iff. the equation
$ alpha_1 v_1 + alpha_2 v_2 + ... + alpha_n v_n = 0 $
has only a unique set of solutions $alpha_1, alpha_2, ..., alpha_n$ such that
all $alpha_n$ are zero.
Equivalently,
$ abs(alpha_1) + abs(alpha_2) + ... + abs(alpha_n) = 0 $
More precisely,
$ sum_(i=1)^k abs(alpha_i) = 0 $
Therefore, a set of vectors $v_1, v_2, ..., v_m$ is linearly dependent if the opposite is true, that is there exists solution $alpha_1, alpha_2, ..., alpha_m$ to the equation
$ alpha_1 v_1 + alpha_2 v_2 + ... + alpha_m v_m = 0 $
such that
$ sum_(i=1)^k abs(alpha_i) != 0 $
=== Basis
We say a system of vectors $v_1, v_2, ..., v_n in V$ is a _basis_ in $V$ if the
system is both linearly independent and spanning. That is, the system must be
able to represent any vector in $V$ as well as satisfy our requirements for
linear independence.
Equivalently, we may say that a system of vectors in $V$ is a basis in $V$ if
any vector $v in V$ admits a _unique representation_ as a linear combination of
vectors in the system. This is equivalent to our previous statement, that the
system must be spanning and linearly independent.
=== Standard basis
We may define a _standard basis_ for a vector space. By convention, the
standard basis in $RR^2$ is
$ vec(1, 0) vec(0, 1) $
Verify that the above is in fact a basis (that is, linearly independent and
generating).
Recalling the definition of the basis, we can represent any vector in $RR^2$ as
the linear combination of the standard basis.
Therefore, for any arbitrary vector $v in RR^2$, we can represent it as
$ v = alpha_1 vec(1, 0) + alpha_2 vec(0,1) $
Let us call $alpha_1$ and $alpha_2$ the _coordinates_ of the vector. Then, we can write $v$ as
$ v = vec(alpha_1, alpha_2) $
For example, the vector
$ vec(1, 2) $
represents
$ 1 dot vec(1, 0) + 2 dot vec(0,1) $
Verify that this aligns with your previous intuition of vectors.
You may recognize the standard basis in $RR^2$ as the familiar unit vectors
$ accent(i, hat), accent(j, hat) $
This aligns with the fact that
$ vec(alpha, beta) = alpha hat(i) + beta hat(j) $
However, we may define a standard basis in any arbitrary vector space. So, let
$ e_1, e_2, ..., e_n $
be a standard basis in $FF^n$. Then, the coordinates $alpha_1, alpha_2, ..., alpha_n$ of a vector $v in FF^n$ represent the following
$
vec(alpha_1, alpha_2, dots.v, alpha_n) = alpha_1 e_1 + alpha_2 + e_2 + alpha_n e_n
$
Using our new notation, the standard basis in $RR^2$ is
$ e_1 = vec(1,0), e_2 = vec(0,1) $
== Matrices
Before discussing any properties of matrices, let's simply reiterate what we
@ -162,4 +291,121 @@ $
mat(1,2,3;4,5,6)^T = mat(1,4;2,5;3,6)
$
Formally, we can say $(A_(j,k))^T = A_(k,j)$.
Formally, we can say $(A^T)_(j,k) = A_(k,j)$
== Linear transformations
A linear transformation $T : V -> W$ is a mapping between two vector spaces $V
-> W$, such that the following axioms are satisfied:
+ $T(v + w) = T(v) + T(w), forall v in V, forall w in W$
+ $T(alpha v) + T(beta w) = alpha T(v) + beta T(w), forall v in V, forall w in W$, for all scalars $alpha, beta$
_Definition_. $T$ is a linear transformation iff.
$ T(alpha v + beta w) = alpha T(v) + beta T(w) $
_Abuse of notation_. From now on, we may elide the parentheses and say that
$ T(v) = T v, forall v in V $
_Remark_. A phrase that you may commonly hear is that linear transformations
preserve _linearity_. Essentially, straight lines remain straight, parallel
lines remain parallel, and the origin remains fixed at 0. Take a moment to
think about why this is true (at least, in lower dimensional spaces you can
visualize).
_Examples_.
+ #[Rotation for $V = W = RR^2$ (i.e. rotation in 2 dimensions). Given $v, w in
RR^2$, and their linear combination $v + w$, a rotation of $gamma$ radians of
$v + w$ is equivalent to first rotating $v$ and $w$ individually by $gamma$ and
then taking their linear combination.]
+ #[Differentiation of polynomials. In this case $V = PP^n$ and $W = PP^(n -
1)$, where $PP^n$ is the field of all polynomials of degree at most $n$.
$
dif / (dif x) (
alpha v + beta w
) = alpha dif / (dif x) v + beta dif / (dif x) w, forall v in V, w in W, forall "scalars" alpha, beta
$
]
== Matrices represent linear transformations
Suppose we wanted to represent a linear transformation $T: FF^n -> FF^m$. I
propose that we need encode how $T$ acts on the standard basis of $FF^n$.
Using our intuition from lower dimensional vector spaces, we know that the
standard basis in $RR^2$ is the unit vectors $hat(i)$ and $hat(j)$. Because
linear transformations preserve linearity (i.e. all straight lines remain
straight and parallel lines remain parallel), we can encode any transformation
as simply changing $hat(i)$ and $hat(j)$. And indeed, if any vector $v in RR^2$
can be represented as the linear combination of $hat(i)$ and $hat(j)$ (this is
the definition of a basis), it makes sense both symbolically and geometrically
that we can represent all linear transformations as the transformations of the
basis vectors.
_Example_. To reflect all vectors $v in RR^2$ across the $y$-axis, we can simply change the standard basis to
$ vec(-1, 0) vec(0,1) $
Then, any vector in $RR^2$ using this new basis will be reflected across the
$y$-axis. Take a moment to justify this geometrically.
=== Writing a linear transformation as matrix
For any linear transformation $T: FF^m -> FF^n$, we can write it as an $n times
m$ matrix $A$. That is, there is a matrix $A$ with $n$ rows and $m$ columns
that can represent any linear transformation from $FF^m -> FF^n$.
How should we write this matrix? Naturally, from our previous discussion, we
should write a matrix with each _column_ being one of our new transformed
_basis_ vectors.
_Example_. Our $y$-axis reflection transformation from earlier. We write the bases in a matrix
$ mat(-1,0; 0,1) $
=== Matrix-vector multiplication
Perhaps you now see why the so-called matrix-vector multiplication is defined
the way it is. Recalling our definition of a basis, given a basis in $V$, any
vector $v in V$ can be written as the linear combination of the vectors in the
basis. Then, given a linear transformation represented by the matrix containing
the new basis, we simply write the linear combination with the new basis
instead.
_Example_. Let us first write a vector in the standard basis in $RR^2$ and then
show how our matrix-vector multiplication naturally corresponds to the
definition of the linear transformation.
$ vec(1, 2) in RR^2 $
is the same as
$ 1 dot vec(1, 0) + 2 dot vec(0, 1) $
Then, to perform our reflection, we need only replace the basis vector $vec(1,
0)$ with $vec(-1, 0)$.
Then, the reflected vector is given by
$ 1 dot vec(-1, 0) + 2 dot vec(0,1) = vec(-1, 2) $
We can clearly see that this is exactly how the matrix multiplication
$ mat(-1, 0; 0, 1) dot vec(1, 2) $ is defined! The _column-by-coordinate_ rule
for matrix-vector multiplication says that we multiply the $n^("th")$ entry of
the vector by the corresponding $n^("th")$ column of the matrix and sum them
all up (take their linear combination). This algorithm intuitively follows from
our definition of matrices.
=== Matrix-matrix multiplication
As you may have noticed, a very similar natural definition arises for the
_matrix-matrix_ multiplication. Multiplying two matrices $A dot B$ is
essentially just taking each column of $B$, and applying the linear
transformation defined by the matrix $A$!