view linear algebra from geometrically 【官方双语/合集】线性代数的本质 - 系列合集_哔哩哔哩_bilibili
L1 Vector
vector points of view:
- fancy word of list [a, b, c] - computer science
- arrow in space - physics
facts about vector (geometrically view)
- root from origin
- in a coordinate system a vector specify each basis steps from origin to tip
- can be view as length and direction
- natural view of adding and scale
2D x-y coordinate system, adding & scale geometrically natural
view as go to tip of then go as coord of , equally go as , scale v is equal to scale to it's coordinate.
L2 Linear Combinations, Span, Basis
basis
In a coord system, each dimension unit vector like length is 1, any vector in coordinate system can be viewed product of those unit vectors , unit vector called basis, any time vector is depending on current basis
linear combination
scaling 2 vectors and adding is linear combination of those 2 vectors
span
span of two of vectors is the set of all possible vectors can be reach with all linear combination of a given pair of vectors
in x-y 2D coordinate system, if and
- tips not on the same line: span is x-y 2D space
- tips line up: span is all vectors whose tips on a certain line
- tips at origin, span is origin
single vector view as arrow in space cross origin, dealing with collection of vectors more easily viewed them as points
liner dependent
adding vector not expend span, the add one is "linear depend" with existing ones. , is linearly dependent, for all possible , then is linear indedenpent with
the basis of a vector is a set of linear independent vectors that span full space.
L3 Matrices as linear transformations
linear transformation
info
transformation a fancy word of function, with one input then generate one output , view this as movement change to , for after linear transformation geometrically:
- all lines must remain lines (without getting curved)
- origin myst fixed in place
as the author puts: "keeping grid line parallel and evenly spaced"
represent
all vector in input space can be transformation in output span, how to represent numerically concisely
vector in input space can be track by basis, if we track how basis transform, then input vector can be match with output vector. , is after transformation of , likewise for . so
as example , now we see output space basis
so , linear transformation is specify/describe by 4 numbers for arbitrary x y(all input space), we see matrix is transformation specify by column vector as basis
formally linear transformation
- where c is constant
if is linear combination of , span is line, since not extra is added
linear transformation are ways to move around space, see a matrix as certain transformation of space, matrix-vector multiplication is just a way to compute what that transformation does to a given vector
L4 Matrix multiplication and composite transformation
multiplying two matrices like geometric meaning of applying one transformation then another
, application is then (right -> left), track where and going, first after applying M2, , since applying then is same (), the overall effect is same, which means , so the composite transformation basis are two columns, geometrically matches numerically
we can view naturally why and
L6 Determinant
2D as example, determinant of transformation: tracking area surround by basis, and all other shapes change the same mount, since grid lines evenly space
when , it squishes all of space onto a line or even a single point. checking Det is zero means span of output is squishes small dimensions.
if after transformation andremain the virtual order Det is positive, otherwise is negative. this natural: at start and stay apart, Det is positive, then towards , Det is closing to 0, after line on Det is 0, at last across , Det is nagetive.
3D coordinate system Det is volumn, sign apply right-hand rule, some times computations does not fall within the essence of linear algebra, prove take some effort, but if view geometrically it's natural composite area just same each area multiply, it's not official prove but it's give us inspire what it should like.
L7 Inverse matrix, column space and null space
Inverse matrix
linear algebra mainly application in graphs, robot, solve linear equations. support we have in matrix form , it's identical find original vector that after transformation lands on , if there exist unique solution, if we define reverse transformation of is , transform then back as nothing happen.
notice we hind in geometrically as transformation numerically as matrix, then like a coin both side, so we got inverse matrix from inverse transformation.
we want solve
- if , not reverse transformation(function one input match one output) exists, you can not turn a line/point into a plane, not exists.
- if , transformation unique(same span to span), exists.
rank
we transform 3D space into 2D plane, one dimensional number line, even a point, in that rank specify number of dimensions in output basis.
column space
for all possible , the sets of all outputs is column space. zero vector always in column space, since linear transformation require "origin remain". rank also view as dimensions in column space. full rank imply rank == columns count
for full rank transformation only zero vector lands on itself, not full rank transformation a bunch of vectors land on zero vector.
the set of vectors land on origin after transformation is null space/kernal of matrix
L8 no square matrix
as same we have two basis, each use 3 coordinate to descripe
-> LT -> column space is a 2D plane slicing through the origin of of 3D space, it use 3D to describe, but it's not cover all 3D space, it's full rank transformation
full rank: column dimension = input space dimension
L9 Dot product and duality
dot product
, it's geometric meaning is projection one to anther then multiply their length. if project vector is opposite from original one the result is negative. if both vector are perpenclicular, dot product is zero.
connection
why dot production connect with projection ?
consider some 1x2 matrix transform 2x2 vector into number, which means 2D plane -> 1D number, which has the same calculation of dot project.
consider this: in 2D x-y coordinate system project unit vector and in to diagonal line, since this transform is linear, there must exist some 1x2 matrix fulfill this job, which not defined in term of numerical vector or vector dot product, since multiplying 1x2 matrix by a 2D vector is the same thing as turning its side and taking a dot product, this transformation inescapably related to to some 2D vector, so in this way dot product has geometrically projecting.
for any 2D vector in original space, , tranform to line of , tracking and , mapping into , ,
for non-unit vector , dot product also multiple some scale factor c
matrix multiply (1x2 as example) is the same thing as tasking a dot product.
duality
duality: natural but surprising correspondence between two types of mathematical things.
- dual of a vector is the linear transformation that it encodes.
- dual of linear transformation from some space to one dimension is a certain vector
dotting 2 vectors together is a way to translate one of them into the world of transformation
view vector as physical embodiment of a linear transformation, vector is really a conceptual shorthand for a certain transformation
C10 Cross product in the light of linear transformation
, geometrically
3D: something combines 2 different 3D vector to get a new 3D vector
- direction fellowed by right-hand rule.
- perpendicular with and .
- length =
, it's strange mess with i-hat in determinant, just symbols for show.
C11Cross product
pre-defined:
- , parallelogram's area over and , surround by
- perpendicular to and
- direction obeys right-hand rule.
why corss product geometrically related #1 #2 #3 properties ?
when there exists a linear transformation to number line, a vector can be found (dual of that transformation), performing the linear transformation is same thing as taking a dot product.
hint of steps:
- define 3D->1D linear transformation in term of and
- find dual of vector of #1 linear transformation.
- show dual vector is
function , f transform 3D -> number line, it's linear,
-> there must exist unique matrix describe this transformation,
-> there must exist unique vector(dual) makes
-> compare both side , plugging is way of signaling interpret those coefficients as the coordinates, thus
-> in this way, we connect cross product and dot product, has geometric interoperation
-> cross product has geometric perpendicular to plane determined by and , is area determinded by and
12 Change of basis
coordinate system function as translate between vectors and sets of number, x, y implicit current basis.
standard coordinate system (SC) ; bob 's coordinate system (BC), (use SC basis describe)
- origin meets
- direction of axes and the spacing of grid lines are different
function as transform bob' s coordinate into current coordinate system(describe by SC), then apply M 100 times(describe by SC), then reverse into BC(describe by BC)
13 Eigenvectors and eigenvalue
math equation : , after linear transformation some vector is equal scaled by some factor, the vector is eigenvector, the factor is eigenvalue.
eigenvector ' s span it unchanged, it geometrically stretch/squish.
application: 3D rotation of axis is eigenvector, it's unchanged durning rotation. it's easy think 3D rotation in terms of some axis rotation and an angle rather then thinking 3x3 matrix associated with transformation.
eigenvector/eigenvalue in in-depend with coordinate system
, zero vector always hold, the only way it's possible for product of a matrix with a non-zero vector to become zero is if the transformation with that squishes space into lower dimension, thus
3D -> plane, line, origin
if every vector has move, thus no eigenvector exists, rotation , thus no real number solution, since rotation lets every vector left there own span.
shear transformation , x-axis vector not move or scale,
eigenvectors may have multiple eigenvalue
eigenbasis: eigenvector lines up with basis.
diagonal matrix: all the basis vector is eigenvector, eigenvalue is diagonal value.
guaranteed to be diagonal with , base vector just get scaled durning transformation.
, use eigen value change basis, form diagonal matrix, then multiple 100 times, reverse to original coordinate system.
C14 Abstract vector space
what's vector
determinant and eigenvectors don't care about the coordinate system, how much a transformation scales area, stay on their own span durning transformation.
vector-ish qualities:
- additivity:
- scaling:
transformation, function(derivative is linear) can preserve the operations of additivity & scaler multiplication, so abstract entity full-fill #1 #2 properties is in vector space
axioms for vector
- for every , exists
grid lines remain parallel and evenly spaced is geometrically equal to 8-axioms more intuitively, the form of vector doesn't really matter. like number 3, can be 3 persons, 3 cars, 3 things, while add, subtract is same thing.
Abstract is price of generality
terms between different context, they mean the same thing.
linear algebra | alternate names when apply functions |
---|---|
linear transformation | linear operations |
dot product | inner product |
eigenvector | eigenfunction |
C15 Cramer's rule
show as example
, slove x, y.
Cramer's rule is not most efficient ways of solving equation, but it's more intuitive.
det(A) = 0 there may be non solution or many solutions
, unique solution exist.
after some transformation T, if , T is orthonormal(正交)
surround with parallelogram area equal to y, this parallelogram after transformation change the same of mount det(A), (parallelogram formed by transformed i-hat and ) = , , for the reason,
it's worthy thing how to work in 3D, volume change the same way