[[Reading Status Button]]
Linear Transformation:
- Transformation - Function.
- Linear Transformation - takes a vector and transforms(moves) that vector into another vector
- Linear Transformation - 2 properites:
- All lines remain lines, without getting curved
- The origin is fixed
- The resultant grid lines are parallel and evenly spaced
- If you apply a linear transformation on a set of vectors, if you know where the vectors and are, you can find any other vector. It will be the same linear combination and , that was before the transformation
will always be the same.
Matrix mulitplication is just a transformation
This will always be true in a linear transformation. So when you apply linear transformation, just with the co-ordinates of and you can get any other vector, if you know the original co-ordinates of the vector before the transformation.
Example:
you have a vector , which is -1 and, suppose after some transformation the lands at and lands at then the
So the original has moved from co-ordinates (-1,2) to (5,-2) after the transformation.
A 2 dimensional linear transformation always relies on just 4 numbers, the the 2 co-ordinates of and .
These 4 numbers are packed into a 2 x 2 matrix.
The columns denote where the vectors and land.
If the lines where and are linearly dependent, then entire 2D space into a single line, where the 2 vectors sit.
matrices is way to describe transformation, and matrix multiplication is the way to find out what the transformation does to a given vector.
How does matrix multiplication relate to Neural Networks:
- so when performing matrix multiplication on the tokenβs embeddings, we ensure that the reflected context is transformed onto the other tokens.
- And when a 2 linearly dependent vectors are multiplied, the resulting vector has only one dimension.
- say for example take 2 tokens, first name, and second name, those 2 are nouns, so will occupy the same dimension. and might have linearly dependent vectors. So when you multiply them the result will a noun.