This content originally appeared on DEV Community and was authored by Terra
Honestly, this is because I forgot that the book shouldn't need a bachelors in mathematics to read :(
A more mathematical proof
So, remember the long breakdown I did before? So, turns out the author has a more fancy way (and probably the correct way) of finding out why there's an inverse in our formula before hand.
We are going to look at
With:
and
What does this mean?
We'll get to that, before that though, remember this?
The linear mapping can be notated as:
With the beggining of the quation bj being the vectors of the new basis B (with a weird eyebrow) of V.
And,
With the beggining of the quation ck being the vectors of the new basis C (with a weird eyebrow) of W.
Finally
being the transformation matrix with respect to B (eyebrow) onto coordinates with respect to B and being the transformation matrix with respect to C (eyebrow) onto coordinates with respect to C respectively
Now, let's find out about the first equation
Note:
The first half of the equation above is what's said in the book, but if it's a transformation of B after a basis change, then it's C
First method:
Apply Mapping
Which results in
where we first expressed the new basis vectors
and then swapped the order of summation.
Second method
We express it as:
Which we'll arrive at
Finally
where we exploited the linearity of matrix transformation which result in
and therefore,
such that
But Terra, does that mean the method in the last blog is wrong?
A simple answer is no, it's still stays true since it's working :D
also because the example stated that the matrix is a homomorphism, making the V and W the same
.
What does that mean?
It means it's when bj and ck is the same, making T and S the same. The use of P is just arbitrary since we can just change it to T and S and it's still the same result.
What's clear is that if the matrix isn't a homomorphism, then the formula to use is this one, where it differentiates each of the basis change to ensure if the shape of the matrix is different the formula can still be used.
Acknowledgement
I can't overstate this: I'm truly grateful for this book being open-sourced for everyone. Many people will be able to learn and understand machine learning on a fundamental level. Whether changing careers, demystifying AI, or just learning in general, this book offers immense value even for fledgling composer such as myself. So, Marc Peter Deisenroth, A. Aldo Faisal, and Cheng Soon Ong, thank you for this book.
Source:
Axler, Sheldon. 2015. Linear Algebra Done Right. Springer
Deisenroth, M. P., Faisal, A. A., & Ong, C. S. (2020). Mathematics for Machine Learning. Cambridge: Cambridge University Press.
https://mml-book.com
This content originally appeared on DEV Community and was authored by Terra
Terra | Sciencx (2024-07-26T15:40:44+00:00) Mathematics for Machine Learning – Day 18. Retrieved from https://www.scien.cx/2024/07/26/mathematics-for-machine-learning-day-18/
Please log in to upload a file.
There are no updates yet.
Click the Upload button above to add an update.