NOTE: This StatQuest was supported by these awesome people who support StatQuest at the Double BAM level: I. Urosev, S. Ágoston, M. Steenbergen, P. Keener, Alex, S. Kundapurkar, JWC, BufferUnderrrun, S. Jeffcoat, S. Handschuh, J. Le, D. Greene, D. Schioberg, Z. Rosenberg, H-M Chang, M. Ayoubieh, Losings, F. Pedemonte, S. Song US, A. Tolkachev, L. Cisterna, J. Alexander
Hi!!
A small fan boy from India. I find your ML and Statistics video series. The concepts are explained with astonishing clarity and simplicity. Simply love the way you teach. Kudos for this!!
Just a small request – Could you please provide the slides or notes for your Statistics playlist ? Would be a great help.
A lot of the stuff from my statistics playlist can be found in my book, The StatQuest Illustrated Guide to Machine Learning: https://statquest.org/statquest-store/
Hi,
Thanks for all the great content in your youtube channel. I have bought the illustrated guide for ML v2.2, however I was surprised that it doesn’t include topics like Transformers, CNN, and other topics that are explained on Youtube. I hope you someday update the guide and share it as a free upgrade to those who already purchased it.
Thanks and keep up.
Regards
Converting the content in my videos into a book form takes a surprising amount of time. However, once I finish two more videos on neural networks, I hope to create an entire book about the topic that also includes code examples. Fingers crossed that I can get it done next year.
Hi Starmer!
Any update on this?
Yes, I’m about 3/4s of the way done writing a rough draft of the book. The hope is to have it out by the end of the year (or early 2025).
Thanks for the video!
In the explanation of the encoding matrix, at minute 7:00 or so, the words “what” and “StatQuest” have the same embedding values. Shouldn’t they be different?
In theory they would be different, but this example is just too simple to show all of the nuances in how transformers work.