Long Short-Term Memory (LSTM), Clearly Explained!!!

NOTE: This StatQuest was supported by these awesome people who support StatQuest at the Double BAM level: S. Kundapurkar, JWC, B. Bellman, BufferUnderrun, S. Jeffcoat, S. Handschuh, J. Le, D. Greene, D. Schioberg, Magpie, Z. Rosenberg, J. N., H-M Chang, F. Pedemonte, S. Song US, A. Tolkachev, L. Cisterna, J. Alexander, J. Varghese, K. Manickam

5 thoughts on “Long Short-Term Memory (LSTM), Clearly Explained!!!

  1. Hi this is absolutely golden for me, also if I’ll have to implement these from scratch for a thing I’m after.

    I’m a newb at ML (I can understand the terminology and stuff after years of casually reading about it).

    While working on an innovative/new type of game I have invented a new fuzzy search algorithm which could theoretically replace CTC or maybe even do more fancy stuff.

    I’d like to test it out as a loss function for an LSTM and then run some benchmarks.

    I’m currently stuck with it in Swift (using Accelerate.vDSP on M1 CPUs) in a very advanced state (it even comes out with 0 to 1 closeness values and lots of possible matches) – prolly I won’t even have to port it to Python (needs a lot of complicated initialization, its a lot of code) and a local machine RPC would suffice for the testing.

    Could you point me in the right direction? Like to working LSTM exmaple with CTC for NLP/chat bot (not speech or handwriting as it would be difficult to set up I think) and a freely available dataset?

Leave a Reply

Your email address will not be published. Required fields are marked *