LSTM Architecture
"Machine intelligence is the last invention that humanity will ever need to make " -- Nick Bostrom As we have already discussed RNNs in my previous post, it's time we explore LSTMs for long memories. Since LSTM's work takes previous knowledge into consideration it would be good for you also to have a look at my previous article on RNNs ( relatable right?). Let's take an example, suppose I show you one image and after 2 mins I ask you about that image you will probably remember that image content, but if I ask about the same image some days later, the information might be fade or totally lost right? The first condition is where we need RNNs ( for shorter memories) while the other one is when we need LSTMs for long memory capacities. For more clarification let's take another one, suppose you are watching a movie without knowing its name ( e.g. Justice League) in one frame you See Ban Affleck and think this might be The Batman Movie, in another frame you see Gal Gadot and think this can be Wonder Women right?
May-28-2021, 22:35:55 GMT