A simple overview of RNN, LSTM and Attention Mechanism



Original Source Here

RNNs to the rescue

There’s something magical about Recurrent Neural Networks.

— Andrej Karpathy

Recurrent Neural Networks address this drawback of vanilla NNs with a simple yet elegant mechanism and are great at modeling sequential data.

https://www.reddit.com/r/plantvstruaxlivesst/
https://www.reddit.com/r/Truaxvplantlivestt/
https://www.reddit.com/r/plantvtruaxlivetrems/
https://www.reddit.com/r/royalrumblelivest/
https://www.reddit.com/live/16eim0y7kjifw/
https://www.reddit.com/live/16eim7sr7ldzk/
https://www.reddit.com/live/16egzf3bnp4bm/
https://www.reddit.com/live/16ejaddovktbe/

https://www.reddit.com/live/16ejakqnhgurl/

https://www.reddit.com/live/16ejaro91lw0n/

https://www.reddit.com/live/16egznd2ekrm5/

https://www.reddit.com/live/16egzxr0q8qct/

https://www.reddit.com/live/16eh04ujjrcoy/

https://www.reddit.com/live/16eh0b61schsf/

https://www.reddit.com/live/16eh0jmrbrwav/

https://app.livestorm.co/czxczxc/watch-caleb-truax-vs-caleb-plant-live-stream-free-boxing-full-fight-online/

https://app.livestorm.co/czxczxc/crackstreams-caleb-plant-vs-caleb-truax-full-fight-live-streams-reddit-buffstreams

https://app.livestorm.co/czxczxc/boxingtv-plant-vs-truax-fight-live-streams-reddit-free-watch-boxing-full-match-2021

https://app.livestorm.co/czxczxc/watch-plant-vs-truax-live-strem-2021-boxing-free

https://medium.com/@ruth_forsyth/i-accidentally-sent-a-sexually-explicit-text-to-my-employer-2ca51dd3ab24

https://ruth-forsyth.medium.com/i-accidentally-sent-a-sexually-explicit-text-to-my-employer-2ca51dd3ab24

https://www.deviantart.com/ppv-live/journal/Buffstreams-Plant-vs-Truax-Live-Stream-On-Reddit-868780556

https://www.deviantart.com/ppv-live/journal/Boxing-Fight-Plant-vs-Truax-Live-Stream-reddit-868780615

https://www.deviantart.com/ppv-live/journal/WATCH-CALEB-TRUAX-VS-CALEB-PLANT-LIVE-STREAM-FRE-868780707

https://www.deviantart.com/ppv-live/journal/BOXING-HD-Watch-Plant-vs-Truax-Live-On-Anywhere-868780806

https://www.deviantart.com/ppv-live/journal/OFFICIAL-LIVE-Truax-vs-Plant-Live-Stream-868780874

https://www.deviantart.com/ppv-live/journal/T-V-Plant-vs-Truax-Live-Stream-Boxing-868780976

https://www.deviantart.com/ppv-live/journal/H-D-Plant-vs-Truax-Live-Stream-Boxing-868781080

https://www.deviantart.com/ppv-live/journal/Watch-Plant-vs-Truax-Live-Stream-2021-868781148

https://www.deviantart.com/ppv-live/journal/BOxiNG-Plant-vs-Truax-Live-Stream-2021-868781186

https://www.deviantart.com/ppv-live/journal/PPV-Plant-vs-Truax-Live-Stream-2021-868781249

https://www.deviantart.com/ppv-live/journal/LIVE-Plant-vs-Truax-Live-Stream-Free-2021-868781309

https://www.deviantart.com/ppv-live/journal/PPV-Livestream-Plant-vs-Truax-Live-Free-On-Fox-Tv-868781400

rolled RNN

Does RNN look weird to you? Let me explain and remove the confusion.

Take a simple feed forward Neural Network first, shown below. It has the input coming in, red dot, to the hidden layer, blue dot, which results in black dot output.

A simple NN

An RNN feeds it’s output to itself at next time-step, forming a loop, passing down much needed information.

RNN feeding hidden state value to itself

To better understand the flow, look at the unrolled version below, where each RNN has different input (token in a sequence) and output at each time step.

Unrolled RNN, from time step 0 to t

The NN A takes in input at each time step while giving output h and passing information to itself for next incoming input t+1 step. The incoming sequential data is encoded using RNN first before being utilized to determine the intent/action via another feed forward network for decision.

RNNs have become the go-to NNs to be used for various tasks involving notion of sequential data, such as: speech recognition, language modeling, translation, image captioning etc.

Let’s say we ask a question to your in-house developed AI Assistant named Piri (or whatever ‘iri’ you prefer), “what time is it?”, here we try to break the sequence and color code it.

How RNNs work for the tokens of sequence

Final query retrieved as a result of processing the entire sequence

Memory: An essential requirement for making Neural Networks smart(er)

Humans tend to retrieve information from memory, short or long, use current information with it and derive logic to take next action (or impulse/habit, again based on previous experiences).

Similar is the idea to make RNN hold on to previous information or state(s). As the output of a recurrent neuron, at a given time step t, is clearly a function of the previous input (or think of it as previous input with accumulated information) till time step t-1, one could consider this mechanism as a form of memory. Any part of a neural network that has the notion of preserving state, even partially, across time steps is usually referred to as a memory-cell.

Each recurrent neuron has an output as well as a hidden state which is passed to next step neuron.

hidden state

Unrolled RNN with hidden state and output at each time step

AI/ML

Trending AI/ML Article Identified & Digested via Granola by Ramsey Elbasheer; a Machine-Driven RSS Bot



via WordPress https://ramseyelbasheer.wordpress.com/2021/01/30/a-simple-overview-of-rnn-lstm-and-attention-mechanism/

Popular posts from this blog

I’m Sorry! Evernote Has A New ‘Home’ Now

Jensen Huang: Racism is one flywheel we must stop

Streamlit — Deploy your app in just a few minutes