The fall of RNN / LSTM

The fall of RNN / LSTM

  • May 13, 2018
Table of Contents

The fall of RNN / LSTM

It is the year 2014 and LSTM and RNN make a great come-back from the dead. We all read Colah’s blog and Karpathy’s ode to RNN. But we were all young and unexperienced.

For a few years this was the way to solve sequence learning, sequence translation (seq2seq), which also resulted in amazing results in speech to text comprehension and the raise of Siri, Cortana, Google voice assistant, Alexa. Also let us not forget machine translation, which resulted in the ability to translate documents into different languages or neural machine translation, but also translate images into text, text into images, and captioning video, and… well you got the idea. Then in the following years (2015–16) came ResNet and Attention.

One could then better understand that LSTM were a clever bypass technique. Also attention showed that MLP network could be replaced by averaging networks influenced by a context vector. But do not take our words for it, also see evidence that Attention based networks are used more and more by Google, Facebook, Salesforce, to name a few.

All these companies have replaced RNN and variants for attention based models, and it is just the beginning. RNN have the days counted in all applications, because they require more resources to train and run than attention-based models. See this post for more info.

Source: towardsdatascience.com

Tags :
Share :
comments powered by Disqus

Related Posts

Automatic Photography with Google Clips

Automatic Photography with Google Clips

How could we train an algorithm to recognize interesting moments? As with most machine learning problems, we started with a dataset. We created a dataset of thousands of videos in diverse scenarios where we imagined Clips being used.

Read More
Artificial Neural Nets Grow Brainlike Navigation Cells

Artificial Neural Nets Grow Brainlike Navigation Cells

Having the sense to take a shortcut, the most direct route from point A to point B, doesn’t sound like a very impressive test of intelligence. Yet according to a new report appearing today in Nature, in which researchers describe the performance of their new navigational artificial intelligence, the system’s ability to explore complex simulated environments and find the shortest route to a goal put it in a class previously reserved for humans and other living things. The surprising key to the system’s performance was that while learning how to navigate, the neural net spontaneously developed the equivalent of “grid cells,” sets of brain cells that enable at least some mammals to track their location in space.

Read More
Intel Starts R&D Effort in Probabilistic Computing for AI

Intel Starts R&D Effort in Probabilistic Computing for AI

Intel announced today that it is forming a strategic research alliance to take artificial intelligence to the next level. Autonomous systems don’t have good enough ways to respond to the uncertainties of the real world, and they don’t have a good enough way to understand how the uncertainties of their sensors should factor into the decisions they need to make. According to Intel CTO Mike Mayberry the answer is “probabilistic computing”, which he says could be AI’s next wave.

Read More