Deep Learning Research: Creating Adaptable Meta-Learning Models

Deep Learning Research: Creating Adaptable Meta-Learning Models

  • May 24, 2018
Table of Contents

Deep Learning Research: Creating Adaptable Meta-Learning Models

Adaptability is one of the key cognitive abilities that defined us as humans. Even as babies, we can intuitively shift between similar tasks even if we don’t have prior training on them. This contrasts with the traditional train-and-test approach of most artificial intelligence(AI) systems which require an agent to go through massive amounts of training before it can master a specific task.

By definition, train-and-test systems are not very adaptable and, consequently, they are not very applicable to scenarios that operate in real word environments. Improving the adaptability of AI systems has been one of the core areas of research of an increasingly popular discipline known as meta-learning that focuses on improving the learning abilities of AI agents. In neuroscience literature, cognitive flexibility or adaptability is commonly ascribed to prefrontal cortex (PFC) and working memory in the brain.

Neuroscientific evidence suggests that these areas use incoming information to support task-specific temporal adaptation and planning. Recently, researchers from the Microsoft AI lab in Montreal published a research paper that mimics some of these neuroscientific principles in a new technique known as cognitive shifted neurons(CSN).

Source: towardsdatascience.com

Tags :
Share :
comments powered by Disqus

Related Posts

A Deep Dive into Monte Carlo Tree Search

A Deep Dive into Monte Carlo Tree Search

The very first Go AIs used multiple modules to handle each aspect of playing Go – life and death, capturing races, opening theory, endgame theory, and so on. The idea was that by having experts program each module using heuristics, the AI would become an expert in all areas of the game. All that came to a grinding halt with the introduction of Monte Carlo Tree Search (MCTS) around 2008.

Read More
Tensor Compilers: Comparing PlaidML, Tensor Comprehensions, and TVM

Tensor Compilers: Comparing PlaidML, Tensor Comprehensions, and TVM

One of the most complex and performance critical parts of any machine learning framework is its support for device specific acceleration. Indeed, without efficient GPU acceleration, much of modern ML research and deployment would not be possible. This acceleration support is also a critical bottleneck, both in terms of adding support for a wider range of hardware targets (including mobile) as well as for writing new research kernels.

Read More
The Nengo Neural Simulator

The Nengo Neural Simulator

Nengo is a graphical and scripting based Python package for simulating large-scale neural networks. Nengo can create sophisticated spiking or non-spiking neural simulations with sensible defaults in a few lines of code. Yet, Nengo is highly extensible and flexible.

Read More