Understanding Convolutional Neural Networks

Understanding Convolutional Neural Networks

  • September 14, 2019
Table of Contents

Understanding Convolutional Neural Networks

A Convolutional Neural Network (CNN) is a class of deep, feed-forward artificial neural networks most commonly applied to analyzing visual imagery. The architecture of these networks was loosely inspired by biological neurons that communicate with each other and generate outputs dependent on the inputs. Though work on CNNs started in the early 1980s, they only became popular with recent technology advancements and computational capabilities that allow the processing of large amounts of data and the training of sophisticated algorithms in a reasonable amount of time.

Some of the applications of CNNs include AI-based virtual assistant, automatic photo tagging, video labeling, and self-driving cars. This blog assumes that you have a basic knowledge of neural networks. You can also check out Introduction to convolutional neural networks, which covers everything you need to know for this post.

Source: cloudera.com

Tags :
Share :
comments powered by Disqus

Related Posts

Deep probabilistic modelling with Pyro

Deep probabilistic modelling with Pyro

Classical machine learning and deep learning algorithms can only propose the most probable solutions and are not able to adequately model uncertainty. The success of deep neural networks in diverse areas as image recognition and natural language processing has been outstanding in recent years. However, classical machine learning and deep learning algorithms can only propose the most probable solutions and are not able to adequately model uncertainty.

Read More
First Programmable Memristor Computer

First Programmable Memristor Computer

Michigan team builds memristors atop standard CMOS logic to demo a system that can do a variety of edge computing AI tasks Hoping to speed AI and neuromorphic computing and cut down on power consumption, startups, scientists, and established chip companies have all been looking to do more computing in memory rather than in a processor’s computing core. Memristors and other nonvolatile memory seem to lend themselves to the task particularly well. However, most demonstrations of in-memory computing have been in standalone accelerator chips that either are built for a particular type of AI problem or that need the off-chip resources of a separate processor in order to operate.

Read More
Replay in biological and artificial neural networks

Replay in biological and artificial neural networks

Our waking and sleeping lives are punctuated by fragments of recalled memories: a sudden connection in the shower between seemingly disparate thoughts, or an ill-fated choice decades ago that haunts us as we struggle to fall asleep. By measuring memory retrieval directly in the brain, neuroscientists have noticed something remarkable: spontaneous recollections, measured directly in the brain, often occur as very fast sequences of multiple memories. These so-called ‘replay’ sequences play out in a fraction of a second–so fast that we’re not necessarily aware of the sequence.

Read More