Sequence Tagging with Tensorflow

Sequence Tagging with Tensorflow

  • March 14, 2018
Table of Contents

Sequence Tagging with Tensorflow

I remember the first time I heard about the magic of Deep Learning for Natural Language Processing (NLP). I was just starting a project with a young French startup Riminder and it was the first time I heard about word embeddings. There are moments in life when the confrontation with a new theory seems to make everything else irrelevant.

Hearing about word vectors that encode similarity and meaning between words was one of these moments. I was baffled by the simplicity of the model as I started to play with these new concepts, building my first recurrent neural network for sentiment analysis. A few months later, as part of the master thesis of my master in the French university Ecole polytechnique I was working on more advanced models for sequence tagging at Proxem.

Source: github.io

Tags :
Share :
comments powered by Disqus

Related Posts

Understanding word vectors in NLP

Understanding word vectors in NLP

In this tutorial, I’m going to show you how word vectors work. This tutorial assumes a good amount of Python knowledge, but even if you’re not a Python expert, you should be able to follow along and make small changes to the examples without too much trouble.

Read More
Building a Next Word Predictor in Tensorflow

Building a Next Word Predictor in Tensorflow

Next Word Prediction or what is also called Language Modeling is the task of predicting what word comes next. It is one of the fundamental tasks of NLP and has many applications. You might be using it daily when you write texts or emails without realizing it.

Read More