3D Face Reconstruction with Position Map Regression Networks

3D Face Reconstruction with Position Map Regression Networks

  • May 24, 2018
Table of Contents

3D Face Reconstruction with Position Map Regression Networks

Position Map Regression Networks (PRN) is a method to jointly regress dense alignment and 3D face shape in an end-to-end manner. In this article, I’ll provide a short explanation and discuss its applications in computer vision. In the last few decades, a lot of important research groups in computer vision have made amazing advances in 3D face reconstruction and face alignment.

Primarily, these groups have used CNNs as the de facto ANN for this task. However, the performance of these methods is restricted because of the limitations of 3D space defined by face model templates used for mapping. In a recent paper, Yao Feng and others proposed an end-to-end method called Position Map Regression Networks (PRN) to jointly predict dense alignment and reconstruct 3D face shape.

They claim their method surpasses all previous attempts at both 3D face alignment and reconstruction on multiple datasets. Specifically, they designed a UV position map, which is a 2D image recording the 3D coordinates of a complete facial point cloud, which maintains the semantic meaning at each UV polygon. They then train a simple encoder-decoder network with a weighted loss that focuses more on discriminative region to regress the UV position map from a single 2D facial image.

In a recent paper, Yao Feng and others proposed an end-to-end method called Position Map Regression Networks (PRN) to jointly predict dense alignment and reconstruct 3D face shape. They claim their method surpasses all previous attempts at both 3D face alignment and reconstruction on multiple datasets.

Source: fritz.ai

Tags :
Share :
comments powered by Disqus

Related Posts

A Deep Dive into Monte Carlo Tree Search

A Deep Dive into Monte Carlo Tree Search

The very first Go AIs used multiple modules to handle each aspect of playing Go – life and death, capturing races, opening theory, endgame theory, and so on. The idea was that by having experts program each module using heuristics, the AI would become an expert in all areas of the game. All that came to a grinding halt with the introduction of Monte Carlo Tree Search (MCTS) around 2008.

Read More
Tensor Compilers: Comparing PlaidML, Tensor Comprehensions, and TVM

Tensor Compilers: Comparing PlaidML, Tensor Comprehensions, and TVM

One of the most complex and performance critical parts of any machine learning framework is its support for device specific acceleration. Indeed, without efficient GPU acceleration, much of modern ML research and deployment would not be possible. This acceleration support is also a critical bottleneck, both in terms of adding support for a wider range of hardware targets (including mobile) as well as for writing new research kernels.

Read More