Difference between revisions of "PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space"

From statwiki
Jump to: navigation, search
(Added to Semantic Scene Labeling)
(Random Input Point Dropout)
Line 79: Line 79:
[[File:grouping.png | 300px|thumb|center|Example of the two ways to perform grouping]]
[[File:grouping.png | 300px|thumb|center|Example of the two ways to perform grouping]]
Moreover, "random input point dropout" was proposed. Each input point has a dropout probability <math>\theta</math>. The authors used a <math>\theta</math> value of 0.95. As shown in the experiments section below, dropout provides robustness to input point density variations. During testing stage all points are used.
== Point Cloud Segmentation ==
== Point Cloud Segmentation ==

Revision as of 12:17, 22 March 2018


This paper builds off of ideas from PointNet (Qi et al., 2017). The name PointNet is derived from the network's input - a point cloud. A point cloud is a set of three dimensional points that each have coordinates [math] (x,y,z) [/math]. These coordinates usually represent the surface of an object. For example, a point cloud describing the shape of a torus is shown below.

Point cloud torus

Processing point clouds is important in applications such as autonomous driving where point clouds are collected from an onboard LiDAR sensor. These point clouds can then be used for object detection. However, point clouds are challenging to process because:

  1. They are unordered. If [math] N [/math] is the number of points in a point cloud, then there are [math] N! [/math] permutations that the point cloud can be represented.
  2. The spatial arrangement of the points contains useful information, thus it needs to be encoded.
  3. The function processing the point cloud needs to be invariant to transformations such as rotation and translations of all points.

Previously, typical point cloud processing methods handled the challenges of point clouds by transforming the data with a 3D voxel grid or by representing the point cloud with multiple 2D images. When PointNet was introduced, it was novel because it directly took points as its input. PointNet++ improves on PointNet by using a hierarchical method to better capture local structures of the point cloud.

Examples of point clouds and their associated task. Classification (left), part segmentation (centre), scene segmentation (right)

Review of PointNet

The PointNet architecture is shown below. The input of the network is [math] n [/math] points, which each have [math] (x,y,z) [/math] coordinates. Each point is processed individually through a multi-layer perceptron (MLP). This network creates an encoding for each point; in the diagram, each point is represented by a 1024 dimension vector. Then, using a max pool layer a vector is created that represents the "global signature" of a point cloud. If classification is the task, this global signature is processed by another MLP to compute the classification scores. If segmentation is the task, this global signature is appended to to each point from the "nx64" layer, and these points are processed by a MLP to compute a semantic category score for each point.

The core idea of the network is to learn a symmetric function on transformed points. Through the T-Nets and the MLP network, a transformation is learned with the hopes of making points invariant to point cloud transformations. Learning a symmetric function solves the challenge imposed by having unordered points; a symmetric function will produce the same value no matter the order of the input. This symmetric function is represented by the max pool layer.

PointNet architecture. The blue highlighted region is when it is used for classification, and the beige highlighted region is when it is used for segmentation.


The motivation for PointNet++ is that PointNet does not capture local, fine-grained details. Since PointNet performs a max pool layer over all of its points, information such as the local interaction between points is lost.

Problem Statement

There is a metric space [math] X = (M,d) [/math] where [math]d[/math] is the metric from a Euclidean space [math]\pmb{\mathbb{R}}^n[/math] and [math] M \subseteq \pmb{\mathbb{R}}^n [/math] is the set of points. The goal is to learn a function that takes [math]X[/math] as the input and output a class or per point label to each member of [math]M[/math].


High Level Overview

PointNet++ architecture

The PointNet++ architecture is shown on the right. The core idea is that a hierarchical architecture is used and at each level of the hierarchy a set of points is processed and abstracted to a new set with less points, i.e.,

\begin{aligned} \text{Input at each level: } N \times (d + c) \text{ matrix} \end{aligned}

where [math]N[/math] is the number of points, [math]d[/math] is the coordinate points [math](x,y,z)[/math] and [math]c[/math] is the feature representation of each point, and

\begin{aligned} \text{Output at each level: } N' \times (d + c') \text{ matrix} \end{aligned}

where [math]N'[/math] is the new number (smaller) of points and [math]c'[/math] is the new feature vector.

Each level has three layers: Sampling, Grouping, and PointNet. The Sampling layer selects points that will act as centroids of local regions within the point cloud. The Grouping layer then finds points near these centroids. Lastly, the PointNet layer performs PointNet on each group to encode local information.

Sampling Layer

The input of this layer is a set of points [math]{\{x_1,x_2,...,x_n}\}[/math]. The goal of this layer is to select a subset of these points [math]{\{\hat{x}_1, \hat{x}_2,...,\hat{x}_m\}} [/math] that will define the centroid of local regions.

To select these points farthest point sampling is used. This is where [math]\hat{x}_j[/math] is the most distant point with regards to [math]{\{\hat{x}_1, \hat{x}_2,...,\hat{x}_{j-1}\}}[/math]. This ensures coverage of the entire point cloud opposed to random sampling.

Grouping Layer

The objective of the grouping layer is to form local regions around each centroid by grouping points near the selected centroids. The input is a point set of size [math]N \times (d + c)[/math] and the coordinates of the centroids [math]N' \times d[/math]. The output is the groups of points within each region [math]N' \times k \times (d+c)[/math] where [math]k[/math] is the number of points in each region.

Note that [math]k[/math] can vary per group. Later, the PointNet layer creates a feature vector that is the same size for all regions at a hierarchical level.

To determine which points belong to a group a ball query is used; all points within a radius of the centroid are grouped. This is advantageous over nearest neighbour because it guarantees a fixed region space, which is important when learning local structure.

PointNet Layer

After grouping, PointNet is applied to the points. However, first the coordinates of points in a local region are converted to a local coordinate frame by [math] x_i = x_i - \bar{x}[/math] where [math]\bar{x}[/math] is the coordinates of the centroid.

Robust Feature Learning under Non-Uniform Sampling Density

The previous description of grouping uses a single scale. This is not optimal because the density varies per section of the point cloud. At each level, it would be better if the PointNet layer was applied to adaptively sized groups depending on the point cloud density.

The two grouping methods the authors propose are shown in the diagram below. Multi-scale grouping (MSG) applies PointNet at various scales per group. The features from the various scales are concatenated. This method, however, is computationally expensive because for each region it always applies PointNet to all points. On the other hand, multi-resolution grouping (MRG) is less computationally expensive but still adaptively collects features. As shown in the diagram, the left vector is obtained by applying PointNet to three points, and these three points obtained information from three groups. This vector is then concatenated by a vector that is created by using PointNet on all the points in the level below. The second vector can be weighed more heavily if the first vector contains a sparse amount of points.

Example of the two ways to perform grouping

Moreover, "random input point dropout" was proposed. Each input point has a dropout probability [math]\theta[/math]. The authors used a [math]\theta[/math] value of 0.95. As shown in the experiments section below, dropout provides robustness to input point density variations. During testing stage all points are used.

Point Cloud Segmentation

If the task is segmentation, the architecture is slightly modified since we want a semantic score for each point. To achieve this, distance-based interpolation and skip-connections are used.

Distance-based Interpolation

Here, point features from [math]N_l \times (d + C)[/math] points are propagated to [math]N_{l-1} \times (d + C)[/math] points where [math]N_{l-1}[/math] is greater than [math]N_l[/math].

To propagate features an inverse distance weighted average based on [math]k[/math] nearest neighbors is used. The [math]p=2[/math] and [math]k=3[/math].

Feature interpolation during segmentation


In addition, skip connections are used (see the PointNet++ architecture diagram). The features from the the skip layers are concatenated with the interpolated features. Next, a "unit-wise" PointNet is applied, which the authors describe as similar to a one-by-one convolution.


To validate the effectiveness of PointNet++, experiments in three areas were performed - classification in Euclidean metric space, semantic scene labelling, and classification in non-Euclidean space.

Point Set Classification in Euclidean Metric Space

The digit dataset, MNIST, was converted to a 2D point cloud. Pixel intensities were normalized in the range of [math][0, 1][/math], and only pixels with intensities larger than 0.5 were considered. The coordinate system was set at the centre of the image. PointNet++ achieved a classification error of 0.51%. The original PointNet had 0.78% classification error. The table below compares these results to the state-of-the-art.

MNIST classification results.

In addition, the ModelNet40 dataset was used. This dataset consists of CAD models. Three dimensional point clouds were sampled from mesh surfaces of the ModelNet40 shapes. The classification results from this dataset are shown below.

ModelNet40 classification results.

An experiment was performed to show how the accuracy was affected by the number of points used. With PointNet++ using multi-scale grouping and dropout, the performance decreased by less than 1% when 1024 test points were reduced to 256. On the other hand, PointNet's performance was impacted by the decrease in points.

An example showing the reduction of points visually. At 256 points, the points making up the object is very spare, however the accuracy is only reduced by 1%
Relationship between accuracy and the number of points used for classification.

Semantic Scene Labelling

The ScanNet dataset was used for experiments in semantic scene labelling. This dataset consists of laser scans of indoor scenes where the goal is to predict a semantic label for each point. Example results are shown below.

Example ScanNet semantic segmentation results.

To compare to other methods, the authors convert their point labels to a voxel format, and accuracy is determined on a per voxel basis. The accuracy compared to other methods is shown below.

ScanNet semantic segmentation classification comparison to other methods.

To test how the trained model performed on scans with non-uniform sampling density, virtual scans of Scannet scenes were synthesized and the network was evaluated on this data. It can be seen from the above figures that SSG performance greatly falls due to the sampling density shift. MRG network, on the other hand, is more robust to the sampling density shift since it is able to automatically switch to features depicting coarser granularity when the sampling is sparse. This proves the effectiveness of the proposed density adaptive layer design.

Classification in Non-Euclidean Metric Space

Example of shapes from the SHREC15 dataset.

Lastly, experiments were performed on the SHREC15 dataset. This dataset contains shapes that have different poses. This experiment shows that PointNet++ is able to generalize to non-Euclidean spaces. Results from this dataset are provided below.

Results from the SHREC15 dataset.

Feature Visualization

The figure below visualizes what is learned by just the first layer kernels of the network. The model is trained on a dataset the mostly consisted of furniture which explains the lines, corners, and planes visible in the visualization. Visualization is performed by creating a voxel grid in space and only aggregating point sets that activate specific neurons the most.

Pointclouds learned from first layer kernels (red is near, blue is far)


It seems clear that PointNet is lacking capturing local context between points. PointNet++ seems to be an important extension, but the improvements in the experimental results seem small. Some computational efficiency experiments would have been nice. For example, the processing speed of the network, and the computational efficiency of MRG over MRG.


Code for PointNet++ can be found at: https://github.com/charlesq34/pointnet2


1. Charles R. Qi, Li Yi, Hao Su, Leonidas J. Guibas. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space, 2017

2. Charles R. Qi, Hao Su, Kaichun Mo, Leonidas J. Guibas. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation, 2017