Home Work 4: FNU ABHIMANYU

(abhiman2@andrew.cmu.edu)

Collaborated with Ananya Bal

1. Point cloud classification

Test accuracy: 0.962

These are the 3 pictures for which the predictions are correct. They are chair, vase and lamp

This is actually a chair but predicted as a vase. I think this is predicted because of the width of the chair is comparable to a vase.

This is actually a vase but predicted as a lamp. I think this is predicted because of the resemblance to the lamp.

This is actually a lamp but predicted as a vase. I think this is predicted because of the resemblance to the vase because of its width similar to the vase

2. Point cloud Segmentation

The test accuracy is: 0.899

The top gif is ground truth and lower gif is prediction

The top gif is ground truth and lower gif is prediction

The top gif is ground truth and lower gif is prediction

The top gif is ground truth and lower gif is prediction. The prediction is wrong for the sitting area of the chair. The ground truth is marked as red and the prediction coes out to be blue.

The top gif is ground truth and lower gif is prediction. The prediction is wrong for the handle of the chair. The ground truth is marked as yellow and the prediction coes out to be red.

3. Analysis

This part of assignment deals with affect of rotation on classification. In this, I will rotate the point cloud in x,y,z axis from [0,3.14]rad in an interval of 0.3 rad

X-axis:

The classification score is max at 0 deg and decreases and becomes minimum around 1.57 deg. Then the classification accuracy incrases till 3.14. This shows the non-generalizibilty of network oon x-axis rotation. The avg accuracy is: 0.572

Y-axis:

The classification score is max at 0 deg and decreases till 3.14. There is slight increase around 1.57 rad. This shows the non-generalizibilty of network on y-axis rotation. The avg accuracy is: 0.59

Z-axis:

The classification score is max at 0 deg and decreases till 3.14. There is slight increase around 1.57 rad. This shows the non-generalizibilty of network on z-axis rotation. The avg accuracy is: 0.502. The curve is very similar to y-axis rotation

One more interesting observation made:

For this lamp, the classification predicts as vase at this perspective(0 deg). But when the lamp is actually rotated and the 4 arms of the lamps are shown, the prediction is a lamp. This shows one of the cases where classification gets correct because more relevant features are visible.

Image1: Original non-rotated image, Pred: Vase

Image2: Rotated image, Pred: Lamp

Points vs accuracy analysis

This comparision is about num points vs classification accuracy. The classification accuracy is robust for most of the range of the point cloud, till num_points=10, where the point cloud loses most of the key features.

Segmentation task

Similarly for segmentation task, the following accuracy drops as we rotate in x,y and z axis from [0,3] rad at an inteval of 0.5 rads

This comparision is about num points vs classification accuracy. The classification accuracy is robust for most of the range of the point cloud, till num_points=10, where the point cloud loses most of the key features.

4.1. PointNet ++

I have implemented PointNet++ model for 10000 points pointcloud. The model looks as follows: SA(K,r,[l1, ..., ld]) is a set abstraction (SA) level with K local regions of ball radius r using PointNet of d fully connected layers with width li (i = 1, ..., d).

SA(1000, 0.2, [64, 64, 128]) → SA(100, 0.4, [128, 128, 256]) → SA([256, 512, 1024]) → F C(512, 0.5) → F C(256, 0.5) → F C(Class_num) where Class_num is 3.

Though the ball radius is set to 0.2, I have taken the best 50 points in the ball radius, due to GPU space limitation.

The accuracy achieved for PoinNet ++ is : 0.947

Counter intuitevly the accuracy is lower than PointNet. This is mostly because of two implementation details:

  1. Number of regions selected per iteration (10k,1k,100). Its different from the paper.
  2. Number of points in the ball. I am taking 50 points because of GPU space constraint

The trend is very similar for num_points vs accuracy as seen in Q1

For rotational ananlysis, I just tried the x rotation and the trend is similar as shown in Q3