16-889: Learning for 3D Vision (Spring 2022)


Assignmenet 5


Qichen Fu

Q1. Classification Model (40 points)


The test accuracy of my best model is 95.70%.

Some random test point clouds with their label and predcition (0: chairs, 1: vases and 2:lamps)


Label: 0, Prediction: 0
Label: 1, Prediction: 1
Label: 0, Prediction: 0
Label: 1, Prediction: 2
Label: 2, Prediction: 2

Some failure cases with their label and predcition (0: chairs, 1: vases and 2:lamps)


Label: 0, Prediction: 2
Label: 1, Prediction: 2
Label: 2, Prediction: 1
Label: 2, Prediction: 0
Label: 0, Prediction: 1

Q2. Segmentation Model (40 points)


The test accuracy of my best model is 89.26%.

Some random predictions (top row) and the corresponding ground truth (bottom row)



Object 246 prediction with an accuracy of 96%
Object 30 prediction with an accuracy of 97%
Object 436 prediction with an accuracy of 95%
Object 85 prediction with an accuracy of 82%
Object 97 prediction with an accuracy of 61%
Object 246 GT
Object 30 GT
Object 436 GT
Object 85 GT
Object 97 GT

Some bad predictions (top row) and the corresponding ground truth (bottom row)



Object 163 prediction with an accuracy of 28%
Object 235 prediction with an accuracy of 53%
Object 26 prediction with an accuracy of 52%
Object 351 prediction with an accuracy of 54%
Object 96 prediction with an accuracy of 54%
Object 163 GT
Object 235 GT
Object 26 GT
Object 351 GT
Object 96 GT

Q3. Robustness Analysis (20 points)


Classification Task

I first test the robustness by rotating each point cloud by a random degree (below the max_rot angle) for each of the x, y, z axis.

The quantitative results are shown below:

Max Random Rotate Angke (degree) Accuracy (%)
0 (Q1) 95.70
30 65.06
60 36.52
120 31.37
180 31.69

Some qualitative results and comparsions are shown below (no rotation points in the top row, random rotation under 60 degrees in the bottom row):



Label: 0, Prediction: 0
Label: 1, Prediction: 1
Label: 0, Prediction: 0
Label: 1, Prediction: 2
Label: 2, Prediction: 2
Label: 0, Prediction: 2
Label: 1, Prediction: 2
Label: 0, Prediction: 2
Label: 1, Prediction: 2
Label: 2, Prediction: 2

I then test the robustness by changing the number of points per object. The results are shown below:

Number of Points Accuracy (%)
10000 (Q1) 95.70
1000 95.49
100 93.39
10 72.29

Some qualitative results and comparsions are shown below (10000 points in the top row, 100 points in the bottom row):



Label: 0, Prediction: 0
Label: 1, Prediction: 1
Label: 0, Prediction: 0
Label: 1, Prediction: 2
Label: 2, Prediction: 2
Label: 0, Prediction: 0
Label: 1, Prediction: 2
Label: 0, Prediction: 0
Label: 1, Prediction: 2
Label: 2, Prediction: 2

Segmentation Task

I first test the robustness by rotating each point cloud by a random degree (below the max_rot angle) for each of the x, y, z axis. The results are shown below:

Max Random Rotate Angke (degree) Accuracy (%)
0 (Q2) 89.26
30 74.08
60 56.87
120 38.96
180 35.17

Some qualitative results and comparsions are shown below (no rotation points in the top row, random rotation under 60 degrees in the second row, ground truth in the bottom row):




Object 246 prediction with an accuracy of 96%
Object 30 prediction with an accuracy of 97%
Object 436 prediction with an accuracy of 95%
Object 85 prediction with an accuracy of 82%
Object 97 prediction with an accuracy of 61%
Object 246 prediction with an accuracy of 71%
Object 30 prediction with an accuracy of 58%
Object 436 prediction with an accuracy of 46%
Object 85 prediction with an accuracy of 13%
Object 97 prediction with an accuracy of 56%
Object 246 GT
Object 30 GT
Object 436 GT
Object 85 GT
Object 97 GT

I then test the robustness by changing the number of points per object. The results are shown below:

Number of Points Accuracy (%)
10000 (Q2) 89.26
1000 88.16
100 80.10
10 62.43

Some qualitative results and comparsions are shown below (10000 points in the top row, 100 points in the second row, ground truth in the bottom row):




Object 246 prediction with an accuracy of 96%
Object 30 prediction with an accuracy of 97%
Object 436 prediction with an accuracy of 95%
Object 85 prediction with an accuracy of 82%
Object 97 prediction with an accuracy of 61%
Object 246 prediction with an accuracy of 91%
Object 30 prediction with an accuracy of 96%
Object 436 prediction with an accuracy of 88%
Object 85 prediction with an accuracy of 58%
Object 97 prediction with an accuracy of 44%
Object 246 GT
Object 30 GT
Object 436 GT
Object 85 GT
Object 97 GT

As we shall observe in these results, the performance drops as the input is rotated for both classification and segmentation. This is because the model does not have rotation invarience. Thus, as the test data is rotated, it deviates from the training data distribution and makes the model hard to predict correctly. Meanwhile, the performance also drops as the number of points is decreased. This is because fewer number of points mean there are too few details to describe the point cloud for the model to make a good prediction.

Q4. Bonus Question - Locality (20 points)


I implemented the DGCNN model. Since DGCNN is taking much GPU memeory, I compute and compare the perfromance of DGCNN and PointNet under the condition where num_points = 1000 (different than num_points = 10000 in Q1&Q2). The accuracies and comparisions with PointNet are shown.

Task Method Accuracy (%)
Classification PointNet 95.28
Classification DGCNN 96.22
Segmentation PointNet 88.52
Segmentation DGCNN 89.87

Some qualitative results and comparsions of the classification task are shown below (PointNet in the top row, DGCNN in the bottom row):



Label: 0, Prediction: 0
Label: 1, Prediction: 1
Label: 0, Prediction: 0
Label: 1, Prediction: 2
Label: 2, Prediction: 2
Label: 0, Prediction: 0
Label: 1, Prediction: 1
Label: 0, Prediction: 0
Label: 1, Prediction: 1
Label: 2, Prediction: 2

Some qualitative results and comparsions of the segmentation task are shown below (PointNet in the top row, DGCNN in the second row, ground truth in the bottom row):




Object 246 prediction with an accuracy of 96%
Object 30 prediction with an accuracy of 97%
Object 436 prediction with an accuracy of 95%
Object 85 prediction with an accuracy of 75%
Object 97 prediction with an accuracy of 61%
Object 246 prediction with an accuracy of 97%
Object 30 prediction with an accuracy of 97%
Object 436 prediction with an accuracy of 96%
Object 85 prediction with an accuracy of 87%
Object 97 prediction with an accuracy of 70%
Object 246 GT
Object 30 GT
Object 436 GT
Object 85 GT
Object 97 GT

As we can observe, the performance of DGCNN is better than the PointNet because it reasons about the local feature of the point cloud.