16-889 Assignment 5

Byeongjoo Ahn (bahn@andrew.cmu.edu)

Late day:

Q1. Classification Model (40 points)


Test accuracy: 97.69%

Correct predictions:

Correct prediction: Chair
Correct prediction: Vase
Correct prediction: Lamp

Incorrect predictions:

GT: Chair | Prediction: Lamp
GT: Vase | Prediction: Lamp
GT: Lamp | Prediction: Vase

Interpretation:

While we achieve high test accuracy of 97.69%, the classification network provides incorrect prediction when the input point cloud has uncommon shapes. For example, when the leg of chair has a shape of X instead of general four legs, the network classifies it as a lamp. The additional feature at the bottom part of the lamp also causes the incorrect prediction of lamp. Similarly, when the bottom part of the lamp is flat, the network classifies it as a vase.

Q2. Segmentation Model (40 points)


Test accuracy: 87.24%

Good predictions:

GT
Prediction (99.09%)
GT
Prediction (98.49%)
GT
Prediction (98.32%)
Prediction (98.07%)

Bad predictions:

GT
Prediction (34.39%)
GT
Prediction (36.80%)
GT
Prediction (38.64%)
GT
Prediction (40.24%)

Interpretation:

The visualization of good and bad predictions show the top 4 best and worst results. Our network provides accurate results when the target shape is simple (e.g., total number of segments is small) as shown in the good predictions. On the other hand, it provides bad results when the number of segments is large and the target shape has multiple parts including stools.

Q3. Robustness Analysis (20 points)


Q3.1. Robustness to Rotation

We analyze how our model is robust to rotations. We rotate the point cloud by 45, 90 degrees in each axis and compare the test accuracy to the result without rotation.

Classification (97.69% without rotation)

angle \ axisX-axisY-axisZ-axis
4545^\circ40.163.06%67.68%
9090^\circ40.82%47.64%24.24%

The failure cases for different rotations are visualized. The worst result is obtained when we rotate the point cloud by 90 degrees with respect to the Z-axis. Adding transformation blocks can help the dependency on rotations.

[4545^\circ in X-axis] GT= Chair | Prediction = Vase
[4545^\circ in Y-axis] GT= Chair | Prediction = Vase
[4545^\circ in Z-axis] GT= Chair | Prediction = Vase
[9090^\circ in X-axis] GT= Chair | Prediction = Vase
[9090^\circ in Y-axis] GT= Vase | Prediction = Lamp
[9090^\circ in Z-axis] GT= Chair | Prediction = Vase

Segmentation (87.24% without rotation)

angle \ axisX-axisY-axisZ-axis
4545^\circ67.95%75.37%65.44%
9090^\circ28.15%65.24%43.43%

The best results for different rotations are visualized. When the number of segments is small, we can achieve good test accuracy even when the input point cloud is rotated.

[4545^\circ in X-axis]
[4545^\circ in Y-axis]
[4545^\circ in Z-axis]

[9090^\circ in X-axis]
[9090^\circ in Y-axis]
[9090^\circ in Z-axis]

The worst results for different rotations are visualized. The worst results are not obtained when the object shape is irregular (e.g., including stools), which is different with the case without rotation. The segmentation results show that they are segmented just based on the location of the points (e.g., yellow at top, red at middle), not considering the global structure.

[4545^\circ in X-axis]
[4545^\circ in Y-axis]
[4545^\circ in Z-axis]
[9090^\circ in X-axis]
[9090^\circ in Y-axis]
[9090^\circ in Z-axis]

Overall, rotating the input point cloud by 90 degrees in a certain axis provide the worst results, but the classification is worst in the rotation w.r.t. Z-axis, and the segmentation is worst in the rotation w.r.t. X-axis. This discrepancy implies good segmentation requires something more than good global feature used in the segmentation. As the current implementation relies on the single global feature for all points, better representation that involves locality can improve the segmentation results.

Q3.2. Robustness to Number of Points

We analyze how our model is robust to the number of points. We change the number of points used for evaluation and compare the test accuracy to the result with the number of points of 10000.

Classification

Number of pointsClassification Accuracy
1024.87%
5072.40%
10090.77%
50096.22%
100096.75%
500097.59%
1000097.69%

The failure cases for different number of points are visualized

[#points = 10] GT = Vase | Prediction = Lamp
[#points = 50] GT = Vase | Prediction = Lamp
[#points = 100] GT = Vase | Prediction = Lamp
[#points = 500] GT = Lamp | Prediction = Vase
[#points = 500] GT = Lamp | Prediction = Vase
[#points = 500] GT = Lamp | Prediction = Vase

Segmentation

Number of pointsSegmentation Accuracy
1052.53%
5074.12%
10079.24%
50085.62%
100086.88%
500087.22%
1000087.24%

The best results for each number points are visualized. When the number of segments is small, we can achieve good test accuracy even when the number of points are small.

#points = 10
#points = 50
#points = 100
#points = 500
#points = 1000
#points = 5000

The worst results for each number of points are visualized. When #points > 100, we observe the worst result is obtained when the object shape is irregular (e.g., including stools). However, when the #points is small, point clouds of simple shape show the worst result because of the lack of global information.

#points = 10
#points = 50
#points = 100
#points = 500
#points = 1000
#points = 5000

Overall, we observe that the test accuracy increases as we increase the number of points both in classification and segmentation. The test accuracy increases rapidly until the number of points becomes 500.