Test accuracy: 0.962
These are the 3 pictures for which the predictions are correct. They are chair, vase and lamp
This is actually a chair but predicted as a vase. I think this is predicted because of the width of the chair is comparable to a vase.
This is actually a vase but predicted as a lamp. I think this is predicted because of the resemblance to the lamp.
This is actually a lamp but predicted as a vase. I think this is predicted because of the resemblance to the vase because of its width similar to the vase
The test accuracy is: 0.899
The top gif is ground truth and lower gif is prediction
The top gif is ground truth and lower gif is prediction
The top gif is ground truth and lower gif is prediction
The top gif is ground truth and lower gif is prediction. The prediction is wrong for the sitting area of the chair. The ground truth is marked as red and the prediction coes out to be blue.
The top gif is ground truth and lower gif is prediction. The prediction is wrong for the handle of the chair. The ground truth is marked as yellow and the prediction coes out to be red.
This part of assignment deals with affect of rotation on classification. In this, I will rotate the point cloud in x,y,z axis from [0,3.14]rad in an interval of 0.3 rad
The classification score is max at 0 deg and decreases and becomes minimum around 1.57 deg. Then the classification accuracy incrases till 3.14. This shows the non-generalizibilty of network oon x-axis rotation. The avg accuracy is: 0.572
The classification score is max at 0 deg and decreases till 3.14. There is slight increase around 1.57 rad. This shows the non-generalizibilty of network on y-axis rotation. The avg accuracy is: 0.59
The classification score is max at 0 deg and decreases till 3.14. There is slight increase around 1.57 rad. This shows the non-generalizibilty of network on z-axis rotation. The avg accuracy is: 0.502. The curve is very similar to y-axis rotation
One more interesting observation made:
For this lamp, the classification predicts as vase at this perspective(0 deg). But when the lamp is actually rotated and the 4 arms of the lamps are shown, the prediction is a lamp. This shows one of the cases where classification gets correct because more relevant features are visible.
Image1: Original non-rotated image, Pred: Vase
Image2: Rotated image, Pred: Lamp
This comparision is about num points vs classification accuracy. The classification accuracy is robust for most of the range of the point cloud, till num_points=10, where the point cloud loses most of the key features.
Similarly for segmentation task, the following accuracy drops as we rotate in x,y and z axis from [0,3] rad at an inteval of 0.5 rads
This comparision is about num points vs classification accuracy. The classification accuracy is robust for most of the range of the point cloud, till num_points=10, where the point cloud loses most of the key features.
I have implemented PointNet++ model for 10000 points pointcloud. The model looks as follows: SA(K,r,[l1, ..., ld]) is a set abstraction (SA) level with K local regions of ball radius r using PointNet of d fully connected layers with width li (i = 1, ..., d).
SA(1000, 0.2, [64, 64, 128]) → SA(100, 0.4, [128, 128, 256]) → SA([256, 512, 1024]) → F C(512, 0.5) → F C(256, 0.5) → F C(Class_num) where Class_num is 3.
Though the ball radius is set to 0.2, I have taken the best 50 points in the ball radius, due to GPU space limitation.
Counter intuitevly the accuracy is lower than PointNet. This is mostly because of two implementation details:
The trend is very similar for num_points vs accuracy as seen in Q1
For rotational ananlysis, I just tried the x rotation and the trend is similar as shown in Q3
[NbConvertApp] Converting notebook index.ipynb to html [NbConvertApp] Writing 3978359 bytes to index.html