Learning for 3D Vision: Assignement 3

Deepti Upmaka

1. Sphere Tracing (30pts)

The sphere tracing algorithm starts off given the origins of each of the rays and their directions. In ray marching, we take t steps towards the surface in the direction of the ray. With sphere tracing, the t or distance we step is determined by the radius of the sphere. This radius is the closest distance to the surface from the origin. This means we can pass our inital origins into the SDF function to get all the distances for the next step. This step is calculated by current_origin + distance_to_SDF * direction_of_ray. We are given a max iterations so we calculate the next step these many times. At the end we mask which rays actually intersect the surface where I defined an epsilon for how much distance I will tolerate deviating from the surface.

Run this part with this command: python -m a4.main --config-name=torus

2. Optimizing a Neural SDF (30pts)

In the below image we can observe the SDF reconstruction on point cloud data using the default hyperparameters which gave the best output.



The MLP structure I used for the distance is as below. There is an additional skip connection from the input to the 3rd Linear layer.
  1. Harmonic Embedding (input=3, output=4)
  2. Linear (input=4, output=128)
  3. ReLU
  4. Linear (input=128, output=128)
  5. ReLU
  6. Linear (input=128, output=128)
  7. ReLU
  8. Linear (input=128, output=128)
  9. ReLU
  10. Linear (input=128, output=128)
  11. ReLU
  12. Linear (input=128, output=128)
  13. ReLU
  14. Linear (input=128, output=1)


The Eikonal loss wants to force the norm of the gradients of the distances to be 1. This means that we want to get closer and closer to the actual surface. So we compute L2 loss across each 3D point, subtract it from 1 and take the average of this.

Run this part with this command: python -m a4.main --config-name=points

3. VolSDF (30 pts)

Alpha is the constant density that is defined by us in this case and beta determines the smoothing as you approch the surface. If the alpha is too high without an approporiate beta then it is not able to learn consistency across the surface resultng in uneven color across all the views and the geometry does not converge to the appropriate shape. In our case we are settng the alpha and beta terms but they are learnable parameters.

  1. How does high `beta` bias your learned SDF? What about low `beta`?

    We can see that beta is the variance since the Cumulative Distribution Function is a Laplacian wth 0 mean and beta variance. A high beta pushes better generalization since it increases variance and it allows to resolve the geometry for further distances from the surface. A lower beta will decrease variance allowing it to fit more specifically to the training set as we try to solve for points closer to the surface for a sharper image.

  2. Would an SDF be easier to train with volume rendering and low `beta` or high `beta`? Why?

    It would be easier to train volume rendering with a high beta because it doesn't overfit to the input samples. There is blurriness because it allows for points farther away from the surface. For a lower beta, the boundary is tighter so it would more strict which points are reconstructed.

  3. Would you be more likely to learn an accurate surface with high `beta` or low `beta`? Why?

    A lower beta will learn a more accurate surface as it decreases the variance and the smoothness. A high beta will have blurriness given the high variance.

I was able to acheieve these results with the default hyperparameters. I did try increasing the learning rate however, the geometry did not converge to an appropriate shape. It had extra artifacts around the surface plane. Below are a couple of examples with alpha constant and changing the beta values. The higher beta is smoother and the lower beta is slightly sharper.

alpha = 10.0, beta= 0.05



alpha = 10.0, beta= 0.03



alpha = 10.0, beta= 0.5



Run this part with this command python -m a4.main --config-name=volsdf

4. Neural Surface Extras

4.2 Fewer Training Views (10 pts)

I decreased the number of samples that the VolSDF uses to only 20 and 30 samples. Below you can observe the color and the geometry for these configurations. For the 20 samples, the VolSDF shows some holes in the geometry and the color is not as sharp. The NeRF is resolving to an accurate shape but the blinking might be because of the missing input viewpoints that is is not able to describe the color and shape of the occluded regions. However, when the NeRF is run with 30 samples, it shows that it can build a pretty sharp image, more so than the color output from VolSDF. There might be a threshold in between where the NeRF performance overtakes the VolSDF performance for the number of samples.

20 samples, VolSDF Color, VolSDF Geometry, NeRF



30 samples, VolSDF Color, VolSDF Geometry, NeRF



4.3 Alternate SDF to Density Conversions (10 pts)

I implemented the naive solutioin from the NeuS paper which is as follows (s*torch.exp(-1*s*signed_distance))/((1+ torch.exp(-1*s*signed_distance))**2) where s is a tunable parameter. Below are a couple of examples from various s values. As you can see there is a trade off between a low s value having better geometry and poor color and high s value that has better color and poor geometry.

s = 15



s = 40