Assignment 4

Anirudh Chakravarthy (achakrav)

Q1

Command to run:

python -m a4.main --config-name=torus

Implementation Details

In order to perform sphere tracing, I ran the algorithm for max_iters iterations. At each iteration, I computed the signed distance function to the nearest surface. For those points whose distance to the surface was greater than a threshold eps, a step in the direction of the ray was taken with a magnitude equal to the SDF value evaluated at the point. If the distance of the current point to it's nearest surface is less than eps, it is considered to intersect that surface.

Hyper-parameters:

  1. Max iterations: 64
  2. Epsilon: 1e-4

Visualization

Torus

Q2

Command to run:

python -m a4.main --config-name=points

Implementation Details

I used an MLP encoder similar to NeRF with the exact same settings (depth, number of neurons. positional embedding, etc) as in the previous assignment. In order to predict distance to the nearest surface, I used one additional linear layer with no non-linearities.

To implement eikonal loss, for each point, I computed the L2-norm of it's gradient at that point. The absolute value of the difference norm - 1 at each point served as the loss function. Finally, I applied a mean reduction across all points.

Design choices:

  1. I used an L1 loss for per-point SDFs and downscaled this loss by cfg.training.point_weight.
  2. I used an L1-like loss term for the eikonal loss.

Visualizations

Bunny input Bunny geometry

Q3

Command to run:

python -m a4.main --config-name=volsdf

Implementation Details

For the colour MLP, I used the output of the encoder and added a colour head with cfg.n_layers_color additional decoding layers and a final linear layer (along with sigmoid activation) to predict colour.

Explanation

In my opinion, $\alpha$ serves as a scale factor to spread out the per-point densities. For example, if we had a very large $\alpha$, two points with very similar laplace CDF would have very different densities since they are scaled to far-away values.

The term $\beta$ has a clustering effect and controls how similar the densities for nearing points would be.

  1. How does high beta bias your learned SDF? What about low beta?
    • High $\beta$ would force all points to have the same density $\alpha$. A low $\beta$ value entails all points on the surface to have large density values, while points outside the surface have constant density $\alpha$. In some sense, $\beta$ biases the smoothness of the transition across the surface boundary.
  2. Would an SDF be easier to train with volume rendering and low beta or high beta? Why?
    • An SDF would probably be easier to train with low $\beta$ since it would be easy to find out when the ray hits the surface due to constant opacity in the medium and a high density value around the surface boundary.
  3. Would you be more likely to learn an accurate surface with high beta or low beta? Why?
    • A low $\beta$ would probably lead to accurate surface since the density changes almost like a step function at the surface. So it would be easier to understand what a surface really means (i.e. a sharp spike in the density).

Visualizations

Bulldozer geometry Bulldozer color

Q4

Q4.1

Command to run:

python -m a4.main --config-name=largeScene

I rendered a box of assorted donuts with 4 flavours: lemon, chocolate, red velvet, and orange. Unfortunately, I couldn't render my favourite flavour: rainbow sprinkles on a vanilla base! Red velvet is still really good :) but I don't really like chocolate to be honest (hope you don't deduct points for this :p).

Donuts

Q 4.2

Command to run:

python main.py --config-name=nerf_lego_fewview
python -m a4.main --config-name=volsdf_fewview

Visualizations

No. of Views VolSDF NeRF
5
10
20
100

It's clearly visible that in the case of sufficient views (20, 100), NeRF and VolSDF give comparable results. In fact, it seems like the details on the board of the bulldozer are better captured by NeRF. However, when we use very few views (5), NeRF is unable to learn any 3D structure associated with the bulldozer. VolSDF rendering is blur and lacks some details, but is still reasonable. On using 10 views, the NeRF result seems a bit blunt at the edges and also consists of a weird reflection artifact when the bulldozer is pointing forwards. On the other hand, VolSDF does better on both these fronts.

Q 4.3

Command to run:

python -m a4.main --config-name=volsdf_neus

I used the naive method in the NeUS paper to convert SDF to density. I noticed using s=100 gave better results. The paper mentioned that having a higher s is expected as training progresses since standard deviation is $1/s$, and visibly having this seemed to worked better for a constant s.

Visualizations

s Geometry Colour
10
100