Assignment 3

NAME: Hiresh Gupta

ANDREW ID: hireshg

Late days used: 0

1. Differentiable Volume Rendering

1.3. Ray sampling (10 points)

Usage:

python main.py --config-name=box

Outputs of grid/ray visualization:

Grid Rays

1.4. Point sampling (10 points)

I have implemented StratifiedSampler in sampler.py for Ray Sampling.

Usage:

python main.py --config-name=box

Point Samples output:

Sample points

1.5. Volume rendering (30 points)

Implement volume rendering for an SDF volume.

Usage:

python main.py --config-name=box

Color & Depth Visualizations:

Spiral Rendering of Part 1 Spiral Rendering of Part 1

2. Optimizing a basic implicit volume

2.1. Random ray sampling (5 points)

I have implemented get_random_pixels_from_image in ray_utils.py for this.

2.2. Loss and training (5 points)

Usage:

python main.py --config-name=train_box

The optimized center and side lengths of the box after training can be found below:

Box centers: (0.2507, 0.2495, 0.0053)
Box side lengths: (1.999, 1.4942, 1.5404)

2.3. Visualization

The rendered volume visualization can be found below:

Spiral Rendering of Part 2

3. Optimizing a Neural Radiance Field (NeRF) (30 points)

I have implemented the NeRF MLP (without view dependence). The model can be trained by the following command:

python main.py --config-name=nerf_lego

Model output Visualization:

Lego Rendering of Part 3

4. NeRF Extras (Choose at least one! More than one is extra credit)

4.1 View Dependence (10 pts)

I have added view dependence to my previous NeRF model. The model can be trained by the following command:

python main.py --config-name=nerf_lego_view_dependence

Model output Visualization:

Lego Rendering of Part 4.1

Discussion:

As seen above, we can see that the model outputs get slightly better on adding view dependence. Without handling view dependence, the previous NeRF model (in Q3) solely predicts the color & density value based on the 3D location. Hence we see some weird artifacts at the back of the lego in Q3 output.

On adding view dependence, we observe a much more realistic output that changes with the viewing direction. This aligns with the observation shared by the authors in their paper that view dependence helps in recreating the specular reflection on the bulldozer tread.

4.3 High Resolution Imagery (10 pts)

Based on the viewing direction results observed earlier, we had established that adding view dependence helps in generating realistic outputs. So I have used it in my other experiments on running NeRF model on higher resolution images. We can run the NeRF model on high resolution imagery by running the following command:

python main.py --config-name=nerf_lego_highres

Additionally, I also played around the NeRF MLP architecture. The command to train a lighter NeRF model can be found below:

python main.py --config-name=nerf_lego_highres_lite

Output Visualizations:

Model output of a NeRF (with view dependence) trained on 400x400 resolution:

Lego Rendering of Part 4.3

Model output of a lighter NeRF model trained on 400x400 resolution with fewer parameters (n_hidden_neurons_xyz=64 & n_hidden_neurons_dir=32):

Lego Rendering of Part 4.3

Discussion:

As seen in the GIFs above, on increasing the network resolution the model outputs much more fine-grained details which were absent in the previously trained version. In addition to this, I also experimented with the network capacity of the model to check the impact of the number of training parameters on the results.

We can easily compare the above two outputs and see that the NeRF model with the higher number of parameters synthesizes a much clearer & detailed output as compared to the one with fewer parameters.