16-889 Assignment 3: Volume Rendering and Neural Radiance Fields

Author: Zhe Huang (zhehuang@andrew.cmu.edu)

1.3. Ray sampling

1.4. Point sampling

1.5. Volume rendering

2.2. Loss and training

After training and rounding to the nearest hundredth:

Box center: (0.25, 0.25, 0.00)
Box side lengths: (2.01, 1.50, 1.50)

2.3. Visualization

3. Optimizing a Neural Radiance Field (NeRF)

The network structure (e.g. layers, types) of my NeRF implementation follows the original NeRF paper, including the positional embeddings. The model is not view dependent so the output layer is just a nn.Linear layer with out put size $(N_{points}, 4)$ where each point has 3 feature values (with sigmoid activation) and 1 density value (with ReLU activation to avoid negative values). All hyperparameters are default values from nerf_lego.yaml.

4.1 View Dependence

Here I tweaked the network structure from part 3 a little bit by removing the simple output layer and adding the directional embedding layers. I adapt a new set of hyperparameters, as is listed below.

n_harmonic_functions_xyz: 9
n_harmonic_functions_dir: 9
n_hidden_neurons_xyz: 256
n_hidden_neurons_dir: 128
density_noise_std: 0.01
n_layers_xyz: 9
append_xyz: [4]

Here is the result after training 100 episodes. It is little blurry due to the lack of training but we can still clearly see that the luminance of the ground change with as viewpoint changes due to the view dependence.

4.3 High Resolution Imagery

Here I still use view independent NeRF as it is fast to train. For full model, I use the same hyperparameters as in part 4.1. For lite model, I use the default setting from nerf_lego_highres.yaml. This is the result after taining 100 epochs.