This assignment was submitted 2 days late.
xy_grid | rays |
![]() |
![]() |
Here are the points as rendered.
I implemented it. It's pretty fast on an RTX 3090.
Depth render | Spinning cube |
![]() |
![]() |
Implemented. Used the nice permutation trick.
Here are the rounded estimates:
Box center: (0.25, 0.25, 0.00)
Box side lengths: (2.00, 1.50, 1.50)
Here's the rendered wonky box.
Initially, I implemented things without view dependence. This resulted in an image that was a tiny bit noisy and a tiny bit imprecise. I ended up using the nice skip architecture provided by the friendly TAs.
3090 go BRRRRR training only took 5min. Who needs instantngp anyway... Here's my shovel:
Notice that there's a bit of noise underneath the shovel. This is because the direction wasn't considered, making things harder to optimize in a global sense.
This was pretty easy, I just followed the architecture from the original NeRF paper, but modified the number of layers from the original one to the one suggested by the TA's in their config file.
Here's my much cleaner shovel. Also it took fewer steps to converge.
Thankfully, my network did not overfit to the training images. In general though, if you also provide the directions, and you have a fixed (and small) number of perspectives, the network can just memorize the viewpoint-dependent color. There's no telling what the network would predict when interpolating between the viewpoints in this case. However, since the viewpoint is not used in connection with the density (how the original NeRF paper does it), then there's no danger of overfitting the density prediction based on viewpoint.
No time.
No time.