zero.png

1. Sphere Tracing

part_1.gif

def sphere_tracing(
    self,
    implicit_fn,
    origins, # Nx3
    directions, # Nx3
):
    '''
    Input:
        implicit_fn: a module that computes a SDF at a query point
        origins: N_rays X 3
        directions: N_rays X 3
    Output:
        points: N_rays X 3 points indicating ray-surface intersections. For rays that do not intersect the surface,
                the point can be arbitrary.
        mask: N_rays X 1 (boolean tensor) denoting which of the input rays intersect the surface.
    '''

    points = origins + self.near*directions
    epsilon = 1e-5
    for i in range(self.max_iters):
        residuals = implicit_fn(points)
        mask = implicit_fn(points) < epsilon
        potential_points = torch.norm((points + ~mask * residuals * directions), dim=1).reshape(-1,1)
        beyond_far = potential_points > self.far
        stop = torch.logical_or(beyond_far, mask)

        if torch.sum(stop) == directions.shape[0]:
            break

        points = stop * points + ~stop * (points + residuals * directions)

    return points, mask.reshape(-1,1)

My implementation follows the sphere tracing algorithm presented in class slides. I didn't want to use nested loops and hence I vetorized my code. My tolerance for having reached the surface is epsilon = 1e-5.

As mentioned I have only one loop uptil max iters, in this loop I take all points from self.near to self.far starting from origins along directions. In each loop I compute the distance of each point from the surface, I stop updating the point locations if one of two things has happened:

I have a break condition before max_iters if all points have attained one of the above conditions.

2. Optimizing NueralSDF

Point Cloud Input Generated Render after training NeuralSDF
part_2_input.gif part_2.gif

MLP: HarmonicEmbeddingLayer(3, 4) -> Linear_with_ReLU(-1, 256) x (5) -> Linear(256,1)

I basically used the NerF architecture I had from the previous assignment, only changing things as necessary like not ReLU after the last layer, changing the output dimension to 1. The output generated was better than that provided in the handout so I was quite pleased with it. The -1 in the Linear_with_ReLU layer just indicates that the input dimension is to be filled in as appropriate.

Eikonal Loss: My eikonal loss couldn't be simpler: torch.mean(torch.abs(torch.norm(gradients, dim=1) - 1)). It is simply the norm of the gradient along the first dimension, a.k.a going from Nx3 to Nx1 as we want the norm to be 1 per training example. Then I take the element-wise difference with 1, absolute value and then mean it across the current batch of size N.

Hyperparameters: Other than the architecture I used the same hyperparams provided in the config.

3. VolSDF

beta controls the transclucence of the surface, the smaller the beta the most sudden the jump in the density indicating the surface is starting more abruptly. If the beta is higher then the slope is more gentle and hence the contribution of the visible colour at that point can be contributed to by more points along the ray and hence it is good for modelling objects like marble which allow light to enter. It's important to note that the SDF has a negative sign before it when being passed to the Laplacian CDF.

Laplace_cdf_mod.svg.png

alpha is the scaling parameter, it controls the peak density value inside the object. Per the formula, $1 - e^{-\sigma \Delta t}$ is the fraction of light that passes through a segment of the material of thickness $\Delta t$ and hence the higher the alpha, the denser to light the object is.

  1. How does high beta bias your learned SDF? What about low beta? At the time of rendering a high beta will bias the render to have a more hazy surface as the surface boundary will not be as well defined, conversely a low beta will encourage the render to show strong surfaces. Hence if we want to model somthing with a solid surface we would prefer lower values of beta however for surfaces with are more transculent like clothes or marble we might consider higher values of beta.

  2. Would an SDF be easier to train with volume rendering and low beta or high beta? Why?

Beta Value Geometry Render (epoch 50) Coloured Render (epoch 50) Geometry Render (epoch 100) Coloured Render (epoch 100)
- - - - -
beta = 0.0005 part_3_geometry.gif part_3_geometry.gif part_3.gif part_3.gif
beta = 5.0 partwe_3.gif alskdwenasl part_w3.gif alskdwkdnasl

Empirically found that lower beta values train faster, probably because the model has to learn a solid surface and hence having a string boundary similar to an indicator function is the most succint way to model it, with higher values of beta the model would be very confused and would have to take several iterations to reason about the actual location of the surface along the gentle slope.

  1. Would you be more likely to learn an accurate surface with high beta or low beta? Why? I would reason it depends on the type of surface one is trying to learn, if we are trying to learn a surface which is permeable to some amount of light like marble or clothes then having a higher beta would model it better however with most real world objects the surfaces is pretty solid (like the lego truck) like a step function and hence having a lower value of beta would be a more accurate representation.
Geometry Render Coloured Render
part_3_geometry.gif part_3_first.gif

4.1 Render a Large Scene with Sphere Tracing

I attempted to render a deathstar shooting at an imperial Cruiser

I wasn't able to get the imperial cruiser's shape right but I have rendered it as if it was 3D ptinted using 100s of box primitives.

The death star missed in it first attempt!

part_4.1.gif

Succeeded in the second attempt!

part_4.1.gif

4.2 Fewer Training Views

20 Training examples

Only 20 training examples: following indexes [35, 93, 46, 26, 29, 18, 57, 23, 10, 81, 54, 76, 13, 6, 12, 94, 88, 67, 82, 5]

Geometry Render (Neural SDF) Coloured Render (Neural SDF) Coloured Render (NeRF with View Dependence) Coloured Render (NeRF without View Dependence)
part_3_geometry.gif part_3.gif part_3.gif alskdnasl

Can't really detect the drop in performance. So if I visualise all of them during training:

Geometry Render (Neural SDF) Coloured Render (Neural SDF) Coloured Render (NeRF with View Dependence) Coloured Render (NeRF without View Dependence)
part_3_geometry-.gif part_3sd.gif part_3ds.gif alskdnaslds

At a higher resolution for 20 images I was able to train a Neural SDF, although only for 200 epochs but I wasn't able to train a NeRF model:

Geometry Render (Neural SDF) Coloured Render (Neural SDF)
part_3_geometry-.gif part_3sd.gif

4.3 Alternate SDF to Density Conversions

Geometry Render (s=40) Coloured Render (s=40) Geometry Render (s=10) Coloured Render (s=10)
part_3_geometryss.gif part_3.gif part_3.gif alskdnasl

s controls the peak density value at the surface, in the formula used in the NeUS paper which has a surface model as follows. As we can see having a lower value for s leads to a hazier/more transclucent rendering as all of the incoming light is not characterized by the few (possibly only one) sample precisely at or near the surface. Hence higher values of s better model a solid surface. But we also see that the geometry render is more unforgiving i.e has a lot of white spaces when using a high value of s.

The-effect-of-weight-ie-a-on-sigmoid-function-and-derivative-of-sigmoid-function.png

I also renderred in high resolution to show the transcluscent nature of the render with lower s

part_3_275_train_images.gif