def sphere_tracing(
self,
implicit_fn,
origins, # Nx3
directions, # Nx3
):
'''
Input:
implicit_fn: a module that computes a SDF at a query point
origins: N_rays X 3
directions: N_rays X 3
Output:
points: N_rays X 3 points indicating ray-surface intersections. For rays that do not intersect the surface,
the point can be arbitrary.
mask: N_rays X 1 (boolean tensor) denoting which of the input rays intersect the surface.
'''
points = origins + self.near*directions
epsilon = 1e-5
for i in range(self.max_iters):
residuals = implicit_fn(points)
mask = implicit_fn(points) < epsilon
potential_points = torch.norm((points + ~mask * residuals * directions), dim=1).reshape(-1,1)
beyond_far = potential_points > self.far
stop = torch.logical_or(beyond_far, mask)
if torch.sum(stop) == directions.shape[0]:
break
points = stop * points + ~stop * (points + residuals * directions)
return points, mask.reshape(-1,1)
My implementation follows the sphere tracing algorithm presented in class slides. I didn't want to use nested loops and hence I vetorized my code. My tolerance for having reached the surface is epsilon = 1e-5
.
As mentioned I have only one loop uptil max iters, in this loop I take all points from self.near
to self.far
starting from origins
along directions
. In each loop I compute the distance of each point from the surface, I stop updating the point locations if one of two things has happened:
residual < epsilon
self.far
from origins
I have a break condition before max_iters
if all points have attained one of the above conditions.
Point Cloud Input | Generated Render after training NeuralSDF |
---|---|
![]() |
![]() |
MLP: HarmonicEmbeddingLayer(3, 4) -> Linear_with_ReLU(-1, 256) x (5) -> Linear(256,1)
I basically used the NerF architecture I had from the previous assignment, only changing things as necessary like not ReLU after the last layer, changing the output dimension to 1. The output generated was better than that provided in the handout so I was quite pleased with it. The -1
in the Linear_with_ReLU
layer just indicates that the input dimension is to be filled in as appropriate.
Eikonal Loss: My eikonal loss couldn't be simpler: torch.mean(torch.abs(torch.norm(gradients, dim=1) - 1))
. It is simply the norm of the gradient along the first dimension, a.k.a going from Nx3
to Nx1
as we want the norm to be 1 per training example. Then I take the element-wise difference with 1, absolute value and then mean it across the current batch of size N
.
Hyperparameters: Other than the architecture I used the same hyperparams provided in the config.
beta controls the transclucence of the surface, the smaller the beta the most sudden the jump in the density indicating the surface is starting more abruptly. If the beta is higher then the slope is more gentle and hence the contribution of the visible colour at that point can be contributed to by more points along the ray and hence it is good for modelling objects like marble which allow light to enter. It's important to note that the SDF has a negative sign before it when being passed to the Laplacian CDF.
alpha is the scaling parameter, it controls the peak density value inside the object. Per the formula, $1 - e^{-\sigma \Delta t}$ is the fraction of light that passes through a segment of the material of thickness $\Delta t$ and hence the higher the alpha, the denser to light the object is.
How does high beta
bias your learned SDF? What about low beta
? At the time of rendering a high beta will bias the render to have a more hazy surface as the surface boundary will not be as well defined, conversely a low beta will encourage the render to show strong surfaces. Hence if we want to model somthing with a solid surface we would prefer lower values of beta however for surfaces with are more transculent like clothes or marble we might consider higher values of beta
.
Would an SDF be easier to train with volume rendering and low beta
or high beta
? Why?
Beta Value | Geometry Render (epoch 50) | Coloured Render (epoch 50) | Geometry Render (epoch 100) | Coloured Render (epoch 100) |
---|---|---|---|---|
- | - | - | - | - |
beta = 0.0005 | ![]() |
![]() |
![]() |
![]() |
beta = 5.0 | ![]() |
![]() |
![]() |
![]() |
Empirically found that lower beta values train faster, probably because the model has to learn a solid surface and hence having a string boundary similar to an indicator function is the most succint way to model it, with higher values of beta the model would be very confused and would have to take several iterations to reason about the actual location of the surface along the gentle slope.
Geometry Render | Coloured Render |
---|---|
![]() |
![]() |
I wasn't able to get the imperial cruiser's shape right but I have rendered it as if it was 3D ptinted using 100s of box primitives.
The death star missed in it first attempt!
Succeeded in the second attempt!
Only 20 training examples: following indexes [35, 93, 46, 26, 29, 18, 57, 23, 10, 81, 54, 76, 13, 6, 12, 94, 88, 67, 82, 5]
Geometry Render (Neural SDF) | Coloured Render (Neural SDF) | Coloured Render (NeRF with View Dependence) | Coloured Render (NeRF without View Dependence) |
---|---|---|---|
![]() |
![]() |
![]() |
![]() |
Can't really detect the drop in performance. So if I visualise all of them during training:
Geometry Render (Neural SDF) | Coloured Render (Neural SDF) | Coloured Render (NeRF with View Dependence) | Coloured Render (NeRF without View Dependence) |
---|---|---|---|
![]() |
![]() |
![]() |
![]() |
At a higher resolution for 20 images I was able to train a Neural SDF, although only for 200 epochs but I wasn't able to train a NeRF model:
Geometry Render (Neural SDF) | Coloured Render (Neural SDF) | |
---|---|---|
![]() |
![]() |
Geometry Render (s=40) | Coloured Render (s=40) | Geometry Render (s=10) | Coloured Render (s=10) |
---|---|---|---|
![]() |
![]() |
![]() |
![]() |
s
controls the peak density value at the surface, in the formula used in the NeUS paper which has a surface model as follows. As we can see having a lower value for s
leads to a hazier/more transclucent rendering as all of the incoming light is not characterized by the few (possibly only one) sample precisely at or near the surface. Hence higher values of s
better model a solid surface. But we also see that the geometry render is more unforgiving i.e has a lot of white spaces when using a high value of s
.
I also renderred in high resolution to show the transcluscent nature of the render with lower s
['part_3_200_train_images.gif', 'part_3_geometry_150_train_images.gif', 'part_3_geometry_200_train_images.gif', 'part_3_geometry_175_train_images.gif', 'part_3_geometry_125_train_images.gif', 'part_3_geometry_100_train_images.gif', 'part_3_125_train_images.gif', 'part_3_100_train_images.gif', 'part_3_150_train_images.gif', 'part_3_175_train_images.gif'] neus_10 ['part_3_geometry_50_train_images.gif', 'part_3_25_train_images.gif', 'part_3_75_train_images.gif', 'part_3_geometry_25_train_images.gif', 'part_3_50_train_images.gif', 'part_3_geometry_75_train_images.gif'] neus_40 ['part_3_geometry_150_train_images.gif', 'part_3_geometry_50_train_images.gif', 'part_3_geometry_175_train_images.gif', 'part_3_25_train_images.gif', 'part_3_geometry_125_train_images.gif', 'part_3_75_train_images.gif', 'part_3_geometry_100_train_images.gif', 'part_3_geometry_25_train_images.gif', 'part_3_125_train_images.gif', 'part_3_50_train_images.gif', 'part_3_100_train_images.gif', 'part_3_150_train_images.gif', 'part_3_175_train_images.gif', 'part_3_geometry_75_train_images.gif'] neus_unknown