First, the points at the nearest distance (self.near) are sampled. Second, if the distances are smaller than threshold, the corresponding points are considered to intersect with the object surface and stored in the points variable. Third, the next points on the rays are sampled based on the distances (e.g., rays = rays + distances * directions). Fourth, iterate the second and third process up to N (self.max_iters) times.
Alpha represents a uniform density on the object surface. Beta is a parameter to specify how blurry (smooth) the dencity nearby the object surface is.
The high beta will make the learned SDF much more blurry. However, a more smooth gradient is given, and training would be more stable. On the contrary, we expect a highly accurate occupancy map with low beta. However, the training can be unstable due to the more sensitive gradient.
High beta would make training easier since it gives a smoother gradient.
Low beta (low variance) would give a step function-like CDF that can accurately reconstruct the original surface.
This is the best results where I set the beta to 0.04. The training is really sensitive to beta. For example, when I tried to decrease beta to 0.03, the training did not converge.
Training with 100 views (Left: VolSDF, Right: NeRF)
Training with 20 views (Left: VolSDF, Right: NeRF)