The learning based approach involves using the DeepSDF architecture
with Lipschitz regularization. The key insight into this approach is
understanding how neural fields are encoded as signed distance
functions (SDFs). In the left image below, we can see that an input
latent vector (blue) concatenated with an input query point (gray)
is fed into the DeepSDF network. The output SDF (yellow) returns a
float value representing how close or far from the surface the query
point is
The right image above shows the DeepSDF architecture. In our
implementation, we used five fully connected layers with TanH
activations, LayerNorm, and Lipschitz regularization. While the
DeepSDF paper notes that using a random latent vector which is
optimized during training provides better results, we used the
one-hot encoding. That is, given let's say three input meshes (A, B,
C), the corresponding latent vectors would be [1, 0, 0], [0, 1, 0],
[0, 0, 1]
We used MSELoss between the ground truth SDF and our network's
predicted SDF. Additionally, we added in a Lipschitz regularization
term for interpolation stability.