zero.png

cow_render.jpg

1. Practicing with Cameras

360 deg pan

360_cow.gif

Dolly Zoom

Just visualising how the camera moves for my own benefit

dolly.gif

2. Practicing with Meshes

2.1 Tetra hedron

tetrahedron.gif

Took 4 vertices and 4 faces to make the tetrahedron

2.2 Cube

cube.gif

8 vertices and 12 triangle faces

3. Re-texturing a mesh

360_cow_gradient.gif

color1 = [0, 1, 0] (green) and color2 = [1, 0, 1] (purple)

4 Camera Transformations

It all makes sense if we think of R_relative applied to the camera and T_relative applied to the world_origin

In the following image R_relative is doing a left hand rotation about the z-axis, meaning looking at the cow from in front of it with the camera in our hand we are turning the camera counter clockwise. If we then take a picture with the camera the cow will have appeared to have rotated clockwise! Which is what we see in the image.

Another way to think about it is: Because of the Pytorch3d convention, what we are defining here is actually $R^T$. Had a discussion with prof Tulsiani about it. $R^T$ would be a clockwise rotation about z and if we apply that to the world then we get the cow rotated clockwise as follows.

T_relative is the change in the world origin with respect to the camera frame and hence the cow is 2 metres further away as expressed in the camera frame.

T_relative is moving the world origin aka the cow 0.5m in POS x and 0.5m in NEG y direction (similar to the above situation).

This was most tricky to reason. R_relative is turning the camera in a positive righthand fashion in the y-direction by 90 degrees. Since the camera was 3m in front of the cow in world coordinates, it would now be completely turned away from the cow with nothing in frame. With respect to the camera, the world origin is now located at [-3,0,0], hence in order to get the view as desired, we must bring the cow back to [0,0,3] in the NEW camera frame, hence T_relative is [-3,0,3].

Again: Another way to think about this is to think about what $R_{rel}^T$ is defining, which is a lefthand rotation about y if we apply that then the cow (which is at [0,0,3]) would move to [-3,0,0] (Same as above). T_rel reasoning is same as above.

5. Rendering generic 3D representations

5.1 Rendering Point Clouds from RGB-D Images

360_plant_side_by_side.gif

5.2 Parametric

360_taurus1False.gif

360_taurus_diff1False.gif

5.3 Implicit

360_taurus_implicit-2.gif

  1. The parametric approach to render the torus was much faster, as can be seen from the timing cells above the parametric form was more than twelve times as fast and in both cases the number of points sampled were 1,000,000.
  2. One of the reasons for this issue is the style of sampling, if we assume that our object lies entirely within say a unit cube then in the parametric case we can uniformly sample along the three axes and each point we would sample would result in a point on the surface, whereas in the implicit version we need to uniformly sample each point in 3d space for the unit cube upto some quantisation and hence most of the points we sample would be empty. This means that for the same number of samples, the quality of the surface we can render is better with the parametric distribution.
  3. Since we have this drawback, we have to apply marching cubes in the implicit case in order to generate the mesh, which is >90% of the time burden in the implicit rendering case!! (I did some timing experiements which are commented out in my function render_taurus_mesh)
  4. It is clear from the renders above that the parametric form sampled with the same number of points, despite being a much bigger taurus (R=5m, r=1m) has better render quality than the implicit one (R=1m, r=0.3m). In the latter you can clearly see the mesh faces (giving it a non-smooth texture quality).
  5. If we were to spend resources on doing marching cubes on the parametric version of the taurus we can further improve the render quality (albeit at a higher time cost) depending on our requirements.
  6. Some benefits like free form learning of a surface with a blackbox function like neural nets is only possible with the implicit form as the parametric form requires us to know the nature of the manifold in order to learn the spcific parameters that form the shape.
  7. If we we fine with a pixelated voxel grid (as opposed to the smoother rendering we achieved with marching cubes) then the implicit representation could be very fast while providing the above benefit.

6. Do something fun

Rendering a human hand with texture

360_fuuunhand.gif

Flying through the rusted city

360_fuuun1.gif

7 Sampling Points on Meshes

360_subsampled_cow_side_by_side-2.gif