Presented by Dijing Zhang
Create a 360-degree gif video that shows many continuous views of the provided cow mesh
The Dolly Zoom is a famous camera effect, first used in the Alfred Hitchcock film Vertigo. The core idea is to change the focal length of the camera while moving the camera in a way such that the subject is the same size in the frame, producing a rather unsettling effect.
vertices = torch.tensor([[1.0, 0.0, 0.0], [0.0, 1.0, 0.0],
[0.0, 0.0, 0.0], [0.0, 0.0, 1.0]])
faces = torch.tensor([[0, 1, 2], [0, 1, 3], [0, 2, 3], [1, 2, 3]])
Vertices: 4 Faces: 4
vertices = 0.5 * torch.tensor([[1.0, 1.0, -1.0], [1.0, -1.0, -1.0], [-1.0, -1.0, -1.0], [-1.0, 1.0, -1.0],
[1.0, 1.0, 1.0], [1.0, -1.0, 1.0], [-1.0, -1.0, 1.0], [-1.0, 1.0, 1.0]])
faces = torch.tensor([[0, 1, 2], [0, 2, 3], [3, 6, 7], [2, 3, 6],
[1, 2, 6], [1, 5, 6], [4, 5, 7], [4, 6, 7],
[0, 4, 7], [0, 3, 7], [1, 4, 0], [1, 5, 4],])
Vertices: 8 Faces: 12
Retexturing the cow mesh such that the color smoothly changes from the front of the cow to the back of the cow.
color1: [0, 0, 1]
color2: [1, 0, 0]
When working with 3D, finding a reasonable camera pose is often the first step to producing a useful visualization, and an important first step toward debugging.
Find a set (R_relative, T_relative) such that the new camera extrinsics with R = R_relative @ R_0 and T = R_relative @ T_0 + T_relative produces each of the following images:
For case1, which rotate the cow in xy plane 90 degrees closewise direction, we set R_relative = [[0, 1, 0], [-1, 0, 0], [0, 0, 1]] to rotate the x-axis and y-axis.
For case2, which move camera more far from object, just simply increase the translation by setting T_relative=[0, 0, 3]
For case3, which move the camera up-right, just simply set T_relative=[0.5, -0.5, 0].
For case4, which rotates the cow in x-z plane (along y-axis) 90 degrees closewisely, we can reset x-aixs to [0, 0, 1] and z-aixs to [-1, 0, 0] and remember to remove camera by setting T_relative = [-3, 0, 3]
Practice rendering point clouds constructed from 2 RGB-D images from the Common Objects in 3D Dataset.
Reconstruct pointcloud from two different images (including image, depth, mask) and merged results.