16-889 Assignment 1: Rendering Basics with PyTorch3D

Name: Rohan Agarwal

Andrew ID: rohanaga

You can run all files using: python main.py

0.1 Rendering your first mesh

If you run python -m q0 --image_size 256 --output_path images/q0/cow_render.jpg, you should see the following output:

Cow render

1. Practicing with Cameras

1.1. 360-degree Renders (5 points)

If you run python -m src.q1_1 --image_size 256 --output_path images/q1/cow_render.gif, you should see the following output:

360-degree render

1.2 Re-creating the Dolly Zoom (10 points)

If you run python -m src.q1_2 --num_frames 10 --image_size 256 --output_file images/q1/dolly_gif.gif, you should see the following output:

Dolly Zoom

2. Practicing with Meshes

2.1 Constructing a Tetrahedron (5 points)

Number of vertices = 4 Number of faces = 4

If you run python -m src.q2_1 --image_size 256 --output_path images/q2/tetrahedron_render.gif, you should see the following output:

Tetrahedron Render

2.2 Constructing a Cube (5 points)

Number of vertices = 8 Number of faces = 12 (2 triangular meshes for each side)

If you run python -m src.q2_2 --image_size 256 --output_path images/q2/cube_render.gif, you should see the following output:

Cube Render

3. Re-texturing a mesh (10 points)

color1 = [0,0,1] (blue) and color2=[1,0,0] (red)

If you run python -m src.q3 --image_size 256 --output_path images/q3/cow_texture.gif, you should see the following output:

Cow Render

4. Camera Transformations (20 points)

Relative Rotation and Relative Transalation:

Say, a camera is moving in a fixed environment. The environment is represented by the world coordinate system. The camera has its own coordinate system at each of its position. This system has its origin at the camera centre i.e. the origin keeps moving. Now, we are interested in defining the relationship between the camera coordinate systems observed in position 1 and position 2. Relative rotation and relative translation define this relationship.

Relative rotation describes the amount of rotation the camera must undergo(from position 1) such that the axes of the camera in position 1 aligns with the axes in position 2. Mathematically, R2 = R_relative @ R1

Relative translation is similary described as the amount of translation the camera must undergo(from position 1) for its centre in position 1 to align with the centre in position 2. Mathematically, T2 = T_relative @ T1

Please note: Pytorch3D accepts the transpose of what has been described as the relative rotation and translation above.

If you run python -m src.q4 --image_size 512 --output_path images/q4/camera_1.jpg images/q4/camera_2.jpg images/q4/camera_3.jpg images/q4/camera_4. jpg, you should see the following output:

Cow render

R_rel = [[0,1,0],[-1,0,0],[0,0,1]] T_rel = [0,0,0]

The relative rotation sets the new X axis --> old Y axis and new Y axis --> negative of old X axis. The relative translation is zero as camera merely rotates in its position

Cow render

R_rel = [[0,0,1],[0,1,0],[0,0,1]] T_rel = [0,0,3]

The relative rotation is identity (actually, the transpose of it) i.e. has no effect on the original rotation of the camera. The relative translation moves the object away by 3 units along the positve z direction.

Cow render

R_rel = [[0,0,1],[0,1,0],[0,0,1]] T_rel = [0.5,-0.5,0]

The relative translation moves the new camera coordinates to (-0.5, 0.5) which is equivalent to the object moving diagonally towards bottom left corner. The rotation matrix remains unchanged

Cow render

R_rel = [[0.,0.,1],[0,1,0],[-1.,0,0]] T_rel = [-3,0,3]

The relative rotation sets the new X-axis --> old Z-axis and the new Z-axis --> old negative X-axis. The relative translation moves the the camera origin diagonally in the X-Z plane.

5. Rendering Generic 3D Representations

5.1 Rendering Point Clouds from RGB-D Images (10 points)

If you run python -m src.q5_1 --render point_cloud --image_size 256 --output_path images/q5/plants1.gif images/q5/plants2.gif images/q5/plants3. gif, you should see the following output:

Plant render Plant render Plant render

5.2 Parametric Functions (10 points)

If you run python -m src.q5_2 --render parametric --num_samples 100 1000 --image_size 256 --output_path images/q5/, you should see the following output:

Torus render

100 samples

Torus render

1000 samples

5.3 Implicit Surfaces (15 points)

Memory Efficiency: Meshes are less memory efficient as they require storing face representations in addition to vertices. Rendering Speed: Points clouds are easier/faster to render as they don't require rendering the surface between the points. Rendering Quality: Given we use sufficient number of faces, mesh will provide a better visually quality as it allows for observing surfaces. It is easier to reason about the shape because of the connectivity. Points clouds may not be able to accurately represent fine details of an object (e.g. horn of the cow )

If you run python -m src.q5_3 --render implicit --num_samples 100 --image_size 256 --output_path images/q5/torus_implicit.gif, you should see the following output:

Torus render

6. Do Something Fun (10 points)

Nothing fancy here, just a parametric rendering of a spiral spiralling.

If you run python -m src.q6 --render parametric --num_samples 100 --image_size 256 --output_path images/q6/spiral_parametric.gif, you should see the following output:

Cow render

(Extra Credit) 7. Sampling Points on Meshes (10 points)

If you run python -m src.q7 --image_size 256 --output_path images/q7/, you should see the following output:

Cow renderCow render

10 samples

Cow renderCow render

100 samples

Cow renderCow render

1000 samples

Cow renderCow render

10000 samples