## Description

2 Overview

In this assignment, you will implement a stereo reconstruction algorithm given two view

images.

(a) Left image (b) Right image

Figure 1: In this assignment, you will implement a stereo reconstruction algorithm

given two images.

You can download the skeletal code and data (left.bmp and right.bmp) from here:

https://www-users.cs.umn.edu/~hspark/csci5561/HW5.zip

You will fill main_stereo.m that takes input images and intrinsic parameters K, and

produces a stereo disparity map.

2

CSCI 5561: Assignment #5

Stereo Reconstruction

3 SIFT Feature Matching

(a) Matching from I1 to I2 (b) Matching from I2 to I1

(c) Matching from I1 to I2 after ratio test (d) Matching from I2 to I1 after ratio test

(e) Bidirectional matching between I1 and I2

Figure 2: You will match points between I1 and I2 using SIFT features.

You will use VLFeat SIFT to extract keypoints and match between two views using

k-nearest neighbor search. The matches will be filtered using the ratio test and bidirectional consistency check.

function [x1, x2] = FindMatch(I1, I2)

Input: two input gray-scale images with uint8 format.

Output: x1 and x2 are n × 2 matrices that specify the correspondence.

Description: Each row of x1 and x2 contains the (x, y) coordinate of the point correspondence in I1 ad I2, respectively, i.e., x1(i,:) ↔ x2(i,:). This matching function

is similar to HW#2 except that bidirectional consistency check is mandatory.

(Note) Except for SIFT extraction, you are not allowed to use VLFeat functions.

3

CSCI 5561: Assignment #5

Stereo Reconstruction

4 Fundamental Matrix Computation

Figure 3: Given matches, you will compute a fundamental matrix to draw epipolar

lines.

function [F] = ComputeF(x1, x2)

Input: x1 and x2 are n × 2 matrices that specify the correspondence.

Output: F ∈ R3×3

is the fundamental matrix.

Description: F is robustly computed by the 8-point algorithm within RANSAC. Note

that the rank of the fundamental matrix needs to be 2 (SVD clean-up should be applied.). You can verify the validity of fundamental matrix by visualizing epipolar line

as shown in Figure 3.

(Note) Given the fundamental matrix, you can run the provided function:

[R1 C1 R2 C2 R3 C3 R4 C4] = ComputeCameraPose(F, K)

This function computes the four sets of camera poses given the fundamental matrix

where R1 C1 · · · R4 C4 are rotation and camera center (represented in the world coordinate system) and K is the intrinsic parameter. These four configurations can be

visualized in 3D as shown in Figure 4.

-0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1 1.2

0

0.8

0.6

0.4

0.2

0

0.5

0.2

0

0

-0.2

-0.5

-0.4

-1

-0.6

-1.5 0.5

0

-0.5

0

0

0.2

-0.2

0.4

0 0.2 0.4 0.6 0.8 1 -1 -0.5 0 0.5

0.2

0

-0.2

-0.4

-1.5

-0.6

0

Figure 4: Four configurations of camera pose from a fundamental matrix.

4

CSCI 5561: Assignment #5

Stereo Reconstruction

5 Triangulation

Given camera pose and correspondences, you will triangulate to reconstruct 3D points.

function [X] = Triangulation(P1, P2, x1, x2)

Input: P1 and P2 are two camera projection matrices (R3×4

). x1 and x2 are n × 2

matrices that specify the correspondence.

Output: X is n × 3 where each row specifies the 3D reconstructed point.

Description: Use the triangulation method by linear solve, i.e.,

u

1

×

P1

v

1

×

P2

X

1

= 0

(Note) Use plot3 to visualize them as shown in Figure 5.

0 50 100

0

-20

-50

-40

-60

-80

0

0

-100

20

-50

40

60

0

80

50

40

20

40 60 80

0

-20 0 20

-20

-80 -60 -40 0

-40

-20

0

20

0-80 -60 -40 -20 0 20 40 60

(a) nValid = 10 (b) nValid = 488

(c) nValid = 0 (d) nValid = 0

Figure 5: You can visualize four camera pose configurations with point cloud.

5

CSCI 5561: Assignment #5

Stereo Reconstruction

6 Pose Disambiguation

Given four configurations of relative camera pose and reconstructed points, you will

find the best camera pose by verifying through 3D point triangulation.

function [R,C,X] = DisambiguatePose(R1,C1,X1,R2,C2,X2,R3,C3,X3,R4,C4,X4)

Input: R1, C1, X1 · · · R4, C4, X4 are four sets of camera rotation, center, and 3D reconstructed points.

Output: R, C, X are the best camera rotation, center, and 3D reconstructed points.

Description: The 3D point must lie in front of the both cameras, which can be tested

by:

r

T

3

(X − C) > 0 (1)

where r3 is the 3rd row of the rotation matrix. In Figure 5, nValid means the number

of points that are in front of both cameras. (b) configuration produces the maximum

number of valid points, and therefore the best configuration is (b).

6

CSCI 5561: Assignment #5

Stereo Reconstruction

7 Stereo

(a) Rectified image 1 (b) Rectified image 2

Figure 6: Stereo rectification.

Given the disambiguated camera pose, you will implement dense stereo matching between two views based on dense SIFT of VLFeat.

function [H1, H2] = ComputeRectification(K, R, C)

Input: The relative camera pose (R and C) and intrinsic parameter K.

Output: H1 and H2 are homographies that rectify the left and right images such that

the epipoles are at infinity.

Description: Given the disambiguated camera pose, you can find the rectification

rotation matrix, Rrect such that the x-axis of the images aligns with the baseline. Find

the rectification homography H = KRrectRTK−1 where R is the rotation matrix of

the camera. The rectified images are shown in Figure 6. This rectification sends the

epipoles to infinity where the epipolar line becomes horizontal.

(Note) You can use the provided image warping function im_w = WarpImage(im, H)

to check your result.

7

CSCI 5561: Assignment #5

Stereo Reconstruction

Figure 7: Visualization of stereo match.

function [disparity] = DenseMatch(im1, im2)

Input: two gray-scale rectified images with uint8 format.

Output: disparity map disparity ∈ RH×W where H and W are the image height and

width.

Description: Compute the dense matches across all pixels. Given a pixel, u in the

left image, sweep along its epipolar line, lu, and find the disparity, d, that produces the

best match, i.e.,

d = arg min

i

kd

1

u − d

2

u+(i,0)k

2 ∀i = 0, 1, · · · , N

where d

1

u

is the dense SIFT descriptor at u on the left image and d

2

u+(i,0) is the SIFT

descriptor at u+ (i, 0) (i pixel displaced along the x-axis) on the right image. Visualize

the disparity map as shown in Figure 7.

8