Vision 3D artificielle - Final exam (duration: 2h30)
P. Monasse and R. Marlet November 18th, 2014
You can choose to answer in French or English, at your convenience.
1 Relative Pose from Line Correspondences
This section proposes to explore a method estimating rotation and translation between two view positions from observed lines in orthogonal directions. The matrix of internal parametersK is supposed known.
You may use the well-known formula relating to the cross product:
(a×b)×c= (aTc)b−(aTb)c. (1) 1. Show that lines in 3D are viewed as lines in 2D by a pinhole camera.
2. Show that parallel lines in 3D along direction vector d, expressed in the coordinate system linked to the camera, are projected as concurrent lines at pointv=Kd, called vanishing point (“point de fuite” in French).
See the figure for examples of vanishing points.
3. Under what geometric conditions is the vanishing point “at infinity”?
4. Show that if we change the world coordinate system to an arbitrary one (still orthonormal), we can write v = KRd; explain the origin of the rotation matrixR.
Figure 1: Original image and three sets of parallel 3D lines in orthogonal direc- tions.
1
5. Show that vanishing points corresponding to orthogonal directions satisfy v1T(K−TK−1)v2= 0. (2) 6. Suppose we take the world coordinate system so thatd1= 1 0 0T
and d2= 0 1 0T
, directions corresponding to vanishing pointsv1 andv2. Show that the rotation matrixR can be written
R=
K−1v1
kK−1v1k
K−1v2
kK−1v2k
K−1v1
kK−1v1k×kKK−1−1vv22k.
(3) 7. Letl1 a projected line in directiond1. Show that
v1=l1×(K−TK−1v2). (4) 8. Supposing we have 3 lines, l1 in direction d1 and l2, l3 in orthogonal
directiond2. Sum up an algorithm to computeR.
9. Suppose directions di, i= 1,2, map to vanishing pointsvi andv0i in two views, linked by relative motion (Rrel, Trel). Show that we have
K−1vi=RTrelK−1v0i. (5) 10. From an estimation ofRrel, we want to evaluate how well a triplet of lines
(l1, l2, l3) and (l01, l02, l30) fits this rotation with the criterion:
2
X
j=1
cos−1 vjTK−TRTrelK−1vj0 kK−1vjkkK−1v0jk
!
. (6)
Justify this criterion.
11. Propose an algorithm of type RANSAC estimating Rrel from detected lines in both images.
12. We writevij andvij0 vanishing points estimated by configurationiof three lines and directionj. We estimate the least square solution:
Rrel= arg min
R N
X
i=1 2
X
j=1
K−1vij
kK−1vijk−RT K−1v0ij kK−1v0ijk
2
2
. (7)
Write this with 3×2N matricesD andD0 and Frobenius norm:
Rrel = arg min
R
D−RTD0
2
F. (8)
(Reminder: kAk2F = Tr(ATA)) 13. Show that
Rrel = arg max
R Tr(RDD0T). (9)
2
14. Writing the SVD decomposition ofD0DT asU SVT, show that
Rrel=U VT. (10)
15. Suppose that two pairs of lines (assumed to be projections of coplanar 3D lines) intersect atp1,p2in left image andp01,p02in right image. Using the epipolar constraint, show thatTrel can be estimated up to scale onceRrel is known.
2 Feature detection and description
The goal of the following questions is to check your understanding of the course.
You can answer them with just one or a few sentences.
2.1 Repeated elements
Consider the case where you have several images of a scene containing repeated elements, e.g., several similar windows on a facade.
1. What is the impact on feature detection?
2. What is the impact on feature description?
3. What is the impact on feature matching in general?
4. What is the impact on feature matching with the SIFT matching strategy?
5. What is the impact on camera calibration?
6. What is the impact on 3D reconstruction?
2.2 Similarity measures
Consider the way Harris features are detected, usingEAC SSD as in the orginal formulation of Harris and Stephens (1988).
1. What is the impact if we now replaceEAC SSD withEAC ZSSD? 2. What is the impact if we now replaceEAC SSD withEAC ZNSSD? 3. What is the impact if we now replaceEAC SSD withEAC ZNCC?
3 Graph cut for disparity map estimation
Consider the setting used for the assignment (lab work) on disparity map esti- mation using a graph cut, with exact multi-label optimization (cf. slides, p. 112).
3
3.1 Dependence of the neighboring system
In this setting, instead of a 4-pixel neighborhood system , consider the case of 8-pixel neighborhoods, i.e., including diagonals .
1. What bias regarding disparity estimation do we get if we still define the smoothness term as Vp,q(dp, dq) = λ|dp −dq| for any two neighboring pixelsp, q?
2. What definition ofVp,q(dp, dq) should we use instead to prevent this bias?
3.2 Implementation
The question is now how to implement this new definition forVp,q in the frame- work of a linear multi-label graph construction (cf. slides, p. 74+):
1. What should be the weight of edge (pj, qj) for two neighboring pixelsp, q?
2. How is the weight of edgetpj affected?
4