presentazione di powerpoint - politecnico di...
TRANSCRIPT
1zyx
⋅−⋅=⋅−⋅= oIMXoIMu
)(1
oxMx
oIMu −⋅=⋅−⋅=
i.e.,
defining
x = [x, y, z]T
oIMoMMmMtKRKP −⋅=⋅−==⋅⋅=
with RKM ⋅= and mMo ⋅−= −1
The locus
of
the points
x whose
image
is
uis
a straight
line through
o having
direction
mMo ⋅−= −1 is
independent
of u
o is
the camera viewpoint
(perspective
projection
center)
uMd ⋅= −1
line(o, d)
= Interpretation
line of image
point
u
Interpretation of o:
u is
image
of x if uMox ⋅=− −1)( λ
i.e., if uMox ⋅+= −1λ
Intrinsic and extrinsic parameters from P
RKM ⋅=
M K and R
1111 −−−− ⋅=⋅= KRKRM T
RQ-decomposition
of a matrix: as
the product
betweenan
orthogonal
matrix
and an
upper triangular
matrix
M and
m t
tRKtKRKtKR)mMo 1T1 T−=−=−=⋅−= −−− (1
Rot −=
Camera center
0P =O
null-space camera projection matrix
Oλ)(1λAX −+=
Oλ)P(1λPAPXx −+==
For all A all points on AO project on image of A,
therefore O is camera center
Image of camera center is (0,0,0)T, i.e. undefined
Finite cameras: ⎟⎟⎠
⎞⎜⎜⎝
⎛−=⎟⎟
⎠
⎞⎜⎜⎝
⎛=
−
1M
1
1moO
Infinite cameras: 0Md,0d
=⎟⎟⎠
⎞⎜⎜⎝
⎛=o
Action of projective camera on point
PXx =
[ ] MdDm|MPDx ===
Forward projection
Back-projection
xPX += ( ) 1PPPP −+ = TT IPP =+
(pseudo-inverse)
0P =O
( ) OλxPλX += +
( ) ( )⎟⎟⎠
⎞⎜⎜⎝
⎛=⎟⎟
⎠
⎞⎜⎜⎝
⎛+⎟⎟⎠
⎞⎜⎜⎝
⎛=
1m-μxM
1mM-
0xM
μλX-1-1-1
xMd -1=
OD
Camera matrix decomposition
Finding the camera center
0P =O (use SVD to find null-space)
[ ]( )m,p,pdet 32=X [ ]( )m,p,pdet 31−=Y
[ ]( )m,p,pdet 21=Z [ ]( )m,p,pdet 21−=TFinding the camera orientation and internal parameters
KRM = (use RQ decomposition ~QR)
Q R=( )-1= -1 -1QR
(if only QR, invert)
Radial distortion
Data normalization
32ii UXX~ =
ii Tuu~ =
(i)
translate
origin
to
gravity
center(ii)
(an)isotropic
scaling
Exterior orientation
Calibrated camera, position and orientation unkown
→ Pose estimation
6 dof
⇒ 3 points minimal (4 solutions in general)
What does calibration give?
xKd 1−=
⎥⎦⎤
⎢⎣⎡= 0d0]|K[Ix
( )( ) ( )( )21-T-T
211-T-T
1
2-1-TT
1
2T
21T
1
2T
1
)xK(Kx)xK(Kx
)xK(Kx
dddd
ddcos ==θ
An image line l defines a plane through the camera center with normal n=KTl measured in the camera’s Euclidean frame. In fact the backprojection of l is PTl n=KTl
The image of the absolute conic
KRd0d
O]|KR[IPXx =⎟⎟⎠
⎞⎜⎜⎝
⎛−== ∞
mapping between π∞
to an image is given by the planar homogaphy
x=Hd, with H=KR
absolute conic (IAC), represented by I3 within
π∞
(w=0)
( ) 1-T-1T KKKKω ==− ( )1TCHHC −−α
(i) IAC depends only on intrinsics(ii) angle between two rays(iii) DIAC=ω*=KKT
(iv) ω⇔ K (Cholesky factorization)(v) image of circular points belong
to ω
(image of absolute conic)
( )( )2T
21T
1
2T
1
ωxxωxx
ωxxcos =θ
∞Ω
its
image
(IAC)
A simple calibration device
(i) compute Hi for each square (corners (0,0),(1,0),(0,1),(1,1))
(ii) compute the imaged circular points Hi [1,±i,0]T
(iii) fit a conic ω
to 6 imaged circular points(iv) compute K from ω=K-T K-1 through Cholesky
factorization(= Zhang’s calibration method)
Properties of perspective transformations
1) vanishing points
V image
of the point
at the ∞
along
direction d
dMd
mMuV ⋅=⋅=0
VuMd ⋅= −1
the interpretation
line of
V is
parallel
to
d
O
V
∞Pd
The images
of parallel
lines
are concurrent
lines
2) cross ratio invariance
Given
four
colinear
points ( )4321 ,,, pppp
( )
42
32
41
31
4321 ,,,
xxxxxxxx
CR
−−−−
=pppp
let ( )4321 ,,, xxxx be
their
abscissae
Properties of perspective transformations ctd.
Cross ratio invariance
under perspective
transformation
TT ],0,0,[],,,[ txtzyx ==Xa point
on the line y=0=z
xPu T ⋅=⋅== 144
1],[tx
wupp
its
image belongs
to
a line
its
coordinate u
XPu ⋅== T],,[ wvu
( )4231
43214321 ,det,det
,det,det,,,
uuuuuuuu
uuuu =CR42143114
43142114
,detdet,detdet,detdet,detdetxxPxxPxxPxxP
=
4231
4321
,det,det,det,detxxxxxxxx
= ( )4321 ,,, xxxxCR=
Object localization 1: three colinear points
geometric
model of an
objecta perspective
image
of the objectposition and orientationof the object
?
A
B
C
C’
A’
B’O
π
calibrated
camera: mMP = A
B
C
known
known interpretation lines
A
B
C
C’
A’
B’O
π
∞
V
a) orientation
( )cbcaCBACRVCBACR
−−
=∞= ),,,(,',','
Cross ratio invariance:
solve for
V (image
of ∞)
V: vanishing
point
of the direction of (A,B,C)
interpretation
line of V parallel
to
(A,B,C) VuM ⋅−1direction
b) position (e.g., distance(O,A))
A
B
C
C’
A’
B’O
π
V
VuM ⋅−1
CuM ⋅−1
AuM ⋅−1
interpretation
lines
angles
α
and γ
α
γ
αγ
sinsinACOA =
Object localization 2: four coplanar points
O
(i)
orientation
of (A,E,C)(ii)
orientation
of (B,E,D)(iii)
distance
(O,A)
E
A
B
CD
a’
b’b”
a”
Find
vanishing
point
of the field-bottom
line direction
Off-side
images
of symmetric
segments
a
and b: abscissae
of the endpoints
of a segmentc=(a+b)/2: abscissa
of segment
midpoint,d=∞: point
at the infinite along
the segment
direction
a c b d
( ) 1,,, −=−−
=
−−−−
=cbca
dbdacbca
dcbaCR
(a’,b’) and (a”, b”) are image
of symmetric
segmentssame image of the midpoint c’, same vanishing point d’
Harmonic
4-tuple (a,b,c,d)
solve
( ) 1
''''
''''
',',',' −=
−−
−−
=
dbda
cbca
dcbaCR
( ) 1
''''''
''''''
',','','' −=
−−
−−
=
dbda
cbca
dcbaCR
{ for
c’, d’
system of two linear equations in (c’d’) and (c’+d’)
two degree equation, whose solutions are c’ and d’
among
the two
solutions, the one for
d’
is
the value
external
to
the range
[a’,b’]
Orthogonality relation
( )( )2T
21T
1
2T
1
ωvvωvv
ωvvcos =θ
0ωvv 2T
1 =
0lωl 2*T
1 =
Calibration from vanishing points and lines
Calibration from vanishing points and lines
Two-view geometry
Epipolar geometry
F-matrix comp.
3D reconstruction
Structure comp.
(i) Correspondence geometry: Given an image point x
in the first view, how does this constrain the position of the corresponding point x’
in the second image?
(ii) Camera geometry (motion): Given a set of corresponding image points {xi ↔x’i
}, i=1,…,n, what are the cameras P and P’
for the two views?
(iii) Scene geometry (structure): Given corresponding image points xi ↔x’i
and cameras P, P’, what is the position of (their pre-image) X in space?
Three questions:
The epipolar geometry
C,C’,x,x’
and X
are coplanar
The epipolar geometry
What if only C,C’,x
are known?
The epipolar geometry
All points on π
project on l
and l’
The epipolar geometry
Family of planes π
and lines l and l’Intersection in e and e’
The epipolar geometry
epipoles
e,e’= intersection of baseline with image plane = projection of projection center in other image= vanishing point of camera motion direction
an epipolar
plane = plane
containing baseline (1-D family)
an epipolar
line = intersection of epipolar
plane with image(always come in corresponding pairs)
Example: converging cameras
Example: motion parallel with image plane
Example: forward motion
e
e’
The fundamental matrix Falgebraic representation of epipolar
geometry
l'x α
we will see that mapping is (singular) correlation (i.e. projective mapping from points to lines) represented by the fundamental matrix F
The fundamental matrix F
geometric derivation
xHx' π=
x'e'l' ×= [ ] FxxHe' π == ×
mapping from 2-D to 1-D family (rank 2)
The fundamental matrix F
algebraic derivation
( ) λC0
xMλX
-1
+⎥⎦
⎤⎢⎣
⎡=
[ ] 1MM'e'F −×=
[ ] [ ] [ ] xMM'Cm'M'0
xMm''MC'm'Ml' 1
1−
−
×=⎥⎦
⎤⎢⎣
⎡×=
(note: doesn’t work for C=C’
⇒ F=0)
xP+
( )λX0Fxx'0 l' x' l' x' T =→=→∈
The fundamental matrix F
correspondence condition
0Fxx'T =
The fundamental matrix satisfies the condition that for any pair of corresponding points x↔x’ in the two images ( )0l'x'T =
The fundamental matrix F
F is the unique 3x3 rank 2 matrix that satisfies x’TFx=0 for all x↔x’
(i) Transpose: if F is fundamental matrix for (P,P’), then FT
is fundamental matrix for (P’,P)(ii) Epipolar lines: l’=Fx
& l=FTx’(iii) Epipoles: on all epipolar
lines, thus e’TFx=0, ∀x ⇒e’TF=0, similarly Fe=0
(iv) F has 7 d.o.f. , i.e. 3x3-1(homogeneous)-1(rank2)(v) F is a correlation, projective mapping from a point x to
a line l’=Fx
(not a proper correlation, i.e. not invertible)
The epipolar line geometry
l,l’
epipolar
lines, k line not through e⇒ l’=F[k]x
l
and symmetrically l=FT[k’]x
l’
lk× e
kl lFk×
e'
(pick k=e, since eTe≠0)
[ ] leFl' ×= [ ] l'e'Fl T×=
Fundamental matrix for pure translation
Fundamental matrix for pure translation
Fundamental matrix for pure translation
[ ] [ ]×∞× == e'He'F ( )RKKH 1−∞ =
×⎥⎥⎦
⎤
⎢⎢⎣
⎡=
0101-00000
F( )T1,0,0e'=
example:
y'y =⇔= 0Fxx'T
0]X|K[IPXx ==
⎥⎦⎤
⎢⎣⎡== Z
xKt]|K[IXP'x'-1
ZKt/xx' +=
ZX,Y,Z x/K)( -1T =
motion starts at x and moves towards e, faster depending on Z
pure translation: F only 2 d.o.f., xT[e]x
x=0 ⇒ auto-epipolar
General motion
Zt/K'xRKK'x' -1 +=
[ ] 0Hxe''x =×T
[ ] 0x̂e''x =×T
Geometric representation of F
( ) 2/FFF TS += ( ) 2/FFF T
A −= ( )AS FFF +=
0FxxT =xx ↔ ( )0xFx AT ≡
0xFx ST =
Fs
: Steiner conic, 5 d.o.f.Fa
=[xa
]x
: pole of line ee’
w.r.t. Fs
, 2 d.o.f.
Geometric representation of F
Pure planar motion Steiner conic Fs
is degenerate (two lines)
Projective transformation and invariance
-1-T FHH'F̂ x'H''x̂ Hx,x̂ =⇒==
Derivation based purely on projective concepts
( )( ) X̂P̂XHPHPXx -1 ===
F invariant to transformations of projective 3-space
( )( ) X̂'P̂XHHP'XP'x' -1 ===
( ) FP'P, α
( )P'P,Fαunique
not uniquecanonical form
m]|[MP'0]|[IP
== [ ] MmF ×=
Projective ambiguity of cameras given Fprevious slide: at least projective ambiguitythis slide: not more!
Show that if F is same for (P,P’) and (P,P’), there exists a projective transformation H so that P=HP and P’=HP’
~ ~
~ ~
]a~|A~['P~ 0]|[IP~ a]|[AP' 0]|[IP ====
[ ] [ ] A~a~AaF ×× ==
( )T1 avAA~ kaa~ +== −klemma:
[ ] kaa~Fa~0AaaaF2rank
==== ⇒×
[ ] [ ] [ ] ( ) ( ) TavA-A~k0A-A~kaA~a~Aa =⇒=⇒= ×××
⎥⎦⎤
⎢⎣⎡= −
−
kkIkH T1
1
v0
( ) 'P~]a|av-A[v0a]|[AHP' T1
T1
1==⎥⎦
⎤⎢⎣⎡= −
−
−
kkkkIk
(22-15=7, ok)
Canonical cameras given FF matrix corresponds to P,P’
iff
P’TFP is skew-symmetric
( )X0,FPXP'X TT ∀=
F matrix, S skew-symmetric matrix
]e'|[SFP' 0]|[IP ==
⎟⎠⎞
⎜⎝⎛
⎥⎦⎤
⎢⎣⎡=⎥⎦
⎤⎢⎣⎡= 00
0FSF0Fe'0FSF0]|F[I]e'|[SF
TT
T
TTT
(fund.matrix=F)
Possible choice:
]e'|F][[e'P' 0]|[IP ×==
Canonical representation:
]λe'|ve'F][[e'P' 0]|[IP T+== ×
The essential matrix~fundamental matrix for calibrated cameras (remove K)
[ ] ×× == t]R[RRtE T
0x̂E'x̂ T =
FKK'E T=
( ) x'K'x̂ x;Kx̂ -1-1 ==
5 d.o.f. (3 for R; 2 for t up to scale)
E is essential matrix if and only iftwo singularvalues
are equal (and third=0)
T0)VUdiag(1,1,E =SVD
⎥⎥⎥
⎦
⎤
⎢⎢⎢
⎣
⎡ −=
100001010
WGiven
Motion from E
TTVUWR'=
TUWV'R' =
and
3Ut'=
3U't' −=
Four
solutions
Four possible reconstructions from E
(only one solution where points is in front of both cameras)
C1
C2
l2
π
l1e1
e20m m 1
T2 =F
Fundamental matrix (3x3 rank 2 matrix)
1. Computable from corresponding points
2. Simplifies matching3. Allows to detect wrong matches4. Related to calibration
Underlying structure in set of matches for rigid scenes
l2
C1m1
L1
m2
L2
M
C2
m1
m2
C1
C2
l2
π
l1e1
e2
m1
L1
m2
L2
M
l2lT1
Epipolar geometry
Canonical representation:
]λe'|ve'F][[e'P' 0]|[IP T+== ×
1T
1 lPπ =
1T
1T
22 lPPl +=
0lx 2T
2 =
[ ] 111 xel ×=
0x][ePPx 1x1T
1T
2T2 =+
3D reconstruction of cameras and structure
given xi ↔x‘i
, compute
P,P‘
and Xi
reconstruction problem:
ii PXx = ii XPx ′=′ for
all i
without additional information possible up to projective ambiguity
The projective reconstruction theorem
If
a set
of point correspondences
in two
views
determine
the
fundamental matrix
uniquely, then
the
scene
and cameras
may
be
reconstructed
from
these
correspondences
alone, and any
two
such reconstructions
from
these
correspondences
are
projectively
equivalent
The projective reconstruction theorem
If
a set
of point correspondences
in two
views
determine
the
fundamental matrix
uniquely, then
the
scene
and cameras
may
be
reconstructed
from
these
correspondences
alone, and any
two
such reconstructions
from
these
correspondences
are
projectively
equivalent
{ }( )i111 X,'P,P { }( )i222 X,'P,Pii xx ′↔-1
12 HPP = -112 HPP ′=′ 12 HXX = ( )0FxFx :except =′= ii
theorem
from
last class
( ) iiiii 22111-1
112 XPxXPHXHPHXP ====⇒ along
same
ray
of
P2, idem for
P‘2
two
possibilities: X2i
=HX1i
, or
points
along
baselinekey result: allows reconstruction from pair of uncalibrated images
ObjectiveGiven
two
uncalibrated
images
compute
(PM
,P‘M
,{XMi
})(i.e. within
similarity
of original scene
and cameras)
Algorithm(i)
Compute
projective
reconstruction
(P,P‘,{Xi
})(a)
Compute
F from
xi
↔x‘i(b)
Compute
P,P‘
from
F(c)
Triangulate
Xi
from
xi
↔x‘i(ii)
Rectify
reconstruction
from
projective
to metricDirect method: compute
H from
control
points
Stratified method:(a) Affine reconstruction: compute
π∞
(b) Metric reconstruction: compute
IAC ω
ii HXXE =-1
M PHP = -1M HPP ′=′ ii HXXM =
⎥⎦⎤
⎢⎣⎡=
∞π0|I H
⎥⎦⎤
⎢⎣⎡= 10
0AH-1
( ) 1TT ωMMAA −=
Stratified reconstruction
(i)
Projective
reconstruction(ii)
Affine reconstruction(iii)
Metric
reconstruction
Projective to affine
remember
2-D case
Projective to affine
{ }( )iX,P'P,
( ) ( )TT 1,0,0,0,,,π αDCBA=∞
( )TT 1,0,0,0πH- =∞
⎥⎦⎤
⎢⎣⎡=
∞π0|I H (if
D≠0)
theorem
says
up to projective
transformation, but
projective
with
fixed
π∞
is
affine transformation
can
be
sufficient
depending
on application, e.g. mid-point, centroid, parallellism
Translational motionpoints
at infinity
are
fixed
for
a pure translation⇒ reconstruction
of xi
↔ xi
is
on π∞
×× == ]e'[]e[F 0]|[IP =]e'|[IP =
Scene constraintsParallel linesparallel lines
intersect
at infinityreconstruction
of corresponding
vanishing
point yieldspoint on plane at infinity
3 sets
of parallel lines
allow
to uniquely
determine
π∞
remark: in presence
of noise
determining
the
intersection
of parallel lines
is
a delicate
problem
remark: obtaining
vanishing
point in one
image can
be
sufficient
Scene constraints
Scene constraintsDistance ratios on a line
known
distance ratio
along
a line
allow
to determine
point at infinity
(same
as 2D case)
Affine to metricidentify
absolute conic
transform
so that ∞∞ =++Ω on π ,0: 222 ZYX
then
projective
transformation
relating
original and reconstruction
is
a similarity
transformation
in practice, find image of Ω∞image ω∞
back-projects
to cone
that
intersects
π∞
in Ω∞
ω*
Ω*
projection
constraints
note
that
image is
independent of particular
reconstruction
Affine to metric
m]|[MP = ω
⎥⎦⎤
⎢⎣⎡= 10
0AH-1
given
possible
transformation
from
affine to metric
is
( ) 1TT ωMMAA −=
(cholesky
factorisation)
m]|[MAPHP -1M ==
proof:
TTTMM
* MMAAMMω ==T-T-1-1 AAMωM =
⎟⎠⎞
⎜⎝⎛
⎥⎦⎤
⎢⎣⎡=Ω 00
0I*
Orthogonality
0ωvv 2T1 =
ωvl =
vanishing
points
corresponding
to orthogonal directions
vanishing
line
and vanishing
point correspondingto plane and normal direction
known internal parameters
-1-TKKω =
0ωω 2112 ==0=s
rectangular
pixels
yx αα =
square
pixels
2211 ωω =
Same camera for all images
same
intrinsics
⇒ same
image of the
absolute conic
e.g. moving
cameras
given
sufficient
images
there
is
in general
only
oneconic
that
projects
to the
same
image in all images,i.e. the
absolute conic
This
approach
is
called
self-calibration, see
later
-1-TωHHω' ∞∞=transfer
of IAC:
Direct metric reconstruction using ω
KKKω -1-T ⇒=
approach
1
calibrated
reconstruction
approach
2
compute
projective
reconstruction
back-project
ω
from
both
images
intersection
defines
Ω∞
and its
support
plane π∞
(in general
two
solutions)
Direct reconstruction using ground truth
ii HXXE =
use
control
points
XEi
with
know
coordinatesto go
from
projective
to metric
Eii XPHx -1=(2 lin. eq. in H-1
per view, 3 for
two
views)
ObjectiveGiven
two
uncalibrated
images
compute
(PM
,P‘M
,{XMi
})(i.e. within
similarity
of original scene
and cameras)
Algorithm(i)
Compute
projective
reconstruction
(P,P‘,{Xi
})(a)
Compute
F from
xi
↔x‘i(b)
Compute
P,P‘
from
F(c)
Triangulate
Xi
from
xi
↔x‘i(ii)
Rectify
reconstruction
from
projective
to metricDirect method: compute
H from
control
points
Stratified method:(a) Affine reconstruction: compute
π∞
(b) Metric reconstruction: compute
IAC ω
ii HXXE =-1
M PHP = -1M HPP ′=′ ii HXXM =
⎥⎦⎤
⎢⎣⎡=
∞π0|I H
⎥⎦⎤
⎢⎣⎡= 10
0AH-1
( ) 1TT ωMMAA −=
Image information provided View relations and projective objects
3-space objects
reconstruction ambiguity
point correspondences F projective
point correspondences including vanishing points
F,H∞ π∞affine
Points correspondences and internal camera calibration F,H∞
ω,ω’π∞Ω∞
metric
UncalibratedUncalibrated
visual visual odometryodometry for ground plane motionfor ground plane motion
(joint work with Simone Gasparini)
Problem formulation
•
Given:–
an uncalibrateduncalibrated camera mounted on a robot
–
the camera is fixedfixed and aims at the floor–
the robot moving on a planarplanar floor (ground plane)
•
Determine:–
the estimate of robot motion from observed features
on the floor
Technique involvedEstimate the ground plane transformation (homography) between images taken before and after robot displacement
Motivations•
Dead reckoning techniques are not reliable and diverge after few steps [Borenstein96]
•
Visual odometry
techniques exploits cameras to recover motion
•
We use a single uncalibrated camera–
3D reconstruction with uncalibrated
camera
usually require auto-calibration–
Non-planar motion is required [Triggs98]
–
Planar motion with different camera attitudes [Knight03]
–
Special devices is required (e.g. PTZ cameras)
Problem formulation
•
Fixed uncalibrated
camera mounted on a robot
•
The pose of the camera w.r.t. the robot is unknown
•
The projective transformation between ground plane and image plane is a homography
T (3x3)
•
T does not change with robot motion••
T is T is uknownuknown
Problem formulation
•
Robot and camera undergo a planar motion–
rotation of unknown angle θ
about unknown
vertical axis–
2D rotation matrix R (3x3 with homogeneous coordinates)
2x2ˆ ˆ orthogonal matrix0 translation vector (3x1 )
R R⎡ ⎤= ⎢ ⎥⎣ ⎦
R tt
Problem
formulation
•
Given 2 images before and after robot displacement determine:–
Rotation centre C
–
Rotation angle θ–
Determine unknown
transformation T
between
ground plane and image plane
Ground reference frame
•
We define a ground reference frame–
O is the origin of the projected reference frame
•
E.g. O is the backprojection
of image point O’(0,0)
–
The vector connecting the origin to a point A is the unit vector along the x-axis
•
E.g. A is the backprojection
of image point A’(100,0)
Estimation of robot displacement
•
Transformation relating the two images is still a homography
•
Eigenvectors of H:
1 ( unknown)−H = TRT T
( )' associated to the eigval, C rotation center '
associated to the eigval, I and J circular points'
C TCeigvec I TI
J TJ
= →⎧⎪ = ⎫⎨ →⎬⎪ = ⎭⎩
H =¡
£
•
Since eig(H)=eig(R)–
Angle θ
can be calculated through the ratio of
imaginary and real part of the complex eigenvalue
of H
–
E.g. the eigenvalue
associated to I’
is μe±iθ
Estimation of robot displacement: degenerate motion
Pure translation: a frequent motion on planar floor
•
Under pure translation, the whole line at the infinity is invariant
ground plane can be only rectified modulo an affine transformation:e.g., orientation of translaton
wrt
ground reference
can not be determined
•
In practice:–
Use a motion including rotations to estimate ground plane to image plane transformation T
–
Use knowledge of T while translating
Estimation of the transformation T
•
Estimating T allows to determine the shape of observed features
•
Four pairs of corresponding points are needed–
The 2 circular points I and J
–
The 2 points A and O used for ground reference
[ ][ ][ ] [ ][ ] [ ]
where 1, ,0
where 1, ,0
where 0,0,1 and ' 0,0,1
where 1,0,1 and ' 100,0,1
T
T
T T
T T
i
i
⎧ =⎪⎪ = −⎪⎨
= =⎪⎪
= =⎪⎩
I'=TI I
J'=TJ J
O'=TO O O
A'=TA A A
Experimental set up
•
Fixed perspective camera placed on turntable ( ground truth)
•
Optical distortion is negligible•
Camera pointing towards the ground floor
•
Camera viewpoint in a generic position wrt rotation axis
•
Basic algorithm:–
Feature (corner) extraction from floor texture [Harris88]
–
Feature tracking and matching in order to determine correspondences
–
Corresponding features used to fit the homography
H
–
Eigendecomposition
of matrix H•
Rotation angle
•
Image of circular points•
Image of rotation centre
•
Ground plane to image plane transformation
Experiments on sequences of images
Sequence of large displacements
•
Image displacements of the order of 10°
•
Features extracted and matched [Torr93]
•
Use best matching pairs to fit a homography using RANSAC [Fischler81]
•
Good overall accuracy (error < 1°)•
Larger displacements affect matching
Experiments on sequences of images
Sequence of small displacements•
Image displacements of about 5°
•
Small rotations may lead to numerical instability
•
Track features over three images–
H13
fitted with correspondences of 1 and 3
•
Good overall accuracy (error < 1°)
HH13131 2 3
Experiments on a mobile platform
Feature extraction and matching between images
Rotation and relevant centre of rotation determination