Rank Constraints for Homographies over Two Views: Revisiting the Rank Four Constraint

被引:13
作者
Chen, Pei [1 ,2 ]
Suter, David [3 ]
机构
[1] Sun Yat Sen Univ, Sch Informat Sci & Technol, Guangzhou 510275, Guangdong, Peoples R China
[2] CAS CUHK, Shenzhen Inst Adv Integrat Technol, Shenzhen, Peoples R China
[3] Monash Univ, ARC Ctr Percept & Intelligent Machines Complex En, Dept Elect & Comp Syst Engn, Melbourne, Vic 3004, Australia
关键词
Homography; Rank constraint; First order perturbation; FACTORIZATION METHOD; COMPUTER VISION; MOTION; SHAPE; IMAGES;
D O I
10.1007/s11263-008-0167-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is well known that one can collect the coefficients of five (or more) homographies between two views into a large, rank deficient matrix. In principle, this implies that one can refine the accuracy of the estimates of the homography coefficients by exploiting the rank constraint. However, the standard rank-projection approach is impractical for two different reasons: it requires many homographies to even score a modest gain; and, secondly, correlations between the errors in the coefficients will lead to poor estimates. In this paper we study these problems and provide solutions to each. Firstly, we show that the matrices of the homography coefficients can be recast into two parts, each consistent with ranks of only one. This immediately establishes the prospect of realistically (that is, with as few as only three or four homographies) exploiting the redundancies of the homographies over two views. We also tackle the remaining issue: correlated coefficients. We compare our approach with the "gold standard"; that is, non-linear bundle adjustment (initialized from the ground truth estimate-the ideal initialization). The results confirm our theory and show one can implement rank-constrained projection and come close to the gold standard in effectiveness. Indeed, our algorithm (by itself), or our algorithm further refined by a bundle adjustment stage; may be a practical algorithm: providing generally better results than the "standard" DLT (direct linear transformation) algorithm, and even better than the bundle adjustment result with the DLT result as the starting point. Our unoptimized version has roughly the same cost as bundle adjustment and yet can generally produce close to the "gold standard" estimate (as illustrated by comparison with bundle adjustment initialized from the ground truth). Independent of the merits or otherwise of our algorithm, we have illuminated why the naive approach of direct rank-projection is relatively doomed to failure. Moreover, in revealing that there are further rank constraints, not previously known; we have added to the understanding of these issues, and this may pave the way for further improvements.
引用
收藏
页码:205 / 225
页数:21
相关论文
共 32 条
[1]  
Aguiar P. M., 1999, Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149), P178, DOI 10.1109/CVPR.1999.786936
[2]   Rank 1 weighted factorization for 3D structure recovery: Algorithms and performance analysis [J].
Aguiar, PMQ ;
Moura, JMF .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2003, 25 (09) :1134-1149
[3]  
Aguiar PMQ, 2000, IEEE IMAGE PROC, P549, DOI 10.1109/ICIP.2000.901017
[4]   Three-dimensional modeling from two-dimensional video [J].
Aguiar, PMQ ;
Moura, JMF .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2001, 10 (10) :1541-1551
[5]  
AGUIAR PMQ, 1999, P INT C IM PROC
[6]   Factorization with uncertainty [J].
Anandan, P ;
Irani, M .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2002, 49 (2-3) :101-116
[7]   Lambertian reflectance and linear subspaces [J].
Basri, R ;
Jacobs, DW .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2003, 25 (02) :218-233
[8]  
BASRI R, 1999, P INT C COMP VIS, P383
[9]  
Brand M, 2002, LECT NOTES COMPUT SC, V2350, P707
[10]   A bilinear approach to the parameter estimation of a general heteroscedastic linear system, with application to conic fitting [J].
Chen, P. ;
Suter, D. .
JOURNAL OF MATHEMATICAL IMAGING AND VISION, 2007, 28 (03) :191-208