Deterministic Hypothesis Generation For Robust Fitting of Multiple Structures
Deterministic Hypothesis Generation For Robust Fitting of Multiple Structures
Deterministic Hypothesis Generation For Robust Fitting of Multiple Structures
Abstract— We present a novel algorithm for generating robust and consistent hypotheses for multiple-structure model fitting.
Most of the existing methods utilize random sampling which produce varying results especially when outlier ratio is high. For a
structure where a model is fitted, the inliers of other structures are regarded as outliers when multiple structures are present.
Global optimization has recently been investigated to provide stable and unique solutions, but the computational cost of the
algorithms is prohibitively high for most image data with reasonable sizes. The algorithm presented in this paper uses a maximum
feasible subsystem (MaxFS) algorithm to generate consistent initial hypotheses only from partial datasets in spatially overlapping
local image regions. Our assumption is that each genuine structure will exist as a dominant structure in at least one of the local
regions. To refine initial hypotheses estimated from partial datasets and to remove residual tolerance dependency of the MaxFS
algorithm, iterative re-weighted L1 (IRL1) minimization is performed for all the image data. Initial weights of IRL1 framework are
determined from the initial hypotheses generated in local regions. Our approach is significantly more efficient than those that use
only global optimization for all the image data. Experimental results demonstrate that the presented method can generate more
reliable and consistent hypotheses than random-sampling methods for estimating single and multiple structures from data with a
large amount of outliers. We clearly expose the influence of algorithm parameter settings on the results in our experiments.
Index Terms—Fitting of multiple structures, hypothesis generation, maximum feasible subsystem (MaxFS), iterative re-
weighted L1(IRL1) minimization
—————————— u ——————————
1 INTRODUCTION
inlier ratio or the number of structures. Many multiple- fitting method for multi-structure data has been pro-
structure model fitting methods also start with random hy- posed. To reduce high computation cost, the MaxFS al-
pothesis generation [9, 10, 11, 13, 12, 14, 17, 8]. Some of gorithm is performed for small subset like the method
these algorithms classify dataset based on randomly gen- presented in this paper. However, while initial subset is
erated hypotheses and find model parameters [9, 10, 11]. obtained by sorting keypoint matching scores in [36],
Due to the same nature of random sampling, however, var- the presented method does not require application-spe-
ying results may be produced from the same dataset. cific knowledge. In [36], there exist dependencies be-
Global optimization has recently been actively investi- tween hypotheses generated, since a sequential “fitting-
gated for model fitting problems in computer vision [6, 2, and-removing” procedure is used. Therefore, it is im-
3, 25, 26]. Li has developed a global optimization method possible to generate hypotheses in parallel. On the other
for the algebraic DLT (Direct Linear Transformation) prob- hand, the presented method can be immediately paral-
lem that has fixed bounded variables [2]. He suggested an lelized since there is no dependency between all of the
exact bilinear-MIP (Mixed Integer Programming) formula- hypotheses generated.
tion and obtained globally optimal solutions using an LP The rest of the paper is organized as follows: Section
(Linear Programming)-based BnB (Branch-and-Bound) al- 2 introduces two deterministic methods for geometric
gorithm. In [3], Yu et al. directly solved Big-M based MILP fitting. Section 3 describes our algorithm based on
problem. While these methods guarantee globally optimal MaxFS and IRL1 frameworks. Section 4 shows the ex-
solution, high computational cost is required in general. perimental results on synthetic and real data, and we
For a large dataset, the global optimization methods re- conclude in Section 5.
quire a great deal more running time than RANSAC. Fur-
thermore, the presence of image features from multiple
2 DETERMINISTIC METHODS FOR ROBUST
structures makes their computation cost even higher.
In this paper, we present a novel approach to reliable GEOMETRIC FITTING
and consistent hypothesis generation for multiple- In this section, we briefly describe two main optimiza-
structure model fitting. Unlike previous random sam- tion techniques that we employ in our method.
pling methods, our method generates hypotheses using
deterministic optimization techniques, and thus pro- 2.1 Maximum Feasible Subsystem (MaxFS)
duces consistent results given a set of images. A whole Problem
image dataset is split into spatially overlapping circular The aim of a MaxFS algorithm is to find the largest car-
regions of subsets, and a maximum feasible subsystem dinality set with constraints that are feasible [4, 2, 3]. In
(MaxFS) problem is solved to generate consistent initial other words, it finds a feasible subsystem containing the
hypotheses in each region. Because of the reduction of largest number of inequalities for an infeasible linear sys-
data size using local regions, the MaxFS algorithm can tem Ax ≤ b with real matrix A∈Â ´ , real vector b∈Â and n k k
generate the initial hypothesis from each local region variable x∈Â . The objective of the MaxFS and RANSAC
n
with reasonable efficiency. Since the MaxFS algorithm are the same. However, unlike the RANSAC, the MaxFS
yields a globally optimal solution only for the image guarantees a global solution. The MaxFS problem admits
subset in a spatially local region and the result is influ- the following mixed integer linear programming (MILP)
enced by a residual tolerance value, an iterative re- formulation by introducing a binary variable y for each of i
Fig. 1. Overview of our approach. (a) Overlapping circular regions for hypothesis generation. (b) Detection of dominant structures in local
regions. (c) IRL1 minimization for refining initial hypothesis.
3: t = t+1
The most common deterministic method for geometric
fitting is the least-squares estimation. The best fit mini-
4: xt = arg min | Wt x |1 , s.t.y = Ax
x
mizes the sum of the squared residuals (L2-norm). This
5: 1
method is optimal for Gaussian noise, but the algorithm wit +1 =
mostly fails in the presence of outliers. Using the sum of | xit | +a
the absolute residuals (L1-norm) relatively brings about
better results than using L2-norm in the presence of outliers, 6: Until convergence or a fixed number of times
since L1 minimization puts less weight on large residuals
than L2 minimization. Nevertheless, L1 minimization still
cannot guarantee robustness in the case of severely con- 3 HYPOTHESIS GENERATION USING
taminated data or multiple structured data with outliers DETERMINISTIC OPTIMIZATION TECHNIQUES
because the influence function has no cut off point. Alt- We present a deterministic algorithm to generate a reli-
hough these methods always have a global solution, relia- able and consistent hypothesis for robust fitting of multiple
ble output cannot be guaranteed in the presence of severe structures. The whole data space is split into densely over-
outliers. lapping circular regions (see Fig. 1(a)). In each region, an
Iterative re-weighted L1 (IRL1) minimization has been initial hypothesis is generated from the maximum inliers.
presented by Candѐs, Wakin and Boyd [31]. The IRL1 algo- However, the initial hypothesis may be slightly biased
rithm solves a sequence of weighted L1-norm problems even if it is generated from pure inliers, since the estimated
where the weights are determined according to the esti- inliers are from the local region. To estimate the best fit for
mated coefficient magnitude. The IRL1 minimization algo- each genuine structure from the initial hypothesis, IRL1
rithm is summarized in Algorithm 1. W is a diagonal minimization is performed for all of the data in the image
weighting matrix from the tth iteration with ith diagonal el- (see Fig. 1(c)). The algorithm of our method is summarized
ement wit and a is a stability parameter which affects the in Algorithm 2.
stability of the algorithm.
In [31], experimental results show that it often outper-
forms standard L1 minimization. Although each iteration
of the algorithm solves a convex optimization problem, the
overall algorithm does not. Therefore, one cannot expect
this algorithm to always find a global minimum. Conse-
quentially, it is important to determine a good starting
point for the algorithm. In [31], initializing with the solu-
tion to standard L1 minimization has been introduced.
However, it is unsuitable for problems with heavily con-
taminated multi-structure data.
4
2: For j = 1 to l
6: Repeat
xjc s
7: Solve the weighted L1 minimization prob- s
lem from Eq. (8)
8: Update the weights from Eq. (9)
and thus it suffices for our algorithm to find only one dom- erence for determining the step size s and Rj s for individual
inant structure in a circular region. The remaining struc- circles. We compute the density of data using the 2D Kernel
tures are missed due to the dominant structure’s initial de- Density Estimate (KDE) described in [7], and Figure 2 (c)
tection, yet they can be found in the other local regions. shows the data density map for the image data points
Figure 1 (b) shows an example where an undetected struc- shown in Fig. 2(b). In Fig. 2 (d), the red circle represents the
ture in a window (Fig. 1 (b), middle) becomes a dominant highest-density region with k-nearest neighbors and the
structure in a neighboring window (Fig. 1 (b), bottom). dotted red line does r . maxdensity
The circles’ sizes and positions should be determined The step size s is determined as follows:
depending on the number of data points to be included in
the circular regions. The number of data points k that a cir- s = rmaxdensity α, (3)
cle covers should be larger than the minimum number re-
quired for fitting a desired model for hypothesis genera- where α is a fractional constant. The subset Dj around the
tion. If k becomes larger, the result is more reliable but the circle center xjc is defined as:
computational cost is higher. The smaller the interval be-
tween the circles is, the slimmer the chance of missing
small structures becomes. To maintain even performance
5
D j = {xi Î D | |xi - xcj | £ R j }, If yi=1, the i data is an outlier and the corresponding con-
th
c T Θ = 1, w ( t +1)
i = exp( i ), i = 1,..., N ,
σ
Θ Î Ân , yi ∈{0,1}, i = 1,...,k . (9)
where Mi is a large positive number (called Big-M value). where riΘ̂ t is the residual of data xi to the current hypothe-
The case where yi =0 indicates that the i data is an inlier. th
sis and σ controls the width of the weight function. In each
iteration, Eq. (8) and Eq. (9) are alternately performed (see
6
Fig. 3. Examples for line fitting. (Top row) Input data with different outlier ratio (50%, 66%, 75%, 20%, 33.3%, 42.8%). (Middle row) The gener-
ated hypotheses using the proposed method. (Bottom row) Fitting results.
Fig. 4. Hypothesis refinement through reweighted L1 iterations. (First column) Initial hypotheses from Max-FS framework. (Second column)
Hypotheses after the first reweighted iteration. (Third column) Hypotheses after the third reweighted iteration. (Last column) Fitting results.
(g) Sene
Fig. 5. Test image-sets used for homography estimation experiment. Yellow crosses indicate the gross outliers randomly generated and other
colored squares indicate the inliers of each structure.
to 5, fractional constant α was fixed to 1 and the number of hypothesis is reliable but may not yield the minimum er-
data points in the circular region k was fixed to 30. rors for true inliers. Note that random sampling-based
Table 1 summarizes the performance of four methods for methods produced large variations in their results.
estimating planar homography for seven datasets (see Fig.
5. (a-g)) with various outlier ratios. The results demon- 4.2.3 Performance under Increasing Outlier Rates
strate that our method yields more reliable and consistent We compared the performance of four algorithms un-
results with reasonable computational efficiency. der different outlier rates. We increased the outlier ratio
Affine Fundamental Matrix Estimation: We also tested from 10% to 80% for the CollegeIII data and the Biscuit-
the performance of our proposed method for estimating an bookbox data. Figs. 10(a) and 10(c) show the re-projection
affine fundamental matrix on real image data. For the af- errors produced by the four methods on the two test da-
fine fundamental matrix estimation, the Big-M value was tasets as the outlier ratio increases and Figs. 10(b) and 10(d)
fixed to 10000, the residual tolerance ε was fixed to 0.001, show the corresponding standard deviations for the re-
the variance of weight function σ was fixed to 3, the num- projection errors. Our algorithm outperforms the other al-
ber of iterations in IRL1 also was fixed to 5, fractional con- gorithms when outlier ratio is high over similar periods of
stant α was fixed to 0.5 and the number of data points in time (elapsed computation time). Since the probability of
the circular region k was fixed to 20. producing an all-inlier subset decreases with random sam-
Table 2 shows the performance of the algorithms for esti- pling-based approaches as the outlier ratio increases, the
mating the affine fundamental matrix for seven datasets maximum errors and standard deviations increase
(see Fig. 6. (a-g)) with varied outlier ratios. The results substantially. On the other hand, more robust results were
show that our algorithm generates high-quality hypothe- yielded with our MaxFS-IRL1 method since an initial hy-
ses with reasonable efficiency for the datasets with high pothesis generated from the MaxFS was rarely influenced
outlier ratios and thus finds all the true structures stably by the outlier ratio.
from all of the datasets. In most cases, our method results
in the smallest errors on all the test datasets except for the 4.2.4 Combination of MaxFS and IRL1
Biscuit book dataset with a relatively low outlier ratio. In Figs. 11 and 12, we compared the results using the
Since our method performs IRL1 minimization for all of conventional IRL1 method and using our MaxFS-IRL1
the data in order to generate a hypothesis while other method. We used the CollegeIII dataset. For the conven-
methods generate a hypothesis from minimal subsets, its tional IRL1 method, the initial weights were generated
10
from standard L1 minimization from local datasets to com- model fitting. For reliable hypothesis generation with rea-
pete on par with the MaxFS-IRL1 method. We can see the sonable computational efficiency, we employ a MaxFS
contribution of each stage from Figs. 11 and 12. (maxium feasible subsystem) algorithm, a global optimiza-
Figures 11(a) and 11(d) show initial fitting results from tion technique, only in spatially localized image regions
standard L1 minimization and the MaxFS algorithm for the and refine the hypotheses using an IRL1 (re-weighted L1)
CollegeIII data with no outliers included. Figures 11(b) and minimization method. To search out all of the structures
11(e) show the fitting results after the first reweighted iter- thoroughly, the local circular regions spatially overlap with
ation. Figs. 11(c) and 11(f) show the final fitting results the neighboring ones, and the number of data in each local
from the conventional IRL1 method and the MaxFS-IRL1 region is distributed evenly for computational efficiency.
method. The model parameters of major structure are estimated in
Figures 12(a-c) show fitting results from the conven- each local image region, and those of the remainder can be
tional IRL1 method in each iteration step and Figures 12(d- found in one of the neighboring regions. The IRL1 minimi-
f) show those from the MaxFS-IRL1 method for the Col- zation is performed over all the image data to refine initial
legeIII data with 30% of the outliers included. hypotheses estimated from subsets and to get rid of resid-
As appears by these results, when the outlier ratio is low, ual tolerance dependency of the MaxFS algorithm.
the standard L1 minimization can provide good results. Our experiments show that without prior knowledge of
When there are severe outliers, however, standard L1 min- the inlier ratio, inlier scale and the number of structures,
Fig. 10. The performance of four algorithms under different outlier rates. Graphs (a) and (c) show the re-projection errors produced by the four
methods on the two test datasts as outlier ratio increases. (b) and (d) show the corresponding standard deviations for the re-projection errors.
imization frequently fails. To obtain good fitting results us- our method generates consistent hypotheses which are
ing the IRL1 method, it is important to determine good in- more reliable than the random sampling-based methods as
itial weights for the algorithm. Therefore, the MaxFS and outlier ratios increase.
the IRL1 algorithms are complementary and we show that
their combination is highly effective for the task of robust
multiple-structure fitting.
5 CONCLUSION
We present a new deterministic approach to reliable and
consistent hypothesis generation for multiple-structure
11
Fig. 11. (a) and (d) show initial fitting results from the standard L1 minimization and the MaxFS algorithm for the CollegeIII data with no outlier
included. (b) and (e) show the fitting results after the first reweighted iteration. (c) and (f) show the final fitting results from the conventional
IRL1 method and the MaxFS-IRL1 method.
Fig. 12. (a) and (d) show initial fitting results from the standard L1 minimization and the MaxFS algorithm for the CollegeIII data with 30% of
the outlier included. (b) and (e) show the fitting results after the first reweighted iteration. (c) and (f) show the final fitting results from the
conventional IRL1 method and the MaxFS-IRL1 method.
TABLE 1
PERFORMANCE OF VARIOUS METHODS ON HOMOGRAPHY ESTIMATION FOR THE SEVERAL REAL DATASETS.
TABLE 2
PERFORMANCE OF VARIOUS METHODS ON AFFINE FUNDAMENTAL MATRIX ESTIMATION FOR THE SEVERAL REAL DATASETS.
[7] Z. I. Botev, J. F. Grotowski and D. P. Kroese Kernel density esti- [35] http://cs.adelaid.edu.au/~hwong/doku.php?id=data
mation via diffusion. Annals of Statistics 2010. Volume 38, Num- [36] Kwang Hee Lee, Sang Wook Lee.: Deterministic Fitting of Mul-
ber 5, Pages 2916—2957 tiple Structures using Iterative MaxFS with Inlier Scale Estima-
[8] H. Li. Two-view motion segmentation from linear programming tion. In ICCV 2013.
relaxation. In CVPR, 2007.
[9] R. Toldo, A. Fusiello. Robust Multiple Structures Estimation
with JLinkage. In ECCV 2008: 537-547.
[10] T.-J. Chin, H. Wang, D. Suter. Robust Fitting of Multiple Struc-
tures: The Statistical Learning Approach. In ICCV 2009 : 413-420.
[11] W. Zhang, J. Kosecka. Nonparametric Estimation of Multiple
Structures with Outliers. LNCS, DVW 2006: 4358. 60-74.
[12] Y. Kanazawa, H. Kawakami. Detection of planar regions with
uncalibrated stereo using distributions of feature points. In:
BMVC 2004.
[13] M. Zuliani, C. Kenney and B. Manjunath. The multiransac algo-
rithm and its application to detect planar homographies. In ICIP
2005 : Volume 3.
[14] H. S. Wong, T.–J. Chin, J.Yu and D. Suter. Efficient Multi-struc-
ture Robust Fitting with Incremental Top-k Lists Comparison. In
ACCV 2010: 553-564.
[15] T.-J. Chin, J. Yu and D. Suter. Accelerated Hypothesis Genera-
tion for Multistructure Data via Preference Analysis. IEEE Trans.
Pattern Anal. Mach. Intell. 34(4): 2012 625-638
[16] H. S. Wong, T.-J. Chin, J. Yu and D. Suter. Dynamic and hierar-
chical multi-structure geometric model fitting. In ICCV 2011:
1044-1051
[17] A. Delong, A. Osokin, H. Isack, and Y. Boykov. Fast approximate
energy minimization with label costs. In CVPR, 2010.
[18] O. Chum, J. Matas and J. Kittler. Locally optimized RANSAC. In:
DAGM 2003.
[19] B. J. Tordo and D. W. Murray. Guided-MLESAC: Faster image
transform estimation by using matching priors. IEEE Trans. Pat-
tern Anal. Mach. Intell. (27) 2005:1523-1535
[20] O. Chum and J. Matas. Matching with PROSAC- progressive
sample consensus. In CVPR.2005.
[21] K. Ni, H. Jin and F. Dellaert. GroupSAC: Efficient consensus in
the presence of groupings. In ICCV. 2009.
[22] T. Sattler, B. Leibe and L. Kobbelt. SCRAMSAC: Improving
RANSAC's efficiency with a spatial consistency filter. In: ICCV
2009.
[23] T.-J. Chin, J. Yu, and D. Suter. Accelerated hypothesis generation
for multi-structure robust fitting. In ECCV 2010.
[24] M. A. Fischler and R. C. Bolles. RANSAC: A paradigm for model
fitting with applications to image analysis and automated car-
tography. Comm. of the ACM 24 1981: 381-395
[25] R. Hartley and F. Kahl. Global Optimization through Rotation
Space Search. IJCV, 82(1):64-79, 2009.
[26] F. Kahl. Multiple view geometry and the L-infinity norm. In
ICCV 2005: 1002–1009.
[27] D. R. Myatt, Philip H. S. Torr, Slawomir J. Nasuto, J. Mark Bishop,
R. Craddock. NAPSAC: High Noise, High Dimensional Robust
Estimation - it's in the Bag. In BMVC 2002.
[28] K. Schindler and D. Suter. Two-View Multibody Structure-and-
Motion with Outliers through Model Selection. IEEE Trans. Pat-
tern Anal. Mach. Intell. 28(6): 983-995, 2006.
[29] R. Raguram, J. M. Frahm and M. Pollefeys. Exploiting uncer-
tainty in random sample consensus. In ICCV 2009: 2074-2081.
[30] Gurobi optimization. http://www.gurobi.com/, 2010
[31] E. J. Cand`es, M. B. Wakin and S. P. Boyd. Enhancing sparsity by
reweighted L1 minimization, The Journal of Fourier Analysis
and Applications, vol. 14, no. 5, pp. 877–905, 2008.
[32] http://www.robots.ox.ac.uk/~vgg/data/
[33] http://www.csse.uwa.edu.au/~pk/research/matlabfns/
[34] http://cs.adelaide.edu.au/~tjchin/doku.php