Generalised Hough transform
The Hough transform was initially developed to detect analytically defined shapes (e.g., line, circle, ellipse etc.). In these cases, we have knowledge of the shape and aim to find out its location and orientation in the image. The Generalized Hough Transform or GHT, introduced by Dana H. Ballard in 1981, is the modification of the Hough Transform using the principle of template matching.[1] This modification enables the Hough Transform to be used for not only the detection of an object described with an analytic function. Instead, it can also be used to detect an arbitrary object described with its model.
The problem of finding the object (described with a model) in an image can be solved by finding the model's position in the image. With the generalized Hough Transform, the problem of finding the model's position is transformed to a problem of finding the transformation's parameter that maps the model into the image. As long as we know the value of the transformation's parameter, the position of the model in the image can be determined.
The original implementation of the GHT uses edge information to define a mapping from orientation of an edge point to a reference point of the shape. In the case of a binary image where pixels can be either black or white, every black pixel of the image can be a black pixel of the desired pattern thus creating a locus of reference points in the Hough Space. Every pixel of the image votes for its corresponding reference points. The maximum points of the Hough Space indicate possible reference points of the pattern in the image. This maximum can be found by scanning the Hough Space or by solving a relaxed set of equations, each of them corresponding to a black pixel. [2]
Earlier Work
Merlin and Farber[3] showed how to use a Hough algorithm when the desired curves could not be described analytically. It was a precursor to Ballard’s algorithm but was restricted to translation and didn’t take into account rotation and scale changes.[4]
The Merlin-Farber algorithm is impractical for real image data as in an image with a large number of edge pixels, there will be many false instances of the desired shape due to similar pixel arrangements.
Theory of generalized Hough Transform
To generalize the Hough algorithm to non-analytic curves, Ballard defines the following parameters for a generalized shape: a={y,s,θ} where y is a reference origin for the shape, θ is its orientation, and s = (sx, sy) describes two orthogonal scale factors. As in the case of initial Hough Transforms, there is an algorithm for computing the best set of parameters for a given shape from edge pixel data. These parameters no longer have equal status. The reference origin location, y, is described in terms of a template table called the R table of possible edge pixel orientations. The computation of the additional parameters s and θ is then accomplished by straightforward transformations to this table. The key to generalizing the Hough algorithm to arbitrary shapes is the use of directional information. Given any shape and a fixed reference point on it, instead of a parametric curve, the information provided by the boundary pixels is stored in the form of the R-table in the transform stage. For every edge point on the test image, the properties of the point are looked up on the R-table and reference point is retrieved and the appropriate cell in a matrix called the Accumulator matrix is incremented. The cell with maximum ‘votes’ in the Accumulator matrix can be a possible point of existence of fixed reference of the object in the test image.
Building the R-Table
Choose a reference point y for the shape (typically chosen inside the shape). For each boundary point x, compute ɸ(x), the gradient direction and r = y – x as shown in the image. Store r as a function of ɸ. Notice that each index of ɸ may have many values of r. One can either store the co-ordinate differences between the fixed reference and the edge point ((xc – xij),( yc - yij)) or as the radial distance and the angle between them (rij , αij) . Having done this for each point, the R-table will fully represent the template object. Also, since the generation phase is invertible, we may use it to localise object occurrences at other places in the image.
i | ɸi | Rɸi |
---|---|---|
1 | 0 | (r11, α11) (r12, α12)… (r1n, α1n) |
2 | Δɸ | (r21, α21) (r22, α22)… (r1m, α1m) |
3 | 2Δɸ | (r31, α31) (r32, α32)… (r3k, α3k) |
... | ... | ... |
Object localization
For each edge pixel x in the image, find the gradient ɸ and increment all the corresponding points x+r in the accumulator array A (initialized to a maximum size of the image) where r is a table entry indexed by ɸ, i.e., r(ɸ). These entry points give us each possible position for the reference point. Although some bogus points may be calculated, given that the object exists in the image, a maximum will occur at the reference point. Maxima in A correspond to possible instances of the shape.
Generalization of scale and orientation
For a fixed orientation of shape, the accumulator array was two-dimensional in the reference point co-ordinates. To search for shapes of arbitrary orientation θ and scale s, these two parameters are added to the shape description. The accumulator array now consists of four dimensions corresponding to the parameters (y, s, θ). The R-table can also be used to increment this larger dimensional space since different orientations and scales correspond to easily computed transformations of the table. Denote a particular R-table for a shape S by R(ɸ). Simple transformations to this table will allow it to detect scaled or rotated instances of the same shape. For example if the shape is scaled by s and this transformation is denoted by Ts. then Ts[R(ɸ)] = sR(ɸ) i.e., all the vectors are scaled by s. Also, if the object is rotated by θ and this transformation is denoted by Tθ, then Tθ[R(ɸ)] = Rot{R[(ɸ-θ)mod2π],θ} i.e., all the indices are incremented by - θ modulo 2π, the appropriate vectors r are found, and then they are rotated by θ. Another property which will be useful in describing the composition of generalized Hough transforms is the change of reference point. If we want to choose a new reference point ỹ such that y-ỹ = z then the modification to the R-table is given by R(ɸ)+ z, i.e. z is added to each vector in the table.
Alternate way using pairs of edges
A pair of edge pixels can be used to reduce the parameter space. Using the R-table and the properties as described above, each edge pixel defines a surface in the four-dimensional accumulator space of a = (y, s, θ). Two edge pixels at different orientations describe the same surface rotated by the same amount with respect to θ. If these two surfaces intersect, points where they intersect will correspond to possible parameters a for the shape. Thus it is theoretically possible to use the two points in image space to reduce the locus in parameter space to a single point. However, the difficulties of finding the intersection points of the two surfaces in parameter space will make this approach unfeasible for most cases.
Composite shapes
If the shape S has a composite structure consisting of subparts S1, S2, .. SN and the reference points for the shapes S, S1, S2, .. SN are y, y1, y2, .. yn, respectively, then for a scaling factor s and orientation θ, the generalized Hough Transform Rs(ɸ) is given by . The concern with this transform is that the choice of reference can greatly affect the accuracy. To overcome this, Ballard has suggested smoothing the resultant accumulator with a composite smoothing template. The composite smoothing template H(y) is given as a composite convolution of individual smoothing templates of the sub-shapes. . Then the improved Accumulator is given by As = A*H and the maxima in As corresponds to possible instances of the shape.
Spatial Decomposition
Observing that the global Hough Transform can be obtained by the summation of local Hough transforms of disjoint sub-region, Heather and Yang[5] proposed a method which involves the recursive subdivision of the image into sub-images, each with their own parameter space, and organized in a quadtree structure. It results in improved efficiency in finding endpoints of line segments and improved robustness and reliability in extracting lines in noisy situations, at a slightly increased cost of memory.
Implementation[6]
Combining the above equations we have:
Constructing the R-Table:
- (0) Convert the sample shape image into an edge image using any edge detecting algorithm like Canny edge detector
- (1) Pick a reference point (e.g., (xc, yc))
- (2) Draw a line from the reference point to the boundary
- (3) Compute ɸ
- (4) Store the reference point (xc, yc) as a function of ɸ in R(ɸ) table.
Detection:
- (0) Convert the sample shape image into an edge image using any edge detecting algorithm like
Canny edge detectors.
- (1) Initialize the Accumulator table: A[xcmin. . . xcmax][ycmin. . . ycmax]
- (2) For each edge point (x, y)
- (2.1) Using the gradient angle ɸ, retrieve from the R-table all the (α, r) values indexed under ɸ.
- (2.2) For each (α,r), compute the candidate reference points:
- xc = x + r cos(α)
- yc = y + r sin(α)
- (2.3) Increase counters (voting):
- ++A([[x<sub>c</sub>]][yc])
- (3) Possible locations of the object contour are given by local maxima in A[xc][yc].
- If A[xc][yc] > T, then the object contour is located at (xc, yc)
General case:
Suppose the object has undergone some rotation ϴ and uniform scaling s:
- (x’, y’) --> (x’’, y’’)
- x" = (x’cos(ϴ) – y’sin(ϴ))s
- y" = (x’sin(ϴ) + y’cos(ϴ))s
- Replacing x’ by x" and y’ by y":
- xc = x – x" or xc = x - (x’cos(ϴ) – y’sin(ϴ))s
- yc = y – y" or yc = y - (x’sin(ϴ) + y’cos(ϴ))s
- (1) Initialize the Accumulator table: A[xcmin. . . xcmax][ycmin. . . ycmax][qmin . . .qmax][smin . . . smax]
- (2) For each edge point (x, y)
- (2.1) Using its gradient angle ɸ, retrieve all the (α, r) values from the R-table
- (2.2) For each (α, r), compute the candidate reference points:
- x' = r cos(α)
- y’ = r sin(α)
- for(ϴ = ϴmin; ϴ ≤ ϴmax; ϴ++)
- for(s = smin; s ≤ smax; s++)
- xc = x - (x’cos(ϴ) – y’sin(ϴ))s
- yc = y - (x’sin(ϴ) + y’cos(ϴ))s
- ++(A[xc][yc][ϴ][s])
- for(s = smin; s ≤ smax; s++)
- (3) Possible locations of the object contour are given by local maxima in A[xc][yc][ϴ][s]
- If A[xc][yc][ϴ][s] > T, then the object contour is located at (xc, yc), has undergone a rotation ϴ, and has been scaled by s.
Advantages and Disadvantages
Advantages:
- It is robust to partial or slightly deformed shapes (i.e., robust to recognition under occlusion).
- It is robust to the presence of additional structures in the image.
- It is tolerant to noise.
- It can find multiple occurrences of a shape during the same processing pass.
Disadvantages:
- It has substantial computational and storage requirements which become acute when object orientation and scale have to be considered.
Related work
Ballard suggested using orientation information of the edge decreasing the cost of the computation. Many efficient GHT techiques have been suggested such as the SC-GHT (Using slope and curvature as local properties).[7] Davis and Yam[8] also suggested an extension of Merlin’s work for orientation and scale invariant matching which complement’s Ballard’s work but does not include Ballard’s utilization of edge-slope information and composite structures
See also
- Hough Transform
- Randomized Hough Transform
- Radon Transform
- Template matching
- Outline of object recognition
References
- ↑ D.H. Ballard, "Generalizing the Hough Transform to Detect Arbitrary Shapes", Pattern Recognition, Vol.13, No.2, p.111-122, 1981
- ↑ Jaulin, L.; Bazeille, S. (2013). Image Shape Extraction using Interval Methods (PDF). In Proceedings of Sysid 2009, Saint-Malo, France.
- ↑ P. M . Merlin and D . J . Farber, “A parallel mechanism for detecting curves in pictures,” IEEE Trans . Comput . C24, 96-98 (1975)
- ↑ L. Davis, "Hierarchical Generalized Hough Transforms and Line Segment Based Generalized Hough Transforms", University of Texas Computer Sciences, Nov 1980
- ↑ J.A. Heather, Xue Dong Yang, "Spatial Decomposition of the Hough Transform", The 2nd Canadian Conference on Computer and Robot Vision, 2005.
- ↑ Ballard and Brown, section 4.3.4, Sonka et al., section 5.2.6
- ↑ A. A. Kassim, T. Tan, K. H. Tan, "A comparative study of efficient generalised Hough transform techniques", Image and Vision Computing, Volume 17, Issue 10, Pages 737-748, August 1999
- ↑ L. Davis and S. Yam, "A generalized hough-like transformation for shape recognition". University of Texas Computer Sciences, TR-134, Feb 1980.
External links
- OpenCV implementation of generalized Hough Transform http://docs.opencv.org/master/dc/d46/classcv_1_1GeneralizedHoughBallard.html
- Tutorial and implementation of generalized Hough transforms http://www.itriacasa.it/generalized-hough-transform/default.html
- Practical Generalized Hough transform implementation http://www.irit.fr/~Julien.Pinquier/Docs/Hough_transform.html
- FPGA implementation of generalized Hough transforms, IEEE Digital Library http://ieeexplore.ieee.org/xpl/login.jsp?tp=&arnumber=5382047&url=http%3A%2F%2Fieeexplore.ieee.org%2Fxpls%2Fabs_all.jsp%3Farnumber%3D5382047
- MATLAB implementation of generalized Hough Transform http://www.mathworks.com/matlabcentral/fileexchange/44166-generalized-hough-transform