US9424486B2 - Method of image processing - Google Patents

Method of image processing Download PDF

Info

Publication number
US9424486B2
US9424486B2 US14/254,851 US201414254851A US9424486B2 US 9424486 B2 US9424486 B2 US 9424486B2 US 201414254851 A US201414254851 A US 201414254851A US 9424486 B2 US9424486 B2 US 9424486B2
Authority
US
United States
Prior art keywords
image
template
data
target object
orientation field
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US14/254,851
Other versions
US20140226859A1 (en
Inventor
Anil Anthony Bharath
Jeffrey Ng Sing Kwong
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zebra Technologies Corp
Original Assignee
Cortexica Vision Systems Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Cortexica Vision Systems Ltd filed Critical Cortexica Vision Systems Ltd
Priority to US14/254,851 priority Critical patent/US9424486B2/en
Publication of US20140226859A1 publication Critical patent/US20140226859A1/en
Assigned to CORTEXICA VISION SYSTEMS LIMITED reassignment CORTEXICA VISION SYSTEMS LIMITED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: IMPERIAL INNOVATIONS LIMITED
Assigned to IMPERIAL INNOVATIONS LTD reassignment IMPERIAL INNOVATIONS LTD ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BHARATH, ANIL ANTHONY, NG SING KWONG, JEFFREY
Publication of US9424486B2 publication Critical patent/US9424486B2/en
Application granted granted Critical
Assigned to ZEBRA TECHNOLOGIES CORPORATION reassignment ZEBRA TECHNOLOGIES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CORTEXICA VISION SYSTEMS LIMITED
Assigned to JPMORGAN CHASE BANK, N.A. reassignment JPMORGAN CHASE BANK, N.A. SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LASER BAND, LLC, TEMPTIME CORPORATION, ZEBRA TECHNOLOGIES CORPORATION
Assigned to ZEBRA TECHNOLOGIES CORPORATION, LASER BAND, LLC, TEMPTIME CORPORATION reassignment ZEBRA TECHNOLOGIES CORPORATION RELEASE OF SECURITY INTEREST - 364 - DAY Assignors: JPMORGAN CHASE BANK, N.A.
Assigned to JPMORGAN CHASE BANK, N.A. reassignment JPMORGAN CHASE BANK, N.A. SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ZEBRA TECHNOLOGIES CORPORATION
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • G06K9/6202
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06K9/00228
    • G06K9/00268
    • G06K9/52
    • G06T7/0044
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • G06T7/74Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation

Definitions

  • the present application relates to a method of identifying a target object in an image using image processing. It further relates to a system and computer software implementing the method.
  • template-matching methods where a learned image template is compared (matched) against all possible locations in the image.
  • Simple template matching methods operate on luminance values of the template and image and compute a distance metric based on normalized cross-correlation. Normalized cross-correlation is invariant to global and linear illumination changes but not to local non-homogenous changes. It is also not invariant to intra-class variations of a family of objects such as relative luminance values between different regions of the object.
  • Template matching between binary edge maps of the image and shape has led to faster, and more selective discrimination of shapes than that using simply the pixel values.
  • European patent application no. EP 1193642 A1 describes a method for recognizing a user defined model object within an image.
  • the model object is searched for in an image by generating a match metric using a normalized cross-correlation technique.
  • a method of identifying a target object in an image including a plurality of image features comprises: storing template data, the template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object; receiving image data representing the image; processing the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features; processing the image data to generate image confidence data based on at least one characteristic for use in identifying the target object in a given image, the characteristic being indicative of an increased likelihood that at least one part of the given image represents at least part of the target object, relative to other parts of the given image; processing the image orientation field using the template orientation field and the image confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and using the match metric to determine whether or not the target object has been identified in the image.
  • the method described in the present application allows a target object to be identified in images, whether still images or video images, both quickly and efficiently.
  • image confidence data allows the target object to be identified with greater certainty, and thus with more accuracy.
  • the image confidence data may be based on one or more characteristics related to the target object so that matching data of a higher quality may be generated.
  • a method of identifying a target object in an image including a plurality of image features comprises: storing template data, the template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object; receiving template confidence data indicating a likelihood that at least one part of the template accurately represents at least part of the target object; receiving image data representing the image; processing the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features; processing the image orientation field using the template orientation field and the template confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and using the match metric to determine whether or not the target object has been identified in the image.
  • the template confidence data is based on known parameters related to the target object and/or other features of the image.
  • the template confidence data is used to improve a quality of matching with the template data such that the target object may be more accurately and reliably identified in the image.
  • apparatus arranged to perform the method of the present invention, computer software arranged to perform the method of the present invention and a data carrier storing the computer software of the present invention.
  • FIG. 1 shows schematically an apparatus for performing the method of the present invention.
  • FIG. 2 shows a flow diagram giving an overview of the method according to one or more embodiments of the present invention.
  • FIG. 3 shows schematically a method of generating a template orientation field, in accordance with an embodiment of the present invention.
  • FIG. 4 shows schematically a method of generating a template orientation field, in accordance with a different embodiment of the present invention.
  • FIGS. 5 a and 5 b show exemplary masks in accordance with an embodiment of the present invention.
  • FIGS. 5 c , 5 d , 5 e show exemplary orientation fields in accordance with the invention.
  • FIG. 6 shows schematically a method of generating a template phase map in accordance with an embodiment of the present invention.
  • FIG. 7 shows schematically a method of generating a match metric in accordance with an embodiment of the present invention.
  • FIG. 8 shows schematically a method of generating a match metric in accordance with a different embodiment of the present invention.
  • Embodiments of the present invention provide a method of identifying a target object in an image.
  • the target object is a human face, however, it is envisaged that the target object may be an alternative object.
  • the method is performed by computer software which the image processing system is arranged to run.
  • the computer software may be stored on a suitable data carrier such as a compact disc (CD).
  • FIG. 1 shows schematically a system 1 arranged to perform the method in accordance with at least one embodiment of the invention.
  • the system 1 includes a computer terminal CT which includes a central processing unit (CPU) 8 , memory 10 , a data storage device such as a hard disc drive 12 and I/O devices 6 which facilitate interconnection of the computer CT with an optional image capture device such as a camera 13 arranged to record image data in accordance with the present invention.
  • the CT may be arranged to receive image data from a remote image data source such as an image database.
  • the I/O devices 6 further facilitate interconnection of a display element 32 of a screen 28 via a screen I/O device 30 .
  • Operating system programs 14 are stored on the hard disc drive 12 and control, in a known manner, low level operation of the computer terminal CT.
  • Program files and data 20 are also stored on the hard disc drive 12 , and control, in a known manner, outputs to an operator via associated devices and output data stored on the hard disc drive 12 .
  • the associated devices include the display 32 as an element of the screen 28 , a pointing device (not shown) and keyboard (not shown), which receive input from, and output information to, the operator via further I/O devices (not shown).
  • Included in the program files 20 stored on the hard drive 12 is a database 22 for storing image data and data related to the image data, including data related to the characteristics described further below, and a database 24 for storing data related to the template data, including the template data.
  • FIG. 2 is a flow diagram indicating steps of the method according to an embodiment of the invention. This will be followed by a description of steps of the embodiment of the method in greater detail
  • template data is stored, in step S 1 , in the template data store 24 .
  • the template data represents a template orientation field indicative of an orientation of each of a plurality of features of a template object. Further details of the template data will be described later.
  • image data is received, in step S 2 .
  • the image data may be recorded by the camera 13 , or may have been received from a remote image data source.
  • the image data represents a digital image which includes a plurality of image features.
  • the image data may represent a colored image or a grayscale image.
  • Each of the image features represents, in the image, a boundary which separates a region of higher light intensity from a region of lower light intensity.
  • Each boundary corresponds with an edge or, in some embodiments, a line in the image.
  • a boundary may define an edge which separates the face from an image background.
  • the boundary may also separate two regions from each other, these two regions each having a similar light intensity, with the boundary being a line of different intensity from these two regions.
  • the boundary may define a structural line of a human face.
  • Template confidence data is received, in step S 3 , which indicates a likelihood that at least one part of the template represents at least part of the target object.
  • the template confidence data is relatively low.
  • the template confidence data is relatively high.
  • step S 4 the image data is processed to generate an image orientation field which indicates an orientation of each of the plurality of image features.
  • step S 5 the image data is processed, in step S 5 , to generate image confidence data which is based on at least one characteristic for use in identifying the target object in a given image. Further details of the characteristic, or characteristics, involved will be explained later.
  • step S 6 the image orientation field is processed using the template orientation field, the image confidence data and the template confidence data to generate, in step S 7 , a match metric which is indicative of an extent of matching between at least part of the template orientation field and at least part of the modified image orientation field.
  • step S 8 the match metric is used to determine whether or not the target object has been identified in the image.
  • At least one of the image data, template data, image confidence data and template confidence data is arranged according to a matrix format.
  • Divisive normalization is a refinement of local contrast normalization. Rather than normalizing image pixels by local image contrast or brightness over a neighborhood of pixels, image gradient values are normalized by the sum of their own squared values on a pixel-by-pixel basis.
  • the template object which is represented by the template orientation field of the template data, corresponds with the target object to be identified in the image.
  • the target object is a human face but it should be appreciated that other target objects may be desired, such as an animal or a vehicle, for example a car.
  • the template object has a plurality of features having similar properties to the image features.
  • the template orientation field may comprise a field of orientation vectors, arranged in this embodiment in a matrix format, which may have a rectangular grid form.
  • FIG. 3 shows schematically a method of generating the template orientation field according to an embodiment of the present invention.
  • Training image data representing a training image 40 is input data for use in generating the template orientation field.
  • the template data has similar properties to the image data.
  • the training image data is convolved, using a suitable convolution function 41 for example, a convolution function C k,l calculated in accordance with equation 1:
  • a m,n and B k,l are exemplary masks for convolving, and the convolution produces results for all values of k and l.
  • each of at least one x axis gradient mask 42 and at least one y axis gradient mask 44 are convolved with the training image data.
  • a gradient mask may alternatively be referred to as an image boundary filter.
  • the x axis gradient mask(s) 42 and the y axis gradient mask(s) 44 may be applied either in parallel or sequentially to the training image data.
  • the training image 40 is, in this embodiment, a mean image of a plurality of human face images. In this way, the template data, and consequently the template object, corresponds with the mean human face image.
  • Data of the at least one x axis gradient mask 42 and the at least one y axis gradient mask 44 is arranged according to a matrix format.
  • the gradient mask(s) 42 , 44 correspond with different possible boundary edges which may be identified in the image, in the x axis or y axis direction, appropriately.
  • each x gradient mask 44 may correspond with a different pattern of higher and lower light intensity regions which are separated by a boundary running at least partly in the x axis direction.
  • each y gradient mask 44 may correspond with a different pattern of higher and lower light intensity regions, separated by a boundary running at least partly in the y axis direction.
  • x axis and y axis gradient masks 42 , 44 yields gradient direction vectors for x axis and y axis features, respectively. These gradient direction vectors are converted into orientation vectors by wrapping their angles to the range from ⁇ /2 to ⁇ /2. Alternatively, a double angle representation may be used to increase the selectivity in generating the orientation vectors.
  • the output of each of these two convolutions may comprise a plurality of orientation vectors corresponding with a plurality of features of the training image along the x axis and the y axis, accordingly, which are similar to the image features.
  • These convolution outputs are each squared, using a squaring function 46 , and then summed together, using a summation function 48 , with a tuneable constant parameter ⁇ , the value of which is determined according to an amount of image noise identified in the image data.
  • the parameter may preferably be set at a value of less than 25%, for example 10% of the maximum gradient of the image data.
  • a square root function 49 is applied to the output of the summation function 48 , and the result of the summation function 48 is used to divisively normalize, using a divisive normalization function 50 , the output data of each of the convolution functions 41 , to produce an x axis template orientation field component t x m,n and a y axis template orientation field component t y m,n .
  • the length of each vector which may otherwise be referred to as the gradient intensity, in the appropriate template orientation field component is thus normalized by the sum of itself and the parameter ⁇ .
  • the two template orientation field components together form a template orientation field t m,n where m and n are template coordinates referring to values of a vector matrix of the template data. These coordinates may be defined with the origin at the center of the template orientation field.
  • FIG. 4 shows schematically an alternative method of generating the template orientation field using a steerable quadrature filter, according to a different embodiment of the present invention.
  • This alternative method are similar to the method described previously using FIG. 3 , and such features will be referred to using the same reference numerals, incremented by 100; corresponding descriptions should be taken to apply also.
  • the training image data representing the training image 140 is convolved, using a convolution function 52 , separately with each of a pair of masks consisting of a symmetric mask and an antisymmetric mask.
  • the symmetric mask responds most strongly to local image structure that is exactly symmetric about an axis that is aligned with the mask direction of the boundary, and has overlapping spatial frequency support, in magnitude, to that of the mask.
  • the associated antisymmetric mask responds most strongly to local image structure that is exactly antisymmetric about an axis aligned with the mask direction of the boundary and has overlapping spatial frequency content, in magnitude, to that of the mask.
  • the masks are used in pairs and an exemplary 0 radian and vertically oriented symmetric mask and antisymmetric mask is shown, respectively, in FIGS.
  • each of these convolutions is wrapped by using an intermediate double angle representation.
  • the output of each convolution is then squared, using the square function 63 , and summed together using the summation function 64 .
  • the steps of generating the template orientation field described so far, using FIG. 4 are performed with respect to one of a total number K of direction channels.
  • Each direction channel provides phase-invariant information concerning identification of boundaries between different light intensity regions of the training image, corresponding with the features of the training image.
  • Each direction channel corresponds to a directional phase invariant unit. As illustrated in FIG.
  • a first directional channel 54 corresponds with 0 radians
  • a second directional channel 56 corresponds with (K ⁇ 1) ⁇ /K radians
  • a further number of K ⁇ 2 directional channels 58 may be utilized, indicated by dashed lines, which correspond with the formula k ⁇ /K radians, where k refers to a certain one of K ⁇ 2 further directional channels.
  • Each direction channel can provide phase-invariant information concerning the boundaries present in the training image 140 and the plurality of direction channels together encode the direction of local energy in the training image 140 . More channels provide a better ability to distinguish between shape boundaries of the training image 140 that have sharp curvatures.
  • each direction channel is fed through a first 60 and a second 62 path. Following the first path 60 , the output of the direction channel is unchanged and is summed with the unchanged output of each of the other direction channels, using the summation function 65 . The output of this summation is fed through the square root function 66 and then added to the constant ⁇ , which was described earlier in further detail, using a further summation function 67 . This summed result is used for divisive normalization upon the x and y axis processing channels as described below.
  • each direction channel is fed through the square-root function 68 , and then is split into an x axis channel and a y axis channel.
  • the total number K of x axis channels are weighted with a gain function W k C , as calculated according to equation 2, with k taking integer values from 1 up to K.
  • W k C cos(2 ⁇ ( k ⁇ 1)/ K ) (2)
  • the total number of K of y-channels are weighted with a gain function W k S , as calculated according to equation 3, with k taking integer values from 1 up to K.
  • W k S sin(2 ⁇ ( k ⁇ 1)/ K ) (3)
  • the weighted x channels are summed, using the summation function 69 and divisively normalized with the output of the first path 60 following summation with the constant ⁇ using the divisive normalizing function 70 .
  • the length of each vector otherwise referred to as the steered energy of the template data, is normalized by the sum of the energy of each basis filter at that position in the template and the tuneable parameter ⁇ to yield the x axis template orientation field component t x m,n .
  • the output of each directional unit is modified by dividing it by the square root of the sum of the squared magnitude outputs at each location in the image. The purpose of this is that each feature measurement used for subsequent detection is normalized by other feature measures that share similar properties.
  • the weighted y channels are summed, using the summation function 69 and divisively normalized with the output of the first path 60 following summation with the constant ⁇ using the divisive normalizing function 70 .
  • the length of each vector is normalized by the sum of the energy of each basis filter at that position in the template and the tuneable parameter ⁇ to yield the y axis template orientation field component t y m,n .
  • FIG. 5 c shows an exemplary template orientation field applied over a training image of a mean human face.
  • FIG. 5 d shows the same as FIG. 5 c , but without the training image.
  • FIG. 5 e shows an exemplary template orientation field with a cosine weighting from the center of the field.
  • the template orientation field will be affected by both the representation of the target object in the training image data and the background near the boundary of the target object in the training image.
  • the template object corresponds to the target object.
  • the effect of background clutter in the template orientation field may be controlled.
  • the template confidence data indicates parts of the template data which have a greater likelihood of representing the template object than other parts of the template data.
  • the template confidence data may indicate a known parameter of the template object such as the size of the template object which may, for example, correspond to an expected extent of a human face (for example, a radial extent).
  • the template confidence data indicates a reliability of each vector in the template orientation field, for example due to background clutter, using a measure proportional to the distance between the position of the vector and the boundary of the template object. Such a measure may, alternatively, be taken between the vector position and any particular location of the template, such as the center point. Further known parameters relating to the template object may also be used to generate the template confidence data.
  • FIG. 6 shows schematically a method of generating template phase data, namely a template phase map t p m,n , indicative of a phase of at least one of the plurality of features of the template data.
  • the template phase map t p m,n includes a plurality of template phase values, each value corresponding with one of the plurality of features of the template data, each feature having certain template coordinate m,n values.
  • the values of the template phase map t p m,n are arranged in this embodiment in a matrix format.
  • FIG. 6 Generation of the template phase map t p m,n according to a steerable quadrature filter method will now be described with reference to FIG. 6 .
  • Some features and method steps described using FIG. 6 are similar to those described previously with reference to FIG. 4 ; corresponding reference numerals will be used, incremented by 200, and corresponding descriptions should be taken to apply here also.
  • Training image data for example the training image 240 , is convolved using the convolution function 252 , separately for each of the pair of symmetric and antisymmetric masks, as described previously. Such convolutions are performed for each of the directional channels.
  • the x axis template orientation field component t x m,n and the y axis template orientation field component t y m,n which correspond with one feature of the training image 240 are used to determine template angle data, namely a template angle a t m,n , of the one feature.
  • a template angle a t m,n is similarly determined for at least one, preferably each, of the plurality of features to generate, for example, a template angle matrix of the template angle a t m,n values.
  • the magnitude of each template angle a t m,n value is used to determine a magnitude of weighting applied to the corresponding feature of the result of the convolution by the symmetric and anti-symmetric masks at the template coordinates m,n for each directional channel.
  • the interpolation weight function 72 weights the value of each feature of the convolution outputs of each directional channel according to the magnitude of the corresponding template angle a t m,n For the directional channel with directional phase invariant unit of 0 radians, this results in a matrix of weight values, represented schematically in FIG. 6 as a grid 74 .
  • a matrix of weight values is generated for each directional channel. For example, for the directional channel 256 having the directional phase invariant unit (K ⁇ 1)pi/K radians, a matrix 75 of weight values is generated. A weight matrix is similarly generated for the further directional channels 258 , as indicated in FIG. 6 using dashed lines.
  • the training image 240 convolved with the symmetric mask is multiplied element-wise by the matrix 74 of weight values using a multiplication function 76 .
  • the multiplication function 76 is an element-wise multiplication of two data arrays.
  • the training image 240 convolved with the symmetric mask is multiplied 76 with the matrix of weight values generated using the angle a t m,n .
  • the training image 240 convolved with the anti-symmetric mask is multiplied 76 with the corresponding matrix of weight values generated using the appropriate interpolation weighting function.
  • each multiplication function 76 using the symmetric masks is summed using summation function 78 to provide a symmetric filter output.
  • the output of each multiplication function 76 corresponding with the antisymmetric masks is summed using summation function 80 , to provide an antisymmetric filter output.
  • the ratio of the antisymmetric filter output to the symmetric filter output is determined, and an arctangent function 82 is applied to the ratio to generate the template phase map t p m,n .
  • the template phase map t p m,n represents the symmetry of the responses for each feature of the template data in a direction perpendicular to that provided by the orientation fields.
  • the methods of generating the template orientation field described previously using FIG. 3 or FIG. 4 may be used to generate the image orientation field from the image data. Instead, however, for generating the image orientation field, the input data should be the received image data instead of the training image data. Consequently, the output of either of the methods described using FIG. 3 or FIG. 4 is an x axis image orientation field component i x k,l and a y axis image orientation field component i y k,l where k and l are image coordinates referring to vector values of the matrix of the image data. Together, these two generated image orientation field components form an image orientation field i k,l .
  • the generated image confidence data may be based on a plurality of different characteristics. Each characteristic is indicative of an increased likelihood that at least one part of a given image, in this example namely the image represented by the received image data, represents at least part of the target object, relative to other parts of the given image.
  • image confidence data may alternatively be used which is generated according to more than one of the different characteristics described below.
  • the characteristic indicated by the image confidence data relates to an extent of image errors expected in the image data and the method includes using the image confidence data to generate the match metric according to the expected extent of image data errors.
  • the image data may include data errors including, for example, errors incurred during transmission of the image data to the CT of the system 1 and errors incurred when recording the image data due to, for example, use of a poor quality camera or dust particles lying on a lens of the camera 13 which thus prevents recording of a true image.
  • error characteristic data indicating an extent of expected image errors is received by the system 1 .
  • the error characteristic data may include data of the quality of the camera and/or statistical data of a proportion of image data which is successfully received when transmitted over a given transmission means, such as a computer network. Further, image error corrections algorithms or image denoising algorithms may be used to generate the image confidence data.
  • this image confidence data may be used to successfully identify the target object from the image data where identification of the target object from the image orientation field without using image confidence data would have been unsuccessful.
  • the characteristic indicated by the image confidence data relates to a possible occlusion of part of the target object in the image and the method includes using the image confidence data to generate the match metric according to the possible occlusion.
  • the image includes an image of the target object
  • part of the target object may be occluded by other objects in the image, such that image data only represents part of the target object.
  • image occlusion characteristic data is received by the apparatus 1 .
  • the image occlusion characteristic data may represent a plurality of possible forms of an occlusion of the target object in an image, for example a shape occluding horizontally and/or vertically a middle part of the target object or a shape occluding horizontally and/or vertically a top, bottom and/or side part of the target object in the image. Further, if details of the type of objects present in the image scene recorded in the image are known, data may be available of shapes of objects commonly found in such a scene which may occlude the target object.
  • the image occlusion characteristic data may include data of a percentage of an area of the object in the image which may be occluded whilst still allowing the target object to be successfully identified in the image using the template data.
  • the image data is video data and the characteristic relates to a possible movement of the target object.
  • the method may include identifying a movement of the target object in the image using, for example, an image background subtraction algorithm or using data relating to a learned image background.
  • Movement of the target represented by the video image data may cause the shape of the target object represented by the video image data to change.
  • the identified movement of the target object is used to determine a modified shape of the target object which is modified from a shape of the target object represented by the template data.
  • the image data represents the image in color, such that the image data includes image color data, and the characteristic relates to at least one color indicative of the target object.
  • the characteristic relates to a color indicative of human skin.
  • object color characteristic data indicating at least one color of human skin is received and is applied to the image data to generate the image confidence data. Parts of the image data which indicate parts of the image having a human skin color are thus identified.
  • Processing the image confidence data based on the color characteristic with the image orientation field allows parts of the image data to be identified which have a greater likelihood to represent the target object than other parts of the image data. In this way, parts of the image data are assigned a higher certainty of may be representative of the target object than other parts of the image data.
  • the method of generating the template phase map t p m,n described previously using FIG. 6 may be used to generate image phase data, namely an image phase map i p k,l using the image data.
  • the image phase data is indicative of a phase of at least one of the plurality of image features.
  • the received image data is used instead of the training image data.
  • the image orientation field components i x k,l , i y k,l for the x and y axes are inputted.
  • the template angles a t m,n are not determined; image angle data, namely image angles a i k,l corresponding with at least one of the plurality of image features is determined instead.
  • a match metric is generated by processing the image orientation field using the template orientation field, the image confidence data and the template confidence data.
  • Matching between template data and image data may be generalized to matching between template and image vector fields, which are normalized by combinations of non-linear functions of the outputs of linear spatial operators at each pixel.
  • the matching may be performed using banks of cross-correlators that operate in parallel upon the normalized orientation field components. This has significance for speed and ease of implementation on a variety of hardware and software platforms.
  • the outputs of such correlators are generally mapped through non-linear functions, then weighted and combined to create accumulator spaces.
  • the accumulator spaces are similar in spirit to those of Hough Transform techniques, but are more selective to the specific shapes of the training. Detection therefore relies on thresholding or detecting local maxima in the accumulator spaces.
  • the image confidence data CI m,n includes x axis image confidence data CI x k,l and y axis image confidence data CI y m,n .
  • the template confidence data CT m,n includes x axis template confidence data CT x m,n and y axis template confidence data CT y m,n .
  • the x axis and y axis image confidence data and the x and y axis template confidence data is arranged according to a matrix format, where, in this example, each value of the confidence data matrix represents a likelihood that a part of the image represents the target object or a likelihood that a part of the template represents the template object, appropriately.
  • Each matrix element is a vector.
  • FIG. 7 shows schematically a method of generating a cross correlation match metric m c k,l at a position k,l in the image data, according to an embodiment of the invention.
  • the match metric m c k,l is based on the cross correlation calculation according to equation 4, which doesn't yet incorporate the image and template confidence data.
  • CCX is the convolution of the 2D-flipped x axis template orientation field component (t x m,n ) f with the x axis image orientation field component i x m,n
  • CCY is the convolution of the 2D-flipped y axis template field component (t u m,n ) f with the y axis image field component i y m,n .
  • Equation 4 namely the inner product and the two norm of difference can be generalized to a kernel K(a,b), as defined in equation 5.
  • Image confidence data CI m,n and template confidence data CT m,n which may otherwise be referred to as an image confidence map and a template confidence map, are introduced into equation 5 to improve the match metric, as defined in equation 6.
  • CCX becomes a convolution of (t x m,n ⁇ CT x m,n ) f and (i x k,l ⁇ CI x k,l )
  • CCY becomes a convolution of (t y m,n ⁇ CT y m,n ) f and (i y k,l ⁇ CI y k,l ).
  • Calculation of the product of the image confidence data CI k,l , either for the x axis or they axis, and the image orientation field, either for the x axis or they axis, respectively, involves applying the image confidence data to the image orientation field to generate the element-wise product.
  • the element-wise product may otherwise be referred to as a modified image orientation field MI k,l (including MI x k,l for the x axis and MI y k,l for the y axis) which indicates both an orientation of each of the plurality of image features and a likelihood that at least one part of the image orientation field represents at least part of the target object.
  • Calculation of the element-wise product of the template confidence data CT m,n either for the x axis or the y axis, and the template orientation field, either for the x axis or the y axis, respectively, involves applying the template confidence data to the template orientation field to generate the element-wise product.
  • the element-wise product may otherwise be referred to as a modified template orientation field MT m,n (MT x m,n for the x axis and MT y m,n for the y axis) which indicates both an orientation of each of the plurality of features of the template object and a likelihood that at least one part of the template orientation field represents at least part of the target object.
  • the 2D flipping converts the convolution operation into correlation that is required for template matching.
  • the operation may comprise a flip along the x-axis followed by a flip along the y-axis or vice-versa.
  • FIG. 8 shows schematically a method of generating a Euclidean distance based match metric m d k,l at a position in the image data, as an alternative to generating the cross correlation match metric described using FIG. 7 , according to a different embodiment of the present application.
  • the cross-correlation match metric described using FIG. 7 can give high values to undesired image orientation field neighborhoods containing relatively longer vectors than the template for unnormalized orientation fields or vectors whose lengths approach unity for normalized orientation fields.
  • the Euclidean distance based match metric m d k,l is more selective to matching of the template data by using a negated Euclidean Distance, which only gives a zero value match metric when the template and the image orientation field neighborhoods are exactly the same. Otherwise the Euclidean distance based match metric m d k,l produces negative values.
  • the Euclidean distance based match metric m d k,l without incorporation of the image and template confidence data, is defined according to equation 7.
  • ⁇ t m,n ⁇ i k+m,l ⁇ n ⁇ 2 ( t x m,n ⁇ i x k+m,l+n +) 2 +( t y m,n ⁇ i y k+m,l+n ) 2 (8)
  • CDX is defined according to equation 9 and CDY is defined according to equation 10.
  • CDX 2 CCX ⁇ CCXT ⁇ CCXI (9)
  • CDY 2 CCY ⁇ CCYT ⁇ CCYI (10)
  • the Euclidean distance based match metric can be generalized to a kernel K(a,b).
  • Image confidence data CI m,n and template confidence data CT m,n are introduced into equation 5 to improve the match metric such that:
  • CCX becomes a convolution of (t x m,n ⁇ CT x m,n ) f and (i x k,l ⁇ CI x k,l ), and CCY becomes a convolution of (t y m,n ⁇ CT y m,n ) f and (i y k,l ⁇ CI y k,l ).
  • CCXT becomes a convolution of the x axis image confidence data CI x k,l and the 2D flipped element-wise product of the x axis template confidence data CT x m,n and the element-wise squared x axis template orientation field (t x m,n ) 2 .
  • CCXI becomes a convolution of the 2D flipped x axis template confidence data (CT x m,n ) f and the element-wise product of the x axis image confidence data CI x k,l and the element-wise squared x axis image orientation field (i x k,l ) 2 .
  • CCXT is a convolution of ((t x m,n ) 2 ⁇ CT x m,n ) f and CI x k,l
  • CCXI is a convolution of ((i i x k,l ) 2 ⁇ CI x k,l ) and (CT x m,n ) f .
  • CCYT becomes a convolution of the y axis image confidence data CI y k,l and the 2D flipped element-wise product of the y axis template confidence data CT y m,n and the element-wise squared y axis template orientation field (t y m,n .
  • CCYI becomes a convolution of the 2D flipped y axis template confidence data (CT y m,n ) f and the element-wise product of the y axis image image confidence data CI y k,l and the element-wise squared y axis image orientation field (i y k,l ) 2 .
  • CCYT is a convolution of ((t y m,n ) 2 ⁇ CT y m,n ) f and CI y k,l
  • CCYI is a convolution of ((i y k,l ) 2 ⁇ CI y k,l ) and (CT y m,n ) f .
  • the x axis or y axis image confidence data CI x k,l , CI y k,l is set to a default value of 1.
  • the described convolution operations involving image confidence data use values of 0 outside of the bounds (in other words valid grid values) of the input masks. Since summation functions described above can use coordinates k+m and l+n that are outside the bounds of the input masks, values of 0 are used for these coordinates, according to a boundary handling technique.
  • the x axis or y axis template confidence data CT x m,n , CT y m,n is set to a default value of 1.
  • the absolute value of the phase difference between the template phase map t p m,n and a portion of the image phase map i p k+m,+l+n , centered on a location with coordinates k,l is multiplied by the absolute value of the cosine of the difference between the template angle a t m,n of each feature of the template orientation field and the image angle a i k+m,l+n of each feature of a portion of the image orientation field, to determine deformation data D k,l m,n .
  • the deformation data D k,l m,n is in this embodiment a matrix including a plurality of scalar deformation values, each value corresponding with the matching of one of the features of the template data at coordinate m,n to a feature from a portion of the image data centered at coordinate k,l.
  • Each deformation value corresponds with a phase difference between the template data and a portion of the image data.
  • the deformation data is indicative of a deformation of the template with respect to the image data. Larger deformation values may indicate a greater deformation to fit the template data to the portion of the image data, and vice versa. Such deformation values may be considered to represent a mismatch between the template data and the portion of the image data. The magnitude of the value indicates the extent of the mismatch.
  • deformation data D k,l m,n described so far relates to matching template data to image data of a still image.
  • the deformation data may also be applied to video image data. Accordingly, deformation data D k,l may be determined for a series of image frames of video image data, and a mismatch between the image data from consecutive frames may be used to indicate motion of a feature at coordinate k,l in the video image data.
  • the sum of the square of each scalar value of the deformation data D k,l m,n matrix is subtracted from the distance match metric m d k,l to determine the modified Euclidean distance match metric m e k,l .
  • the calculated match metric m e k,l advantageously takes into account any deformation between the template data and the image data, to allow a target object to be identified in the image data more accurately.
  • the deformation data D k,l m,n therefore provides for deformable or warping matching of a template with image data, rather than being limited to rigid matching.
  • use of the deformation data D k,l m,n provides for improved identification of a target object when phase characteristics of the target object are known. For example, if a target object is sought in image data representing a line drawing, the known phase characteristics can allow the system to be made more responsive to line drawings than to solid objects.
  • One of the match metrics described above may be computed for each vector of the matrix of the image data, as determined by the matrix coordinates, or for each vector value of a region of interest (ROI) which is determined in the image data by a user of by a computer vision algorithm.
  • ROI region of interest
  • each value of the calculated match metric is thresholded by a threshold value that can be set by a user or by a different object detector. For example, if one of the calculated match metrics has a value greater than the threshold, then it is determined that the target object has been identified in the image, with the center of the target object being represented by that match metric value. If, however, the match metric value is lower than the threshold, then that match metric does not identify the target object.
  • the image and/or template confidence data influences the values of the calculated match metrics, such that, in this example, the value of the match metric is increased where the corresponding part of the image has a relatively high likelihood of including the target object, as indicated by the confidence data. Further, the value of the match metric is decreased for parts of the image having a relatively low likelihood of including the target object. In this way the confidence data improves identification of the target object in the image.
  • one of the image confidence data and the template confidence data may not be used to generate the match metric.
  • only template confidence data may be used, or only image confidence data may be used.
  • an orientation field may be determined for each of a plurality of different training images representing, for example, different human faces, and a mean orientation field may be generated from the plurality of determined orientation fields.
  • the template and/or image phase map may be used to provide certainty data based upon detecting patterns of deformation. For example, higher deformation values localized to a region of image data compared with other regions of the image data during the matching of a template to a portion of the image data may be used to indicate occlusion of a target object in the image data.
  • image processing methods including methods for generating image orientation fields and template orientation fields methods for generating a match metric.
  • Further image processing methods may be applied in the method of the present invention, either additionally, or alternatively, to those described.
  • the image data of any of the embodiments described may include video image data.
  • the method of the present application may, for example, be utilized to provide a focus of attention operator.

Abstract

Disclosed embodiments relate to a systems and methods of identifying a target object in an image using image processing. Template data is stored which represents a template orientation field indicative of an orientation of each of a plurality of features of a template object. Image data may be received and processed to generate an image orientation field indicating an orientation corresponding to the plurality of image features. The image data may be further processed to generate image confidence data based on at least one characteristic for use in identifying the target object in a given image, and the characteristic may be indicative of an increased likelihood that at least one part of the given image represents at least part of the target object, relative to other parts of the given image. The image orientation field is processed using the template orientation field and the image confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field. The match metric may then be used to determine whether or not the target object has been identified in the image.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This application is a continuation of U.S. patent application Ser. No. 12/377,490 filed Jun. 12, 2009, which is a U.S. national stage entry under 35 U.S.C. §371 of PCT International Patent Application No. PCT/EP2007/058547, filed Aug. 16, 2007, which claims priority to GB Patent Application No. 0616293.7, filed Aug. 16, 2006. Each of the above-referenced patent applications is hereby incorporated by reference in its entirety.
BACKGROUND OF THE INVENTION
1. Field of the Invention
The present application relates to a method of identifying a target object in an image using image processing. It further relates to a system and computer software implementing the method.
2. Description of the Related Technology
Amongst the most generic of techniques for the detection of planar projections of shapes in digital images are the so-called template-matching methods, where a learned image template is compared (matched) against all possible locations in the image. Simple template matching methods operate on luminance values of the template and image and compute a distance metric based on normalized cross-correlation. Normalized cross-correlation is invariant to global and linear illumination changes but not to local non-homogenous changes. It is also not invariant to intra-class variations of a family of objects such as relative luminance values between different regions of the object.
An important issue is that of the property used to match between image and template. The most widely used techniques make use of the image pixel values themselves, and use cross-correlation between a template containing the image pixel values from one or more averaged examples of the shape.
Template matching between binary edge maps of the image and shape has led to faster, and more selective discrimination of shapes than that using simply the pixel values.
European patent application no. EP 1193642 A1 describes a method for recognizing a user defined model object within an image. The model object is searched for in an image by generating a match metric using a normalized cross-correlation technique.
It is an object of the application to provide improved object detection in image processing techniques.
SUMMARY
In accordance with one embodiment, there is provided a method of identifying a target object in an image including a plurality of image features, wherein the method comprises: storing template data, the template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object; receiving image data representing the image; processing the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features; processing the image data to generate image confidence data based on at least one characteristic for use in identifying the target object in a given image, the characteristic being indicative of an increased likelihood that at least one part of the given image represents at least part of the target object, relative to other parts of the given image; processing the image orientation field using the template orientation field and the image confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and using the match metric to determine whether or not the target object has been identified in the image.
The method described in the present application allows a target object to be identified in images, whether still images or video images, both quickly and efficiently.
Use of image confidence data allows the target object to be identified with greater certainty, and thus with more accuracy. The image confidence data may be based on one or more characteristics related to the target object so that matching data of a higher quality may be generated.
In accordance with a further embodiment, there is provided a method of identifying a target object in an image including a plurality of image features, wherein the method comprises: storing template data, the template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object; receiving template confidence data indicating a likelihood that at least one part of the template accurately represents at least part of the target object; receiving image data representing the image; processing the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features; processing the image orientation field using the template orientation field and the template confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and using the match metric to determine whether or not the target object has been identified in the image.
The template confidence data is based on known parameters related to the target object and/or other features of the image. The template confidence data is used to improve a quality of matching with the template data such that the target object may be more accurately and reliably identified in the image.
In accordance with yet further aspects of the present invention, there is provided apparatus arranged to perform the method of the present invention, computer software arranged to perform the method of the present invention and a data carrier storing the computer software of the present invention.
Further features and advantages will become apparent from the following description of preferred embodiments, given by way of example only, which is made with reference to the accompanying drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 shows schematically an apparatus for performing the method of the present invention.
FIG. 2 shows a flow diagram giving an overview of the method according to one or more embodiments of the present invention.
FIG. 3 shows schematically a method of generating a template orientation field, in accordance with an embodiment of the present invention.
FIG. 4 shows schematically a method of generating a template orientation field, in accordance with a different embodiment of the present invention.
FIGS. 5a and 5b show exemplary masks in accordance with an embodiment of the present invention.
FIGS. 5c, 5d, 5e show exemplary orientation fields in accordance with the invention.
FIG. 6 shows schematically a method of generating a template phase map in accordance with an embodiment of the present invention.
FIG. 7 shows schematically a method of generating a match metric in accordance with an embodiment of the present invention.
FIG. 8 shows schematically a method of generating a match metric in accordance with a different embodiment of the present invention.
DETAILED DESCRIPTION OF CERTAIN INVENTIVE EMBODIMENTS
Embodiments of the present invention provide a method of identifying a target object in an image. In the exemplary embodiments of the invention described below, the target object is a human face, however, it is envisaged that the target object may be an alternative object.
Image Processing System
In the embodiments of the method of the application to be described, the method is performed by computer software which the image processing system is arranged to run. The computer software may be stored on a suitable data carrier such as a compact disc (CD).
FIG. 1 shows schematically a system 1 arranged to perform the method in accordance with at least one embodiment of the invention. The system 1 includes a computer terminal CT which includes a central processing unit (CPU) 8, memory 10, a data storage device such as a hard disc drive 12 and I/O devices 6 which facilitate interconnection of the computer CT with an optional image capture device such as a camera 13 arranged to record image data in accordance with the present invention. Alternatively, the CT may be arranged to receive image data from a remote image data source such as an image database.
The I/O devices 6 further facilitate interconnection of a display element 32 of a screen 28 via a screen I/O device 30. Operating system programs 14 are stored on the hard disc drive 12 and control, in a known manner, low level operation of the computer terminal CT. Program files and data 20 are also stored on the hard disc drive 12, and control, in a known manner, outputs to an operator via associated devices and output data stored on the hard disc drive 12. The associated devices include the display 32 as an element of the screen 28, a pointing device (not shown) and keyboard (not shown), which receive input from, and output information to, the operator via further I/O devices (not shown). Included in the program files 20 stored on the hard drive 12 is a database 22 for storing image data and data related to the image data, including data related to the characteristics described further below, and a database 24 for storing data related to the template data, including the template data.
Method of Identifying a Target Object in an Image
An overview of methods according to embodiments of the invention will first be given, with reference to FIG. 2 which is a flow diagram indicating steps of the method according to an embodiment of the invention. This will be followed by a description of steps of the embodiment of the method in greater detail
With reference to FIG. 2, template data is stored, in step S1, in the template data store 24. The template data represents a template orientation field indicative of an orientation of each of a plurality of features of a template object. Further details of the template data will be described later.
Next, image data is received, in step S2. The image data may be recorded by the camera 13, or may have been received from a remote image data source. The image data represents a digital image which includes a plurality of image features. The image data may represent a colored image or a grayscale image.
Each of the image features represents, in the image, a boundary which separates a region of higher light intensity from a region of lower light intensity. Each boundary corresponds with an edge or, in some embodiments, a line in the image. For the example where the image includes a human face, a boundary may define an edge which separates the face from an image background. In embodiments of the invention to be described, the boundary may also separate two regions from each other, these two regions each having a similar light intensity, with the boundary being a line of different intensity from these two regions. For example, the boundary may define a structural line of a human face.
Template confidence data is received, in step S3, which indicates a likelihood that at least one part of the template represents at least part of the target object. In parts of the template where correspondence with the template orientation field is not highly indicative of a target object, the template confidence data is relatively low. In parts of the template where correspondence with the template orientation field is highly indicative of a target object, the template confidence data is relatively high.
In step S4, the image data is processed to generate an image orientation field which indicates an orientation of each of the plurality of image features.
Next, the image data is processed, in step S5, to generate image confidence data which is based on at least one characteristic for use in identifying the target object in a given image. Further details of the characteristic, or characteristics, involved will be explained later.
In step S6, the image orientation field is processed using the template orientation field, the image confidence data and the template confidence data to generate, in step S7, a match metric which is indicative of an extent of matching between at least part of the template orientation field and at least part of the modified image orientation field.
In step S8, the match metric is used to determine whether or not the target object has been identified in the image.
At least one of the image data, template data, image confidence data and template confidence data is arranged according to a matrix format.
In the description below, a divisive normalization function is used. Divisive normalization is a refinement of local contrast normalization. Rather than normalizing image pixels by local image contrast or brightness over a neighborhood of pixels, image gradient values are normalized by the sum of their own squared values on a pixel-by-pixel basis.
Template Data
The template object, which is represented by the template orientation field of the template data, corresponds with the target object to be identified in the image. In the embodiments to be described, the target object is a human face but it should be appreciated that other target objects may be desired, such as an animal or a vehicle, for example a car. The template object has a plurality of features having similar properties to the image features. The template orientation field may comprise a field of orientation vectors, arranged in this embodiment in a matrix format, which may have a rectangular grid form.
FIG. 3 shows schematically a method of generating the template orientation field according to an embodiment of the present invention. Training image data representing a training image 40 is input data for use in generating the template orientation field. The template data has similar properties to the image data. The training image data is convolved, using a suitable convolution function 41 for example, a convolution function Ck,l calculated in accordance with equation 1:
C k , l = m n A m , n · B k + m , l + n ( 1 )
where Am,n and Bk,l are exemplary masks for convolving, and the convolution produces results for all values of k and l.
In the present embodiment, each of at least one x axis gradient mask 42 and at least one y axis gradient mask 44 are convolved with the training image data. Such a gradient mask may alternatively be referred to as an image boundary filter. The x axis gradient mask(s) 42 and the y axis gradient mask(s) 44 may be applied either in parallel or sequentially to the training image data. The training image 40 is, in this embodiment, a mean image of a plurality of human face images. In this way, the template data, and consequently the template object, corresponds with the mean human face image.
Data of the at least one x axis gradient mask 42 and the at least one y axis gradient mask 44 is arranged according to a matrix format. The gradient mask(s) 42, 44 correspond with different possible boundary edges which may be identified in the image, in the x axis or y axis direction, appropriately.
Where a plurality of different x axis gradient masks 42 are applied to the training image data, each x gradient mask 44 may correspond with a different pattern of higher and lower light intensity regions which are separated by a boundary running at least partly in the x axis direction.
Similarly, where a plurality of different y axis gradient masks 44 are applied to the training image data, each y gradient mask 44 may correspond with a different pattern of higher and lower light intensity regions, separated by a boundary running at least partly in the y axis direction.
Application of the x axis and y axis gradient masks 42, 44 yields gradient direction vectors for x axis and y axis features, respectively. These gradient direction vectors are converted into orientation vectors by wrapping their angles to the range from −π/2 to π/2. Alternatively, a double angle representation may be used to increase the selectivity in generating the orientation vectors.
The output of each of these two convolutions may comprise a plurality of orientation vectors corresponding with a plurality of features of the training image along the x axis and the y axis, accordingly, which are similar to the image features. These convolution outputs are each squared, using a squaring function 46, and then summed together, using a summation function 48, with a tuneable constant parameter γ, the value of which is determined according to an amount of image noise identified in the image data. The parameter may preferably be set at a value of less than 25%, for example 10% of the maximum gradient of the image data.
A square root function 49 is applied to the output of the summation function 48, and the result of the summation function 48 is used to divisively normalize, using a divisive normalization function 50, the output data of each of the convolution functions 41, to produce an x axis template orientation field component tx m,n and a y axis template orientation field component ty m,n. The length of each vector, which may otherwise be referred to as the gradient intensity, in the appropriate template orientation field component is thus normalized by the sum of itself and the parameter γ. The two template orientation field components together form a template orientation field tm,n where m and n are template coordinates referring to values of a vector matrix of the template data. These coordinates may be defined with the origin at the center of the template orientation field.
FIG. 4 shows schematically an alternative method of generating the template orientation field using a steerable quadrature filter, according to a different embodiment of the present invention. Features of this alternative method are similar to the method described previously using FIG. 3, and such features will be referred to using the same reference numerals, incremented by 100; corresponding descriptions should be taken to apply also.
The training image data representing the training image 140 is convolved, using a convolution function 52, separately with each of a pair of masks consisting of a symmetric mask and an antisymmetric mask. The symmetric mask responds most strongly to local image structure that is exactly symmetric about an axis that is aligned with the mask direction of the boundary, and has overlapping spatial frequency support, in magnitude, to that of the mask. The associated antisymmetric mask responds most strongly to local image structure that is exactly antisymmetric about an axis aligned with the mask direction of the boundary and has overlapping spatial frequency content, in magnitude, to that of the mask. The masks are used in pairs and an exemplary 0 radian and vertically oriented symmetric mask and antisymmetric mask is shown, respectively, in FIGS. 5a and 5b . Further details may be found in “A Steerable Complex Wavelet Construction and Its Application to Image Denoising”, A. Bharath and J. Ng, IEEE Transactions on Image Processing. 14(7):948-959. July 2005, the contents of which are included herein by reference.
The output of each of these convolutions is wrapped by using an intermediate double angle representation. The output of each convolution is then squared, using the square function 63, and summed together using the summation function 64. The steps of generating the template orientation field described so far, using FIG. 4, are performed with respect to one of a total number K of direction channels. Each direction channel provides phase-invariant information concerning identification of boundaries between different light intensity regions of the training image, corresponding with the features of the training image. Each direction channel corresponds to a directional phase invariant unit. As illustrated in FIG. 4, a first directional channel 54 corresponds with 0 radians, a second directional channel 56 corresponds with (K−1)π/K radians and a further number of K−2 directional channels 58 may be utilized, indicated by dashed lines, which correspond with the formula kπ/K radians, where k refers to a certain one of K−2 further directional channels.
Each direction channel can provide phase-invariant information concerning the boundaries present in the training image 140 and the plurality of direction channels together encode the direction of local energy in the training image 140. More channels provide a better ability to distinguish between shape boundaries of the training image 140 that have sharp curvatures.
The output from each direction channel is fed through a first 60 and a second 62 path. Following the first path 60, the output of the direction channel is unchanged and is summed with the unchanged output of each of the other direction channels, using the summation function 65. The output of this summation is fed through the square root function 66 and then added to the constant γ, which was described earlier in further detail, using a further summation function 67. This summed result is used for divisive normalization upon the x and y axis processing channels as described below.
Following the second path 62, the output of each direction channel is fed through the square-root function 68, and then is split into an x axis channel and a y axis channel. The total number K of x axis channels are weighted with a gain function Wk C, as calculated according to equation 2, with k taking integer values from 1 up to K.
W k C=cos(2π(k−1)/K)  (2)
Similarly, the total number of K of y-channels are weighted with a gain function Wk S, as calculated according to equation 3, with k taking integer values from 1 up to K.
W k S=sin(2π(k−1)/K)  (3)
The weighted x channels are summed, using the summation function 69 and divisively normalized with the output of the first path 60 following summation with the constant γ using the divisive normalizing function 70. In this way the length of each vector, otherwise referred to as the steered energy of the template data, is normalized by the sum of the energy of each basis filter at that position in the template and the tuneable parameter γ to yield the x axis template orientation field component tx m,n. For the divisive normalization function, the output of each directional unit is modified by dividing it by the square root of the sum of the squared magnitude outputs at each location in the image. The purpose of this is that each feature measurement used for subsequent detection is normalized by other feature measures that share similar properties.
The weighted y channels are summed, using the summation function 69 and divisively normalized with the output of the first path 60 following summation with the constant γ using the divisive normalizing function 70. Similarly, in this way the length of each vector is normalized by the sum of the energy of each basis filter at that position in the template and the tuneable parameter γ to yield the y axis template orientation field component ty m,n.
The x axis template orientation field component tx m,n and the y axis template orientation field component ty m,n in combination form the template orientation field tm,n.
In accordance with this embodiment, FIG. 5c shows an exemplary template orientation field applied over a training image of a mean human face. FIG. 5d shows the same as FIG. 5c , but without the training image. FIG. 5e shows an exemplary template orientation field with a cosine weighting from the center of the field.
Template Confidence Data
The template orientation field will be affected by both the representation of the target object in the training image data and the background near the boundary of the target object in the training image. The template object corresponds to the target object. The effect of background clutter in the template orientation field may be controlled.
The template confidence data indicates parts of the template data which have a greater likelihood of representing the template object than other parts of the template data. The template confidence data may indicate a known parameter of the template object such as the size of the template object which may, for example, correspond to an expected extent of a human face (for example, a radial extent). In this way the template confidence data indicates a reliability of each vector in the template orientation field, for example due to background clutter, using a measure proportional to the distance between the position of the vector and the boundary of the template object. Such a measure may, alternatively, be taken between the vector position and any particular location of the template, such as the center point. Further known parameters relating to the template object may also be used to generate the template confidence data.
Template Phase Map
In accordance with an embodiment of the invention, FIG. 6 shows schematically a method of generating template phase data, namely a template phase map tp m,n, indicative of a phase of at least one of the plurality of features of the template data. The template phase map tp m,n includes a plurality of template phase values, each value corresponding with one of the plurality of features of the template data, each feature having certain template coordinate m,n values. The values of the template phase map tp m,n are arranged in this embodiment in a matrix format.
Generation of the template phase map tp m,n according to a steerable quadrature filter method will now be described with reference to FIG. 6. Some features and method steps described using FIG. 6 are similar to those described previously with reference to FIG. 4; corresponding reference numerals will be used, incremented by 200, and corresponding descriptions should be taken to apply here also.
Training image data, for example the training image 240, is convolved using the convolution function 252, separately for each of the pair of symmetric and antisymmetric masks, as described previously. Such convolutions are performed for each of the directional channels.
To generate the template phase map tp m,n, the x axis template orientation field component tx m,n and the y axis template orientation field component ty m,n which correspond with one feature of the training image 240 are used to determine template angle data, namely a template angle at m,n, of the one feature. A template angle at m,n is similarly determined for at least one, preferably each, of the plurality of features to generate, for example, a template angle matrix of the template angle at m,n values.
Using an interpolation weight function 72, the magnitude of each template angle at m,n value is used to determine a magnitude of weighting applied to the corresponding feature of the result of the convolution by the symmetric and anti-symmetric masks at the template coordinates m,n for each directional channel. Thus, the interpolation weight function 72 weights the value of each feature of the convolution outputs of each directional channel according to the magnitude of the corresponding template angle at m,n For the directional channel with directional phase invariant unit of 0 radians, this results in a matrix of weight values, represented schematically in FIG. 6 as a grid 74.
A matrix of weight values is generated for each directional channel. For example, for the directional channel 256 having the directional phase invariant unit (K−1)pi/K radians, a matrix 75 of weight values is generated. A weight matrix is similarly generated for the further directional channels 258, as indicated in FIG. 6 using dashed lines.
For the directional channel having a directional phase invariant unit of 0 radians, the training image 240 convolved with the symmetric mask is multiplied element-wise by the matrix 74 of weight values using a multiplication function 76. The multiplication function 76 is an element-wise multiplication of two data arrays. Similarly, for the other directional channels, including the directional channel with directional phase invariant unit (K−1)pi/K radians, the training image 240 convolved with the symmetric mask is multiplied 76 with the matrix of weight values generated using the angle at m,n.
For each directional channel, the training image 240 convolved with the anti-symmetric mask is multiplied 76 with the corresponding matrix of weight values generated using the appropriate interpolation weighting function.
For each directional channel, the output of each multiplication function 76 using the symmetric masks is summed using summation function 78 to provide a symmetric filter output. Similarly, the output of each multiplication function 76 corresponding with the antisymmetric masks is summed using summation function 80, to provide an antisymmetric filter output.
The ratio of the antisymmetric filter output to the symmetric filter output is determined, and an arctangent function 82 is applied to the ratio to generate the template phase map tp m,n. Thus, the template phase map tp m,n represents the symmetry of the responses for each feature of the template data in a direction perpendicular to that provided by the orientation fields.
Image Orientation Field
The methods of generating the template orientation field described previously using FIG. 3 or FIG. 4 may be used to generate the image orientation field from the image data. Instead, however, for generating the image orientation field, the input data should be the received image data instead of the training image data. Consequently, the output of either of the methods described using FIG. 3 or FIG. 4 is an x axis image orientation field component ix k,l and a y axis image orientation field component iy k,l where k and l are image coordinates referring to vector values of the matrix of the image data. Together, these two generated image orientation field components form an image orientation field ik,l.
Image Confidence Data
The generated image confidence data may be based on a plurality of different characteristics. Each characteristic is indicative of an increased likelihood that at least one part of a given image, in this example namely the image represented by the received image data, represents at least part of the target object, relative to other parts of the given image.
Generation of image confidence data according to a different characteristic will now be described for different embodiments. In further embodiments, image confidence data may alternatively be used which is generated according to more than one of the different characteristics described below.
i. Image Error Characteristic
In an embodiment of the present invention, the characteristic indicated by the image confidence data relates to an extent of image errors expected in the image data and the method includes using the image confidence data to generate the match metric according to the expected extent of image data errors.
The image data may include data errors including, for example, errors incurred during transmission of the image data to the CT of the system 1 and errors incurred when recording the image data due to, for example, use of a poor quality camera or dust particles lying on a lens of the camera 13 which thus prevents recording of a true image.
In an embodiment of the invention, error characteristic data indicating an extent of expected image errors is received by the system 1. The error characteristic data may include data of the quality of the camera and/or statistical data of a proportion of image data which is successfully received when transmitted over a given transmission means, such as a computer network. Further, image error corrections algorithms or image denoising algorithms may be used to generate the image confidence data.
By using the error characteristic data to generate image confidence data, this image confidence data may be used to successfully identify the target object from the image data where identification of the target object from the image orientation field without using image confidence data would have been unsuccessful.
ii. Image Occlusion Characteristic
In an alternative embodiment of the present invention, the characteristic indicated by the image confidence data relates to a possible occlusion of part of the target object in the image and the method includes using the image confidence data to generate the match metric according to the possible occlusion.
Where the image includes an image of the target object, it is possible that part of the target object may be occluded by other objects in the image, such that image data only represents part of the target object.
In one embodiment of the invention, image occlusion characteristic data is received by the apparatus 1. The image occlusion characteristic data may represent a plurality of possible forms of an occlusion of the target object in an image, for example a shape occluding horizontally and/or vertically a middle part of the target object or a shape occluding horizontally and/or vertically a top, bottom and/or side part of the target object in the image. Further, if details of the type of objects present in the image scene recorded in the image are known, data may be available of shapes of objects commonly found in such a scene which may occlude the target object.
Additionally, the image occlusion characteristic data may include data of a percentage of an area of the object in the image which may be occluded whilst still allowing the target object to be successfully identified in the image using the template data.
iii. Object Motion Characteristic
In a further alternative embodiment of the present invention, the image data is video data and the characteristic relates to a possible movement of the target object. In one embodiment, the method may include identifying a movement of the target object in the image using, for example, an image background subtraction algorithm or using data relating to a learned image background.
Movement of the target represented by the video image data may cause the shape of the target object represented by the video image data to change. The identified movement of the target object is used to determine a modified shape of the target object which is modified from a shape of the target object represented by the template data.
Without using the image confidence data based on the object motion characteristic, such movement, which consequently changes the shape of the target object to a different shape from that of the template object, would prevent successful identification of the target object in the video image data. Use of the image confidence data based on the object motion characteristic therefore allows the target object to be successfully identified in video image data.
iv. Object Color Characteristic
In another alternative embodiment of the present invention, the image data represents the image in color, such that the image data includes image color data, and the characteristic relates to at least one color indicative of the target object. In this example the characteristic relates to a color indicative of human skin.
In some embodiments, object color characteristic data indicating at least one color of human skin is received and is applied to the image data to generate the image confidence data. Parts of the image data which indicate parts of the image having a human skin color are thus identified.
Processing the image confidence data based on the color characteristic with the image orientation field allows parts of the image data to be identified which have a greater likelihood to represent the target object than other parts of the image data. In this way, parts of the image data are assigned a higher certainty of may be representative of the target object than other parts of the image data.
Image Phase Map
The method of generating the template phase map tp m,n described previously using FIG. 6 may be used to generate image phase data, namely an image phase map ip k,l using the image data. The image phase data is indicative of a phase of at least one of the plurality of image features. For generating the image phase map ip m,n, the received image data is used instead of the training image data. Further, instead of inputting the template orientation field components tx m,n, ty m,n for the x and y axes, the image orientation field components ix k,l, iy k,l for the x and y axes are inputted. Further, the template angles at m,n are not determined; image angle data, namely image angles ai k,l corresponding with at least one of the plurality of image features is determined instead.
Generation of Match Metric
As described previously, a match metric is generated by processing the image orientation field using the template orientation field, the image confidence data and the template confidence data.
Matching between template data and image data may be generalized to matching between template and image vector fields, which are normalized by combinations of non-linear functions of the outputs of linear spatial operators at each pixel.
The matching may be performed using banks of cross-correlators that operate in parallel upon the normalized orientation field components. This has significance for speed and ease of implementation on a variety of hardware and software platforms. The outputs of such correlators are generally mapped through non-linear functions, then weighted and combined to create accumulator spaces. The accumulator spaces are similar in spirit to those of Hough Transform techniques, but are more selective to the specific shapes of the training. Detection therefore relies on thresholding or detecting local maxima in the accumulator spaces.
In the embodiments to be described, at least one of the image confidence data and the template confidence data described previously is used. The image confidence data CIm,n includes x axis image confidence data CIx k,l and y axis image confidence data CIy m,n. The template confidence data CTm,n includes x axis template confidence data CTx m,n and y axis template confidence data CTy m,n. The x axis and y axis image confidence data and the x and y axis template confidence data is arranged according to a matrix format, where, in this example, each value of the confidence data matrix represents a likelihood that a part of the image represents the target object or a likelihood that a part of the template represents the template object, appropriately. Each matrix element is a vector.
FIG. 7 shows schematically a method of generating a cross correlation match metric mc k,l at a position k,l in the image data, according to an embodiment of the invention.
The match metric mc k,l is based on the cross correlation calculation according to equation 4, which doesn't yet incorporate the image and template confidence data.
m k , l c = m n t m , n , i k + m , l + n ( 4 )
The computation of the match metric mc k,l over the image, according to equation 4, can be broken into the sum of two convolution operations: CCX is the convolution of the 2D-flipped x axis template orientation field component (tx m,n)f with the x axis image orientation field component ix m,n and CCY is the convolution of the 2D-flipped y axis template field component (tu m,n)f with the y axis image field component iy m,n.
The operators of equation 4, namely the inner product and the two norm of difference can be generalized to a kernel K(a,b), as defined in equation 5.
m k , l = m n K ( t m , n , i k + m , l + n ) ( 5 )
Image confidence data CIm,n and template confidence data CTm,n which may otherwise be referred to as an image confidence map and a template confidence map, are introduced into equation 5 to improve the match metric, as defined in equation 6.
m k , l cf = m n CT m , n · CI k , l · K ( t m , n , i k + m , l + n ) ( 6 )
Thus, CCX becomes a convolution of (tx m,n·CTx m,n)f and (ix k,l·CIx k,l), and CCY becomes a convolution of (ty m,n·CTy m,n)f and (iy k,l·CIy k,l).
Calculation of the product of the image confidence data CIk,l, either for the x axis or they axis, and the image orientation field, either for the x axis or they axis, respectively, involves applying the image confidence data to the image orientation field to generate the element-wise product. The element-wise product may otherwise be referred to as a modified image orientation field MIk,l (including MIx k,l for the x axis and MIy k,l for the y axis) which indicates both an orientation of each of the plurality of image features and a likelihood that at least one part of the image orientation field represents at least part of the target object.
Calculation of the element-wise product of the template confidence data CTm,n either for the x axis or the y axis, and the template orientation field, either for the x axis or the y axis, respectively, involves applying the template confidence data to the template orientation field to generate the element-wise product. The element-wise product may otherwise be referred to as a modified template orientation field MTm,n (MTx m,n for the x axis and MTy m,n for the y axis) which indicates both an orientation of each of the plurality of features of the template object and a likelihood that at least one part of the template orientation field represents at least part of the target object.
The 2D flipping converts the convolution operation into correlation that is required for template matching. The operation may comprise a flip along the x-axis followed by a flip along the y-axis or vice-versa.
The advantage of breaking down the computation of the match metric into convolution operations is that fast optimized convolution implementations already exist in both software libraries and hardware and only minimal change to these implementations is required.
FIG. 8 shows schematically a method of generating a Euclidean distance based match metric md k,l at a position in the image data, as an alternative to generating the cross correlation match metric described using FIG. 7, according to a different embodiment of the present application.
The cross-correlation match metric described using FIG. 7 can give high values to undesired image orientation field neighborhoods containing relatively longer vectors than the template for unnormalized orientation fields or vectors whose lengths approach unity for normalized orientation fields.
The Euclidean distance based match metric md k,l is more selective to matching of the template data by using a negated Euclidean Distance, which only gives a zero value match metric when the template and the image orientation field neighborhoods are exactly the same. Otherwise the Euclidean distance based match metric md k,l produces negative values. The Euclidean distance based match metric md k,l, without incorporation of the image and template confidence data, is defined according to equation 7.
m k , l d = m n - t m , n - i k - m , l - n 2 ( 7 )
t m,n −i k+m,l−nμ2=(t x m,n −i x k+m,l+n+)2+(t y m,n −i y k+m,l+n)2  (8)
The computation of the match metric md k,l over the whole image involves a non-linear inner kernel, defined in equation 8, that can be broken down into the sum of two series of linear convolutions, md=CDX+CDY, where CDX and CDY can be obtained by the addition of two convolution operations each to CCX and CCY.
CDX is defined according to equation 9 and CDY is defined according to equation 10.
CDX=2CCX−CCXT−CCXI  (9)
CDY=2CCY−CCYT−CCYI  (10)
With reference to equations 5 and 6, the Euclidean distance based match metric can be generalized to a kernel K(a,b). Image confidence data CIm,n and template confidence data CTm,n, are introduced into equation 5 to improve the match metric such that:
CCX becomes a convolution of (tx m,n·CTx m,n)f and (ix k,l·CIx k,l), and CCY becomes a convolution of (ty m,n·CTy m,n)f and (iy k,l·CIy k,l).
CCXT becomes a convolution of the x axis image confidence data CIx k,l and the 2D flipped element-wise product of the x axis template confidence data CTx m,n and the element-wise squared x axis template orientation field (tx m,n)2. CCXI becomes a convolution of the 2D flipped x axis template confidence data (CTx m,n)f and the element-wise product of the x axis image confidence data CIx k,l and the element-wise squared x axis image orientation field (ix k,l)2. Namely, CCXT is a convolution of ((tx m,n)2·CTx m,n)f and CIx k,l and CCXI is a convolution of ((i ix k,l)2·CIx k,l) and (CTx m,n)f.
CCYT becomes a convolution of the y axis image confidence data CIy k,l and the 2D flipped element-wise product of the y axis template confidence data CTy m,n and the element-wise squared y axis template orientation field (ty m,n. CCYI becomes a convolution of the 2D flipped y axis template confidence data (CTy m,n)f and the element-wise product of the y axis image image confidence data CIy k,l and the element-wise squared y axis image orientation field (iy k,l)2. Namely, CCYT is a convolution of ((ty m,n)2·CTy m,n)f and CIy k,l, and CCYI is a convolution of ((iy k,l)2·CIy k,l) and (CTy m,n)f.
When generating the Euclidean distance based match metric without image confidence data, the x axis or y axis image confidence data CIx k,l, CIy k,l is set to a default value of 1. The described convolution operations involving image confidence data use values of 0 outside of the bounds (in other words valid grid values) of the input masks. Since summation functions described above can use coordinates k+m and l+n that are outside the bounds of the input masks, values of 0 are used for these coordinates, according to a boundary handling technique.
Similarly, generating the Euclidean distance based match metric without template confidence data, the x axis or y axis template confidence data CTx m,n, CTy m,n is set to a default value of 1.
In accordance with a further embodiment of the present application, the Euclidean distance match metric md k,l may be modified to obtain an alternative match metric md k,l, by using the template phase map tp m,n and the image phase map ip k,l generated as described above.
abs(t p m,n −i p k+m,l+nabs(cos(a t m,n −a i k+m,l+n))=D k,l m,n  (11)
Using equation 11, the absolute value of the phase difference between the template phase map tp m,n and a portion of the image phase map ip k+m,+l+n, centered on a location with coordinates k,l, is multiplied by the absolute value of the cosine of the difference between the template angle at m,n of each feature of the template orientation field and the image angle ai k+m,l+n of each feature of a portion of the image orientation field, to determine deformation data Dk,l m,n. If an angle of a feature of the portion of the image orientation field is the same as the corresponding angle of a feature of the template data, the cosine of the difference of the orientation angles equals 1, and thus this uses the entire value of the displacement provided by the difference of the template and image phase at the location of the feature. If, however, a template and an image angle are perpendicular to each other, the deformation cannot be resolved and there is no contribution of the displacement at the location of this feature to the match metric.
The deformation data Dk,l m,n is in this embodiment a matrix including a plurality of scalar deformation values, each value corresponding with the matching of one of the features of the template data at coordinate m,n to a feature from a portion of the image data centered at coordinate k,l. Each deformation value corresponds with a phase difference between the template data and a portion of the image data. The deformation data is indicative of a deformation of the template with respect to the image data. Larger deformation values may indicate a greater deformation to fit the template data to the portion of the image data, and vice versa. Such deformation values may be considered to represent a mismatch between the template data and the portion of the image data. The magnitude of the value indicates the extent of the mismatch.
The deformation data Dk,l m,n described so far relates to matching template data to image data of a still image. The deformation data may also be applied to video image data. Accordingly, deformation data Dk,l may be determined for a series of image frames of video image data, and a mismatch between the image data from consecutive frames may be used to indicate motion of a feature at coordinate k,l in the video image data.
In this embodiment, and using a similar method as that described using FIG. 8, a modified Euclidean distance match metric me k,l is generated using the deformation data Dk,l m,n. Accordingly, the non-linear inner kernel described above with reference to FIG. 8 is modified to equation 12:
m k , l e = m k , l d - m n ( D m , n k , l ) 2 . ( 12 )
In the non-linear inner kernel of equation 12, the sum of the square of each scalar value of the deformation data Dk,l m,n matrix is subtracted from the distance match metric md k,l to determine the modified Euclidean distance match metric me k,l. Thus, the calculated match metric me k,l advantageously takes into account any deformation between the template data and the image data, to allow a target object to be identified in the image data more accurately. The deformation data Dk,l m,n therefore provides for deformable or warping matching of a template with image data, rather than being limited to rigid matching. In particular, use of the deformation data Dk,l m,n provides for improved identification of a target object when phase characteristics of the target object are known. For example, if a target object is sought in image data representing a line drawing, the known phase characteristics can allow the system to be made more responsive to line drawings than to solid objects.
Identification of Target Object
One of the match metrics described above may be computed for each vector of the matrix of the image data, as determined by the matrix coordinates, or for each vector value of a region of interest (ROI) which is determined in the image data by a user of by a computer vision algorithm.
In order to determine whether the target object has been identified in the image data, each value of the calculated match metric is thresholded by a threshold value that can be set by a user or by a different object detector. For example, if one of the calculated match metrics has a value greater than the threshold, then it is determined that the target object has been identified in the image, with the center of the target object being represented by that match metric value. If, however, the match metric value is lower than the threshold, then that match metric does not identify the target object.
Use of the image and/or template confidence data influences the values of the calculated match metrics, such that, in this example, the value of the match metric is increased where the corresponding part of the image has a relatively high likelihood of including the target object, as indicated by the confidence data. Further, the value of the match metric is decreased for parts of the image having a relatively low likelihood of including the target object. In this way the confidence data improves identification of the target object in the image.
The above embodiments are to be understood as illustrative examples of the invention. Further embodiments of the invention are envisaged. For example, further characteristics of the target object may be used to generate the image confidence data. Also, further known parameters of the target object may be used to generate template confidence data.
In further embodiments of the present invention, one of the image confidence data and the template confidence data may not be used to generate the match metric. For example, only template confidence data may be used, or only image confidence data may be used.
The determination of an orientation field using a training image that may be a mean human face image has been described above. In alternative embodiments, an orientation field may be determined for each of a plurality of different training images representing, for example, different human faces, and a mean orientation field may be generated from the plurality of determined orientation fields.
The template and/or image phase map may be used to provide certainty data based upon detecting patterns of deformation. For example, higher deformation values localized to a region of image data compared with other regions of the image data during the matching of a template to a portion of the image data may be used to indicate occlusion of a target object in the image data.
Examples of image processing methods have been described, including methods for generating image orientation fields and template orientation fields methods for generating a match metric. Further image processing methods, known in the art, may be applied in the method of the present invention, either additionally, or alternatively, to those described. Further, the image data of any of the embodiments described may include video image data. In such an embodiment, the method of the present application may, for example, be utilized to provide a focus of attention operator.
It is to be understood that any feature described in relation to any one embodiment may be used alone, or in combination with other features described, and may also be used in combination with one or more features of any other of the embodiments, or any combination of any other of the embodiments. Furthermore, equivalents and modifications not described above may also be employed without departing from the scope of the invention, which is defined in the accompanying claims.

Claims (20)

What is claimed is:
1. A method of identifying a target object in an image including a plurality of image features, the method comprising:
storing template data, the template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object;
receiving image data representing the image;
processing the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features;
processing the image data to generate image confidence data based on at least one characteristic for use in identifying the target object in a given image, the characteristic being indicative of an increased likelihood that at least one part of the given image represents at least part of the target object, relative to other parts of the given image;
processing the image orientation field using the template orientation field and the image confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and
using the match metric to determine whether or not the target object has been identified in the image.
2. The method of claim 1, the method comprises applying the image confidence data to the image orientation field to generate a modified image orientation field indicating both an orientation corresponding to the plurality of image features and a likelihood that at least one part of the image orientation field represents at least part of the target object.
3. The method of claim 1, wherein the characteristic indicated by the image confidence data relates to an extent of image data errors expected in the image data, and wherein the method further comprises using the image confidence data to generate the match metric according to the expected extent of image data errors.
4. The method of claim 1, wherein the characteristic indicated by the image confidence data relates to a possible occlusion of part of the target object in the image, and wherein the method further comprises using the image confidence data to generate the match metric according to the possible occlusion.
5. The method of claim 1, wherein the image data is video image data and wherein the characteristic relates to a possible movement of the target object.
6. The method of claim 5, wherein the method further comprises using the characteristic relating to the possible movement of the target object and an identified movement of the target object to determine a modified shape of the target object, and wherein the target object is modified from a shape of the target object represented by the template data.
7. The method of claim 1, wherein the image data represents the image in color, and wherein the characteristic relates to at least one color indicative of the target object.
8. The method of claim 1, wherein generating the match metric comprises using deformation data indicative of a deformation of template data with respect to image data.
9. The method of claim 1, wherein at least one of the image data, the template data, the image confidence data, and deformation data used to generate the match metric, is arranged according to a matrix format.
10. The method of claim 1, wherein each of the image features represents, in the image, a boundary which separates a region of higher light intensity from a region of lower light intensity.
11. The method of claim 10, wherein the boundary separates two regions, each having a similar light intensity, from each other.
12. A system for identifying a target object in an image including a plurality of image features, the system comprising:
a data store comprising executable software;
a processor in data communication with the data store, the processor configured to execute the software and cause a computing device to:
store template data, said template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object;
receive image data representing the image;
process the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features;
process the image data to generate image confidence data based on at least one characteristic for use in identifying the target object in a given image, the characteristic being indicative of an increased likelihood that at least one part of the given image represents at least part of the target object, relative to other parts of the given image;
process the image orientation field using the template orientation field and the image confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and
use the match metric to determine whether or not the target object has been identified in the image.
13. A non-transitory computer-readable storage medium having computer readable instructions stored thereon, wherein, when executed, cause the computerized device to perform a method of identifying a target object in an image including a plurality of image features, the method comprising, the method comprising:
storing template data, said template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object;
receiving image data representing the image;
processing the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features;
processing the image data to generate image confidence data based on at least one characteristic for use in identifying the target object in a given image, the characteristic being indicative of an increased likelihood that at least one part of the given image represents at least part of the target object, relative to other parts of the given image;
processing the image orientation field using the template orientation field and the image confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and
using the match metric to determine whether or not the target object has been identified in the image.
14. A method of identifying a target object in an image including a plurality of image features, the method comprising:
storing template data, the template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object;
receiving template confidence data indicating a likelihood that at least one part of the template accurately represents at least part of the target object;
receiving image data representing the image;
processing the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features;
processing the image orientation field using the template orientation field and the template confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and
using the match metric to determine whether or not the target object has been identified in the image.
15. The method of claim 14, wherein the method further comprises applying the template confidence data to the template orientation field to generate a modified template orientation field indicating both an orientation of each of the plurality of features of the template object and a likelihood that at least one part of the template orientation field represents at least part of the target object.
16. The method of claim 14, wherein generating the match metric comprises using deformation data indicative of a deformation of template data with respect to image data.
17. The method of claim 14, wherein each of the image features represents a boundary in the image which separates a region of higher light intensity from a region of lower light intensity.
18. The method of claim 17, wherein the boundary separates two regions, each having a similar light intensity, from each other.
19. A system for identifying a target object in an image including a plurality of image features, the system comprising:
a data store comprising executable software;
a processor in data communication with the data store, the processor configured to execute the software and cause a computing device to:
store template data, the template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object;
receive template confidence data indicating a likelihood that at least one part of the template accurately represents at least part of the target object;
receive image data representing the image;
process the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features;
process the image orientation field using the template orientation field and the template confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and
use the match metric to determine whether or not the target object has been identified in the image.
20. A non-transitory computer-readable storage medium having computer readable instructions stored thereon, wherein, when executed, cause the computerized device to perform a method of identifying a target object in an image including a plurality of image features, the method comprising:
storing template data, the template data representing a template orientation field indicative of an orientation of each of a plurality of features of a template object;
receiving template confidence data indicating a likelihood that at least one part of the template accurately represents at least part of the target object;
receiving image data representing the image;
processing the image data to generate an image orientation field indicating an orientation corresponding to the plurality of image features;
processing the image orientation field using the template orientation field and the template confidence data to generate a match metric indicative of an extent of matching between at least part of the template orientation field and at least part of the image orientation field; and
using the match metric to determine whether or not the target object has been identified in the image.
US14/254,851 2006-08-16 2014-04-16 Method of image processing Active US9424486B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/254,851 US9424486B2 (en) 2006-08-16 2014-04-16 Method of image processing

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
GB0616293.7 2006-08-16
GBGB0616293.7A GB0616293D0 (en) 2006-08-16 2006-08-16 Method of image processing
PCT/EP2007/058547 WO2008020068A1 (en) 2006-08-16 2007-08-16 Method of image processing
US37749009A 2009-06-12 2009-06-12
US14/254,851 US9424486B2 (en) 2006-08-16 2014-04-16 Method of image processing

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
PCT/EP2007/058547 Continuation WO2008020068A1 (en) 2006-08-16 2007-08-16 Method of image processing
US12/377,490 Continuation US8718321B2 (en) 2006-08-16 2007-08-16 Method of image processing

Publications (2)

Publication Number Publication Date
US20140226859A1 US20140226859A1 (en) 2014-08-14
US9424486B2 true US9424486B2 (en) 2016-08-23

Family

ID=37081088

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/377,490 Active 2029-05-06 US8718321B2 (en) 2006-08-16 2007-08-16 Method of image processing
US14/254,851 Active US9424486B2 (en) 2006-08-16 2014-04-16 Method of image processing

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/377,490 Active 2029-05-06 US8718321B2 (en) 2006-08-16 2007-08-16 Method of image processing

Country Status (8)

Country Link
US (2) US8718321B2 (en)
EP (1) EP2064652B1 (en)
JP (1) JP2010500677A (en)
AT (1) ATE484034T1 (en)
AU (1) AU2007285683A1 (en)
DE (1) DE602007009705D1 (en)
GB (1) GB0616293D0 (en)
WO (1) WO2008020068A1 (en)

Families Citing this family (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5111255B2 (en) * 2008-06-20 2013-01-09 キヤノン株式会社 Image processing apparatus, image processing method, computer program, and recording medium
GB0813406D0 (en) * 2008-07-22 2008-08-27 Univ City Image analysis system and method
GB0820273D0 (en) 2008-11-05 2008-12-10 Imp Innovations Ltd Improvements in relation to image processing
US8977629B2 (en) * 2011-05-24 2015-03-10 Ebay Inc. Image-based popularity prediction
WO2013066601A1 (en) * 2011-10-17 2013-05-10 Kimmel Zebadiah M Method and apparatus for monitoring individuals while protecting their privacy
US9974466B2 (en) 2011-10-17 2018-05-22 Atlas5D, Inc. Method and apparatus for detecting change in health status
WO2013058978A1 (en) 2011-10-17 2013-04-25 Kimmel Zebadiah M Method and apparatus for sizing and fitting an individual for apparel, accessories, or prosthetics
US9330339B2 (en) 2012-06-11 2016-05-03 Hi-Tech Solutions Ltd. System and method for detecting cargo container seals
US9092667B2 (en) * 2012-08-27 2015-07-28 Symbol Technologies, Llc Arrangement for and method of reading forms in correct orientation by image capture
CN103077377B (en) * 2012-12-31 2015-07-29 清华大学 Based on the fingerprint correction method of field of direction distribution
US9639867B2 (en) 2013-05-01 2017-05-02 Cloudsight, Inc. Image processing system including image priority
US10223454B2 (en) 2013-05-01 2019-03-05 Cloudsight, Inc. Image directed search
US9569465B2 (en) 2013-05-01 2017-02-14 Cloudsight, Inc. Image processing
US10140631B2 (en) 2013-05-01 2018-11-27 Cloudsignt, Inc. Image processing server
US9665595B2 (en) * 2013-05-01 2017-05-30 Cloudsight, Inc. Image processing client
US9575995B2 (en) 2013-05-01 2017-02-21 Cloudsight, Inc. Image processing methods
US9830522B2 (en) 2013-05-01 2017-11-28 Cloudsight, Inc. Image processing including object selection
CN103500334B (en) * 2013-09-17 2016-08-31 中国人民解放军国防科学技术大学 A kind of method and apparatus asking for the overlapped fingerprint field of direction
US9940506B2 (en) * 2013-11-25 2018-04-10 Ehsan FAZL ERSI System and method for face recognition
US9600993B2 (en) 2014-01-27 2017-03-21 Atlas5D, Inc. Method and system for behavior detection
EP2980669B1 (en) * 2014-08-01 2017-09-20 AGCO Corporation Determining field characterisitics using optical recognition
US10013756B2 (en) 2015-03-13 2018-07-03 Atlas5D, Inc. Methods and systems for measuring use of an assistive device for ambulation
CN105827931B (en) * 2015-06-19 2019-04-12 维沃移动通信有限公司 It is a kind of based on the audio-frequency inputting method and device taken pictures
ES2946807T3 (en) 2015-09-01 2023-07-26 Dream It Get It Ltd Media Drive Recovery and Related Processes
CN105678834B (en) * 2016-02-26 2019-12-17 腾讯科技(深圳)有限公司 Method and device for distinguishing objects
CN106127713B (en) * 2016-06-30 2019-02-26 山东中安科创光电科技有限公司 Small and weak gas leakage target enhancing and Fast implementation
JP7083809B2 (en) 2016-08-02 2022-06-13 アトラス5ディー, インコーポレイテッド Systems and methods for identifying and / or identifying and / or pain, fatigue, mood, and intent with privacy protection
CN106296676A (en) * 2016-08-04 2017-01-04 合肥景昇信息科技有限公司 The object positioning method that view-based access control model is mutual
US11131767B2 (en) * 2017-06-22 2021-09-28 The Boeing Company Synthetic aperture radar mapping and registration systems and methods
CN109685058B (en) * 2017-10-18 2021-07-09 杭州海康威视数字技术股份有限公司 Image target identification method and device and computer equipment
CN108764297B (en) * 2018-04-28 2020-10-30 北京猎户星空科技有限公司 Method and device for determining position of movable equipment and electronic equipment
CN108596923B (en) * 2018-05-14 2020-10-16 江苏哈工智能机器人股份有限公司 Three-dimensional data acquisition method and device and electronic equipment
US10825148B2 (en) * 2018-11-29 2020-11-03 Adobe Inc. Boundary-aware object removal and content fill

Citations (62)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH03201078A (en) 1989-12-27 1991-09-02 Honda Motor Co Ltd Method and device for hough transform
JPH04349583A (en) 1991-05-27 1992-12-04 Nippon Telegr & Teleph Corp <Ntt> Generalized hough transform circuit
US5233670A (en) 1990-07-31 1993-08-03 Thomson Trt Defense Method and device for the real-time localization of rectilinear contours in a digitized image, notably for shape recognition in scene analysis processing
US5430810A (en) 1990-11-20 1995-07-04 Imra America, Inc. Real time implementation of the hough transform
EP0722149A1 (en) 1995-01-13 1996-07-17 STMicroelectronics S.r.l. Hough transform with fuzzy gradient and fuzzy voting
EP0762326A2 (en) 1995-09-12 1997-03-12 Matsushita Electric Industrial Co., Ltd. Object detecting apparatus in which the position of a planar object is estimated by using hough transform
JPH11351842A (en) 1998-06-12 1999-12-24 East Japan Railway Co Wheel shape measuring device
EP1017019A2 (en) 1998-12-31 2000-07-05 Eastman Kodak Company Method for automatic determination of main subjects in photographic images
US6173066B1 (en) 1996-05-21 2001-01-09 Cybernet Systems Corporation Pose determination and tracking by matching 3D objects to a 2D sensor
WO2001029770A2 (en) 1999-10-21 2001-04-26 Arch Development Corporation Detectnig asymetric abnormalities in chest radiography by contralateral and temporal subtraction technique using elastic matching
WO2001033497A1 (en) 1999-10-29 2001-05-10 Microsoft Corporation A system and method for face detection through geometric distribution of a non-intensity image property
US6263088B1 (en) 1997-06-19 2001-07-17 Ncr Corporation System and method for tracking movement of objects in a scene
JP3201078B2 (en) 1993-07-14 2001-08-20 三協アルミニウム工業株式会社 Partition method of building exterior wall material
US20010055425A1 (en) 2000-02-01 2001-12-27 Ming-Yee Chiu Method and apparatus for monitoring an analog meter
US6351573B1 (en) 1994-01-28 2002-02-26 Schneider Medical Technologies, Inc. Imaging device and method
US6356647B1 (en) 1996-07-19 2002-03-12 Telefonaktiebolaget Lm Ericsson Hough transform based method of estimating parameters
EP1193642A1 (en) 2000-09-27 2002-04-03 MVTec Software GmbH System and method for object recognition
WO2002029712A2 (en) 2000-10-05 2002-04-11 Coreco Imaging, Inc. System and method for pattern identification
US20020057838A1 (en) 2000-09-27 2002-05-16 Carsten Steger System and method for object recognition
WO2003023695A1 (en) 2001-09-13 2003-03-20 Honeywell International Inc. Near-infrared method and system for use in face detection
US6546137B1 (en) 1999-01-25 2003-04-08 Siemens Corporate Research, Inc. Flash system for fast and accurate pattern localization
US6556692B1 (en) 1998-07-14 2003-04-29 Daimlerchrysler Ag Image-processing method and apparatus for recognizing objects in traffic
US20030123713A1 (en) 2001-12-17 2003-07-03 Geng Z. Jason Face recognition system and method
US20030123736A1 (en) 2001-12-12 2003-07-03 Xun Xu Imlementation of hough transform and its application in line detection and video motion analysis
US6591004B1 (en) 1998-09-21 2003-07-08 Washington University Sure-fit: an automated method for modeling the shape of cerebral cortex and other complex structures using customized filters and transformations
GB2384067A (en) 2002-01-10 2003-07-16 Hewlett Packard Co Method of associating two record sets comprising a set of processor states and a set of notes
US20030198398A1 (en) 2002-02-08 2003-10-23 Haike Guan Image correcting apparatus and method, program, storage medium, image reading apparatus, and image forming apparatus
US20030235327A1 (en) 2002-06-20 2003-12-25 Narayan Srinivasa Method and apparatus for the surveillance of objects in images
EP1394727A1 (en) 2002-08-30 2004-03-03 MVTec Software GmbH Hierarchical component based object recognition
US6732046B1 (en) 2001-10-03 2004-05-04 Navigation Technologies Corp. Application of the hough transform to modeling the horizontal component of road geometry and computing heading and curvature
WO2004038659A2 (en) 2002-10-21 2004-05-06 Sarnoff Corporation Method and system for performing surveillance
WO2004084054A2 (en) 2003-03-21 2004-09-30 Queen's University At Kingston Method and apparatus for communication between humans and devices
JP2004279187A (en) 2003-03-14 2004-10-07 Mitsubishi Electric Corp Method for measuring shape
US6807286B1 (en) 2000-04-13 2004-10-19 Microsoft Corporation Object recognition using binary image quantization and hough kernels
US6829370B1 (en) 1999-09-03 2004-12-07 Honeywell International Inc. Near-IR human detector
US6850646B1 (en) 1997-12-31 2005-02-01 Cognex Corporation Fast high-accuracy multi-dimensional pattern inspection
EP1058909B1 (en) 1997-12-08 2005-02-02 Intel Corporation A new perceptual thresholding for gradient-based local edge detection
US6856698B1 (en) 1997-11-26 2005-02-15 Cognex Corporation Fast high-accuracy multi-dimensional pattern localization
EP1515277A1 (en) 2003-04-28 2005-03-16 Sony Corporation Image recognition device and method, and robot device
WO2005024708A1 (en) 2003-09-08 2005-03-17 Jiris Co., Ltd The pupil detection method and shape descriptor extraction method for a iris recognition, iris feature extraction apparatus and method, and iris recognition system and method using its
US6876779B2 (en) 2000-01-24 2005-04-05 Sony Côrporation Method and apparatus of reconstructing audio/video/image data from higher moment data
US20050074156A1 (en) 2003-01-30 2005-04-07 Holupka Edward J. Automatic detection of radioactive seeds for CT based post-planning for prostate seed implantation based on the hough transform
US20050105780A1 (en) 2003-11-14 2005-05-19 Sergey Ioffe Method and apparatus for object recognition using probability models
US20050152588A1 (en) 2003-10-28 2005-07-14 University Of Chicago Method for virtual endoscopic visualization of the colon by shape-scale signatures, centerlining, and computerized detection of masses
US20050169531A1 (en) 2004-01-30 2005-08-04 Jian Fan Image processing methods and systems
US20050169529A1 (en) 2004-02-03 2005-08-04 Yuri Owechko Active learning system for object fingerprinting
WO2005074314A1 (en) 2004-02-02 2005-08-11 Koninklijke Philips Electronics, N.V. Method and device using randomized hough transform for detecting radio systems with periodic emission pattern
WO2005098751A1 (en) 2004-04-08 2005-10-20 Mobileye Technologies Limited Crowd detection
US20050251347A1 (en) 2004-05-05 2005-11-10 Pietro Perona Automatic visual recognition of biological particles
US6975334B1 (en) 2003-03-27 2005-12-13 Systems Paving Method and apparatus for simulating the appearance of paving stone on an existing driveway
US6985620B2 (en) 2000-03-07 2006-01-10 Sarnoff Corporation Method of pose estimation and model refinement for video representation of a three dimensional scene
EP1301896B1 (en) 2000-06-27 2006-01-25 Solystic Segmentation of a postal object digital image by hough transform
US7016532B2 (en) 2000-11-06 2006-03-21 Evryx Technologies Image capture and identification system and process
WO2006036842A2 (en) 2004-09-24 2006-04-06 The University Of North Carolina At Chapel Hill Methods, systems, and computer program products for hierarchical registration between a blood vessel and tissue surface model for a subject and blood vessel and tissue surface image for the subject
US7035434B2 (en) 2000-12-11 2006-04-25 Texas Instruments Incorporated Hough transform based motion detection image recording system and method
US20060093998A1 (en) 2003-03-21 2006-05-04 Roel Vertegaal Method and apparatus for communication between humans and devices
US7054505B2 (en) 2000-01-05 2006-05-30 The United States Of America As Represented By The Secretary Of The Navy Composite Hough transform for multitarget multisensor tracking
US7352892B2 (en) 2003-03-20 2008-04-01 Micron Technology, Inc. System and method for shape reconstruction from optical images
US7444002B2 (en) 2004-06-02 2008-10-28 Raytheon Company Vehicular target acquisition and tracking using a generalized hough transform for missile guidance
US7486812B2 (en) 2003-11-25 2009-02-03 Icad, Inc. Shape estimates and temporal registration of lesions and nodules
JP4349583B2 (en) 2005-04-11 2009-10-21 興研株式会社 Hazardous gas exposure prevention device for anatomy laboratory
US7986820B2 (en) 2001-10-19 2011-07-26 Mitsubishi Electric Research Laboratories, Inc. Method for comparing features extracted from images of fingerprints

Patent Citations (74)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH03201078A (en) 1989-12-27 1991-09-02 Honda Motor Co Ltd Method and device for hough transform
US5233670A (en) 1990-07-31 1993-08-03 Thomson Trt Defense Method and device for the real-time localization of rectilinear contours in a digitized image, notably for shape recognition in scene analysis processing
US5430810A (en) 1990-11-20 1995-07-04 Imra America, Inc. Real time implementation of the hough transform
JPH04349583A (en) 1991-05-27 1992-12-04 Nippon Telegr & Teleph Corp <Ntt> Generalized hough transform circuit
JP3201078B2 (en) 1993-07-14 2001-08-20 三協アルミニウム工業株式会社 Partition method of building exterior wall material
US6351573B1 (en) 1994-01-28 2002-02-26 Schneider Medical Technologies, Inc. Imaging device and method
EP0741540B1 (en) 1994-01-28 2004-04-21 Schneider Medical Technologies, Inc. Imaging device and method
EP0722149A1 (en) 1995-01-13 1996-07-17 STMicroelectronics S.r.l. Hough transform with fuzzy gradient and fuzzy voting
US5825915A (en) 1995-09-12 1998-10-20 Matsushita Electric Industrial Co., Ltd. Object detecting apparatus in which the position of a planar object is estimated by using hough transform
EP0762326A2 (en) 1995-09-12 1997-03-12 Matsushita Electric Industrial Co., Ltd. Object detecting apparatus in which the position of a planar object is estimated by using hough transform
US6173066B1 (en) 1996-05-21 2001-01-09 Cybernet Systems Corporation Pose determination and tracking by matching 3D objects to a 2D sensor
US6356647B1 (en) 1996-07-19 2002-03-12 Telefonaktiebolaget Lm Ericsson Hough transform based method of estimating parameters
US6263088B1 (en) 1997-06-19 2001-07-17 Ncr Corporation System and method for tracking movement of objects in a scene
US6856698B1 (en) 1997-11-26 2005-02-15 Cognex Corporation Fast high-accuracy multi-dimensional pattern localization
EP1058909B1 (en) 1997-12-08 2005-02-02 Intel Corporation A new perceptual thresholding for gradient-based local edge detection
US6850646B1 (en) 1997-12-31 2005-02-01 Cognex Corporation Fast high-accuracy multi-dimensional pattern inspection
JPH11351842A (en) 1998-06-12 1999-12-24 East Japan Railway Co Wheel shape measuring device
US6556692B1 (en) 1998-07-14 2003-04-29 Daimlerchrysler Ag Image-processing method and apparatus for recognizing objects in traffic
US6591004B1 (en) 1998-09-21 2003-07-08 Washington University Sure-fit: an automated method for modeling the shape of cerebral cortex and other complex structures using customized filters and transformations
EP1017019A2 (en) 1998-12-31 2000-07-05 Eastman Kodak Company Method for automatic determination of main subjects in photographic images
US6546137B1 (en) 1999-01-25 2003-04-08 Siemens Corporate Research, Inc. Flash system for fast and accurate pattern localization
US6829370B1 (en) 1999-09-03 2004-12-07 Honeywell International Inc. Near-IR human detector
WO2001029770A2 (en) 1999-10-21 2001-04-26 Arch Development Corporation Detectnig asymetric abnormalities in chest radiography by contralateral and temporal subtraction technique using elastic matching
US6594378B1 (en) 1999-10-21 2003-07-15 Arch Development Corporation Method, system and computer readable medium for computerized processing of contralateral and temporal subtraction images using elastic matching
US6792135B1 (en) 1999-10-29 2004-09-14 Microsoft Corporation System and method for face detection through geometric distribution of a non-intensity image property
WO2001033497A1 (en) 1999-10-29 2001-05-10 Microsoft Corporation A system and method for face detection through geometric distribution of a non-intensity image property
US7054505B2 (en) 2000-01-05 2006-05-30 The United States Of America As Represented By The Secretary Of The Navy Composite Hough transform for multitarget multisensor tracking
US6876779B2 (en) 2000-01-24 2005-04-05 Sony Côrporation Method and apparatus of reconstructing audio/video/image data from higher moment data
US6845177B2 (en) 2000-02-01 2005-01-18 Setrix Aktiengesellschaft Method and apparatus for monitoring an analog meter
US20010055425A1 (en) 2000-02-01 2001-12-27 Ming-Yee Chiu Method and apparatus for monitoring an analog meter
US6985620B2 (en) 2000-03-07 2006-01-10 Sarnoff Corporation Method of pose estimation and model refinement for video representation of a three dimensional scene
US6807286B1 (en) 2000-04-13 2004-10-19 Microsoft Corporation Object recognition using binary image quantization and hough kernels
EP1301896B1 (en) 2000-06-27 2006-01-25 Solystic Segmentation of a postal object digital image by hough transform
EP1193642A1 (en) 2000-09-27 2002-04-03 MVTec Software GmbH System and method for object recognition
US20020057838A1 (en) 2000-09-27 2002-05-16 Carsten Steger System and method for object recognition
WO2002029712A2 (en) 2000-10-05 2002-04-11 Coreco Imaging, Inc. System and method for pattern identification
US7016532B2 (en) 2000-11-06 2006-03-21 Evryx Technologies Image capture and identification system and process
US7035434B2 (en) 2000-12-11 2006-04-25 Texas Instruments Incorporated Hough transform based motion detection image recording system and method
US20030053664A1 (en) 2001-09-13 2003-03-20 Ioannis Pavlidis Near-infrared method and system for use in face detection
US7027619B2 (en) 2001-09-13 2006-04-11 Honeywell International Inc. Near-infrared method and system for use in face detection
WO2003023695A1 (en) 2001-09-13 2003-03-20 Honeywell International Inc. Near-infrared method and system for use in face detection
US6732046B1 (en) 2001-10-03 2004-05-04 Navigation Technologies Corp. Application of the hough transform to modeling the horizontal component of road geometry and computing heading and curvature
US7986820B2 (en) 2001-10-19 2011-07-26 Mitsubishi Electric Research Laboratories, Inc. Method for comparing features extracted from images of fingerprints
US20030123736A1 (en) 2001-12-12 2003-07-03 Xun Xu Imlementation of hough transform and its application in line detection and video motion analysis
US20030123713A1 (en) 2001-12-17 2003-07-03 Geng Z. Jason Face recognition system and method
WO2003058496A2 (en) 2002-01-10 2003-07-17 Hewlett-Packard Company Accessing a remotely-stored data set and associating notes with that data set
GB2384067A (en) 2002-01-10 2003-07-16 Hewlett Packard Co Method of associating two record sets comprising a set of processor states and a set of notes
US20030198398A1 (en) 2002-02-08 2003-10-23 Haike Guan Image correcting apparatus and method, program, storage medium, image reading apparatus, and image forming apparatus
US20030235327A1 (en) 2002-06-20 2003-12-25 Narayan Srinivasa Method and apparatus for the surveillance of objects in images
US20040042661A1 (en) 2002-08-30 2004-03-04 Markus Ulrich Hierarchical component based object recognition
EP1394727A1 (en) 2002-08-30 2004-03-03 MVTec Software GmbH Hierarchical component based object recognition
WO2004038659A2 (en) 2002-10-21 2004-05-06 Sarnoff Corporation Method and system for performing surveillance
US20040119819A1 (en) 2002-10-21 2004-06-24 Sarnoff Corporation Method and system for performing surveillance
WO2004059573A2 (en) 2002-12-17 2004-07-15 Genex Technologies, Inc. Face recognition system and method
US20050074156A1 (en) 2003-01-30 2005-04-07 Holupka Edward J. Automatic detection of radioactive seeds for CT based post-planning for prostate seed implantation based on the hough transform
JP2004279187A (en) 2003-03-14 2004-10-07 Mitsubishi Electric Corp Method for measuring shape
US7352892B2 (en) 2003-03-20 2008-04-01 Micron Technology, Inc. System and method for shape reconstruction from optical images
US20060093998A1 (en) 2003-03-21 2006-05-04 Roel Vertegaal Method and apparatus for communication between humans and devices
WO2004084054A2 (en) 2003-03-21 2004-09-30 Queen's University At Kingston Method and apparatus for communication between humans and devices
US6975334B1 (en) 2003-03-27 2005-12-13 Systems Paving Method and apparatus for simulating the appearance of paving stone on an existing driveway
US20050213818A1 (en) 2003-04-28 2005-09-29 Sony Corporation Image recognition device and method, and robot device
EP1515277A1 (en) 2003-04-28 2005-03-16 Sony Corporation Image recognition device and method, and robot device
WO2005024708A1 (en) 2003-09-08 2005-03-17 Jiris Co., Ltd The pupil detection method and shape descriptor extraction method for a iris recognition, iris feature extraction apparatus and method, and iris recognition system and method using its
US20050152588A1 (en) 2003-10-28 2005-07-14 University Of Chicago Method for virtual endoscopic visualization of the colon by shape-scale signatures, centerlining, and computerized detection of masses
US20050105780A1 (en) 2003-11-14 2005-05-19 Sergey Ioffe Method and apparatus for object recognition using probability models
US7486812B2 (en) 2003-11-25 2009-02-03 Icad, Inc. Shape estimates and temporal registration of lesions and nodules
US20050169531A1 (en) 2004-01-30 2005-08-04 Jian Fan Image processing methods and systems
WO2005074314A1 (en) 2004-02-02 2005-08-11 Koninklijke Philips Electronics, N.V. Method and device using randomized hough transform for detecting radio systems with periodic emission pattern
US20050169529A1 (en) 2004-02-03 2005-08-04 Yuri Owechko Active learning system for object fingerprinting
WO2005098751A1 (en) 2004-04-08 2005-10-20 Mobileye Technologies Limited Crowd detection
US20050251347A1 (en) 2004-05-05 2005-11-10 Pietro Perona Automatic visual recognition of biological particles
US7444002B2 (en) 2004-06-02 2008-10-28 Raytheon Company Vehicular target acquisition and tracking using a generalized hough transform for missile guidance
WO2006036842A2 (en) 2004-09-24 2006-04-06 The University Of North Carolina At Chapel Hill Methods, systems, and computer program products for hierarchical registration between a blood vessel and tissue surface model for a subject and blood vessel and tissue surface image for the subject
JP4349583B2 (en) 2005-04-11 2009-10-21 興研株式会社 Hazardous gas exposure prevention device for anatomy laboratory

Non-Patent Citations (18)

* Cited by examiner, † Cited by third party
Title
"Srinivasa, N. ""Learning and Generlization of Noisy Mappings Using a Modified PROBART Neural Network"" IEEE Transactionson Signal Processing, vol. 45, Oct. 1997, pp. 2533-2550".
Artolazabal, Jar., et al., "3DSVHT: Extraction of 3D Linear Motion via Multi-view, Temporal Evidence Accumulation," ACIVS 2005, LNCS 3708, pp. 563-570, 2005.
Basalamah, Saleh, et al., "Contrast Marginalized Gradient Template Matching," ECCV 2004, LNCS 3023, pp. 417-429,2004, (c) Springer-Verlag Berlin Heidelberg 2004.
Bharath, Anil Anthony, et al., "A Steerable Complex Wavelet Construction and Its Application to Image Denoising," IEEE Transactions on Image Processing, vol. 14, No. 7, Jul. 2005, pp. 948-959.
Chandrasekaran, V., et al., "Facial Feature Detection Using Compact Vector-Field Canonical Templates," Systems, Man and Cybernetics and Simulation, IEEE, vol. 3, Oct. 12, 1997, pp. 2022-2027, XP010249240 ISBN: 0-7803-4053-1.
Chen. Teh-Chuan. et al., "An Efficient Randomized Algorithm for Detecting Circles," Computer Vision and Image Understanding, vol. 83, 172-191 (2001).
Crouzil, Alain, et al., "A New Correlation Criterion Based on Gradient Fields Similarity," Proceedings of the International Conference on Pattern Recognition, IEEE, vol. 1, 1996, pp. 632-636, XP002419454.
Froba, Bernhard, et al., "Real-Time Face Detection Using Edge-Orientation Matching," AVBPA 2001, LNCS 2091, pp. 78-83, XP001077539, (c) Springer-Verlag Berlin Heidelberg 2001.
International Preliminary Report on Patentability issued Feb. 26, 2009 on related application PCT/EP2007/058547 filed Aug. 16, 2007.
International Search Report issued Nov. 6, 2007 on related application PCT/EP2007/058547 filed Aug. 16, 2007.
Kimura, Akio, et al., "An Extension of the Generalized Hough Transform to Realize Affine-Invariant Two-dimensional (2D) Shape Detection," Proceedings 16th International Conference on Pattern Recognition, vol. 4, pp. 65-69, 2002, IEEE.
Kimura, Akio, et al., "Generalized Hough Transform to Be Extended as an Affine-Invariant Detector of Arbitrary Shapes," Electronics and Communications in Japan, Part 2, vol. 87, No. 6, 2004, pp. 58-68.
Lappas, Pelopidas, et al., "Feature Tracking in an Energy Maximization Framework," Proceedings of the Second International Conference on Machine Learning and Cybernetics, Xi'an, 2003, pp. 3109-3114, vol. 5, IEEE Cat. No. 03EX693.
Scharstein, Daniel, "Matching Images by Comparing their Gradient Fields," IEEE, vol. 1, Oct. 9, 1994, pp. 572-575, XP000515226, ISBN: 0-8186-6267-0.
Tian. Qi-Chuan, et al., "Fast Algorithm and Application of Hough Transform in Iris Segmentation," Proceedings of the Third International Conference on Machine Learning and Cybernetics, Shanghai, Aug. 26-29, 2004, 0-7803-8403-2/04, (c) 2004, IEEE Cat. No. 04EX826), pp. 3977-3980, vol. 7.
Yabuta, Yoshito, et al., "Binocular robot vision system with autonomous movement of viewpoint," Proc. of SPIE Int. Sco. Opt. Eng., vol. 5302, (c) 2004, pp. 61-68.
Yager, Neil, et al., "Coarse Fingerprint Registration Using Orientation Fields," EURASIP Journal on Applied Signal Processing 2005: 13, 2043-2053, (c) 2005 Hindawi Publishing Corporation, XP-002455428.
Zheshen, Wang, et al., "Building Detection and Recognition Via the Improved Hough Transform," Proc. Int. Corn put. Congr. Wavelet Anal. Appl. Act. Media Technol., vol. 2, 2004, pp. 1075-1080.

Also Published As

Publication number Publication date
GB0616293D0 (en) 2006-09-27
EP2064652A1 (en) 2009-06-03
EP2064652B1 (en) 2010-10-06
US20100226533A1 (en) 2010-09-09
AU2007285683A1 (en) 2008-02-21
US20140226859A1 (en) 2014-08-14
WO2008020068A1 (en) 2008-02-21
JP2010500677A (en) 2010-01-07
DE602007009705D1 (en) 2010-11-18
US8718321B2 (en) 2014-05-06
ATE484034T1 (en) 2010-10-15

Similar Documents

Publication Publication Date Title
US9424486B2 (en) Method of image processing
US11410323B2 (en) Method for training convolutional neural network to reconstruct an image and system for depth map generation from an image
US10325151B1 (en) Method of extracting image of port wharf through multispectral interpretation
Christiansen et al. Unsuperpoint: End-to-end unsupervised interest point detector and descriptor
US9968845B2 (en) Image processing device and image processing method, and program
US7623683B2 (en) Combining multiple exposure images to increase dynamic range
US8538077B2 (en) Detecting an interest point in an image using edges
KR101117837B1 (en) Multi-image feature matching using multi-scale oriented patches
US20190197693A1 (en) Automated detection and trimming of an ambiguous contour of a document in an image
US7912321B1 (en) Image registration with uncertainty analysis
CN110148157B (en) Picture target tracking method and device, storage medium and electronic equipment
Schmidt et al. COMPARATIVE ASSESSMENT OF POINT FEATURE DETECTORS AND DESCRIPTORS IN THE CONTEXT OF ROBOT NAVIGATION.
US20130051626A1 (en) Method And Apparatus For Object Pose Estimation
US20100086220A1 (en) Image registration using rotation tolerant correlation method
US20130148860A1 (en) Motion aligned distance calculations for image comparisons
US9007481B2 (en) Information processing device and method for recognition of target objects within an image
US20150186753A1 (en) Image Processing Apparatus, Image Processing Method, and Image Processing Program
US11227178B2 (en) Back-propagation image visual saliency detection method based on depth image mining
CA3136674C (en) Methods and systems for crack detection using a fully convolutional network
CN110084743B (en) Image splicing and positioning method based on multi-flight-zone initial flight path constraint
CN103700082B (en) Image split-joint method based on dual quaterion relative orientation
Shokouh et al. Ridge detection by image filtering techniques: A review and an objective analysis
US8126275B2 (en) Interest point detection
US9547913B2 (en) Information processing device, information processing apparatus, information processing method, and program
CN112070077B (en) Deep learning-based food identification method and device

Legal Events

Date Code Title Description
AS Assignment

Owner name: CORTEXICA VISION SYSTEMS LIMITED, UNITED KINGDOM

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:IMPERIAL INNOVATIONS LIMITED;REEL/FRAME:039160/0596

Effective date: 20121115

Owner name: IMPERIAL INNOVATIONS LTD, UNITED KINGDOM

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BHARATH, ANIL ANTHONY;NG SING KWONG, JEFFREY;REEL/FRAME:039160/0581

Effective date: 20090805

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: ZEBRA TECHNOLOGIES CORPORATION, ILLINOIS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CORTEXICA VISION SYSTEMS LIMITED;REEL/FRAME:051565/0503

Effective date: 20200101

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

AS Assignment

Owner name: JPMORGAN CHASE BANK, N.A., NEW YORK

Free format text: SECURITY INTEREST;ASSIGNORS:ZEBRA TECHNOLOGIES CORPORATION;LASER BAND, LLC;TEMPTIME CORPORATION;REEL/FRAME:053841/0212

Effective date: 20200901

AS Assignment

Owner name: LASER BAND, LLC, ILLINOIS

Free format text: RELEASE OF SECURITY INTEREST - 364 - DAY;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:056036/0590

Effective date: 20210225

Owner name: ZEBRA TECHNOLOGIES CORPORATION, ILLINOIS

Free format text: RELEASE OF SECURITY INTEREST - 364 - DAY;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:056036/0590

Effective date: 20210225

Owner name: TEMPTIME CORPORATION, NEW JERSEY

Free format text: RELEASE OF SECURITY INTEREST - 364 - DAY;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:056036/0590

Effective date: 20210225

AS Assignment

Owner name: JPMORGAN CHASE BANK, N.A., NEW YORK

Free format text: SECURITY INTEREST;ASSIGNOR:ZEBRA TECHNOLOGIES CORPORATION;REEL/FRAME:056471/0906

Effective date: 20210331

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8