US20110102553A1 - Enhanced real-time face models from stereo imaging - Google Patents

Enhanced real-time face models from stereo imaging Download PDF

Info

Publication number
US20110102553A1
US20110102553A1 US12/824,204 US82420410A US2011102553A1 US 20110102553 A1 US20110102553 A1 US 20110102553A1 US 82420410 A US82420410 A US 82420410A US 2011102553 A1 US2011102553 A1 US 2011102553A1
Authority
US
United States
Prior art keywords
face
model
region
camera
applying
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/824,204
Inventor
Peter Corcoran
Petronel Bigioi
Mircea C. Ionita
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fotonation Ltd
Original Assignee
Tessera Technologies Ireland Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US12/038,147 external-priority patent/US8509561B2/en
Application filed by Tessera Technologies Ireland Ltd filed Critical Tessera Technologies Ireland Ltd
Priority to US12/824,204 priority Critical patent/US20110102553A1/en
Publication of US20110102553A1 publication Critical patent/US20110102553A1/en
Priority to US14/673,246 priority patent/US9639775B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/97Determining parameters from multiple pictures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/593Depth or shape recovery from multiple images from stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/60Extraction of image or video features relating to illumination properties, e.g. using a reflectance or lighting model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/755Deformable models or variational models, e.g. snakes or active contours
    • G06V10/7557Deformable models or variational models, e.g. snakes or active contours based on appearance, e.g. active appearance models [AAM]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/15Processing image signals for colour aspects of image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/74Circuitry for compensating brightness variation in the scene by influencing the scene brightness using illuminating means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/08Indexing scheme for image data processing or generation, in general involving all processing steps from image acquisition to 3D model generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Definitions

  • FIGS. 1A and 1B illustrate an example of annotations used for the Yale B Database.
  • FIG. 2A illustrates variation between individuals.
  • FIG. 2B illustrates estimated albedo of the individuals of FIG. 2A .
  • FIG. 2C illustrates albedo eigen-textures with 95% energy preservation.
  • FIG. 3A illustrates a reference sample subset of images with various directional lighting effects.
  • FIG. 3B illustrate face samples from FIG. 3A with the contribution of directional lighting removed by filtering (see equation 5).
  • FIG. 4 illustrates images of FIG. 3B subtracted from images of FIG. 3A to yield a set of difference (residual) images.
  • FIG. 5 illustrates process steps to build a color extension of the combined DLS+ULS model for face recognition.
  • FIG. 6 illustrates a general architecture for real-time stereo video capture.
  • FIG. 7 illustrates an internal architecture for real-time stereo video capture.
  • FIG. 8 illustrates a stereo face image pair example.
  • FIG. 9 illustrates the Parallax Effect.
  • FIG. 10 illustrates a depth map result for the stereo image pair of FIG. 8 .
  • FIG. 11 illustrates a fitted AAM face model on the stereo pair of FIG. 8 .
  • FIG. 12 illustrates corresponding triangulated meshes for a fitted model.
  • FIG. 13 illustrates generating a 3D shape from 2D stereo data with triangulation-based warping.
  • FIG. 14A illustrates progressive blurring
  • FIG. 14B illustrates selective blurring
  • FIG. 15A illustrates Frontal Face, with simple Directional Lighting.
  • FIG. 15B illustrates Directional Lighting—Note Shadows from Eyelashes and Nose demonstrating sophisticated post-acquisition effects possible with 3D model.
  • FIG. 15C illustrates Directional Lighting—Note cheek regions is strongly shaded although it is to the foreground, demonstrating the selective application of the directional lighting effect to the cheek and eye regions. Here too we see the eyelash shadows.
  • FIG. 16 illustrates an estimated 3D profile from 2D stereo data using Thin Plate Spline—based warping.
  • a particular class of 2D affine models are involved in certain embodiments, known as active appearance models (AAM), which are relatively fast and are sufficiently optimal to be suitable for in-camera implementations.
  • AAM active appearance models
  • improvements are provided for example to (i) deal with directional lighting effects and (ii) make use of the full color range to improve accuracy and convergence of model to a detected face region.
  • stereo imaging provides improved model registration by using two real-time video images with slight variations in spatial perspective.
  • AAM models may comprise 2D affine models
  • the use of a real-time stereo video stream opens interesting possibilities to advantageously create full 3D face model from the 2D real-time models.
  • a method for performing face recognition and comparative analysis shows that results from improved models according to certain embodiments are significantly better than those obtained from a conventional AAM or from a conventional eigenfaces method for performing face recognition.
  • a differential stereo model is also provided which can be used to further enhance model registration and which offers the means to extend a 2D real-time model to a pseudo 3D model.
  • An approach is also provided for generating realistic 3D avatars based on a computationally reduced thin plate spline warping technique.
  • the method incorporates modeling enhancements also described herein.
  • Embodiments that involve the use of AAM models across a range of gaming applications is also provided.
  • T. F. Cootes see, e.g., T. F. Cootes, G. J. Edwards, and C. J. Taylor, “Active appearance models”, Lecture Notes in Computer Science, vol. 1407, pp. 484-, 1998, incorporated by reference
  • Statistical Models of Appearance represent both the shape and texture variations and the correlations between them for a particular class of objects.
  • Example members of the class of objects to be modeled are annotated by a number of landmark points. The shape is defined by the number of landmarks chosen to best depict the contour of the object of interest, in our case a person's face.
  • an AAM algorithm can be employed to fit the model to a new, unseen, image region.
  • the statistical model is linear in both shape and texture.
  • fitting the model to a new image region is a non-linear optimization process.
  • the fitting algorithm works by minimizing the error between a query image and the equivalent model-synthesized image.
  • the reference shape used to generate the texture vectors should be the same one for all models, i.e. either identity or directional lighting models.
  • Our goal is to determine specialized subspaces, such as the identity subspace or the directional lighting subspace.
  • FIG. 1 illustrates examples of annotations used for the Yale B Database.
  • each landmark point should have the same face geometry correspondence for all images.
  • the landmarks should predominantly target fiducial points, which permit a good description of facial geometry, allowing as well the extraction of geometrical differences between different individuals.
  • the facial textures corresponding to images of individuals in the Yale database with frontal illumination are represented in FIGS. 2A , 2 B and 2 C.
  • FIG. 2A illustrates variation between individuals.
  • FIG. 2B illustrates estimated albedo of the individuals.
  • FIG. 2C illustrates albedo eigen-textures with 95% energy preservation.
  • the identity model can now be generated from the albedo images based on the standard PCA technique.
  • the factor g contains both identity and directional lighting information.
  • the same reference shape may be used to obtain the new texture vectors g, which ensures that the previous and new texture vectors have all equal lengths.
  • FIG. 3A a random selection of faces is shown as a reference sample subset of images with various directional lighting effects.
  • the projection of the texture vectors g onto ULS gives the sets of optimal texture parameter vectors as in:
  • the back-projection stage returns the texture vector, optimally synthesized by the identity model.
  • the projection/back-projection process filters out all the variations which could not be explained by the identity model. Thus, for this case, directional lighting variations are filtered out by this process,
  • FIG. 3B illustrates face samples from FIG. 3A with the contribution of directional lighting removed by filtering (per equation 5, below).
  • the residual texture is further obtained as the difference between the original texture and the synthesized texture which retained only the identity information.
  • This residual texture normally retains the information other than identity.
  • the residual images give the directional lighting information, as illustrated at FIG. 4 which includes the images of FIG. 3B subtracted from the images of FIG. 3A to yield a set of difference (residual) images. These residuals are then modeled using PCA in order to generate a directional lighting subspace.
  • the shape model of the face As described above, three separate components of the face model have been generated. These are: (i) the shape model of the face, (ii) texture model encoding identity information, and (iii) the texture model for directional lighting.
  • the resulting texture subspaces are also orthogonal due to the approach described above.
  • the fusion between the two texture models can be realized by a weighted concatenation of parameters:
  • W lighting and Ws are two vectors of weights used to compensate for the differences in units between the two sets of texture parameters, and for the differences in units between shape and texture parameters, respectively.
  • the conventional AAM algorithm uses a gradient estimate built from training images and thus cannot be successfully applied to images where there are significant variations in illumination conditions.
  • the solution proposed by Batur et al. is based on using an adaptive gradient AAM (see, e.g., F. Kahraman, M. Gokmen, S. Darkner, and R. Larsen, “An active illumination and appearance (AIA) model for face alignment,” Computer Vision and Pattern Recognition, 2007. CVPR '07. IEEE Conference on, pp. 1-7, June 2007.
  • the gradient matrix is linearly adapted according to texture composition of the target image.
  • I 1 R + G + B 3 ( 5 ⁇ a )
  • I 2 R - B 2 ( 5 ⁇ b )
  • I 3 2 ⁇ G - R - B 4 ( 5 ⁇ c )
  • I1 represents the achromatic (intensity) component
  • I2 and I3 are the chromatic components.
  • Ohta's space the AAM search algorithm becomes more robust to variations in lighting levels and color distributions.
  • Tables I and II see also M. C. Ionita, P. Corcoran, and V. Buzuloiu, “On color texture normalization for active appearance models,” IEEE Transactions on Image Processing, vol. 18, issue 6, pp. 1372-1378, June 2009 and M. Ionita, “Advances in the design of statistical face modeling techniques for face recognition”, PhD Thesis, NUI Galway, 2009, which are incorporated by reference).
  • An advantageous AAM model may be used in face recognition.
  • face recognition there are a multitude of alternative applications for such models. These models have been widely used for face tracking (see, e.g., P. Corcoran, M. C. Ionita, I. Barcivarov, “Next generation face tracking technology using AAM techniques,” Signals, Circuits and Systems, ISSCS 2007, International Symposium on, Volume 1, p 1-4, 13-14 Jul. 2007, incorporated by reference), and for measuring facial pose and orientation.
  • FIG. 5 shows a process steps to build a color extension of the combined DLS+ULS model for face recognition.
  • Step 1 involves N persons and uniform lighting.
  • Step 2 involves N persons and 30 directional lighting conditions.
  • step 1 moves to uniform lighting space (ULS)
  • step 2 moves to the ULS.
  • step 3 involves an image difference fed by the step 2 images and the resultant N ⁇ 3-uniform light filtered images.
  • Step 4 involves N ⁇ 30 difference lighting images.
  • a directional lighting space (DLS) is achieved.
  • the right-hand side process diagram of FIG. 5 shows how the DLS subspace can be used to train a color ULS, implemented in the other color space.
  • Step 5 involves M persons in color with random lighting conditions. These are fed to the directional lighting subspace (DLS).
  • LDS directional lighting subspace
  • M identity filtered color images are achieved and fed at step 6 to an image difference along with the m persons in color with random lighting conditions.
  • Step 7 then involves M difference (identity) images.
  • texture PCA a uniform lighting (identity) color texture space (ULCTS) is achieved.
  • the example processes illustrated at FIG. 5 yield a full color ULS which retains the orthogonality with the DLS and when combined with it yields an enhanced AAM model incorporating shape+DLS+color ULS subspaces.
  • the color ULS has the same improved fitting characteristics as the color model (see, M. C. Ionita, P. Corcoran, and V. Buzuloiu, “On color texture normalization for active appearance models,” IEEE Transactions on Image Processing, vol. 18, issue 6, pp. 1372-1378, June 2009, incorporated by reference).
  • This combined model exhibits both improved registration and robustness to directional lighting.
  • the color AAM techniques based on RGB color space generally cannot compete with the conventional eigenface method of face recognition.
  • the I1I2I3 based models perform at least as well as the eigenface method, even when the model has been trained on a different database.
  • the I1I2I3 SChN model outperforms the eigenface method by at least 10% when the first 50 components are used. If we restrict our model to the first 5 or 10 components then the differential is about 20% in favor of the improved AAM model.
  • FIG. 6 An example of a general architecture of a stereo imaging system is illustrated at FIG. 6 , which shows two CMOS sensors and a VGA monitor connected to a power PC with Xilinx Virtex4 FPGA and DDR SDRAM.
  • the two CMOS sensors are connected to an FPGA which incorporates a PowerPC core and associated SDRAM.
  • Additional system components can be added to implement a dual stereo image processing pipeline (see, e.g., I. Andorko and P. Corcoran, “FPGA Based Stereo Imaging System with Applications in Computer Gaming”, at International IEEE Consumer Electronics Society's Games Innovations Conference 2009 (ICE-GIC 09), London, UK, incorporated by reference).
  • the development board is a Xilinx ML405 development board, with a Virtex 4 FPGA, a 64 MB DDR SDRAM memory, and a PowerPC RISC processor.
  • the clock frequency of the system is 100 MHz.
  • FIG. 7 An example internal architecture of the system in accordance with certain embodiments is illustrated at FIG. 7 , which shows two conversion blocks respectively coupled to camera units 1 and 2 .
  • the camera units 1 and 2 feed a PLB that feeds a VGA controller.
  • a DCR is connected to the camera units 1 and 2 , the VGA controller, an I2C controller and a Power PC.
  • the PLB is also coupled with DDR SDRAM.
  • the sensor used in this embodiment includes a 1 ⁇ 3 inch SXGA CMOS sensor made by Micron.
  • FIG. 8 illustrates a stereo face image pair example.
  • Parallax is an apparent displacement or difference of orientation of an object viewed along two different lines of sight, and is measured by the angle or semi-angle of inclination between those two lines.
  • the advantage of the parallax effect is that with the help of this, depth maps can be computed.
  • the computation in certain embodiments involves use of pairs of rectified images (see, K. Muhlmann, D. Maier, J. Hesser, R. Manner, “Calculating Dense Disparity Maps from Color Stereo Images, an Efficient Implementation”, International Journal of Computer Vision, vol. 47, numbers 1-3, pp. 79-88, April 2002, incorporated by reference).
  • This means that corresponding epipolar lines are horizontal and on the same height.
  • the search of corresponding pictures takes place in horizontal direction only in certain embodiments. For every pixel in the left image, the goal is to find the corresponding pixel in the right image, or vice-versa.
  • FIG. 9 illustrates the parallax effect.
  • is the average grayscale value of image window
  • N is the selected square window size
  • the first local variation calculation may be made over a 3 ⁇ 3 window. After this, the points with a value under a certain threshold are marked for further processing. The same operation is done for 5 ⁇ 5 and 7 ⁇ 7 windows as well. The size of each of the windows is stored for use in the depth map computation.
  • the operation to compute the depth map is the Sum of Absolute Differences for RGB images (SAD).
  • SAD Sum of Absolute Differences for RGB images
  • the value of SAD is computed for up to a maximum value of d on the x line. After all the SAD values have been computed, the minimum value of SAD(x,y,d) is chosen, and the value of d from this minimum will be the value of the pixel in the depth map. At searching the minimum, there are some problems that we should be aware of.
  • FIG. 10 illustrates a depth map result for the stereo image pair illustrated in FIG. 8 .
  • stereo image pairs should contain strong contrast between the colors within the image and there should not be large areas of nearly uniform color.
  • Other researchers who attempted the implementation of this algorithm used computer generated stereo image pairs which contained multiple colors (see Georgoulas et al. and L. Di Stefano, M. Marchionni, and S. Mattoccia, “A Fast Area-Based Stereo Matching Algorithm”, Image and Vision Computing, pp. 983-1005, 2004, which are incorporated by reference).
  • the results after applying the algorithm for faces can be sub-optimal, because the color of facial skin is uniform across most of the face region and the algorithm may not be able to find exactly similar pixels in the stereo image pair.
  • a face model may involve two, orthogonal texture spaces.
  • the development of a dual orthogonal shape subspace is described below which may be derived from the difference and averaged values of the landmark points derived from the right-hand and left hand stereo face images. This separation provides us with an improved 2D registration estimate from the averaged landmark point locations and an orthogonal subspace derived from the different values.
  • FIG. 11 illustrates a fitted AAM face model on the stereo pair of FIG. 8 , and represents an example of fitting the model on the stereo image pair, and illustrates identified positions of considered facial landmarks.
  • An example of corresponding triangulated shapes is then illustrated in FIG. 12 .
  • the landmarks are used as control points for generating the 3D shape, based on their relative 2D displacement in the two images.
  • the result is illustrated at FIG. 13 as corresponding triangulated meshes for the fitted model of FIG. 11 .
  • the 3D shape model allows for 3D constraints to be imposed, making the face model more robust to pose variations; it also reduces the possibility of generating unnatural shape instances during the fitting process, subsequently reducing the risk of an erroneous convergence.
  • Examples of efficient fitting algorithms for the new, so called 2D+3D, model are described at J. Xiao, S. Baker, I. Matthews, and T. Kanade, “Real-Time Combined 2D+3D Active Appearance Models,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR'04), pp. 535-542, 2004; C. Hu, J. Xiao, I. Matthews, S. Baker, J. Cohn, and T.
  • Kanade “Fitting a single active appearance model simultaneously to multiple images,” in Proc. of the British Machine Vision Conference, September 2004; and S. C. Koterba, S. Baker, I. Matthews, C. Hu, J. Xiao, J. Cohn, and T. Kanade, “Multi-View AAM Fitting and Camera Calibration,” in Proc. International Conference on Computer Vision, October, 2005, pp. 511-518, which are each incorporated by reference.
  • FIG. 13 illustrates a 3D shape generated from 2D stereo data with triangulation-based warping (see also FIG. 16 ).
  • 3D faces may be used for gaming applications.
  • a 3D model may also be created within a camera from multiple acquired images. This model then allows enhancements of portrait images in particular by enabling refinements of the facial region based on distance from camera and the determination of the specific regions of a face (cheek, forehead, eye, hair, chin, nose, and so on.)
  • FIGS. 14A-14B and 15 A, 15 B and 15 C are sample portrait images that illustrate certain effects that can result when a 3D face model is created within a camera.
  • a “generic” model may already be available in the camera and the stereo images may be used to refine this generic model to match an individual.
  • a stereo camera may be used in certain embodiments, while in others a stereo camera is not needed.
  • a sequence of sweep panorama images are acquired, which involves moving “around” the subject, rather than “across” a panoramic scene. Unlike a panorama image, the camera would be pointed continuously at the subject, albeit from different perspectives (two such perspectives are illustrated at FIG. 8 ).
  • Scanning may be started, for example, from a left profile, followed by a sweep around the subject.
  • a main (full res) image may be captured from a fully frontal perspective.
  • the sweep may then continue to capture a right profile image.
  • the various preview images may be used to construct a pseudo-3D depth map that may be applied to a post-process to enhance the main image.
  • a sweep in the context of depth of field (DOF), in a portrait enhancement mode, can be performed as just-described or alternatively similar to a sweep that may be performed when acquiring a panorama image, i.e., moving the camera along a linear or curvilinear path. While doing that, the camera can be continuously pointed at the same subject, rather than pointing it each time at a new scene overlapping and adjacent the previous one. At the end, after the camera acquires enough info, a full res image can be captured, or alternatively it can use one of the few images from the sweep, including initializing the sensor in continuous mode at sufficient resolution. Depth from parallax can be advantageously used. A good 3d map can be advantageously created for foreground/background separation. In the process, the camera may be configured to determine to fire the flash as well (i.e., if the light is too low, then flash could help for this).
  • Another way to obtain a 3D depth map is to use depth from defocus (DFD), which involves capturing at least two images of the same scene with different focal depths.
  • DFD depth from defocus
  • This can be a more difficult approach than the others, but it may be used to generate a 3D depth map.
  • advantages can be realized using a combination of DFD and stereoscopic images.
  • FIG. 14A illustrates progressive blurring
  • FIG. 14B illustrates selective blurring
  • a technique may involve obtaining a stereoscopic image of a face using a dual-lens camera, or alternatively by moving the camera to capture facial images from more than one perspective, or alternatively employing a method such as depth from defocus (i.e., capturing at least two differently focused images of the same scene), or through combinations of these.
  • a depth map may be created from these images.
  • a 3D model of the face region may be generated from these images and the depth map.
  • This 3D face model may be used to perform one or more of the following: improving foreground background separation of the modeled face; applying progressive blurring to the face region based on the distance of different portions of the face model from the camera as determined from either the depth map, or the 3D model or both; applying selective blurring to the face based on a combination of distance from the camera and the type of face region (e.g., hair, eyes, nose, mouth, cheek, chin, or regions and/or combinations thereof.
  • the type of face region e.g., hair, eyes, nose, mouth, cheek, chin, or regions and/or combinations thereof.
  • FIG. 15A illustrates an image acquired of a frontal face pose, with simple directional lighting, e.g., from the left.
  • FIG. 15B further illustrates directional lighting.
  • the shadows are even apparent from eyelashes and from the nose demonstrating sophisticated post-acquisition effects that are achieved with 3D modeling.
  • FIG. 15C also illustrates directional lighting.
  • cheek regions are strongly shaded although it is to the foreground, demonstrating the selective application of the directional lighting effect to the cheek and eye regions, and eyelash shadows are again apparent.
  • a technique may involve obtaining a stereoscopic image of the face using a dual-lens camera, or alternatively by moving the camera to capture facial images from more than one perspective or alternatively employing a method such as depth from defocus (i.e. capturing at least two differently focused images of the same scene) or through combinations of these.
  • a depth map may be created from these images.
  • a 3D model may be generated of a face region (or another object or region) from these images and the depth map.
  • the 3D model may include a first set of illuminations components corresponding to a frontally illuminated face and a second set of illumination components corresponding to a directionally illuminated face.
  • the 3D face model may be used to perform one or more of the following: improving foreground background separation of the modeled face; applying progressive directional illumination to the face region based on the distance of different portions of the face model from the camera as determined from either the depth map, or the 3D model or both; applying selective directional illumination to the face based on a combination of distance from the camera and the type of face region (hair, eyes, nose, mouth, cheek, chin, and/or regions and/or combinations thereof).
  • a digital camera may be set into a “portrait acquisition” mode.
  • the user aims the camera at a subject and captures an image.
  • the user is then prompted to move (sweep) the camera slightly to the left or right, keeping the subject at the center of the image.
  • the camera has either a motion sensor, or alternatively may use a frame-to-frame registration engine, such as those that may also be used in sweep panorama techniques, to determine the frame-to-frame displacement.
  • a camera Once a camera has moved approximately 6-7 cm from its original position, the camera acquires a second image of the subject thus simulating the effect of a stereo camera.
  • the acquisition of this second image is automatic, but may be associated with a cue for the user, such as an audible “beep” which informs that the acquisition has been successful.
  • a depth map is next constructed and a 3D face model is generated.
  • a larger distance may be used, or more than two images may be acquired, each at different displacement distances. It may also be useful to acquire a dual image (e.g. flash+no-flash) at each acquisition point to further refine the face model. This approach can be particularly advantageous in certain embodiments for indoor images, or images acquired in low lighting levels or where backlighting is prevalent.
  • the distance to the subject may be advantageously known or determined, e.g., from the camera focusing light, from the detected size of the face region or from information derived within the camera autofocus engine or using methods of depth from defocus, or combinations thereof. Additional methods such as an analysis of the facial shadows or of directional illumination on the face region (see, e.g., US published applications nos. 2008/0013798, 2008/0205712, and 2009/0003661, which are each incorporated by reference and relate to orthogonal lighting models) may additionally be used to refine this information and create an advantageously accurate depth map and subsequently, a 3D face model.
  • a triangulation-based, piecewise affine method may be used for generating and fitting statistical face models. Such may have advantageously efficient computational requirements.
  • the Delauney triangulation technique may be used in certain embodiments, particularly for partitioning a convex hull of control points. The points inside triangles may be mapped via an affine transformation which uniquely assigns the corners of a triangle to their new positions.
  • a different warping method that yields a denser 3D representation, may be based on thin plate splines (TPS) (see, e.g., F. Bookstein, “Principal warps: Thin-plate splines and the decomposition of deformations,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 11, no. 6, pp. 567-585, June 1989, incorporated by reference).
  • TPS thin plate splines
  • TPS-based warping represents a nonrigid registration method, built upon an analogy with a theory in mechanics. Namely, the analogy is made with minimizing the bending energy of a thin metal plate on which pressure is exerted using some point constraints. The bending energy is then given by a quadratic form; the spline is represented as a linear combination (superposition) of eigenvectors of the bending energy matrix:
  • I f ⁇ ⁇ R 2 ⁇ ( ( ⁇ 2 ⁇ f ⁇ x 2 ) 2 + 2 ⁇ ( ⁇ 2 ⁇ f ⁇ x ⁇ ⁇ y ) 2 + ( ⁇ 2 ⁇ f ⁇ 2 ⁇ y ) 2 ) ⁇ ⁇ x ⁇ ⁇ y . ( 8 )
  • the surface is deformed such that to have minimum bending energy.
  • the conditions that need to be met so that (7) is valid, i.e., so that f (x, y) has second-order derivatives, are given by
  • FIG. 16 illustrates an estimated 3D profile from 2D stereo data using Thin Plate Spline-based warping.
  • the solution involves the inversion of a p x p matrix (the bending energy matrix) which has a computational complexity of O(N 3 ), where p is the number of points in the dataset (i.e., the number of pixels in the image); and furthermore, the evaluation process is O(N 2 ).
  • O(N 2 ) the computational complexity of O(N 3 )
  • MLFMM multilevel fast multipole method
  • Embodiments have been described to build improved AAM facial models which condense significant information about facial regions within a relatively small data model. Methods have been described which allow models to be constructed with orthogonal texture and shape subspaces. These allow compensation for directional lighting effects and improved model registration using color information.

Abstract

A stereoscopic image of a face is generated. A depth map is created based on the stereoscopic image. A 3D face model of the face region is generated from the stereoscopic image and the depth map. The 3D face model is applied to process an image.

Description

    PRIORITY
  • This application claims priority to U.S. provisional patent application Ser. No. 61/221,425, filed Jun. 29, 2009. This application is also a continuation in part (CIP) of U.S. patent application no. 12/038,147, filed Feb. 27, 2008, which claims priority to U.S. provisional 60/892,238, filed Feb. 28, 2007. These priority applications are incorporated by reference.
  • BACKGROUND
  • Face detection and tracking technology has become commonplace in digital cameras in the last year or so. All of the practical embodiments of this technology are based on Haar classifiers and follow some variant of the classifier cascade originally proposed by Viola and Jones (see P. A. Viola, M. J. Jones, “Robust real-time face detection”, International Journal of Computer Vision, vol. 57, no. 2, pp. 137-154, 2004, incorporated by reference). These Haar classifiers are rectangular and by computing a grayscale integral image mapping of the original image it is possible to implement a highly efficient multi-classifier cascade. These techniques are also well suited for hardware implementations (see A. Bigdeli, C. Sim, M. Biglari-Abhari and B. C. Lovell, Face Detection on Embedded Systems, Proceedings of the 3rd international conference on Embedded Software and Systems, Springer Lecture Notes In Computer Science; Vol. 4523, p 295-308, May 2007, incorporated by reference).
  • Now, despite the rapid adoption of such in-camera face tracking, the tangible benefits are primarily in improved enhancement of the global image. An analysis of the face regions in an image enables improved exposure and focal settings to be achieved. However current techniques can only determine the approximate face region and do not permit any detailed matching to facial orientation or pose. Neither do they permit matching to local features within the face region. Matching to such detailed characteristics of a face region would enable more sophisticated use of face data and the creation of real-time facial animations for use in, for example, gaming avatars. Another field of application for next-generation gaming technology would be the use of real-time face models for novel user interfaces employing face data to initiate game events, or to modify difficult levels based on the facial expression of a gamer.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
  • FIGS. 1A and 1B illustrate an example of annotations used for the Yale B Database.
  • FIG. 2A illustrates variation between individuals.
  • FIG. 2B illustrates estimated albedo of the individuals of FIG. 2A.
  • FIG. 2C illustrates albedo eigen-textures with 95% energy preservation.
  • FIG. 3A illustrates a reference sample subset of images with various directional lighting effects.
  • FIG. 3B illustrate face samples from FIG. 3A with the contribution of directional lighting removed by filtering (see equation 5).
  • FIG. 4 illustrates images of FIG. 3B subtracted from images of FIG. 3A to yield a set of difference (residual) images.
  • FIG. 5 illustrates process steps to build a color extension of the combined DLS+ULS model for face recognition.
  • FIG. 6 illustrates a general architecture for real-time stereo video capture.
  • FIG. 7 illustrates an internal architecture for real-time stereo video capture.
  • FIG. 8 illustrates a stereo face image pair example.
  • FIG. 9 illustrates the Parallax Effect.
  • FIG. 10 illustrates a depth map result for the stereo image pair of FIG. 8.
  • FIG. 11 illustrates a fitted AAM face model on the stereo pair of FIG. 8.
  • FIG. 12 illustrates corresponding triangulated meshes for a fitted model.
  • FIG. 13 illustrates generating a 3D shape from 2D stereo data with triangulation-based warping.
  • FIG. 14A illustrates progressive blurring.
  • FIG. 14B illustrates selective blurring.
  • FIG. 15A illustrates Frontal Face, with simple Directional Lighting.
  • FIG. 15B illustrates Directional Lighting—Note Shadows from Eyelashes and Nose demonstrating sophisticated post-acquisition effects possible with 3D model.
  • FIG. 15C illustrates Directional Lighting—Note cheek regions is strongly shaded although it is to the foreground, demonstrating the selective application of the directional lighting effect to the cheek and eye regions. Here too we see the eyelash shadows.
  • FIG. 16 illustrates an estimated 3D profile from 2D stereo data using Thin Plate Spline—based warping.
  • DETAILED DESCRIPTIONS OF THE EMBODIMENTS
  • Techniques for improved 2D active appearance face models are described below. When these are applied to stereoscopic image pairs we show that sufficient information on image depth is obtained to generate an approximate 3D face model. Two techniques are investigated, the first based on 2D+3D AAMs and the second using methods based on thin plate splines. The resulting 3D models can offer a practical real-time face model which is suitable for a range of applications in computer gaming. Due to the compact nature of AAMs these are also very suitable for use in embedded devices such as gaming peripherals.
  • A particular class of 2D affine models are involved in certain embodiments, known as active appearance models (AAM), which are relatively fast and are sufficiently optimal to be suitable for in-camera implementations. To improve the speed and robustness of these models, several enhancements are described. Improvements are provided for example to (i) deal with directional lighting effects and (ii) make use of the full color range to improve accuracy and convergence of model to a detected face region.
  • Additionally, the use of stereo imaging provides improved model registration by using two real-time video images with slight variations in spatial perspective. As AAM models may comprise 2D affine models, the use of a real-time stereo video stream opens interesting possibilities to advantageously create full 3D face model from the 2D real-time models.
  • An overview of these models is provided below along with example steps in constructing certain AAM models. Embodiments are also provided with regard to handling directional lighting. The use of the full color range is provided in example models and it is demonstrated below that color information can be used advantageously to improve both the accuracy and speed of convergence of the model. A method is provided for performing face recognition, and comparative analysis shows that results from improved models according to certain embodiments are significantly better than those obtained from a conventional AAM or from a conventional eigenfaces method for performing face recognition. A differential stereo model is also provided which can be used to further enhance model registration and which offers the means to extend a 2D real-time model to a pseudo 3D model. An approach is also provided for generating realistic 3D avatars based on a computationally reduced thin plate spline warping technique. The method incorporates modeling enhancements also described herein. Embodiments that involve the use of AAM models across a range of gaming applications is also provided.
  • AAM Overview
  • This section explains the fundamentals of creating a statistical model of appearance and of fitting the model to image regions.
  • Statistical Models of Appearance
  • AAM was proposed by T. F. Cootes (see, e.g., T. F. Cootes, G. J. Edwards, and C. J. Taylor, “Active appearance models”, Lecture Notes in Computer Science, vol. 1407, pp. 484-, 1998, incorporated by reference), as a deformable model, capable of interpreting and synthesizing new images of the object of interest. Statistical Models of Appearance represent both the shape and texture variations and the correlations between them for a particular class of objects. Example members of the class of objects to be modeled are annotated by a number of landmark points. The shape is defined by the number of landmarks chosen to best depict the contour of the object of interest, in our case a person's face.
  • Model Fitting to an Image Region
  • After a statistical model of appearance is created, an AAM algorithm can be employed to fit the model to a new, unseen, image region. The statistical model is linear in both shape and texture. However, fitting the model to a new image region is a non-linear optimization process. The fitting algorithm works by minimizing the error between a query image and the equivalent model-synthesized image.
  • In this paper we use an optimization scheme which is robust to directional variations in illumination. This relies on the fact that lighting information is decoupled from facial identity information. This can be seen as an adaptation of the method(s) proposed at A. U. Batur and M. H. Hayes, “Adaptive active appearance models,” IEEE Transactions on Image Processing, vol. 14, no. 11, pp. 1707-1721, 2005, incorporated by reference. These authors use an adaptive gradient where the gradient matrix is linearly adapted according to the texture composition of the target image, generating an improved estimate of the actual gradient. In our model the separation of texture into lighting dependent and lighting independent subspaces enables a faster adaptation of the gradient.
  • Initialization of the Model within an Image
  • Prior to implementing that AAM fitting procedure it is necessary to initialize the model within an image. To detect faces we employ a modified Viola-Jones face detector (see J. J. Gerbrands, “On the relationships between SVD, KLT and PCA.” Pattern Recognition, vol. 14, no. 1-6, pp. 375-381, 1981, incorporated by reference) which can accurately estimate the position of the eye regions within a face region. Using the separation of the eye regions also provides an initial size estimate for the model fitting. The speed and accuracy of this detector enables us to apply the AAM model to large unconstrained image sets without a need to pre-filter or crop face regions from the input image set.
  • Model Enhancements Illumination and Multi-Channel Colour Registration Building an Initial Identity Model
  • The reference shape used to generate the texture vectors should be the same one for all models, i.e. either identity or directional lighting models. Our goal is to determine specialized subspaces, such as the identity subspace or the directional lighting subspace.
  • We first need to model only the identity variation between individuals. For training this identity-specific model we only use images without directional lighting variation. Ideally these face images should be obtained in diffuse lighting conditions. Textures are extracted by projecting the pixel intensities across the facial region, as defined by manual annotation, into the reference shape—chosen as the mean shape of the training data. FIG. 1 illustrates examples of annotations used for the Yale B Database.
  • The number of landmark points used should be kept fixed over the training data set. In addition to this, each landmark point must have the same face geometry correspondence for all images. The landmarks should predominantly target fiducial points, which permit a good description of facial geometry, allowing as well the extraction of geometrical differences between different individuals. The facial textures corresponding to images of individuals in the Yale database with frontal illumination are represented in FIGS. 2A, 2B and 2C. FIG. 2A illustrates variation between individuals. FIG. 2B illustrates estimated albedo of the individuals. FIG. 2C illustrates albedo eigen-textures with 95% energy preservation. The identity model can now be generated from the albedo images based on the standard PCA technique.
  • Building a Model for Directional Lighting Variations
  • Consider now all facial texture which exhibit directional lighting variations from all four (4) subsets. These textures are firstly projected onto the previously built subspace of individual variation, ULS. These texture vectors contain some directional lighting information, with g.
  • In equation 1 below, the factor g contains both identity and directional lighting information. The same reference shape may be used to obtain the new texture vectors g, which ensures that the previous and new texture vectors have all equal lengths. In FIG. 3A, a random selection of faces is shown as a reference sample subset of images with various directional lighting effects. The projection of the texture vectors g onto ULS gives the sets of optimal texture parameter vectors as in:

  • b ident (opt)ident T(g− t )  (1)
  • The back-projection stage returns the texture vector, optimally synthesized by the identity model. The projection/back-projection process filters out all the variations which could not be explained by the identity model. Thus, for this case, directional lighting variations are filtered out by this process,

  • g filt = t ident b ident (opt)  (2)
  • Continuing with the procedure for the examples in FIG. 3A, their filtered versions are illustrated in the example of FIG. 3B, which illustrates face samples from FIG. 3A with the contribution of directional lighting removed by filtering (per equation 5, below).
  • The residual texture is further obtained as the difference between the original texture and the synthesized texture which retained only the identity information. This residual texture normally retains the information other than identity.

  • t res =g−g filt =g− t−Φ ident b ident (opt)  (3)
  • The residual images give the directional lighting information, as illustrated at FIG. 4 which includes the images of FIG. 3B subtracted from the images of FIG. 3A to yield a set of difference (residual) images. These residuals are then modeled using PCA in order to generate a directional lighting subspace.
  • Creating a Merged Face Model
  • As described above, three separate components of the face model have been generated. These are: (i) the shape model of the face, (ii) texture model encoding identity information, and (iii) the texture model for directional lighting. The resulting texture subspaces are also orthogonal due to the approach described above. The fusion between the two texture models can be realized by a weighted concatenation of parameters:
  • c = [ W S b S b ident W light b light ] ( 4 )
  • where Wlighting and Ws are two vectors of weights used to compensate for the differences in units between the two sets of texture parameters, and for the differences in units between shape and texture parameters, respectively.
  • Fitting the Lighting Enhanced Model
  • The conventional AAM algorithm uses a gradient estimate built from training images and thus cannot be successfully applied to images where there are significant variations in illumination conditions. The solution proposed by Batur et al. is based on using an adaptive gradient AAM (see, e.g., F. Kahraman, M. Gokmen, S. Darkner, and R. Larsen, “An active illumination and appearance (AIA) model for face alignment,” Computer Vision and Pattern Recognition, 2007. CVPR '07. IEEE Conference on, pp. 1-7, June 2007. The gradient matrix is linearly adapted according to texture composition of the target image. We further modify the approach of Batur (cited above) to handle our combined ULS and DLS texture subspace (see, e.g., M. Ionita, “Advances in the design of statistical face modeling techniques for face recognition”, PhD Thesis, NUI Galway, 2009, and M. Ionita and P. Corcoran, “A Lighting Enhanced Facial Model: Training and Fast Optimization Scheme”, submitted to Pattern Recognition, May 2009, which are incorporated by reference.
  • Colour Space Enhancements
  • When a typical multi-channel image is represented in a conventional color space such as RGB, there are correlations between its channels. For natural images, the cross-correlation coefficient between B and R channels is ˜0.78, between R and G channels is ˜0.98, and for G and B channels is ˜0.94 (see M. Tkalcic and J. F. Tasic, “Colour spaces—perceptual, historical and applicational background,” in IEEE, EUROCON, 2003, incorporated by reference). This inter-channel correlation explains why previous authors (G. J. Edwards, T. F. Cootes, and C. J. Taylor, “Advances in active appearance models,” in International Conference on Computer Vision (ICCV'99), 1999, pp. 137-142, incorporated by reference) obtained poor results using RGB AAM models.
  • Ohta's space (see Y. Ohta, T. Kanade, and T. Sakai, “Color Information for Region Segmentation”, Ó Comput. Graphics Image Process., vol. 13, pp. 222-240, 1980, incorporated by reference) realizes a statistically optimal minimization of the inter-channel correlations, i.e. decorrelation of the color components, for natural images. The conversion from RGB to I1I2I3 is given by the simple linear transformations in (5a-c).
  • I 1 = R + G + B 3 ( 5 a ) I 2 = R - B 2 ( 5 b ) I 3 = 2 G - R - B 4 ( 5 c )
  • I1 represents the achromatic (intensity) component, while I2 and I3 are the chromatic components. By using Ohta's space the AAM search algorithm becomes more robust to variations in lighting levels and color distributions. A summary of comparative results across different color spaces is provided in Tables I and II (see also M. C. Ionita, P. Corcoran, and V. Buzuloiu, “On color texture normalization for active appearance models,” IEEE Transactions on Image Processing, vol. 18, issue 6, pp. 1372-1378, June 2009 and M. Ionita, “Advances in the design of statistical face modeling techniques for face recognition”, PhD Thesis, NUI Galway, 2009, which are incorporated by reference).
  • TABLE I
    TEXTURE NORMALISATION RESULTS
    ON (PIE) SUBSET 2 (Unseen)
    Success Pt-Crv Pt-Pt PCTE
    Model [%] (Mean/Std) (Mean/Std) (Mean/Std)
    Grey scale 88.46 3.93/2.00 6.91/5.45
    RGB ON 80.77 3.75/1.77 7.09/4.99 7.20/2.25
    CIELAB GN 100 2.70/0.93 4.36/1.63 5.91/1.19
    I1I2I3 SChN 100 2.60/0.93 4.20/1.45 5.87/1.20
    RGB SChN 73.08 4.50/2.77 8.73/7.20 7.25/2.67
    CIELAB SChN 88.46 3.51/2.91 6.70/8.29 6.28/2.09
    I1I2I3 GN 92.31 3.23/1.21 5.55/2.72 6.58/1.62
  • TABLE II
    CONVERGENCE RESULTS ON UNSEEN DATABASES
    Success Pt-Crv PTE
    Model Rate [%] (Mean/Std/Median) (Mean/Std/Median)
    db1-Grayscale* 92.17 5.10 1.66 4.90 4.28 1.03 4.21
    db1-RGB-name 99.13 4.94 1.37 4.82 10.09 1.58 9.93
    db1-RGB-G 98.26 4.98 1.44 4.65 7.49 1.98 7.02
    db1-RGB-Ch 87.83 5.32 1.65 5.08 6.33 1.40 5.95
    db1-I1I2I3-Ch 99.13 3.60 1.32 3.32 5.10 1.01 4.85
    db1-I1I2-Ch 99.13 4.25 1.65 3.79 8.26 4.11 6.10
    db2-Grayscale* 75.73 4.17 1.44 3.67 5.12 4.24 4.03
    db2-RGB-name 84.47 4.02 1.40 3.69 12.43 3.43 12.41
    db2-RGB-G 94.17 3.74 1.45 3.23 9.04 1.83 8.97
    db2-RGB-Ch 62.14 4.01 1.60 3.46 7.70 4.26 6.06
    db2-I1I2I3-Ch 88.35 3.31 1.26 2.98 6.16 2.28 5.73
    db2-I1I2-Ch 87.38 3.60 1.55 3.04 10.00 3.41 8.94
    db3-Grayscale* 63.89 4.85 2.12 4.26 4.90 3.44 3.98
    db3-RGB-name 75.22 4.44 1.79 3.99 14.23 4.79 13.34
    db3-RGB-G 65.28 4.55 2.03 4.01 9.68 2.81 9.27
    db3-RGB-Ch 59.72 5.02 2.04 4.26 7.16 4.91 5.74
    db3-I1I2I3-Ch 86.81 3.53 1.49 3.15 6.04 2.56 5.20
    db3-I1I2-Ch 86.81 3.90 1.66 3.41 6.60 1.94 6.30
  • Uses of Statistical Models and AAM
  • An advantageous AAM model may be used in face recognition. However there are a multitude of alternative applications for such models. These models have been widely used for face tracking (see, e.g., P. Corcoran, M. C. Ionita, I. Barcivarov, “Next generation face tracking technology using AAM techniques,” Signals, Circuits and Systems, ISSCS 2007, International Symposium on, Volume 1, p 1-4, 13-14 Jul. 2007, incorporated by reference), and for measuring facial pose and orientation.
  • In other research we have demonstrated the use of AAM models for detecting phenomena such as eye-blink, analysis and characterization of mouth regions, and facial expressions (see I. Bacivarov, M. Ionita, P. Corcoran, “Statistical Models of Appearance for Eye Tracking and Eye-Blink Detection and Measurement”. IEEE Transactions on Consumer Electronics, August 2008; I. Bacivarov, M. C. Ionita, and P. Corcoran, A Combined Approach to Feature Extraction for Mouth Characterization and Tracking, in Signals and Systems Conference, 208. (ISSC 2008). IET Irish, Volume 1, p 156-161, Galway, Ireland 18-19 Jun. 2008; and J. Shi, A. Samal, and D. Marx, “How effective are landmarks and their geometry for face recognition?” Comput. Vis. Image Underst., vol. 102, no. 2, pp. 117-133, 2006, respectively, which are incorporated by reference). In such context these models are more sophisticated than other pattern recognition methods which can only determine if, for example, an eye is in an open or closed state. Our models can determine other metrics such as the degree to which an eye region is open or closed or the gaze direction of the eye. This opens the potential for sophisticated game avatars or novel gaming UI methods.
  • Building a Combined Model
  • A notable applicability of the directional lighting sub-model, generated from a grayscale training database, is that it can be efficiently incorporated into a color face model. This process is illustrated in FIG. 5 which shows a process steps to build a color extension of the combined DLS+ULS model for face recognition.
  • The left-hand process diagram of FIG. 5 illustrates the partitioning of the model texture space into orthogonal ULS and DLS subspaces. Step 1 involves N persons and uniform lighting. Step 2 involves N persons and 30 directional lighting conditions. Using texture PCA, step 1 moves to uniform lighting space (ULS), and using projection, step 2 moves to the ULS. Using back projection, N×30 uniform light filtered images are the result. Step 3 involves an image difference fed by the step 2 images and the resultant N×3-uniform light filtered images. Step 4 involves N×30 difference lighting images. Using texture PCA, a directional lighting space (DLS) is achieved.
  • The right-hand side process diagram of FIG. 5 shows how the DLS subspace can be used to train a color ULS, implemented in the other color space. Step 5 involves M persons in color with random lighting conditions. These are fed to the directional lighting subspace (DLS). Using back projection, M identity filtered color images are achieved and fed at step 6 to an image difference along with the m persons in color with random lighting conditions. Step 7 then involves M difference (identity) images. Using texture PCA, a uniform lighting (identity) color texture space (ULCTS) is achieved.
  • The example processes illustrated at FIG. 5 yield a full color ULS which retains the orthogonality with the DLS and when combined with it yields an enhanced AAM model incorporating shape+DLS+color ULS subspaces. The color ULS has the same improved fitting characteristics as the color model (see, M. C. Ionita, P. Corcoran, and V. Buzuloiu, “On color texture normalization for active appearance models,” IEEE Transactions on Image Processing, vol. 18, issue 6, pp. 1372-1378, June 2009, incorporated by reference). This combined model exhibits both improved registration and robustness to directional lighting.
  • Model Application to Face Recognition Benchmarking for AAM-Based Face Recognition
  • The recognition tests which follow have been performed by considering the large gallery test performance (see P. J. Phillips, P. Rauss, and S. Der, “FERET recognition algorithm development and test report,” U.S. Army Research Laboratory, Tech. Rep., 1996, incorporated by reference). As a benchmark with other methods we decided to compare relative performance with respect to the well-known eigenfaces method (see M. A. Turk and A. P. Pentland, “Face recognition using eigenfaces,” in Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR'91), 586-591, 1991, incorporated by reference). Detailed results of these tests are reported in M. Ionita, “Advances in the design of statistical face modeling techniques for face recognition”, PhD Thesis, NUI Galway, 2009, incorporated by reference. There is a reported modest improvement of 5%-8% to be achieved in using a color AAM method (RGB) over a grayscale AAM. The performance of the color AAM is approximately equal to that of both grayscale and color eigenfaces methods.
  • Tests on the Improved AAM Model
  • The color AAM techniques based on RGB color space generally cannot compete with the conventional eigenface method of face recognition. Conversely, the I1I2I3 based models perform at least as well as the eigenface method, even when the model has been trained on a different database. When trained on the same database we conclude that the I1I2I3 SChN model outperforms the eigenface method by at least 10% when the first 50 components are used. If we restrict our model to the first 5 or 10 components then the differential is about 20% in favor of the improved AAM model.
  • Model Enhancements Differential AAM from Real-Time Stereo Channels Hardware Architecture of Stereo Imaging System
  • An example of a general architecture of a stereo imaging system is illustrated at FIG. 6, which shows two CMOS sensors and a VGA monitor connected to a power PC with Xilinx Virtex4 FPGA and DDR SDRAM. The two CMOS sensors are connected to an FPGA which incorporates a PowerPC core and associated SDRAM. Additional system components can be added to implement a dual stereo image processing pipeline (see, e.g., I. Andorko and P. Corcoran, “FPGA Based Stereo Imaging System with Applications in Computer Gaming”, at International IEEE Consumer Electronics Society's Games Innovations Conference 2009 (ICE-GIC 09), London, UK, incorporated by reference).
  • The development board is a Xilinx ML405 development board, with a Virtex 4 FPGA, a 64 MB DDR SDRAM memory, and a PowerPC RISC processor. The clock frequency of the system is 100 MHz. An example internal architecture of the system in accordance with certain embodiments is illustrated at FIG. 7, which shows two conversion blocks respectively coupled to camera units 1 and 2. The camera units 1 and 2 feed a PLB that feeds a VGA controller. A DCR is connected to the camera units 1 and 2, the VGA controller, an I2C controller and a Power PC. The PLB is also coupled with DDR SDRAM. The sensor used in this embodiment includes a ⅓ inch SXGA CMOS sensor made by Micron. It has an active zone of 1280×1024 pixels. It is programmable through the I2C interface. It works at 13.9 fps and the clock frequency is 25 MHz. This sensor was selected because of its small size, low cost and the specifications of these sensors are satisfactory for this project. This system enables real-time stereo video capture with a fixed distance between the two imaging sensors. FIG. 8 illustrates a stereo face image pair example.
  • Determination of a Depth Map
  • When using two sensors for stereo imaging, the problem of parallax effect appears. Parallax is an apparent displacement or difference of orientation of an object viewed along two different lines of sight, and is measured by the angle or semi-angle of inclination between those two lines.
  • The advantage of the parallax effect is that with the help of this, depth maps can be computed. The computation in certain embodiments involves use of pairs of rectified images (see, K. Muhlmann, D. Maier, J. Hesser, R. Manner, “Calculating Dense Disparity Maps from Color Stereo Images, an Efficient Implementation”, International Journal of Computer Vision, vol. 47, numbers 1-3, pp. 79-88, April 2002, incorporated by reference). This means that corresponding epipolar lines are horizontal and on the same height. The search of corresponding pictures takes place in horizontal direction only in certain embodiments. For every pixel in the left image, the goal is to find the corresponding pixel in the right image, or vice-versa. FIG. 9 illustrates the parallax effect.
  • It is difficult or at least computationally expensive to find corresponding single pixels, and so windows of different sizes (3×3; 5×5; 7×7) may be used. The size of window is computed based on the value of the local variation of each pixel (see C. Georgoulas, L. Kotoulas, G. Ch. Sirakoulis, I. Andreadis, A. Gasteratos, “Real-Time Disparity Map Computation Module”, Microprocessors and Microsystems 32, pp. 159-170, 2008, incorporated by reference). A formula that may be used for the computation of the local variation per Georgoulas et al. is shown below n equation 6:
  • LV ( p ) = i = 1 N j = 1 N I ( i , j ) - μ ( 6 )
  • where μ is the average grayscale value of image window, and N is the selected square window size.
  • The first local variation calculation may be made over a 3×3 window. After this, the points with a value under a certain threshold are marked for further processing. The same operation is done for 5×5 and 7×7 windows as well. The size of each of the windows is stored for use in the depth map computation. The operation to compute the depth map is the Sum of Absolute Differences for RGB images (SAD). The value of SAD is computed for up to a maximum value of d on the x line. After all the SAD values have been computed, the minimum value of SAD(x,y,d) is chosen, and the value of d from this minimum will be the value of the pixel in the depth map. At searching the minimum, there are some problems that we should be aware of. If the minimum is not unique, or its position is dmin or dmax, the value is discarded. Instead of just seeking the minimum, it is helpful to track the three smallest SAD values as well. The minimum defines a threshold above which the third smallest value must lie. Otherwise, the value is discarded. FIG. 10 illustrates a depth map result for the stereo image pair illustrated in FIG. 8.
  • One of the conditions for a depth map computation technique to work properly is that the stereo image pairs should contain strong contrast between the colors within the image and there should not be large areas of nearly uniform color. Other researchers who attempted the implementation of this algorithm used computer generated stereo image pairs which contained multiple colors (see Georgoulas et al. and L. Di Stefano, M. Marchionni, and S. Mattoccia, “A Fast Area-Based Stereo Matching Algorithm”, Image and Vision Computing, pp. 983-1005, 2004, which are incorporated by reference). In some cases, the results after applying the algorithm for faces can be sub-optimal, because the color of facial skin is uniform across most of the face region and the algorithm may not be able to find exactly similar pixels in the stereo image pair.
  • AAM Enhanced Shape Model
  • A face model may involve two, orthogonal texture spaces. The development of a dual orthogonal shape subspace is described below which may be derived from the difference and averaged values of the landmark points derived from the right-hand and left hand stereo face images. This separation provides us with an improved 2D registration estimate from the averaged landmark point locations and an orthogonal subspace derived from the different values.
  • This second subspace enables an improved determination of the SAD values and the estimation of an enhanced 3D surface view over the face region. FIG. 11 illustrates a fitted AAM face model on the stereo pair of FIG. 8, and represents an example of fitting the model on the stereo image pair, and illustrates identified positions of considered facial landmarks. An example of corresponding triangulated shapes is then illustrated in FIG. 12. The landmarks are used as control points for generating the 3D shape, based on their relative 2D displacement in the two images. The result is illustrated at FIG. 13 as corresponding triangulated meshes for the fitted model of FIG. 11.
  • The 3D shape model allows for 3D constraints to be imposed, making the face model more robust to pose variations; it also reduces the possibility of generating unnatural shape instances during the fitting process, subsequently reducing the risk of an erroneous convergence. Examples of efficient fitting algorithms for the new, so called 2D+3D, model are described at J. Xiao, S. Baker, I. Matthews, and T. Kanade, “Real-Time Combined 2D+3D Active Appearance Models,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR'04), pp. 535-542, 2004; C. Hu, J. Xiao, I. Matthews, S. Baker, J. Cohn, and T. Kanade, “Fitting a single active appearance model simultaneously to multiple images,” in Proc. of the British Machine Vision Conference, September 2004; and S. C. Koterba, S. Baker, I. Matthews, C. Hu, J. Xiao, J. Cohn, and T. Kanade, “Multi-View AAM Fitting and Camera Calibration,” in Proc. International Conference on Computer Vision, October, 2005, pp. 511-518, which are each incorporated by reference.
  • Examples of full 3D face models, called 3D morphable models (3DMM), are described at V. Blanz and T. Vetter, “A morphable model for the synthesis of 3D faces,” in Proceedings of the 26th annual conference on Computer graphics and interactive techniques, pp. 187-194, 1999, incorporated by reference. Yet, these models have a high complexity and significant computational requirements, thus in certain embodiments the approaches based on the simpler AAM techniques are alternatively used, particularly for implementation in embedded systems. FIG. 13 illustrates a 3D shape generated from 2D stereo data with triangulation-based warping (see also FIG. 16).
  • Portrait Enhancement with 3D Face Modeling
  • In certain embodiments, 3D faces may be used for gaming applications. In further embodiments, a 3D model may also be created within a camera from multiple acquired images. This model then allows enhancements of portrait images in particular by enabling refinements of the facial region based on distance from camera and the determination of the specific regions of a face (cheek, forehead, eye, hair, chin, nose, and so on.)
  • FIGS. 14A-14B and 15A, 15B and 15C are sample portrait images that illustrate certain effects that can result when a 3D face model is created within a camera. In some embodiments, a “generic” model may already be available in the camera and the stereo images may be used to refine this generic model to match an individual. Also, a stereo camera may be used in certain embodiments, while in others a stereo camera is not needed. In one alternative embodiment, a sequence of sweep panorama images are acquired, which involves moving “around” the subject, rather than “across” a panoramic scene. Unlike a panorama image, the camera would be pointed continuously at the subject, albeit from different perspectives (two such perspectives are illustrated at FIG. 8).
  • Scanning may be started, for example, from a left profile, followed by a sweep around the subject. A main (full res) image may be captured from a fully frontal perspective. The sweep may then continue to capture a right profile image. The various preview images may be used to construct a pseudo-3D depth map that may be applied to a post-process to enhance the main image.
  • In the context of depth of field (DOF), in a portrait enhancement mode, a sweep can be performed as just-described or alternatively similar to a sweep that may be performed when acquiring a panorama image, i.e., moving the camera along a linear or curvilinear path. While doing that, the camera can be continuously pointed at the same subject, rather than pointing it each time at a new scene overlapping and adjacent the previous one. At the end, after the camera acquires enough info, a full res image can be captured, or alternatively it can use one of the few images from the sweep, including initializing the sensor in continuous mode at sufficient resolution. Depth from parallax can be advantageously used. A good 3d map can be advantageously created for foreground/background separation. In the process, the camera may be configured to determine to fire the flash as well (i.e., if the light is too low, then flash could help for this).
  • Another way to obtain a 3D depth map is to use depth from defocus (DFD), which involves capturing at least two images of the same scene with different focal depths. For digital cameras that have a very uniform focal depth, this can be a more difficult approach than the others, but it may be used to generate a 3D depth map. In other embodiments, advantages can be realized using a combination of DFD and stereoscopic images.
  • FIG. 14A illustrates progressive blurring, while FIG. 14B illustrates selective blurring. In accordance with this embodiment, a technique may involve obtaining a stereoscopic image of a face using a dual-lens camera, or alternatively by moving the camera to capture facial images from more than one perspective, or alternatively employing a method such as depth from defocus (i.e., capturing at least two differently focused images of the same scene), or through combinations of these. A depth map may be created from these images. A 3D model of the face region may be generated from these images and the depth map. This 3D face model may be used to perform one or more of the following: improving foreground background separation of the modeled face; applying progressive blurring to the face region based on the distance of different portions of the face model from the camera as determined from either the depth map, or the 3D model or both; applying selective blurring to the face based on a combination of distance from the camera and the type of face region (e.g., hair, eyes, nose, mouth, cheek, chin, or regions and/or combinations thereof. The following are incorporated by reference as disclosing various alternative embodiments and applications of described embodiments: U.S. Pat. Nos. 7,606,417, 7,680,342, 7,692,696, 7,469,071, 7,515,740 and 7,565,030, and US published applications nos. 2010/0126831. 2009/0273685, 2009/0179998, 2009/0003661, 2009/0196466, 2009/0244296, 2009/0190803, 2009/0263022, 2009/0179999, 2008/0292193, 2008/0175481, 2007/0147820, and 2007/0269108, and U.S. Ser. No. 12/636,647.
  • FIG. 15A illustrates an image acquired of a frontal face pose, with simple directional lighting, e.g., from the left. FIG. 15B further illustrates directional lighting. The shadows are even apparent from eyelashes and from the nose demonstrating sophisticated post-acquisition effects that are achieved with 3D modeling. FIG. 15C also illustrates directional lighting. In this example, cheek regions are strongly shaded although it is to the foreground, demonstrating the selective application of the directional lighting effect to the cheek and eye regions, and eyelash shadows are again apparent. In accordance with these embodiments, a technique may involve obtaining a stereoscopic image of the face using a dual-lens camera, or alternatively by moving the camera to capture facial images from more than one perspective or alternatively employing a method such as depth from defocus (i.e. capturing at least two differently focused images of the same scene) or through combinations of these. A depth map may be created from these images. A 3D model may be generated of a face region (or another object or region) from these images and the depth map. The 3D model may include a first set of illuminations components corresponding to a frontally illuminated face and a second set of illumination components corresponding to a directionally illuminated face. The 3D face model may be used to perform one or more of the following: improving foreground background separation of the modeled face; applying progressive directional illumination to the face region based on the distance of different portions of the face model from the camera as determined from either the depth map, or the 3D model or both; applying selective directional illumination to the face based on a combination of distance from the camera and the type of face region (hair, eyes, nose, mouth, cheek, chin, and/or regions and/or combinations thereof).
  • In an alternative embodiment, a digital camera may be set into a “portrait acquisition” mode. In this mode the user aims the camera at a subject and captures an image. The user is then prompted to move (sweep) the camera slightly to the left or right, keeping the subject at the center of the image. The camera has either a motion sensor, or alternatively may use a frame-to-frame registration engine, such as those that may also be used in sweep panorama techniques, to determine the frame-to-frame displacement. Once a camera has moved approximately 6-7 cm from its original position, the camera acquires a second image of the subject thus simulating the effect of a stereo camera. The acquisition of this second image is automatic, but may be associated with a cue for the user, such as an audible “beep” which informs that the acquisition has been successful.
  • After aligning the two images a depth map is next constructed and a 3D face model is generated. In alternative embodiments, a larger distance may be used, or more than two images may be acquired, each at different displacement distances. It may also be useful to acquire a dual image (e.g. flash+no-flash) at each acquisition point to further refine the face model. This approach can be particularly advantageous in certain embodiments for indoor images, or images acquired in low lighting levels or where backlighting is prevalent.
  • The distance to the subject may be advantageously known or determined, e.g., from the camera focusing light, from the detected size of the face region or from information derived within the camera autofocus engine or using methods of depth from defocus, or combinations thereof. Additional methods such as an analysis of the facial shadows or of directional illumination on the face region (see, e.g., US published applications nos. 2008/0013798, 2008/0205712, and 2009/0003661, which are each incorporated by reference and relate to orthogonal lighting models) may additionally be used to refine this information and create an advantageously accurate depth map and subsequently, a 3D face model.
  • Model Application 3D Gaming Avatars
  • A triangulation-based, piecewise affine method may be used for generating and fitting statistical face models. Such may have advantageously efficient computational requirements. The Delauney triangulation technique may be used in certain embodiments, particularly for partitioning a convex hull of control points. The points inside triangles may be mapped via an affine transformation which uniquely assigns the corners of a triangle to their new positions.
  • A different warping method, that yields a denser 3D representation, may be based on thin plate splines (TPS) (see, e.g., F. Bookstein, “Principal warps: Thin-plate splines and the decomposition of deformations,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 11, no. 6, pp. 567-585, June 1989, incorporated by reference). Further examples of the use of TPS for improving the convergence accuracy of color AAMs are provided at M. C. Ionita and P. Corcoran, “Benefits of Using Decorrelated Color Information for Face Segmentation/Tracking,” Advances in Optical Technologies, vol. 2008, Article ID 583687, 8 pages, 2008. doi:10.1155/2008/583687, incorporated by reference. TPS-based warping may be used for estimating 3D face profiles.
  • In the context of generating realistic 3D avatars, the choice of TPS-based warping technique offers an advantageous solution. This technique is more complex that the piecewise linear warping employed for example above; yet simplifies versions are possible with reduced computational complexity. TPS-based warping represents a nonrigid registration method, built upon an analogy with a theory in mechanics. Namely, the analogy is made with minimizing the bending energy of a thin metal plate on which pressure is exerted using some point constraints. The bending energy is then given by a quadratic form; the spline is represented as a linear combination (superposition) of eigenvectors of the bending energy matrix:
  • f ( x , y ) = a 1 + a x x + a y y + i = 1 p w i U ( ( x i , y i ) - ( x , y ) ) . ( 7 )
  • where U(r)=r2 log(r); (xi, yi) are the initial control points. a=(a1 ax ay) defines the affine part, while w defines the nonlinear part of the deformation. The total bending energy is expressed as
  • I f = R 2 ( ( 2 f x 2 ) 2 + 2 ( 2 f x y ) 2 + ( 2 f 2 y ) 2 ) x y . ( 8 )
  • The surface is deformed such that to have minimum bending energy. The conditions that need to be met so that (7) is valid, i.e., so that f (x, y) has second-order derivatives, are given by
  • i = 1 v w i = 0 and ( 9 ) i = 1 p w i x i = 0 ; i = 1 p w i y i = 0. ( 10 )
  • Adding to this the interpolation conditions f(xi, yi)=vi, (7) can now be written as the linear system in (10):
  • [ K P P T O ] [ w a ] = [ v o ] , ( 11 )
  • where Kij=U(∥(xi, yi)−(xj, yj)∥), O is a 3×3 matrix of zeros, o is a 3×1 vector of zeros, Pij=(1, xi, yi); w and v are the column vectors formed by wi and vi, respectively, while a=[a1 ax ay]T.
  • FIG. 16 illustrates an estimated 3D profile from 2D stereo data using Thin Plate Spline-based warping. The main drawback of using the thin plate splines used to be considered their high computational load. The solution involves the inversion of a p x p matrix (the bending energy matrix) which has a computational complexity of O(N3), where p is the number of points in the dataset (i.e., the number of pixels in the image); and furthermore, the evaluation process is O(N2). However, progress has been made and will continue to be made that serves to speed this process up. For example, an approximation approach was proved in G. Donato and S. Belongie, “Approximate thin plate spline mappings,” in ECCV (3), 2002, pp. 21-31, incorporated by reference, and such has been observed to be very efficient in dealing with the first problem, reducing greatly the computational burden. As far as the evaluation process is concerned, the multilevel fast multipole method (MLFMM) framework was described in R. K. Beatson and W. A. Light, “Fast evaluation of radial basis functions: methods for two-dimensional polyharmonic splines,” IMA Journal of Numerical Analysis, vol. 17, no. 3, pp. 343-372,1997, incorporated by reference, for the evaluation of two-dimensional polyharmonic splines. Meanwhile, in A. Zandifar, S.-N. Lim, R. Duraiswami, N. A. Gumerov, and L. S. Davis, “Multi-level fast multipole method for thin plate spline evaluation.” In ICIP, 2004, pp. 1683-1686, incorporated by reference, this work was extended for the specific case of TPS, showing that a reduction of the computational complexity from O(N2) to O(N logN) is indeed possible. Thus the computational difficulties involving the use of TPS have been greatly reduced. Based on this warping technique, 3D facial profiles may be generated as illustrated at FIG. 16.
  • Embodiments have been described to build improved AAM facial models which condense significant information about facial regions within a relatively small data model. Methods have been described which allow models to be constructed with orthogonal texture and shape subspaces. These allow compensation for directional lighting effects and improved model registration using color information.
  • These improved models may then be applied to stereo image pairs to deduce 3D facial depth data. This enables the extension of the AAM to provide a 3D face model. Two approaches have been described, one based on 2D+3D AAM and a second approach based on thin plate spline warpings. Those based on thin plate splines are shown to produce a particularly advantageous 3D rendering of the face data. These extended AAM based techniques may be combined with stereoscopic image data offering improved user interface methods and the generation of dynamic real-time avatars for computer gaming applications.
  • While exemplary drawings and specific embodiments of the present invention have been described and illustrated, it is to be understood that that the scope of the present invention is not to be limited to the particular embodiments discussed. Thus, the embodiments shall be regarded as illustrative rather than restrictive, and it should be understood that variations may be made in those embodiments by workers skilled in the arts without departing from the scope of the present invention.
  • In addition, in methods that may be performed according to preferred embodiments herein and that may have been described above, the operations have been described in selected typographical sequences. However, the sequences have been selected and so ordered for typographical convenience and are not intended to imply any particular order for performing the operations, except for those where a particular order may be expressly set forth or where those of ordinary skill in the art may deem a particular order to be necessary.
  • In addition, all references cited above and below herein, as well as the background, invention summary, abstract and brief description of the drawings, are all incorporated by reference into the detailed description of the preferred embodiments as disclosing alternative embodiments.
  • The following are incorporated by reference: U.S. Pat. Nos. 7,715,597, 7,702,136, 7,692,696, 7,684,630, 7,680,342, 7,676,108, 7,634,109, 7,630,527, 7,620,218, 7,606,417, 7,587,068, 7,403,643, 7,352,394, 6,407,777, 7,269,292, 7,308,156, 7,315,631, 7,336,821, 7,295,233, 6,571,003, 7,212,657, 7,039,222, 7,082,211, 7,184,578, 7,187,788, 6,639,685, 6,628,842, 6,256,058, 5,579,063, 6,480,300, 5,781,650, 7,362,368, 7,551,755, 7,515,740, 7,469,071, 5,978,519, 7,630,580, 7,567,251, 6,940,538, 6,879,323, 6,456,287, 6,552,744, 6,128,108, 6,349,153, 6,385,349, 6,246,413, 6,604,399 and 6,456,323; and
    • U.S. published application nos. 2002/0081003, 2003/0198384, 2003/0223622, 2004/0080631, 2004/0170337, 2005/0041121, 2005/0068452, 2006/0268130, 2006/0182437, 2006/0077261, 2006/0098890, 2006/0120599, 2006/0140455, 2006/0153470, 2006/0204110, 2006/0228037, 2006/0228038, 2006/0228040, 2006/0276698, 2006/0285754, 2006/0188144, 2007/0071347, 2007/0110305, 2007/0147820, 2007/0189748, 2007/0201724, 2007/0269108, 2007/0296833, 2008/0013798, 2008/0031498, 2008/0037840, 2008/0106615, 2008/0112599, 2008/0175481, 2008/0205712, 2008/0219517, 2008/0219518, 2008/0219581, 2008/0220750, 2008/0232711, 2008/0240555, 2008/0292193, 2008/0317379, 2009/0022422, 2009/0021576, 2009/0080713, 2009/0080797, 2009/0179998, 2009/0179999, 2009/0189997, 2009/0189998, 2009/0189998, 2009/0190803, 2009/0196466, 2009/0263022, 2009/0263022, 2009/0273685, 2009/0303342, 2009/0303342, 2009/0303343, 2010/0039502, 2009/0052748, 2009/0144173, 2008/0031327, 2007/0183651, 2006/0067573, 2005/0063582, PCT/US2006/021393; and
    • U.S. patent applications Nos. 60/829,127, 60/914,962, 61/019,370, 61/023,855, 61/221,467, 61/221,425, 61/221,417, 61/106,910, 61/182,625, 61/221,455, 61/091,700, and 61/120,289; and
    • Kampmann, M. [Markus], Ostermann, J. [Jörn], Automatic adaptation of a face model in a layered coder with an object-based analysis-synthesis layer and a knowledge-based layer, Signal Processing: Image Communication, (9), No. 3, March 1997, pp. 201-220.
    • Markus, Ostermann: Estimation of the Chin and Cheek Contours for Precise Face Model Adaptation, IEEE International Conference on Image Processing, '97 (III: 300-303).
    • Lee, K. S. [Kam-Sum], Wong, K. H. [Kin-Hong], Or, S. H. [Siu-Hang], Fung, Y. F. [Yiu-Fai], 3D Face Modeling from Perspective-Views and Contour-Based Generic-Model, Real Time Imaging, (7), No. 2, April 2001, pp. 173-182.
    • Grammalidis, N., Sarris, N., Varzokas, C., Strintzis, M. G., Generation of 3-d Head Models from Multiple Images Using Ellipsoid Approximation for the Rear Part, IEEE International Conference on Image Processing, '00 (Vol I: 284-287).
    • Sarris, N. [Nikos], Grammalidis, N. [Nikos], Strintzis, M. G. [Michael G.], Building Three Dimensional Head Models, GM(63), No. 5, September 2001, pp. 333-368.
    • Grammalidis, N., Sarris, N., Varzokas, C., Strintzis, M. G., Generation of 3-d Head Models from Multiple Images Using Ellipsoid Approximation for the Rear Part, ICIP00(Vol I: 284-287).
    • M. Kampmann, L. Zhang, Liang Zhang, Estimation of Eye, Eyebrow and Nose Features in Videophone Sequences, Proc. International Workshop on Very Low Bitrate Video Coding, 1998.
    • Yin, L. and Basu, A., Nose shape estimation and tracking for model-based coding, IEEE International Conference on Acoustics, Speech, and Signal Processing, 2001 Vol 3 (ISBN: 0-7803-7041-4).
    • Markus Kampmann, Segmentation of a Head into Face, Ears, Neck and Hair for Knowledge-Based Analysis-Synthesis Coding of Videophone Sequences, Int. Conf. on Image Processing, 1998.

Claims (25)

1. An image processing method using a 3D face model, comprising:
generating a stereoscopic image of a face, including using a dual-lens camera, or using a method including moving a camera relative to the face to capture facial images from more than one perspective, or applying a depth from defocus process including capturing at least two differently focused images of an approximately same scene, or combinations thereof;
creating a depth map based on the stereoscopic image;
generating a 3D face model of the face region from the stereoscopic image and the depth map; and
applying the 3D face model to process an image.
2. The method of claim 1, further comprising applying a foreground/background separation operation, wherein the modeled face comprises a foreground region.
3. The method of claim 1, further comprising applying progressive blurring to the face region based on distances of different portions of the face model from the camera as determined from either the depth map, or the 3D model, or both.
4. The method of claim 3, further comprising applying selective blurring to the face based on a combination of distance from the camera and the type of face region.
5. The method of claim 4, wherein the type of face region comprises a hair region, one or both eyes, a nose or nose region, a mouth or mouth region, a cheek portion, a chin or chin region, or combinations thereof.
6. The method of claim 3, further comprising applying selective blurring to the face based on a combination of distance from the camera and the type of face region.
7. The method of claim 1, wherein said 3D face model comprises a first set of one or more illumination components corresponding to a frontally illuminated face and a second set of one or more illumination components corresponding to a directionally illuminated face.
8. The method of claim 7, further comprising applying a foreground/background separation operation, wherein the modeled face comprises a foreground region.
9. The method of claim 7, further comprising applying progressive directional illumination to the face based on distances of different portions of the face from the camera as determined from the depth map or the 3D model, or both.
10. The method of claim 7, further comprising applying selective directional illumination to the face based on a combination of distance from the camera and type of face region.
11. The method of claim 10, wherein the type of face region comprises a hair region, one or both eyes, a nose or nose region, a mouth or mouth region, a cheek portion, a chin or chin region, or combinations thereof.
12. A method of determining a characteristic of a face within a scene captured in a digital image, comprising:
acquiring digital images from at least two perspectives including a face within a scene, and generating a stereoscopic image based thereon;
generating and applying a 3D face model based on the stereoscopic image, the 3D face model comprising a class of objects including a set of model components;
obtaining a fit of said model to said face including adjusting one or more individual values of one or more of the model components of said 3D face model;
based on the obtained fit of the model to said face in the scene, determining at least one characteristic of the face; and
electronically storing, transmitting, applying face recognition to, editing, or displaying a modified version of at least one of the digital images or a 3D image based on the acquired digital images including the determined characteristic or a modified value thereof, or combinations thereof.
13. The method of claim 12, wherein the model components comprise eigenvectors, and the individual values comprise eigenvalues of the eigenvectors.
14. The method of claim 12, wherein the at least one determined characteristic comprises a feature that is independent of directional lighting.
15. The method of claim 12, further comprising determining an exposure value for the face, including obtaining a fit to the face to a second 3D model that comprises a class of objects including a set of model components that exhibit a dependency on exposure value variations.
16. The method of claim 15, further comprising reducing an effect of a background region or density contrast caused by shadow, or both.
17. The method of claim 12, further comprising controlling a flash to accurately reflect a lighting condition, including obtaining a flash control condition by referring to a reference table and controlling a flash light emission according to the flash control condition.
18. The method of claim 17, further comprising reducing an effect of contrasting density caused by shadow or black compression or white compression or combinations thereof.
19. The method of claim 12, wherein the set of model components comprises a first subset of model components that exhibit a dependency on directional lighting variations and a second subset of model components which are independent of directional lighting variations.
20. The method of claim 19, further comprising applying a foreground/background separation operation, wherein the modeled face comprises a foreground region.
21. The method of claim 19, further comprising applying progressive directional illumination to the face based on distances of different portions of the face from the camera as determined from the depth map or the 3D model, or both.
22. The method of claim 19, further comprising applying selective directional illumination to the face based on a combination of distance from the camera and type of face region.
23. The method of claim 22, wherein the type of face region comprises a hair region, one or both eyes, a nose or nose region, a mouth or mouth region, a cheek portion, a chin or chin region, or combinations thereof.
24. A digital image acquisition device including an optoelectonic system for acquiring a digital image, and a digital memory having stored therein processor-readable code for programming the processor to perform a method as recited at any of claims 1-23.
25. One or more computer readable storage media having code embedded therein for programming a processor to perform a method as recited at any of claims 1-23.
US12/824,204 2004-12-29 2010-06-27 Enhanced real-time face models from stereo imaging Abandoned US20110102553A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US12/824,204 US20110102553A1 (en) 2007-02-28 2010-06-27 Enhanced real-time face models from stereo imaging
US14/673,246 US9639775B2 (en) 2004-12-29 2015-03-30 Face or other object detection including template matching

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US89223807P 2007-02-28 2007-02-28
US12/038,147 US8509561B2 (en) 2007-02-28 2008-02-27 Separating directional lighting variability in statistical face modelling based on texture space decomposition
US22142509P 2009-06-29 2009-06-29
US12/824,204 US20110102553A1 (en) 2007-02-28 2010-06-27 Enhanced real-time face models from stereo imaging

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12/038,147 Continuation-In-Part US8509561B2 (en) 2004-12-29 2008-02-27 Separating directional lighting variability in statistical face modelling based on texture space decomposition

Publications (1)

Publication Number Publication Date
US20110102553A1 true US20110102553A1 (en) 2011-05-05

Family

ID=43925007

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/824,204 Abandoned US20110102553A1 (en) 2004-12-29 2010-06-27 Enhanced real-time face models from stereo imaging

Country Status (1)

Country Link
US (1) US20110102553A1 (en)

Cited By (104)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110141300A1 (en) * 2009-12-11 2011-06-16 Fotonation Ireland Limited Panorama Imaging Using a Blending Map
US20110141225A1 (en) * 2009-12-11 2011-06-16 Fotonation Ireland Limited Panorama Imaging Based on Low-Res Images
US20110141229A1 (en) * 2009-12-11 2011-06-16 Fotonation Ireland Limited Panorama imaging using super-resolution
US20120007939A1 (en) * 2010-07-06 2012-01-12 Tessera Technologies Ireland Limited Scene Background Blurring Including Face Modeling
US20120121126A1 (en) * 2010-11-17 2012-05-17 Samsung Electronics Co., Ltd. Method and apparatus for estimating face position in 3 dimensions
US20120206458A1 (en) * 2011-02-10 2012-08-16 Edge 3 Technologies, Inc. Near-Touch Interaction with a Stereo Camera Grid Structured Tessellations
US20120219180A1 (en) * 2011-02-25 2012-08-30 DigitalOptics Corporation Europe Limited Automatic Detection of Vertical Gaze Using an Embedded Imaging Device
EP2515526A2 (en) 2011-04-08 2012-10-24 DigitalOptics Corporation Europe Limited Display device with image capture and analysis module
US20120268571A1 (en) * 2011-04-19 2012-10-25 University Of Southern California Multiview face capture using polarized spherical gradient illumination
US20120306991A1 (en) * 2011-06-06 2012-12-06 Cisco Technology, Inc. Diminishing an Appearance of a Double Chin in Video Communications
EP2538388A1 (en) * 2011-06-20 2012-12-26 Alcatel Lucent Method and arrangement for image model construction
JP2013097588A (en) * 2011-11-01 2013-05-20 Dainippon Printing Co Ltd Three-dimensional portrait creation device
CN103218612A (en) * 2013-05-13 2013-07-24 苏州福丰科技有限公司 3D (Three-Dimensional) face recognition method
US8509561B2 (en) 2007-02-28 2013-08-13 DigitalOptics Corporation Europe Limited Separating directional lighting variability in statistical face modelling based on texture space decomposition
US20130215112A1 (en) * 2012-02-17 2013-08-22 Etron Technology, Inc. Stereoscopic Image Processor, Stereoscopic Image Interaction System, and Stereoscopic Image Displaying Method thereof
WO2013136053A1 (en) 2012-03-10 2013-09-19 Digitaloptics Corporation Miniature camera module with mems-actuated autofocus
US8587665B2 (en) 2011-02-15 2013-11-19 DigitalOptics Corporation Europe Limited Fast rotation estimation of objects in sequences of acquired digital images
US8587666B2 (en) 2011-02-15 2013-11-19 DigitalOptics Corporation Europe Limited Object detection from image profiles within sequences of acquired digital images
US20140022249A1 (en) * 2012-07-12 2014-01-23 Cywee Group Limited Method of 3d model morphing driven by facial tracking and electronic device using the method the same
WO2014033099A2 (en) 2012-08-27 2014-03-06 Digital Optics Corporation Europe Limited Rearview imaging systems for vehicle
US20140064579A1 (en) * 2012-08-29 2014-03-06 Electronics And Telecommunications Research Institute Apparatus and method for generating three-dimensional face model for skin analysis
US8705894B2 (en) 2011-02-15 2014-04-22 Digital Optics Corporation Europe Limited Image rotation from local motion estimates
US20140118507A1 (en) * 2012-10-26 2014-05-01 Korea Advanced Institute Of Science And Technology Apparatus and method for depth manipulation of stereoscopic 3d image
US8723959B2 (en) 2011-03-31 2014-05-13 DigitalOptics Corporation Europe Limited Face and other object tracking in off-center peripheral regions for nonlinear lens geometries
WO2014072837A2 (en) 2012-06-07 2014-05-15 DigitalOptics Corporation Europe Limited Mems fast focus camera module
US8750578B2 (en) 2008-01-29 2014-06-10 DigitalOptics Corporation Europe Limited Detecting facial expressions in digital images
US20140176548A1 (en) * 2012-12-21 2014-06-26 Nvidia Corporation Facial image enhancement for video communication
US20140200417A1 (en) * 2010-06-07 2014-07-17 Affectiva, Inc. Mental state analysis using blink rate
US20140285637A1 (en) * 2013-03-20 2014-09-25 Mediatek Inc. 3d image capture method with 3d preview of preview images generated by monocular camera and related electronic device thereof
US8948461B1 (en) * 2005-04-29 2015-02-03 Hewlett-Packard Development Company, L.P. Method and system for estimating the three dimensional position of an object in a three dimensional physical space
US20150042760A1 (en) * 2013-08-06 2015-02-12 Htc Corporation Image processing methods and systems in accordance with depth information
US20150042840A1 (en) * 2013-08-12 2015-02-12 Canon Kabushiki Kaisha Image processing apparatus, distance measuring apparatus, imaging apparatus, and image processing method
US8982180B2 (en) 2011-03-31 2015-03-17 Fotonation Limited Face and other object detection and tracking in off-center peripheral regions for nonlinear lens geometries
US8995715B2 (en) 2010-10-26 2015-03-31 Fotonation Limited Face or other object detection including template matching
US9001268B2 (en) 2012-08-10 2015-04-07 Nan Chang O-Film Optoelectronics Technology Ltd Auto-focus camera module with flexible printed circuit extension
US9007520B2 (en) 2012-08-10 2015-04-14 Nanchang O-Film Optoelectronics Technology Ltd Camera module with EMI shield
WO2015066628A1 (en) * 2013-11-04 2015-05-07 Facebook, Inc. Systems and methods for facial representation
TWI485654B (en) * 2012-06-28 2015-05-21 Imec Taiwan Co Imaging system and method
US9087408B2 (en) 2011-08-16 2015-07-21 Google Inc. Systems and methods for generating depthmaps
US20150206354A1 (en) * 2012-08-30 2015-07-23 Sharp Kabushiki Kaisha Image processing apparatus and image display apparatus
US20150220773A1 (en) * 2012-05-04 2015-08-06 Commonwealth Scientific And Industrial Research Organisation System and method for eye alignment in video
US20150228081A1 (en) * 2014-02-10 2015-08-13 Electronics And Telecommunications Research Institute Method and apparatus for reconstructing 3d face with stereo camera
US9208608B2 (en) 2012-05-23 2015-12-08 Glasses.Com, Inc. Systems and methods for feature tracking
US9236024B2 (en) 2011-12-06 2016-01-12 Glasses.Com Inc. Systems and methods for obtaining a pupillary distance measurement using a mobile computing device
US20160011322A1 (en) * 2014-07-11 2016-01-14 Korea Atomic Energy Research Institute Symmetrical-type mono-sensor three-dimensional radiation detection and visualization system and method thereof
US20160019688A1 (en) * 2014-07-18 2016-01-21 University Of Georgia Research Foundation, Inc. Method and system of estimating produce characteristics
US20160026878A1 (en) * 2014-07-23 2016-01-28 GM Global Technology Operations LLC Algorithm to extend detecting range for avm stop line detection
US9251562B1 (en) * 2011-08-04 2016-02-02 Amazon Technologies, Inc. Registration of low contrast images
US9265458B2 (en) 2012-12-04 2016-02-23 Sync-Think, Inc. Application of smooth pursuit cognitive testing paradigms to clinical drug development
US9286715B2 (en) 2012-05-23 2016-03-15 Glasses.Com Inc. Systems and methods for adjusting a virtual try-on
US20160191788A1 (en) * 2013-01-31 2016-06-30 Canon Kabushiki Kaisha Image processing apparatus and image pickup apparatus
US9380976B2 (en) 2013-03-11 2016-07-05 Sync-Think, Inc. Optical neuroinformatics
US20160205389A1 (en) * 2015-01-13 2016-07-14 Boe Technology Group Co., Ltd. Control method and a control apparatus for a naked eye 3d display apparatus and a naked eye 3d display apparatus
US20160316185A1 (en) * 2015-04-27 2016-10-27 Microsoft Technology Licensing, Llc Trigger zones for objects in projected surface model
US20160316113A1 (en) * 2015-04-27 2016-10-27 Microsoft Technology Licensing, Llc Integrated processing and projection device with object detection
US9483853B2 (en) 2012-05-23 2016-11-01 Glasses.Com Inc. Systems and methods to display rendered images
US20160330469A1 (en) * 2015-05-04 2016-11-10 Ati Technologies Ulc Methods and apparatus for optical blur modeling for improved video encoding
WO2016196275A1 (en) * 2015-05-29 2016-12-08 Indiana University Research And Technology Corporation Method and apparatus for 3d facial recognition
US9525807B2 (en) 2010-12-01 2016-12-20 Nan Chang O-Film Optoelectronics Technology Ltd Three-pole tilt control system for camera module
US20170163876A1 (en) * 2012-05-17 2017-06-08 Canon Kabushiki Kaisha Image processing apparatus, image processing method, image processing program, and image pickup apparatus acquiring a focusing distance from a plurality of images
US20170185825A1 (en) * 2014-06-23 2017-06-29 British Telecommunications Public Limited Company Biometric identification
US9817206B2 (en) 2012-03-10 2017-11-14 Digitaloptics Corporation MEMS auto focus miniature camera module with fixed and movable lens groups
CN107563359A (en) * 2017-09-29 2018-01-09 重庆市智权之路科技有限公司 Recognition of face temperature analysis generation method is carried out for dense population
CN107909045A (en) * 2017-11-24 2018-04-13 合肥博焱智能科技有限公司 Face identification system based on FPGA
US10080006B2 (en) 2009-12-11 2018-09-18 Fotonation Limited Stereoscopic (3D) panorama creation on handheld device
US10101636B2 (en) 2012-12-31 2018-10-16 Digitaloptics Corporation Auto-focus camera module with MEMS capacitance estimator
CN109165487A (en) * 2018-06-28 2019-01-08 努比亚技术有限公司 A kind of method, mobile terminal and the computer readable storage medium of face unlock
US20190012578A1 (en) * 2017-07-07 2019-01-10 Carnegie Mellon University 3D Spatial Transformer Network
US10282623B1 (en) * 2015-09-25 2019-05-07 Apple Inc. Depth perception sensor data processing
US20190143221A1 (en) * 2017-11-15 2019-05-16 Sony Interactive Entertainment America Llc Generation and customization of personalized avatars
US10310060B2 (en) 2015-11-06 2019-06-04 Artilux Corporation High-speed light sensing apparatus
US10372973B2 (en) 2014-06-23 2019-08-06 British Telecommunications Public Limited Company Biometric identification
US10418407B2 (en) 2015-11-06 2019-09-17 Artilux, Inc. High-speed light sensing apparatus III
US10564718B2 (en) * 2015-08-04 2020-02-18 Artilux, Inc. Eye gesture tracking
US10615219B2 (en) 2015-07-23 2020-04-07 Artilux, Inc. High efficiency wide spectrum sensor
US10620454B2 (en) 2017-12-22 2020-04-14 Optikam Tech, Inc. System and method of obtaining fit and fabrication measurements for eyeglasses using simultaneous localization and mapping of camera images
US10628998B2 (en) 2013-10-25 2020-04-21 Onevisage Sa System and method for three dimensional object reconstruction and quality monitoring
US10643383B2 (en) 2017-11-27 2020-05-05 Fotonation Limited Systems and methods for 3D facial modeling
US10685994B2 (en) 2015-08-04 2020-06-16 Artilux, Inc. Germanium-silicon light sensing apparatus
US10707260B2 (en) 2015-08-04 2020-07-07 Artilux, Inc. Circuit for operating a multi-gate VIS/IR photodiode
US10739443B2 (en) 2015-11-06 2020-08-11 Artilux, Inc. High-speed light sensing apparatus II
US10741598B2 (en) 2015-11-06 2020-08-11 Atrilux, Inc. High-speed light sensing apparatus II
US10770504B2 (en) 2015-08-27 2020-09-08 Artilux, Inc. Wide spectrum optical sensor
US10777692B2 (en) 2018-02-23 2020-09-15 Artilux, Inc. Photo-detecting apparatus and photo-detecting method thereof
US10796480B2 (en) 2015-08-14 2020-10-06 Metail Limited Methods of generating personalized 3D head models or 3D body models
US10854770B2 (en) 2018-05-07 2020-12-01 Artilux, Inc. Avalanche photo-transistor
US10861888B2 (en) 2015-08-04 2020-12-08 Artilux, Inc. Silicon germanium imager with photodiode in trench
US10886309B2 (en) 2015-11-06 2021-01-05 Artilux, Inc. High-speed light sensing apparatus II
US10886311B2 (en) 2018-04-08 2021-01-05 Artilux, Inc. Photo-detecting apparatus
US10893231B1 (en) 2020-04-14 2021-01-12 International Business Machines Corporation Eye contact across digital mediums
US10969877B2 (en) 2018-05-08 2021-04-06 Artilux, Inc. Display apparatus
US11238302B2 (en) 2018-08-01 2022-02-01 Samsung Electronics Co., Ltd. Method and an apparatus for performing object illumination manipulation on an image
US11270110B2 (en) 2019-09-17 2022-03-08 Boston Polarimetrics, Inc. Systems and methods for surface modeling using polarization cues
US11290658B1 (en) 2021-04-15 2022-03-29 Boston Polarimetrics, Inc. Systems and methods for camera exposure control
US11302012B2 (en) 2019-11-30 2022-04-12 Boston Polarimetrics, Inc. Systems and methods for transparent object segmentation using polarization cues
US11315224B2 (en) * 2019-02-20 2022-04-26 Samsung Electronics Co., Ltd. Electronic device applying bokeh effect to image and controlling method thereof
US11393256B2 (en) * 2018-12-29 2022-07-19 Beijing Sensetime Technology Development Co., Ltd. Method and device for liveness detection, and storage medium
US11525906B2 (en) 2019-10-07 2022-12-13 Intrinsic Innovation Llc Systems and methods for augmentation of sensor systems and imaging systems with polarization
US11579472B2 (en) 2017-12-22 2023-02-14 Optikam Tech, Inc. System and method of obtaining fit and fabrication measurements for eyeglasses using depth map scanning
US11580667B2 (en) 2020-01-29 2023-02-14 Intrinsic Innovation Llc Systems and methods for characterizing object pose detection and measurement systems
US11630212B2 (en) 2018-02-23 2023-04-18 Artilux, Inc. Light-sensing apparatus and light-sensing method thereof
US11689813B2 (en) 2021-07-01 2023-06-27 Intrinsic Innovation Llc Systems and methods for high dynamic range imaging using crossed polarizers
US11797863B2 (en) 2020-01-30 2023-10-24 Intrinsic Innovation Llc Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images
US11954886B2 (en) 2021-04-15 2024-04-09 Intrinsic Innovation Llc Systems and methods for six-degree of freedom pose estimation of deformable objects

Citations (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5572596A (en) * 1994-09-02 1996-11-05 David Sarnoff Research Center, Inc. Automated, non-invasive iris recognition system and method
US5852672A (en) * 1995-07-10 1998-12-22 The Regents Of The University Of California Image system for three dimensional, 360 DEGREE, time sequence surface mapping of moving objects
US6072903A (en) * 1997-01-07 2000-06-06 Kabushiki Kaisha Toshiba Image processing apparatus and image processing method
US6301440B1 (en) * 2000-04-13 2001-10-09 International Business Machines Corp. System and method for automatically setting image acquisition controls
US20010038714A1 (en) * 2000-04-25 2001-11-08 Daiki Masumoto Picture recognition apparatus and method
US6456737B1 (en) * 1997-04-15 2002-09-24 Interval Research Corporation Data processing system and method
US20030068100A1 (en) * 2001-07-17 2003-04-10 Covell Michele M. Automatic selection of a visual image or images from a collection of visual images, based on an evaluation of the quality of the visual images
US20030160879A1 (en) * 2002-02-28 2003-08-28 Robins Mark Nelson White eye portraiture system and method
US20030190090A1 (en) * 2002-04-09 2003-10-09 Beeman Edward S. System and method for digital-image enhancement
US20040088272A1 (en) * 2002-11-01 2004-05-06 Nebojsa Jojic Method and apparatus for fast machine learning using probability maps and fourier transforms
US20040197013A1 (en) * 2001-12-14 2004-10-07 Toshio Kamei Face meta-data creation and face similarity calculation
US20040213482A1 (en) * 1997-07-12 2004-10-28 Kia Silverbrook Method of capturing and processing sensed images
US20040223629A1 (en) * 2003-05-06 2004-11-11 Viswis, Inc. Facial surveillance system and method
US20050018925A1 (en) * 2003-05-29 2005-01-27 Vijayakumar Bhagavatula Reduced complexity correlation filters
US20050068452A1 (en) * 2003-09-30 2005-03-31 Eran Steinberg Digital camera with built-in lens calibration table
US20050102246A1 (en) * 2003-07-24 2005-05-12 Movellan Javier R. Weak hypothesis generation apparatus and method, learning apparatus and method, detection apparatus and method, facial expression learning apparatus and method, facial expression recognition apparatus and method, and robot apparatus
US20050105805A1 (en) * 2003-11-13 2005-05-19 Eastman Kodak Company In-plane rotation invariant object detection in digitized images
US6920283B2 (en) * 2001-07-30 2005-07-19 Hewlett-Packard Development Company, L.P. System and method for controlling electronic devices
US7103225B2 (en) * 2002-11-07 2006-09-05 Honda Motor Co., Ltd. Clustering appearances of objects under varying illumination conditions
US20060210256A1 (en) * 2003-03-28 2006-09-21 Satoshi Fukui Photographing apparatus photographing method and computer program
US20060257047A1 (en) * 2005-05-11 2006-11-16 Fuji Photo Film Co., Ltd. Image processing apparatus, image processing method, and image processing program
US20060269270A1 (en) * 2005-05-11 2006-11-30 Fuji Photo Film Co., Ltd. Photography apparatus, photography method and photography program
US20060268150A1 (en) * 2005-05-16 2006-11-30 Fuji Photo Film Co., Ltd. Photography apparatus, photography method, and photography program
US20060280380A1 (en) * 2005-06-14 2006-12-14 Fuji Photo Film Co., Ltd. Apparatus, method, and program for image processing
US20060291739A1 (en) * 2005-06-24 2006-12-28 Fuji Photo Film Co., Ltd. Apparatus, method and program for image processing
US20070053590A1 (en) * 2005-09-05 2007-03-08 Tatsuo Kozakaya Image recognition apparatus and its method
US20070071347A1 (en) * 2005-09-26 2007-03-29 Fuji Photo Film Co., Ltd. Image processing method, image processing apparatus, and computer-readable recording medium storing image processing program
US20070070440A1 (en) * 2005-09-27 2007-03-29 Fuji Photo Film Co., Ltd. Image processing method, image processing apparatus, and computer-readable recording medium storing image processing program
US20070075996A1 (en) * 2005-10-03 2007-04-05 Konica Minolta Holdings, Inc. Modeling system, and modeling method and program
US20070216777A1 (en) * 2006-03-17 2007-09-20 Shuxue Quan Systems, methods, and apparatus for exposure control
US7324688B2 (en) * 2005-02-14 2008-01-29 Mitsubishi Electric Research Laboratories, Inc. Face relighting for normalization of directional lighting
US20080175446A1 (en) * 2006-08-28 2008-07-24 Colorado State University Research Foundation Set to set pattern recognition
US7412081B2 (en) * 2002-09-27 2008-08-12 Kabushiki Kaisha Toshiba Personal authentication apparatus and personal authentication method
US20080205712A1 (en) * 2007-02-28 2008-08-28 Fotonation Vision Limited Separating Directional Lighting Variability in Statistical Face Modelling Based on Texture Space Decomposition
US20090190803A1 (en) * 2008-01-29 2009-07-30 Fotonation Ireland Limited Detecting facial expressions in digital images
US7792335B2 (en) * 2006-02-24 2010-09-07 Fotonation Vision Limited Method and apparatus for selective disqualification of digital images
US7804983B2 (en) * 2006-02-24 2010-09-28 Fotonation Vision Limited Digital image acquisition control and correction method and apparatus
US20120219180A1 (en) * 2011-02-25 2012-08-30 DigitalOptics Corporation Europe Limited Automatic Detection of Vertical Gaze Using an Embedded Imaging Device

Patent Citations (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5572596A (en) * 1994-09-02 1996-11-05 David Sarnoff Research Center, Inc. Automated, non-invasive iris recognition system and method
US5852672A (en) * 1995-07-10 1998-12-22 The Regents Of The University Of California Image system for three dimensional, 360 DEGREE, time sequence surface mapping of moving objects
US6072903A (en) * 1997-01-07 2000-06-06 Kabushiki Kaisha Toshiba Image processing apparatus and image processing method
US6456737B1 (en) * 1997-04-15 2002-09-24 Interval Research Corporation Data processing system and method
US20040213482A1 (en) * 1997-07-12 2004-10-28 Kia Silverbrook Method of capturing and processing sensed images
US6301440B1 (en) * 2000-04-13 2001-10-09 International Business Machines Corp. System and method for automatically setting image acquisition controls
US20010038714A1 (en) * 2000-04-25 2001-11-08 Daiki Masumoto Picture recognition apparatus and method
US20030068100A1 (en) * 2001-07-17 2003-04-10 Covell Michele M. Automatic selection of a visual image or images from a collection of visual images, based on an evaluation of the quality of the visual images
US6920283B2 (en) * 2001-07-30 2005-07-19 Hewlett-Packard Development Company, L.P. System and method for controlling electronic devices
US20040197013A1 (en) * 2001-12-14 2004-10-07 Toshio Kamei Face meta-data creation and face similarity calculation
US20030160879A1 (en) * 2002-02-28 2003-08-28 Robins Mark Nelson White eye portraiture system and method
US20030190090A1 (en) * 2002-04-09 2003-10-09 Beeman Edward S. System and method for digital-image enhancement
US7412081B2 (en) * 2002-09-27 2008-08-12 Kabushiki Kaisha Toshiba Personal authentication apparatus and personal authentication method
US20040088272A1 (en) * 2002-11-01 2004-05-06 Nebojsa Jojic Method and apparatus for fast machine learning using probability maps and fourier transforms
US7103225B2 (en) * 2002-11-07 2006-09-05 Honda Motor Co., Ltd. Clustering appearances of objects under varying illumination conditions
US20060210256A1 (en) * 2003-03-28 2006-09-21 Satoshi Fukui Photographing apparatus photographing method and computer program
US20040223629A1 (en) * 2003-05-06 2004-11-11 Viswis, Inc. Facial surveillance system and method
US20050018925A1 (en) * 2003-05-29 2005-01-27 Vijayakumar Bhagavatula Reduced complexity correlation filters
US20050102246A1 (en) * 2003-07-24 2005-05-12 Movellan Javier R. Weak hypothesis generation apparatus and method, learning apparatus and method, detection apparatus and method, facial expression learning apparatus and method, facial expression recognition apparatus and method, and robot apparatus
US20050068452A1 (en) * 2003-09-30 2005-03-31 Eran Steinberg Digital camera with built-in lens calibration table
US7590305B2 (en) * 2003-09-30 2009-09-15 Fotonation Vision Limited Digital camera with built-in lens calibration table
US20050105805A1 (en) * 2003-11-13 2005-05-19 Eastman Kodak Company In-plane rotation invariant object detection in digitized images
US7324688B2 (en) * 2005-02-14 2008-01-29 Mitsubishi Electric Research Laboratories, Inc. Face relighting for normalization of directional lighting
US20060257047A1 (en) * 2005-05-11 2006-11-16 Fuji Photo Film Co., Ltd. Image processing apparatus, image processing method, and image processing program
US20060269270A1 (en) * 2005-05-11 2006-11-30 Fuji Photo Film Co., Ltd. Photography apparatus, photography method and photography program
US20060268150A1 (en) * 2005-05-16 2006-11-30 Fuji Photo Film Co., Ltd. Photography apparatus, photography method, and photography program
US20060280380A1 (en) * 2005-06-14 2006-12-14 Fuji Photo Film Co., Ltd. Apparatus, method, and program for image processing
US20060291739A1 (en) * 2005-06-24 2006-12-28 Fuji Photo Film Co., Ltd. Apparatus, method and program for image processing
US20070053590A1 (en) * 2005-09-05 2007-03-08 Tatsuo Kozakaya Image recognition apparatus and its method
US20070071347A1 (en) * 2005-09-26 2007-03-29 Fuji Photo Film Co., Ltd. Image processing method, image processing apparatus, and computer-readable recording medium storing image processing program
US20070070440A1 (en) * 2005-09-27 2007-03-29 Fuji Photo Film Co., Ltd. Image processing method, image processing apparatus, and computer-readable recording medium storing image processing program
US20070075996A1 (en) * 2005-10-03 2007-04-05 Konica Minolta Holdings, Inc. Modeling system, and modeling method and program
US7804983B2 (en) * 2006-02-24 2010-09-28 Fotonation Vision Limited Digital image acquisition control and correction method and apparatus
US7792335B2 (en) * 2006-02-24 2010-09-07 Fotonation Vision Limited Method and apparatus for selective disqualification of digital images
US7995795B2 (en) * 2006-02-24 2011-08-09 Tessera Technologies Ireland Limited Method and apparatus for selective disqualification of digital images
US8005268B2 (en) * 2006-02-24 2011-08-23 Tessera Technologies Ireland Limited Digital image acquisition control and correction method and apparatus
US20110280446A1 (en) * 2006-02-24 2011-11-17 Tessera Technologies Ireland Limited Method and Apparatus for Selective Disqualification of Digital Images
US20110279700A1 (en) * 2006-02-24 2011-11-17 Tessera Technologies Ireland Limited Digital Image Acquisition Control and Correction Method and Apparatus
US20070216777A1 (en) * 2006-03-17 2007-09-20 Shuxue Quan Systems, methods, and apparatus for exposure control
US20080175446A1 (en) * 2006-08-28 2008-07-24 Colorado State University Research Foundation Set to set pattern recognition
US20080205712A1 (en) * 2007-02-28 2008-08-28 Fotonation Vision Limited Separating Directional Lighting Variability in Statistical Face Modelling Based on Texture Space Decomposition
US20090003661A1 (en) * 2007-02-28 2009-01-01 Fotonation Vision Limited Separating a Directional Lighting Variability In Statistical Face Modelling Based On Texture Space Decomposition
US20090190803A1 (en) * 2008-01-29 2009-07-30 Fotonation Ireland Limited Detecting facial expressions in digital images
US20120219180A1 (en) * 2011-02-25 2012-08-30 DigitalOptics Corporation Europe Limited Automatic Detection of Vertical Gaze Using an Embedded Imaging Device
US20120218398A1 (en) * 2011-02-25 2012-08-30 Tessera Technologies Ireland Limited Automatic Detection of Vertical Gaze Using an Embedded Imaging Device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Fransens et al., "Parametric stereo for multi-pose face recognition and 3D-face modeling," AMFG 2005, LNCS 3723, pp. 109-124, 2005 *

Cited By (177)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9639775B2 (en) 2004-12-29 2017-05-02 Fotonation Limited Face or other object detection including template matching
US8948461B1 (en) * 2005-04-29 2015-02-03 Hewlett-Packard Development Company, L.P. Method and system for estimating the three dimensional position of an object in a three dimensional physical space
US8509561B2 (en) 2007-02-28 2013-08-13 DigitalOptics Corporation Europe Limited Separating directional lighting variability in statistical face modelling based on texture space decomposition
US8582896B2 (en) 2007-02-28 2013-11-12 DigitalOptics Corporation Europe Limited Separating directional lighting variability in statistical face modelling based on texture space decomposition
US8565550B2 (en) 2007-02-28 2013-10-22 DigitalOptics Corporation Europe Limited Separating directional lighting variability in statistical face modelling based on texture space decomposition
US8542913B2 (en) 2007-02-28 2013-09-24 DigitalOptics Corporation Europe Limited Separating directional lighting variability in statistical face modelling based on texture space decomposition
US11470241B2 (en) 2008-01-27 2022-10-11 Fotonation Limited Detecting facial expressions in digital images
US11689796B2 (en) 2008-01-27 2023-06-27 Adeia Imaging Llc Detecting facial expressions in digital images
US9462180B2 (en) 2008-01-27 2016-10-04 Fotonation Limited Detecting facial expressions in digital images
US8750578B2 (en) 2008-01-29 2014-06-10 DigitalOptics Corporation Europe Limited Detecting facial expressions in digital images
US20110141229A1 (en) * 2009-12-11 2011-06-16 Fotonation Ireland Limited Panorama imaging using super-resolution
US11115638B2 (en) 2009-12-11 2021-09-07 Fotonation Limited Stereoscopic (3D) panorama creation on handheld device
US10080006B2 (en) 2009-12-11 2018-09-18 Fotonation Limited Stereoscopic (3D) panorama creation on handheld device
US20110141300A1 (en) * 2009-12-11 2011-06-16 Fotonation Ireland Limited Panorama Imaging Using a Blending Map
US20110141225A1 (en) * 2009-12-11 2011-06-16 Fotonation Ireland Limited Panorama Imaging Based on Low-Res Images
US8294748B2 (en) 2009-12-11 2012-10-23 DigitalOptics Corporation Europe Limited Panorama imaging using a blending map
US9723992B2 (en) * 2010-06-07 2017-08-08 Affectiva, Inc. Mental state analysis using blink rate
US20140200417A1 (en) * 2010-06-07 2014-07-17 Affectiva, Inc. Mental state analysis using blink rate
US8723912B2 (en) * 2010-07-06 2014-05-13 DigitalOptics Corporation Europe Limited Scene background blurring including face modeling
US20120007939A1 (en) * 2010-07-06 2012-01-12 Tessera Technologies Ireland Limited Scene Background Blurring Including Face Modeling
US8995715B2 (en) 2010-10-26 2015-03-31 Fotonation Limited Face or other object detection including template matching
US20120121126A1 (en) * 2010-11-17 2012-05-17 Samsung Electronics Co., Ltd. Method and apparatus for estimating face position in 3 dimensions
US8805021B2 (en) * 2010-11-17 2014-08-12 Samsung Electronics Co., Ltd. Method and apparatus for estimating face position in 3 dimensions
US9525807B2 (en) 2010-12-01 2016-12-20 Nan Chang O-Film Optoelectronics Technology Ltd Three-pole tilt control system for camera module
US10061442B2 (en) 2011-02-10 2018-08-28 Edge 3 Technologies, Inc. Near touch interaction
US9323395B2 (en) * 2011-02-10 2016-04-26 Edge 3 Technologies Near touch interaction with structured light
US20150130803A1 (en) * 2011-02-10 2015-05-14 Edge 3 Technologies, Inc. Near Touch Interaction with Structured Light
US20120206458A1 (en) * 2011-02-10 2012-08-16 Edge 3 Technologies, Inc. Near-Touch Interaction with a Stereo Camera Grid Structured Tessellations
US9652084B2 (en) * 2011-02-10 2017-05-16 Edge 3 Technologies, Inc. Near touch interaction
US8970589B2 (en) * 2011-02-10 2015-03-03 Edge 3 Technologies, Inc. Near-touch interaction with a stereo camera grid structured tessellations
US10599269B2 (en) 2011-02-10 2020-03-24 Edge 3 Technologies, Inc. Near touch interaction
US8587665B2 (en) 2011-02-15 2013-11-19 DigitalOptics Corporation Europe Limited Fast rotation estimation of objects in sequences of acquired digital images
US8705894B2 (en) 2011-02-15 2014-04-22 Digital Optics Corporation Europe Limited Image rotation from local motion estimates
US8587666B2 (en) 2011-02-15 2013-11-19 DigitalOptics Corporation Europe Limited Object detection from image profiles within sequences of acquired digital images
US8836777B2 (en) 2011-02-25 2014-09-16 DigitalOptics Corporation Europe Limited Automatic detection of vertical gaze using an embedded imaging device
US20120219180A1 (en) * 2011-02-25 2012-08-30 DigitalOptics Corporation Europe Limited Automatic Detection of Vertical Gaze Using an Embedded Imaging Device
US8982180B2 (en) 2011-03-31 2015-03-17 Fotonation Limited Face and other object detection and tracking in off-center peripheral regions for nonlinear lens geometries
US8723959B2 (en) 2011-03-31 2014-05-13 DigitalOptics Corporation Europe Limited Face and other object tracking in off-center peripheral regions for nonlinear lens geometries
EP2515526A2 (en) 2011-04-08 2012-10-24 DigitalOptics Corporation Europe Limited Display device with image capture and analysis module
US9123116B2 (en) * 2011-04-19 2015-09-01 University Of Southern California Multiview face capture using polarized spherical gradient illumination
US20120268571A1 (en) * 2011-04-19 2012-10-25 University Of Southern California Multiview face capture using polarized spherical gradient illumination
US20120306991A1 (en) * 2011-06-06 2012-12-06 Cisco Technology, Inc. Diminishing an Appearance of a Double Chin in Video Communications
US8687039B2 (en) * 2011-06-06 2014-04-01 Cisco Technology, Inc. Diminishing an appearance of a double chin in video communications
EP2538388A1 (en) * 2011-06-20 2012-12-26 Alcatel Lucent Method and arrangement for image model construction
CN103608847A (en) * 2011-06-20 2014-02-26 阿尔卡特朗讯 Method and arrangement for image model construction
WO2012175320A1 (en) * 2011-06-20 2012-12-27 Alcatel Lucent Method and arrangement for image model construction
US9324191B2 (en) 2011-06-20 2016-04-26 Alcatel Lucent Method and arrangement for image model construction
US9251562B1 (en) * 2011-08-04 2016-02-02 Amazon Technologies, Inc. Registration of low contrast images
US9530208B1 (en) 2011-08-04 2016-12-27 Amazon Technologies, Inc. Registration of low contrast images
US9087408B2 (en) 2011-08-16 2015-07-21 Google Inc. Systems and methods for generating depthmaps
JP2013097588A (en) * 2011-11-01 2013-05-20 Dainippon Printing Co Ltd Three-dimensional portrait creation device
US9236024B2 (en) 2011-12-06 2016-01-12 Glasses.Com Inc. Systems and methods for obtaining a pupillary distance measurement using a mobile computing device
US20130215112A1 (en) * 2012-02-17 2013-08-22 Etron Technology, Inc. Stereoscopic Image Processor, Stereoscopic Image Interaction System, and Stereoscopic Image Displaying Method thereof
WO2013136053A1 (en) 2012-03-10 2013-09-19 Digitaloptics Corporation Miniature camera module with mems-actuated autofocus
US9817206B2 (en) 2012-03-10 2017-11-14 Digitaloptics Corporation MEMS auto focus miniature camera module with fixed and movable lens groups
US9424463B2 (en) * 2012-05-04 2016-08-23 Commonwealth Scientific And Industrial Research Organisation System and method for eye alignment in video
US20150220773A1 (en) * 2012-05-04 2015-08-06 Commonwealth Scientific And Industrial Research Organisation System and method for eye alignment in video
US10021290B2 (en) * 2012-05-17 2018-07-10 Canon Kabushiki Kaisha Image processing apparatus, image processing method, image processing program, and image pickup apparatus acquiring a focusing distance from a plurality of images
US20170163876A1 (en) * 2012-05-17 2017-06-08 Canon Kabushiki Kaisha Image processing apparatus, image processing method, image processing program, and image pickup apparatus acquiring a focusing distance from a plurality of images
US9235929B2 (en) 2012-05-23 2016-01-12 Glasses.Com Inc. Systems and methods for efficiently processing virtual 3-D data
US9208608B2 (en) 2012-05-23 2015-12-08 Glasses.Com, Inc. Systems and methods for feature tracking
US9483853B2 (en) 2012-05-23 2016-11-01 Glasses.Com Inc. Systems and methods to display rendered images
US10147233B2 (en) 2012-05-23 2018-12-04 Glasses.Com Inc. Systems and methods for generating a 3-D model of a user for a virtual try-on product
US9286715B2 (en) 2012-05-23 2016-03-15 Glasses.Com Inc. Systems and methods for adjusting a virtual try-on
US9311746B2 (en) 2012-05-23 2016-04-12 Glasses.Com Inc. Systems and methods for generating a 3-D model of a virtual try-on product
US9378584B2 (en) 2012-05-23 2016-06-28 Glasses.Com Inc. Systems and methods for rendering virtual try-on products
WO2014072837A2 (en) 2012-06-07 2014-05-15 DigitalOptics Corporation Europe Limited Mems fast focus camera module
TWI485654B (en) * 2012-06-28 2015-05-21 Imec Taiwan Co Imaging system and method
US9262869B2 (en) * 2012-07-12 2016-02-16 UL See Inc. Method of 3D model morphing driven by facial tracking and electronic device using the method the same
US20140022249A1 (en) * 2012-07-12 2014-01-23 Cywee Group Limited Method of 3d model morphing driven by facial tracking and electronic device using the method the same
US9007520B2 (en) 2012-08-10 2015-04-14 Nanchang O-Film Optoelectronics Technology Ltd Camera module with EMI shield
US9001268B2 (en) 2012-08-10 2015-04-07 Nan Chang O-Film Optoelectronics Technology Ltd Auto-focus camera module with flexible printed circuit extension
WO2014033099A2 (en) 2012-08-27 2014-03-06 Digital Optics Corporation Europe Limited Rearview imaging systems for vehicle
US20140064579A1 (en) * 2012-08-29 2014-03-06 Electronics And Telecommunications Research Institute Apparatus and method for generating three-dimensional face model for skin analysis
US20150206354A1 (en) * 2012-08-30 2015-07-23 Sharp Kabushiki Kaisha Image processing apparatus and image display apparatus
US9445074B2 (en) * 2012-10-26 2016-09-13 Korea Advanced Institute Of Science And Technology Apparatus and method for depth manipulation of stereoscopic 3D image
US20140118507A1 (en) * 2012-10-26 2014-05-01 Korea Advanced Institute Of Science And Technology Apparatus and method for depth manipulation of stereoscopic 3d image
US9265458B2 (en) 2012-12-04 2016-02-23 Sync-Think, Inc. Application of smooth pursuit cognitive testing paradigms to clinical drug development
US20140176548A1 (en) * 2012-12-21 2014-06-26 Nvidia Corporation Facial image enhancement for video communication
US10101636B2 (en) 2012-12-31 2018-10-16 Digitaloptics Corporation Auto-focus camera module with MEMS capacitance estimator
US20160191788A1 (en) * 2013-01-31 2016-06-30 Canon Kabushiki Kaisha Image processing apparatus and image pickup apparatus
US9998650B2 (en) * 2013-01-31 2018-06-12 Canon Kabushiki Kaisha Image processing apparatus and image pickup apparatus for adding blur in an image according to depth map
US9380976B2 (en) 2013-03-11 2016-07-05 Sync-Think, Inc. Optical neuroinformatics
US20140285637A1 (en) * 2013-03-20 2014-09-25 Mediatek Inc. 3d image capture method with 3d preview of preview images generated by monocular camera and related electronic device thereof
US9967549B2 (en) * 2013-03-20 2018-05-08 Mediatek Inc. 3D image capture method with 3D preview of preview images generated by monocular camera and related electronic device thereof
CN103218612A (en) * 2013-05-13 2013-07-24 苏州福丰科技有限公司 3D (Three-Dimensional) face recognition method
US20150042760A1 (en) * 2013-08-06 2015-02-12 Htc Corporation Image processing methods and systems in accordance with depth information
US9445073B2 (en) * 2013-08-06 2016-09-13 Htc Corporation Image processing methods and systems in accordance with depth information
US20150042840A1 (en) * 2013-08-12 2015-02-12 Canon Kabushiki Kaisha Image processing apparatus, distance measuring apparatus, imaging apparatus, and image processing method
US9413952B2 (en) * 2013-08-12 2016-08-09 Canon Kabushiki Kaisha Image processing apparatus, distance measuring apparatus, imaging apparatus, and image processing method
US9811909B2 (en) * 2013-08-12 2017-11-07 Canon Kabushiki Kaisha Image processing apparatus, distance measuring apparatus, imaging apparatus, and image processing method
US20160314591A1 (en) * 2013-08-12 2016-10-27 Canon Kabushiki Kaisha Image processing apparatus, distance measuring apparatus, imaging apparatus, and image processing method
US10628998B2 (en) 2013-10-25 2020-04-21 Onevisage Sa System and method for three dimensional object reconstruction and quality monitoring
CN105874474A (en) * 2013-11-04 2016-08-17 脸谱公司 Systems and methods for facial representation
WO2015066628A1 (en) * 2013-11-04 2015-05-07 Facebook, Inc. Systems and methods for facial representation
US11210503B2 (en) 2013-11-04 2021-12-28 Facebook, Inc. Systems and methods for facial representation
US10095917B2 (en) 2013-11-04 2018-10-09 Facebook, Inc. Systems and methods for facial representation
US20150228081A1 (en) * 2014-02-10 2015-08-13 Electronics And Telecommunications Research Institute Method and apparatus for reconstructing 3d face with stereo camera
KR20150093972A (en) * 2014-02-10 2015-08-19 한국전자통신연구원 Method and apparatus for reconstructing 3d face with stereo camera
KR102135770B1 (en) * 2014-02-10 2020-07-20 한국전자통신연구원 Method and apparatus for reconstructing 3d face with stereo camera
US10043278B2 (en) * 2014-02-10 2018-08-07 Electronics And Telecommunications Research Institute Method and apparatus for reconstructing 3D face with stereo camera
US20170185825A1 (en) * 2014-06-23 2017-06-29 British Telecommunications Public Limited Company Biometric identification
US10289896B2 (en) * 2014-06-23 2019-05-14 British Telecommunications Public Limited Company Biometric identification
US10372973B2 (en) 2014-06-23 2019-08-06 British Telecommunications Public Limited Company Biometric identification
US20160011322A1 (en) * 2014-07-11 2016-01-14 Korea Atomic Energy Research Institute Symmetrical-type mono-sensor three-dimensional radiation detection and visualization system and method thereof
US9910164B2 (en) * 2014-07-11 2018-03-06 Korea Atomic Energy Research Institute Symmetrical-type mono-sensor three-dimensional radiation detection and visualization system and method thereof
US20160019688A1 (en) * 2014-07-18 2016-01-21 University Of Georgia Research Foundation, Inc. Method and system of estimating produce characteristics
US20160026878A1 (en) * 2014-07-23 2016-01-28 GM Global Technology Operations LLC Algorithm to extend detecting range for avm stop line detection
US10318824B2 (en) * 2014-07-23 2019-06-11 GM Global Technology Operations LLC Algorithm to extend detecting range for AVM stop line detection
US10321125B2 (en) * 2015-01-13 2019-06-11 Boe Technology Group Co., Ltd. Control method and a control apparatus for a naked eye 3D display apparatus and a naked eye 3D display apparatus
US20160205389A1 (en) * 2015-01-13 2016-07-14 Boe Technology Group Co., Ltd. Control method and a control apparatus for a naked eye 3d display apparatus and a naked eye 3d display apparatus
US20160316185A1 (en) * 2015-04-27 2016-10-27 Microsoft Technology Licensing, Llc Trigger zones for objects in projected surface model
US10306193B2 (en) * 2015-04-27 2019-05-28 Microsoft Technology Licensing, Llc Trigger zones for objects in projected surface model
US20160316113A1 (en) * 2015-04-27 2016-10-27 Microsoft Technology Licensing, Llc Integrated processing and projection device with object detection
US10979704B2 (en) * 2015-05-04 2021-04-13 Advanced Micro Devices, Inc. Methods and apparatus for optical blur modeling for improved video encoding
US20160330469A1 (en) * 2015-05-04 2016-11-10 Ati Technologies Ulc Methods and apparatus for optical blur modeling for improved video encoding
WO2016196275A1 (en) * 2015-05-29 2016-12-08 Indiana University Research And Technology Corporation Method and apparatus for 3d facial recognition
US11335725B2 (en) 2015-07-23 2022-05-17 Artilux, Inc. High efficiency wide spectrum sensor
US10615219B2 (en) 2015-07-23 2020-04-07 Artilux, Inc. High efficiency wide spectrum sensor
US11755104B2 (en) 2015-08-04 2023-09-12 Artilux, Inc. Eye gesture tracking
US10564718B2 (en) * 2015-08-04 2020-02-18 Artilux, Inc. Eye gesture tracking
US10964742B2 (en) 2015-08-04 2021-03-30 Artilux, Inc. Germanium-silicon light sensing apparatus II
US10756127B2 (en) 2015-08-04 2020-08-25 Artilux, Inc. Germanium-silicon light sensing apparatus
US10685994B2 (en) 2015-08-04 2020-06-16 Artilux, Inc. Germanium-silicon light sensing apparatus
US10707260B2 (en) 2015-08-04 2020-07-07 Artilux, Inc. Circuit for operating a multi-gate VIS/IR photodiode
US10861888B2 (en) 2015-08-04 2020-12-08 Artilux, Inc. Silicon germanium imager with photodiode in trench
US11756969B2 (en) 2015-08-04 2023-09-12 Artilux, Inc. Germanium-silicon light sensing apparatus
US10761599B2 (en) * 2015-08-04 2020-09-01 Artilux, Inc. Eye gesture tracking
US10796480B2 (en) 2015-08-14 2020-10-06 Metail Limited Methods of generating personalized 3D head models or 3D body models
US10770504B2 (en) 2015-08-27 2020-09-08 Artilux, Inc. Wide spectrum optical sensor
US10282623B1 (en) * 2015-09-25 2019-05-07 Apple Inc. Depth perception sensor data processing
US10353056B2 (en) 2015-11-06 2019-07-16 Artilux Corporation High-speed light sensing apparatus
US10739443B2 (en) 2015-11-06 2020-08-11 Artilux, Inc. High-speed light sensing apparatus II
US11637142B2 (en) 2015-11-06 2023-04-25 Artilux, Inc. High-speed light sensing apparatus III
US10795003B2 (en) 2015-11-06 2020-10-06 Artilux, Inc. High-speed light sensing apparatus
US10418407B2 (en) 2015-11-06 2019-09-17 Artilux, Inc. High-speed light sensing apparatus III
US11579267B2 (en) 2015-11-06 2023-02-14 Artilux, Inc. High-speed light sensing apparatus
US10310060B2 (en) 2015-11-06 2019-06-04 Artilux Corporation High-speed light sensing apparatus
US10886309B2 (en) 2015-11-06 2021-01-05 Artilux, Inc. High-speed light sensing apparatus II
US11749696B2 (en) 2015-11-06 2023-09-05 Artilux, Inc. High-speed light sensing apparatus II
US10886312B2 (en) 2015-11-06 2021-01-05 Artilux, Inc. High-speed light sensing apparatus II
US11131757B2 (en) 2015-11-06 2021-09-28 Artilux, Inc. High-speed light sensing apparatus
US10741598B2 (en) 2015-11-06 2020-08-11 Atrilux, Inc. High-speed light sensing apparatus II
US11747450B2 (en) 2015-11-06 2023-09-05 Artilux, Inc. High-speed light sensing apparatus
US20190012578A1 (en) * 2017-07-07 2019-01-10 Carnegie Mellon University 3D Spatial Transformer Network
US10755145B2 (en) * 2017-07-07 2020-08-25 Carnegie Mellon University 3D spatial transformer network
CN107563359A (en) * 2017-09-29 2018-01-09 重庆市智权之路科技有限公司 Recognition of face temperature analysis generation method is carried out for dense population
US20190143221A1 (en) * 2017-11-15 2019-05-16 Sony Interactive Entertainment America Llc Generation and customization of personalized avatars
CN107909045A (en) * 2017-11-24 2018-04-13 合肥博焱智能科技有限公司 Face identification system based on FPGA
US10643383B2 (en) 2017-11-27 2020-05-05 Fotonation Limited Systems and methods for 3D facial modeling
US11830141B2 (en) 2017-11-27 2023-11-28 Adela Imaging LLC Systems and methods for 3D facial modeling
US11257289B2 (en) 2017-11-27 2022-02-22 Fotonation Limited Systems and methods for 3D facial modeling
US11579472B2 (en) 2017-12-22 2023-02-14 Optikam Tech, Inc. System and method of obtaining fit and fabrication measurements for eyeglasses using depth map scanning
US10620454B2 (en) 2017-12-22 2020-04-14 Optikam Tech, Inc. System and method of obtaining fit and fabrication measurements for eyeglasses using simultaneous localization and mapping of camera images
US11630212B2 (en) 2018-02-23 2023-04-18 Artilux, Inc. Light-sensing apparatus and light-sensing method thereof
US10777692B2 (en) 2018-02-23 2020-09-15 Artilux, Inc. Photo-detecting apparatus and photo-detecting method thereof
US11329081B2 (en) 2018-04-08 2022-05-10 Artilux, Inc. Photo-detecting apparatus
US10886311B2 (en) 2018-04-08 2021-01-05 Artilux, Inc. Photo-detecting apparatus
US10854770B2 (en) 2018-05-07 2020-12-01 Artilux, Inc. Avalanche photo-transistor
US10969877B2 (en) 2018-05-08 2021-04-06 Artilux, Inc. Display apparatus
CN109165487A (en) * 2018-06-28 2019-01-08 努比亚技术有限公司 A kind of method, mobile terminal and the computer readable storage medium of face unlock
US11238302B2 (en) 2018-08-01 2022-02-01 Samsung Electronics Co., Ltd. Method and an apparatus for performing object illumination manipulation on an image
US11393256B2 (en) * 2018-12-29 2022-07-19 Beijing Sensetime Technology Development Co., Ltd. Method and device for liveness detection, and storage medium
US11315224B2 (en) * 2019-02-20 2022-04-26 Samsung Electronics Co., Ltd. Electronic device applying bokeh effect to image and controlling method thereof
US11699273B2 (en) 2019-09-17 2023-07-11 Intrinsic Innovation Llc Systems and methods for surface modeling using polarization cues
US11270110B2 (en) 2019-09-17 2022-03-08 Boston Polarimetrics, Inc. Systems and methods for surface modeling using polarization cues
US11525906B2 (en) 2019-10-07 2022-12-13 Intrinsic Innovation Llc Systems and methods for augmentation of sensor systems and imaging systems with polarization
US11842495B2 (en) 2019-11-30 2023-12-12 Intrinsic Innovation Llc Systems and methods for transparent object segmentation using polarization cues
US11302012B2 (en) 2019-11-30 2022-04-12 Boston Polarimetrics, Inc. Systems and methods for transparent object segmentation using polarization cues
US11580667B2 (en) 2020-01-29 2023-02-14 Intrinsic Innovation Llc Systems and methods for characterizing object pose detection and measurement systems
US11797863B2 (en) 2020-01-30 2023-10-24 Intrinsic Innovation Llc Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images
US10893231B1 (en) 2020-04-14 2021-01-12 International Business Machines Corporation Eye contact across digital mediums
US11290658B1 (en) 2021-04-15 2022-03-29 Boston Polarimetrics, Inc. Systems and methods for camera exposure control
US11683594B2 (en) 2021-04-15 2023-06-20 Intrinsic Innovation Llc Systems and methods for camera exposure control
US11954886B2 (en) 2021-04-15 2024-04-09 Intrinsic Innovation Llc Systems and methods for six-degree of freedom pose estimation of deformable objects
US11953700B2 (en) 2021-05-27 2024-04-09 Intrinsic Innovation Llc Multi-aperture polarization optical systems using beam splitters
US11689813B2 (en) 2021-07-01 2023-06-27 Intrinsic Innovation Llc Systems and methods for high dynamic range imaging using crossed polarizers

Similar Documents

Publication Publication Date Title
US20110102553A1 (en) Enhanced real-time face models from stereo imaging
US11830141B2 (en) Systems and methods for 3D facial modeling
Piotraschke et al. Automated 3d face reconstruction from multiple images using quality measures
Thies et al. Real-time expression transfer for facial reenactment.
US8326025B2 (en) Method for determining a depth map from images, device for determining a depth map
Wechsler Reliable Face Recognition Methods: System Design, Impementation and Evaluation
US8086027B2 (en) Image processing apparatus and method
Kadambi et al. 3d depth cameras in vision: Benefits and limitations of the hardware: With an emphasis on the first-and second-generation kinect models
US7103211B1 (en) Method and apparatus for generating 3D face models from one camera
Morency et al. Pose estimation using 3d view-based eigenspaces
US20130335535A1 (en) Digital 3d camera using periodic illumination
Strom et al. Real time tracking and modeling of faces: An ekf-based analysis by synthesis approach
KR20170008638A (en) Three dimensional content producing apparatus and three dimensional content producing method thereof
JP2011039869A (en) Face image processing apparatus and computer program
JP6207210B2 (en) Information processing apparatus and method
JPWO2006049147A1 (en) Three-dimensional shape estimation system and image generation system
US10229534B2 (en) Modeling of a user's face
KR20170079680A (en) Apparatus and method for synthesizing facial expression using weighted interpolation map
JP4539519B2 (en) Stereo model generation apparatus and stereo model generation method
Chen et al. Monogaussianavatar: Monocular gaussian point-based head avatar
KR101812664B1 (en) Method and apparatus for extracting multi-view object with fractional boundaries
EP4036858A1 (en) Volumetric imaging
Ishiyama et al. Fast and accurate facial pose estimation by aligning a 3D appearance model
CN112990047A (en) Multi-pose face verification method combining face angle information
Bartoli et al. Augmenting images of non-rigid scenes using point and curve correspondences

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION