WO1999017254A1 - Digital redeye removal - Google Patents
Digital redeye removal Download PDFInfo
- Publication number
- WO1999017254A1 WO1999017254A1 PCT/US1998/019428 US9819428W WO9917254A1 WO 1999017254 A1 WO1999017254 A1 WO 1999017254A1 US 9819428 W US9819428 W US 9819428W WO 9917254 A1 WO9917254 A1 WO 9917254A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- redeye
- pixels
- hue
- luminance
- values
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/20—Image enhancement or restoration by the use of local operators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/193—Preprocessing; Feature extraction
Definitions
- redeye effect causes pupils to appear red rather than black, brown, blue, green or gray. It is generally accepted that the redeye effect is caused by light from
- the electronic flash that is reflected from the retina of the subject's eye.
- light from the flash is imaged at the retina where a fraction of it is reflected back out the pupil of the eye. If the eye were
- the redeye effect can be avoided by locating the flash unit some distance (perhaps six
- a system and method for correcting redeye in a digital image includes the functions or
- At least one pixel of the image for comparison with the threshold values; identifying redeye pixels of the image by comparing luminance, hue and saturation values which are derived from color
- weighted functions of the color components subject to a first set of color weights identifying edge pixels as the redeye pixels which border non-redeye pixels; and smoothing each redeye region by
- Figure 1 is a diagram of an image acquisition and rendering system including an image
- FIG. 2 is a block diagram of a preferred sequence of steps of the inventive method.
- Figure 3 is a representation of a redeye region composed of redeye pixels.
- a digital image consists of pixels each containing color and luminance information which
- the pixel information can be resident in RGB (red, green and blue) device-dependent color space having R, G and B color components for use with a color monitor, or in CMYK (cyan, magenta, yellow black) device-dependent color space having C, M, Y and K color components for use with a color printer.
- RGB red, green and blue
- CMYK cyan, magenta, yellow black
- the properties of a pixel are represented in RGB color space and the pixel values are transformed into YIQ color space as defined in the following equations (1) through (5) where, for each pixel, Y represents luminance, I and Q represent chrominance, H represents hue and S represents saturation.
- Maximum and minimum threshold values of luminance, hue and saturation for defining a redeye region of a pixel are determined in block 30 of Figure 2 by testing a database of images
- MaxLum represents the maximum luminance threshold value
- MinLum represents the- minimum luminance threshold value
- MaxHue represents the maximum hue threshold value
- MinHue represents the minimum hue threshold value
- MaxSat represents the maximum saturation
- MinSat represents the minimum saturation threshold value. Threshold ranges having both upper and lower limits are necessary for preventing spurious redeye readings.
- threshold values are preferably precalculated and prestored in a configuration file, database, look-up table or other memory structure prior to
- the above threshold values may be varied if
- One or more pixels of the image are selected in block 32 for comparison against the
- predetermined threshold values MaxLum, MinLum, MaxHue, MinHue, MaxSat and MinSat.
- comparison starts with one seed pixel and continues in some predetermined pattern until the luminance, hue and saturation properties of all pixels for a given area (which could be a part, or the
- redeye pixel i.e. a pixel which exhibits the redeye effect.
- a pixel at a spatial location (x,y) could be flagged or labeled as a redeye pixel by setting
- Flags(x,y) the appropriate element in the array Flags(x,y) equal to 3 whereas a non-redeye pixel could be indicated by setting Flags(x,y) equal to 0.
- Redeye regions are identified in block 36 as regions of the image having one or more adjacent redeye pixels which are surrounded by non-redeye pixels. After all redeye pixels and
- values of those pixels are replaced in block 38 with values of luminance, hue and saturation derived from weighted functions of the color components.
- the weighted functions of the color components are determined in equations (7), (8) and (9), and the replacement luminance, hue and
- saturation values are derived by inserting the weighted functions R', G' and B' into equations (1),
- the density of each pixel after insertion of replacement values is typically dark neutral
- the edges of the redeye regions are smoothed or feathered so that the colors of the pupil
- each redeye pixel of each redeye region to dark neutral gray as noted above, then blending the tones of edge pixels of
- the edge pixels are defined and identified in block 40 as redeye pixels in a redeye region
- FIG. 3 illustrates an 8x8 block 50 of redeye pixels, i.e. a redeye region 50, where each pixel 1 is an edge pixel.
- Each pixel is tested to determine if it should be designated as a redeye pixel, and if so, then the particular tone or level of gray which should be assigned to each layer of pixels is
- This feathering produces a more natural, gradual blend between the colors of the pupil and the retina.
- Equations (6) through (9) calculate weighted functions R', G', and B' from the R, G and B
- luminance, hue and saturation values for any pixel which is identified as a redeye pixel are identified as a redeye pixel.
- R' ⁇ C*n + R*(4-n) ⁇ /4 (7)
- G' ⁇ C*n + G*(4-n) ⁇ /4 (8)
- R', G' and B' are the weighted functions of the color components corresponding to
- the red color component R, the blue color component B and the green color component G are compared with one another and the numerical minimum value of the three color components is set to C, which is then applied to equations (7), (8) and (9).
- the predetermined threshold values include a maximum luminance threshold value
- MaxLum a minimum luminance threshold value MinLum, a maximum hue threshold value
- MaxHue a minimum hue threshold value MinHue, a maximum saturation threshold value MaxSat, and a minimum saturation threshold value MinSat.
- the predetermined threshold values establish the redeye range for each of the designated pixel properties.
- any potential redeye region having an aspect ratio greater than, say, 2: 1 or 1 :2 would be dismissed as a non-redeye region.
- the unassisted mode identifies all potential redeye pixels, it is relatively slow, computationally extensive and prone to false redeye
- the user via a data structure input to the algorithm. For instance, the user will move the cursor to a
- the predetermined threshold values for the fiilly assisted mode are represented as:
- MinLum Y - DeltaLumM (13)
- MaxLum Y + DeltaLumP (14)
- MinHue H - DeltaHueM (15)
- H and S are, respectively, the luminance, hue and saturation values of the
- DeltaSatP are preselected by testing images from a database.
- the user clicks the mouse cursor near the corner of an eye then pulls the cursor away from that point to create a rectangle around the area of interest where the user identifies redeye.
- the method as described herein automatically identifies and replaces any luminance, hue and saturation values of redeye pixels within the selected rectangular region.
- the saturation criterion is augmented by a function based on the color of the pixel.
- pixel is determined as the pixel within the selected rectangle which maximizes the expression (19):
- Figure 1 is a diagram of an image acquisition and rendering system 26 built in
- the system 26 includes a computer 14, one or more image acquisition devices such as the camera 12 and the scanner 10 which are useful for
- the system 26 includes conventional operator interface elements such as the keyboard
- the computer 14 processes certain programs which enable selected image processing operations such as the redeye removal algorithm described hereinbefore.
- system 26 may include a variety of components in addition to or instead of those depicted in Figure 1.
- the various elements of the system 26 may be replaced by special-purpose hardware, firmware or any combination thereof.
- those skilled in the art understand that many, if not all, of the processing capabilities of the computer 14 can alternatively be performed in intelligent system components, such as the camera 12, the scanner 10, the monitor 16, the keyboard 18 and the printer 20.
- the redeye removal algorithm as described above, is loaded into the computer 14, for instance, via a CDROM 24 and then run by the user who interacts with the computer 14 by way of the keyboard 18 and the monitor 16. First, an image is acquired, for instance, through an image acquisition device 10 or 12, or from a file stored within the computer 14.
- Means for selecting one or more pixels for comparison with the threshold values is defined as the computer 14 when the algorithm is operating in the unassisted mode, or as the mouse 22 when the algorithm is operating in either the assisted and partially assisted modes.
- the computer 14 then performs the functions of identifying redeye pixels, identifying redeye regions, replacing redeye pixel values, identifying edge pixels and the smoothing.
Abstract
A system and method for correcting redeye in a digital image includes the functions or steps, respectively, of: presetting threshold values of luminance, hue and saturation; selecting at least one pixel of the image for comparison with the threshold values; identifying redeye pixels of the image by comparing luminance, hue and saturation values which are derived from color components of the selected pixels with the threshold values; indentifying redeye regions of the image having one or more adjacent redeye pixels which are surrounded by non-redeye pixels; replacing the luminance, hue and saturation values of the identified redeye pixels with values derived from weighted functions of the color components subject to a first set of color weights; identifying edge pixels as the redeye pixels which border non-redeye pixels; and smoothing each redeye region by replacing the luminance, hue and saturation values of the edge pixels with the weighted functions of the color components subject to a second set of color weights.
Description
DIGITAL REDEYE REMOVAL
BACKGROUND OF THE INVENTION
In some color and black-and-white pictures taken by an electronic flash the eyes of a human
or an animal have an unnatural "redeye" effect which causes pupils to appear red rather than black, brown, blue, green or gray. It is generally accepted that the redeye effect is caused by light from
the electronic flash that is reflected from the retina of the subject's eye. Thus, light from the flash is imaged at the retina where a fraction of it is reflected back out the pupil of the eye. If the eye were
a perfect optical system, this light would be reflected entirely back into the flash source. However,
since the eye is not perfect, and since the original path of the light originating at the electronic flash is proximate to the lens axis, then some of the spreading reflected light from the retina overlaps the
lens axis and is sensed by the lens. Since the reflectivity of the retina is high in the red light portion of the visible light spectrum, the pupil spot effected by the reflected light is highly red. Although significant subject variations in susceptibility to redeye have been noted,-these can be explained by
variability in retinal-reflectaance, accommodation, pupil size and ocular optics of the subjects. The redeye effect can be avoided by locating the flash unit some distance (perhaps six
inches or more) from the lens axis, viewing the subject from an angle, having the subject look away from the camera, or raising the ambient light level so that the pupils will close to a significant
degree. Many other methods and systems have been utilized for minimizing redeye such as
reducing the size of the pupil by sending a pulsating light from the electronic flash just prior to taking the picture.
It is a primary object of the present invention to provide a system and method for redeye removal in a digital image.
SUMMARY OF THE INVENTION
A system and method for correcting redeye in a digital image includes the functions or
steps, respectively, of: presetting threshold values of luminance, hue and saturation; selecting at
least one pixel of the image for comparison with the threshold values; identifying redeye pixels of the image by comparing luminance, hue and saturation values which are derived from color
components of the selected pixels with the threshold values; identifying redeye regions of the image having one or more adjacent redeye pixels which are surrounded by non-redeye pixels; replacing
the luminance, hue and saturation values of the identified redeye pixels with values derived from
weighted functions of the color components subject to a first set of color weights; identifying edge pixels as the redeye pixels which border non-redeye pixels; and smoothing each redeye region by
replacing the luminance, hue and saturation values of the edge pixels with the weighted functions of the color components subject to a second set of color weights.
BRIEF DESCRIPTION OF THE DRAWINGS
The aforementioned aspects and other features of the invention are described in detail in
conjunction with the accompanying drawings where:
Figure 1 is a diagram of an image acquisition and rendering system including an image
processing system constructed in accordance with the principles of the invention;
Figure 2 is a block diagram of a preferred sequence of steps of the inventive method; and
Figure 3 is a representation of a redeye region composed of redeye pixels.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
A digital image consists of pixels each containing color and luminance information which
can be represented in a variety of domains. For instance, the pixel information can be resident in RGB (red, green and blue) device-dependent color space having R, G and B color components for use with a color monitor, or in CMYK (cyan, magenta, yellow black) device-dependent color space having C, M, Y and K color components for use with a color printer. When transferring
information between such device-dependent color spaces, use is made of device-independent color
space such as, but not limited to, proprietary color spaces, CIE L*a*b*, CIE XYZ, YUV and YIQ.
In one embodiment of the redeye removal method, the properties of a pixel are represented in RGB color space and the pixel values are transformed into YIQ color space as defined in the following equations (1) through (5) where, for each pixel, Y represents luminance, I and Q represent chrominance, H represents hue and S represents saturation.
Y = 0.30R + 0.59G + 0.11B (1)
I = 0.5R - 0.23G - 0.27B (2) Q = 0.203R - 0.5G + 0.297B (3)
H = tan-1(Q/I) (4)
S^ I' + Q2}^ (5)
The transformations of equations (1) through (5) are conventional as described in "Colour Space
In Television And Display Systems" by W. N. Sposon, Adam Hilger Ltd., Bristol (1983), pages 19 and 64.
Maximum and minimum threshold values of luminance, hue and saturation for defining a redeye region of a pixel are determined in block 30 of Figure 2 by testing a database of images
where: MaxLum represents the maximum luminance threshold value; MinLum represents the-
minimum luminance threshold value; MaxHue represents the maximum hue threshold value; MinHue represents the minimum hue threshold value; MaxSat represents the maximum saturation
threshold value; and MinSat represents the minimum saturation threshold value. Threshold ranges having both upper and lower limits are necessary for preventing spurious redeye readings. These
threshold values, the calculation of which will be detailed further on, are preferably precalculated and prestored in a configuration file, database, look-up table or other memory structure prior to
initiation of the redeye removal method. Furthermore, the above threshold values may be varied if
desired by the user.
One or more pixels of the image are selected in block 32 for comparison against the
predetermined threshold values, MaxLum, MinLum, MaxHue, MinHue, MaxSat and MinSat. The
comparison starts with one seed pixel and continues in some predetermined pattern until the luminance, hue and saturation properties of all pixels for a given area (which could be a part, or the
whole, of an image) are compared against the predetermined threshold values. If each of the luminance, hue and saturation values of a pixel being tested is both greater than the respective
minimum threshold value and less than the respective maximum threshold value, then that pixel in
block 34 is identified and labeled as a redeye pixel, i.e. a pixel which exhibits the redeye effect. For
instance, a pixel at a spatial location (x,y) could be flagged or labeled as a redeye pixel by setting
the appropriate element in the array Flags(x,y) equal to 3 whereas a non-redeye pixel could be indicated by setting Flags(x,y) equal to 0.
Redeye regions are identified in block 36 as regions of the image having one or more adjacent redeye pixels which are surrounded by non-redeye pixels. After all redeye pixels and
redeye regions have been identified in block 36, then the original luminance, hue and saturation
values of those pixels are replaced in block 38 with values of luminance, hue and saturation derived
from weighted functions of the color components. The weighted functions of the color components are determined in equations (7), (8) and (9), and the replacement luminance, hue and
saturation values are derived by inserting the weighted functions R', G' and B' into equations (1),
(4) and (5). The density of each pixel after insertion of replacement values is typically dark neutral
gray to closely resemble a pupil. Other Y, H and S replacement values corresponding to! different
shades of gray, or different eye colors such as blue, green, brown or gray could be preselected by
the user if desired.
In order to provide a more natural effect for the resulting pupils which have had redeye
removed, the edges of the redeye regions are smoothed or feathered so that the colors of the pupil
and the retina will gradually blend together. This is accomplished by first setting each redeye pixel of each redeye region to dark neutral gray as noted above, then blending the tones of edge pixels of
the redeye region from dark neutral gray to a lighter gray to account for flare and specular reflection.
The edge pixels are defined and identified in block 40 as redeye pixels in a redeye region
which border non-redeye pixels. Figure 3 illustrates an 8x8 block 50 of redeye pixels, i.e. a redeye region 50, where each pixel 1 is an edge pixel. Thus feathering or smoothing of redeye region 50 in
the simplest form amounts to setting all pixels 1, 2, 3 and 4 to dark neutral gray, and then replacing the edge pixels 1 with an intermediate gray value. However, the extent of feathering can be
extended beyond-one layer of edge pixels. This is accomplished in block 42 by flagging layers of
edge pixels. Each pixel is tested to determine if it should be designated as a redeye pixel, and if so, then the particular tone or level of gray which should be assigned to each layer of pixels is
determined by the amount of feathering desired. For inst-ance in Figure 3, the redeye region 50 is surrounded by non-redeye pixels (not shown) at locations (x,y) which are flagged as Flags(x,y) = 0.
All redeye pixels 1, 2, 3 and 4 are initially flagged as Flags(x,y) = 3 corresponding to pixels to be replaced with dark neutral gray. Thereafter, pixels 3 and 4 remain flagged as Flags(x,y) = 3 corresponding to dark neutral gray for producing the most correction, pixels 2 are flagged as
Flags(x,y) = 2 corresponding to an intermediate gray for producing medium correction, and pixels 1 are flagged as Flags(x,y) = 1 corresponding to a light gray for producing a small correction. Each
set of pixels designated by a different Flags value will have its original luminance, hue and
saturation values replaced. This feathering, as exemplified by the following pseudocode, produces a more natural, gradual blend between the colors of the pupil and the retina.
for each x,y in oldBorder for xl ,y 1 = each neighbor pixel of x,y and Flags(xl,yl) = 0 add xl,yl to newBorder setFlags(xl,yl) = 2 swap oldBorder and newBorder pointers for each x,y in oldBorder for xl,yl = each neighbor pixel of x,y and Flags(xl,yl) = 0 set Flags(xl,yl) = 1
Equations (6) through (9) calculate weighted functions R', G', and B' from the R, G and B
color components which, in turn, are inserted into equations (1) through (5) to derive replacement
luminance, hue and saturation values for any pixel which is identified as a redeye pixel.
C = min(R, G, B) (6)
R' = {C*n + R*(4-n)}/4 (7)
G' = {C*n + G*(4-n)}/4 (8)
B' = {C*n + B*(4-n)}/4 (9)
where: R', G' and B' are the weighted functions of the color components corresponding to
feathered colorant values which replace the original values R, G and B; and n is a variable defined as the color weighting value so that n = 4 for Flags = 3, n = 2 for Flags = 2, and n = 1 for
Flags = 1. The red color component R, the blue color component B and the green color component G are compared with one another and the numerical minimum value of the three color components is set to C, which is then applied to equations (7), (8) and (9).
Three modes are discussed for locating or selecting pbcels in the image for comparison with the threshold values. They are the unassisted mode, the partially assisted mode and the fully
assisted mode.
The predetermined threshold values include a maximum luminance threshold value
MaxLum, a minimum luminance threshold value MinLum, a maximum hue threshold value
MaxHue, a minimum hue threshold value MinHue, a maximum saturation threshold value MaxSat, and a minimum saturation threshold value MinSat. In the unassisted mode, for a given pixel having
a luminance Y, a hue H and a saturation S to be designated as a redeye pixel, the conditions of equations (10), (11) and (12) must be met.
MinLum < Y < MaxLum (10)
" MinHue ≤ H ≤ MaxHue (11)
MinSat < S < MaxSat (12)
In other words, the predetermined threshold values establish the redeye range for each of the designated pixel properties.
Some of the pixels identified as potential redeye pixels could actually be non-redeye pixels
which exhibit the same over-threshold attributes but which are located in parts of the image away
from the pupils. For instance, a bright red shirt could possibly result in erroneous redeye pixel findings. In order to minimize the occurrence of this type of error, additional threshold values are applied on the size and shape of the potential redeye regions. Since a pupil has an aspect ratio of approximately 1:1, then any potential redeye region having an aspect ratio greater than, say, 2: 1 or
1 :2 would be dismissed as a non-redeye region. Although the unassisted mode identifies all potential redeye pixels, it is relatively slow, computationally extensive and prone to false redeye
pixel identification.
In the fully assisted mode, the coordinates of the seed pixels of each region are supplied by
the user via a data structure input to the algorithm. For instance, the user will move the cursor to a
redeye region in the image, then click the cursor on the redeye region with the mouse. The predetermined threshold values for the fiilly assisted mode are represented as:
MinLum = Y - DeltaLumM (13)
MaxLum = Y + DeltaLumP (14) MinHue = H - DeltaHueM (15)
MaxHue = H + DeltaHueP ( 16)
MinSat = S - DeltaSatM ( 17)
MaxSat = S + DeltaSatP (18)
where as before Y, H and S are, respectively, the luminance, hue and saturation values of the
selected-pixels. The values DeltaLumM, DeltaLumP, DeltaHueM DeltaHueP, DeltaSatM and
DeltaSatP are preselected by testing images from a database.
After the threshold values have been preset as previously described, then a number of points are selected by the user as seed pixels for suspected redeye regions. Although the maximum
number of selected points is theoretically limited by the number of pixels in the image, the practical limitation is in the order of 10 to 20. As the number of selected points increases, the computational
time and processing requirements also increase, so it is preferred to select as few points as possible
for identification of redeye. For instance, in an image having two people with redeye, a minimum number of 4 points is preferably selected, one for each eye.
MISSING UPON TIME OF PUBLICATION
AUSENTE EN EL MOMENTO DE LA PUBLICACION
MANQUANT AU MOMENT DE LA PUBLICATION
In the partially assisted mode, the user clicks the mouse cursor near the corner of an eye then pulls the cursor away from that point to create a rectangle around the area of interest where the user identifies redeye. Upon selection by the user of a redeye removal button on the screen, the
method as described herein automatically identifies and replaces any luminance, hue and saturation values of redeye pixels within the selected rectangular region. However, for the partially assisted mode the saturation criterion is augmented by a function based on the color of the pixel. The seed
pixel is determined as the pixel within the selected rectangle which maximizes the expression (19):
S + HueFactor * (HueRedMax - abs(H - HueRed))/HueRedMax (19)
Only the pixels located within the user-specified rectangle are tested for redeye, so computational efficiency and processing times are minimized. The above steps can be repeated as desired for additional redeye regions with the image.
Figure 1 is a diagram of an image acquisition and rendering system 26 built in
accordance with the principles of the invention. The system 26 includes a computer 14, one or more image acquisition devices such as the camera 12 and the scanner 10 which are useful for
acquiring an image and converting it to digital image data, and one or more image rendering devices such as the monitor 16 and the printer 20 which are useful for generating an output
image.
The system 26 includes conventional operator interface elements such as the keyboard
18, the mouse 22, the monitor 16 and a compact or floppy disk 24. During operation of the system 26, the computer 14 processes certain programs which enable selected image processing operations such as the redeye removal algorithm described hereinbefore. The
system 26 may include a variety of components in addition to or instead of those depicted in Figure 1. For instance, the various elements of the system 26 may be replaced by
special-purpose hardware, firmware or any combination thereof. Also, those skilled in the art understand that many, if not all, of the processing capabilities of the computer 14 can alternatively be performed in intelligent system components, such as the camera 12, the scanner 10, the monitor 16, the keyboard 18 and the printer 20. The redeye removal algorithm, as described above, is loaded into the computer 14, for instance, via a CDROM 24 and then run by the user who interacts with the computer 14 by way of the keyboard 18 and the monitor 16. First, an image is acquired, for instance, through an image acquisition device 10 or 12, or from a file stored within the computer 14. Means for selecting one or more pixels for comparison with the threshold values is defined as the computer 14 when the algorithm is operating in the unassisted mode, or as the mouse 22 when the algorithm is operating in either the assisted and partially assisted modes. The computer 14 then performs the functions of identifying redeye pixels, identifying redeye regions, replacing redeye pixel values, identifying edge pixels and the smoothing.
The above embodiments of the present invention are merely illustrative of the many possible specific embodiments and variations of the invention as claimed. Many other equivalents of the claimed devices and methods are within the purview of the claims as understood by those of ordinary skill in the art.
Claims
1. A method for correcting redeye in a digital image, comprising the steps of: presetting threshold values of luminance, hue and saturation; selecting at least one pbcel of the image for comparison with said threshold values; identifying redeye pixels of the image by comparing luminance, hue and saturation values which are derived from color components of said selected pixels with said threshold values;
identifying redeye regions of the image having one or more adjacent redeye pixels which
are surrounded by non-redeye pixels; replacing the luminance, hue and saturation values of the identified redeye pixels with values derived from weighted functions of the color components;
identifying edge pixels as said redeye pixels which border non-redeye pixels; and smoothing each said redeye region by replacing the luminance, hue and saturation values of
the edge pixels with the weighted functions of the color components subject to a second set of color weights.
2. The method of claim 1 wherein said threshold values of luminance, hue and saturation are determined by testing on a database of images.
3. A system for correcting redeye in a digital image, comprising: means for selecting at least one pixel of the image for comparison with preset threshold values of luminance, hue and saturation; means for identifying redeye pixels of the image by comparing luminance, hue and saturation values which are derived from color components of said selected pixels with said threshold values; means for identifying redeye regions of the image having one or more adjacent redeye pixels which are surrounded by non-redeye pixels; means for replacing the luminance, hue and saturation values of the identified redeye pixels
with values derived from weighted functions of the color components subject to a first set of color weights;
means for identifying edge pixels as said redeye pixels which border non-redeye pixels; and means for smoothing each said redeye region by replacing the luminance, hue and
saturation values of the edge pixels with the weighted functions of the color components subject to a second set of color weights.
4. The system of claim 3 wherein said threshold values of luminance, hue and saturation are determined by testing on a database of images.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US93838497A | 1997-09-26 | 1997-09-26 | |
US08/938,384 | 1997-09-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO1999017254A1 true WO1999017254A1 (en) | 1999-04-08 |
Family
ID=25471338
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US1998/019428 WO1999017254A1 (en) | 1997-09-26 | 1998-09-17 | Digital redeye removal |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO1999017254A1 (en) |
Cited By (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1199672A2 (en) * | 2000-10-16 | 2002-04-24 | Xerox Corporation | Red-eye detection method |
GB2379819A (en) * | 2001-09-14 | 2003-03-19 | Pixology Software & Systems | Image processing to remove red-eye features |
GB2384639A (en) * | 2002-01-24 | 2003-07-30 | Pixology Software & Systems | Image processing to remove red-eye features |
GB2385736A (en) * | 2002-02-22 | 2003-08-27 | Pixology Software & Systems | Detection of red eye features in digital images by determination of the profile of image pixel values along a row of values |
WO2003071781A1 (en) * | 2002-02-22 | 2003-08-28 | Pixology Software Limited | Detection and correction of red-eye features in digital images |
US7042505B1 (en) | 1997-10-09 | 2006-05-09 | Fotonation Ireland Ltd. | Red-eye filter method and apparatus |
US7116820B2 (en) * | 2003-04-28 | 2006-10-03 | Hewlett-Packard Development Company, Lp. | Detecting and correcting red-eye in a digital image |
US7352394B1 (en) | 1997-10-09 | 2008-04-01 | Fotonation Vision Limited | Image modification based on red-eye filter analysis |
US7747071B2 (en) | 2005-10-27 | 2010-06-29 | Hewlett-Packard Development Company, L.P. | Detecting and correcting peteye |
US7764846B2 (en) | 2006-12-12 | 2010-07-27 | Xerox Corporation | Adaptive red eye correction |
US7787022B2 (en) | 1997-10-09 | 2010-08-31 | Fotonation Vision Limited | Red-eye filter method and apparatus |
US7907786B2 (en) * | 2005-06-06 | 2011-03-15 | Xerox Corporation | Red-eye detection and correction |
US7953252B2 (en) | 2005-11-18 | 2011-05-31 | Tessera Technologies Ireland Limited | Two stage detection for photographic eye artifacts |
US7962629B2 (en) | 2005-06-17 | 2011-06-14 | Tessera Technologies Ireland Limited | Method for establishing a paired connection between media devices |
US8000526B2 (en) | 2007-11-08 | 2011-08-16 | Tessera Technologies Ireland Limited | Detecting redeye defects in digital images |
US8000505B2 (en) | 2004-09-01 | 2011-08-16 | Eastman Kodak Company | Determining the age of a human subject in a digital image |
US8055067B2 (en) | 2007-01-18 | 2011-11-08 | DigitalOptics Corporation Europe Limited | Color segmentation |
US8064694B2 (en) | 2006-06-21 | 2011-11-22 | Hewlett-Packard Development Company, L.P. | Nonhuman animal integument pixel classification |
US8126217B2 (en) | 2005-11-18 | 2012-02-28 | DigitalOptics Corporation Europe Limited | Two stage detection for photographic eye artifacts |
US8126208B2 (en) | 2003-06-26 | 2012-02-28 | DigitalOptics Corporation Europe Limited | Digital image processing using face detection information |
US8160308B2 (en) | 2005-11-18 | 2012-04-17 | DigitalOptics Corporation Europe Limited | Two stage detection for photographic eye artifacts |
US8233674B2 (en) | 2007-03-05 | 2012-07-31 | DigitalOptics Corporation Europe Limited | Red eye false positive filtering using face location and orientation |
US20120243783A1 (en) * | 2011-03-21 | 2012-09-27 | Apple Inc. | Red-Eye Removal Using Multiple Recognition Channels |
US8786735B2 (en) | 2011-03-21 | 2014-07-22 | Apple Inc. | Red-eye removal using multiple recognition channels |
US8811683B2 (en) | 2011-06-02 | 2014-08-19 | Apple Inc. | Automatic red-eye repair using multiple recognition channels |
US8837785B2 (en) | 2011-03-21 | 2014-09-16 | Apple Inc. | Red-eye removal using multiple recognition channels |
US8837827B2 (en) | 2011-03-21 | 2014-09-16 | Apple Inc. | Red-eye removal using multiple recognition channels |
US9412007B2 (en) | 2003-08-05 | 2016-08-09 | Fotonation Limited | Partial face detector red-eye filter method and apparatus |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5130789A (en) * | 1989-12-13 | 1992-07-14 | Eastman Kodak Company | Localized image recoloring using ellipsoid boundary function |
EP0635972A2 (en) * | 1993-07-19 | 1995-01-25 | Eastman Kodak Company | Automated detection and correction of eye color defects due to flash illumination |
-
1998
- 1998-09-17 WO PCT/US1998/019428 patent/WO1999017254A1/en active Application Filing
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5130789A (en) * | 1989-12-13 | 1992-07-14 | Eastman Kodak Company | Localized image recoloring using ellipsoid boundary function |
EP0635972A2 (en) * | 1993-07-19 | 1995-01-25 | Eastman Kodak Company | Automated detection and correction of eye color defects due to flash illumination |
Cited By (48)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8203621B2 (en) | 1997-10-09 | 2012-06-19 | DigitalOptics Corporation Europe Limited | Red-eye filter method and apparatus |
US7787022B2 (en) | 1997-10-09 | 2010-08-31 | Fotonation Vision Limited | Red-eye filter method and apparatus |
US7619665B1 (en) | 1997-10-09 | 2009-11-17 | Fotonation Ireland Limited | Red eye filter for in-camera digital image processing within a face of an acquired subject |
US7352394B1 (en) | 1997-10-09 | 2008-04-01 | Fotonation Vision Limited | Image modification based on red-eye filter analysis |
US7042505B1 (en) | 1997-10-09 | 2006-05-09 | Fotonation Ireland Ltd. | Red-eye filter method and apparatus |
EP1199672A2 (en) * | 2000-10-16 | 2002-04-24 | Xerox Corporation | Red-eye detection method |
EP1199672A3 (en) * | 2000-10-16 | 2004-01-02 | Xerox Corporation | Red-eye detection method |
US6718051B1 (en) | 2000-10-16 | 2004-04-06 | Xerox Corporation | Red-eye detection method |
GB2379819B (en) * | 2001-09-14 | 2005-09-07 | Pixology Ltd | Image processing to remove red-eye features |
GB2379819A (en) * | 2001-09-14 | 2003-03-19 | Pixology Software & Systems | Image processing to remove red-eye features |
WO2003026278A1 (en) * | 2001-09-14 | 2003-03-27 | Pixology Limited | Image processing to remove red-eye features |
GB2379819A8 (en) * | 2001-09-14 | 2005-04-25 | Pixology Ltd | Image processing to remove red-eye features |
WO2003063081A3 (en) * | 2002-01-24 | 2003-11-06 | Pixology Software And Systems | Image processing to remove red-eye features without user interaction |
GB2384639B (en) * | 2002-01-24 | 2005-04-13 | Pixology Ltd | Image processing to remove red-eye features |
GB2384639A (en) * | 2002-01-24 | 2003-07-30 | Pixology Software & Systems | Image processing to remove red-eye features |
WO2003063081A2 (en) * | 2002-01-24 | 2003-07-31 | Pixology Software Limited | Image processing to remove red-eye features without user interaction |
GB2385736A (en) * | 2002-02-22 | 2003-08-27 | Pixology Software & Systems | Detection of red eye features in digital images by determination of the profile of image pixel values along a row of values |
WO2003071781A1 (en) * | 2002-02-22 | 2003-08-28 | Pixology Software Limited | Detection and correction of red-eye features in digital images |
GB2385736B (en) * | 2002-02-22 | 2005-08-24 | Pixology Ltd | Detection and correction of red-eye features in digital images |
US7116820B2 (en) * | 2003-04-28 | 2006-10-03 | Hewlett-Packard Development Company, Lp. | Detecting and correcting red-eye in a digital image |
US8224108B2 (en) | 2003-06-26 | 2012-07-17 | DigitalOptics Corporation Europe Limited | Digital image processing using face detection information |
US8126208B2 (en) | 2003-06-26 | 2012-02-28 | DigitalOptics Corporation Europe Limited | Digital image processing using face detection information |
US8131016B2 (en) | 2003-06-26 | 2012-03-06 | DigitalOptics Corporation Europe Limited | Digital image processing using face detection information |
US9412007B2 (en) | 2003-08-05 | 2016-08-09 | Fotonation Limited | Partial face detector red-eye filter method and apparatus |
US8000505B2 (en) | 2004-09-01 | 2011-08-16 | Eastman Kodak Company | Determining the age of a human subject in a digital image |
US7907786B2 (en) * | 2005-06-06 | 2011-03-15 | Xerox Corporation | Red-eye detection and correction |
US7962629B2 (en) | 2005-06-17 | 2011-06-14 | Tessera Technologies Ireland Limited | Method for establishing a paired connection between media devices |
US7747071B2 (en) | 2005-10-27 | 2010-06-29 | Hewlett-Packard Development Company, L.P. | Detecting and correcting peteye |
US7970184B2 (en) | 2005-11-18 | 2011-06-28 | Tessera Technologies Ireland Limited | Two stage detection for photographic eye artifacts |
US8126217B2 (en) | 2005-11-18 | 2012-02-28 | DigitalOptics Corporation Europe Limited | Two stage detection for photographic eye artifacts |
US8126218B2 (en) | 2005-11-18 | 2012-02-28 | DigitalOptics Corporation Europe Limited | Two stage detection for photographic eye artifacts |
US8131021B2 (en) | 2005-11-18 | 2012-03-06 | DigitalOptics Corporation Europe Limited | Two stage detection for photographic eye artifacts |
US8160308B2 (en) | 2005-11-18 | 2012-04-17 | DigitalOptics Corporation Europe Limited | Two stage detection for photographic eye artifacts |
US8175342B2 (en) | 2005-11-18 | 2012-05-08 | DigitalOptics Corporation Europe Limited | Two stage detection for photographic eye artifacts |
US8180115B2 (en) | 2005-11-18 | 2012-05-15 | DigitalOptics Corporation Europe Limited | Two stage detection for photographic eye artifacts |
US7970183B2 (en) | 2005-11-18 | 2011-06-28 | Tessera Technologies Ireland Limited | Two stage detection for photographic eye artifacts |
US7953252B2 (en) | 2005-11-18 | 2011-05-31 | Tessera Technologies Ireland Limited | Two stage detection for photographic eye artifacts |
US8064694B2 (en) | 2006-06-21 | 2011-11-22 | Hewlett-Packard Development Company, L.P. | Nonhuman animal integument pixel classification |
US7764846B2 (en) | 2006-12-12 | 2010-07-27 | Xerox Corporation | Adaptive red eye correction |
US8055067B2 (en) | 2007-01-18 | 2011-11-08 | DigitalOptics Corporation Europe Limited | Color segmentation |
US8233674B2 (en) | 2007-03-05 | 2012-07-31 | DigitalOptics Corporation Europe Limited | Red eye false positive filtering using face location and orientation |
US8000526B2 (en) | 2007-11-08 | 2011-08-16 | Tessera Technologies Ireland Limited | Detecting redeye defects in digital images |
US8786735B2 (en) | 2011-03-21 | 2014-07-22 | Apple Inc. | Red-eye removal using multiple recognition channels |
US8818091B2 (en) * | 2011-03-21 | 2014-08-26 | Apple Inc. | Red-eye removal using multiple recognition channels |
US8837785B2 (en) | 2011-03-21 | 2014-09-16 | Apple Inc. | Red-eye removal using multiple recognition channels |
US8837827B2 (en) | 2011-03-21 | 2014-09-16 | Apple Inc. | Red-eye removal using multiple recognition channels |
US20120243783A1 (en) * | 2011-03-21 | 2012-09-27 | Apple Inc. | Red-Eye Removal Using Multiple Recognition Channels |
US8811683B2 (en) | 2011-06-02 | 2014-08-19 | Apple Inc. | Automatic red-eye repair using multiple recognition channels |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO1999017254A1 (en) | Digital redeye removal | |
US6016354A (en) | Apparatus and a method for reducing red-eye in a digital image | |
US7519219B2 (en) | Method and program for correcting facial image and recording medium | |
JP2830871B2 (en) | Image color correction device and recording medium storing color correction program | |
JP4006347B2 (en) | Image processing apparatus, image processing system, image processing method, storage medium, and program | |
US6535301B1 (en) | Image processing apparatus, image processing method, image processing program recording medium, color adjustment method, color adjustment device, and color adjustment control program recording medium | |
JP4194133B2 (en) | Image processing method and apparatus, and storage medium | |
US7545976B2 (en) | Method and apparatus for associating image enhancement with color | |
US7468812B2 (en) | Image processing apparatus and its method for color correction | |
US20090027732A1 (en) | Image processing apparatus, image processing method, and computer program | |
US6262778B1 (en) | Image processing system | |
JP2002051231A (en) | Method for reproducing color image of scene by using preferential color mapping and scene-dependent tone scale constitution | |
JP4368513B2 (en) | Image processing method and apparatus, and recording medium | |
JP2000134486A (en) | Image processing unit, image processing method and storage medium | |
JP2001186325A (en) | Method and device for processing picture, recording medium and transmitting medium | |
JP2004509486A (en) | Color cast detection and removal method | |
GB2381689A (en) | Finding and filtering smaller and larger noise areas in an image | |
JPH05342344A (en) | Method and system for picture processing | |
JP3918257B2 (en) | Image processing apparatus, image processing method, and medium storing image processing program | |
JPH1091761A (en) | Picture processor | |
JP4281244B2 (en) | Image forming apparatus, image data processing method, and recording medium recording image data processing program | |
CN109064431A (en) | A kind of picture luminance adjusting method, equipment and its storage medium | |
JP2000013625A (en) | Image processing method, device and recording medium | |
JPH11232451A (en) | Adapting method of color representation to image observed in environments of different relative luminance levels | |
Hardeberg | Red eye removal using digital color image processing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
122 | Ep: pct application non-entry in european phase |