WO2005091651A2 - Interactive video display system - Google Patents

Interactive video display system Download PDF

Info

Publication number
WO2005091651A2
WO2005091651A2 PCT/US2005/008984 US2005008984W WO2005091651A2 WO 2005091651 A2 WO2005091651 A2 WO 2005091651A2 US 2005008984 W US2005008984 W US 2005008984W WO 2005091651 A2 WO2005091651 A2 WO 2005091651A2
Authority
WO
WIPO (PCT)
Prior art keywords
camera
interactive
recited
display screen
video display
Prior art date
Application number
PCT/US2005/008984
Other languages
French (fr)
Other versions
WO2005091651A3 (en
Inventor
Matthew Bell
Philip Gleckman
Joshua Zide
Helen Shaughnessy
Original Assignee
Reactrix Systems, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US10/946,414 external-priority patent/US7710391B2/en
Priority claimed from US10/946,084 external-priority patent/US20050122308A1/en
Priority claimed from US10/946,263 external-priority patent/US8035612B2/en
Application filed by Reactrix Systems, Inc. filed Critical Reactrix Systems, Inc.
Publication of WO2005091651A2 publication Critical patent/WO2005091651A2/en
Publication of WO2005091651A3 publication Critical patent/WO2005091651A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • G06F3/0425Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means using a single imaging device like a video camera for tracking the absolute position of a single or a plurality of objects with respect to an imaged reference surface, e.g. video camera imaging a display or a projection screen, a table or a wall surface, on which a computer generated image is displayed or projected
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • G06F3/0428Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means by sensing at the edges of the touch surface the interruption of optical paths, e.g. an illumination plane, parallel to the touch surface which may be virtual
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0486Drag-and-drop
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1087Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera
    • A63F2300/1093Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera using visible light

Definitions

  • the present invention relates to the field of visual electronic displays. Specifically, embodiments of the present invention relate to an interactive video display system.
  • buttons have been used for presenting information on flat surfaces. For example, an image may be displayed on a touchscreen, and a user may interact with the image by touching the touchscreen, causing the image to change.
  • the user in order to interact with the image displayed on the touchscreen, the user must actually come in contact with the touchscreen.
  • typically touchscreens can only receive one input at any time, and are not able to discern the shape of the input.
  • the present invention provides a self-contained interactive video display system.
  • a display screen is for displaying a visual image for presentation to a user.
  • the display screen is a cathode ray tube (CRT).
  • the display screen is a flat-panel display screen.
  • the flat-panel display screen is a liquid crystal display (LCD) screen.
  • the display screen is a plasma screen.
  • the self-contained interactive video display system further includes a projector for projecting the visual image and wherein the display screen is a screen for displaying the visual image.
  • a camera is for detecting an object in an interactive area located in front of the display screen, where the camera is operable to capture three- dimensional information about the object.
  • the camera is a time-of-flight camera.
  • the object is a user.
  • the distance information is used for person tracking. In one embodiment, the distance information is used for feature tracking.
  • the self-contained interactive video display system further includes a second camera for detecting the object.
  • the second camera is a time-of-flight camera.
  • the camera and the second camera operate in conjunction to create the interactive area.
  • the camera is operable to create the interactive area proximate the display screen, and the second camera is operable to create a second interactive area farther from the display screen than the interactive area.
  • the self-contained interactive video display system further includes an infrared camera for creating a second interactive area.
  • the camera provides the computer system with three- dimensional information at a first frame rate and the infrared camera provides the computer system with two-dimensional information at a second frame rate, wherein the first frame rate is slower than the second frame rate, such that the computer system is operable to generate high quality three-dimensional information based on the three-dimensional information and the two- dimensional information.
  • the computer system can identify a portion of people near the self-contained interactive video display system actively interact with the self-contained interactive video display system by determining people in the second interactive area that do not enter the interactive area.
  • the interactive area comprises a plurality of interactive regions including volumetric regions at particular distances from the display screen.
  • a first volumetric region extends from the display screen to a particular distance from the display screen
  • a second volumetric region extends from the particular distance away from the display screen.
  • the second volumetric region extends to a second particular distance away from the display screen.
  • the second particular distance is dynamic.
  • a user may interact with the visual image by moving into and out of the first volumetric region.
  • the present invention provides an interactive video display system.
  • a display screen is for displaying a visual image for presentation to a user.
  • a camera is for detecting an object in an interactive area located in front of the display screen, the camera operable to capture three-dimensional information about the object.
  • the camera is a time-of-flight camera.
  • a computer system is for directing the display screen to change the visual image in response to the object.
  • the camera is separate from the display screen, such that the interactive area is between the camera and the display screen.
  • the interactive video display system includes a second camera for detecting the object, wherein the second camera is separate from the display screen.
  • the second camera is a time-of-flight camera.
  • the camera and the second camera operate in conjunction to create the interactive area.
  • the camera is operable to create the interactive area proximate the display screen, and the second camera is operable to create a second interactive area farther from the display screen than the interactive area.
  • the present invention provides a method for presenting an interactive visual image using an interactive video display system.
  • a visual image is displayed on a display screen, wherein the visual image is for presentation to a user on a front side of the display screen.
  • Interaction of an object with the visual image is detected using a camera, the camera operable to capture three-dimensional information about the object.
  • the visual image is changed in response to the interaction.
  • Figure 1 illustrates a side view of an interactive display including multiple time-of-flight cameras, in accordance with an embodiment of the present invention.
  • Figure 2 illustrates a top view of an interactive display including multiple time-of-flight cameras, in accordance with an embodiment of the present invention.
  • Figure 3 illustrates a top view of an interactive display including multiple time-of-flight cameras, in accordance with an embodiment of the present
  • Figure 4 illustrates a side view of an interactive display including multiple time-of-flight cameras for detecting interactions at different distances from the display, in accordance with an embodiment of the present invention.
  • Figure 5A illustrates a side view of an interactive floor display including multiple time-of-flight cameras, in accordance with an embodiment of the present invention
  • Figure 5B illustrates a top view of the surface covered by the interactive floor display of Figure 5A, in accordance with an embodiment of the present invention.
  • Figure 6 illustrates a side view of an interactive display including an infrared camera and a time-of-flight camera, in accordance with an embodiment of the present invention.
  • Figure 7 illustrates an interactive display system including different volumetric interactive regions, in accordance with an embodiment of the present invention.
  • the interactive video display system of the described embodiments require the use of a video camera which produces a three-dimensional (3D) image of the objects it views.
  • Time-of-flight cameras have this property.
  • laser scanners, stereo cameras, and other technologies are also capable of producing 3D image data.
  • the embodiments of the present invention are described as including at least one time-of-flight camera, it should be appreciated that the present invention may be implemented using any camera or combination of cameras that are operable to determine three- dimensional information of the imaged object, such as a laser scanners and stereo cameras.
  • Embodiments of the present invention may be implemented using time- of-flight cameras.
  • a time-of-flight camera has a built-in capability to detect distance information for each pixel of its image. This distance information for each pixel may consist of the distance from the camera to the object seen at that pixel.
  • Time-of-flight cameras obtain this distance data by timing how long it takes an invisible light pulse to travel from an illuminator next to the camera to the object being imaged and then back to the camera. The light pulses can be rapidly repeated, allowing the time-of-flight camera to have a frame rate similar to a video camera.
  • Time of flight cameras typically have a range of 1-2 meters at 30 frames per second. However, the range can be significantly increased by lengthening the exposure time, which lowers the frame rate.
  • Manufacturers of time-of-flight cameras include Canesta Inc of Sunnyvale, CA, and 3DV systems of Israel.
  • the time-of-flight camera or cameras are positioned near a video display such that they can see objects on and in front of the display.
  • the objects viewed by the camera or cameras consist of the hands, arms, and bodies of human users interacting with the display.
  • the time-of-flight camera may work with any display (e.g., an LCD panel, a cathode- ray tube display, etc.) without modifications.
  • Data from the time-of-flight camera or cameras may be processed in real time by a computer that uses the data to create interactive video effects on the display.
  • the interactive video effects may be calibrated such that a user making a hand motion in front of a particular spot on the display will cause interactive video effects at that spot on the display.
  • an "augmented reality" system is created, whereby users can use their hands and other parts of their bodies to manipulate virtual objects and cause digital visual effects on the display.
  • a single time-of-flight camera may be used.
  • a single time-of- flight camera may not be able to detect objects that are blocked by objects closer to the camera. For example, if a user had one hand in front of the other relative to the camera, then the camera would not be able to see the other hand. Therefore, embodiments of the present invention may utilize multiple time-of-flight cameras, as shown in Figure 1. The multiple cameras, with multiple perspectives, reduce the chance that an object will not be seen.
  • the interactive space in front of the display has a coordinate system for specifying the location of users or objects in front of the display.
  • the system should know the position the user or object is on the screen and how far it is away from the screen it is. Having user/object data in the aforementioned coordinate system is very useful as it allows the easy creation of interactive effects that are physically matched to positions on the display, as well as having interactive effects that depend on the distance from the user/object to the display.
  • a coordinate transform is performed on each pixel of each time-of- flight camera to put it in a common coordinate space with each other camera and the display.
  • One such space is defined by: (x, y) - the position of the point projected perpendicularly onto the display surface, and (z) - the perpendicular distance of the point from the display.
  • This coordinate space transformation can be determined by looking at the angle and position of each camera relative to the display. Alternatively, the transformation may be determined by a calibration process, in which an object of known size, shape and position is placed in front of the display. By having each of the cameras image the object, the appropriate transformation function can be determined from points viewed by each camera into points in the. common coordinate space. If the camera coordinate transforms are done in real time, then a real-time picture of the area in front of the camera in 3D is achieved.
  • FIG. 3 illustrates a front view of an interactive display 3000 including multiple time-of-flight cameras, in accordance with an embodiment of the present invention.
  • Interactive display 3000 includes time-of-flight cameras 3010 and 3020 that create an interactive area overlaying the surface of display 3030. It should be appreciated that multiple time-of-flight cameras may be placed at variations locations and positions around display 3030 to create the interactive area. The interactive area is bounded by the viewed region of the time-of-flight cameras, as shown by the dotted lines.
  • time-of-flight cameras may be placed in two corners of display 3030, four corners of display 3030 (as shown in Figure 2), next to each other on one side of display 3030 so that the interactive areas of each camera are substantially parallel, or any other variation for creating an interactive area overlaying display 3030.
  • Figure 4 illustrates a side view of an interactive display 3100 including multiple time-of-flight cameras for detecting interactions at different distances from the display, in accordance with an embodiment of the present invention.
  • Interactive display 3100 includes time-of-flight cameras 3110 and 3120 that create an interactive area overlaying and proximate the surface of display 3140.
  • Time-of-flight camera 3130 creates an interactive area for a region farther from display 3140.
  • Time-of-flight camera 3130 is operable to capture interactions farther from display 3140 than time-of-flight cameras 3110 and 3120. For instance, time-of-flight camera 3130 provides the ability to detect a person passing by interactive display 3100 at a certain distance and change the visual image shown on display 3140 to attract their attention. As the person approaches the display, they are picked up by time-of-flight cameras 3110 and 3120. In one embodiment, time-of-flight cameras 3110, 3120 and 3130 are the same type of time-of-flight camera, where time-of-flight camera 3130 operates at a lower frame rate and longer range than time-of-flight cameras 3110 and 3120.
  • a time-of-flight camera or cameras are placed at a distance from the display to create the interactive region.
  • the display may be placed on a wall and the time-of-flight camera(s) may be placed on the opposite wall or on the ceiling, creating the interactive area between the time-of-flight camera (s) and the display.
  • the present invention may provide an interactive table, where the display is placed substantially horizontal and the time-of-flight camera(s) are above the table or to the side of the table. It should be appreciated that there are many different permutations of configurations of the present invention, and that the present invention is not limited to the described embodiments.
  • the distance information supplied by time-of-flight cameras can also be used to identify a user's hands. Identifying hands is important because hands are often used to provide specific interactions with the interactive display. In one embodiment, the hands are determined to be the points on each object viewed by the cameras that are closest to the display. In another embodiment, 3D shape recognition can be used to identify the hands. For example, simple cylinders or more complex skeletal models could be matched (using a variety of processes such as gradient descent) to users' hands, arms, and bodies. Other algorithms may also be used to identify the hands.
  • FIG. 5A illustrates a side view of an interactive floor display 3200 including multiple time-of-flight cameras, in accordance with an embodiment of the present invention.
  • Interactive floor display 3200 includes six time-of-flight cameras 3210a-f placed above display surface 3220. Each time-of-flight camera is directed at a different portion of display surface 3220.
  • the smaller the area imaged by a time-of-flight camera to create an interactive area the higher the frame rate possible by the time-of-flight camera.
  • the more time-of-flight cameras covering a region of display surface 3220 the better the performance of interactive floor display 3200.
  • Figure 5B illustrates a top view of the surface covered by interactive floor display 3200 of Figure 5A. Time-of-flight cameras 3210a-f each image a respective region 3230a-f for creating the interactive area.
  • Figure 6 illustrates a side view of an interactive display 3300 including an infrared camera 3310 and a time-of-flight camera 3320, in accordance with an embodiment of the present invention.
  • Infrared camera 3310 is operable to record images illuminated by infrared light of an infrared illuminator (not shown).
  • Infrared camera 3310 and the infrared illuminator create interactive area 3315 above display surface 3330 (e.g., a floor).
  • Time-of-flight camera 3320 is operable to create interactive area 3325 above display surface 3330.
  • interactive area 3325 covers a larger area than interactive area 3315.
  • Infrared camera 3310 is operable to provide a succession of two- dimensional (2D) images of objects in interactive area 3315.
  • time-of-flight camera 3325 operates at a lower frame rate than infrared camera 3310 to provide three-dimensional image information of interactive area 3325 (including interactive area 3315).
  • the three-dimensional image information may be used to label the two- dimensional image to provide high quality three-dimensional image information.
  • the three-dimensional image information of time-of-flight camera 3325 is superimposed on the two-dimensional image.
  • the 3D camera will have a lower frame rate than the 2D camera.
  • 3D information can be estimated for two-dimensional images taken without a corresponding three-dimensional image. Since a feature tracking algorithm operating on the stream of 2D images can track the movement of several patches of an object, the 3D information from the most recent 3D camera image can be tracked to its current positions in the two- dimensional image, thus providing estimated 3D information even at times when no 3D image is available.
  • the three-dimensional image information can also be used to provide for person tracking.
  • the height of the highest point of an object can be used to identify the object as a person.
  • the camera is on the ceiling and images a region on the floor
  • the point closest to the camera for each person can be identified.
  • the height, size, and 3D shape of each object viewed by the cameras can be used to identify whether it is a person. This can be useful in determining how many people are interacting with interactive display 3300.
  • height information can be used to distinguish between adults and children, which may be useful information for marketing purposes.
  • Other information can also be acquired, such as the orientation of people (e.g., standing, sitting, lying down).
  • interactive area 3325 is bigger than interactive area 3315, people can be tracked that are not interacting with interactive area 3315. This information can be used to identify what portion of people passing interactive display 3300 actually end up interacting with interactive display 3300. It should be appreciate that person tracking and feature tracking may be performed without the use of an infrared camera, because the three-dimensional information provided by a time-of-flight camera is sufficient. However, the infrared camera is useful for getting higher-quality 2D outline and shape information, which may be useful in creating the interactive effects.
  • FIG. 7 illustrates an interactive display system 3400 including different volumetric interactive regions, in accordance with an embodiment of the present invention. While interactive display system 3400 is shown as a wall display, it should be appreciated that other types of displays, such as floor displays and table displays, can be implemented.
  • Interactive display 3400 includes time-of-flight cameras 3410 and 3420 that create an interactive area including different interactive areas in front of the surface of display 3430. Since time-of-flight cameras include distance information for viewed objects, different interactive areas that include specific volumetric regions at particular distances from display 3430 may be defined.
  • Interactive display 3400 includes interactive regions that include specific volumetric regions. For instance, region 3440, also referred to herein as the touch region, is proximate to display 3430. In one embodiment, the touch region extends up to six inches out from display 3430. Region 3442, also referred to herein as the interactive region, extends from the end of the touch region. In one embodiment, the interactive region may be dynamic. For instance, if no people are interacting with interactive display 3400, region 3442 may extend several feet or to the end of a room to attempt to cause people to interact, thereby attracting their attention. Once a person engages interactive display 3400, the back of region 3442 may move in so as to ensure that other more distant people passing by do not affect the display.
  • the interactive region ends a particular distance (e.g., one foot) behind the nearest person in the interactive region.
  • interactive display 3400 may include any number of volumetric interactive regions (e.g., region 3444), and is not limited to the described embodiments.
  • the touch zone also allows for a user to perform certain types of gestures.
  • display 3430 may display an object that a user can move through hand gestures. This movement of objects can be thought of as being analogous to the "drag and drop" operation commonly performed with a computer mouse, except in this case it is done using a hand gesture.
  • the user selects the object by moving their hand toward the object and crossing through the threshold separating the touch zone from the interactive zone. Staying in the touch zone, the user moves their hand to another location, and the object follows their hand's movement. The user can then release the object by pulling their hand back, moving it out of the touch zone.
  • the object can be selected by moving into and out of the touch zone such that the object may be moved while the hand is in the interactive zone. Then, the user can release the object by moving their hand back into and out of the touch zone.
  • the object can be selected by the user making a grabbing motion (closing their hand). The object could then be moved by the user moving their hand, and finally released by the user by opening their hand.
  • the touch zone can also be used for pressing virtual buttons or selecting objects on the display.
  • a user can press a button by moving their hand over the button's image on the display and then moving their hand towards the display until their hand crosses the threshold from the interactive zone to the touch zone. When their hand crosses this threshold, the button is pushed.
  • Information from the coordinate-transformed 3D data about objects or users in front of the display can serve as input to visual effects on the display.
  • the outline (from the display's point of view) of users in front of the display can be shown on the display to give users a sense of where they are as they are using the system.
  • display 3430 changes in response to an object being selected and/or released.
  • display 3430 includes one or more icons that are associated with the hands of a user and are located at the position(s) on the screen corresponding to the user's hand(s). It should be appreciated that there can be multiple icons active on display 3430 at any time.
  • the icon can transform (e.g., change shape, color, or size) when a user moves from the interactive zone to the touch zone, signifying a selection or release of an object.
  • the interactive video display system can be used in many different applications.
  • the system's capability to have touchscreen-like behavior for multiple users as well as full or partial body outline interaction increases its appeal for information interfaces which require more precise selection and manipulation of buttons and objects.
  • Uses of these interactive display systems include, but are not limited to, interactive video games in which users move their bodies to play the game, interactive menu, catalog, and browsing systems that let users browse through pages of informational content using gestures, systems that allow users to "try on” clothing using an image of themselves, pure entertainment applications in which images or outlines of users serve as input to a video effects system, interactive characters that interact with the motions of users in front of the screen, and virtual playlands and storybooks that users interact with by moving their bodies.
  • Other uses of the present invention include, but are not limited to: allowing users to customize or view available options for customizing the product on display, allowing the product on display to be ordered at the display, using either the display interface, a keyboard, a credit card swiper, or a combination of the three, comparing the features of multiple products on the display, showing combinations or compatibilities between multiple products on the display, and placing a product in different virtual settings on the screen to demonstrate the features (e.g., water, forest, asphalt, etc.)
  • These interactive display systems can incorporate additional inputs and outputs, including, but not limited to, microphones, touchscreens, keyboards, mice, radio frequency identification (RFID) tags, pressure pads, cellular telephone signals, personal digital assistants (PDAs), and speakers.
  • RFID radio frequency identification
  • PDAs personal digital assistants
  • These interactive display systems can be tiled together to create a single larger screen or interactive area. Tiled or physically separate screens can also be networked together, allowing actions on one screen to affect the image on another screen.
  • the present invention is implemented using a combination of hardware and software in the form of control logic, in either an integrated or a modular manner. Based on the disclosure and teachings provided herein, a person of ordinary skill in the art will know of other ways and/or methods to implement the present invention.
  • the present invention as described above provides a system that allows a camera to view an area in front of a display.
  • a system is provided to create a reactive space in front of the display.
  • the present invention can be used to capture information from the reactive space.

Abstract

An interactive video display system. A display screen is for displaying a visual image for presentation to a user. A camera is for detecting an object in an interactive area located in front of the display screen, the camera operable to capture three-dimensional information about the object. A computer system is for directing the display screen to change the visual image in response to the object.

Description

INTERACTIVE VIDEO DISPLAY SYSTEM
CROSS-REFERENCE TO RELATED APPLICATION
[0001] This application is a Continuation-in-Part Patent Application claiming priority from co-pending U.S. Patent Application No. 10/160,217, filed on May 28, 2002, entitled "INTERACTIVE VIDEO DISPLAY SYSTEM," by Bell, and assigned to the assignee of the present application, co-pending U.S. Patent Application No. 10/946,263, filed on September 20, 2004, entitled "SELF- CONTAINED INTERACTIVE VIDEO DISPLAY SYSTEM," by Bell et al., and assigned to the assignee of the present application, co-pending U.S. Patent Application No. 10/946,084, filed on September 20, 2004, entitled "SELF- . CONTAINED INTERACTIVE VIDEO DISPLAY SYSTEM," by Bell, and assigned to the assignee of the present application, and co-pending U.S. Patent Application No. 10/946,414, filed on September 20, 2004, entitled "INTERACTIVE VIDEO WINDOW DISPLAY SYSTEM," by Bell, and assigned to the assignee of the present application, all of which are herein incorporated by reference. This application also claims priority from co-pending U.S. Provisional Patent Application No. 60/554,520, filed on March 18, 2004, entitled "METHOD AND SYSTEM FOR ALLOWING A CAMERA TO VIEW AN AREA IN FRONT OF A DISPLAY BY IMAGING IT THROUGH THE DISPLAY," by Bell et al., and assigned to the assignee of the present application, all of which are herein incorporated by reference. FIELD
[0002] The present invention relates to the field of visual electronic displays. Specifically, embodiments of the present invention relate to an interactive video display system.
BACKGROUND
[0003] For many years, information was typically conveyed to an audience by use of static displays. For example, product advertisements were presented using print ads and posters. With the advent of television and movies, information could be presented using a dynamic display (e.g., commercials). While more engaging than static displays, dynamic displays do not typically provide interactivity between a user and the display.
[0004] More recently, interactive touchscreens have been used for presenting information on flat surfaces. For example, an image may be displayed on a touchscreen, and a user may interact with the image by touching the touchscreen, causing the image to change. However, in order to interact with the image displayed on the touchscreen, the user must actually come in contact with the touchscreen. Moreover, typically touchscreens can only receive one input at any time, and are not able to discern the shape of the input.
Essentially, current touchscreens are only able to receive the input of one finger contact. [0005] In some applications, such as point-of-sale, retail advertising, promotions, arcade entertainment sites, etc., it is desirable to provide. an interactive interface for displaying information to a user. This interactivity provides an even more engaging interface for presenting information (e.g., media, advertisements, etc.). By catching the attention of a person, for even a few moments, the person may be more likely to absorb the information presented in the interactive display than in previous displays.
[0006] As described above, current interactive displays typically require a user to physically contact a touchscreen surface. By requiring contact with a touchscreen to provide interactivity, a large number of potential users are not interested in or intimidated by current interactive displays. Moreover, since only one user may interact with a touchscreen at a time, more users are excluded. Furthermore, because current touchscreens cannot discern the shape of input, they are limited in the type of information that can be presented in response to interaction.
SUMMARY
[0007] Various embodiments of the present invention, an interactive video display system, are described herein. In one embodiment, the present invention provides a self-contained interactive video display system. A display screen is for displaying a visual image for presentation to a user. In one embodiment, the display screen is a cathode ray tube (CRT). In another embodiment, the display screen is a flat-panel display screen. In one embodiment, the flat-panel display screen is a liquid crystal display (LCD) screen. In another embodiment, the display screen is a plasma screen. In one embodiment, the self-contained interactive video display system further includes a projector for projecting the visual image and wherein the display screen is a screen for displaying the visual image.
[0008] A camera is for detecting an object in an interactive area located in front of the display screen, where the camera is operable to capture three- dimensional information about the object. In one embodiment, the camera is a time-of-flight camera. In one embodiment, the object is a user. In one embodiment, the distance information is used for person tracking. In one embodiment, the distance information is used for feature tracking.
[0009] In one embodiment, the self-contained interactive video display system further includes a second camera for detecting the object. In one embodiment, the second camera is a time-of-flight camera. In one embodiment, the camera and the second camera operate in conjunction to create the interactive area. In another embodiment, the camera is operable to create the interactive area proximate the display screen, and the second camera is operable to create a second interactive area farther from the display screen than the interactive area.
[0010] In one embodiment, the self-contained interactive video display system further includes an infrared camera for creating a second interactive area. In one embodiment, the camera provides the computer system with three- dimensional information at a first frame rate and the infrared camera provides the computer system with two-dimensional information at a second frame rate, wherein the first frame rate is slower than the second frame rate, such that the computer system is operable to generate high quality three-dimensional information based on the three-dimensional information and the two- dimensional information. In one embodiment, the computer system can identify a portion of people near the self-contained interactive video display system actively interact with the self-contained interactive video display system by determining people in the second interactive area that do not enter the interactive area.
[0011] In one embodiment, the interactive area comprises a plurality of interactive regions including volumetric regions at particular distances from the display screen. In one embodiment, a first volumetric region extends from the display screen to a particular distance from the display screen, and a second volumetric region extends from the particular distance away from the display screen. In one embodiment, the second volumetric region extends to a second particular distance away from the display screen. In one embodiment, the second particular distance is dynamic. In one embodiment, a user may interact with the visual image by moving into and out of the first volumetric region.
[0012] In another embodiment, the present invention provides an interactive video display system. A display screen is for displaying a visual image for presentation to a user. A camera is for detecting an object in an interactive area located in front of the display screen, the camera operable to capture three-dimensional information about the object. In one embodiment, the camera is a time-of-flight camera. A computer system is for directing the display screen to change the visual image in response to the object. In one embodiment, the camera is separate from the display screen, such that the interactive area is between the camera and the display screen. In one embodiment, the interactive video display system includes a second camera for detecting the object, wherein the second camera is separate from the display screen. In one embodiment, the second camera is a time-of-flight camera. In one embodiment, the camera and the second camera operate in conjunction to create the interactive area. In another embodiment, the camera is operable to create the interactive area proximate the display screen, and the second camera is operable to create a second interactive area farther from the display screen than the interactive area.
[0013] In another embodiment, the present invention provides a method for presenting an interactive visual image using an interactive video display system. A visual image is displayed on a display screen, wherein the visual image is for presentation to a user on a front side of the display screen. Interaction of an object with the visual image is detected using a camera, the camera operable to capture three-dimensional information about the object. The visual image is changed in response to the interaction.
BRIEF DESCRIPTION OF THE DRAWINGS
[0014] The accompanying drawings, which are incorporated in and form a part of this specification, illustrate embodiments of the invention and, together with the description, serve to explain the principles of the invention:
[0015] Figure 1 illustrates a side view of an interactive display including multiple time-of-flight cameras, in accordance with an embodiment of the present invention.
[0016] Figure 2 illustrates a top view of an interactive display including multiple time-of-flight cameras, in accordance with an embodiment of the present invention.
[0017] Figure 3 illustrates a top view of an interactive display including multiple time-of-flight cameras, in accordance with an embodiment of the present
invention.
[0018] Figure 4 illustrates a side view of an interactive display including multiple time-of-flight cameras for detecting interactions at different distances from the display, in accordance with an embodiment of the present invention. [0019] Figure 5A illustrates a side view of an interactive floor display including multiple time-of-flight cameras, in accordance with an embodiment of the present invention
[0020] Figure 5B illustrates a top view of the surface covered by the interactive floor display of Figure 5A, in accordance with an embodiment of the present invention.
[0021] Figure 6 illustrates a side view of an interactive display including an infrared camera and a time-of-flight camera, in accordance with an embodiment of the present invention.
[0022] Figure 7 illustrates an interactive display system including different volumetric interactive regions, in accordance with an embodiment of the present invention.
DETAILED DESCRIPTION
[0023] Reference will now be made in detail to various embodiments of the invention, an electronic device for monitoring the presence of objects around a second electronic device, examples of which are illustrated in the accompanying drawings. While the invention will be described in conjunction with these embodiments, it is understood that they are not intended to limit the invention to these embodiments. On the contrary, the invention is intended to cover alternatives, modifications and equivalents, which may be included within the spirit and scope of the invention as defined by the appended claims. Furthermore, in the following detailed description of the invention, numerous specific details are set forth in order to provide a thorough understanding of the invention. However, it will be recognized by one of ordinary skill in the art that the invention may be practiced without these specific details. In other instances, well known methods, procedures, components, and circuits have not been described in detail as not to unnecessarily obscure aspects of the invention.
[0024] Some portions of the detailed descriptions, which follow, are presented in terms of procedures, steps, logic blocks, processing, and other symbolic representations of operations on data bits that can be performed on computer memory. These descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. A procedure, computer executed step, logic block, process, etc., is here, and generally, conceived to be a self-consistent sequence of steps or instructions leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated in a computer system. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
[0025] It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the following discussions, it is appreciated that throughout the present invention, discussions utilizing terms such as "projecting" or "detecting" or "changing" or "illuminating" or "correcting" or "eliminating" or the like, refer to the action and processes of an electronic system (e.g., an interactive video system), or similar electronic computing device, that manipulates and transforms data represented as physical (electronic) quantities within the electronic device's registers and memories into other data similarly represented as physical quantities within the electronic device memories or registers or other such information storage, transmission or display devices. [0026] The interactive video display system of the described embodiments require the use of a video camera which produces a three-dimensional (3D) image of the objects it views. Time-of-flight cameras have this property. However, laser scanners, stereo cameras, and other technologies are also capable of producing 3D image data. Moreover, while the embodiments of the present invention are described as including at least one time-of-flight camera, it should be appreciated that the present invention may be implemented using any camera or combination of cameras that are operable to determine three- dimensional information of the imaged object, such as a laser scanners and stereo cameras.
TlME-OF-FLIGHT CAMERA INTERACTIVE DISPLAY
[0027] Embodiments of the present invention may be implemented using time- of-flight cameras. A time-of-flight camera has a built-in capability to detect distance information for each pixel of its image. This distance information for each pixel may consist of the distance from the camera to the object seen at that pixel. Time-of-flight cameras obtain this distance data by timing how long it takes an invisible light pulse to travel from an illuminator next to the camera to the object being imaged and then back to the camera. The light pulses can be rapidly repeated, allowing the time-of-flight camera to have a frame rate similar to a video camera. Time of flight cameras typically have a range of 1-2 meters at 30 frames per second. However, the range can be significantly increased by lengthening the exposure time, which lowers the frame rate. Manufacturers of time-of-flight cameras include Canesta Inc of Sunnyvale, CA, and 3DV systems of Israel.
[0028] The time-of-flight camera or cameras are positioned near a video display such that they can see objects on and in front of the display. In a preferred embodiment, the objects viewed by the camera or cameras consist of the hands, arms, and bodies of human users interacting with the display. The time-of-flight camera may work with any display (e.g., an LCD panel, a cathode- ray tube display, etc.) without modifications. Data from the time-of-flight camera or cameras may be processed in real time by a computer that uses the data to create interactive video effects on the display. The interactive video effects may be calibrated such that a user making a hand motion in front of a particular spot on the display will cause interactive video effects at that spot on the display. Thus, an "augmented reality" system is created, whereby users can use their hands and other parts of their bodies to manipulate virtual objects and cause digital visual effects on the display.
[0029] A single time-of-flight camera may be used. However, a single time-of- flight camera may not be able to detect objects that are blocked by objects closer to the camera. For example, if a user had one hand in front of the other relative to the camera, then the camera would not be able to see the other hand. Therefore, embodiments of the present invention may utilize multiple time-of-flight cameras, as shown in Figure 1. The multiple cameras, with multiple perspectives, reduce the chance that an object will not be seen.
[0030] With redundancy of cameras, there is no longer a need to worry about one camera not being able to detect all the objects because of one object occluding another object. For example, as shown in Figure 2, four time-of-flight cameras may be placed at the corners of a display, ensuring that the entire area of the display is interactive.
[0031] In some embodiments of the system, the interactive space in front of the display has a coordinate system for specifying the location of users or objects in front of the display. At a conceptual level, the system should know the position the user or object is on the screen and how far it is away from the screen it is. Having user/object data in the aforementioned coordinate system is very useful as it allows the easy creation of interactive effects that are physically matched to positions on the display, as well as having interactive effects that depend on the distance from the user/object to the display.
[0032] In order to use this time-of-flight implementation for one or more cameras, a coordinate transform is performed on each pixel of each time-of- flight camera to put it in a common coordinate space with each other camera and the display. One such space is defined by: (x, y) - the position of the point projected perpendicularly onto the display surface, and (z) - the perpendicular distance of the point from the display. This coordinate space transformation can be determined by looking at the angle and position of each camera relative to the display. Alternatively, the transformation may be determined by a calibration process, in which an object of known size, shape and position is placed in front of the display. By having each of the cameras image the object, the appropriate transformation function can be determined from points viewed by each camera into points in the. common coordinate space. If the camera coordinate transforms are done in real time, then a real-time picture of the area in front of the camera in 3D is achieved.
[0033] Variations of the system of the present invention may be implemented using one or multiple time-of-flight cameras with a display. Figure 3 illustrates a front view of an interactive display 3000 including multiple time-of-flight cameras, in accordance with an embodiment of the present invention. Interactive display 3000 includes time-of-flight cameras 3010 and 3020 that create an interactive area overlaying the surface of display 3030. It should be appreciated that multiple time-of-flight cameras may be placed at variations locations and positions around display 3030 to create the interactive area. The interactive area is bounded by the viewed region of the time-of-flight cameras, as shown by the dotted lines. For example, time-of-flight cameras may be placed in two corners of display 3030, four corners of display 3030 (as shown in Figure 2), next to each other on one side of display 3030 so that the interactive areas of each camera are substantially parallel, or any other variation for creating an interactive area overlaying display 3030. [0034] Figure 4 illustrates a side view of an interactive display 3100 including multiple time-of-flight cameras for detecting interactions at different distances from the display, in accordance with an embodiment of the present invention. Interactive display 3100 includes time-of-flight cameras 3110 and 3120 that create an interactive area overlaying and proximate the surface of display 3140. Time-of-flight camera 3130 creates an interactive area for a region farther from display 3140. Time-of-flight camera 3130 is operable to capture interactions farther from display 3140 than time-of-flight cameras 3110 and 3120. For instance, time-of-flight camera 3130 provides the ability to detect a person passing by interactive display 3100 at a certain distance and change the visual image shown on display 3140 to attract their attention. As the person approaches the display, they are picked up by time-of-flight cameras 3110 and 3120. In one embodiment, time-of-flight cameras 3110, 3120 and 3130 are the same type of time-of-flight camera, where time-of-flight camera 3130 operates at a lower frame rate and longer range than time-of-flight cameras 3110 and 3120.
[0035] Various other implementations of the present invention using time-of- flight cameras and a display are possible. In one embodiment, a time-of-flight camera or cameras are placed at a distance from the display to create the interactive region. For example, the display may be placed on a wall and the time-of-flight camera(s) may be placed on the opposite wall or on the ceiling, creating the interactive area between the time-of-flight camera (s) and the display. In another embodiment, the present invention may provide an interactive table, where the display is placed substantially horizontal and the time-of-flight camera(s) are above the table or to the side of the table. It should be appreciated that there are many different permutations of configurations of the present invention, and that the present invention is not limited to the described embodiments.
[0036] The distance information supplied by time-of-flight cameras can also be used to identify a user's hands. Identifying hands is important because hands are often used to provide specific interactions with the interactive display. In one embodiment, the hands are determined to be the points on each object viewed by the cameras that are closest to the display. In another embodiment, 3D shape recognition can be used to identify the hands. For example, simple cylinders or more complex skeletal models could be matched (using a variety of processes such as gradient descent) to users' hands, arms, and bodies. Other algorithms may also be used to identify the hands.
[0037] Figure 5A illustrates a side view of an interactive floor display 3200 including multiple time-of-flight cameras, in accordance with an embodiment of the present invention. Interactive floor display 3200 includes six time-of-flight cameras 3210a-f placed above display surface 3220. Each time-of-flight camera is directed at a different portion of display surface 3220. In one embodiment, the smaller the area imaged by a time-of-flight camera to create an interactive area, the higher the frame rate possible by the time-of-flight camera. Thus, the more time-of-flight cameras covering a region of display surface 3220, the better the performance of interactive floor display 3200. Figure 5B illustrates a top view of the surface covered by interactive floor display 3200 of Figure 5A. Time-of-flight cameras 3210a-f each image a respective region 3230a-f for creating the interactive area.
[0038] Figure 6 illustrates a side view of an interactive display 3300 including an infrared camera 3310 and a time-of-flight camera 3320, in accordance with an embodiment of the present invention. Infrared camera 3310 is operable to record images illuminated by infrared light of an infrared illuminator (not shown). Infrared camera 3310 and the infrared illuminator create interactive area 3315 above display surface 3330 (e.g., a floor). Time-of-flight camera 3320 is operable to create interactive area 3325 above display surface 3330. In one embodiment, interactive area 3325 covers a larger area than interactive area 3315. Infrared camera 3310 is operable to provide a succession of two- dimensional (2D) images of objects in interactive area 3315. In one embodiment, time-of-flight camera 3325 operates at a lower frame rate than infrared camera 3310 to provide three-dimensional image information of interactive area 3325 (including interactive area 3315). [0039] The three-dimensional image information may be used to label the two- dimensional image to provide high quality three-dimensional image information. In one embodiment, the three-dimensional image information of time-of-flight camera 3325 is superimposed on the two-dimensional image. Thus, the high resolution of the two-dimensional image is preserved as the 3D information is incorporated. In some embodiments, the 3D camera will have a lower frame rate than the 2D camera. However, using techniques such as feature tracking, 3D information can be estimated for two-dimensional images taken without a corresponding three-dimensional image. Since a feature tracking algorithm operating on the stream of 2D images can track the movement of several patches of an object, the 3D information from the most recent 3D camera image can be tracked to its current positions in the two- dimensional image, thus providing estimated 3D information even at times when no 3D image is available.
[0040] The three-dimensional image information can also be used to provide for person tracking. For example, the height of the highest point of an object can be used to identify the object as a person. In one embodiment, where the camera is on the ceiling and images a region on the floor, the point closest to the camera for each person can be identified. The height, size, and 3D shape of each object viewed by the cameras can be used to identify whether it is a person. This can be useful in determining how many people are interacting with interactive display 3300. Also, height information can be used to distinguish between adults and children, which may be useful information for marketing purposes. Other information can also be acquired, such as the orientation of people (e.g., standing, sitting, lying down). In one embodiment, since interactive area 3325 is bigger than interactive area 3315, people can be tracked that are not interacting with interactive area 3315. This information can be used to identify what portion of people passing interactive display 3300 actually end up interacting with interactive display 3300. It should be appreciate that person tracking and feature tracking may be performed without the use of an infrared camera, because the three-dimensional information provided by a time-of-flight camera is sufficient. However, the infrared camera is useful for getting higher-quality 2D outline and shape information, which may be useful in creating the interactive effects.
[0041] The 3D nature of the information from time-of-flight cameras provides an interactive display system of the present invention the ability to ignore or change the processing of data from objects that fall inside different volumetric regions in front of the display. Figure 7 illustrates an interactive display system 3400 including different volumetric interactive regions, in accordance with an embodiment of the present invention. While interactive display system 3400 is shown as a wall display, it should be appreciated that other types of displays, such as floor displays and table displays, can be implemented. Interactive display 3400 includes time-of-flight cameras 3410 and 3420 that create an interactive area including different interactive areas in front of the surface of display 3430. Since time-of-flight cameras include distance information for viewed objects, different interactive areas that include specific volumetric regions at particular distances from display 3430 may be defined.
[0042] Interactive display 3400 includes interactive regions that include specific volumetric regions. For instance, region 3440, also referred to herein as the touch region, is proximate to display 3430. In one embodiment, the touch region extends up to six inches out from display 3430. Region 3442, also referred to herein as the interactive region, extends from the end of the touch region. In one embodiment, the interactive region may be dynamic. For instance, if no people are interacting with interactive display 3400, region 3442 may extend several feet or to the end of a room to attempt to cause people to interact, thereby attracting their attention. Once a person engages interactive display 3400, the back of region 3442 may move in so as to ensure that other more distant people passing by do not affect the display. In one embodiment, the interactive region ends a particular distance (e.g., one foot) behind the nearest person in the interactive region. It should be appreciated that interactive display 3400 may include any number of volumetric interactive regions (e.g., region 3444), and is not limited to the described embodiments.
[0043] The touch zone also allows for a user to perform certain types of gestures. For example, display 3430 may display an object that a user can move through hand gestures. This movement of objects can be thought of as being analogous to the "drag and drop" operation commonly performed with a computer mouse, except in this case it is done using a hand gesture. In one embodiment, the user selects the object by moving their hand toward the object and crossing through the threshold separating the touch zone from the interactive zone. Staying in the touch zone, the user moves their hand to another location, and the object follows their hand's movement. The user can then release the object by pulling their hand back, moving it out of the touch zone. In another embodiment, the object can be selected by moving into and out of the touch zone such that the object may be moved while the hand is in the interactive zone. Then, the user can release the object by moving their hand back into and out of the touch zone. In another embodiment, the object can be selected by the user making a grabbing motion (closing their hand). The object could then be moved by the user moving their hand, and finally released by the user by opening their hand.
[0044] The touch zone can also be used for pressing virtual buttons or selecting objects on the display. In one embodiment, a user can press a button by moving their hand over the button's image on the display and then moving their hand towards the display until their hand crosses the threshold from the interactive zone to the touch zone. When their hand crosses this threshold, the button is pushed.
[0045] Information from the coordinate-transformed 3D data about objects or users in front of the display, such as their shape or outline, can serve as input to visual effects on the display. For example, the outline (from the display's point of view) of users in front of the display can be shown on the display to give users a sense of where they are as they are using the system.
[0046] In one embodiment, display 3430 changes in response to an object being selected and/or released. In one embodiment, display 3430 includes one or more icons that are associated with the hands of a user and are located at the position(s) on the screen corresponding to the user's hand(s). It should be appreciated that there can be multiple icons active on display 3430 at any time. The icon can transform (e.g., change shape, color, or size) when a user moves from the interactive zone to the touch zone, signifying a selection or release of an object.
USES
[0047] The interactive video display system can be used in many different applications. The system's capability to have touchscreen-like behavior for multiple users as well as full or partial body outline interaction increases its appeal for information interfaces which require more precise selection and manipulation of buttons and objects.
[0048] Uses of these interactive display systems include, but are not limited to, interactive video games in which users move their bodies to play the game, interactive menu, catalog, and browsing systems that let users browse through pages of informational content using gestures, systems that allow users to "try on" clothing using an image of themselves, pure entertainment applications in which images or outlines of users serve as input to a video effects system, interactive characters that interact with the motions of users in front of the screen, and virtual playlands and storybooks that users interact with by moving their bodies.
[0049] Other uses of the present invention include, but are not limited to: allowing users to customize or view available options for customizing the product on display, allowing the product on display to be ordered at the display, using either the display interface, a keyboard, a credit card swiper, or a combination of the three, comparing the features of multiple products on the display, showing combinations or compatibilities between multiple products on the display, and placing a product in different virtual settings on the screen to demonstrate the features (e.g., water, forest, asphalt, etc.)
PERIPHERALS
[0050] These interactive display systems can incorporate additional inputs and outputs, including, but not limited to, microphones, touchscreens, keyboards, mice, radio frequency identification (RFID) tags, pressure pads, cellular telephone signals, personal digital assistants (PDAs), and speakers. [0051] These interactive display systems can be tiled together to create a single larger screen or interactive area. Tiled or physically separate screens can also be networked together, allowing actions on one screen to affect the image on another screen.
[0052] In an exemplary implementation, the present invention is implemented using a combination of hardware and software in the form of control logic, in either an integrated or a modular manner. Based on the disclosure and teachings provided herein, a person of ordinary skill in the art will know of other ways and/or methods to implement the present invention.
[0053] In one exemplary aspect, the present invention as described above provides a system that allows a camera to view an area in front of a display. In a related invention, a system is provided to create a reactive space in front of the display. The present invention can be used to capture information from the reactive space.
[0054] It is understood that the examples and embodiments described herein are for illustrative purposes only and that various modifications or changes in light thereof will be suggested to persons skilled in the art and are to be
included, within the purview of this application and scope ofthe appended claims. All publications, patents, and patent applications cited herein are hereby incorporated by reference for all purposes in their entirety. [0055] Various embodiments of the invention, an interactive video display system, are thus described. While the present invention has been described in particular embodiments, it should be appreciated that the invention should not be construed as limited by such embodiments, but rather construed according to the below claims.

Claims

CLAIMSWhat is claimed is:
1. An interactive video display system comprising: a display screen for displaying a visual image for presentation to a user; a camera for detecting an object in an interactive area located in front of said display screen, said camera operable to capture three-dimensional information about said object; and a computer system for directing said display screen to change said visual image in response to said object.
2. The interactive video display system as recited in Claim 1 wherein said camera is a time-of-flight camera.
3. The interactive video display system as recited in Claim 1 wherein said display screen is a cathode ray tube.
4. The interactive video display system as recited in Claim 1 wherein said display screen is a flat-panel display screen.
5. The interactive video display system as recited in Claim 4 wherein said flat-panel display screen is a liquid crystal display screen.
6. The interactive video display system as recited in Claim 1 wherein said display screen is a plasma screen.
7. The interactive video display system as recited in any of Claims 1-6, further comprising a projector for projecting said visual image and wherein said display screen is a screen for displaying said visual image.
8. The interactive video display system as recited in any of Claims 1-7, further comprising a second camera for detecting said object, said second camera operable to capture three-dimensional information about said object.
9. The interactive video display system as recited in Claim 8 wherein said second camera is a time-of-flight camera.
10. The interactive video display system as recited in Claim 8 wherein said camera and said second camera operate in conjunction to create said interactive area.
11. The interactive video display system as recited in Claim 8 wherein said camera is operable to create said interactive area proximate said display screen, and wherein said second camera is operable to create a second interactive area farther from said display screen than said interactive area.
12. The interactive video display system as recited in any of Claims 1-7, further comprising an infrared camera for creating a second interactive area.
13. The interactive video display system as recited in Claim 12 wherein said camera provides said computer system with three-dimensional information at a first frame rate and said infrared camera provides said computer system with two-dimensional information at a second frame rate, wherein said first frame rate is slower than said second frame rate, such that said computer system is operable to generate high quality three-dimensional information based on said three-dimensional information and said two- dimensional information.
14. The interactive video display system as recited in Claim 13 wherein said computer system can identify a portion of people near said interactive video display system actively interact with said interactive video display system by determining people in said second interactive area that do not enter said interactive area.
15. The interactive video display system as recited in any of Claims 1-14, wherein said three-dimensional information is used for person tracking.
16. The interactive video display system as recited in any of Claims 1-15, wherein said interactive area comprises a plurality of interactive regions including volumetric regions at particular distances from said display screen.
17. The interactive video display system as recited in Claim 16 wherein a first volumetric region extends from said display screen to a particular distance from said display screen, and a second volumetric region extends from said particular distance away from said display screen.
18. The interactive video display system as recited in Claim 17 wherein said second volumetric region extends to a second particular distance away from said display screen.
19. The interactive video display system as recited in Claim 18 wherein said second particular distance is dynamic.
20. The interactive video display system as recited in Claim 16 wherein a user may interact with said visual image by moving into and out of said first volumetric region.
21. The interactive video display system as recited in any of Claims 1 -20, 'wherein said object is a user.
22. The interactive video display system as recited in any of Claims 1 -7 or 15-21, wherein said camera is separate from said display screen, such that said interactive area is between said camera and said display screen.
23. The interactive video display system as recited in Claim 22 further comprising a second camera for detecting said object, wherein said second camera is separate from said display screen.
24. The interactive video display system as recited in Claim 23 wherein said camera and said second camera operate in conjunction to create said interactive area.
25. The interactive video display system as recited in Claim 23 wherein said camera is operable to create said interactive area proximate said display screen, and wherein said second camera is operable to create a second interactive area farther from said display screen than said interactive area.
26. The interactive video display system as recited in Claim 22 further comprising an infrared camera for creating a second interactive area.
27. The interactive video display system as recited in any of Claims 1-26, wherein
said three-dimensional information is for providing interactions of said object physically matched to a location on said display screen.
28. A method for presenting an interactive visual image using an interactive video display system, said method comprising: displaying a visual image on a display screen, wherein said visual image is for presentation to a user on a front side of said display screen; detecting interaction of an object with said visual image using a three- dimensional camera, said three-dimensional camera operable to capture three-dimensional information about said object; and changing said visual image in response to said interaction.
29. The method as recited in Claim 28 wherein said three- dimensional camera is a time-of-flight camera.
30. An interactive video display system comprising: a display screen for displaying a visual image for presentation to a user; a time-of-flight camera for detecting an object in an interactive area located in front of said display screen, said time-of-flight camera operable to capture three-dimensional information about said object; and a computer system for directing said display screen to change said visual image in response to said object, wherein said three-dimensional information is for providing interactions of said object physically matched to a location on said display screen.
PCT/US2005/008984 2004-03-18 2005-03-18 Interactive video display system WO2005091651A2 (en)

Applications Claiming Priority (8)

Application Number Priority Date Filing Date Title
US55452004P 2004-03-18 2004-03-18
US60/554,520 2004-03-18
US10/946,414 US7710391B2 (en) 2002-05-28 2004-09-20 Processing an image utilizing a spatially varying pattern
US10/946,084 US20050122308A1 (en) 2002-05-28 2004-09-20 Self-contained interactive video display system
US10/946,263 US8035612B2 (en) 2002-05-28 2004-09-20 Self-contained interactive video display system
US10/946,414 2004-09-20
US10/946,084 2004-09-20
US10/946,263 2004-09-20

Publications (2)

Publication Number Publication Date
WO2005091651A2 true WO2005091651A2 (en) 2005-09-29
WO2005091651A3 WO2005091651A3 (en) 2006-01-12

Family

ID=34679515

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2005/008984 WO2005091651A2 (en) 2004-03-18 2005-03-18 Interactive video display system

Country Status (1)

Country Link
WO (1) WO2005091651A2 (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2007105246A2 (en) 2006-03-10 2007-09-20 Link Formazione S.R.L. Interactive multimedia system
EP1845490A1 (en) * 2006-04-06 2007-10-17 Vis-A-Pix GmbH System and method for automatic actuation of a virtual recognition event
US7710391B2 (en) 2002-05-28 2010-05-04 Matthew Bell Processing an image utilizing a spatially varying pattern
US7809167B2 (en) 2003-10-24 2010-10-05 Matthew Bell Method and system for processing captured image information in an interactive video display system
US7834846B1 (en) 2001-06-05 2010-11-16 Matthew Bell Interactive video display system
EP2348390A1 (en) * 2010-01-20 2011-07-27 Evoluce Ag Input device with a camera
US8035612B2 (en) 2002-05-28 2011-10-11 Intellectual Ventures Holding 67 Llc Self-contained interactive video display system
US8081822B1 (en) 2005-05-31 2011-12-20 Intellectual Ventures Holding 67 Llc System and method for sensing a feature of an object in an interactive video display
US8098277B1 (en) 2005-12-02 2012-01-17 Intellectual Ventures Holding 67 Llc Systems and methods for communication between a reactive video system and a mobile communication device
US8159682B2 (en) 2007-11-12 2012-04-17 Intellectual Ventures Holding 67 Llc Lens system
US8199108B2 (en) 2002-12-13 2012-06-12 Intellectual Ventures Holding 67 Llc Interactive directed light/sound system
US8230367B2 (en) 2007-09-14 2012-07-24 Intellectual Ventures Holding 67 Llc Gesture-based user interactions with status indicators for acceptable inputs in volumetric zones
US8259163B2 (en) 2008-03-07 2012-09-04 Intellectual Ventures Holding 67 Llc Display with built in 3D sensing
US8300042B2 (en) 2001-06-05 2012-10-30 Microsoft Corporation Interactive video display system using strobed light
US8487866B2 (en) 2003-10-24 2013-07-16 Intellectual Ventures Holding 67 Llc Method and system for managing an interactive video display system
US8595218B2 (en) 2008-06-12 2013-11-26 Intellectual Ventures Holding 67 Llc Interactive display management systems and methods
US9128519B1 (en) 2005-04-15 2015-09-08 Intellectual Ventures Holding 67 Llc Method and system for state-based control of objects
ES2563105A1 (en) * 2014-09-10 2016-03-10 Universidad De Valladolid Interactive system that combines movement, sound and color as therapeutic support to develop perceptive-motor skills, stimulate the sensory-perceptive area, encourage self-expression, diagnose emotional processes of the user (Machine-translation by Google Translate, not legally binding)
ES2699999A1 (en) * 2018-04-25 2019-02-13 Mendez Francisco Jose Cuadrado PROCEDURE FOR THE CREATION AND HANDLING OF MUSIC AND SOUND FROM THE INTERACTION WITH TANGIBLE OBJECTS (TUI) AND A SYSTEM OF INCREASED REALITY, WITH SPECIAL APPLICABILITY TO THE SCOPE OF TEACHING. (Machine-translation by Google Translate, not legally binding)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0913790A1 (en) * 1997-10-29 1999-05-06 Takenaka Corporation Hand pointing apparatus
US20010012001A1 (en) * 1997-07-07 2001-08-09 Junichi Rekimoto Information input apparatus
US20040046744A1 (en) * 1999-11-04 2004-03-11 Canesta, Inc. Method and apparatus for entering data using a virtual input device
US20040046736A1 (en) * 1997-08-22 2004-03-11 Pryor Timothy R. Novel man machine interfaces and applications
WO2004055776A1 (en) * 2002-12-13 2004-07-01 Reactrix Systems Interactive directed light/sound system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20010012001A1 (en) * 1997-07-07 2001-08-09 Junichi Rekimoto Information input apparatus
US20040046736A1 (en) * 1997-08-22 2004-03-11 Pryor Timothy R. Novel man machine interfaces and applications
EP0913790A1 (en) * 1997-10-29 1999-05-06 Takenaka Corporation Hand pointing apparatus
US20040046744A1 (en) * 1999-11-04 2004-03-11 Canesta, Inc. Method and apparatus for entering data using a virtual input device
WO2004055776A1 (en) * 2002-12-13 2004-07-01 Reactrix Systems Interactive directed light/sound system

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8300042B2 (en) 2001-06-05 2012-10-30 Microsoft Corporation Interactive video display system using strobed light
US7834846B1 (en) 2001-06-05 2010-11-16 Matthew Bell Interactive video display system
US8035614B2 (en) 2002-05-28 2011-10-11 Intellectual Ventures Holding 67 Llc Interactive video window
US7710391B2 (en) 2002-05-28 2010-05-04 Matthew Bell Processing an image utilizing a spatially varying pattern
US8035612B2 (en) 2002-05-28 2011-10-11 Intellectual Ventures Holding 67 Llc Self-contained interactive video display system
US8035624B2 (en) 2002-05-28 2011-10-11 Intellectual Ventures Holding 67 Llc Computer vision based touch screen
US8199108B2 (en) 2002-12-13 2012-06-12 Intellectual Ventures Holding 67 Llc Interactive directed light/sound system
US8487866B2 (en) 2003-10-24 2013-07-16 Intellectual Ventures Holding 67 Llc Method and system for managing an interactive video display system
US7809167B2 (en) 2003-10-24 2010-10-05 Matthew Bell Method and system for processing captured image information in an interactive video display system
US9128519B1 (en) 2005-04-15 2015-09-08 Intellectual Ventures Holding 67 Llc Method and system for state-based control of objects
US8081822B1 (en) 2005-05-31 2011-12-20 Intellectual Ventures Holding 67 Llc System and method for sensing a feature of an object in an interactive video display
US8098277B1 (en) 2005-12-02 2012-01-17 Intellectual Ventures Holding 67 Llc Systems and methods for communication between a reactive video system and a mobile communication device
WO2007105246A3 (en) * 2006-03-10 2007-11-22 Link Formazione S R L Interactive multimedia system
WO2007105246A2 (en) 2006-03-10 2007-09-20 Link Formazione S.R.L. Interactive multimedia system
EP1845490A1 (en) * 2006-04-06 2007-10-17 Vis-A-Pix GmbH System and method for automatic actuation of a virtual recognition event
US9811166B2 (en) 2007-09-14 2017-11-07 Intellectual Ventures Holding 81 Llc Processing of gesture-based user interactions using volumetric zones
US9058058B2 (en) 2007-09-14 2015-06-16 Intellectual Ventures Holding 67 Llc Processing of gesture-based user interactions activation levels
US8230367B2 (en) 2007-09-14 2012-07-24 Intellectual Ventures Holding 67 Llc Gesture-based user interactions with status indicators for acceptable inputs in volumetric zones
US10990189B2 (en) 2007-09-14 2021-04-27 Facebook, Inc. Processing of gesture-based user interaction using volumetric zones
US10564731B2 (en) 2007-09-14 2020-02-18 Facebook, Inc. Processing of gesture-based user interactions using volumetric zones
US9229107B2 (en) 2007-11-12 2016-01-05 Intellectual Ventures Holding 81 Llc Lens system
US8159682B2 (en) 2007-11-12 2012-04-17 Intellectual Ventures Holding 67 Llc Lens system
US8810803B2 (en) 2007-11-12 2014-08-19 Intellectual Ventures Holding 67 Llc Lens system
US8259163B2 (en) 2008-03-07 2012-09-04 Intellectual Ventures Holding 67 Llc Display with built in 3D sensing
US9247236B2 (en) 2008-03-07 2016-01-26 Intellectual Ventures Holdings 81 Llc Display with built in 3D sensing capability and gesture control of TV
US10831278B2 (en) 2008-03-07 2020-11-10 Facebook, Inc. Display with built in 3D sensing capability and gesture control of tv
US8595218B2 (en) 2008-06-12 2013-11-26 Intellectual Ventures Holding 67 Llc Interactive display management systems and methods
EP2348390A1 (en) * 2010-01-20 2011-07-27 Evoluce Ag Input device with a camera
ES2563105A1 (en) * 2014-09-10 2016-03-10 Universidad De Valladolid Interactive system that combines movement, sound and color as therapeutic support to develop perceptive-motor skills, stimulate the sensory-perceptive area, encourage self-expression, diagnose emotional processes of the user (Machine-translation by Google Translate, not legally binding)
ES2699999A1 (en) * 2018-04-25 2019-02-13 Mendez Francisco Jose Cuadrado PROCEDURE FOR THE CREATION AND HANDLING OF MUSIC AND SOUND FROM THE INTERACTION WITH TANGIBLE OBJECTS (TUI) AND A SYSTEM OF INCREASED REALITY, WITH SPECIAL APPLICABILITY TO THE SCOPE OF TEACHING. (Machine-translation by Google Translate, not legally binding)

Also Published As

Publication number Publication date
WO2005091651A3 (en) 2006-01-12

Similar Documents

Publication Publication Date Title
US7170492B2 (en) Interactive video display system
US7348963B2 (en) Interactive video display system
WO2005091651A2 (en) Interactive video display system
US10990189B2 (en) Processing of gesture-based user interaction using volumetric zones
US10761612B2 (en) Gesture recognition techniques
US6195104B1 (en) System and method for permitting three-dimensional navigation through a virtual reality environment using camera-based gesture inputs
US8081822B1 (en) System and method for sensing a feature of an object in an interactive video display
CN104471511B (en) Identify device, user interface and the method for pointing gesture
CN103858074B (en) The system and method interacted with device via 3D display device
Lee et al. Occlusion based interaction methods for tangible augmented reality environments
KR101481880B1 (en) A system for portable tangible interaction
US20110164032A1 (en) Three-Dimensional User Interface
CN108885533A (en) Combine virtual reality and augmented reality
US20100259610A1 (en) Two-Dimensional Display Synced with Real World Object Movement
US20060248462A1 (en) Remote control of on-screen interactions
CN105229720A (en) Display control unit, display control method and recording medium
JP2008521110A (en) Personal device with image capture function for augmented reality resources application and method thereof
NZ525717A (en) A method of tracking an object of interest using multiple cameras
EP1695196A2 (en) Self-contained interactive video display system
CN113168228A (en) Systems and/or methods for parallax correction in large area transparent touch interfaces
Wischgoll Display systems for visualization and simulation in virtual environments
CN107452288A (en) Electronic device and electronic advertisement playing and interaction method
Chan et al. Gesture-based interaction for a magic crystal ball
WO2016102948A1 (en) Coherent touchless interaction with stereoscopic 3d images
Beardsley et al. Handheld projectors for mixing physical and digital textures

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase in:

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS RULE 69(1) EPC (EPO FORM 1205A 04.12.2006)

122 Ep: pct application non-entry in european phase