US20090235151A1 - Method and apparatus for associating the color of an object with an event - Google Patents

Method and apparatus for associating the color of an object with an event Download PDF

Info

Publication number
US20090235151A1
US20090235151A1 US12/472,761 US47276109A US2009235151A1 US 20090235151 A1 US20090235151 A1 US 20090235151A1 US 47276109 A US47276109 A US 47276109A US 2009235151 A1 US2009235151 A1 US 2009235151A1
Authority
US
United States
Prior art keywords
color
color value
video frames
video frame
gui
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/472,761
Inventor
Daniel Bates
Jorge Geaga
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Creative Frontier Inc
Original Assignee
Creative Frontier Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Creative Frontier Inc filed Critical Creative Frontier Inc
Priority to US12/472,761 priority Critical patent/US20090235151A1/en
Publication of US20090235151A1 publication Critical patent/US20090235151A1/en
Priority to US16/168,093 priority patent/US20190080193A1/en
Priority to US16/454,743 priority patent/US20190318187A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/74Browsing; Visualisation therefor
    • G06F16/748Hypervideo
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/4722End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting additional data associated with the content
    • H04N21/4725End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting additional data associated with the content using interactive regions of the image, e.g. hot spots
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/858Linking data to content, e.g. by linking an URL to a video object, by creating a hotspot
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/858Linking data to content, e.g. by linking an URL to a video object, by creating a hotspot
    • H04N21/8583Linking data to content, e.g. by linking an URL to a video object, by creating a hotspot by creating hot-spots

Definitions

  • the present invention relates generally to event processing, and in particular to associating the color of an object with an event.
  • Portions of the disclosure of this patent document contain material that is subject to copyright protection.
  • the copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure as it appears in the Patent and Trademark Office file or records, but otherwise reserves all copyright rights whatsoever.
  • a user can interact with the device in various ways.
  • One way is to use a mouse.
  • the user utilizes the mouse to position a pointer somewhere on a screen.
  • the user then depresses a mouse button and depending on where the pointer is, something pre-programmed will happen if that area of the screen is “hot”.
  • the user can position the mouse pointer over a hot spot “hyperlinked” to a web page and depress the mouse button. This will cause the computer, set-top box, or other device to send a signal to a server to retrieve that web page and display it on the computer screen.
  • Event processing refers to the association between some type of “action” and the occurrence of an “event”. Such a term is used herein in the context of a user's computing experience. The user performs an action followed by an event, followed by a series of other actions followed by other events.
  • An action occurs when a user positions a cursor in a specified location on a computer screen, for instance using a mouse, and depressing one of the mouse buttons.
  • This example shows a common type of action, but many other actions exist as well within the context of computer event processing, including the input of text, a voice command, or any other type of action.
  • the example action may take place in any computing environment, but one common environment that the action occurs is in a web browser.
  • a web browser is a graphical user interface that enables users to utilize the Internet, for instance by viewing web pages. Normally the user will position the mouse pointer over a “hyperlink” and depress the mouse button there. This, in turn, triggers an event.
  • the event in this example is to cause a server to retrieve another web page.
  • the user interacts with the Internet by a series of actions followed by events.
  • the information network known as the “Internet” has become increasingly popular.
  • the Internet provides a body of information which may be accessed by users via computers for such purposes as business, education, and entertainment.
  • the Internet includes a large collection of interconnected documents, images, sounds, videos, and other forms of media which are stored in computers in a system known as the World Wide Web or simply “the web”.
  • Internet resources are typically accessed in a two-way environment via a network connection.
  • connectivity to the network may be via a conventional twisted-pair telephone line which has a relatively low data-carrying capacity (e.g., bandwidth), or via a higher bandwidth path, which may comprise optical fiber, coaxial cable, ISDN, DSL, wireless connections, or other transmission mechanisms.
  • a network destination e.g., an Internet site
  • the network destination typically includes a computer known as a server.
  • the server then sends a request signal to a source function, which may be a memory which is coupled to the server.
  • the source function includes the user-requested information which may comprise, for example, text, graphics, audio and/or video data.
  • the source function provides the requested information to the client.
  • HTML Hypertext Markup Language
  • HTML Hypertext Markup Language
  • HTML is a system for marking documents to indicate how the document should be displayed, and how various documents should be linked together.
  • HTML is a form of Standard Generalized Markup Language (SGML), defined by the International Standards Organization. HTML specifies the grammar and syntax of markup tags which are inserted into a data file to define how the data will be presented when read by a computer program known as a “web browser”. Conventional web browsers include Internet Explorer, Netscape Navigator, and others.
  • the data file which is typically stored on a server, includes one or more web pages which are visited by users who have computers which may ran different browsers.
  • HTML data output from the server is downloaded to the client computer.
  • the client computer's browser processes the data to format a layout for the page so the page may be viewed by the user on a computer screen.
  • HTML tags provide text formatting, hypertext links to other pages, and links to sound and image elements. HTML tags also define input fields for interactive web pages.
  • HTML application is made available to users on the web by storing the HTML file in a directory that is accessible to a server.
  • a server is typically a web server which conforms to a web browser-supported protocol known as Hypertext Transfer Protocol (HTTP). Servers that conform to other protocols such as the File Transfer Protocol (FTP) or GOPHER may also be used, but do not support interactive HTML files.
  • HTTP Hypertext Transfer Protocol
  • FTP File Transfer Protocol
  • GOPHER GeneHER
  • HTTP defines a set of rules that servers and browsers follow when communicating with each other.
  • the process begins when a user accesses an icon in an HTML page which is the anchor of a hyperlink, (for instance, by positioning a cursor on the icon and depressing a mouse button), or the user inputs a Uniform Resource Locator (URL) to his or her web browser, described below.
  • a connection is then made to the server at the address and port number specified by the URL.
  • the browser sends a request to retrieve an object from the server, or to post data to an object on the server.
  • the server sends a response to the browser including a status code and the response data.
  • the connection between the browser and server is then closed.
  • a URL is a unique address which identifies virtually all files and resources on the Internet.
  • a URL has the form: method://server:port/path/file# anchor.
  • the “method” of accessing the resource is the web browser-supported protocol, and may include, for example, HTTP, FTP, GOPHER, TELNET, NEWS, or MAILTO.
  • the “server:port” indicates the name of the server which is providing the resource, and is alternatively known as the Internet domain name. For example, many businesses will use their business name as part of the server field.
  • the port designation is the port number on the server, but is usually not used since a default port is assumed.
  • the “path” indicates the directory path to the resource.
  • the file indicates the file name of the resource.
  • the “anchor” indicates the named element in the HTML document. Not all fields are required.
  • the web evolved As the web evolved, other types of event processing have emerged. In particular, the web evolved to have a more active nature.
  • the text-based web became merged with images.
  • the images resided in various formats and could also be used to associate an action with an event.
  • the images suffered the same drawback as the text and essentially performed the same function. For instance, the images were used in the same event context (e.g., depressing a mouse button on the image) and the same actions resulted (e.g., the image caused a server to send another web page to the user).
  • each of the 100 images is an image map specifying that in frame 1 the boundary of the baton is in a first position and in frame 2 , the boundary of the baton is in a second position.
  • the prior art essentially draws a wire frame around the baton in each frame. Therefore, 100 wire frames must be drawn in varying locations.
  • the baton in a moving image varies in position and angle. This in turn varies the shape of the wire frame in each frame of the image.
  • the prior scheme is extremely disadvantageous because it requires a tremendous amount of manual pre-processing to draw 100 varied size wire frames. Such pre-processing contains a large lag on overhead. This scheme also must store each and every potential action in a computer and then wait for the action to occur. This technique uses a large amount of the computer's storage space and a large amount of time to create the wire frames. This scheme, in addition, is bound to a web browser implementation and may only trigger events relating to the accessing of a web page at a specified URL.
  • the present invention provides a method and apparatus for associating the color of an object with an event.
  • the invention takes a streaming or static digital image frames and associates an active, dynamic, or static object with an event by its color.
  • an event is the retrieval of a document at a specified URL and the display of the document in a web browser.
  • an event includes the retrieval of textual or audio information associated with the object.
  • the event comprises switching to a separate video stream in an environment using streaming media, such as a digital video disc (DVD) player or recorder, or other suitable video environment.
  • streaming media such as a digital video disc (DVD) player or recorder, or other suitable video environment.
  • the invention defines the color in a location (termed a region of interest) by its digital color mathematical properties. For instance, one embodiment uses the RGB standard which separates the red, green, and blue properties of the color as numerical values. Another embodiment also uses the HSV characteristics of the color, which define hue, saturation, and value as components of color. Another embodiment uses both RGB and HSV characteristics. In one embodiment, the color characteristics are obtained from a digital RGB frame buffer
  • the event links to another URL (i.e., to facilitate a transaction over the Internet) or in the case of a DVD or other video environment, the event causes the video stream to transition to a different data track (i.e., to pull up different information) or to a different video stream
  • the invention uses an edge detection scheme to help define the boundaries for the colors in a digital RGB frame.
  • the algorithm developed gives those pixels an “edge” which is not visible to the user, nor drawn on the screen, nor pre-processed prior to the user viewing the images formed by the pixels.
  • the values of the pixels defines an edge rather than the actual drawing of a wire frame on the screen.
  • one or more embodiments of the present invention create the associations between objects and events dynamically in that the data which creates the association is a property of the image itself (i.e., its color values).
  • the present invention requires no alteration or pre-processing of the image to create the associations.
  • FIG. 1 is a flowchart showing how to associate the color of an object with an event according to an embodiment of the present invention.
  • FIG. 2 is a flowchart showing how to define the color characteristics of an object according to an embodiment of the present invention.
  • FIG. 3 is a flowchart showing how to define the color characteristics of an object according to another embodiment of the present invention.
  • FIG. 4 is a flowchart showing how to define the color characteristics of an object according to another embodiment of the present invention.
  • FIG. 5 is a flowchart showing how to associate the color of an object with an event using a table according to an embodiment of the present invention.
  • FIG. 6 is a computer execution environment where one or more embodiments of the present invention may be implemented.
  • FIG. 7 shows an embodiment of the present invention where the frame buffer is utilized.
  • FIG. 8 shows an embodiment of the present invention where color ranges are utilized.
  • the invention is a method and apparatus for associating the color of an object with an event.
  • numerous specific details are set forth to provide a more thorough description of embodiments of the invention. It is apparent, however, to one skilled in the art, that the invention may be practiced without these specific details. In other instances, well known features have not been described in detail so as not to obscure the invention.
  • One or more embodiments of the present invention are directed to enhancing a user's ability to interact with a computer, set-top box, interactive television set, or other processing device.
  • the user provides input on some area of a display device, for instance by depressing a mouse button at a specified location.
  • the invention determines where the user has depressed the mouse button and what color value appears on the display device at the time the user has depressed the mouse button. Then, the invention initiates an event that corresponds to that color value.
  • the invention takes a streaming or static media and associates an active, dynamic, or static object with an event by its given color.
  • the event may be, for instance, the retrieval of a document at a specified URL in a web browser or the retrieval of other information in non-web browser implementations.
  • Other implementations include, for instance, the viewing of media, via a television, DVD or other media player, set-top devices, web-TV, interactive TV, and video-games.
  • a web page implementing streaming video Take, for example, a web page implementing streaming video.
  • a game show host is wearing a green shirt. If a user provides input to the screen by depressing a mouse with the cursor in a location where the green shirt is located, the invention links to a URL associated with the green shirt.
  • the moving color may substitute for a conventional, static, text link.
  • the game show host with the green shirt may be in a television show. In this embodiment, providing input on the moving green shirt may initiate a different event, such as providing information about the shirt or the game show host.
  • FIG. 1 One embodiment of the present invention is shown in FIG. 1 .
  • an action occurs.
  • the location of the action is determined. Locating a region of interest is well known in the art.
  • One method to locate a region of interest is to obtain an ordered pair corresponding to the location of the cursor at the moment of the action.
  • a computer screen may be represented as a two-dimensional graph wherein the axis of the graph resides in the lower left hand corner of the computer screen.
  • all ordered pairs will correspond to positive values of x,y pairs on the graph.
  • the value associated with a color in the region of interest is determined at step 120 . Thereafter, at step 130 , an event associated with that color is initiated.
  • the invention defines the color by its mathematical properties. For instance, one embodiment uses the RGB standard which separates the red, green, and blue properties of the color as numerical values. Another embodiment also uses the HSV characteristics of the color, which define hue, saturation, and value as components of color. HSV characteristics further define the properties of a color, for instance V represents an analog of the intensity of a color or how bright the color appears. Another embodiment of the present invention uses both RGB and HSV characteristics.
  • step 200 it is determined whether an action has occurred. If no action occurred, step 200 repeats until an action occurs. If the action has occurred, it is determined where the action has occurred at step 210 .
  • pixel N The location is shown herein with respect to a single pixel, “pixel N’.
  • a red component of a color for pixel A is determined.
  • a green component of the color for pixel A is determined.
  • a blue component of the color for pixel A is determined.
  • the color characteristics are combined to form a color value for pixel A
  • step 250 it is determined whether any event is associated with the color value at step 255 . If no event is associated with that color value, no event is initiated and the process repeats at step 200 . If an event is associated with the color value, then at step 260 an event associated with the color value is initiated.
  • step 300 it is determined whether an action has occurred. If no action occurred, step 300 repeats until an action occurs. If the action has occurred, it is determined where the action has occurred at step 310 . The location is shown herein with respect to a single pixel, “pixel B”.
  • pixel B a hue component of a color for pixel B is determined.
  • step 330 a saturation component of the color for pixel B is determined.
  • step 340 a value component of the color for pixel B is determined.
  • step 350 the color characteristics are combined to form a color value for pixel B.
  • step 355 it is determined whether any event is associated with the color value at step 355 . If no event is associated with that color value, no event is initiated and the process repeats at step 300 . If an event is associated with the color value, then at step 360 an event associated with the color value is initiated.
  • step 400 it is determined whether an action has occurred. If no action occurred, step 400 repeats until an action occurs. If the action has occurred, it is determined where the action has occurred at step 410 . The location is shown herein with respect to a single pixel, “pixel C’.
  • a hue component of a color for pixel C is determined.
  • a saturation component of the color for pixel C is determined.
  • a value component of the color for pixel C is determined.
  • step 450 a red component of a color for pixel C is determined.
  • step 460 a green component of the color for pixel C is determined.
  • step 470 a blue component of the color for pixel C is determined.
  • step 480 the color characteristics are combined to form a color value for pixel B.
  • step 480 it is determined whether any event is associated with the color value at step 485 . If no event is associated with that color value, no event is initiated and the process repeats at step 400 . If an event is associated with the color value, then at step 490 an event associated with the color value is initiated.
  • Frame Buffer In a computer or other comparable device, images are often used as a form of output. A common way that such devices handle images or streams of images (e.g., video) is by using a “frame buffer”. For offline pre-processing or for real-time processing using high speed computers, the RGB frame buffers reside in memory. Alternatively, a special high speed video board can be used to process an RGB frame buffer for real-time processing.
  • a video signal for instance, comprises a series of still images shown sequentially where the series of images appears so quickly that a user perceives the series of pictures to be a continuous stream of motion.
  • data relating to the image is stored in the frame buffer. Such data includes, for instance, the color values associated with each pixel at each location on the screen.
  • FIG. 7 shows an embodiment of the present invention where the frame buffer is utilized.
  • an action occurs.
  • the location of the action (a region of interest) is determined.
  • the invention access the frame buffer to obtain color values at the location of the action when it occurred. Then, it is determined whether any event is associated with the color value at step 730 . If no event is associated with that color value, no event is initiated and the process repeats at step 700 . If an event is associated with the color value, then at step 740 an event associated with the color value is initiated.
  • one event is a link to another URL (i.e., to facilitate a transaction over the Internet).
  • one event may comprise the retrieval of a web page from a server computer connected to the Internet.
  • an event may be the retrieval of a text page. For instance, assume the user is watching a DVD program showing several of the great pyramids of Egypt. If the user causes an event to occur in a region of interest defined by a specific pyramid, then the event may comprise the retrieval of a text page telling the user exactly which pyramid this is, when it was built, the history associated with this particular pyramid, and any other pertinent information. Thus, the event would cause the DVD to move to a different data track (i.e., to pull up different information).
  • another event may comprise the retrieval of audio information, for instance a narrator providing information similar to that given in the text track in the above example.
  • the event may comprise the retrieval of a new media stream, such as the close up of the pyramid, alternate angles showing different views of the same pyramid, or views from inside the pyramid, for example.
  • an event comprises the retrieval of a data track.
  • a data track comprises information associated with the color acted upon. Assume for instance, that this embodiment of the present invention takes place using an interactive television set. While the interactive television set is operating, a game show host appears wearing a green shirt. Then assume, for instance, that a user caused an action to occur at the location of the green shirt. In this instance, the data track may be information about the shirt, the game show host, or any other relevant information.
  • Whether an event is associated with a color value may be determined in a plurality of ways.
  • One such way is to use a table.
  • An example of one embodiment of a table is shown in Table 1.
  • the table may be arranged in a plurality of ways and may be accessed in methods well known to those skilled in the art, including hash tables, hash functions, pointers, indices, or any other suitable form of table lookup.
  • different color values may initiate the same event.
  • every color value on an output device may be associated with the same event.
  • FIG. 5 shows an embodiment of the present invention where a table is implemented.
  • an action occurs.
  • the location of the action (a region of interest) is determined.
  • the value associated with a color in the region of interest is determined.
  • a table is searched to locate the color value to obtain an event that is paired with the color value at step 530 .
  • regions of interest i.e., those regions where a user has performed an action
  • a color range for instance, includes those shades of blue that correspond to the body of a bluebird in a video stream.
  • Realistic image and video streams will not have a uniform shade of blue at all areas of the bluebird's body since light and other factors will cause natural variations in the perceived color reflected from an object. In this embodiment, this reality is accounted for, which gives this embodiment specific application in dynamic and high-quality video streams.
  • Color ranges are defined by color characteristics. For instance, a color range may translate into values on an RGB or HSV scale, or could be combinations of both. Typically, in a computer system such values range from 0-255. Thus, the bluebird's body may be defined as a region of interest where the color range includes a B value on the RGB scale of 50-75, for instance.
  • the colors may be very similar, yet they may be aspects of different components of the image and associated with different events.
  • the ranges are defined with respect to a plurality of color characteristics including RGB and HSV.
  • a range may be defined where the R is selected to have a range of 50-80, the G is selected to have a range of 150-160, and the B is selected to have a range of 200-220.
  • An example color is selected by the user where the example color has RGB characteristics of 60, 155, and 205 respectively.
  • Another color on the screen that is associated with a different event may appear to be substantially identical to the example color but has RGB characteristics of 190, 100,33. Since the ranges have been defined by a plurality of characteristics, however, such minute differences in color shading are accounted for by this embodiment of the present invention.
  • the characteristics of the color are defined by R, G, B, H, S, V, and R-G, and G-B, where R-G and G-B have red and green and green and blue components respectively.
  • any three of these variables is sufficient to uniquely define all colors displayed on an output device or stored in a frame buffer. To determine whether the variables are within the color range for the purpose of determining whether an event will be associated with the area of interest, the following algorithm is used (shown in FIG. 8 ).
  • step 800 it is determined whether an action has occurred. If no action occurred, step 800 repeats until an action occurs. If the action has occurred, it is determined where the action has occurred at step 810 . Then at step 820 , three variables are chosen from a list of possible color components, for instance, R, G, B, H, S, V, R-G, or G-B. For the chosen variables, histograms are produced at step 830.
  • a histogram is a graphical method for describing data. It is a presentation of quantitative data (i.e., color components) that have been summarized in a frequency distribution. The values of the quantitative variable are shown on the horizontal axis. A rectangle is drawn above each class such that the base of the rectangle is equal to the width of the class interval and its height is proportional to the number of data values in the class.
  • the color ranges are determined at step 840 , for instance by determining the variable peak on the histogram + or ⁇ three times its variance. After step 840 , it is determined whether any event is associated with the color range at step 850 . If no event is associated with that color range, no event is initiated and the process repeats at step 800 . If an event, is associated with that color range, then at step 860 an event associated with the color range is initiated.
  • a region of interest may be a plaid tablecloth. In this case, it is not adequate to use a single color value, nor a color range to define the region of interest. To account for this, one embodiment of the present invention uses color patterns to define the region of interest.
  • a color pattern comprises a representation of the colors in the region of interest, for instance a graphical representation, defined by a plot on an x,y axis.
  • the varying colors in the region of interest create peaks and valleys in the color pattern, graph representation.
  • multi-color patterns such as the plaid tablecloth, are used to associate actions with events.
  • Edge Detection uses an edge detection scheme to define the boundaries for the colors.
  • the color component of an object is described by its pixels.
  • An algorithm gives those pixels an “edge” which is not visible to the user, nor drawn on the screen, nor pre-processed prior to the user viewing the images formed by the pixels. For instance, an invisible wire frame is created to describe the shape of the given colored object to associate with the event. But the value of the pixels defines the wire frame rather than actually drawing a wire frame on the screen.
  • One embodiment of the present invention uses Quicktime, a software package available from Apple ComputersTM.
  • Quicktime is used to support the use of “sprites”.
  • a sprite is a clickable element within a video stream
  • Quicktime a sprite is modified to be defined by its color component.
  • one or more embodiments of the present invention create the associations between objects and events dynamically, in that the data which creates the association is a property of the image itself (i.e., its color values).
  • the present invention requires no alteration or pre-processing of the image to create the associations.
  • Embodiment of Computer Execution Environment may be implemented as computer software in the form of computer readable program code executed in a general purpose computing environment such as environment 600 illustrated in FIG. 6 , or in the form of byte code class files executable within a Java runtime EnvironmentTM running in such an environment, or in the form of byte codes running on a processor (or devices enabled to process byte codes) existing in a distributed environment (e.g., one or more processors on a network).
  • a keyboard 610 and mouse 611 are coupled to a system bus 618 . The keyboard and mouse are for introducing user input to the computer system and communicating that user input to central processing unit (CPU) 613 .
  • CPU central processing unit
  • I/O (input/output) unit 619 coupled to bidirectional system bus 618 represents such I/O elements as a printer, A/V (audio/video) I/O, etc.
  • Computer 601 may include a communication interface 620 coupled to bus 618 .
  • Communication interface 620 provides a two-way data communication coupling via a network link 621 to a local network 622 .
  • ISDN integrated services digital network
  • communication interface 620 provides a data communication connection to the corresponding type of telephone line, which comprises part of network link 621 .
  • LAN local area network
  • communication interface 620 provides a data communication connection via network link 621 to a compatible LAN.
  • Wireless links are also possible.
  • communication interface 620 sends and receives electrical, electromagnetic or optical signals which carry digital data streams representing various types of information.
  • Network link 621 typically provides data communication through one or more networks to other data devices.
  • network link 621 may provide a connection through local network to local server computer 623 or to data equipment operated by ISP 624 .
  • ISP 624 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the “Internet” 625 .
  • Internet 625 uses electrical, electromagnetic or optical signals which carry digital data streams.
  • the signals through the various networks and the signals on network link 621 and through communication interface 620 which carry the digital data to and from computer 600 , are exemplary forms of carrier waves transporting the information.
  • Processor 613 may reside wholly on client computer 601 or wholly on server 626 or processor 613 may have its computational power distributed between computer 601 and server 626 .
  • Server 626 symbolically is represented in FIG. 6 as one unit, but server 626 may also be distributed between multiple “tiers”.
  • server 626 comprises a middle and back tier where application logic executes in the middle tier and persistent data is obtained in the back tier.
  • processor 613 resides wholly on server 626
  • the results of the computations performed by processor 613 are transmitted to computer 601 via Internet 625 , Internet Service Provider (ISP) 624 , local network 622 and communication interface 620 .
  • ISP Internet Service Provider
  • computer 601 is able to display the results of the computation to a user in the form of output.
  • Computer 601 includes a video memory 614 , main memory 615 and mass storage 612 , all coupled to bidirectional system bus 618 along with keyboard 610 , mouse 611 and processor 613 .
  • main memory 615 and mass storage 612 may reside wholly on server 626 or computer 601 , or they may be distributed between the two. Examples of systems where processor 613 , main memory 615 , and mass storage 612 are distributed between computer 601 and server 626 include the thin-client computing architecture developed by Sun Microsystems, Inc., the palm pilot computing device and other personal digital assistants, Internet ready cellular phones and other Internet computing devices, and in platform independent computing environments, such as those which utilize the Java technologies also developed by Sun Microsystems, Inc.
  • the mass storage 612 may include both fixed and removable media, such as magnetic, optical or magnetic optical storage systems or any other available mass storage technology.
  • Bus 618 may contain, for example, thirty-two address lines for addressing video memory 614 or main memory 615 .
  • the system bus 618 also includes, for example, a 32-bit data bus for transferring data between and among the components; such as processor 613 , main memory 615 , video memory 614 and mass storage 612 .
  • multiplex data/address lines may be used instead of separate data and address lines.
  • the processor 613 is a microprocessor manufactured by Motorola, such as the 680 ⁇ 0 processor or a microprocessor manufactured by Intel, such as the 80 ⁇ 86, or Pentium processor, or a SPARC microprocessor from Sun Microsystems, Inc.
  • Main memory 615 is comprised of dynamic random access memory (DRAM).
  • Video memory 614 is a dual-ported video random access memory. One port of the video memory 614 is coupled to video amplifier 616 .
  • the video amplifier 616 is used to drive the cathode ray tube (CR 1) raster monitor 617 .
  • Video amplifier 616 is well known in the art and may be implemented by any suitable apparatus. This circuitry converts pixel data stored in video memory 614 to a raster signal suitable for use by monitor 617 .
  • Monitor 617 is a type of monitor suitable for displaying graphic images.
  • Computer 601 may send messages and receive data, including program code, through the network(s), network link 621 , and communication interface 620 .
  • remote server computer 626 might transmit a requested code for an application program through Internet 625 , ISP 624 , local network 622 and communication interface 620 .
  • the received code may be executed by processor 613 as it is received, and/or stored in mass storage 612 , or other non-volatile storage for later execution.
  • computer 600 may obtain application code in the form of a carrier wave.
  • remote server computer 626 may execute applications using processor 613 , and utilize mass storage 612 , and/or video main memory 615 .
  • the results of the execution at server 626 are then transmitted through Internet 625 , ISP 624 , local network 622 and communication interface 620 .
  • computer 601 performs only input and output functions.
  • Application code maybe embodied in any form of computer program product.
  • a computer program product comprises a medium configured to, store or transport computer readable code, or in which computer readable code may be embedded.
  • Some examples of computer program products are CD-ROM disks, ROM cards, floppy disks, magnetic tapes, computer hard drives, servers on a network, and carrier waves.

Abstract

Method and apparatus for associating the color of an object with an event preferably provides linked information for the object in any one of a plurality of succeeding video frames based only on the color value of the object. Preferably, at least one computer is coupled to a GUI and to a pointing structure, and is configured to: (i) identify coordinates of the object in a first video frame where the pointing structure has pointed; (ii) determine the color values of the identified object; (iii) store the color values of the identified object; (iv) detect an action of the object being selected in any one of the plurality of succeeding video frames; (v) determine the color values of the selected object; (vi) determine whether the selected object color values correspond to the stored color values; and (vii) automatically provide the linked information to the GUI only when the color values correspond.

Description

  • This is a continuation of U.S. patent application Ser. No. 09/679,391, filed Oct. 3, 2000 (abandoned), the entire contents of which are incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates generally to event processing, and in particular to associating the color of an object with an event. Portions of the disclosure of this patent document contain material that is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure as it appears in the Patent and Trademark Office file or records, but otherwise reserves all copyright rights whatsoever.
  • 2. Background Art
  • When using a computer or other interactive devices a user can interact with the device in various ways. One way is to use a mouse. The user utilizes the mouse to position a pointer somewhere on a screen. The user then depresses a mouse button and depending on where the pointer is, something pre-programmed will happen if that area of the screen is “hot”. In the context of the Internet, the user can position the mouse pointer over a hot spot “hyperlinked” to a web page and depress the mouse button. This will cause the computer, set-top box, or other device to send a signal to a server to retrieve that web page and display it on the computer screen.
  • We foresee the desirability to interact in this manner in many different contexts, not only linking with web pages. Current schemes are limited in the contexts that a user may interact. Before further describing the limitations of current schemes, a discussion of event processing is instructive.
  • Event Processing. “Event processing” refers to the association between some type of “action” and the occurrence of an “event”. Such a term is used herein in the context of a user's computing experience. The user performs an action followed by an event, followed by a series of other actions followed by other events.
  • An action occurs when a user positions a cursor in a specified location on a computer screen, for instance using a mouse, and depressing one of the mouse buttons. This example shows a common type of action, but many other actions exist as well within the context of computer event processing, including the input of text, a voice command, or any other type of action. The example action may take place in any computing environment, but one common environment that the action occurs is in a web browser. A web browser is a graphical user interface that enables users to utilize the Internet, for instance by viewing web pages. Normally the user will position the mouse pointer over a “hyperlink” and depress the mouse button there. This, in turn, triggers an event. The event in this example is to cause a server to retrieve another web page. Thus, in this example, the user interacts with the Internet by a series of actions followed by events.
  • Current event processing is disadvantageous, in part, because the types of actions available and the events initiated by the actions are limited. Before further discussing the drawbacks of current event processing, an overview of the Internet is provided below.
  • Internet. The information network known as the “Internet” has become increasingly popular. The Internet provides a body of information which may be accessed by users via computers for such purposes as business, education, and entertainment. In particular, the Internet includes a large collection of interconnected documents, images, sounds, videos, and other forms of media which are stored in computers in a system known as the World Wide Web or simply “the web”.
  • Internet resources are typically accessed in a two-way environment via a network connection. For example, connectivity to the network may be via a conventional twisted-pair telephone line which has a relatively low data-carrying capacity (e.g., bandwidth), or via a higher bandwidth path, which may comprise optical fiber, coaxial cable, ISDN, DSL, wireless connections, or other transmission mechanisms. To retrieve data from a network destination (e.g., an Internet site), a user sends a message to the network destination to request the transfer of information to the user's computer, referred to as a client. The network destination typically includes a computer known as a server. The server then sends a request signal to a source function, which may be a memory which is coupled to the server. The source function includes the user-requested information which may comprise, for example, text, graphics, audio and/or video data. In response to the request signal from the server, the source function provides the requested information to the client.
  • HTML. Source information which is stored in the source function is often stored in a format known as “Hypertext Markup Language (HTML)”. This file or script format allows the display of text, graphics and audio information, and provides links to other pages of information through “hyperlinks.” Hyperlinks are strings of characters in a particular format that specify the address of the desired page of information.
  • HTML is a system for marking documents to indicate how the document should be displayed, and how various documents should be linked together. HTML is a form of Standard Generalized Markup Language (SGML), defined by the International Standards Organization. HTML specifies the grammar and syntax of markup tags which are inserted into a data file to define how the data will be presented when read by a computer program known as a “web browser”. Conventional web browsers include Internet Explorer, Netscape Navigator, and others.
  • The data file, which is typically stored on a server, includes one or more web pages which are visited by users who have computers which may ran different browsers. When a page is visited, HTML data output from the server is downloaded to the client computer. The client computer's browser processes the data to format a layout for the page so the page may be viewed by the user on a computer screen. Generally, HTML tags provide text formatting, hypertext links to other pages, and links to sound and image elements. HTML tags also define input fields for interactive web pages.
  • An HTML application is made available to users on the web by storing the HTML file in a directory that is accessible to a server. Such a server is typically a web server which conforms to a web browser-supported protocol known as Hypertext Transfer Protocol (HTTP). Servers that conform to other protocols such as the File Transfer Protocol (FTP) or GOPHER may also be used, but do not support interactive HTML files.
  • HTTP defines a set of rules that servers and browsers follow when communicating with each other. Typically, the process begins when a user accesses an icon in an HTML page which is the anchor of a hyperlink, (for instance, by positioning a cursor on the icon and depressing a mouse button), or the user inputs a Uniform Resource Locator (URL) to his or her web browser, described below. A connection is then made to the server at the address and port number specified by the URL. Next, the browser sends a request to retrieve an object from the server, or to post data to an object on the server. The server sends a response to the browser including a status code and the response data. The connection between the browser and server is then closed. A URL is a unique address which identifies virtually all files and resources on the Internet. A URL has the form: method://server:port/path/file# anchor. The “method” of accessing the resource is the web browser-supported protocol, and may include, for example, HTTP, FTP, GOPHER, TELNET, NEWS, or MAILTO. The “server:port” indicates the name of the server which is providing the resource, and is alternatively known as the Internet domain name. For example, many businesses will use their business name as part of the server field. The port designation is the port number on the server, but is usually not used since a default port is assumed. The “path” indicates the directory path to the resource. The file indicates the file name of the resource. The “anchor” indicates the named element in the HTML document. Not all fields are required.
  • Consider the following URL, for example: http://www.business.com/information/profits.html# section1. The access method is HTTP, the server is www.business.com. there is no port specified, the path is information, the file is profits.html, and the anchor is section1. Thus, it may be seen that web browsers operate in a two-way communication environment to access information by sending a request signal defined by a URL command to a server, and receiving information in return.
  • Current Event Processing Schemes. In the context of the Internet, event processing has been limited by the nature of the web. Originally, the web was largely a textual body of interconnected information. Actions were limited to depressing a mouse button on a text link which initiated the event of retrieving another web page from a specified URL. These methods are disadvantageous because they are static. Such events have no application in an active environment, for instance where streaming video or television signals are involved, and such events have no application outside of a web browser (i.e., where a GUI is translating HTML commands into actions).
  • As the web evolved, other types of event processing have emerged. In particular, the web evolved to have a more active nature. First, the text-based web became merged with images. The images resided in various formats and could also be used to associate an action with an event. The images, however, suffered the same drawback as the text and essentially performed the same function. For instance, the images were used in the same event context (e.g., depressing a mouse button on the image) and the same actions resulted (e.g., the image caused a server to send another web page to the user).
  • After images came more active web environments. One example came in the form of animation where a sequential series of images are displayed in the web browser which cause the image to appear to move on the screen. To allow the user to act upon an animated image, one scheme takes each frame of the animated image and pre-determines where on that individual frame a potential action should be associated with an event.
  • For instance, imagine an animation of a runner holding a baton where the runner is moving horizontally across the screen and the desired action is to depress a mouse button when the pointer is over the baton. In this example, assume the animation consists of 100 images. Using one scheme, the 100 images are each individually stored in a computer and each image is analyzed to find where the baton was on the screen.
  • Along with each of the 100 images is an image map specifying that in frame 1 the boundary of the baton is in a first position and in frame 2, the boundary of the baton is in a second position. To map to these positions, the prior art essentially draws a wire frame around the baton in each frame. Therefore, 100 wire frames must be drawn in varying locations. The baton in a moving image varies in position and angle. This in turn varies the shape of the wire frame in each frame of the image.
  • The prior scheme is extremely disadvantageous because it requires a tremendous amount of manual pre-processing to draw 100 varied size wire frames. Such pre-processing contains a large lag on overhead. This scheme also must store each and every potential action in a computer and then wait for the action to occur. This technique uses a large amount of the computer's storage space and a large amount of time to create the wire frames. This scheme, in addition, is bound to a web browser implementation and may only trigger events relating to the accessing of a web page at a specified URL.
  • As the web has developed more recently, such schemes have become even less advantageous. Specifically, the web has expanded away from a static model and now incorporates other technologies, such as television and other signals. The web through this expansion has also moved to other media forms, such as “set-top boxes” and interactive television sets where the user interacts with the web using a conventional television set as an output device. With these other media forms and other environments, other types of GUIs (not only web browsers) and other types of actions (not only the retrieval of a web page from a specific URL) are desired. However, there is currently no efficient way to process events in such environments.
  • SUMMARY OF THE INVENTION
  • The present invention provides a method and apparatus for associating the color of an object with an event. In one or more embodiments, the invention takes a streaming or static digital image frames and associates an active, dynamic, or static object with an event by its color.
  • In one embodiment, an event is the retrieval of a document at a specified URL and the display of the document in a web browser. In another embodiment, an event includes the retrieval of textual or audio information associated with the object. In another embodiment, the event comprises switching to a separate video stream in an environment using streaming media, such as a digital video disc (DVD) player or recorder, or other suitable video environment.
  • The invention defines the color in a location (termed a region of interest) by its digital color mathematical properties. For instance, one embodiment uses the RGB standard which separates the red, green, and blue properties of the color as numerical values. Another embodiment also uses the HSV characteristics of the color, which define hue, saturation, and value as components of color. Another embodiment uses both RGB and HSV characteristics. In one embodiment, the color characteristics are obtained from a digital RGB frame buffer
  • As described above, the event links to another URL (i.e., to facilitate a transaction over the Internet) or in the case of a DVD or other video environment, the event causes the video stream to transition to a different data track (i.e., to pull up different information) or to a different video stream
  • In one embodiment the invention uses an edge detection scheme to help define the boundaries for the colors in a digital RGB frame. The algorithm developed gives those pixels an “edge” which is not visible to the user, nor drawn on the screen, nor pre-processed prior to the user viewing the images formed by the pixels. The values of the pixels defines an edge rather than the actual drawing of a wire frame on the screen.
  • Thus, one or more embodiments of the present invention create the associations between objects and events dynamically in that the data which creates the association is a property of the image itself (i.e., its color values). The present invention requires no alteration or pre-processing of the image to create the associations.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • These and other features, aspects and advantages of the present invention will become better understood with regard to the following description, appended claims and accompanying drawings where:
  • FIG. 1 is a flowchart showing how to associate the color of an object with an event according to an embodiment of the present invention.
  • FIG. 2 is a flowchart showing how to define the color characteristics of an object according to an embodiment of the present invention.
  • FIG. 3 is a flowchart showing how to define the color characteristics of an object according to another embodiment of the present invention.
  • FIG. 4 is a flowchart showing how to define the color characteristics of an object according to another embodiment of the present invention.
  • FIG. 5 is a flowchart showing how to associate the color of an object with an event using a table according to an embodiment of the present invention.
  • FIG. 6 is a computer execution environment where one or more embodiments of the present invention may be implemented.
  • FIG. 7 shows an embodiment of the present invention where the frame buffer is utilized.
  • FIG. 8 shows an embodiment of the present invention where color ranges are utilized.
  • DETAILED DESCRIPTION OF THE INVENTION
  • The invention is a method and apparatus for associating the color of an object with an event. In the following description, numerous specific details are set forth to provide a more thorough description of embodiments of the invention. It is apparent, however, to one skilled in the art, that the invention may be practiced without these specific details. In other instances, well known features have not been described in detail so as not to obscure the invention.
  • One or more embodiments of the present invention are directed to enhancing a user's ability to interact with a computer, set-top box, interactive television set, or other processing device. According to embodiments of the present invention, the user provides input on some area of a display device, for instance by depressing a mouse button at a specified location. The invention then determines where the user has depressed the mouse button and what color value appears on the display device at the time the user has depressed the mouse button. Then, the invention initiates an event that corresponds to that color value.
  • Event Processing. In one or more embodiments, the invention takes a streaming or static media and associates an active, dynamic, or static object with an event by its given color. The event may be, for instance, the retrieval of a document at a specified URL in a web browser or the retrieval of other information in non-web browser implementations. Other implementations include, for instance, the viewing of media, via a television, DVD or other media player, set-top devices, web-TV, interactive TV, and video-games.
  • Take, for example, a web page implementing streaming video. In the example, a game show host is wearing a green shirt. If a user provides input to the screen by depressing a mouse with the cursor in a location where the green shirt is located, the invention links to a URL associated with the green shirt. Thus, the moving color may substitute for a conventional, static, text link. In another embodiment, the game show host with the green shirt may be in a television show. In this embodiment, providing input on the moving green shirt may initiate a different event, such as providing information about the shirt or the game show host.
  • One embodiment of the present invention is shown in FIG. 1. At step 100, an action occurs. At step 110, the location of the action (a region of interest) is determined. Locating a region of interest is well known in the art.
  • One method to locate a region of interest is to obtain an ordered pair corresponding to the location of the cursor at the moment of the action. For instance, a computer screen may be represented as a two-dimensional graph wherein the axis of the graph resides in the lower left hand corner of the computer screen. Thus all ordered pairs will correspond to positive values of x,y pairs on the graph.
  • After defining the region of interest, the value associated with a color in the region of interest is determined at step 120. Thereafter, at step 130, an event associated with that color is initiated.
  • Color Characteristics. In one embodiment, the invention defines the color by its mathematical properties. For instance, one embodiment uses the RGB standard which separates the red, green, and blue properties of the color as numerical values. Another embodiment also uses the HSV characteristics of the color, which define hue, saturation, and value as components of color. HSV characteristics further define the properties of a color, for instance V represents an analog of the intensity of a color or how bright the color appears. Another embodiment of the present invention uses both RGB and HSV characteristics.
  • One embodiment defines color characteristics as shown in FIG. 2. At step 200, it is determined whether an action has occurred. If no action occurred, step 200 repeats until an action occurs. If the action has occurred, it is determined where the action has occurred at step 210.
  • The location is shown herein with respect to a single pixel, “pixel N’. At step 220, a red component of a color for pixel A is determined. Then, at step 230, a green component of the color for pixel A is determined. Next, at step 240, a blue component of the color for pixel A is determined. Thereafter, at step 250, the color characteristics are combined to form a color value for pixel A
  • After step 250, it is determined whether any event is associated with the color value at step 255. If no event is associated with that color value, no event is initiated and the process repeats at step 200. If an event is associated with the color value, then at step 260 an event associated with the color value is initiated.
  • Another embodiment defines color characteristics as shown in FIG. 3. At step 300, it is determined whether an action has occurred. If no action occurred, step 300 repeats until an action occurs. If the action has occurred, it is determined where the action has occurred at step 310. The location is shown herein with respect to a single pixel, “pixel B”. At step 320, a hue component of a color for pixel B is determined. Then, at step 330, a saturation component of the color for pixel B is determined. Next, at step 340, a value component of the color for pixel B is determined.
  • Thereafter, at step 350, the color characteristics are combined to form a color value for pixel B. After step 350, it is determined whether any event is associated with the color value at step 355. If no event is associated with that color value, no event is initiated and the process repeats at step 300. If an event is associated with the color value, then at step 360 an event associated with the color value is initiated.
  • Another embodiment defines color characteristics as shown in FIG. 4. At step 400, it is determined whether an action has occurred. If no action occurred, step 400 repeats until an action occurs. If the action has occurred, it is determined where the action has occurred at step 410. The location is shown herein with respect to a single pixel, “pixel C’. At step 420, a hue component of a color for pixel C is determined. Then, at step 430, a saturation component of the color for pixel C is determined. Next, at step 440, a value component of the color for pixel C is determined. At step 450, a red component of a color for pixel C is determined. Then, at step 460, a green component of the color for pixel C is determined. Next, at step 470, a blue component of the color for pixel C is determined.
  • Thereafter, at step 480, the color characteristics are combined to form a color value for pixel B. After step 480, it is determined whether any event is associated with the color value at step 485. If no event is associated with that color value, no event is initiated and the process repeats at step 400. If an event is associated with the color value, then at step 490 an event associated with the color value is initiated.
  • Frame Buffer. In a computer or other comparable device, images are often used as a form of output. A common way that such devices handle images or streams of images (e.g., video) is by using a “frame buffer”. For offline pre-processing or for real-time processing using high speed computers, the RGB frame buffers reside in memory. Alternatively, a special high speed video board can be used to process an RGB frame buffer for real-time processing.
  • A video signal, for instance, comprises a series of still images shown sequentially where the series of images appears so quickly that a user perceives the series of pictures to be a continuous stream of motion. As the computer processes each image, data relating to the image is stored in the frame buffer. Such data includes, for instance, the color values associated with each pixel at each location on the screen.
  • As a new image is displayed, new data about the image is stored in the frame buffer. Depending on the size of the frame buffer, many sequential images may be stored there before it fills up and old image data must be overwritten.
  • One embodiment of the present invention takes advantage of the data already existing in the frame buffer, for instance to perform step 120 of FIG. 1. FIG. 7 shows an embodiment of the present invention where the frame buffer is utilized.
  • At step 700, an action occurs. At step 710, the location of the action (a region of interest) is determined. Next, at step 720, the invention access the frame buffer to obtain color values at the location of the action when it occurred. Then, it is determined whether any event is associated with the color value at step 730. If no event is associated with that color value, no event is initiated and the process repeats at step 700. If an event is associated with the color value, then at step 740 an event associated with the color value is initiated.
  • Events. As described above, one event is a link to another URL (i.e., to facilitate a transaction over the Internet). Thus, one event may comprise the retrieval of a web page from a server computer connected to the Internet.
  • In the context of streaming media, an event may be the retrieval of a text page. For instance, assume the user is watching a DVD program showing several of the great pyramids of Egypt. If the user causes an event to occur in a region of interest defined by a specific pyramid, then the event may comprise the retrieval of a text page telling the user exactly which pyramid this is, when it was built, the history associated with this particular pyramid, and any other pertinent information. Thus, the event would cause the DVD to move to a different data track (i.e., to pull up different information).
  • In the same great pyramid example, another event may comprise the retrieval of audio information, for instance a narrator providing information similar to that given in the text track in the above example. Furthermore, the event may comprise the retrieval of a new media stream, such as the close up of the pyramid, alternate angles showing different views of the same pyramid, or views from inside the pyramid, for example.
  • In one embodiment an event comprises the retrieval of a data track. A data track comprises information associated with the color acted upon. Assume for instance, that this embodiment of the present invention takes place using an interactive television set. While the interactive television set is operating, a game show host appears wearing a green shirt. Then assume, for instance, that a user caused an action to occur at the location of the green shirt. In this instance, the data track may be information about the shirt, the game show host, or any other relevant information.
  • Whether an event is associated with a color value may be determined in a plurality of ways. One such way is to use a table. An example of one embodiment of a table is shown in Table 1.
  • TABLE 1
    Color Value Event
    100 retrieve document at URL www.info.comltest
    300 obtain data track at a specified location
    200 initiate media stream at specified location
    288 initiate audio track at specified location
  • The table may be arranged in a plurality of ways and may be accessed in methods well known to those skilled in the art, including hash tables, hash functions, pointers, indices, or any other suitable form of table lookup. In addition, different color values may initiate the same event. Alternatively, every color value on an output device may be associated with the same event.
  • FIG. 5 shows an embodiment of the present invention where a table is implemented. At step 500, an action occurs. At step 510, the location of the action (a region of interest) is determined. Then at step 520, the value associated with a color in the region of interest is determined. Next, a table is searched to locate the color value to obtain an event that is paired with the color value at step 530. Then, at step 535, it is determined whether the color value is in the table. If it is not, then that color value is not linked to an event and no event is initiated. In that case the process repeats at step 500. Otherwise, at step 540, the event is initiated.
  • Color Ranges. In another embodiment of the present invention regions of interest (i.e., those regions where a user has performed an action) are associated with events based on color ranges. A color range, for instance, includes those shades of blue that correspond to the body of a bluebird in a video stream. Realistic image and video streams will not have a uniform shade of blue at all areas of the bluebird's body since light and other factors will cause natural variations in the perceived color reflected from an object. In this embodiment, this reality is accounted for, which gives this embodiment specific application in dynamic and high-quality video streams.
  • Color ranges are defined by color characteristics. For instance, a color range may translate into values on an RGB or HSV scale, or could be combinations of both. Typically, in a computer system such values range from 0-255. Thus, the bluebird's body may be defined as a region of interest where the color range includes a B value on the RGB scale of 50-75, for instance.
  • Thus, whether the action occurred at an upper left or lower right portion of the bluebird's body, the same event is initiated, despite the potential that at these corners of the regions of interest, the color values differed. All that is required in this embodiment is that both differing color values fall within the defined color range.
  • In many images, the colors may be very similar, yet they may be aspects of different components of the image and associated with different events. In one embodiment, the ranges are defined with respect to a plurality of color characteristics including RGB and HSV.
  • For instance, a range may be defined where the R is selected to have a range of 50-80, the G is selected to have a range of 150-160, and the B is selected to have a range of 200-220. An example color is selected by the user where the example color has RGB characteristics of 60, 155, and 205 respectively. Another color on the screen that is associated with a different event may appear to be substantially identical to the example color but has RGB characteristics of 190, 100,33. Since the ranges have been defined by a plurality of characteristics, however, such minute differences in color shading are accounted for by this embodiment of the present invention.
  • In one embodiment, the characteristics of the color are defined by R, G, B, H, S, V, and R-G, and G-B, where R-G and G-B have red and green and green and blue components respectively. In this embodiment, any three of these variables is sufficient to uniquely define all colors displayed on an output device or stored in a frame buffer. To determine whether the variables are within the color range for the purpose of determining whether an event will be associated with the area of interest, the following algorithm is used (shown in FIG. 8).
  • At step 800, it is determined whether an action has occurred. If no action occurred, step 800 repeats until an action occurs. If the action has occurred, it is determined where the action has occurred at step 810. Then at step 820, three variables are chosen from a list of possible color components, for instance, R, G, B, H, S, V, R-G, or G-B. For the chosen variables, histograms are produced at step 830
  • A histogram is a graphical method for describing data. It is a presentation of quantitative data (i.e., color components) that have been summarized in a frequency distribution. The values of the quantitative variable are shown on the horizontal axis. A rectangle is drawn above each class such that the base of the rectangle is equal to the width of the class interval and its height is proportional to the number of data values in the class.
  • Using the histogram, the color ranges are determined at step 840, for instance by determining the variable peak on the histogram + or − three times its variance. After step 840, it is determined whether any event is associated with the color range at step 850. If no event is associated with that color range, no event is initiated and the process repeats at step 800. If an event, is associated with that color range, then at step 860 an event associated with the color range is initiated.
  • Color Patterns. It is often beneficial to define a regions of interest that is not uniform in color. A region of interest, for instance, may be a plaid tablecloth. In this case, it is not adequate to use a single color value, nor a color range to define the region of interest. To account for this, one embodiment of the present invention uses color patterns to define the region of interest.
  • A color pattern comprises a representation of the colors in the region of interest, for instance a graphical representation, defined by a plot on an x,y axis. The varying colors in the region of interest create peaks and valleys in the color pattern, graph representation. By locating patterns of peaks and valleys in such a color pattern, multi-color patterns, such as the plaid tablecloth, are used to associate actions with events.
  • Edge Detection. One embodiment of the present invention uses an edge detection scheme to define the boundaries for the colors. The color component of an object is described by its pixels. An algorithm gives those pixels an “edge” which is not visible to the user, nor drawn on the screen, nor pre-processed prior to the user viewing the images formed by the pixels. For instance, an invisible wire frame is created to describe the shape of the given colored object to associate with the event. But the value of the pixels defines the wire frame rather than actually drawing a wire frame on the screen.
  • One embodiment of the present invention uses Quicktime, a software package available from Apple Computers™. In this embodiment, Quicktime is used to support the use of “sprites”. A sprite is a clickable element within a video stream Using Quicktime, a sprite is modified to be defined by its color component. Thus, one or more embodiments of the present invention create the associations between objects and events dynamically, in that the data which creates the association is a property of the image itself (i.e., its color values). The present invention requires no alteration or pre-processing of the image to create the associations.
  • Embodiment of Computer Execution Environment (Hardware). An embodiment of the invention may be implemented as computer software in the form of computer readable program code executed in a general purpose computing environment such as environment 600 illustrated in FIG. 6, or in the form of byte code class files executable within a Java runtime Environment™ running in such an environment, or in the form of byte codes running on a processor (or devices enabled to process byte codes) existing in a distributed environment (e.g., one or more processors on a network). A keyboard 610 and mouse 611 are coupled to a system bus 618. The keyboard and mouse are for introducing user input to the computer system and communicating that user input to central processing unit (CPU) 613. Other suitable input devices may be used in addition to, or in place of, the mouse 611 and keyboard 610. I/O (input/output) unit 619 coupled to bidirectional system bus 618 represents such I/O elements as a printer, A/V (audio/video) I/O, etc.
  • Computer 601 may include a communication interface 620 coupled to bus 618. Communication interface 620 provides a two-way data communication coupling via a network link 621 to a local network 622. For example, if communication interface 620 is an integrated services digital network (ISDN) card or a modem, communication interface 620 provides a data communication connection to the corresponding type of telephone line, which comprises part of network link 621. If communication interface 620 is a local area network (LAN) card, communication interface 620 provides a data communication connection via network link 621 to a compatible LAN. Wireless links are also possible. In any such implementation, communication interface 620 sends and receives electrical, electromagnetic or optical signals which carry digital data streams representing various types of information.
  • Network link 621 typically provides data communication through one or more networks to other data devices. For example, network link 621 may provide a connection through local network to local server computer 623 or to data equipment operated by ISP 624. ISP 624 in turn provides data communication services through the world wide packet data communication network now commonly referred to as the “Internet” 625. Local network 622 and Internet 625 both use electrical, electromagnetic or optical signals which carry digital data streams. The signals through the various networks and the signals on network link 621 and through communication interface 620, which carry the digital data to and from computer 600, are exemplary forms of carrier waves transporting the information.
  • Processor 613 may reside wholly on client computer 601 or wholly on server 626 or processor 613 may have its computational power distributed between computer 601 and server 626. Server 626 symbolically is represented in FIG. 6 as one unit, but server 626 may also be distributed between multiple “tiers”. In one embodiment, server 626 comprises a middle and back tier where application logic executes in the middle tier and persistent data is obtained in the back tier. In the case where processor 613 resides wholly on server 626, the results of the computations performed by processor 613 are transmitted to computer 601 via Internet 625, Internet Service Provider (ISP) 624, local network 622 and communication interface 620. In this way, computer 601 is able to display the results of the computation to a user in the form of output.
  • Computer 601 includes a video memory 614, main memory 615 and mass storage 612, all coupled to bidirectional system bus 618 along with keyboard 610, mouse 611 and processor 613. As with processor 613, in various computing environments, main memory 615 and mass storage 612, may reside wholly on server 626 or computer 601, or they may be distributed between the two. Examples of systems where processor 613, main memory 615, and mass storage 612 are distributed between computer 601 and server 626 include the thin-client computing architecture developed by Sun Microsystems, Inc., the palm pilot computing device and other personal digital assistants, Internet ready cellular phones and other Internet computing devices, and in platform independent computing environments, such as those which utilize the Java technologies also developed by Sun Microsystems, Inc.
  • The mass storage 612 may include both fixed and removable media, such as magnetic, optical or magnetic optical storage systems or any other available mass storage technology. Bus 618 may contain, for example, thirty-two address lines for addressing video memory 614 or main memory 615. The system bus 618 also includes, for example, a 32-bit data bus for transferring data between and among the components; such as processor 613, main memory 615, video memory 614 and mass storage 612. Alternatively, multiplex data/address lines may be used instead of separate data and address lines.
  • In one. embodiment of the invention, the processor 613 is a microprocessor manufactured by Motorola, such as the 680×0 processor or a microprocessor manufactured by Intel, such as the 80×86, or Pentium processor, or a SPARC microprocessor from Sun Microsystems, Inc. However, any other suitable microprocessor or microcomputer may be utilized. Main memory 615 is comprised of dynamic random access memory (DRAM). Video memory 614 is a dual-ported video random access memory. One port of the video memory 614 is coupled to video amplifier 616. The video amplifier 616 is used to drive the cathode ray tube (CR 1) raster monitor 617. Video amplifier 616 is well known in the art and may be implemented by any suitable apparatus. This circuitry converts pixel data stored in video memory 614 to a raster signal suitable for use by monitor 617. Monitor 617 is a type of monitor suitable for displaying graphic images.
  • Computer 601 may send messages and receive data, including program code, through the network(s), network link 621, and communication interface 620. In the Internet example, remote server computer 626 might transmit a requested code for an application program through Internet 625, ISP 624, local network 622 and communication interface 620. The received code may be executed by processor 613 as it is received, and/or stored in mass storage 612, or other non-volatile storage for later execution. In this manner, computer 600 may obtain application code in the form of a carrier wave. Alternatively, remote server computer 626 may execute applications using processor 613, and utilize mass storage 612, and/or video main memory 615. The results of the execution at server 626 are then transmitted through Internet 625, ISP 624, local network 622 and communication interface 620. In this example, computer 601 performs only input and output functions.
  • Application code maybe embodied in any form of computer program product. A computer program product comprises a medium configured to, store or transport computer readable code, or in which computer readable code may be embedded. Some examples of computer program products are CD-ROM disks, ROM cards, floppy disks, magnetic tapes, computer hard drives, servers on a network, and carrier waves.
  • The computer systems described above are for purposes of example only. An embodiment of the invention may be implemented in any type of computer system or programming or processing environment.
  • Thus, a method and apparatus for associating the color of an object with an event is described in conjunction with one or more specific embodiments. The invention is defined by the claims and their full scope of equivalents.

Claims (8)

1. A method for automatically identifying an object in a plurality of video frames and associating the object with a link action, comprising:
determining a location in one of said video frames where an action by a pointing device on the object has occurred;
determining a color value of the determined location on the object; and
automatically associating a link action with said color value of said determined location on the object in said one video frame, and automatically associating said link action with said color value on the object in succeeding video frames.
2. A method for automatically identifying an object in a plurality of video frames and associating the object with a URL, comprising:
determining the x-y coordinates of the object in a video frame that is selected by a pointing device;
determining a color value range for the determined object; and
automatically associating a URL with said color value range of said determined object in said one video frame, and automatically associating the URL with said color value range of said determined object in succeeding video frames.
3. A method for automatically identifying an object in a plurality of video frames and associating the object with an link decision, comprising:
determining the coordinates of an object in one of said video frames selected by a pointing device;
determining a color value pattern for the selected object; and
automatically associating the link decision with said color value pattern of the selected object in said one video frame, and automatically associating the link decision with said color value pattern in succeeding video frames.
4. A method for providing linked information for an object in any one of a plurality of succeeding video frames based only on the color value of the object, comprising:
using a pointing device to identify coordinates of the object in a first video frame;
determining the color values of the identified object in the first video frame;
storing the color values of the identified object in the first video frame;
detecting an action of the object being selected in any one of the plurality of video frames succeeding the first video frame;
determining the color values of the selected object in the any one of the plurality of video frames succeeding the first video frame;
determining whether the determined color values correspond to the stored color values; and
automatically providing the linked information only when the determined color values correspond to the stored color values.
5. Apparatus for automatically identifying an object in a plurality of video frames and associating the object with a link action, comprising:
a graphical user interface (GUI);
user pointing structure; and
at least one computer coupled to the graphical user interface and to said user pointing structure, said at least one computer being configured to:
(i) determine a location in one of said video frames displayed on said GUI where said user pointing structure has identified the object;
(ii) determine a color value of the object in said one video frame displayed on said GUI;
(iii) automatically associate a link action with said color value of said determined location on the object in said one video frame; and
(iv) automatically associate said link action with said color value on the object in succeeding video frames.
6. Apparatus for automatically identifying an object in a plurality of video frames and associating the object with a URL, comprising:
a graphical user interface (GUI);
user pointing structure; and
at least one computer coupled to the graphical user interface and to said user pointing structure, said at least one computer being configured to:
(i) determine the x-y coordinates of the object in a video frame displayed on said GUI that is selected by said user pointing structure;
(ii) determine a color value range for the determined object;
(iii) automatically associate a URL with said color value range of said determined object in said one video frame; and
(iv) automatically associate the URL with said color value range of said determined object in succeeding video frames.
7. Apparatus for automatically identifying an object in a plurality of video frames and associating the object with an link decision, comprising:
a graphical user interface (GUI);
user pointing structure; and
at least one computer coupled to the graphical user interface and to said user pointing structure, said at least one computer being configured to:
(i) determine the coordinates of an object in one of said video frames displayed on said GUI selected by said user pointing structure;
(ii) determining a color value pattern for the selected object;
(iii) automatically associate the link decision with said color value pattern of the selected object in said one video frame; and
(iv) automatically associate the link decision with said color value pattern in succeeding video frames.
8. Apparatus for providing linked information for an object in any one of a plurality of succeeding video frames based only on the color value of the object, comprising:
a graphical user interface (GUI);
user pointing structure; and
at least one computer coupled to the graphical user interface and to said user pointing structure, said at least one computer being configured to:
(i) identify coordinates of the object in a first video frame displayed on said GUI where said pointing structure has pointed;
(ii) determine the color values of the identified object in the first video frame displayed on said GUI;
(iii) store the color values of the identified object in the first video frame displayed on said GUI;
(iv) detect an action of the object being selected by said pointing structure in any one of the plurality of video frames displayed on said GUI succeeding the first video frame;
(v) determine the color values of the selected object in the any one of the plurality of video frames succeeding the first video frame;
(vi) determine whether the determined color values correspond to the stored color values; and
(vii) automatically provide the linked information to said GUI only when the determined color values correspond to the stored color values.
US12/472,761 2000-10-03 2009-05-27 Method and apparatus for associating the color of an object with an event Abandoned US20090235151A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US12/472,761 US20090235151A1 (en) 2000-10-03 2009-05-27 Method and apparatus for associating the color of an object with an event
US16/168,093 US20190080193A1 (en) 2000-10-03 2018-10-23 Method and apparatus for associating the color of an object with an event
US16/454,743 US20190318187A1 (en) 2000-10-03 2019-06-27 Method and apparatus for associating the color of an object with an event

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US67939100A 2000-10-03 2000-10-03
US12/472,761 US20090235151A1 (en) 2000-10-03 2009-05-27 Method and apparatus for associating the color of an object with an event

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US67939100A Continuation 2000-10-03 2000-10-03

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/168,093 Continuation US20190080193A1 (en) 2000-10-03 2018-10-23 Method and apparatus for associating the color of an object with an event

Publications (1)

Publication Number Publication Date
US20090235151A1 true US20090235151A1 (en) 2009-09-17

Family

ID=24726731

Family Applications (6)

Application Number Title Priority Date Filing Date
US09/944,543 Expired - Lifetime US6774908B2 (en) 2000-10-03 2001-08-31 System and method for tracking an object in a video and linking information thereto
US10/219,175 Expired - Fee Related US7773093B2 (en) 2000-10-03 2002-08-15 Method and apparatus for associating the color of an object with an event
US10/786,777 Expired - Fee Related US7804506B2 (en) 2000-10-03 2004-02-25 System and method for tracking an object in a video and linking information thereto
US12/472,761 Abandoned US20090235151A1 (en) 2000-10-03 2009-05-27 Method and apparatus for associating the color of an object with an event
US16/168,093 Abandoned US20190080193A1 (en) 2000-10-03 2018-10-23 Method and apparatus for associating the color of an object with an event
US16/454,743 Abandoned US20190318187A1 (en) 2000-10-03 2019-06-27 Method and apparatus for associating the color of an object with an event

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US09/944,543 Expired - Lifetime US6774908B2 (en) 2000-10-03 2001-08-31 System and method for tracking an object in a video and linking information thereto
US10/219,175 Expired - Fee Related US7773093B2 (en) 2000-10-03 2002-08-15 Method and apparatus for associating the color of an object with an event
US10/786,777 Expired - Fee Related US7804506B2 (en) 2000-10-03 2004-02-25 System and method for tracking an object in a video and linking information thereto

Family Applications After (2)

Application Number Title Priority Date Filing Date
US16/168,093 Abandoned US20190080193A1 (en) 2000-10-03 2018-10-23 Method and apparatus for associating the color of an object with an event
US16/454,743 Abandoned US20190318187A1 (en) 2000-10-03 2019-06-27 Method and apparatus for associating the color of an object with an event

Country Status (4)

Country Link
US (6) US6774908B2 (en)
AT (1) ATE509311T1 (en)
ES (1) ES2370893T3 (en)
HK (1) HK1059483A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100275224A1 (en) * 2009-04-24 2010-10-28 Samuel Sheng System and Method for Information Delivery Including Delivery Via Video-Based Keyed or Tagged Content
US20130279744A1 (en) * 2012-04-23 2013-10-24 Apple Inc. Systems and methods for controlling output of content based on human recognition data detection
US20140240260A1 (en) * 2013-02-25 2014-08-28 Samsung Electronics Co., Ltd. Method and apparatus for providing user interface
WO2015197651A1 (en) * 2014-06-25 2015-12-30 Thomson Licensing Annotation method and corresponding device, computer program product and storage medium

Families Citing this family (155)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8028314B1 (en) 2000-05-26 2011-09-27 Sharp Laboratories Of America, Inc. Audiovisual information management system
US7076445B1 (en) 2000-06-20 2006-07-11 Cartwright Shawn D System and methods for obtaining advantages and transacting the same in a computer gaming environment
US6636237B1 (en) * 2000-07-31 2003-10-21 James H. Murray Method for creating and synchronizing links to objects in a video
US8020183B2 (en) 2000-09-14 2011-09-13 Sharp Laboratories Of America, Inc. Audiovisual management system
US6774908B2 (en) 2000-10-03 2004-08-10 Creative Frontier Inc. System and method for tracking an object in a video and linking information thereto
US8234218B2 (en) * 2000-10-10 2012-07-31 AddnClick, Inc Method of inserting/overlaying markers, data packets and objects relative to viewable content and enabling live social networking, N-dimensional virtual environments and/or other value derivable from the content
US8316450B2 (en) * 2000-10-10 2012-11-20 Addn Click, Inc. System for inserting/overlaying markers, data packets and objects relative to viewable content and enabling live social networking, N-dimensional virtual environments and/or other value derivable from the content
US20030038796A1 (en) 2001-02-15 2003-02-27 Van Beek Petrus J.L. Segmentation metadata for audio-visual content
US6952224B2 (en) * 2001-03-30 2005-10-04 Intel Corporation Object trackability via parametric camera tuning
US7904814B2 (en) 2001-04-19 2011-03-08 Sharp Laboratories Of America, Inc. System for presenting audio-video content
US7499077B2 (en) * 2001-06-04 2009-03-03 Sharp Laboratories Of America, Inc. Summarization of football video content
US7203620B2 (en) * 2001-07-03 2007-04-10 Sharp Laboratories Of America, Inc. Summarization of video content
US7432940B2 (en) * 2001-10-12 2008-10-07 Canon Kabushiki Kaisha Interactive animation of sprites in a video production
US7474698B2 (en) * 2001-10-19 2009-01-06 Sharp Laboratories Of America, Inc. Identification of replay segments
US20030079224A1 (en) * 2001-10-22 2003-04-24 Anton Komar System and method to provide additional information associated with selectable display areas
US20030098869A1 (en) * 2001-11-09 2003-05-29 Arnold Glenn Christopher Real time interactive video system
US7120873B2 (en) 2002-01-28 2006-10-10 Sharp Laboratories Of America, Inc. Summarization of sumo video content
US8214741B2 (en) 2002-03-19 2012-07-03 Sharp Laboratories Of America, Inc. Synchronization of video and data
EP1527534A1 (en) * 2002-07-01 2005-05-04 Nokia Corporation A system and method for delivering representative media objects of a broadcast media stream to a terminal
JP2004054435A (en) * 2002-07-17 2004-02-19 Toshiba Corp Hypermedia information presentation method, hypermedia information presentation program and hypermedia information presentation device
US7042471B2 (en) * 2002-07-22 2006-05-09 International Business Machines Corporation Method and system for displaying descriptive information associated with a defined video object
US7657836B2 (en) 2002-07-25 2010-02-02 Sharp Laboratories Of America, Inc. Summarization of soccer video content
US7657907B2 (en) 2002-09-30 2010-02-02 Sharp Laboratories Of America, Inc. Automatic user profiling
US7643550B2 (en) * 2002-10-09 2010-01-05 Hewlett-Packard Development Company, L.P. Method for presenting streaming media for an event
KR100513294B1 (en) * 2003-04-09 2005-09-09 삼성전자주식회사 Method, apparatus and system for providing information of an object included in multimedia content
US20040233233A1 (en) * 2003-05-21 2004-11-25 Salkind Carole T. System and method for embedding interactive items in video and playing same in an interactive environment
CN100429641C (en) * 2003-08-20 2008-10-29 英业达股份有限公司 Interactive video signal generating system and method
JP2005167517A (en) * 2003-12-01 2005-06-23 Olympus Corp Image processor, calibration method thereof, and image processing program
US7742077B2 (en) * 2004-02-19 2010-06-22 Robert Bosch Gmbh Image stabilization system and method for a video camera
US8356317B2 (en) 2004-03-04 2013-01-15 Sharp Laboratories Of America, Inc. Presence based technology
US8949899B2 (en) 2005-03-04 2015-02-03 Sharp Laboratories Of America, Inc. Collaborative recommendation system
US7594245B2 (en) 2004-03-04 2009-09-22 Sharp Laboratories Of America, Inc. Networked video devices
US7590310B2 (en) 2004-05-05 2009-09-15 Facet Technology Corp. Methods and apparatus for automated true object-based image analysis and retrieval
US7890604B2 (en) * 2004-05-07 2011-02-15 Microsoft Corproation Client-side callbacks to server events
US9026578B2 (en) 2004-05-14 2015-05-05 Microsoft Corporation Systems and methods for persisting data between web pages
EP1792314A1 (en) * 2004-08-23 2007-06-06 Sherpa Technologies, LLC Selective displaying of item information in videos
US7594177B2 (en) * 2004-12-08 2009-09-22 Microsoft Corporation System and method for video browsing using a cluster index
US7970771B2 (en) * 2004-12-20 2011-06-28 Microsoft Corporation Method and system for tracking objects associated with an activity
GB0502369D0 (en) * 2005-02-04 2005-03-16 British Telecomm Classifying an object in a video frame
US20060218618A1 (en) * 2005-03-22 2006-09-28 Lorkovic Joseph E Dual display interactive video
WO2006121986A2 (en) 2005-05-06 2006-11-16 Facet Technology Corp. Network-based navigation system having virtual drive-thru advertisements integrated with actual imagery from along a physical route
WO2006127951A2 (en) * 2005-05-23 2006-11-30 Gilley Thomas S Distributed scalable media environment
US8145528B2 (en) 2005-05-23 2012-03-27 Open Text S.A. Movie advertising placement optimization based on behavior and content analysis
US9648281B2 (en) 2005-05-23 2017-05-09 Open Text Sa Ulc System and method for movie segment bookmarking and sharing
US8141111B2 (en) 2005-05-23 2012-03-20 Open Text S.A. Movie advertising playback techniques
EP1758398A1 (en) * 2005-08-23 2007-02-28 Syneola SA Multilevel semiotic and fuzzy logic user and metadata interface means for interactive multimedia system having cognitive adaptive capability
US8150743B2 (en) * 2005-09-20 2012-04-03 Gadge Brown Method and apparatus for the identification of products displayed in media programs
US8689253B2 (en) 2006-03-03 2014-04-01 Sharp Laboratories Of America, Inc. Method and system for configuring media-playing sets
KR20080100434A (en) 2006-03-09 2008-11-18 톰슨 라이센싱 Content access tree
WO2008024190A2 (en) * 2006-08-03 2008-02-28 Nexplore Corporation System and method to provide a search and advertisement dragging system
JP4939890B2 (en) * 2006-10-02 2012-05-30 株式会社東芝 Video encoding apparatus, video decoding apparatus, and video decoding method
US20080088639A1 (en) * 2006-10-13 2008-04-17 Sony Ericsson Mobile Communications Ab Method for generating a graphical user interface
US8139115B2 (en) * 2006-10-30 2012-03-20 International Business Machines Corporation Method and apparatus for managing parking lots
US20080140523A1 (en) * 2006-12-06 2008-06-12 Sherpa Techologies, Llc Association of media interaction with complementary data
US7559017B2 (en) 2006-12-22 2009-07-07 Google Inc. Annotation framework for video
US9071730B2 (en) 2007-04-14 2015-06-30 Viap Limited Product information display and purchasing
AU2008260048B2 (en) * 2007-05-30 2012-09-13 Creatier Interactive, Llc Method and system for enabling advertising and transaction within user generated video content
US9047374B2 (en) * 2007-06-08 2015-06-02 Apple Inc. Assembling video content
EP2174226A1 (en) * 2007-06-26 2010-04-14 Ooyala, Inc. Object tracking and content monetization
US20090009532A1 (en) * 2007-07-02 2009-01-08 Sharp Laboratories Of America, Inc. Video content identification using ocr
US8091103B2 (en) * 2007-07-22 2012-01-03 Overlay.Tv Inc. Server providing content directories of video signals and linkage to content information sources
US20090078774A1 (en) * 2007-09-21 2009-03-26 Symbol Technologies, Inc. Graphical User Interface for Use in Programming a Barcode Reader
US8312486B1 (en) 2008-01-30 2012-11-13 Cinsay, Inc. Interactive product placement system and method therefor
US11227315B2 (en) 2008-01-30 2022-01-18 Aibuy, Inc. Interactive product placement system and method therefor
US20110191809A1 (en) 2008-01-30 2011-08-04 Cinsay, Llc Viral Syndicated Interactive Product System and Method Therefor
US8112702B2 (en) 2008-02-19 2012-02-07 Google Inc. Annotating video intervals
WO2009137368A2 (en) 2008-05-03 2009-11-12 Mobile Media Now, Inc. Method and system for generation and playback of supplemented videos
US8331685B2 (en) 2008-05-28 2012-12-11 Apple Inc. Defining a border for an image
US8452105B2 (en) * 2008-05-28 2013-05-28 Apple Inc. Selecting a section of interest within an image
US8548251B2 (en) 2008-05-28 2013-10-01 Apple Inc. Defining a border for an image
US8280171B2 (en) * 2008-05-28 2012-10-02 Apple Inc. Tools for selecting a section of interest within an image
US8566353B2 (en) * 2008-06-03 2013-10-22 Google Inc. Web-based system for collaborative generation of interactive videos
US8949718B2 (en) 2008-09-05 2015-02-03 Lemi Technology, Llc Visual audio links for digital audio content
US8386314B2 (en) 2008-12-11 2013-02-26 Accenture Global Services Limited Online ad detection and ad campaign analysis
US9665965B2 (en) * 2009-01-14 2017-05-30 Innovid Inc. Video-associated objects
US9183881B2 (en) 2009-02-02 2015-11-10 Porto Technology, Llc System and method for semantic trick play
US8200602B2 (en) * 2009-02-02 2012-06-12 Napo Enterprises, Llc System and method for creating thematic listening experiences in a networked peer media recommendation environment
US9947017B2 (en) 2009-03-03 2018-04-17 Accenture Global Services Limited Online content campaign classification
US9940631B2 (en) * 2009-03-03 2018-04-10 Accenture Global Services Limited Online content collection
US8826117B1 (en) 2009-03-25 2014-09-02 Google Inc. Web-based system for video editing
US8132200B1 (en) 2009-03-30 2012-03-06 Google Inc. Intra-video ratings
US8885977B2 (en) * 2009-04-30 2014-11-11 Apple Inc. Automatically extending a boundary for an image to fully divide the image
US9094714B2 (en) 2009-05-29 2015-07-28 Cognitive Networks, Inc. Systems and methods for on-screen graphics detection
US10116972B2 (en) 2009-05-29 2018-10-30 Inscape Data, Inc. Methods for identifying video segments and displaying option to view from an alternative source and/or on an alternative device
US10949458B2 (en) 2009-05-29 2021-03-16 Inscape Data, Inc. System and method for improving work load management in ACR television monitoring system
US10375451B2 (en) 2009-05-29 2019-08-06 Inscape Data, Inc. Detection of common media segments
US8595781B2 (en) 2009-05-29 2013-11-26 Cognitive Media Networks, Inc. Methods for identifying video segments and displaying contextual targeted content on a connected television
US9449090B2 (en) 2009-05-29 2016-09-20 Vizio Inscape Technologies, Llc Systems and methods for addressing a media database using distance associative hashing
US20120047119A1 (en) * 2009-07-21 2012-02-23 Porto Technology, Llc System and method for creating and navigating annotated hyperlinks between video segments
US20110261258A1 (en) * 2009-09-14 2011-10-27 Kumar Ramachandran Systems and methods for updating video content with linked tagging information
CN102033893A (en) * 2009-09-25 2011-04-27 鸿富锦精密工业(深圳)有限公司 Picture inquiry system and method
US9111287B2 (en) * 2009-09-30 2015-08-18 Microsoft Technology Licensing, Llc Video content-aware advertisement placement
US10192138B2 (en) 2010-05-27 2019-01-29 Inscape Data, Inc. Systems and methods for reducing data density in large datasets
US9838753B2 (en) 2013-12-23 2017-12-05 Inscape Data, Inc. Monitoring individual viewing of television events using tracking pixels and cookies
US9906838B2 (en) 2010-07-12 2018-02-27 Time Warner Cable Enterprises Llc Apparatus and methods for content delivery and message exchange across multiple content delivery networks
US20120017150A1 (en) * 2010-07-15 2012-01-19 MySongToYou, Inc. Creating and disseminating of user generated media over a network
US8819557B2 (en) 2010-07-15 2014-08-26 Apple Inc. Media-editing application with a free-form space for organizing or compositing media clips
WO2012012214A1 (en) * 2010-07-23 2012-01-26 The Directv Group, Inc. Method for creating interactive applications for television
US9792363B2 (en) * 2011-02-01 2017-10-17 Vdopia, INC. Video display method
US8842911B2 (en) 2011-02-04 2014-09-23 Apple Inc. Luma-based color matching
US8611655B2 (en) 2011-02-04 2013-12-17 Apple Inc. Hue-based color matching
US8594426B2 (en) 2011-02-04 2013-11-26 Apple Inc. Color matching using color segmentation
US8891864B2 (en) 2011-02-16 2014-11-18 Apple Inc. User-aided image segmentation
US8823726B2 (en) 2011-02-16 2014-09-02 Apple Inc. Color balance
US8854370B2 (en) 2011-02-16 2014-10-07 Apple Inc. Color waveform
US20120304062A1 (en) * 2011-05-23 2012-11-29 Speakertext, Inc. Referencing content via text captions
US8929588B2 (en) * 2011-07-22 2015-01-06 Honeywell International Inc. Object tracking
RU2604670C2 (en) 2011-08-29 2016-12-10 Синсэй, Инк. Containerized software for virally copying from one endpoint to another
US9183546B2 (en) * 2011-10-07 2015-11-10 Hewlett-Packard Development Company, L.P. Methods and systems for a reminder servicer using visual recognition
ITTO20110946A1 (en) * 2011-10-19 2013-04-20 Emisfera Societa Cooperativa SYSTEM TO ALLOW A USER TO INTERACT IN REAL TIME WITH A VIDEO CONTENT
EP2769424A1 (en) * 2011-10-19 2014-08-27 E. I. Du Pont de Nemours and Company Organic electronic device for lighting
EP2782330A4 (en) * 2011-11-17 2015-03-11 Panasonic Ip Corp America Image processing device, imaging device, and image processing method
US9438883B2 (en) * 2012-04-09 2016-09-06 Intel Corporation Quality of experience reporting for combined unicast-multicast/broadcast streaming of media content
US10789631B2 (en) 2012-06-21 2020-09-29 Aibuy, Inc. Apparatus and method for peer-assisted e-commerce shopping
US9607330B2 (en) 2012-06-21 2017-03-28 Cinsay, Inc. Peer-assisted shopping
WO2014014963A1 (en) 2012-07-16 2014-01-23 Questionmine, LLC Apparatus and method for synchronizing interactive content with multimedia
US9186793B1 (en) 2012-08-31 2015-11-17 Brain Corporation Apparatus and methods for controlling attention of a robot
WO2014144882A1 (en) * 2013-03-15 2014-09-18 E-Connect Visual analysis of transactions
US9521438B2 (en) * 2013-03-29 2016-12-13 Microsoft Technology Licensing, Llc Custom data indicating nominal range of samples of media content
US9953347B2 (en) 2013-09-11 2018-04-24 Cinsay, Inc. Dynamic binding of live video content
JP6531105B2 (en) 2013-09-27 2019-06-12 アイバイ,インコーポレイテッド N-level duplication of supplemental content
KR102344237B1 (en) 2013-09-27 2021-12-27 에이아이바이, 인크. Apparatus and method for supporting relationships associated with content provisioning
TW201519161A (en) * 2013-11-07 2015-05-16 Inst Information Industry Electronic device and video object motion trajectory modification method thereof
TWI570666B (en) * 2013-11-15 2017-02-11 財團法人資訊工業策進會 Electronic device and video object tracking method thereof
US9955192B2 (en) 2013-12-23 2018-04-24 Inscape Data, Inc. Monitoring individual viewing of television events using tracking pixels and cookies
BR112016014598B1 (en) * 2013-12-23 2022-08-23 Inscape Data, Inc METHOD AND SYSTEM FOR TRACKING PIXELS AND COOKIES FOR VIEWING TELEVISION EVENTS
WO2015118563A1 (en) * 2014-02-06 2015-08-13 Gurbrinder Singh Bindra A method and system for providing information on one or more frames selected from a video by a user
US20150244772A1 (en) * 2014-02-24 2015-08-27 Triple It B.V. Fast rendering of content using a mobile user device
US9533413B2 (en) 2014-03-13 2017-01-03 Brain Corporation Trainable modular robotic apparatus and methods
US9987743B2 (en) 2014-03-13 2018-06-05 Brain Corporation Trainable modular robotic apparatus and methods
BR102014007666B1 (en) * 2014-03-28 2023-01-10 Samsung Eletrônica Da Amazônia Ltda METHOD FOR AUTHENTICING MOBILE TRANSACTIONS USING VIDEO ENCRYPTION AND METHOD FOR VIDEO ENCRYPTION
US20150339589A1 (en) * 2014-05-21 2015-11-26 Brain Corporation Apparatus and methods for training robots utilizing gaze-based saliency maps
US9462239B2 (en) * 2014-07-15 2016-10-04 Fuji Xerox Co., Ltd. Systems and methods for time-multiplexing temporal pixel-location data and regular image projection for interactive projection
US9852759B2 (en) * 2014-10-25 2017-12-26 Yieldmo, Inc. Methods for serving interactive content to a user
US11809811B2 (en) * 2014-10-25 2023-11-07 Yieldmo, Inc. Methods for serving interactive content to a user
EP3029631A1 (en) * 2014-12-04 2016-06-08 Thomson Licensing A method and apparatus for assisted object selection in video sequences
JP5818050B1 (en) * 2015-01-28 2015-11-18 ビックリック株式会社 Status judgment system
MX2017009738A (en) 2015-01-30 2017-11-20 Inscape Data Inc Methods for identifying video segments and displaying option to view from an alternative source and/or on an alternative device.
CN107949849B (en) 2015-04-17 2021-10-08 构造数据有限责任公司 System and method for reducing data density in large data sets
US11432046B1 (en) * 2015-06-12 2022-08-30 Veepio Holdings, Llc Interactive, personalized objects in content creator's media with e-commerce link associated therewith
US9840003B2 (en) 2015-06-24 2017-12-12 Brain Corporation Apparatus and methods for safe navigation of robotic devices
AU2016293601B2 (en) 2015-07-16 2020-04-09 Inscape Data, Inc. Detection of common media segments
CA2992529C (en) 2015-07-16 2022-02-15 Inscape Data, Inc. Prediction of future views of video segments to optimize system resource utilization
US10080062B2 (en) 2015-07-16 2018-09-18 Inscape Data, Inc. Optimizing media fingerprint retention to improve system resource utilization
CA2992519A1 (en) 2015-07-16 2017-01-19 Inscape Data, Inc. Systems and methods for partitioning search indexes for improved efficiency in identifying media segments
US10445885B1 (en) 2015-10-01 2019-10-15 Intellivision Technologies Corp Methods and systems for tracking objects in videos and images using a cost matrix
CN108124167A (en) * 2016-11-30 2018-06-05 阿里巴巴集团控股有限公司 A kind of play handling method, device and equipment
KR20190134664A (en) 2017-04-06 2019-12-04 인스케이프 데이터, 인코포레이티드 System and method for using media viewing data to improve device map accuracy
US11006188B2 (en) 2017-12-29 2021-05-11 Comcast Cable Communications, Llc Secondary media insertion systems, methods, and apparatuses
US10999640B2 (en) 2018-11-29 2021-05-04 International Business Machines Corporation Automatic embedding of information associated with video content
US11141656B1 (en) * 2019-03-29 2021-10-12 Amazon Technologies, Inc. Interface with video playback
US11554324B2 (en) * 2020-06-25 2023-01-17 Sony Interactive Entertainment LLC Selection of video template based on computer simulation metadata
US11263385B1 (en) * 2021-03-24 2022-03-01 Ebay Inc. Web browser extension for linking images to webpages
CN114943936B (en) * 2022-06-17 2023-06-20 北京百度网讯科技有限公司 Target behavior recognition method and device, electronic equipment and storage medium
US11805588B1 (en) 2022-07-29 2023-10-31 Electronic Theatre Controls, Inc. Collision detection for venue lighting

Citations (99)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4719584A (en) * 1985-04-01 1988-01-12 Hughes Aircraft Company Dual mode video tracker
US4758714A (en) * 1986-10-06 1988-07-19 Carlson Steven R Point-of-sale mechanism
US4797920A (en) * 1987-05-01 1989-01-10 Mastercard International, Inc. Electronic funds transfer system with means for verifying a personal identification number without pre-established secret keys
US4969189A (en) * 1988-06-25 1990-11-06 Nippon Telegraph & Telephone Corporation Authentication system and apparatus therefor
US5053607A (en) * 1986-10-06 1991-10-01 Carlson Steven R Point-of-sale device particularly adapted for processing checks
US5125022A (en) * 1990-05-15 1992-06-23 Vcs Industries, Inc. Method for recognizing alphanumeric strings spoken over a telephone network
US5297194A (en) * 1990-05-15 1994-03-22 Vcs Industries, Inc. Simultaneous speaker-independent voice recognition and verification over a telephone network
US5343529A (en) * 1993-09-28 1994-08-30 Milton Goldfine Transaction authentication using a centrally generated transaction identifier
US5430809A (en) * 1992-07-10 1995-07-04 Sony Corporation Human face tracking system
US5463728A (en) * 1993-03-10 1995-10-31 At&T Corp. Electronic circuits for the graphical display of overlapping windows with transparency
USH1506H (en) * 1991-12-11 1995-12-05 Xerox Corporation Graphical user interface for editing a palette of colors
US5504318A (en) * 1991-09-13 1996-04-02 Symbol Technologies, Inc. Analog waveform decoder using peak locations
US5517605A (en) * 1993-08-11 1996-05-14 Ast Research Inc. Method and apparatus for managing browsing, and selecting graphic images
US5596654A (en) * 1987-04-20 1997-01-21 Fuji Photo Film Co., Ltd. Method of determining desired image signal range based on histogram data
US5659742A (en) * 1995-09-15 1997-08-19 Infonautics Corporation Method for storing multi-media information in an information retrieval system
US5729741A (en) * 1995-04-10 1998-03-17 Golden Enterprises, Inc. System for storage and retrieval of diverse types of information obtained from different media sources which includes video, audio, and text transcriptions
US5751286A (en) * 1992-11-09 1998-05-12 International Business Machines Corporation Image query system and method
US5767894A (en) * 1995-01-26 1998-06-16 Spectradyne, Inc. Video distribution system
US5794249A (en) * 1995-12-21 1998-08-11 Hewlett-Packard Company Audio/video retrieval system that uses keyword indexing of digital recordings to display a list of the recorded text files, keywords and time stamps associated with the system
US5819286A (en) * 1995-12-11 1998-10-06 Industrial Technology Research Institute Video database indexing and query method and system
US5893110A (en) * 1996-08-16 1999-04-06 Silicon Graphics, Inc. Browser driven user interface to a media asset database
US5929849A (en) * 1996-05-02 1999-07-27 Phoenix Technologies, Ltd. Integration of dynamic universal resource locators with television presentations
US5987454A (en) * 1997-06-09 1999-11-16 Hobbs; Allen Method and apparatus for selectively augmenting retrieved text, numbers, maps, charts, still pictures and/or graphics, moving pictures and/or graphics and audio information from a network resource
US6006241A (en) * 1997-03-14 1999-12-21 Microsoft Corporation Production of a video stream with synchronized annotations over a computer network
US6008820A (en) * 1995-08-04 1999-12-28 Microsoft Corporation Processor for controlling the display of rendered image layers and method for controlling same
US6008816A (en) * 1996-04-25 1999-12-28 Microsoft Corporation Method and system for managing color specification using attachable palettes and palettes that refer to other palettes
US6016504A (en) * 1996-08-28 2000-01-18 Infospace.Com, Inc. Method and system for tracking the purchase of a product and services over the Internet
US6016349A (en) * 1995-12-20 2000-01-18 Musa; Lorenzo Cellular phone provided with legal identification means of the owner of the cellular phone
US6018748A (en) * 1996-05-28 2000-01-25 Sun Microsystems, Inc. Dynamic linkable labels in a network browser page
US6026411A (en) * 1997-11-06 2000-02-15 International Business Machines Corporation Method, apparatus, and computer program product for generating an image index and for internet searching and querying by image colors
US6032181A (en) * 1998-02-25 2000-02-29 U.S. Philips Corporation Interactive distribution system for digital audio-video signals
US6031541A (en) * 1996-08-05 2000-02-29 International Business Machines Corporation Method and apparatus for viewing panoramic three dimensional scenes
US6067401A (en) * 1993-01-11 2000-05-23 Abecassis; Max Playing a version of and from within a video by means of downloaded segment information
US6069630A (en) * 1997-08-22 2000-05-30 International Business Machines Corporation Data processing system and method for creating a link map
US6070176A (en) * 1997-01-30 2000-05-30 Intel Corporation Method and apparatus for graphically representing portions of the world wide web
US6070161A (en) * 1997-03-19 2000-05-30 Minolta Co., Ltd. Method of attaching keyword or object-to-key relevance ratio and automatic attaching device therefor
US6076104A (en) * 1997-09-04 2000-06-13 Netscape Communications Corp. Video data integration system using image data and associated hypertext links
US6154879A (en) * 1994-11-28 2000-11-28 Smarttouch, Inc. Tokenless biometric ATM access system
US6154771A (en) * 1998-06-01 2000-11-28 Mediastra, Inc. Real-time receipt, decompression and play of compressed streaming video/hypervideo; with thumbnail display of past scenes and with replay, hyperlinking and/or recording permissively intiated retrospectively
US6161108A (en) * 1997-04-28 2000-12-12 Justsystem Corp. Method and apparatus for managing images, a method and apparatus for retrieving images, and a computer-readable recording medium with a program for making a computer execute the methods stored therein
US6164528A (en) * 1996-12-31 2000-12-26 Chequemark Patent, Inc. Check writing point of sale system
US6169573B1 (en) * 1997-07-03 2001-01-02 Hotv, Inc. Hypervideo system and method with object tracking in a compressed digital video environment
US6184878B1 (en) * 1997-12-23 2001-02-06 Sarnoff Corporation Interactive world wide web access using a set top terminal in a video on demand system
US6188777B1 (en) * 1997-08-01 2001-02-13 Interval Research Corporation Method and apparatus for personnel detection and tracking
US6198833B1 (en) * 1998-09-16 2001-03-06 Hotv, Inc. Enhanced interactive video with object tracking and hyperlinking
US6202151B1 (en) * 1997-05-09 2001-03-13 Gte Service Corporation System and method for authenticating electronic transactions using biometric certificates
US6205231B1 (en) * 1995-05-10 2001-03-20 Identive Corporation Object identification in a moving video image
US6204843B1 (en) * 1991-11-25 2001-03-20 Actv, Inc. Compressed digital-data interactive program system
US6209025B1 (en) * 1998-11-09 2001-03-27 John C Bellamy Integrated video system
US6222937B1 (en) * 1996-02-16 2001-04-24 Microsoft Corporation Method and system for tracking vantage points from which pictures of an object have been taken
US6240555B1 (en) * 1996-03-29 2001-05-29 Microsoft Corporation Interactive entertainment system for presenting supplemental interactive content together with continuous video programs
US6253238B1 (en) * 1998-12-02 2001-06-26 Ictv, Inc. Interactive cable television system with frame grabber
US6256785B1 (en) * 1996-12-23 2001-07-03 Corporate Media Patners Method and system for providing interactive look-and-feel in a digital broadcast via an X-Y protocol
US6269348B1 (en) * 1994-11-28 2001-07-31 Veristar Corporation Tokenless biometric electronic debit and credit transactions
US6272250B1 (en) * 1999-01-20 2001-08-07 University Of Washington Color clustering for scene change detection and object tracking in video sequences
US6275989B1 (en) * 1996-07-01 2001-08-14 Opentv, Inc. Interactive television system and method for displaying web-like stills with hyperlinks
US6337969B1 (en) * 1999-09-22 2002-01-08 Toshiba Tec Kabushiki Kaisha Fixing device
US6337916B1 (en) * 1998-04-03 2002-01-08 Discreet Logic Inc. Modifying image data
US6343141B1 (en) * 1996-10-08 2002-01-29 Lucent Technologies Inc. Skin area detection for video image systems
US6357042B2 (en) * 1998-09-16 2002-03-12 Anand Srinivasan Method and apparatus for multiplexing separately-authored metadata for insertion into a video data stream
US20020056136A1 (en) * 1995-09-29 2002-05-09 Wistendahl Douglass A. System for converting existing TV content to interactive TV programs operated with a standard remote control and TV set-top box
US6397181B1 (en) * 1999-01-27 2002-05-28 Kent Ridge Digital Labs Method and apparatus for voice annotation and retrieval of multimedia data
US20020069218A1 (en) * 2000-07-24 2002-06-06 Sanghoon Sull System and method for indexing, searching, identifying, and editing portions of electronic multimedia files
US6411725B1 (en) * 1995-07-27 2002-06-25 Digimarc Corporation Watermark enabled video objects
US20020087530A1 (en) * 2000-12-29 2002-07-04 Expresto Software Corp. System and method for publishing, updating, navigating, and searching documents containing digital video data
US6424370B1 (en) * 1999-10-08 2002-07-23 Texas Instruments Incorporated Motion based event detection system and method
US6457018B1 (en) * 1996-04-30 2002-09-24 International Business Machines Corporation Object oriented information retrieval framework mechanism
US6478223B1 (en) * 2000-01-12 2002-11-12 Intermec Ip Corporation Machine-readable color symbology and method and apparatus for reading same with standard readers such as laser scanners
US6493707B1 (en) * 1999-10-29 2002-12-10 Verizon Laboratories Inc. Hypervideo: information retrieval using realtime buffers
US6496981B1 (en) * 1997-09-19 2002-12-17 Douglass A. Wistendahl System for converting media content for interactive TV use
US6526169B1 (en) * 1999-03-15 2003-02-25 Grass Valley (Us), Inc. Histogram-based segmentation of objects from a video signal via color moments
US6536043B1 (en) * 1996-02-14 2003-03-18 Roxio, Inc. Method and systems for scalable representation of multimedia data for progressive asynchronous transmission
US6549643B1 (en) * 1999-11-30 2003-04-15 Siemens Corporate Research, Inc. System and method for selecting key-frames of video data
US20030122860A1 (en) * 2001-12-05 2003-07-03 Yuji Ino Video data searching method and video data searching system as well as video data editing method and video data editing system
US6603921B1 (en) * 1998-07-01 2003-08-05 International Business Machines Corporation Audio/video archive system and method for automatic indexing and searching
US6625315B2 (en) * 1998-10-23 2003-09-23 Facet Technology Corp. Method and apparatus for identifying objects depicted in a videostream
US6637032B1 (en) * 1997-01-06 2003-10-21 Microsoft Corporation System and method for synchronizing enhancing content with a video program using closed captioning
US6639998B1 (en) * 1999-01-11 2003-10-28 Lg Electronics Inc. Method of detecting a specific object in an image signal
US6642940B1 (en) * 2000-03-03 2003-11-04 Massachusetts Institute Of Technology Management of properties for hyperlinked video
US20030226150A1 (en) * 2000-01-27 2003-12-04 Berberet Suzanne M. System and method for providing broadcast programming, a virtual vcr, and a video scrapbook to programming subscribers
US6697796B2 (en) * 2000-01-13 2004-02-24 Agere Systems Inc. Voice clip search
US6724933B1 (en) * 2000-07-28 2004-04-20 Microsoft Corporation Media segmentation system and related methods
US6724915B1 (en) * 1998-03-13 2004-04-20 Siemens Corporate Research, Inc. Method for tracking a video object in a time-ordered sequence of image frames
US6741655B1 (en) * 1997-05-05 2004-05-25 The Trustees Of Columbia University In The City Of New York Algorithms and system for object-oriented content-based video search
US6744908B2 (en) * 2000-02-29 2004-06-01 Kabushiki Kaisha Toshiba Traffic density analysis apparatus based on encoded video
US20040215660A1 (en) * 2003-02-06 2004-10-28 Canon Kabushiki Kaisha Image search method and apparatus
US6819797B1 (en) * 1999-01-29 2004-11-16 International Business Machines Corporation Method and apparatus for classifying and querying temporal and spatial information in video
US20040227768A1 (en) * 2000-10-03 2004-11-18 Creative Frontier, Inc. System and method for tracking an object in a video and linking information thereto
US20040233233A1 (en) * 2003-05-21 2004-11-25 Salkind Carole T. System and method for embedding interactive items in video and playing same in an interactive environment
US6859799B1 (en) * 1998-11-30 2005-02-22 Gemstar Development Corporation Search engine for video and graphics
US20050086703A1 (en) * 1999-07-08 2005-04-21 Microsoft Corporation Skimming continuous multimedia content
US6925474B2 (en) * 2000-12-07 2005-08-02 Sony United Kingdom Limited Video information retrieval
US6990448B2 (en) * 1999-03-05 2006-01-24 Canon Kabushiki Kaisha Database annotation and retrieval including phoneme data
US7003156B1 (en) * 1999-01-29 2006-02-21 Kabushiki Kaisha Toshiba Object detection method and a video data retrieval method
US7020192B1 (en) * 1998-07-31 2006-03-28 Kabushiki Kaisha Toshiba Method of retrieving video picture and apparatus therefor
US7024020B2 (en) * 2001-01-20 2006-04-04 Samsung Electronics Co., Ltd. Apparatus and method for generating object-labeled image in video sequence
US7032182B2 (en) * 2000-12-20 2006-04-18 Eastman Kodak Company Graphical user interface adapted to allow scene content annotation of groups of pictures in a picture database to promote efficient database browsing
US7054812B2 (en) * 2000-05-16 2006-05-30 Canon Kabushiki Kaisha Database annotation and retrieval
US7158676B1 (en) * 1999-02-01 2007-01-02 Emuse Media Limited Interactive system

Family Cites Families (70)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US659643A (en) * 1900-07-11 1900-10-16 Vincenzo Bonzagni Apparatus for cleansing water-pipes.
US3878557A (en) 1974-03-15 1975-04-15 Int Video Corp Color framing videotape recording apparatus and method
US5204749A (en) 1984-05-25 1993-04-20 Canon Kabushiki Kaisha Automatic follow-up focus detecting device and automatic follow-up device
US4924303A (en) 1988-09-06 1990-05-08 Kenneth Dunlop Method and apparatus for providing interactive retrieval of TV still frame images and audio segments
US5885086A (en) 1990-09-12 1999-03-23 The United States Of America As Represented By The Secretary Of The Navy Interactive video delivery system
DE4309802A1 (en) * 1993-03-28 1994-09-29 Robert Prof Dr Ing Massen Color control close to production with imaging sensors
TW318315B (en) 1993-05-03 1997-10-21 At & T Corp
US5590261A (en) 1993-05-07 1996-12-31 Massachusetts Institute Of Technology Finite-element method for image alignment and morphing
US5735744A (en) 1993-05-10 1998-04-07 Yugengaisha Adachi International Interactive communication system for communicating video game and karaoke software
US5841978A (en) 1993-11-18 1998-11-24 Digimarc Corporation Network linking method using steganographically embedded data objects
US5475756A (en) 1994-02-17 1995-12-12 At&T Corp. Method of authenticating a terminal in a transaction execution system
US5907597A (en) 1994-08-05 1999-05-25 Smart Tone Authentication, Inc. Method and system for the secure communication of data
DE69522924T2 (en) 1994-10-11 2002-04-11 Koninkl Philips Electronics Nv METHOD AND ARRANGEMENT FOR TRANSMITTING AN INTERACTIVE AUDIOVISUAL PROGRAM
US5774868A (en) * 1994-12-23 1998-06-30 International Business And Machines Corporation Automatic sales promotion selection system and method
US5677955A (en) 1995-04-07 1997-10-14 Financial Services Technology Consortium Electronic funds transfer instruments
US5752160A (en) 1995-05-05 1998-05-12 Dunn; Matthew W. Interactive entertainment network system and method with analog video startup loop for video-on-demand
US5907323A (en) 1995-05-05 1999-05-25 Microsoft Corporation Interactive program summary panel
US5684715A (en) 1995-06-07 1997-11-04 Canon Information Systems, Inc. Interactive video system with dynamic video object descriptors
US5912980A (en) 1995-07-13 1999-06-15 Hunke; H. Martin Target acquisition and tracking
US5737553A (en) 1995-07-14 1998-04-07 Novell, Inc. Colormap system for mapping pixel position and color index to executable functions
AU723079B2 (en) * 1995-07-19 2000-08-17 Veracel Inc. Window texture extraction for image analysis
US5745103A (en) 1995-08-02 1998-04-28 Microsoft Corporation Real-time palette negotiations in multimedia presentations
US5874985A (en) 1995-08-31 1999-02-23 Microsoft Corporation Message delivery method for interactive televideo system
US6005602A (en) 1995-08-31 1999-12-21 Microsoft Corporation Message delivery method for interactive televideo system
US5781228A (en) 1995-09-07 1998-07-14 Microsoft Corporation Method and system for displaying an interactive program with intervening informational segments
US5708845A (en) * 1995-09-29 1998-01-13 Wistendahl; Douglass A. System for mapping hot spots in media content for interactive digital media program
US5724901A (en) 1995-11-02 1998-03-10 Gaz Metropolitan And Company Limited Oxygen-enriched gas burner for incinerating waste materials
US5793414A (en) 1995-11-15 1998-08-11 Eastman Kodak Company Interactive video communication system
US5822530A (en) 1995-12-14 1998-10-13 Time Warner Entertainment Co. L.P. Method and apparatus for processing requests for video on demand versions of interactive applications
US5867584A (en) 1996-02-22 1999-02-02 Nec Corporation Video object tracking method for interactive multimedia applications
US6006256A (en) 1996-03-11 1999-12-21 Opentv, Inc. System and method for inserting interactive program content within a television signal originating at a remote network
US5778187A (en) 1996-05-09 1998-07-07 Netcast Communications Corp. Multicasting method and apparatus
US5886705A (en) * 1996-05-17 1999-03-23 Seiko Epson Corporation Texture memory organization based on data locality
US5872857A (en) 1996-05-22 1999-02-16 Raytheon Company Generalized biased centroid edge locator
US5900905A (en) 1996-06-05 1999-05-04 Microsoft Corporation System and method for linking video, services and applications in an interactive television system
US5774191A (en) * 1996-06-26 1998-06-30 Intel Corporation Chroma-key color range determination
US5929850A (en) 1996-07-01 1999-07-27 Thomson Consumer Electronices, Inc. Interactive television system and method having on-demand web-like navigational capabilities for displaying requested hyperlinked web-like still images associated with television content
US5903830A (en) 1996-08-08 1999-05-11 Joao; Raymond Anthony Transaction security apparatus and method
US5983244A (en) 1996-09-27 1999-11-09 International Business Machines Corporation Indicating when clickable image link on a hypertext image map of a computer web browser has been traversed
US5774666A (en) 1996-10-18 1998-06-30 Silicon Graphics, Inc. System and method for displaying uniform network resource locators embedded in time-based medium
JP3658893B2 (en) * 1996-11-15 2005-06-08 カシオ計算機株式会社 Delivery package management device
US5931908A (en) 1996-12-23 1999-08-03 The Walt Disney Corporation Visual object present within live programming as an actionable event for user selection of alternate programming wherein the actionable event is selected by human operator at a head end for distributed data and programming
US5973692A (en) 1997-03-10 1999-10-26 Knowlton; Kenneth Charles System for the capture and indexing of graphical representations of files, information sources and the like
KR100321839B1 (en) 1997-04-02 2002-03-08 포만 제프리 엘 Method and apparatus for integrating hyperlinks in video
US5818440A (en) 1997-04-15 1998-10-06 Time Warner Entertainment Co. L.P. Automatic execution of application on interactive television
US5937404A (en) 1997-04-23 1999-08-10 Appaloosa Interactive Corporation Apparatus for bleaching a de-activated link in a web page of any distinguishing color or feature representing an active link
US6268852B1 (en) 1997-06-02 2001-07-31 Microsoft Corporation System and method for facilitating generation and editing of event handlers
US5933005A (en) 1997-07-29 1999-08-03 Brunswick Corporation Throttle position monitor with one stationary sensor and one movable sensor
GB2327837B (en) 1997-07-29 1999-09-15 Microsoft Corp Providing enhanced content with broadcast video
JPH1169330A (en) 1997-08-19 1999-03-09 Fujitsu Ltd Image communication equipment provided with automatic answering function
JPH11103441A (en) * 1997-09-26 1999-04-13 Matsushita Electric Ind Co Ltd Method for displaying clip and its device
US5867208A (en) 1997-10-28 1999-02-02 Sun Microsystems, Inc. Encoding system and method for scrolling encoded MPEG stills in an interactive television application
US6002168A (en) 1997-11-25 1999-12-14 Tessera, Inc. Microelectronic component with rigid interposer
JPH11184782A (en) * 1997-12-19 1999-07-09 Matsushita Electric Ind Co Ltd Device for receiving and displaying file and method therefor and recording medium for recording the same method
US6222541B1 (en) * 1998-01-20 2001-04-24 International Business Machines Corporation Method and apparatus for fast-path location and selection of links
US6154528A (en) * 1999-02-11 2000-11-28 Ameritech Corporation System and method for storing and transferring information tokens in a low network communication
US7181691B2 (en) 1999-09-16 2007-02-20 Sharp Laboratories Of America, Inc. Audiovisual information management system with presentation service
US6757866B1 (en) 1999-10-29 2004-06-29 Verizon Laboratories Inc. Hyper video: information retrieval using text from multimedia
US6937744B1 (en) * 2000-06-13 2005-08-30 Microsoft Corporation System and process for bootstrap initialization of nonparametric color models
US6822663B2 (en) * 2000-09-12 2004-11-23 Adaptview, Inc. Transform rule generator for web-based markup languages
US20030098869A1 (en) 2001-11-09 2003-05-29 Arnold Glenn Christopher Real time interactive video system
US6536046B1 (en) * 2002-05-09 2003-03-25 Akadema, Inc. Baseball glove
US7446803B2 (en) 2003-12-15 2008-11-04 Honeywell International Inc. Synchronous video and data annotations
US20050044105A1 (en) 2003-08-19 2005-02-24 Kelly Terrell System and method for delivery of content-specific video clips
US7191164B2 (en) 2003-08-19 2007-03-13 Intel Corporation Searching for object images with reduced computation
US20050223031A1 (en) 2004-03-30 2005-10-06 Andrew Zisserman Method and apparatus for retrieving visual object categories from a database containing images
JP4304108B2 (en) 2004-03-31 2009-07-29 株式会社東芝 METADATA DISTRIBUTION DEVICE, VIDEO REPRODUCTION DEVICE, AND VIDEO REPRODUCTION SYSTEM
US8132204B2 (en) 2004-04-07 2012-03-06 Visible World, Inc. System and method for enhanced video selection and categorization using metadata
US7590310B2 (en) 2004-05-05 2009-09-15 Facet Technology Corp. Methods and apparatus for automated true object-based image analysis and retrieval
US20060082662A1 (en) 2004-10-15 2006-04-20 Brian Isaacson System and process for digitizing and tracking audio, video and text information

Patent Citations (101)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4719584A (en) * 1985-04-01 1988-01-12 Hughes Aircraft Company Dual mode video tracker
US4758714A (en) * 1986-10-06 1988-07-19 Carlson Steven R Point-of-sale mechanism
US5053607A (en) * 1986-10-06 1991-10-01 Carlson Steven R Point-of-sale device particularly adapted for processing checks
US5596654A (en) * 1987-04-20 1997-01-21 Fuji Photo Film Co., Ltd. Method of determining desired image signal range based on histogram data
US4797920A (en) * 1987-05-01 1989-01-10 Mastercard International, Inc. Electronic funds transfer system with means for verifying a personal identification number without pre-established secret keys
US4969189A (en) * 1988-06-25 1990-11-06 Nippon Telegraph & Telephone Corporation Authentication system and apparatus therefor
US5297194A (en) * 1990-05-15 1994-03-22 Vcs Industries, Inc. Simultaneous speaker-independent voice recognition and verification over a telephone network
US5125022A (en) * 1990-05-15 1992-06-23 Vcs Industries, Inc. Method for recognizing alphanumeric strings spoken over a telephone network
US5504318A (en) * 1991-09-13 1996-04-02 Symbol Technologies, Inc. Analog waveform decoder using peak locations
US6204843B1 (en) * 1991-11-25 2001-03-20 Actv, Inc. Compressed digital-data interactive program system
US6252586B1 (en) * 1991-11-25 2001-06-26 Actv, Inc. Compressed digital-data interactive program system
USH1506H (en) * 1991-12-11 1995-12-05 Xerox Corporation Graphical user interface for editing a palette of colors
US5430809A (en) * 1992-07-10 1995-07-04 Sony Corporation Human face tracking system
US5751286A (en) * 1992-11-09 1998-05-12 International Business Machines Corporation Image query system and method
US6067401A (en) * 1993-01-11 2000-05-23 Abecassis; Max Playing a version of and from within a video by means of downloaded segment information
US5463728A (en) * 1993-03-10 1995-10-31 At&T Corp. Electronic circuits for the graphical display of overlapping windows with transparency
US5517605A (en) * 1993-08-11 1996-05-14 Ast Research Inc. Method and apparatus for managing browsing, and selecting graphic images
US5343529A (en) * 1993-09-28 1994-08-30 Milton Goldfine Transaction authentication using a centrally generated transaction identifier
US6154879A (en) * 1994-11-28 2000-11-28 Smarttouch, Inc. Tokenless biometric ATM access system
US6269348B1 (en) * 1994-11-28 2001-07-31 Veristar Corporation Tokenless biometric electronic debit and credit transactions
US5767894A (en) * 1995-01-26 1998-06-16 Spectradyne, Inc. Video distribution system
US5729741A (en) * 1995-04-10 1998-03-17 Golden Enterprises, Inc. System for storage and retrieval of diverse types of information obtained from different media sources which includes video, audio, and text transcriptions
US6205231B1 (en) * 1995-05-10 2001-03-20 Identive Corporation Object identification in a moving video image
US6411725B1 (en) * 1995-07-27 2002-06-25 Digimarc Corporation Watermark enabled video objects
US6008820A (en) * 1995-08-04 1999-12-28 Microsoft Corporation Processor for controlling the display of rendered image layers and method for controlling same
US5659742A (en) * 1995-09-15 1997-08-19 Infonautics Corporation Method for storing multi-media information in an information retrieval system
US20020056136A1 (en) * 1995-09-29 2002-05-09 Wistendahl Douglass A. System for converting existing TV content to interactive TV programs operated with a standard remote control and TV set-top box
US5819286A (en) * 1995-12-11 1998-10-06 Industrial Technology Research Institute Video database indexing and query method and system
US6016349A (en) * 1995-12-20 2000-01-18 Musa; Lorenzo Cellular phone provided with legal identification means of the owner of the cellular phone
US5794249A (en) * 1995-12-21 1998-08-11 Hewlett-Packard Company Audio/video retrieval system that uses keyword indexing of digital recordings to display a list of the recorded text files, keywords and time stamps associated with the system
US6536043B1 (en) * 1996-02-14 2003-03-18 Roxio, Inc. Method and systems for scalable representation of multimedia data for progressive asynchronous transmission
US6222937B1 (en) * 1996-02-16 2001-04-24 Microsoft Corporation Method and system for tracking vantage points from which pictures of an object have been taken
US6240555B1 (en) * 1996-03-29 2001-05-29 Microsoft Corporation Interactive entertainment system for presenting supplemental interactive content together with continuous video programs
US6008816A (en) * 1996-04-25 1999-12-28 Microsoft Corporation Method and system for managing color specification using attachable palettes and palettes that refer to other palettes
US6457018B1 (en) * 1996-04-30 2002-09-24 International Business Machines Corporation Object oriented information retrieval framework mechanism
US5929849A (en) * 1996-05-02 1999-07-27 Phoenix Technologies, Ltd. Integration of dynamic universal resource locators with television presentations
US6018748A (en) * 1996-05-28 2000-01-25 Sun Microsystems, Inc. Dynamic linkable labels in a network browser page
US6275989B1 (en) * 1996-07-01 2001-08-14 Opentv, Inc. Interactive television system and method for displaying web-like stills with hyperlinks
US6031541A (en) * 1996-08-05 2000-02-29 International Business Machines Corporation Method and apparatus for viewing panoramic three dimensional scenes
US5893110A (en) * 1996-08-16 1999-04-06 Silicon Graphics, Inc. Browser driven user interface to a media asset database
US6016504A (en) * 1996-08-28 2000-01-18 Infospace.Com, Inc. Method and system for tracking the purchase of a product and services over the Internet
US6343141B1 (en) * 1996-10-08 2002-01-29 Lucent Technologies Inc. Skin area detection for video image systems
US6256785B1 (en) * 1996-12-23 2001-07-03 Corporate Media Patners Method and system for providing interactive look-and-feel in a digital broadcast via an X-Y protocol
US6164528A (en) * 1996-12-31 2000-12-26 Chequemark Patent, Inc. Check writing point of sale system
US6637032B1 (en) * 1997-01-06 2003-10-21 Microsoft Corporation System and method for synchronizing enhancing content with a video program using closed captioning
US6070176A (en) * 1997-01-30 2000-05-30 Intel Corporation Method and apparatus for graphically representing portions of the world wide web
US6006241A (en) * 1997-03-14 1999-12-21 Microsoft Corporation Production of a video stream with synchronized annotations over a computer network
US6070161A (en) * 1997-03-19 2000-05-30 Minolta Co., Ltd. Method of attaching keyword or object-to-key relevance ratio and automatic attaching device therefor
US6161108A (en) * 1997-04-28 2000-12-12 Justsystem Corp. Method and apparatus for managing images, a method and apparatus for retrieving images, and a computer-readable recording medium with a program for making a computer execute the methods stored therein
US6741655B1 (en) * 1997-05-05 2004-05-25 The Trustees Of Columbia University In The City Of New York Algorithms and system for object-oriented content-based video search
US6202151B1 (en) * 1997-05-09 2001-03-13 Gte Service Corporation System and method for authenticating electronic transactions using biometric certificates
US5987454A (en) * 1997-06-09 1999-11-16 Hobbs; Allen Method and apparatus for selectively augmenting retrieved text, numbers, maps, charts, still pictures and/or graphics, moving pictures and/or graphics and audio information from a network resource
US6169573B1 (en) * 1997-07-03 2001-01-02 Hotv, Inc. Hypervideo system and method with object tracking in a compressed digital video environment
US6188777B1 (en) * 1997-08-01 2001-02-13 Interval Research Corporation Method and apparatus for personnel detection and tracking
US6069630A (en) * 1997-08-22 2000-05-30 International Business Machines Corporation Data processing system and method for creating a link map
US6076104A (en) * 1997-09-04 2000-06-13 Netscape Communications Corp. Video data integration system using image data and associated hypertext links
US6496981B1 (en) * 1997-09-19 2002-12-17 Douglass A. Wistendahl System for converting media content for interactive TV use
US6026411A (en) * 1997-11-06 2000-02-15 International Business Machines Corporation Method, apparatus, and computer program product for generating an image index and for internet searching and querying by image colors
US6184878B1 (en) * 1997-12-23 2001-02-06 Sarnoff Corporation Interactive world wide web access using a set top terminal in a video on demand system
US6032181A (en) * 1998-02-25 2000-02-29 U.S. Philips Corporation Interactive distribution system for digital audio-video signals
US6724915B1 (en) * 1998-03-13 2004-04-20 Siemens Corporate Research, Inc. Method for tracking a video object in a time-ordered sequence of image frames
US6337916B1 (en) * 1998-04-03 2002-01-08 Discreet Logic Inc. Modifying image data
US6154771A (en) * 1998-06-01 2000-11-28 Mediastra, Inc. Real-time receipt, decompression and play of compressed streaming video/hypervideo; with thumbnail display of past scenes and with replay, hyperlinking and/or recording permissively intiated retrospectively
US6603921B1 (en) * 1998-07-01 2003-08-05 International Business Machines Corporation Audio/video archive system and method for automatic indexing and searching
US7020192B1 (en) * 1998-07-31 2006-03-28 Kabushiki Kaisha Toshiba Method of retrieving video picture and apparatus therefor
US6357042B2 (en) * 1998-09-16 2002-03-12 Anand Srinivasan Method and apparatus for multiplexing separately-authored metadata for insertion into a video data stream
US6198833B1 (en) * 1998-09-16 2001-03-06 Hotv, Inc. Enhanced interactive video with object tracking and hyperlinking
US6625315B2 (en) * 1998-10-23 2003-09-23 Facet Technology Corp. Method and apparatus for identifying objects depicted in a videostream
US6209025B1 (en) * 1998-11-09 2001-03-27 John C Bellamy Integrated video system
US6859799B1 (en) * 1998-11-30 2005-02-22 Gemstar Development Corporation Search engine for video and graphics
US6253238B1 (en) * 1998-12-02 2001-06-26 Ictv, Inc. Interactive cable television system with frame grabber
US6639998B1 (en) * 1999-01-11 2003-10-28 Lg Electronics Inc. Method of detecting a specific object in an image signal
US6272250B1 (en) * 1999-01-20 2001-08-07 University Of Washington Color clustering for scene change detection and object tracking in video sequences
US6397181B1 (en) * 1999-01-27 2002-05-28 Kent Ridge Digital Labs Method and apparatus for voice annotation and retrieval of multimedia data
US7003156B1 (en) * 1999-01-29 2006-02-21 Kabushiki Kaisha Toshiba Object detection method and a video data retrieval method
US6819797B1 (en) * 1999-01-29 2004-11-16 International Business Machines Corporation Method and apparatus for classifying and querying temporal and spatial information in video
US7158676B1 (en) * 1999-02-01 2007-01-02 Emuse Media Limited Interactive system
US6990448B2 (en) * 1999-03-05 2006-01-24 Canon Kabushiki Kaisha Database annotation and retrieval including phoneme data
US6526169B1 (en) * 1999-03-15 2003-02-25 Grass Valley (Us), Inc. Histogram-based segmentation of objects from a video signal via color moments
US20050086703A1 (en) * 1999-07-08 2005-04-21 Microsoft Corporation Skimming continuous multimedia content
US6337969B1 (en) * 1999-09-22 2002-01-08 Toshiba Tec Kabushiki Kaisha Fixing device
US6424370B1 (en) * 1999-10-08 2002-07-23 Texas Instruments Incorporated Motion based event detection system and method
US6493707B1 (en) * 1999-10-29 2002-12-10 Verizon Laboratories Inc. Hypervideo: information retrieval using realtime buffers
US6549643B1 (en) * 1999-11-30 2003-04-15 Siemens Corporate Research, Inc. System and method for selecting key-frames of video data
US6478223B1 (en) * 2000-01-12 2002-11-12 Intermec Ip Corporation Machine-readable color symbology and method and apparatus for reading same with standard readers such as laser scanners
US6697796B2 (en) * 2000-01-13 2004-02-24 Agere Systems Inc. Voice clip search
US20030226150A1 (en) * 2000-01-27 2003-12-04 Berberet Suzanne M. System and method for providing broadcast programming, a virtual vcr, and a video scrapbook to programming subscribers
US6744908B2 (en) * 2000-02-29 2004-06-01 Kabushiki Kaisha Toshiba Traffic density analysis apparatus based on encoded video
US6642940B1 (en) * 2000-03-03 2003-11-04 Massachusetts Institute Of Technology Management of properties for hyperlinked video
US7054812B2 (en) * 2000-05-16 2006-05-30 Canon Kabushiki Kaisha Database annotation and retrieval
US20020069218A1 (en) * 2000-07-24 2002-06-06 Sanghoon Sull System and method for indexing, searching, identifying, and editing portions of electronic multimedia files
US6724933B1 (en) * 2000-07-28 2004-04-20 Microsoft Corporation Media segmentation system and related methods
US20050162439A1 (en) * 2000-10-03 2005-07-28 Creative Frontier, Inc. Method and apparatus for associating the color of an object with an event
US20040227768A1 (en) * 2000-10-03 2004-11-18 Creative Frontier, Inc. System and method for tracking an object in a video and linking information thereto
US6925474B2 (en) * 2000-12-07 2005-08-02 Sony United Kingdom Limited Video information retrieval
US7032182B2 (en) * 2000-12-20 2006-04-18 Eastman Kodak Company Graphical user interface adapted to allow scene content annotation of groups of pictures in a picture database to promote efficient database browsing
US20020087530A1 (en) * 2000-12-29 2002-07-04 Expresto Software Corp. System and method for publishing, updating, navigating, and searching documents containing digital video data
US7024020B2 (en) * 2001-01-20 2006-04-04 Samsung Electronics Co., Ltd. Apparatus and method for generating object-labeled image in video sequence
US20030122860A1 (en) * 2001-12-05 2003-07-03 Yuji Ino Video data searching method and video data searching system as well as video data editing method and video data editing system
US20040215660A1 (en) * 2003-02-06 2004-10-28 Canon Kabushiki Kaisha Image search method and apparatus
US20040233233A1 (en) * 2003-05-21 2004-11-25 Salkind Carole T. System and method for embedding interactive items in video and playing same in an interactive environment

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"National Audubon Society". 19 July 1997 and 9 October 1997. Retrieved from http://web.archive.org/web/19970719101311/http://www.audubon.org/ and http://web.archive.org/web/19971009035653/http://www.audubon.org/ *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100275224A1 (en) * 2009-04-24 2010-10-28 Samuel Sheng System and Method for Information Delivery Including Delivery Via Video-Based Keyed or Tagged Content
US20130279744A1 (en) * 2012-04-23 2013-10-24 Apple Inc. Systems and methods for controlling output of content based on human recognition data detection
US9633186B2 (en) * 2012-04-23 2017-04-25 Apple Inc. Systems and methods for controlling output of content based on human recognition data detection
US20170277875A1 (en) * 2012-04-23 2017-09-28 Apple Inc. Systems and methods for controlling output of content based on human recognition data detection
US10360360B2 (en) * 2012-04-23 2019-07-23 Apple Inc. Systems and methods for controlling output of content based on human recognition data detection
US20140240260A1 (en) * 2013-02-25 2014-08-28 Samsung Electronics Co., Ltd. Method and apparatus for providing user interface
WO2015197651A1 (en) * 2014-06-25 2015-12-30 Thomson Licensing Annotation method and corresponding device, computer program product and storage medium

Also Published As

Publication number Publication date
ES2370893T3 (en) 2011-12-23
US20040227768A1 (en) 2004-11-18
ATE509311T1 (en) 2011-05-15
HK1059483A1 (en) 2004-07-02
US6774908B2 (en) 2004-08-10
US20190080193A1 (en) 2019-03-14
US20190318187A1 (en) 2019-10-17
US20020122042A1 (en) 2002-09-05
US20050162439A1 (en) 2005-07-28
US7773093B2 (en) 2010-08-10
US7804506B2 (en) 2010-09-28

Similar Documents

Publication Publication Date Title
US20190318187A1 (en) Method and apparatus for associating the color of an object with an event
JP4037457B2 (en) Network-based access system
US7761326B2 (en) Token type content providing system and token type content providing method and portable user terminal
US6961737B2 (en) Serving signals
US6973619B1 (en) Method for generating display control information and computer
US6151622A (en) Method and system for portably enabling view synchronization over the world-wide web using frame hierarchies
US20040034622A1 (en) Applications software and method for authoring and communicating multimedia content in a multimedia object communication and handling platform
US20110209046A1 (en) Optimizing web content display on an electronic mobile reader
US7127453B1 (en) Gathering data from a database for display
JPH10261106A (en) Hot area in interactive dynamic image
EP1332427B1 (en) System and method for tracking an object in a video and linking information thereto
CA2397337C (en) Serving signals
KR20090101748A (en) Time delaying keyword advertisement contents provding method
KR20090101749A (en) Time delaying keyword advertisement contents provding method
JPH11306196A (en) Information display device
CA2618611A1 (en) Serving signals
JP2002202909A (en) Access method to terminal file by homepage

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION