US20090274437A1 - Embedding data in media metadata tracks during playback - Google Patents

Embedding data in media metadata tracks during playback Download PDF

Info

Publication number
US20090274437A1
US20090274437A1 US12/245,642 US24564208A US2009274437A1 US 20090274437 A1 US20090274437 A1 US 20090274437A1 US 24564208 A US24564208 A US 24564208A US 2009274437 A1 US2009274437 A1 US 2009274437A1
Authority
US
United States
Prior art keywords
media stream
data
network
user
track
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/245,642
Inventor
Todd Stiers
Kent Karlsson
Ulf Windisch
Lars Olsson
Dag Henriksson
Niklas Westin
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
MobiTv Inc
Original Assignee
MobiTv Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by MobiTv Inc filed Critical MobiTv Inc
Priority to US12/245,642 priority Critical patent/US20090274437A1/en
Assigned to MOBITV, INC. reassignment MOBITV, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HENRIKSSON, DAG, OLSSON, LARS, WESTIN, NIKLAS, WINDISCH, ULF, KARLSSON, KENT, STIERS, TODD
Publication of US20090274437A1 publication Critical patent/US20090274437A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/82Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
    • H04N9/8205Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/82Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
    • H04N9/8205Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
    • H04N9/8211Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being a sound signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/82Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
    • H04N9/8205Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
    • H04N9/8233Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being a character code signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/804Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
    • H04N9/8042Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction

Definitions

  • the present disclosure relates to embedding data in media metadata tracks during playback.
  • a variety of conventional mechanisms obtain data and statistics associated with the playback of media streams.
  • the information is logged to allow later retrieval and analysis of statistics.
  • the statistics may include viewing characteristics, network performance information, user settings, etc.
  • FIG. 1 illustrates an exemplary system for use with embodiments of the present invention.
  • FIG. 2 illustrates one example of a Real-Time Transport Protocol (RTP) packet.
  • RTP Real-Time Transport Protocol
  • FIG. 3 illustrates one example of an RTP stream.
  • FIG. 4 illustrates one example of modification of an RTP stream including removal and insertion of packets.
  • FIG. 5 illustrates one example of embedding data in a media stream.
  • FIG. 6 illustrates an example of modifying a media stream based on data embedded in a track.
  • FIG. 7 illustrates one example of a content server.
  • a system uses a processor in a variety of contexts.
  • a system can use multiple processors can while remaining within the scope of the present invention unless otherwise noted.
  • the techniques and mechanisms of the present invention will sometimes describe a connection between two entities. It should be noted that a connection between two entities does not necessarily mean a direct, unimpeded connection, as a variety of other entities may reside between the two entities.
  • a processor may be connected to memory, but it will be appreciated that a variety of bridges and controllers may reside between the processor and memory. Consequently, a connection does not necessarily mean a direct, unimpeded connection unless otherwise noted.
  • Data is embedded onto new or existing media metadata tracks during playback of a media stream.
  • a content server provides a media stream to a mobile device.
  • Data associated with the playback of the media stream on the mobile device is obtained by the content server and saved to the media stream itself.
  • Data may include playback statistics, viewing characteristics, channel changes, comment logs, etc.
  • the information can be stored in a time-correlated manner to allow extraction and analysis of data.
  • a content server provides a media stream to a mobile device. Numerous mobile devices playing the media stream generates a large amount of data that can be transmitted back to the content server. In many instances, the content server collects user characteristics such as channel changes, offers selected, volume light/dark controls, etc., as well as network characteristics such as packet loss, throughput, retransmission, etc. As media streams are played by multiple users in different environments, the amount of data becomes substantial and potentially unwieldy for analysis, particularly real-time playback analysis.
  • Streaming servers typically use logging to maintain data associated with media stream playback on devices such as mobile phones. Some network devices also log network performance metrics to track packet loss, throughput, retransmission rates, etc. However, no existing systems allow efficient storage and management of data associated with media stream playback.
  • the media streams that are delivered usually include at least an audio track and a video track, but the media streams may include additional tracks.
  • media streams may also include captions, subtitles, and real-time metadata describing activity relating directly to the content presented on the other tracks.
  • a content server or other network device records data to a new or existing track in a media stream in real-time or near real-time as the media stream is being played by one or more users.
  • the data includes channel changes, configuration selections, volume and screen brightness changes, option selections, chat logs, packet drops, throughput, retransmission rate, etc.
  • Location, temperature, battery condition, network performance, local-time, and/or anything else happening on the device itself may be reported to a content server or other network device and embedded onto a media stream in a track.
  • the data may be recorded in a time-correlated manner on the media stream. For example, indications that many users adjusted the volume when a particular program ends may be recorded in the track corresponding to the point in the media stream where the program ends.
  • a particularly high number of channel changes may be recorded at the point where a commercial begins.
  • a large number of packet drops may be recorded at the beginning of a program sequence.
  • network transmission rates and commercial selection can be modified during playback to account for user feedback. The modifications can be made during particular points of media stream play back based on the data embedded in the media track.
  • available metadata is analyzed to allow a content server to dynamically modify the media stream provided to a mobile device.
  • Commercials identified by a data track as unappealing by particular demographic groups are replaced with alternate commercials.
  • more appropriate advertising can be provided based on real-time feedback.
  • Programs and commercials can be better produced/validated/verified.
  • parameters collected from a media playback experience can be tightly bound to the media stream itself. Instead of using separate systems to deliver data and generate logs and summaries, collected data can be correlated directly to media being played.
  • data derived from a variety of sources can be embedded in new or existing tracks within the media stream itself.
  • the recording and retrieval of the data happens in real-time or near real-time on every playback of a particular media stream so that data can be used to adjust current playback parameters of the media as it is played a subsequent time.
  • packetization of the media segment can be adjusted to prevent future packet loss.
  • contents of a chat server are correlated with events in a media stream. Ratings buttons or other question and answer responses are elicited.
  • the chat session and questionnaire statistics can be stored with the media stream and used later for search, replay, archival or other ratings and analysis purposes. By placing data associated with playback in the media stream itself, cumbersome correlation of events across different streams can be reduced.
  • the embedded data can be continuously modified and/or expanded as needed.
  • playback options such as charts and graphs can be dynamically generated during playback.
  • FIG. 1 is a diagrammatic representation illustrating one example of a system that can use the techniques and mechanisms of the present invention.
  • content servers 119 , 121 , 123 , and 125 are configured to provide media content to a mobile device 101 using protocols such as RTP and RTCP.
  • protocols such as RTP and RTCP.
  • a mobile device 101 is shown, it should be recognized that other devices such as set top boxes and computer systems can also be used.
  • the content servers 119 , 121 , 123 , and 125 can themselves establish sessions with mobile devices and stream video and audio content to mobile devices.
  • the content servers 119 , 121 , 123 , and 125 have access to a campaign server 143 .
  • the campaign server 143 provides profile information for various mobile devices 101 .
  • the campaign server 143 is itself a content server or a controller.
  • the campaign server 143 can receive information from external sources about devices such as mobile device 101 .
  • the information can be profile information associated with various users of the mobile device including interests and background.
  • the campaign server 143 can also monitor the activity of various devices to gather information about the devices.
  • the content servers 119 , 121 , 123 , and 125 can obtain information about the various devices from the campaign server 143 .
  • a content server 125 uses the campaign server 143 to determine the types of media clips a user on a mobile device 101 would be interested in viewing.
  • the content servers 119 , 121 , 123 , and 125 are receiving media streams from content providers such as satellite providers or cable providers, manipulating the streams, and sending the streams to devices using RTP 131 .
  • content servers 119 , 121 , 123 , and 125 access database 141 to obtain desired content that can be used to supplement streams from satellite and cable providers.
  • the content servers also read, write, and analyze track data in real-time or near real-time as media content is being streamed.
  • a mobile device 101 requests a particular stream.
  • a controller 107 establishes a session with the mobile device 101 using RTSP 133 and the content server 125 begins streaming the content to the mobile device 101 using RTP 131 .
  • Another controller 105 connected to a load balancer may also be used.
  • the content server 125 obtains profile information from campaign server 143 .
  • the content server 125 can also obtain profile information from other sources, such as from the mobile device 101 itself. Using the profile information, the content server can select a clip from a database 141 to provide to a user. In some instances, the clip is injected into a live stream without affecting mobile device application performance. In other instances, the live stream itself is replaced with another live stream. The content server handles processing to make the transition between streams and clips seamless from the point of view of a mobile device application. In still other examples, advertisements from a database 141 can be intelligently selected from the database 141 using profile information from a campaign server 143 and used to seamlessly replace default advertisements in a live stream.
  • Content servers 119 , 121 , 123 , and 125 have the capability to manipulate RTP packets to allow introduction and removal of media content.
  • the content server 125 continually obtains information from a variety of sources.
  • the content server 125 obtains statistical information from mobile device 101 including time duration, channel switching events and frequency, screen brightness, volume settings, commercial interest levels, etc.
  • the content server 125 may also obtain network statistics such as packet transmission rates, throughput, latency, etc.
  • the content server 125 can store this information in a database.
  • the techniques of the present invention recognize that much of this information is time dependent. For example, it may be useful to store information about how many channel changes occur when programming changes or a particular commercial is shown. The techniques and mechanisms of the present invention allow storage of this information in an efficient and easily accessible manner.
  • time dependent information can be embedded in a media stream in an existing track or a new track.
  • an RTP stream is modified to include statistical information aggregated from numerous sources, such as the average number of users in a particular market viewing a program or the average packet transmission rates for a specific sequence.
  • the data is stored in a time-correlated manner. For example, the number of channel changes occurring when a commercial begins can be recorded in the media track at the location when the commercial begins.
  • the content server records the information in real time or near real-time to allow extraction of the information during analysis or subsequent playback.
  • the content server may determine that packet drop rates at a particular point in a stream are too high and the content server may switch to a lower transmission rate stream.
  • the content server may determine that a large percentage of viewers change channels when a particular commercial is played and may insert an alternative commercial.
  • RTP streams and packets may be manipulated based on information embedded in media tracks.
  • FIG. 2 illustrates one example of an RTP packet.
  • An RTP packet 201 includes a header 211 .
  • the header 211 includes information such as the version number, amount of padding, protocol extensions, application level, payload format, etc.
  • the RTP packet 201 also includes a sequence number 213 . Client applications receiving RTP packets expect that the sequence numbers for received packets be unique. If different packets have the same sequence number, erroneous operation can occur.
  • RTP packets also have a timestamp 215 that allows jitter and synchronization calculations.
  • Fields 217 and 219 identify the synchronization source and the contributing source. Extensions are provided in field 221 . Data is provided at 223 .
  • data 231 holds actual media data such as MPEG frames.
  • a single RTP packet 201 holds a single MPEG frame.
  • many RTP packets are required to hold a single MPEG frame.
  • the sequence numbers change across RTP packets while the timestamp 215 remains the same across the different RTP packets.
  • Different MPEG frames include I-frames, P-frames, and B-frames.
  • I-frames are intraframes coded completely by itself.
  • P-frames are predicted frames which require information from a previous I-frame or P-frame.
  • B-frames are bi-directionally predicted frames that require information from surrounding I-frames and P-frames.
  • each track of metadata can be represented in a stream of RTP packets for transport/recording and playback within a subsequent RTSP session.
  • the client player negotiates which RTP tracks to set up during negotiation of an RTSP session with a RTSP/RTP server.
  • the client player has the ability to synchronize and use tracks it is requesting. It should be recognized that a variety of mechanisms can be used to packetize media in their native track formats into RTP, and many ways to record new metadata back into a file are contemplated. It should be noted that a new metadata track can be added to the disk content as new streams of RTP packets are synchronized to the audio and video RTP packet streams. Recording metadata tracks can occur on a client recording during playback or on the server during delivery, or in combination.
  • FIG. 3 illustrates one example of an RTP packet stream that may be used with the techniques of the present invention.
  • An RTP packet stream 301 includes individual packets having a variety of fields and payload data.
  • the fields include a timestamp 303 , sequence 305 , marker 307 , etc.
  • the packets also include payload data 309 holding MPEG frames such as I, P, and B-frames. Timestamps for different packets may be the same. In particular examples, several packets carrying portions of the same I-frame have the same time stamp. However, sequence numbers are different for each packet.
  • Marker bits 307 can be used for different purposes, such as signaling the starting point of an advertisement.
  • packets with sequence numbers 4303 , 4304 , and 4305 carrying potions of the same I-frame and have the same timestamp of 6 .
  • Packets with sequence numbers 4306 , 4307 , 4308 , and 4309 carry P, B, P, and P-frames and have timestamps of 7 , 8 , 9 , and 10 respectively.
  • Packets with sequence numbers 4310 and 4311 carry different portions of the same I-frame and both have the same timestamp of 11.
  • Packets with sequence numbers 4312 , 4313 , 4314 , 4315 , and 4316 carry P, P, B, P, and B-frames respectively and have timestamps 12 , 13 , 14 , 15 , and 16 .
  • the timestamps shown in FIG. 3 are merely representational. Actual timestamps can be computed using a variety of mechanisms.
  • the timestamp is incremented by the packetization interval multiplied by the sampling rate. For example, for audio packets having 20 ms of audio sampled at 8,000 Hz, the timestamp for each block of audio increases by 160. The actual sampling rate may also differ slightly from this nominal rate.
  • the timestamps generated depend on whether the application can determine the frame number. If the application can determine the frame number, the timestamp is governed by the nominal frame rate. Thus, for a 30 f/s video, timestamps would increase by 3,000 for each frame. If a frame is transmitted as several RTP packets, these packets would all bear the same timestamp. If the frame number cannot be determined or if frames are sampled a periodically, as is typically the case for software codecs, the timestamp may be computed from the system clock
  • sequence numbers are used to detect loss. Sequence numbers increase by one for each RTP packet transmitted, timestamps increase by the time “covered” by a packet. For video formats where a video frame is split across several RTP packets, several packets may have the same timestamp. For example, packets with sequence numbers 4317 and 4318 have the same timestamp 17 and carry portions of the same I-frame.
  • FIG. 4 illustrates one example of RTP packet stream modification.
  • An RTP packet stream 401 includes individual packets having a variety of fields and payload data.
  • the fields include a timestamp 403 , sequence 405 , marker 407 , etc.
  • the packets also include payload data 409 holding MPEG frames such as I, P, and B-frames. Timestamps for different packets may be the same. In particular examples, several packets carrying portions of the same I-frame have the same time stamp. However, sequence numbers are different for each packet.
  • Marker bits 407 can be used for different purposes, such as signaling the starting point of an advertisement.
  • metadata searches allow the introduction of targeted advertising that can be inserted seamlessly into an RTP stream.
  • packets with sequence numbers 4303 , 4304 , and 4305 carrying potions of the same I-frame and have the same timestamp of 6 .
  • Packets with sequence numbers 4306 , 4307 , 4308 , and 4309 carry P, B, P, and P-frames and have timestamps of 7 , 8 , 9 , and 10 respectively.
  • a content server removes multiple packets from an RTP packet stream 401 , including packets with sequence numbers 4310 through 4316 .
  • the packets with sequence numbers 4310 and 4311 carry different portions of the same I-frame and both have the same timestamp of 11 .
  • Packets with sequence numbers 4312 , 4313 , 4314 , 4315 , 4316 carry P, P, B, P, and B-frames respectively and have timestamps 12 , 13 , 14 , 15 , and 16 .
  • the spliced stream now ends at packet with sequence number 4309 carrying a P-frame.
  • a B-frame is included in packet having sequence number 4307 . It should be noted that B-frames sometimes may depend on information included in a subsequent I-frame which has been removed. Although having a few B-frames lacking reference frames is not extremely disruptive, it can sometimes be noticed. Therefore, the techniques of the present invention recognize that in some embodiments, the last packets left in a stream prior to splicing should be an I-frame or a P-frame.
  • an RTP sequence 411 can be inserted.
  • the RTP sequence inserted 411 begins with an I-frame for subsequent P and B-frames to reference. Without an I-frame for reference, an RTP sequence inserted may begin with a partial or incomplete picture.
  • the packets for insertion are modified to have sequence numbers following the last sequence number of spliced packet stream 401 .
  • RTP insertion sequence 411 has sequence numbers 4310 - 4317 corresponding to packets carrying I, I, I, B, P, P, B, B, frames respectively, with the I-frame carried in three packets with the same time stamp of 11 and the B, P, P, B, an B-frames having timestamps of 12 - 16 respectively.
  • packets with sequence numbers 4317 and 4318 have the same timestamp 17 and carry portions of the same I-frame.
  • the number of packets in the RTP sequence removed 421 will be exactly the same as the number of packets in the RTP sequence for insertion 411 .
  • the number of packets removed and inserted will differ. For example, some frames may require more than one packet for transmission.
  • timestamps can be configured to be the same, so that a 5 second clip can be replaced with another 5 second clip, the number of packets and consequently the sequence numbers can be thrown askew.
  • packet with sequence number 4309 is referred to herein as a data stream end point packet.
  • Packet with sequence number 4318 is referred to herein as a data stream restart point packet. Packets with sequence numbers 4310 and 4316 in removed sequence are referred to herein as the removed sequence start packet and the removed sequence end packet respectively. Packets with sequence numbers 4310 and 4316 in the insertion sequence are referred to herein as the insertion sequence start packet and the insertion sequence end packet respectively.
  • the content server maintains a current sequence number per RTP data stream and modified subsequent packets after removing and inserting streams. For example, packets having timestamp 17 are modified to have sequence numbers 4318 and 4319 instead of 4317 and 4318 .
  • the content server then proceeds to update subsequent timestamps in the RTP data stream.
  • this operation is uniquely performed at a content server because the content server has information about individual mobile devices and also is able to know information about the sequence numbers of an entire content stream.
  • a content provider may not know information about individual mobile devices, whereas a network device or network switch may not receive all data packets in a sequence. Some packets may have been dropped while others may have been transmitted on different paths.
  • FIG. 5 is a process flow diagram showing one technique for embedding data in media tracks.
  • metadata content for a media stream is extracted and analyzed.
  • the data is obtained from client devices.
  • the data may include user device configurations, user events, channel changes, notes, volume and brightness settings, client player information, client feedback, etc.
  • network data is obtained from network devices.
  • Network data may also be obtained from client devices as well.
  • Network data my include transmission rates, packet drop rates, throughput, etc.
  • data is aggregated from client and network devices. In particular embodiments, data may also be received from other sources, such as chat servers with information corresponding to particular programming.
  • Data is aggregated and written in a time-correlated manner to a new or existing track in the media stream at 511 .
  • the data may be written in real-time or near real-time.
  • the media stream is stored.
  • media stream may continually have new data written as track information. In particular instances, it may be useful to prune some of this data to allow efficient retrieval and access during subsequent playback and analysis.
  • FIG. 6 is a process flow diagram showing a technique for modifying a media stream based on information embedded in a track.
  • metadata content for a particular media stream is analyzed. According to various embodiments, analysis is done for only metadata content for an immediate time period such as 10 minutes of a most recently presented stream.
  • a media stream is modified based on metadata content.
  • advertisements corresponding to search patterns, tags, strings, and sequences may be selected and inserted into the media stream based on user data and network data included in the media stream. For example, if user data indicates that a large percentage of users changed channels when a particular advertisement was shown, an alternative advertisement may be selected.
  • user data for a particular demographic or geographic group may indicate that an alternative advertisement should be selected.
  • alternative advertising is selected.
  • media stream with different transmission rates may be selected.
  • network data may indicate that a particular sequence in a media stream causes a large number of packet drops.
  • a lower bit rate stream may be selected for transmission to user devices in a seamless and efficient manner.
  • the media stream is provided for subsequent playback.
  • FIG. 7 illustrates one example of a content server that can perform live stream modification.
  • a system 700 suitable for implementing particular embodiments of the present invention includes a processor 701 , a memory 703 , an interface 711 , and a bus 715 (e.g., a PCI bus or other interconnection fabric) and operates as a streaming server.
  • the processor 701 When acting under the control of appropriate software or firmware, the processor 701 is responsible for modifying and transmitting live media data to a client.
  • Various specially configured devices can also be used in place of a processor 701 or in addition to processor 701 .
  • the interface 711 is typically configured to end and receive data packets or data segments over a network.
  • interfaces supports include Ethernet interfaces, frame relay interfaces, cable interfaces, DSL interfaces, token ring interfaces, and the like.
  • various very high-speed interfaces may be provided such as fast Ethernet interfaces, Gigabit Ethernet interfaces, ATM interfaces, HSSI interfaces, POS interfaces, FDDI interfaces and the like.
  • these interfaces may include ports appropriate for communication with the appropriate media.
  • they may also include an independent processor and, in some instances, volatile RAM.
  • the independent processors may control such communications intensive tasks as packet switching, media control and management.
  • the system 700 is a content server that also includes a transceiver, streaming buffers, an program guide information.
  • the content server may also be associated with subscription management, logging and report generation, and monitoring capabilities.
  • functionality for allowing operation with mobile devices such as cellular phones operating in a particular cellular network and providing subscription management.
  • an authentication module verifies the identity of devices including mobile devices.
  • a logging and report generation module tracks mobile device requests and associated responses.
  • a monitor system allows an administrator to view usage patterns and system availability.
  • the content server 791 handles requests and responses for media content related transactions while a separate streaming server provides the actual media streams.
  • a particular content server 791 is described, it should be recognized that a variety of alternative configurations are possible. For example, some modules such as a report and logging module 753 and a monitor 751 may not be needed on every server. Alternatively, the modules may be implemented on another device connected to the server. In another example, the server 791 may not include an interface to an abstract buy engine and may in fact include the abstract buy engine itself. A variety of configurations are possible.

Abstract

Data is embedded onto new or existing media metadata tracks during playback of a media stream. A content server provides a media stream to a mobile device. Data associated with the playback of the media stream on the mobile device is obtained by the content server and saved to the media stream itself. Data may include playback statistics, viewing characteristics, channel changes, comment logs, etc. The information can be stored in a time-correlated manner to allow extraction and analysis of data.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • The present invention claims priority to co-pending provisional U.S. patent application No. 61/049,739, filed May 1, 2008, which is incorporated herein by reference in its entirety.
  • TECHNICAL FIELD
  • The present disclosure relates to embedding data in media metadata tracks during playback.
  • DESCRIPTION OF RELATED ART
  • A variety of conventional mechanisms obtain data and statistics associated with the playback of media streams. In many instances, the information is logged to allow later retrieval and analysis of statistics. The statistics may include viewing characteristics, network performance information, user settings, etc.
  • However, conventional mechanisms for maintaining information associated with the playback of media streams are limited. Consequently, it is desirable to provide improved techniques and mechanisms for maintaining information associated with the playback of media streams.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The disclosure may best be understood by reference to the following description taken in conjunction with the accompanying drawings, which illustrate particular embodiments.
  • FIG. 1 illustrates an exemplary system for use with embodiments of the present invention.
  • FIG. 2 illustrates one example of a Real-Time Transport Protocol (RTP) packet.
  • FIG. 3 illustrates one example of an RTP stream.
  • FIG. 4 illustrates one example of modification of an RTP stream including removal and insertion of packets.
  • FIG. 5 illustrates one example of embedding data in a media stream.
  • FIG. 6 illustrates an example of modifying a media stream based on data embedded in a track.
  • FIG. 7 illustrates one example of a content server.
  • DESCRIPTION OF EXAMPLE EMBODIMENTS
  • Reference will now be made in detail to some specific examples of the invention including the best modes contemplated by the inventors for carrying out the invention. Examples of these specific embodiments are illustrated in the accompanying drawings. While the invention is described in conjunction with these specific embodiments, it will be understood that it is not intended to limit the invention to the described embodiments. On the contrary, it is intended to cover alternatives, modifications, and equivalents as may be included within the spirit and scope of the invention as defined by the appended claims.
  • For example, the techniques of the present invention will be described in the context of particular networks and devices. However, it should be noted that the techniques of the present invention apply to a variety of networks and devices. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. Particular example embodiments of the present invention may be implemented without some or all of these specific details. In other instances, well known process operations have not been described in detail in order not to unnecessarily obscure the present invention.
  • Various techniques and mechanisms of the present invention will sometimes be described in singular form for clarity. However, it should be noted that some embodiments include multiple iterations of a technique or multiple instantiations of a mechanism unless noted otherwise. For example, a system uses a processor in a variety of contexts. However, it will be appreciated that a system can use multiple processors can while remaining within the scope of the present invention unless otherwise noted. Furthermore, the techniques and mechanisms of the present invention will sometimes describe a connection between two entities. It should be noted that a connection between two entities does not necessarily mean a direct, unimpeded connection, as a variety of other entities may reside between the two entities. For example, a processor may be connected to memory, but it will be appreciated that a variety of bridges and controllers may reside between the processor and memory. Consequently, a connection does not necessarily mean a direct, unimpeded connection unless otherwise noted.
  • Overview
  • Data is embedded onto new or existing media metadata tracks during playback of a media stream. A content server provides a media stream to a mobile device. Data associated with the playback of the media stream on the mobile device is obtained by the content server and saved to the media stream itself. Data may include playback statistics, viewing characteristics, channel changes, comment logs, etc. The information can be stored in a time-correlated manner to allow extraction and analysis of data.
  • Example Embodiments
  • In many systems, a content server provides a media stream to a mobile device. Numerous mobile devices playing the media stream generates a large amount of data that can be transmitted back to the content server. In many instances, the content server collects user characteristics such as channel changes, offers selected, volume light/dark controls, etc., as well as network characteristics such as packet loss, throughput, retransmission, etc. As media streams are played by multiple users in different environments, the amount of data becomes substantial and potentially unwieldy for analysis, particularly real-time playback analysis.
  • Streaming servers typically use logging to maintain data associated with media stream playback on devices such as mobile phones. Some network devices also log network performance metrics to track packet loss, throughput, retransmission rates, etc. However, no existing systems allow efficient storage and management of data associated with media stream playback.
  • The media streams that are delivered usually include at least an audio track and a video track, but the media streams may include additional tracks. For example, media streams may also include captions, subtitles, and real-time metadata describing activity relating directly to the content presented on the other tracks.
  • According to various embodiments, a content server or other network device records data to a new or existing track in a media stream in real-time or near real-time as the media stream is being played by one or more users. According to various embodiments, the data includes channel changes, configuration selections, volume and screen brightness changes, option selections, chat logs, packet drops, throughput, retransmission rate, etc. Location, temperature, battery condition, network performance, local-time, and/or anything else happening on the device itself may be reported to a content server or other network device and embedded onto a media stream in a track. The data may be recorded in a time-correlated manner on the media stream. For example, indications that many users adjusted the volume when a particular program ends may be recorded in the track corresponding to the point in the media stream where the program ends.
  • According to various embodiments, a particularly high number of channel changes may be recorded at the point where a commercial begins. In other embodiments, a large number of packet drops may be recorded at the beginning of a program sequence. In particular embodiments, network transmission rates and commercial selection can be modified during playback to account for user feedback. The modifications can be made during particular points of media stream play back based on the data embedded in the media track.
  • According to various embodiments, as media is played, available metadata is analyzed to allow a content server to dynamically modify the media stream provided to a mobile device. Commercials identified by a data track as unappealing by particular demographic groups are replaced with alternate commercials.
  • In particular embodiments, more appropriate advertising can be provided based on real-time feedback. Programs and commercials can be better produced/validated/verified. According to various embodiments, parameters collected from a media playback experience can be tightly bound to the media stream itself. Instead of using separate systems to deliver data and generate logs and summaries, collected data can be correlated directly to media being played. According to various embodiments, data derived from a variety of sources can be embedded in new or existing tracks within the media stream itself. In particular embodiments, the recording and retrieval of the data happens in real-time or near real-time on every playback of a particular media stream so that data can be used to adjust current playback parameters of the media as it is played a subsequent time.
  • In a particular example, if a particular network condition was observed in the last playback of a media segment that indicated packet loss, packetization of the media segment can be adjusted to prevent future packet loss. In yet another example, contents of a chat server are correlated with events in a media stream. Ratings buttons or other question and answer responses are elicited. The chat session and questionnaire statistics can be stored with the media stream and used later for search, replay, archival or other ratings and analysis purposes. By placing data associated with playback in the media stream itself, cumbersome correlation of events across different streams can be reduced. The embedded data can be continuously modified and/or expanded as needed. In some examples, playback options such as charts and graphs can be dynamically generated during playback.
  • FIG. 1 is a diagrammatic representation illustrating one example of a system that can use the techniques and mechanisms of the present invention. According to various embodiments, content servers 119, 121, 123, and 125 are configured to provide media content to a mobile device 101 using protocols such as RTP and RTCP. Although a mobile device 101 is shown, it should be recognized that other devices such as set top boxes and computer systems can also be used. In particular examples, the content servers 119, 121, 123, and 125 can themselves establish sessions with mobile devices and stream video and audio content to mobile devices.
  • According to various embodiments, the content servers 119, 121, 123, and 125 have access to a campaign server 143. The campaign server 143 provides profile information for various mobile devices 101. In some examples, the campaign server 143 is itself a content server or a controller. The campaign server 143 can receive information from external sources about devices such as mobile device 101. The information can be profile information associated with various users of the mobile device including interests and background. The campaign server 143 can also monitor the activity of various devices to gather information about the devices. The content servers 119, 121, 123, and 125 can obtain information about the various devices from the campaign server 143. In particular examples, a content server 125 uses the campaign server 143 to determine the types of media clips a user on a mobile device 101 would be interested in viewing.
  • According to various embodiments, the content servers 119, 121, 123, and 125 are receiving media streams from content providers such as satellite providers or cable providers, manipulating the streams, and sending the streams to devices using RTP 131. In particular examples, content servers 119, 121, 123, and 125 access database 141 to obtain desired content that can be used to supplement streams from satellite and cable providers. According to various embodiments, the content servers also read, write, and analyze track data in real-time or near real-time as media content is being streamed. In one example, a mobile device 101 requests a particular stream. A controller 107 establishes a session with the mobile device 101 using RTSP 133 and the content server 125 begins streaming the content to the mobile device 101 using RTP 131. Another controller 105 connected to a load balancer may also be used. In particular examples, the content server 125 obtains profile information from campaign server 143.
  • In some examples, the content server 125 can also obtain profile information from other sources, such as from the mobile device 101 itself. Using the profile information, the content server can select a clip from a database 141 to provide to a user. In some instances, the clip is injected into a live stream without affecting mobile device application performance. In other instances, the live stream itself is replaced with another live stream. The content server handles processing to make the transition between streams and clips seamless from the point of view of a mobile device application. In still other examples, advertisements from a database 141 can be intelligently selected from the database 141 using profile information from a campaign server 143 and used to seamlessly replace default advertisements in a live stream. Content servers 119, 121, 123, and 125 have the capability to manipulate RTP packets to allow introduction and removal of media content.
  • According to various embodiments, the content server 125 continually obtains information from a variety of sources. In particular embodiments, the content server 125 obtains statistical information from mobile device 101 including time duration, channel switching events and frequency, screen brightness, volume settings, commercial interest levels, etc. The content server 125 may also obtain network statistics such as packet transmission rates, throughput, latency, etc. According to various embodiments, the content server 125 can store this information in a database. However, the techniques of the present invention recognize that much of this information is time dependent. For example, it may be useful to store information about how many channel changes occur when programming changes or a particular commercial is shown. The techniques and mechanisms of the present invention allow storage of this information in an efficient and easily accessible manner. In particular embodiments, time dependent information can be embedded in a media stream in an existing track or a new track. According to various embodiments, an RTP stream is modified to include statistical information aggregated from numerous sources, such as the average number of users in a particular market viewing a program or the average packet transmission rates for a specific sequence. In particular embodiments, the data is stored in a time-correlated manner. For example, the number of channel changes occurring when a commercial begins can be recorded in the media track at the location when the commercial begins.
  • The content server records the information in real time or near real-time to allow extraction of the information during analysis or subsequent playback. In particular embodiments, the content server may determine that packet drop rates at a particular point in a stream are too high and the content server may switch to a lower transmission rate stream. In another example, the content server may determine that a large percentage of viewers change channels when a particular commercial is played and may insert an alternative commercial. According to various embodiments, RTP streams and packets may be manipulated based on information embedded in media tracks.
  • FIG. 2 illustrates one example of an RTP packet. An RTP packet 201 includes a header 211. According to various embodiments, the header 211 includes information such as the version number, amount of padding, protocol extensions, application level, payload format, etc. The RTP packet 201 also includes a sequence number 213. Client applications receiving RTP packets expect that the sequence numbers for received packets be unique. If different packets have the same sequence number, erroneous operation can occur. RTP packets also have a timestamp 215 that allows jitter and synchronization calculations. Fields 217 and 219 identify the synchronization source and the contributing source. Extensions are provided in field 221. Data is provided at 223.
  • According to various embodiments, data 231 holds actual media data such as MPEG frames. In some examples, a single RTP packet 201 holds a single MPEG frame. In many instances, many RTP packets are required to hold a single MPEG frame. In instances where multiple RTP packets are required for a single MPEG frame, the sequence numbers change across RTP packets while the timestamp 215 remains the same across the different RTP packets. Different MPEG frames include I-frames, P-frames, and B-frames. I-frames are intraframes coded completely by itself. P-frames are predicted frames which require information from a previous I-frame or P-frame. B-frames are bi-directionally predicted frames that require information from surrounding I-frames and P-frames.
  • Because different MPEG frames require different numbers of RTP packets for transmission, two different streams of the same time duration may require different numbers of RTP packets for transmission. Simply replacing a clip with another clip would not work, as the clips may have different numbers of RTP packets and having different impacts on the sequence numbers of subsequent packets.
  • According to various embodiments, each track of metadata can be represented in a stream of RTP packets for transport/recording and playback within a subsequent RTSP session. As background, the client player negotiates which RTP tracks to set up during negotiation of an RTSP session with a RTSP/RTP server. In particular embodiments, the client player has the ability to synchronize and use tracks it is requesting. It should be recognized that a variety of mechanisms can be used to packetize media in their native track formats into RTP, and many ways to record new metadata back into a file are contemplated. It should be noted that a new metadata track can be added to the disk content as new streams of RTP packets are synchronized to the audio and video RTP packet streams. Recording metadata tracks can occur on a client recording during playback or on the server during delivery, or in combination.
  • FIG. 3 illustrates one example of an RTP packet stream that may be used with the techniques of the present invention. An RTP packet stream 301 includes individual packets having a variety of fields and payload data. According to various embodiments, the fields include a timestamp 303, sequence 305, marker 307, etc. The packets also include payload data 309 holding MPEG frames such as I, P, and B-frames. Timestamps for different packets may be the same. In particular examples, several packets carrying portions of the same I-frame have the same time stamp. However, sequence numbers are different for each packet. Marker bits 307 can be used for different purposes, such as signaling the starting point of an advertisement.
  • According to various embodiments, packets with sequence numbers 4303, 4304, and 4305 carrying potions of the same I-frame and have the same timestamp of 6. Packets with sequence numbers 4306, 4307, 4308, and 4309 carry P, B, P, and P-frames and have timestamps of 7, 8, 9, and 10 respectively. Packets with sequence numbers 4310 and 4311 carry different portions of the same I-frame and both have the same timestamp of 11. Packets with sequence numbers 4312, 4313, 4314, 4315, and 4316 carry P, P, B, P, and B-frames respectively and have timestamps 12, 13, 14, 15, and 16. It should be noted that the timestamps shown in FIG. 3 are merely representational. Actual timestamps can be computed using a variety of mechanisms.
  • For many audio encodings, the timestamp is incremented by the packetization interval multiplied by the sampling rate. For example, for audio packets having 20 ms of audio sampled at 8,000 Hz, the timestamp for each block of audio increases by 160. The actual sampling rate may also differ slightly from this nominal rate. For many video encodings, the timestamps generated depend on whether the application can determine the frame number. If the application can determine the frame number, the timestamp is governed by the nominal frame rate. Thus, for a 30 f/s video, timestamps would increase by 3,000 for each frame. If a frame is transmitted as several RTP packets, these packets would all bear the same timestamp. If the frame number cannot be determined or if frames are sampled a periodically, as is typically the case for software codecs, the timestamp may be computed from the system clock
  • While the timestamp is used by a receiver to place the incoming media data in the correct timing order and provide playout delay compensation, the sequence numbers are used to detect loss. Sequence numbers increase by one for each RTP packet transmitted, timestamps increase by the time “covered” by a packet. For video formats where a video frame is split across several RTP packets, several packets may have the same timestamp. For example, packets with sequence numbers 4317 and 4318 have the same timestamp 17 and carry portions of the same I-frame.
  • FIG. 4 illustrates one example of RTP packet stream modification. An RTP packet stream 401 includes individual packets having a variety of fields and payload data. According to various embodiments, the fields include a timestamp 403, sequence 405, marker 407, etc. The packets also include payload data 409 holding MPEG frames such as I, P, and B-frames. Timestamps for different packets may be the same. In particular examples, several packets carrying portions of the same I-frame have the same time stamp. However, sequence numbers are different for each packet. Marker bits 407 can be used for different purposes, such as signaling the starting point of an advertisement. According to various embodiments, metadata searches allow the introduction of targeted advertising that can be inserted seamlessly into an RTP stream.
  • According to various embodiments, packets with sequence numbers 4303, 4304, and 4305 carrying potions of the same I-frame and have the same timestamp of 6. Packets with sequence numbers 4306, 4307, 4308, and 4309 carry P, B, P, and P-frames and have timestamps of 7, 8, 9, and 10 respectively. According to various embodiments, a content server removes multiple packets from an RTP packet stream 401, including packets with sequence numbers 4310 through 4316. The packets with sequence numbers 4310 and 4311 carry different portions of the same I-frame and both have the same timestamp of 11.
  • Packets with sequence numbers 4312, 4313, 4314, 4315, 4316 carry P, P, B, P, and B-frames respectively and have timestamps 12, 13, 14, 15, and 16. The spliced stream now ends at packet with sequence number 4309 carrying a P-frame. A B-frame is included in packet having sequence number 4307. It should be noted that B-frames sometimes may depend on information included in a subsequent I-frame which has been removed. Although having a few B-frames lacking reference frames is not extremely disruptive, it can sometimes be noticed. Therefore, the techniques of the present invention recognize that in some embodiments, the last packets left in a stream prior to splicing should be an I-frame or a P-frame.
  • According to various embodiments, now that a portion of the RTP stream has been removed, an RTP sequence 411 can be inserted. In particular examples, the RTP sequence inserted 411 begins with an I-frame for subsequent P and B-frames to reference. Without an I-frame for reference, an RTP sequence inserted may begin with a partial or incomplete picture. The packets for insertion are modified to have sequence numbers following the last sequence number of spliced packet stream 401. RTP insertion sequence 411 has sequence numbers 4310 -4317 corresponding to packets carrying I, I, I, B, P, P, B, B, frames respectively, with the I-frame carried in three packets with the same time stamp of 11 and the B, P, P, B, an B-frames having timestamps of 12-16 respectively.
  • For example, packets with sequence numbers 4317 and 4318 have the same timestamp 17 and carry portions of the same I-frame. In some instances, the number of packets in the RTP sequence removed 421 will be exactly the same as the number of packets in the RTP sequence for insertion 411. However, in many instances, the number of packets removed and inserted will differ. For example, some frames may require more than one packet for transmission. Although timestamps can be configured to be the same, so that a 5 second clip can be replaced with another 5 second clip, the number of packets and consequently the sequence numbers can be thrown askew. According to various embodiments, packet with sequence number 4309 is referred to herein as a data stream end point packet. Packet with sequence number 4318 is referred to herein as a data stream restart point packet. Packets with sequence numbers 4310 and 4316 in removed sequence are referred to herein as the removed sequence start packet and the removed sequence end packet respectively. Packets with sequence numbers 4310 and 4316 in the insertion sequence are referred to herein as the insertion sequence start packet and the insertion sequence end packet respectively.
  • Consequently, the content server maintains a current sequence number per RTP data stream and modified subsequent packets after removing and inserting streams. For example, packets having timestamp 17 are modified to have sequence numbers 4318 and 4319 instead of 4317 and 4318. The content server then proceeds to update subsequent timestamps in the RTP data stream. According to various embodiments, this operation is uniquely performed at a content server because the content server has information about individual mobile devices and also is able to know information about the sequence numbers of an entire content stream. A content provider may not know information about individual mobile devices, whereas a network device or network switch may not receive all data packets in a sequence. Some packets may have been dropped while others may have been transmitted on different paths.
  • FIG. 5 is a process flow diagram showing one technique for embedding data in media tracks. At 501, metadata content for a media stream is extracted and analyzed. At 503, the data is obtained from client devices. The data may include user device configurations, user events, channel changes, notes, volume and brightness settings, client player information, client feedback, etc. At 505, network data is obtained from network devices. Network data may also be obtained from client devices as well. Network data my include transmission rates, packet drop rates, throughput, etc. At 509, data is aggregated from client and network devices. In particular embodiments, data may also be received from other sources, such as chat servers with information corresponding to particular programming. Data is aggregated and written in a time-correlated manner to a new or existing track in the media stream at 511. The data may be written in real-time or near real-time. At 513, the media stream is stored. In some examples, media stream may continually have new data written as track information. In particular instances, it may be useful to prune some of this data to allow efficient retrieval and access during subsequent playback and analysis.
  • FIG. 6 is a process flow diagram showing a technique for modifying a media stream based on information embedded in a track. At 601, metadata content for a particular media stream is analyzed. According to various embodiments, analysis is done for only metadata content for an immediate time period such as 10 minutes of a most recently presented stream. At 603, a media stream is modified based on metadata content. In particular embodiments, advertisements corresponding to search patterns, tags, strings, and sequences may be selected and inserted into the media stream based on user data and network data included in the media stream. For example, if user data indicates that a large percentage of users changed channels when a particular advertisement was shown, an alternative advertisement may be selected. In particular embodiments, user data for a particular demographic or geographic group may indicate that an alternative advertisement should be selected. At 605, alternative advertising is selected. At 607, media stream with different transmission rates may be selected. According to various embodiments, network data may indicate that a particular sequence in a media stream causes a large number of packet drops. A lower bit rate stream may be selected for transmission to user devices in a seamless and efficient manner. At 609, the media stream is provided for subsequent playback.
  • FIG. 7 illustrates one example of a content server that can perform live stream modification. According to particular embodiments, a system 700 suitable for implementing particular embodiments of the present invention includes a processor 701, a memory 703, an interface 711, and a bus 715 (e.g., a PCI bus or other interconnection fabric) and operates as a streaming server. When acting under the control of appropriate software or firmware, the processor 701 is responsible for modifying and transmitting live media data to a client. Various specially configured devices can also be used in place of a processor 701 or in addition to processor 701. The interface 711 is typically configured to end and receive data packets or data segments over a network.
  • Particular examples of interfaces supports include Ethernet interfaces, frame relay interfaces, cable interfaces, DSL interfaces, token ring interfaces, and the like. In addition, various very high-speed interfaces may be provided such as fast Ethernet interfaces, Gigabit Ethernet interfaces, ATM interfaces, HSSI interfaces, POS interfaces, FDDI interfaces and the like. Generally, these interfaces may include ports appropriate for communication with the appropriate media. In some cases, they may also include an independent processor and, in some instances, volatile RAM. The independent processors may control such communications intensive tasks as packet switching, media control and management.
  • According to various embodiments, the system 700 is a content server that also includes a transceiver, streaming buffers, an program guide information. The content server may also be associated with subscription management, logging and report generation, and monitoring capabilities. In particular embodiments, functionality for allowing operation with mobile devices such as cellular phones operating in a particular cellular network and providing subscription management. According to various embodiments, an authentication module verifies the identity of devices including mobile devices. A logging and report generation module tracks mobile device requests and associated responses. A monitor system allows an administrator to view usage patterns and system availability. According to various embodiments, the content server 791 handles requests and responses for media content related transactions while a separate streaming server provides the actual media streams.
  • Although a particular content server 791 is described, it should be recognized that a variety of alternative configurations are possible. For example, some modules such as a report and logging module 753 and a monitor 751 may not be needed on every server. Alternatively, the modules may be implemented on another device connected to the server. In another example, the server 791 may not include an interface to an abstract buy engine and may in fact include the abstract buy engine itself. A variety of configurations are possible.
  • In the foregoing specification, the invention has been described with reference to specific embodiments. However, one of ordinary skill in the art appreciates that various modifications and changes can be made without departing from the scope of the invention as set forth in the claims below. Accordingly, the specification and figures are to be regarded in an illustrative rather than a restrictive sense, and all such modifications are intended to be included within the scope of invention.

Claims (20)

1. A method, comprising:
receiving a media stream, the media stream including a video track and an audio track;
transmitting the media stream to a plurality of user devices over a network including a plurality of network devices;
receiving user data from the plurality of user devices;
embedding the user data in a time-correlated manner in a track associated with the media stream;
storing the media stream for subsequent playback.
2. The method of claim 1, wherein user data comprises channel changes and device settings.
3. The method of claim 2, wherein the user data is embedded in a new track in the media stream.
4. The method of claim 1, further comprising receiving network data from the plurality of network devices.
5. The method of claim 4, wherein network data comprises transmissions rates.
6. The method of claim 5, wherein the network data is embedded in a new track in the media stream.
7. The method of claim 1, further comprising receiving network data from the plurality of network devices.
8. The method of claim 1, wherein the user data indicates the percentages of user devices switching channels when a particular portion of the media stream is played.
9. The method of claim 1, wherein user data comprises commercial feedback and user ratings.
10. A system, comprising:
means for receiving a media stream, the media stream including a video track and an audio track;
means for transmitting the media stream to a plurality of user devices over a network including a plurality of network devices;
means for receiving user data from the plurality of user devices;
means for embedding the user data in a time-correlated manner in a track associated with the media stream;
means for storing the media stream for subsequent playback.
11. The system of claim 10, wherein user data comprises channel changes and device settings.
12. The system of claim 11, wherein the user data is embedded in a new track in the media stream.
13. The system of claim 10, further comprising means for receiving network data from the plurality of network devices.
14. The system of claim 13, wherein network data comprises transmissions rates.
15. The system of claim 14, wherein the network data is embedded in a new track in the media stream.
16. The system of claim 10, further comprising means for receiving network data from the plurality of network devices.
17. The system of claim 10, wherein the user data indicates the percentages of user devices switching channels when a particular portion of the media stream is played.
18. The system of claim 10, wherein user data comprises commercial feedback and user ratings.
19. A computer readable medium having computer code embodied therein, the computer readable medium comprising:
computer code for receiving a media stream, the media stream including a video track and an audio track;
computer code for transmitting the media stream to a plurality of user devices over a network including a plurality of network devices;
computer code for receiving user data from the plurality of user devices;
computer code for embedding the user data in a time-correlated manner in a track associated with the media stream;
computer code for storing the media stream for subsequent playback.
20. The computer readable medium of claim 19, wherein user data comprises channel changes and device settings.
US12/245,642 2008-05-01 2008-10-03 Embedding data in media metadata tracks during playback Abandoned US20090274437A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/245,642 US20090274437A1 (en) 2008-05-01 2008-10-03 Embedding data in media metadata tracks during playback

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US4973908P 2008-05-01 2008-05-01
US12/245,642 US20090274437A1 (en) 2008-05-01 2008-10-03 Embedding data in media metadata tracks during playback

Publications (1)

Publication Number Publication Date
US20090274437A1 true US20090274437A1 (en) 2009-11-05

Family

ID=41257147

Family Applications (5)

Application Number Title Priority Date Filing Date
US12/181,186 Abandoned US20090276402A1 (en) 2008-05-01 2008-07-28 Search system using media metadata tracks
US12/245,642 Abandoned US20090274437A1 (en) 2008-05-01 2008-10-03 Embedding data in media metadata tracks during playback
US13/530,310 Active US10250841B2 (en) 2008-05-01 2012-06-22 System and method for modifying media streams using metadata
US16/372,294 Active US11917323B2 (en) 2008-05-01 2019-04-01 System and method for modifying media streams using metadata
US18/132,712 Pending US20230319229A1 (en) 2008-05-01 2023-04-10 System and method for modifying media streams using metadata

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/181,186 Abandoned US20090276402A1 (en) 2008-05-01 2008-07-28 Search system using media metadata tracks

Family Applications After (3)

Application Number Title Priority Date Filing Date
US13/530,310 Active US10250841B2 (en) 2008-05-01 2012-06-22 System and method for modifying media streams using metadata
US16/372,294 Active US11917323B2 (en) 2008-05-01 2019-04-01 System and method for modifying media streams using metadata
US18/132,712 Pending US20230319229A1 (en) 2008-05-01 2023-04-10 System and method for modifying media streams using metadata

Country Status (1)

Country Link
US (5) US20090276402A1 (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110302618A1 (en) * 2010-06-04 2011-12-08 Mobitv, Inc. Fragmented file structure for live media stream delivery
US20120099588A1 (en) * 2009-08-24 2012-04-26 Xiangping Liu Channel change method, device, and system
US20130007576A1 (en) * 2011-06-30 2013-01-03 Cable Television Laboratories, Inc. Synchronization of web applications and media
US20140351256A1 (en) * 2013-05-22 2014-11-27 Nice-Systems Ltd. System and method for generating metadata for a recorded session
US20150082463A1 (en) * 2010-08-30 2015-03-19 Mobitv, Inc. Media rights management on multiple devices
US20160094501A1 (en) * 2014-09-26 2016-03-31 Line Corporation Method, system and recording medium for providing video contents in social platform and file distribution system
US9332302B2 (en) 2008-01-30 2016-05-03 Cinsay, Inc. Interactive product placement system and method therefor
WO2016085700A1 (en) * 2014-11-24 2016-06-02 Roku, Inc. Apparatus and method for content playback utilizing crowd sourced statistics
US20160269455A1 (en) * 2015-03-10 2016-09-15 Mobitv, Inc. Media seek mechanisms
US10055768B2 (en) 2008-01-30 2018-08-21 Cinsay, Inc. Interactive product placement system and method therefor
US11218741B2 (en) 2019-06-03 2022-01-04 Alibaba Group Holding Limited Method, apparatus, electronic device and computer-readable storage medium for video buffering
US11227315B2 (en) 2008-01-30 2022-01-18 Aibuy, Inc. Interactive product placement system and method therefor

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8888592B1 (en) 2009-06-01 2014-11-18 Sony Computer Entertainment America Llc Voice overlay
US9349201B1 (en) 2006-08-03 2016-05-24 Sony Interactive Entertainment America Llc Command sentinel
US8968087B1 (en) 2009-06-01 2015-03-03 Sony Computer Entertainment America Llc Video game overlay
US8613673B2 (en) 2008-12-15 2013-12-24 Sony Computer Entertainment America Llc Intelligent game loading
US9498714B2 (en) 2007-12-15 2016-11-22 Sony Interactive Entertainment America Llc Program mode switching
US20090276402A1 (en) 2008-05-01 2009-11-05 Mobitv, Inc. Search system using media metadata tracks
US20100174603A1 (en) * 2008-10-14 2010-07-08 Robert Hughes System and Method for Advertising Placement and/or Web Site Optimization
US20100094962A1 (en) * 2008-10-15 2010-04-15 Patentvc Ltd. Internet backbone servers with edge compensation
US9049198B2 (en) * 2008-10-15 2015-06-02 Aster Risk Management Llc Methods and systems for distributing pull protocol requests via a relay server
US8506402B2 (en) * 2009-06-01 2013-08-13 Sony Computer Entertainment America Llc Game execution environments
US9426502B2 (en) 2011-11-11 2016-08-23 Sony Interactive Entertainment America Llc Real-time cloud-based video watermarking systems and methods
US20110145370A1 (en) * 2009-08-31 2011-06-16 Bruno Nieuwenhuys Methods and systems to personalize content streams
US8560331B1 (en) 2010-08-02 2013-10-15 Sony Computer Entertainment America Llc Audio acceleration
JP5652642B2 (en) * 2010-08-02 2015-01-14 ソニー株式会社 Data generation apparatus, data generation method, data processing apparatus, and data processing method
KR102126910B1 (en) 2010-09-13 2020-06-25 소니 인터랙티브 엔터테인먼트 아메리카 엘엘씨 Add-on Management
KR102003007B1 (en) 2010-09-13 2019-07-23 소니 인터랙티브 엔터테인먼트 아메리카 엘엘씨 A Method and System of Providing a Computer Game at a Computer Game System Including a Video Server and a Game Server
US20120151079A1 (en) * 2010-12-13 2012-06-14 Jan Besehanic Methods and apparatus to measure media exposure
US9785639B2 (en) * 2012-04-27 2017-10-10 Mobitv, Inc. Search-based navigation of media content
US8863175B2 (en) * 2012-12-19 2014-10-14 Ebay Inc. Method and system for targeted commerce in network broadcasting
US9712885B2 (en) * 2014-03-17 2017-07-18 Cbs Interactive Inc. Techniques to select advertisements using closed captioning data
US11250872B2 (en) 2019-12-14 2022-02-15 International Business Machines Corporation Using closed captions as parallel training data for customization of closed captioning systems

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060090179A1 (en) * 2004-10-26 2006-04-27 Ya-Ling Hsu System and method for embedding supplemental information into a digital stream of a work of content
US7295752B1 (en) * 1997-08-14 2007-11-13 Virage, Inc. Video cataloger system with audio track extraction
US20080092168A1 (en) * 1999-03-29 2008-04-17 Logan James D Audio and video program recording, editing and playback systems using metadata
US7403224B2 (en) * 1998-09-01 2008-07-22 Virage, Inc. Embedded metadata engines in digital capture devices

Family Cites Families (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7372976B2 (en) * 1998-04-16 2008-05-13 Digimarc Corporation Content indexing and searching using content identifiers and associated metadata
US6226618B1 (en) * 1998-08-13 2001-05-01 International Business Machines Corporation Electronic content delivery system
US20020124262A1 (en) * 1999-12-01 2002-09-05 Andrea Basso Network based replay portal
US20010053996A1 (en) * 2000-01-06 2001-12-20 Atkinson Paul D. System and method for distributing and controlling the output of media in public spaces
US6721741B1 (en) * 2000-01-24 2004-04-13 Friskit, Inc. Streaming media search system
US7281034B1 (en) * 2000-01-24 2007-10-09 Friskit, Inc. System and method for media playback over a network using links that contain control signals and commands
US7222163B1 (en) * 2000-04-07 2007-05-22 Virage, Inc. System and method for hosting of video content over a network
US6760721B1 (en) 2000-04-14 2004-07-06 Realnetworks, Inc. System and method of managing metadata data
US20020103920A1 (en) * 2000-11-21 2002-08-01 Berkun Ken Alan Interpretive stream metadata extraction
US7685105B2 (en) * 2001-04-05 2010-03-23 Envirospectives, Inc. System and method for indexing, organizing, storing and retrieving environmental information
WO2002091186A1 (en) * 2001-05-08 2002-11-14 Ipool Corporation Privacy protection system and method
US7266553B1 (en) * 2002-07-01 2007-09-04 Microsoft Corporation Content data indexing
WO2004090899A1 (en) * 2003-04-07 2004-10-21 Internet Pro Video Limited Electronic device with media manipulation capabilities
JP5356652B2 (en) * 2004-02-12 2013-12-04 コア ワイアレス ライセンシング エス アー アール エル Classified media experience quality
US7571181B2 (en) * 2004-04-05 2009-08-04 Hewlett-Packard Development Company, L.P. Network usage analysis system and method for detecting network congestion
FR2870661A1 (en) * 2004-05-19 2005-11-25 France Telecom METHOD FOR OPTIMALLY MANAGING THE CHARGE OF A BATTERY OF A MOBILE TELECOMMUNICATION TERMINAL
US9286388B2 (en) * 2005-08-04 2016-03-15 Time Warner Cable Enterprises Llc Method and apparatus for context-specific content delivery
US20070136742A1 (en) * 2005-12-13 2007-06-14 General Instrument Corporation Method, apparatus and system for replacing advertisements in recorded video content
US7827191B2 (en) * 2005-12-14 2010-11-02 Microsoft Corporation Discovering web-based multimedia using search toolbar data
US7613691B2 (en) * 2006-06-21 2009-11-03 Microsoft Corporation Dynamic insertion of supplemental video based on metadata
US7796779B1 (en) * 2006-08-24 2010-09-14 SugarSync, Inc. Efficient synchronization of changes to images
US20080147608A1 (en) * 2006-12-14 2008-06-19 Yahoo! Inc. Video search and indexing systems and methods
US20080284910A1 (en) * 2007-01-31 2008-11-20 John Erskine Text data for streaming video
US20090094113A1 (en) * 2007-09-07 2009-04-09 Digitalsmiths Corporation Systems and Methods For Using Video Metadata to Associate Advertisements Therewith
US20090187941A1 (en) * 2008-01-21 2009-07-23 Att Knowledge Ventures L.P. System and method for targeted advertising
US9955122B2 (en) * 2008-04-11 2018-04-24 Mobitv, Inc. Dynamic advertisement stream replacement
US20090276402A1 (en) 2008-05-01 2009-11-05 Mobitv, Inc. Search system using media metadata tracks
US9084075B2 (en) * 2011-09-15 2015-07-14 Qualcomm Incorporated Tracking management systems and methods

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7295752B1 (en) * 1997-08-14 2007-11-13 Virage, Inc. Video cataloger system with audio track extraction
US7403224B2 (en) * 1998-09-01 2008-07-22 Virage, Inc. Embedded metadata engines in digital capture devices
US20080092168A1 (en) * 1999-03-29 2008-04-17 Logan James D Audio and video program recording, editing and playback systems using metadata
US20060090179A1 (en) * 2004-10-26 2006-04-27 Ya-Ling Hsu System and method for embedding supplemental information into a digital stream of a work of content

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"Track, n." entry. Oxford English Dictionary. Accessed 1/17/11 http://oed.com/view/Entry/204216 *

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9344754B2 (en) 2008-01-30 2016-05-17 Cinsay, Inc. Interactive product placement system and method therefor
US10055768B2 (en) 2008-01-30 2018-08-21 Cinsay, Inc. Interactive product placement system and method therefor
US10438249B2 (en) 2008-01-30 2019-10-08 Aibuy, Inc. Interactive product system and method therefor
US9332302B2 (en) 2008-01-30 2016-05-03 Cinsay, Inc. Interactive product placement system and method therefor
US9674584B2 (en) 2008-01-30 2017-06-06 Cinsay, Inc. Interactive product placement system and method therefor
US11227315B2 (en) 2008-01-30 2022-01-18 Aibuy, Inc. Interactive product placement system and method therefor
US10425698B2 (en) 2008-01-30 2019-09-24 Aibuy, Inc. Interactive product placement system and method therefor
US9351032B2 (en) 2008-01-30 2016-05-24 Cinsay, Inc. Interactive product placement system and method therefor
US9986305B2 (en) 2008-01-30 2018-05-29 Cinsay, Inc. Interactive product placement system and method therefor
US9338499B2 (en) 2008-01-30 2016-05-10 Cinsay, Inc. Interactive product placement system and method therefor
US9338500B2 (en) 2008-01-30 2016-05-10 Cinsay, Inc. Interactive product placement system and method therefor
US20120099588A1 (en) * 2009-08-24 2012-04-26 Xiangping Liu Channel change method, device, and system
WO2011153366A1 (en) * 2010-06-04 2011-12-08 Mobitv, Inc. Fragmented file structure for live media stream delivery
US20110302618A1 (en) * 2010-06-04 2011-12-08 Mobitv, Inc. Fragmented file structure for live media stream delivery
US9596522B2 (en) * 2010-06-04 2017-03-14 Mobitv, Inc. Fragmented file structure for live media stream delivery
US9223944B2 (en) * 2010-08-30 2015-12-29 Mobitv, Inc. Media rights management on multiple devices
US20150082463A1 (en) * 2010-08-30 2015-03-19 Mobitv, Inc. Media rights management on multiple devices
US8713420B2 (en) * 2011-06-30 2014-04-29 Cable Television Laboratories, Inc. Synchronization of web applications and media
US20130007576A1 (en) * 2011-06-30 2013-01-03 Cable Television Laboratories, Inc. Synchronization of web applications and media
US10334101B2 (en) * 2013-05-22 2019-06-25 Nice Ltd. System and method for generating metadata for a recorded session
US20140351256A1 (en) * 2013-05-22 2014-11-27 Nice-Systems Ltd. System and method for generating metadata for a recorded session
US20160094501A1 (en) * 2014-09-26 2016-03-31 Line Corporation Method, system and recording medium for providing video contents in social platform and file distribution system
US10944707B2 (en) * 2014-09-26 2021-03-09 Line Corporation Method, system and recording medium for providing video contents in social platform and file distribution system
US9794150B2 (en) 2014-11-24 2017-10-17 Roku, Inc. Apparatus and method for content playback utilizing crowd sourced statistics
WO2016085700A1 (en) * 2014-11-24 2016-06-02 Roku, Inc. Apparatus and method for content playback utilizing crowd sourced statistics
US10516591B2 (en) 2014-11-24 2019-12-24 Roku, Inc. Generating playback configurations based on aggregated crowd-sourced statistics
US11489748B2 (en) 2014-11-24 2022-11-01 Roku, Inc. Generating playback configurations based on aggregated crowd-sourced statistics
US20160269455A1 (en) * 2015-03-10 2016-09-15 Mobitv, Inc. Media seek mechanisms
US10440076B2 (en) * 2015-03-10 2019-10-08 Mobitv, Inc. Media seek mechanisms
US11405437B2 (en) 2015-03-10 2022-08-02 Tivo Corporation Media seek mechanisms
US11218741B2 (en) 2019-06-03 2022-01-04 Alibaba Group Holding Limited Method, apparatus, electronic device and computer-readable storage medium for video buffering

Also Published As

Publication number Publication date
US20090276402A1 (en) 2009-11-05
US20230319229A1 (en) 2023-10-05
US20230403368A9 (en) 2023-12-14
US20190230313A1 (en) 2019-07-25
US11917323B2 (en) 2024-02-27
US10250841B2 (en) 2019-04-02
US20120263438A1 (en) 2012-10-18

Similar Documents

Publication Publication Date Title
US20090274437A1 (en) Embedding data in media metadata tracks during playback
US20200162699A1 (en) Dynamic advertisement stream replacement
US10826958B2 (en) Content server media stream management
US9031078B2 (en) Modification of live streams
US10250659B2 (en) Contextually aware client buffer thresholds
US10263875B2 (en) Real-time processing capability based quality adaptation
RU2543568C2 (en) Smooth, stateless client media streaming
US9462302B2 (en) Efficient delineation and distribution of media segments
US20120011267A1 (en) Live streaming media delivery for mobile audiences
US20110302618A1 (en) Fragmented file structure for live media stream delivery
US20100064054A1 (en) Remote fast forward and rewind functionality for client devices
US20110299586A1 (en) Quality adjustment using a fragmented media stream
US20090259764A1 (en) Intro outro merger with bit rate variation support

Legal Events

Date Code Title Description
AS Assignment

Owner name: MOBITV, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:STIERS, TODD;KARLSSON, KENT;WINDISCH, ULF;AND OTHERS;REEL/FRAME:021632/0950;SIGNING DATES FROM 20080826 TO 20080901

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION