US20100046932A1 - Recording medium, playback apparatus and method, recording method, and computer-readable program - Google Patents

Recording medium, playback apparatus and method, recording method, and computer-readable program Download PDF

Info

Publication number
US20100046932A1
US20100046932A1 US12/607,765 US60776509A US2010046932A1 US 20100046932 A1 US20100046932 A1 US 20100046932A1 US 60776509 A US60776509 A US 60776509A US 2010046932 A1 US2010046932 A1 US 2010046932A1
Authority
US
United States
Prior art keywords
page
button
graphics
dsn
stream
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/607,765
Inventor
Joseph McCrossan
Tomoyuki Okada
Masayuki Kozuka
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Panasonic Corp
Original Assignee
Panasonic Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Panasonic Corp filed Critical Panasonic Corp
Priority to US12/607,765 priority Critical patent/US20100046932A1/en
Publication of US20100046932A1 publication Critical patent/US20100046932A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/30Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/102Programmed access in sequence to addressed parts of tracks of operating record carriers
    • G11B27/105Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/12Formatting, e.g. arrangement of data block or words on the record carriers
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/30Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
    • G11B27/3027Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording used signal is digitally coded
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/30Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
    • G11B27/309Table of contents
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/34Indicating arrangements 
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/20Disc-shaped record carriers
    • G11B2220/25Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
    • G11B2220/2537Optical discs
    • G11B2220/2541Blu-ray discs; Blue laser DVR discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/20Disc-shaped record carriers
    • G11B2220/25Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
    • G11B2220/2537Optical discs
    • G11B2220/2562DVDs [digital versatile discs]; Digital video discs; MMCDs; HDCDs

Definitions

  • the present invention relates to a field of user interface control. More particularly, the present invention relates to improvements in recording mediums used for consumer movie distribution, as well as in consumer playback apparatuses having a user interface for playback of the recording mediums.
  • a user interface using a menu structure is an essential function for receiving user operations, such as selection of languages and chapters for playback.
  • the interactive control that a DVD provides has the following disadvantage.
  • For each menu page to be displayed an access needs to be made to the DVD to read page information. Because of the access, the menu page can not be displayed without interruption to video playback. The interruption makes it difficult for the user to concentrate on, and enjoy the movie.
  • the present invention aims to provide a recording medium supporting presentation of multi-page menus which will not interfere with the video playback.
  • the present invention provides a recording medium having a video stream and a graphics stream recorded thereon.
  • the video stream represents a moving picture made up of a plurality of pictures.
  • the graphic stream is used for overlaying a multi-page menu on the moving picture, and includes interactive control information and graphics data.
  • the graphics data is used for generating the multi-page menu.
  • the interactive control information includes time information used for controlling behavior of the multi-page menu in accordance with a playback proceeding of the video stream.
  • the interactive control information multiplexed in the graphics stream includes control information.
  • the control information defines the multi-page menu behavior in accordance with the proceeding of the movie playback.
  • This structure makes it possible to automatically present a number of pages during the time the movie playback is showing scenery, for example.
  • the multi-page menu is presented without interfering with viewing the movie playback. Consequently, the user can enjoy the convenience provided by the multi-page menus, while watching the movie playback without obstructive display.
  • the information used for controlling the multi-page menu behavior is included in the graphics stream, and the graphics stream is multiplexed with the video stream.
  • This structure allows a playback apparatus to read the information needed for presentation of the multi-page menu, without interrupting the video stream reading.
  • pages of the multi-page menu are smoothly presented in synchronization with the video frame rate.
  • FIG. 1 is a view showing a usage pattern of a recording medium according to the present invention
  • FIG. 2 is a view showing an example structure of a BD-ROM
  • FIG. 3 is a view schematically showing the structure of an AV Clip
  • FIG. 4A is a view showing the structure of an interactive graphics stream
  • FIG. 4B is a view showing PES packets carrying functional segments
  • FIG. 5 is a view showing the logical structure defined by various types of functional segments
  • FIG. 6 is a view showing the time axis of AV clip playback to which a DS n is allocated;
  • FIGS. 7A and 7B are views showing the relations between interactive_composition structures and ICSs
  • FIG. 8 is a view showing the internal structure of an ICS
  • FIG. 9 is a view showing the processing performed on an ICS in the case where a stream_model field shows a “Multiplexed” type
  • FIG. 10 is a view showing the processing performed on an ICS in the case where the stream_model field shows a “Preload” type
  • FIG. 12A is a view showing the state transition of a Multi-page menu when a Pop-Up U/I is used;
  • FIG. 12B is a view showing the state transition of the Multi-page menu when an Always-On U/I is used
  • FIG. 13 is the display scenario incorporating the Multi-Page menu
  • FIGS. 14A-14C show a display example defined by the selection_time_out_pts
  • FIGS. 15A-15D show a display example defined by the user_time_out_duration
  • FIG. 16 is a view showing the internal structure of page information associated with an arbitrary one of multiple pages (x th page) of the Multi-Page Menu;
  • FIG. 17 is a view showing the state transition of buttons O-A, O-B, O-C, and O-D;
  • FIG. 18 is a view showing an example of the button_info description defining the button state transition shown in FIG. 17 ;
  • FIG. 19 is a view showing graphical images derived from ODSs 11 - 19 ;
  • FIG. 20 is a view showing an effect_sequence structure
  • FIG. 21A is a view schematically illustrating how an effect_duration is determined
  • FIG. 21B shows an equation for calculating the effect_duration
  • FIG. 22 is a view showing a specific example of an in-effect animation
  • FIG. 23 is a view showing an out-effect animation presented in accordance with the window_width, window_height, and effect_duration field values;
  • FIG. 24 is a view showing the internal structure of an arbitrary composition_object(i) in detail
  • FIG. 25 is a view showing a specific example of an in-effect presenting a Right-Scroll animation effect
  • FIG. 26 is a view showing an example of the composition_object ( 0 ) structure realizing the in-effect shown in FIG. 25 ;
  • FIG. 27 is a view showing an example of the composition_object( 1 ) structure realizing the in-effect shown in FIG. 25 ;
  • FIG. 28 is a view showing an example of the composition_object( 2 ) structure realizing the in-effect shown in FIG. 25 ;
  • FIG. 29 is a view showing an example of the composition_object( 3 ) structure realizing the in-effect shown in FIG. 25 ;
  • FIG. 30 is a view showing the order of ODSs within the DS n ;
  • FIG. 31 is a view showing how the order of ODSs within S-ODSs differs between the cases where the default_selected_button_id ref field is set to “00” and to the valid value specifying a button B;
  • FIG. 32A is a view showing a LASTODSPTS in the case where the default selected button is statically determined
  • FIG. 32B is a view showing a LASTODSPTS in the case where the default selected button is dynamically determined
  • FIG. 33A is a view showing a LASTODSPTS in the case where the DS includes no ODS associated with the selected state;
  • FIG. 33B is a view showing a LASTODSPTS in the case where the DS includes no ODS associated with the normal state;
  • FIG. 33C is a view showing a LASTODSPTS in the case where the DS includes no ODS at all;
  • FIG. 34A is a view showing an algorithm for obtaining the DECODEDURATION value from the value calculated by LASTODSPTS (DS n ) ⁇ DTS (DS n [ICS]) and the value of PLANECLERATIME;
  • FIGS. 34B and 34C are views showing examples of how to calculate the DECODEDURATION value
  • FIG. 35A is a view showing an equation for calculating the EFFECTTD
  • FIG. 35B is a view showing an equation for calculating the PAGDEFAULTTD in the case where no in-effect is provided and the default selected button is statically determined;
  • FIG. 35C is a view showing an equation for calculating the PAGENODEFAULTTD in the case where no in-effect is provided and the default selected button is dynamically determined;
  • FIG. 36 is a view showing an algorithm for selectively determining one of EFFECTTD, PAGEDEFAULTTD, and PAGENODEFAULTTD as TRANSFERDURATION;
  • FIG. 37 is a view showing the internal structure of a playback apparatus according to the present invention.
  • FIG. 38A is a flowchart of “procedure when playback condition is changed”, preformed with respect to a PSR 11 ;
  • FIG. 38B is a flowchart of “procedure when change is requested”, preformed with respect to the PSR 11 ;
  • FIG. 39A is a flowchart of “procedure when playback condition is changed”, performed with respect to a PSR 10 ;
  • FIG. 39B is a flowchart of “procedure when change is requested”, performed with respect to the PSR 10 ;
  • FIG. 40 is a view illustrating the internal structure of an I-Graphics decoder 9 ;
  • FIG. 41 is a flowchart of the main routine of the procedure performed by a graphics controller 37 ;
  • FIG. 42 is a flowchart of the procedure for the first presentation
  • FIG. 43 is a flowchart of the procedure for playback of an in-effect animation
  • FIG. 44 is a flowchart of the display update procedure for presentation buttons in animation
  • FIG. 45 is a flowchart of the procedure for executing a navigation command
  • FIG. 46 shows the procedure for processing a UO
  • FIG. 47 is a flowchart of the current button change procedure
  • FIG. 48 is a flowchart of the procedure for auto-activation of a button.
  • FIG. 49 is a flowchart of the manufacturing process of the BD-ROMs according to the first to fifth embodiments.
  • FIG. 1 shows a usage pattern of the recording medium according to the present invention.
  • a BD-ROM 100 is are cording medium according to the present invention.
  • the BD-ROM 100 is used for providing movies to a home theater system composed of a playback apparatus 200 , a remote controller 300 , and a television 400 .
  • FIG. 2 shows an example structure of the BD-ROM 100 .
  • the BD-ROM 100 is shown on the fourth level, and a BD-ROM's track is shown on the third level.
  • the track is stretched out into a straight line in the figure, although the track in practice spirals outwards from the center of the BD-ROM.
  • the track is composed of a lead-in area, a volume area, and a lead-out area.
  • the volume area has a layer model of a physical layer, a file system layer, and an application layer.
  • the first level shows, in a directory structure, a format of the application layer (application format) of the BD-ROM.
  • the first level shows the format of the application layer (application format).
  • the BD-ROM has a ROOT directory, and a BDMV directory under the ROOT directory, and a STREAM directory under the BDMV directory.
  • the STREAM directory contains an AV clip and Sub clips.
  • the AV clip and Sub clips are contained in files in the STREAM directory, such as “00001.m2ts”, “00002.m2ts”, and “00003.m2ts”.
  • Each file in the STREAM directory has a filename of a unified format, “xxxxx.m2ts” where x is an arbitrary integer.
  • FIG. 3 is a view schematically showing the structure of the AV Clip.
  • the middle level shows the AV clip.
  • This AV clip can be created as follows.
  • the video stream made up of a plurality of video frames (pictures pj 1 , pj 2 , pj 3 , . . . ) and the audio stream made up of a plurality of audio frames both shown on the upper first level are separately converted to PES packets shown on the upper second level, and further converted to TS packets shown on the upper third level.
  • the presentation graphics stream hereinafter, “PG stream”
  • the interactive graphics stream hereinafter, “IG stream” both shown on the lower first level are separately converted to PES packets shown on the lower second level, and further converted to TS packets shown on the lower third level.
  • These TS packets of the video, audio, PG and IG streams are multiplexed to form the AV clip.
  • the PG and IG streams are data in stream format used for compositing graphics with each picture of the video stream.
  • An entity of data carried by the PG and IG streams for overlaying on a picture is referred to as a “graphics object”.
  • the PG stream is a graphics stream used for presentation of subtitles using the graphics objects.
  • the IG stream is a graphics stream used for interactive presentation of the graphics objects. This concludes the description of the AV clip structure. Next, a description is given to the AV clip and the Sub clips.
  • the Sub clips are not multiplexed. Rather, the Sub clips are files exclusively containing the IG stream, audio stream, and text data.
  • the IG stream stored as a Sub clip is loaded to a memory in advance for synchronous playback with an AV clip. This operation of loading a Sub clip to a memory in advance is referred to as “preloading”.
  • FIG. 4A is a view showing the structure of the IG stream.
  • a string of TS packets constituting the AV clip is shown on the first level.
  • a string of PES packets constituting the graphics stream is shown on the second level.
  • the PES packets shown on the second level are formed by concatenating payloads of TS packets having a predetermined PID within the TS packet string shown on the first level. Note that no further description is given to the PG stream because it is not the gist of the present invention.
  • the graphics stream is composed of functional segments that include an ICS (Interactive Composition Segment), a PDS (Palette Definition Segment), an ODS (Object Definition Segment), and an END (End of Display Set Segment).
  • the ICS is a display composition segment
  • the PDS, ODS, and END are definition segments.
  • Each functional segment is either in one-to-one or one-to-multiple correspondence with PES packets. That is, one functional segment is recorded on the BD-ROM 100 after being converted to a single PES packet, or fragmented and converted to a plurality of PES packets.
  • the ICS is a functional segment used to define an interactive display composition of graphics.
  • One example of the interactive display composition is a Multi-Page menu.
  • the ODS is a run-length encoded graphics object composed of a plurality of pieces of run-length data.
  • the run-length data expresses a pixel string using a Pixel Code which shows a pixel value and a continuous length of the pixel value.
  • the Pixel Code is an 8-bit value showing one of the values from 1 to 255. Through the use of this Pixel Code, the run-length data sets arbitrary 256 pixel colors out of full color (16,777,216 colors).
  • the PDS is a functional segment storing pallet data.
  • the pallet data shows combinations of Pixel Codes of 1 to 255 and pixel values.
  • the pixel value used herein is made up of a red color difference component (Cr value), a blue color difference component (Cb value), a luminance component (Y value), and a transparency (T value). Substituting a Pixel Code of each piece of run-length data into a pixel value shown by the palette data produces a color.
  • the END provides an explicit indication that the transmission of functional segments is complete.
  • the END is inserted into the graphics stream as the last segment. This concludes the description of each functional segment.
  • FIG. 4B is a view showing PES packets carrying functional segments.
  • each PES packet is composed of a packet header and a payload.
  • the payload is an entity carrying a functional segment
  • the packet header carries a DTS and a PTS associated with the functional segment.
  • a DTS and a PTS in a packet header of a PES packet carrying a functional segment are simply described as a DTS and a PTS of that functional segment.
  • the first level shows Epochs
  • the second level shows DSs (Display Sets)
  • the third level shows types of the DSs.
  • the functional segments shown on the third level in FIG. 4A is shown on the fourth level in this figure.
  • An Epoch on the first level refers to one time unit of continuous memory management on a time axis of the AV clip playback, and also refers to a group of data allocated to that time unit.
  • Memory mentioned herein includes a graphics plane for storing one screen of graphics objects and an object buffer for storing uncompressed graphics data.
  • Continuous memory management means that neither the graphics plane nor the object buffer is flushed throughout the Epoch, and that clearing and rendering of graphics are performed only within a predetermined rectangular area on the graphics plane (to flush means to clear the entire graphics plane or the entire object buffer). The size and position of this rectangular area are fixed throughout the Epoch. As long as clearing and rendering of graphics are performed within this fixed rectangular area on the graphics plane, seamless playback is guaranteed.
  • the Epoch is a time unit of seamless playback on the time axis of the AV clip playback.
  • To change the graphics clearing/rendering area on the graphics plane it is necessary to define a point on the playback time axis at which the change is effected and to set a new Epoch from the point onward. In this case, the boundary between two Epochs is not seamless.
  • the seamless playback mentioned herein means that the clearing and rendering of graphics are each completed with a predetermined number of video frames.
  • the number of video frames is four to five.
  • the number of video frames is determined based on the ratio of the fixed rectangular area to the entire graphics plane and the transfer rate between the object buffer and the graphics plane.
  • a DS on the second level is a group of functional segments included in the graphics stream and provides one complete display composition.
  • dashed lines hk 1 show to which Epoch the DSs on the second level belong.
  • the DS 1 , DS 2 , DS 3 . . . , and DS n are shown to belong to the first Epoch on the first level.
  • the third level shows the types of DSs.
  • the first DS in an Epoch is an “Epoch Start” DS.
  • Each DS other then the first one in the Epoch is either an “Acquisition Point” DS, a “Normal Case” DS, or an “Epoch Continue” DS.
  • the DSs are shown in the order of Acquisition Point, Normal Case, and Epoch Continue, this order is merely an example and the DSs may be arranged in a different order.
  • the Epoch Start DS provides a display effect “new display”, and indicates a start of a new Epoch.
  • the Epoch Start DS contains all functional segments necessary for the next screen composition.
  • the Epoch Start DS is provided at a point to which a skip operation is likely to be made, such as a start of a chapter in a movie.
  • the Acquisition Point DS provides a display effect “display refresh”, and is identical in contents to the preceding Epoch Start DS.
  • the Acquisition Point DS is not the start of the Epoch, but contains all functional segments necessary for the next screen composition. Therefore, graphics can be displayed reliably when playback is started from the Acquisition Point DS. That is to say, the Acquisition Point DS enables a display composition to be made from a midpoint in the Epoch.
  • the Acquisition Point DS is provided at a point to which a skip operation may be made, such as a point that may be designated by a time search.
  • the time search is an operation of locating a playback point corresponding to a time input by a user in minutes/seconds.
  • the time input is made in a relatively large unit such as ten minutes and ten seconds.
  • the Acquisition Point DS is provided at a point that can be designated by a time search made in units of 10 minutes and 10 seconds.
  • the Normal Case DS provides a display effect “display update”, and contains only a difference from the immediately preceding display composition. For example, if a DS v is identical in contents to an immediately preceding DS u but different in display composition, the DS v is a Normal Case DS which contains only an ICS or an ODS. This eliminates the need to provide overlapping ODSs in DSs, so that the amount of data stored on the BD-ROM 100 is reduced. Since the Normal Case DS contains the difference without the overlapping data, no display composition can be displayed with the Normal Case DS alone.
  • the Epoch Continue DS indicates at the start of an AV clip that one Epoch continues from a different AV clip.
  • a DS n contains a composition state, field having a value indicating “Epoch Continue”.
  • the DS n belongs to the same Epoch as an immediately preceding DS n ⁇ 1 even if the two DSs are on different AV clips. Since the DS n and DS n ⁇ 1 belong to the same Epoch, there will be no flushing of the graphics plane and of the object buffer even if there is a boundary of AV clips between the two DSs.
  • dashed lines kz 1 show to which DS the functional segments on the fourth level belong. Since the functional segments shown in FIGS. 4A and 5 are the same, it is known that the functional segments shown in FIG. 4A belong to the Epoch Start DS. Functional segments belonging to the Acquisition Point DS are identical to those belonging to the Epoch Start DS. In addition, functional segments belonging to the Normal Case DS are identical to those belonging to the Epoch Start DS, except that some of the functional segments are omitted in the Normal Case DS.
  • the Epoch is a time unit of continuous memory management and includes one or more DSs, it is important how those DSs are allocated to the time axis of the AV clip playback.
  • the “time axis of the AV clip playback” refers to the time axis used for defining the decode timing and playback timing of individual pictures multiplexed in the AV clip. The decode timing and playback timing are expressed in the time accuracy of 90 KHz.
  • the DTS and PTS of an ICS or of an ODS within a DS show the timing based on which the synchronous control is realized.
  • DSs are allocated to the time axis of the AV clip playback by carrying out the synchronous control in accordance with each DTS and PTS attached to an ICS and to an ODS.
  • DS n an arbitrary one of DSs within an Epoch is denoted as a DS n , and that the DS n is allocated to the time axis of the AV clip playback based on the DTS and PTS set as shown in FIG. 6 .
  • FIG. 6 shows the time axis of AV clip playback to which the DS n is allocated.
  • the duration of DS n starts at the time shown by the value of DTS of the ICS belonging to the DS n (DTS (DS n [ICS])), and ends at the time shown by the value of PTS of the END belonging to the DS n (PTS(DS n [ICS])).
  • the display timing for the first presentation of the DS n is shown by the value of PTS of the ICS(PTS(DS n [ICS])).
  • the PTS(DS n [ICS]) value is the sum of the time taken for decoding the ODS (DECODEDURATION) and the time taken for transferring the decoded graphics object (TRANSFERDURATION).
  • the decoding of ODS needed for the first presentation is completed within the DECODEDURATION.
  • the duration mc 1 shown in FIG. 6 is the time required for decoding an ODS m , an arbitrary one of ODSs within the DS n .
  • the start of the decode duration mc 1 is shown by the value of DTS(ODS n [ODS m ]), and the end of the decode duration mc 1 is shown by the value of PTS (ODS n [ODS m ]).
  • An Epoch is defined by allocating every ODS within the Epoch to the time axis AV playback as described above. This concludes the description of the allocation to the time axis of AV playback.
  • One feature of the first embodiment is to control the behavior of a Multi-page menu in accordance with the proceeding of AV playback.
  • the novel structure realizing the feature resides in an interactive_composition structure provided in the ICS.
  • an interactive_composition structure provided in the ICS.
  • a description is given to the interactive_composition structure.
  • FIGS. 7A and 7B are views showing the relations between the interactive_composition structures and the ICSs.
  • the interactive_composition structure and the ICS may be in one-to-one correspondence as shown in FIG. 7A , or one-to-multiple correspondence as shown in FIG. 7B .
  • the one-to-one correspondence is applicable to the case where the interactive_composition structure is small enough to be contained in one ICS.
  • the one-to-multiple correspondence is applicable to the case where the interactive_composition structure is too large in size to be carried by one ICS and thus fragmented to be stored in multiple ICSs. Since the interactive_composition structure can be fragmented to be stored in multiple ICSs, there is no restriction imposed on the size of interactive_composition structure.
  • the interactive_composition structure may be as large as 512 Kbytes, 1 Mbyte, or any number of bytes. As above, one interactive_composition structure may correspond to multiple ICSs. For the simplicity sake, however, a description hereinafter is given on precondition that the ICS and interactive_composition structure is in one-to-one correspondence.
  • FIG. 8 is a view showing the internal structure of an ICS.
  • the ICS contains one interactive_composition structure or a fragment of one interactive_composition structure.
  • the ICS is composed of the following fields: “segment descriptor” showing the segment type “ICS”; “video_descriptor” showing the width and height of the video plane in pixels and the frame rate of the associated video stream; “composition_descriptor” showing the composition_state, which is information showing the type of DS to which the ICS belongs is whether a Normal Case DS, an Acquisition Point DS, an Epoch Start DS, or an Effect_Sequence; and “interactive_composition data_fragment” showing that whether the ICS is carrying an entire interactive_composition structure or a fragment of interactive_composition structure.
  • FIG. 8 shows the internal structure of an interactive_composition structure in detail.
  • the interactive_composition structure is composed of the following fields: “interactive_composition_length”, “stream_model”, “user_interface_model”, “composition_time_out_pts”, “selection_time_out_pts”, “user_time_out_duration”, “page information ( 1 ), ( 2 ) . . . (i) . . . (number_of_page- 1 )” for each page of the Multi-Page menu available for presentation.
  • the interactive_composition_length field shows the length of the interactive_composition structure.
  • the “stream_model” field shows the type of stream model of the associated interactive_composition structure.
  • the stream model shows how the interactive_composition structure is stored on the BD-ROM and how the data is to be processed by a buffer (referred to as a “composition buffer”) on a playback apparatus. Specifically, the stream model shows whether (i) the graphics stream is multiplexed with an AV clip and separated to be loaded to the composition buffer or (ii) a graphics stream is stored as a Sub clip in a separate file from the AV clip and preloaded to the composition buffer.
  • the stream_model field is provided in the interactive_composition structure because the composition buffer needs to perform different processing depending on whether the interactive_composition structure is preloaded or multiplexed.
  • FIG. 9 is a view showing the processing performed on the “Multiplexed” ICS.
  • the video stream multiplexed in the AV clip is shown on the fourth level
  • the graphic stream multiplexed in the AV clip is shown on the third level.
  • the graphics stream includes three DSs, which are a DS 1 , a DS 2 , and DS 8 .
  • the second level shows the proceeding of AV playback.
  • the first level shows contents stored in the composition buffer at different times of the AV playback.
  • the composition buffer stores DS 1 [ICS].
  • the composition buffer stores DS 2 [ICS]. That is, in the case where the stream_model is “Multiplexed”, there are more than one interactive_composition structures.
  • the playback apparatus discards a currently stored interactive_composition structure and newly loads another interactive_composition structure residing at the current playback point.
  • the composition buffer stores one interactive_composition structure at a time.
  • This discarding and loading operation is especially effective at the time of a skip operation.
  • the playback apparatus discards the interactive_composition structure currently on the composition buffer.
  • the composition buffer discards the stored contents at the execution timing am 2 of the skip operation.
  • the DS 8 [ICS] residing on the skip destination is loaded to the composition buffer as indicated by an arrow pf 3 .
  • FIG. 10 is a view showing the contents of the composition buffer.
  • FIG. 10 is similar to FIG. 9 with respect to what is shown in each of the first to fourth level. Yet, FIG. 10 is totally different from FIG.
  • the interactive_composition structure is provided with the stream_model field for allowing a controlling subject (the graphics controller) to accurately judge whether to discard the stored interactive_composition structure as the proceeding of AV playback or to retain the stored interactive_composition structure irrespective of the proceeding of AV playback. This concludes the description of the stream_model field.
  • the “user_interface_model” shows the type of user interface model to be used for the interactive_composition structure.
  • the user_interface_model shows either “Always-On U/I” or “Pop-Up U/I”.
  • the Always-On U/I is used, a menu appears on screen and disappears in accordance with the proceeding of AV clip playback.
  • the Pop-Up U/I is used, a menu appears on screen and disappears in response to a user operation.
  • the composition_time_out_pts field shows the end time of an Epoch (Epoch END) to which the ICS belongs.
  • the interactive control defined by the ICS is no longer valid at the Epoch END. That is, a point of time shown by the composition_time_out_pts field is the end of valid interactive control period.
  • the “selection_time_out_pts” field shows the button selection termination time. At the time shown by the selection_time_out_pts value, the button that is currently ins the selected is automatically activated. Note that buttons are elements representing options in the multi-page menu.
  • the selection_time out_pts field defines the valid interaction period for user interaction to activate the buttons.
  • the user_time_out_duration field shows the timeout at which on-screen pages presented in response to a user operation are removed.
  • pages from the second page onward referred to as sub pages
  • all subpages are removed from the screen leaving the first page on-screen.
  • the Pop-Up I/O not only the sub pages but also every page of the Multi-page menu is presented in response to user operations.
  • all on-screen pages are removed leaving no page remain presented (No Menu Display).
  • the timing chart shows that the composition_time_out_pts value specifies the end of Epoch (Epoch END).
  • the interactive display is valid for a period from the PTS(DS n [ICS]) to the Epoch END. This period is referred to as the “Interactive Display Period”.
  • the selection_time_out_pts specifies a time preceding the Epoch End. User interaction may be made during a period from the PTS(DS n [ICS]) to the time specified by the selection_time_out_pts value. This period is referred to as the “valid interaction period”.
  • the selection_time_out_pts field defines the end time of a period which starts from the first presentation of the interactive display and during which the interactive display is available for user interaction.
  • the composition_time_out_pts field defines a period starting from the first presentation of the interactive display during which the interactive display in an Epoch remains valid and thus presented. This concludes the description of the selection_time out_pts and composition_time_out_pts fields. Next, a description is given to the Multi-page menu state transition defined by the selection_time_out_pts, composition_time_out_pts, and user_time_out_duration fields.
  • FIG. 12A is a view showing the state transition of the Multi-page menu based on the selection_time_out_pts, composition_time_out_pts, and user_time_out_duration fields.
  • FIG. 12A is a view showing the state transition of the Multi-Page menu when the Pop-Up U/I is used.
  • the Multi-page menu shown in the figure has the following three states: No Menu Display, 1 st Page Only, Multi-Page Exist.
  • the 1 st page only is the state where only the first page of the Multi-Page menu is presented.
  • the Multi-Page exist is the state where pages from the second page onward (i.e. sub pages) are presented in addition to the first page.
  • An arrow jt 1 represents the transition from the No Menu Display state to the 1 st Page Only state. This state transition is triggered by an event “Pop-Up_On”.
  • the “Pop-Up_On” is an event generated within the playback apparatus in response to a user operation for a menu call.
  • An arrow jt 2 represents the transition from the 1 st Page Only state to the No Menu Display state. This state transition is triggered by an event “Pop-Up_Off”.
  • the “Pop-Up_Off” is an event generated within the playback apparatus in response to a user operation for a menu removal.
  • Arrows jt 3 and jt 4 represent the transition from the 1 st Page Only state to the Multi-Page Exist state, and then from the Multi-Page Exist state to the 1 st Page Only state.
  • the state transition is triggered by an event “ChangeButtonPage” generated upon the timeout defined by the selection_time_out_pts.
  • the ChangeButtonPage is an event generated within the playback apparatus in response to a user operation for switching a menu page. The event is generated on precondition that a user operation for switching a menu page is made.
  • the selection_time_out_pts shows the time for forcefully invalidating a button for receiving a user operation for switching a menu page.
  • the description of the selection_time_out_pts field realizes the state transition of “1 st Page Only ⁇ Multi Page Exist ⁇ 1 st Page Only”.
  • An arrow jt 5 represents the transition from the 1 st Page Only state to the No Menu Display state.
  • An arrow jt 6 represents the transition from the Multi-Page Exist state to the No Menu Display state.
  • the state transition jt 6 is triggered by the timeout of the user_time out_duration.
  • the state transitions to the No Menu Display state from the 1 st Page Only state (jt 5 ) and from the Multi-Page Exist state are defined by the user_time_out_duration field. That is to say, it is possible to define such control that on-screen menu pages are automatically removed when the AV playback reaches to a climax scene. With this control, it is avoided that the menu pages interferes with viewing the video stream playback.
  • FIG. 12B shows the state transition of Multi-page menu when the Always-On U/I is used.
  • An arrow at 1 represents the transition from the No Menu Display state to the 1 st Page Only state. This state transition is triggered by an event “Epoch Start”.
  • the “Epoch Start” is an event showing that the current playback point reaches the point corresponding to the PTS associated with the interactive_composition structure. It is shown that the 1 st page is automatically presented as the video stream playback proceeds and without waiting for any user operation.
  • An arrow at 2 shows the transition from the 1 st Page Only state to the Multi-Page Exist state
  • an arrow at 3 shows the transition from the Multi-Page Exist state to the 1 st Page Only state.
  • An arrow at 4 represents the transition from the Multi-Page Exist state to the 1 st Page Only state. This state transition is triggered by the timeout of the user_time_out_duration. As apparent also from this state transition, when the Always-On U/I is used, the state transition is made to the 1 st Page Only state not to the No Menu Display state upon the time but of the user_time out_duration.
  • An arrow at 5 represents the transition from the 1 st Page Only state to the No Menu Display state. This state transition is triggered by the time out of the composition_time_out_pts. Note that the composition_time_out_pts shows the end of Epoch (EpochEND).
  • the Multi-Page menu behavior is defined by the selection_time_out_pts, user_time_out_duration, and composition_time_out_pts fields provided in the interactive_composition structure.
  • FIGS. 13-15 show the screen display presented under control defined as above.
  • FIG. 13 shows the display scenario incorporating the Multi-Page menu.
  • the Multi-Page menu is not presented on screen at first.
  • the 1 st page of the Multi-Page menu is presented when the current playback point reaches the point t 1 , and a Sub page is additionally presented at the point t 2 .
  • the on-screen Sub page is removed leaving only the 1 st page on screen.
  • the 1 st page is then removed at the point t 4 leaving no menu page displayed on screen.
  • the PTS(DS n [ICS]) is set to the value corresponding to the point t 1 , the selection_time_out_pts to the value corresponding to the point t 2 , the user_time_out_duration to the value corresponding to the point t 3 , and the composition_time_out_pts to the value corresponding to the point t 4 .
  • the Multi-Page menu undergoes the state transitions shown in FIG. 13 .
  • FIGS. 14A-14C show a display example defined by the selection_time_out_pts.
  • the first page of Multi-Page Menu is overlaid with a picture of the video stream.
  • a picture presented at the point t 2 is of a scenic shot and is of less importance in the story of the AV clip. Since the picture presents no valuable contents for users, the selection_time_out_pts value is set to automatically activate a button presented on the 1 st Page at the point t 2 . Consequently, a Sub Page is overlaid with the scenic short shown in FIG. 14B , so that a composite image as shown in FIG. 14C appears on screen.
  • FIGS. 15A-15D show a display example defined by the user_time_out_duration.
  • the Multi-Page menu is in the Multi-Page Exist state_where multiple Sub Pages (2 nd , 3 rd , and 4 th pages) are presented on screen as shown in FIG. 15A .
  • the pages of the Multi-Page menu are overlaid with a picture included in the video stream.
  • the picture to be presented at the point t 3 is a character image as shown in FIG. 15B .
  • the multiple Sub Pages overlaid with the picture end up covering most of the character image as shown in FIG. 15C .
  • the facial expression of the character image can not be seen.
  • the user_time_out_duration is set to the value corresponding to a time immediately before the point t 3 .
  • the Multi-Page menu undergoes the state transition from the Multi-Page Exist state to the 1 st Page Only state.
  • the composite image as shown in FIG. 14D is presented at the point t 3 .
  • the 1 st page is solely presented on screen without blocking the character image too much.
  • the 1 st page remains on screen rather than removing every page of the Multi-Page Menu, there is no need to call the menu again.
  • the ICS multiplexed in the graphics stream includes control information defining behavior of a Multi-Page Menu in accordance with the proceeding of AV playback.
  • control information defining behavior of a Multi-Page Menu in accordance with the proceeding of AV playback.
  • a second embodiment relates to display compositions for presenting pages of a Multi-Page Menu.
  • a display composition is designed to present a plurality of graphical elements called button on screen.
  • an animation effect is presented on screen before introducing and/or after removing a menu page.
  • FIG. 16 is a view showing the internal structure of page information associated with an arbitrary one of multiple pages (x th page) of the Multi-Page Menu.
  • page information (x) is composed of the following fields: “page_id” showing an identifier uniquely identifying the page(x); “UO_mask_table”; “in_effects” specifying an animation effect to be presented as introduction of the page x to the screen; “out_effects” specifying an animation effect to be presented after removal of the page(x) from the screen; “animation_frame_rate_code” showing an animation frame rate to be followed at the time of presenting animation effects for the page(x); “default_selected_button_id_ref”, “default_activated_button_id_ref”; “pallet_id_ref”; and “button_info ( 1 ), ( 2 ) . . . (number_of_buttons- 1 )” each associated with one of the buttons presented on the
  • the UO_mask_table field shows permissions and prohibitions of user operations on the page (x). If this field is set to a value showing “prohibited”, a corresponding user operation to the playback apparatus is invalid while the page (x) is active.
  • the default_selected_button_id_ref field shows whether a button to be selected as default upon starting the presentation of page (x) is determined dynamically or statically.
  • this field is set to “0xFF”
  • the default selected button is determined dynamically.
  • the value held in a Player Status Register (PSR) of the playback apparatus is interpreted with priority, and the button shown by the PSR is put into the selected state.
  • PSR Player Status Register
  • this field is set to the value other than “0xFF”
  • the default selected button is determined statically. In such a case, the PRS is overwritten with the default_selected_button_id_ref field value, and the button specified by this field is put into the selected state.
  • the default_activated_button_id_ref field shows a button to be automatically activated at the time defined by the selection_time_out_pts field value.
  • the default_activated_button_id_ref field value is set to “FF”
  • the button that is in the selected state at the time of the timeout is automatically activated.
  • this field is set to “00”
  • no button is automatically activated.
  • this field is set to the value that is neither “00” or “FF”
  • the value is interpreted as a valid button number statically specifying the button to be automatically activated.
  • the pallet_id_ref field shows the ID associated with a pallet to be set in a CLUT unit, which will be described later.
  • Each button_info structure is information defining a button to be presented on the page (x). With the above fields, each page of the Multi-Page menu is defined. Next, a description is given to the internal structure of button_info structure. In the description, an arbitrary one of the buttons on the page (x) is denoted as a button(i). In FIG. 16 , arrows cx 1 indicate that the internal structure of button_info (i) is excerpted to be shown in detail.
  • Each button presented on the page has three states which are a normal state, a selected state, and an activated state.
  • the button When a button is in the normal state, the button is simply displayed.
  • the button When a button is in the selected stated, the button is currently focused as a result of a user operation but not yet activated.
  • the button When a button is in the activated state, the button has been activated. Since each button has these three states, the button_info(i) is provided with the following information.
  • the button_id field shows a value uniquely identifying the button(i) in the interactive_composition structure.
  • the button_numeric_select_value field shows a flag indicating whether the button(i) is numerically selectable.
  • the auto_action_flag field shows whether the button(i) is automatically activated.
  • the auto_action_flag field is set to ON (bit value of “1”), the button(i) is transferred not to the selected state but directly to the activated state when selected.
  • the auto_action_flag field is set to OFF (bit value of “0”), the button(i) is transferred not to the activated state but to the selected state when selected.
  • the button_horizontal_position field and button_vertical_position field respectively specify the horizontal and vertical positions of the top left pixel of the button(i) on the interactive display.
  • the neighbor_info structure shows buttons to receive the selected state in response to user operations made to move a focus in an up, down, right, and left direction when the button (i) is in the selected state.
  • the neighbor_info structure is composed of upper_button_id_ref, lower_button_id_ref, left_button_id_ref, and right_button_id_ref fields.
  • the upper_button_id_ref field specifies the button number of a button to receive the selected state if a user operation instructing to move the focus in the up direction is made while the button (i) is in the selected state.
  • the move up operation is executed at a push of a Move Up key of the remote controller. If this field is set to the same button number as that of the button(i), the user operation to the Move Up key is ignored.
  • the lower_button_id_ref field, the left_button_id_ref field, the right_button_id_ref field each specify the button numbers of a button to receive the selected state if a user operation instructing to move the focus to the down, left, or right direction is made, while the button(i) is in the selected state.
  • the respective operations are executed at a push of the Move Down key, Move Left key, and Move Right key of the remote controller. If these fields are set to the same button number as that of the button (i), the user operations to the those keys are ignored.
  • the normal_state_info structure defines the normal state of the button(i), and is composed of the normal_start_object_id_ref, normal_end_object_id_ref, and normal_repeat_flag fields.
  • the normal_start_object_id_ref field specifies the first one of the object_id values that are serially assigned to a sequence of ODSs used to present the normal state of button (i) in animation.
  • the normal_end_object_id_ref field specifies the last one of the object_id values that are serially assigned to the sequence of ODSs used to present the normal state of button(i) in animation. If the normal_end_object_id_ref field specifies the same ID value as that of the normal_start_object_id_ref, the static image of a graphics object identified by that ID is presented as the button(i).
  • the normal_repeat_flag field specifies whether the animation of the button(i) in the normal state is to be continuously repeated.
  • the selected_state_info structure defines the selected state of the button(i).
  • the selected_state_info structure is composed of the selected_state_sound_id_ref, selected_start_object_id_ref, selected_end_object_id_ref, and selected_repeat_flag fields.
  • the selected_state_sound_id_ref field specifies sound data to be reproduced as a click sound when the button(i) is put to the selected state.
  • the sound data is specified by the sound_id of a piece of sound data contained in the file called sound.bdmv.
  • this field is set to “0xFF”, no sound data is associated with the selected state of the button (i) and thus no click sound is reproduced when the button(i) is changed to the selected state.
  • the selected_start_object_id_ref field specifies the first one of the object_id values that are serially assigned to a sequence of ODSs used to present the selected state of button(i) in animation.
  • the selcted_end_object_id_ref field specifies the last one of the object_id values that are serially as signed to the sequence of ODSs used to present the normal state of button(i) in animation.
  • the selected_end_object_id_ref field specifies the same ID value as that of the selected_start_object_id_ref, the static image of a graphics object identified by the ID is presented as the button(i).
  • the selected_repeat flag field specifies whether the animation of the button(i) in the selected state is to be continuously repeated. If the selcted_start_object_id_ref and selcted_end_object_id_ref fields have the same value, the selected_repeat_flag field is set to the value of “00”.
  • the activated_state_info structure defines the activated state of the button(i), and is composed of the activated_state_sound_id_ref, activated_start_object_id_ref, and activated_end_object_id_ref fields.
  • the activated_state_sound_id_ref field specifies sound data to be reproduced as a click sound when the button(i) is put to the activated state.
  • the sound data is specified by the sound_id of a piece of sound data contained in the sound.bdmv file. When this field is set to “0xFF”, no sound data is associated with the activated state of the button(i) and thus no click sound is reproduced when the button(i) is changed to the activated state.
  • the activated_start_object_id_ref field specifies the first one of the object_id values that are serially assigned to a sequence of ODSs used to present the activated state of button(i) in animation.
  • the activated end_object_id_ref field specifies the last one of the object_id values that are serially assigned to the sequence of ODSs used to present the activated state of button (i) in animation.
  • the “navigation_command” structure shows a navigation command to be executed when the button(i) is activated.
  • a representative example of the navigation command is, a SetButtonPage command.
  • the SetButtonPage command instructs the playback apparatus to display a desired page of the Multi-Page menu with a desired one of buttons on the page in the selected state.
  • buttons O-A to O-D are associated with the button O-A.
  • arrows hh 1 and hh 2 represent the state transitions defined by the neighbor_info ( ) structure of the button info( 1 ), which is associated with the button O-A.
  • the lower_button_id_ref field is set to the value specifying the button O-C.
  • the right_button_id_ref field in the button info( 1 ) is set to the value specifying the button O-B.
  • An arrow hh 3 in FIG. 17 represents the button state transition defined by the neighbor_info( ) structure in the button_info( 2 ), which is associated with the button O-C.
  • the upper_button_id_ref field is set to the value specifying the button O-A.
  • a DS n provided with the ICS shown in FIG. 18 includes ODSs 11 - 19 corresponding to the graphical images shown in FIG. 19 .
  • the normal_start_object_id_ref and normal_end_object_id_ref fields in the button_info( 0 ) specify the ODSs 11 and 13 , respectively.
  • the normal state of the button O-A is presented in animation using the sequence of the ODSs 11 - 13 .
  • the selected_start_object_id_ref and selected_end_object_id_ref fields in the button_info( 0 ) specify the ODSs 14 and 16 , respectively.
  • the selected state of the button O-A is presented in animation using the sequence of the ODSs 14 - 16 .
  • the graphical image serving as the button O-A is changed from the one presented using the ODSs 11 - 13 to the one presented using the ODSs 14 - 16 .
  • the normal_repeat_flag field in the normal_state_info( ) structure is set to the value “1”
  • the animated presentation of the ODSs 11 - 13 is repeated as indicated by “ ⁇ (A)” and “ ⁇ (A)” in the figure.
  • the selected_repeat_flag field in the selected_state_info( ) structure is set to the value “1”
  • the animated presentation of the ODSs 14 - 16 is repeated as indicated by “ ⁇ (B)” and “ ⁇ (B)” in the figure.
  • buttons O-A, O-B, O-C, and O-D are associated with the buttons O-A, O-B, O-C, and O-D.
  • a character image serving as a button changes its facial expression in response to user operations.
  • FIG. 16 is a view showing an effect_sequence structure.
  • the effect_sequence structure is composed of: window_info( 0 ), ( 1 ), ( 2 ) . . . (number_of_windows- 1 ); and the effect_info( 0 ), ( 1 ), ( 2 ) . . . (number_of_effects- 1 ).
  • the animation effect is presented by updating display compositions at fixed intervals.
  • Each effect_info structure is a piece of information defining one display composition.
  • the effect_info structure is composed of: the effect_duration field showing the above-mentioned fixed interval, i.e. a time period for which the associated display composition is to be presented; the pallet_id_ref field specifying a pallet used for the associated display composition; and the composition_object ( 0 ), ( 1 ), ( 2 ) . . . (number_of_composition_object- 1 ) specifying the details of the display composition.
  • Each window_info structure defines a window or an area on the graphics plane within which the display compositions are rendered. As indicated by arrows wc 1 , the window_info is excerpted to be shown in detail. As shown in the figure, the window_info is composed of the following fields: “window_id” uniquely identifying a widow on the graphics plane; “window_horizontal_position” specifying the horizontal position of the top left pixel of the window; “window_vertical_position” specifying the vertical position of the top left pixel of the window; “window_width” specifying the width of the window; and “window_height” specifying the height of the window.
  • window_horizontal_position a description is given to values that may be set in each of the window_horizontal_position, window_vertical_position, window_width, and window_height fields. These fields are set to the value corresponding to the coordinates within the two-dimensional graphics plane having the height and width defined by the video_height and video_width fields.
  • window_horizontal_position field shows the horizontal address of the pixel on the graphics plane
  • this field takes on a value from “1” to the video_width value.
  • window_vertical_position shows the vertical address of the pixel on the graphics plane
  • this field takes on a value from “1” to the video_height value.
  • window_width field shows the width of a window on the graphics plane
  • this field takes on a value from “1” to the value calculated by subtracting the window_horizontal_position value from the video_width value.
  • window_height field shows the height of window on the graphics plane
  • this field takes on a value from “1” to the value calculated by subtracting the window_vertical_position value from the video height value.
  • window_info is provided with window_horizontal_position, window_vertical_position, window_width, window_height fields defining the size and postion of a window on the graphics plane.
  • Rc denote the transfer rate from the object buffer to the graphics plane.
  • Rc the transfer rate from the object buffer to the graphics plane.
  • the clearing and rendering of the window need to be performed at Rc.
  • the clearing and rendering need to be performed at half of Rc (Rc/2).
  • the total capacity of the graphics plane is 2 Mbytes ( ⁇ 1920 ⁇ 1080 ⁇ 8).
  • the effect_duration showing the rate of clearing and rendering a window is 1 ⁇ 2 or 1 ⁇ 4 of the video frame rate.
  • the window size can be as large as 32% or 64% of the entire graphics plane even if Rc is the same. That is to say, the effect_sequence structure has a feature that the effect_duration is determined to be longer for a larger window size, and shorter for a smaller window size. This concludes the description of the window size.
  • a description is given to the position of window. As mentioned above, the position and the size of a window are fixed throughout an Epoch.
  • the position and the size of a window are fixed throughout the duration of an Epoch for the following reason.
  • the write address of the graphics plane needs to be changed, which incurs overhead.
  • the overhead inevitably reduces the transfer rate from the object buffer to the graphics plane.
  • the number of graphics objects is restricted in order to reduce overhead incurred at the time of transferring decoded graphics objects.
  • the overhead referred herein is incurred at the time of setting addresses of edge parts of the graphics objects. This overhead increases if the number of edge parts is greater.
  • the transfer rate Rc can be set on an assumption that the number of instances of overhead is at most four. Hence a minimum standard for the transfer rate Rc can be determined easily.
  • effect_duration is determined depending on the window size. Now, a description is given to how the effect_duration is determined.
  • FIG. 21A is a view schematically illustrating how the effect_duration is determined in the figure, the (j+1) th display composition is presented after a plurality of video frames from the presentation of the j th display composition
  • the number of the plurality of video frames is the effect_duration.
  • the total size of data to be rendered for presentation of each display composition must be known.
  • the effect_sequence is provided with windows of the number that is indicated by the number_of_windows field value.
  • the total size of data needed to be rendered for presentation of the j th display composition is equal to the total size of the windows ( 0 ) through (number_of_windows- 1 ), as indicated by the brace br.
  • the j th display composition is provided with the windows ( 0 ) through (number_of_windows- 1 ).
  • the total data size of the windows is divided by the transfer rate of the playback apparatus (128 Mbps) and further multiplied by the time accuracy of PTS (90 KHz), so that the result is expressed in 90 KHz.
  • the resulting value is the duration for which the j th display composition of the effect_sequence is presented.
  • FIG. 21B shows the equation for calculating the effect_duration.
  • FIG. 22 is a view showing a specific example of an in-effect animation.
  • the in-effect animation presents an image such that a flash of light appears on the palm of hand of the character image on screen and a menu gradually rises up after the flash of light disappears.
  • the size of graphics data is the largest within the in-effect at the time when the menu is fully shown on screen (effect ( 4 )).
  • the window_width and window_height of a window are determined based on the data size of the effect ( 4 ).
  • the window size determined herein is about 50% of the entire graphics plane. This window size is nearly three times as large as the above-mentioned window size of 16% of the graphics plane.
  • FIG. 23 is a view showing the out-effect animation for which window_width, window_height, and effect_duration field values are determined in the manner described above.
  • the window_info defines an area in the graphics plane for rendering
  • the effect_info defines a duration at intervals of which the rendering area is updated.
  • the duration may be adjusted to be longer when the rendering area is set to be larger, and shorter when the rendering area is set to be smaller.
  • a third embodiment of the present invention relates to an improvement on the presentation of animation effects, such as Scroll, Wipe-In/Out, and Cut-In/Out.
  • the structure for achieving the animation effects lies in the composition_object structures shown in FIG. 20 .
  • FIG. 24 is a view showing the internal structure of an arbitrary composition_object(i) in detail.
  • the composition_object(i) is composed of the following fields: “object_id_ref”; “window_id_ref”, “object_cropped_flag”; “composition_object_horizontal_position”; “composition_object vertical_position”, and “cropping_rectangle_info( 1 ), ( 2 ) . . . (n)”.
  • the object_id_ref field shows a reference value for a graphics object identifier (object_id).
  • the reference value corresponds to the object_id identifying the graphics object that is used for generating a display composition according to the composition_object(i).
  • the window_id_ref field shows a reference value for a window identifier (window_id).
  • the reference value corresponds to the window_id identifying the window in which the display composition according to the composition_object(i) is rendered.
  • the object_cropped flag field shows a flag indicating whether the graphics object cropped on the object buffer is to be presented or not to be presented. When this field is set to “1”, the graphics object cropped on the object buffer is presented. On the other hand, when this field is set to “0”, the cropped graphics object is not presented.
  • composition_object_horizontal_position field shows the horizontal postion of the top left pixel of the graphics object on the graphics plane.
  • composition_object_vertical_position field shows the vertical position of the top left pixel of the graphics object on the graphics plane.
  • the cropping_rectangle structure is valid when the object_cropped_flag field is set to “1”.
  • arrows wd 2 indicate that the internal structure of cropping_rectangle is excerpted to be shown in detail.
  • the cropping_rectangle structure is composed of the following fields: “object_cropping_horizontal_position”; “object_cropping_vertical_position”; “object_cropping_width”; and “object_cropping_height”.
  • the object_cropping_horizontal_position field shows the horizontal position of the top left corner of the cropping rectangle.
  • the cropping rectangle defines the region, of the graphics object to be cropped.
  • the object_cropping_vertical_postion field shows the vertical postion of the top left corner of the cropping rectangle.
  • the object_cropping_width field shows the width of the cropping rectangle.
  • the object_cropping_height field shows the height of the cropping rectangle.
  • composition_object This concludes the description of the structure of composition_object.
  • a description is given to a specific example of the composition_object structure.
  • FIG. 25 is a view showing a specific example of an in-effect presenting a Right-Scroll animation effect.
  • the in-effect animation of this example presents an image such that a language selection menu gradually appears from the right edge of the screen.
  • composition_object( 0 ), composition_object( 1 ), composition_object( 2 ), and composition_object( 3 ) are allocated to points t 0 , t 1 , t 2 , and t 3 , respectively.
  • the ICS and the effect_info structure associated with each of the DSs are defined to include the following window_info structures and composition_object structures.
  • FIGS. 26-29 each illustrate an example setting of the composition_object structure.
  • FIG. 26 is a view showing an example of the composition_object( 0 ) structure.
  • the window_horizontal_position and window_vertical_position fields specify the coordinates of the top left pixel LPO of a window on the graphics plane.
  • the window_width and window_height fields specify the width and height of the window.
  • the object_cropping_horizontal_position and object_cropping vertical_position fields specify a reference point ST 1 used for determining a region of the graphics object to be cropped.
  • the reference point ST 1 is addressed based on the coordinate system which has an origin at the top left corner of the graphics object on the object buffer.
  • the region having the length and width specified by the object_cropping_width and object_cropping_height fields from the reference point ST 1 is determined as the cropping rectangle (solid-line box in the figure).
  • the cropped part of the graphics object is rendered on the graphics plane at the position having the top left pixel LPO specified by the composition_object_horizontal_position and composition_object_vertical_position fields.
  • about 1 ⁇ 4 of the language selection menu from the right is rendered within the window on the graphics plane.
  • the right one-quarter of the language selection menu is presented as a composite image overlaid on the video.
  • FIGS. 27 , 28 , and 29 are views showing examples of the composition_object structure( 1 ), composition_object structure( 2 ), and composition_object structure( 3 ), respectively.
  • the window_info structure shown in each of the figures is identical with the one shown in FIG. 26 . Thus, the description is omitted.
  • the composition_object ( 1 ) structure shown in FIG. 27 differs from the one shown in FIG. 26 .
  • the object_cropping_horizontal_position and object_cropping_vertical_position fields specify the coordinates of the top left pixel of about a right half, rather than 1 ⁇ 4, of the language selection menu stored on the object buffer.
  • the object_cropping_height and object_cropping_width fields specify the height and width of the right half of the language selection menu.
  • the respective fields of the composition_object ( 2 ) structure together specify about 3 ⁇ 4 of the language selection menu from the right.
  • the respective fields of the composition_object( 3 ) structure together specify the entire part of the language selection menu.
  • an animation effect is easily defined as desired.
  • a graphics object stored on the object buffer is rendered so as to make a gradual appearance on screen or gradual disappearance from the screen.
  • a fourth embodiment of the present invention relates to an improvement made through optimizing the order of ODSs within a DS.
  • the ODSs within a DS are referenced by the interactive_composition so as to present an in-effect or out-effect animation or one state of a button.
  • the order of ODSs within a DS is determined depending on whether the ODSs are used for an in-effect animation, an out effect animation, or a button state.
  • the ODSs within a DS are grouped into (1) ODSs used for presentation of an in-effect animation for Page[ 0 ], (2) ODSs used for presentation of the normal state of a button, (3) ODSs used for presentation of the selected state of a button, (4) ODSs used for presentation of the activated state of a button, (5) ODSs used for presentation of an in-effect for Page[ 0 ] and an in-effect and out-effect animations for Page[ 1 ] onward. That is to say, ODSs associated with the same presentation is put into the same group.
  • the group of ODSs associated with the in-effect is referred to as an “in-effect group”.
  • the group of ODSs associated with each button state is referred to as a “button-state group”.
  • the groups of ODSs associated with the out-effect of Page[ 0 ], in-effects and out-effects of Page[ 1 ] onward are referred to as a “remain group”.
  • Those groups of ODSs are arranged in the following order: in-effect group—normal-state group ⁇ seleted-state group—activated-state group ⁇ remain group. In this way, the order of ODSs within a DS is determined based on the presentation targets with which the ODSs are associated.
  • FIG. 30 is a view showing the order of ODSs within the DS n .
  • ODS 1 -ODS g belong to a group of ODSs associated with the in-effect for Page[ 0 ] (in-effect group).
  • ODS g+1 -ODS h are a group of ODSs associated with the normal state of buttons (ODSs for Normal state).
  • ODS h+1 -ODS i belong to a group of ODSs associated with the selected state of the buttons (ODSs for Selected state).
  • ODS i+1 -ODS j belong to a group of ODSs associated with the activated state of the buttons (ODSs for Activated state).
  • ODS j+1 -ODS k belong to a group of ODSs associated with the out-effect of Page[ 0 ] and ODSs associated with the in-effects and out-effects of pages from Page[ 1 ] onward (remain group).
  • the groups of ODSs are arranged in the following order: in-effect group ⁇ normal-state group ⁇ selected-state group ⁇ activated-state group ⁇ remain group. This order is determined so that ODSs needed for the first presentation of interactive display are read first, whereas ODSs needed only after a display update are read later.
  • multiple-reference means that the same object_id is referenced by two or more pieces of normal_state_info, selected_state_info, and activated_state info within an ICS.
  • a specific graphics object used to render a button in the normal state is commonly used to render another button in the selected state.
  • the graphics object is shared, so that the number of ODSs can be reduced.
  • a problem arises as to which button-state group an ODS having the multiple-reference belongs.
  • the ODS is placed in the normal-state group (N-ODSs) but not in the selected-state group (S-ODSs). Also, if an ODS is referenced by both the selected-state and activated-state groups, the ODS is placed in the selected-state group (S-ODSs) but not in the activated-state group (A-ODSs). In this manner, each ODS referenced by two or more different state_info structures, the ODS appears only once in the DS and placed in the first one of the button-state groups.
  • the DS n may not contain all the ODS groups mentioned above.
  • the DS n may not contain the selected-state group ODSs due to a multiple-reference.
  • the DS n may not contain the in-effect group ODSs. This is also because an in-effect is not essential to the presentation of interactive display.
  • S-ODSs the order of ODSs within the button-state group for the selected state.
  • S-ODSs which ODSs should be placed at the beginning depends on whether a default selected button is determined statically or dynamically.
  • the default selected button is statically determined if the default_selected_button_id_ref field in the interactive_composition is set to a valid value other than “00”. This valid value specifies the button to be selected as default.
  • the default_selected_button_id_ref field value is valid and that ODSs associated with the specified default selected button are not present in the N-ODSs, the ODSs are placed at the beginning of S-ODSs.
  • buttons_id_ref field if the default_selected_button_id_ref field is set to “00”, a different button is dynamically selected as default depending on the state of playback apparatus.
  • the default_selected button_id_ref field is set to the value “00” in the case, for example, where an AV Clip into which the DS is multiplexed serves as a merge point of a plurality of playback paths. In such a case, a different button needs to be selected as default depending on which playback path has been taken. Thus, it is not necessary to place specific ODSs at the beginning of S-ODSs.
  • FIG. 31 is a view showing how the order of ODSs within S-ODSs differs between the cases where the default_selected_button_id_ref field is set to “00” and to the valid value specifying the button B.
  • arrows ss 1 indicate the order of ODSs within S-ODSs in the case where the default_selected_button_id_ref field specifies the button_id identifying the button B.
  • Arrows ss 2 indicate the order of ODSs within 8-ODSs in the case where the default_selected_button_id_ref field is set to “00”.
  • the ODSs associated with the selected state of button B are placed at the beginning of the S-ODSs, followed by the ODSs associated with the other buttons (ODSs for Buttons A, C, and D).
  • the default_selected_button_id_ref field is set to “00”
  • the ODSs associated with the selected state of button A are placed at the beginning of the S-ODSs.
  • the order of ODSs within S-ODSs differs depending on whether the default_selected_button_id_ref field value is valued or not.
  • the ODSs within an DS are arranged in the order of appearance in the DS.
  • the display composition containing a number of graphics elements can be smoothly presented.
  • the DTS shows a time, with the time accuracy of 90 KHz, at which decoding of the associated ODS needs to be started.
  • the PTS shows a deadline for completing the decoding.
  • the decoding of ODS can not be completed instantly, and takes a certain duration.
  • the DTS and PTS show the decoding start time and decoding deadline.
  • the PTS value shows the deadline by which decoding of the associated ODS needs to be completed and the resulting uncompressed graphics object needs to be available on the object buffer of the playback apparatus.
  • the decoding start time of an arbitrary ODS j belonging to a DS n is shown by DTS (DS n [ODS]) with the time accuracy of 90 KHz.
  • the decoding deadline of the ODS j is determined by a sum of the DTS (DS n [ODS]) value and a maximum time which may be taken for the decoding.
  • SIZE ODS n [ODS j ]
  • Rd denotes the decoding rate for ODS j
  • the maximum time taken for the decoding is ceil(SIZE (DS n [ODS j ])).
  • the PTS of the ODS j belonging to the DS n can be expressed by the following equation.
  • the ICS needs to be loaded to the composition buffer of the playback apparatus before the decoding start time of the first ODS within the DS n (i.e. before the time shown by DTS(DS n [ODS 1 ])) and also before the first presentation of the DS, becomes valid (i.e. the time shown by PTS (DS n [PDS 1 ])).
  • the PTS (DS n [ICS]) value is set to satisfy the following equation.
  • the DECODEDURATION shows the time taken for decoding graphics objects needed for the first presentation of the DS n (ICS).
  • the decoding start time is equal to the DTS (ICS) value at the earliest.
  • LASTODSPTS denote the PTS associated with the graphics object of which decoding is completed at the latest among graphics objects needed for the first presentation of DS n (ICS)
  • the DECODEDURATION is equal to LASTODSPTS(DS n ) ⁇ DTS(DS n [ICS]).
  • the LASTODSPTS value takes on a different value depending on the default_selected_button_id_ref setting and the order of ODSs within the DS.
  • FIGS. 32A , 32 B, 33 A, 33 B, and 33 C show different ODS orders to illustrate LASTODSPTS setting using the DTS of an ODS.
  • FIG. 32A shows the case where the default selected button is statically determined, i.e. where the default_selected_button_id_ref field is set to a non-zero value.
  • the LASTOBSPTS takes on the PTS value of the first ODS within the S-ODSs (i.e. the value of PTS(DS n [ODS h+1 ])).
  • the ODS h+1 is enclosed within the solid line box.
  • FIG. 32B shows the case where the default selected button is determined dynamically, i.e. where the default_selected_button_id_ref is set to “00”.
  • the LASTODSPTS takes on the PTS value of the last ODS among the S-ODSs (i.e. the value of PTS (DS n [ODS i ])).
  • the ODS i is enclosed within the solid line box.
  • FIG. 33A shows a DS that includes no ODS associated with the selected state.
  • the LASTODSPTS takes on the PTS value of the last ODS among the N-ODSs (i.e. the value of PTS (DS n [ODS h ])).
  • the ODS h is enclosed within the solid line box.
  • FIG. 33B shows a DS that includes no ODS associated with the normal state.
  • the LASTODSPTS takes on the PTS value of the last ODS among the in-effect ODSs (i.e. the value of PTS(DS n [ODS g ]))
  • the ODS g is enclosed within the solid line box.
  • FIG. 33C shows a DS that includes no ODS at all.
  • the LASTODSPTS takes on the DTS value specified in the ICS of the DS n (i.e. the value of DTS(DS n [ICS])).
  • the ICS is enclosed within the solid line box.
  • the above description applies to the case where the DS n is not an Epoch Start DS. If the DS n is an Epoch Start DS n a different description is applied.
  • the graphics plane may be cleared entirely.
  • the time taken for clearing the graphics plane may be longer than the time calculated by LASTODSPTS (DS n ) ⁇ DTS(DS n [ICS]).
  • the graphics decoding and the graphics plane clearing are performed by different processing units (the graphics processor performs the decoding).
  • the graphics decoding and the plane clearing are performed in parallel.
  • the DECODEDURATION needs to set to the value equal to the plane clear duration.
  • the graphics plane has the width and height equal to the video_width and video_height field values, respectively. Furthermore, the transfer rate between the object buffer and the graphics plane is 128 Mbps. In order to clear the entire graphics plane, the playback apparatus is required to overwrite the graphics plane having the size of 8 ⁇ video_width ⁇ video_height at the transfer rate of 128 Mbps. Expressing the duration taken for this overwriting in 90 KHz, the time taken for clearing the graphics plane (referred to as “PLANECLEARTIME”) is calculated in the following equation.
  • PLANECLEARTIME ceil((90000 ⁇ 8 ⁇ DS n [ICS].video_width ⁇ DS n [ICS].video_height)/128,000,000)
  • FIG. 34A shows an algorithm for obtaining the DECODEDURATION value from the value calculated by LASTODSPTS(DS n ) ⁇ DTS(DS n [ICS]) and the value of PLANECLERATIME.
  • the larger one of the LASTODSPTS(DS n ) ⁇ DTS(DS n [ICS]) value and the PLANECLERATIME value is determined as the DECODEDURATION value.
  • the value of LASTODSPTS(DS n ) ⁇ DTS(DS n [ICS]) is determined as the DECODEDURATION value (return (LASTODSPTS (DS n ) ⁇ DTS (DS n [ICS]))).
  • the larger one of the LASTODSPTS(DS n ) ⁇ DTS(DS n [ICS]) value and the PLANECLERATIME value is determined as the DECODEDURATION value (return(max(LASTODSPTS (DS n ) ⁇ DTS(DS n [ICS])))).
  • FIGS. 34B and 34C are views showing examples of DECODEDURATION setting depending on whether or not the PLANECLEARTIME value is larger than the LASTODSPTS(DS n ) ⁇ DTS(DS n [ICS]) value. If the PLANECLEARTIME value is larger than the LASTODSPTS(DS n ) ⁇ DTS(DS n [ICS]) value, the PLANECLEARTIME is determined as the DECODEDURATION value, as shown in FIG. 34B .
  • the PLANECLEARTIME value is smaller than the LASTODSPTS(DS n ) ⁇ DTS(DS n [ICS]) value
  • the LASTODSPTS(DS n ) ⁇ DTS(DS n [ICS]) value is determined as the DECODEDURATION value, as shown in FIG. 34C .
  • the TRANSFERDURATION shows the time taken for transferring graphics data needed for presentation of the first display composition from the object buffer to the graphics plane.
  • the first display composition within a DS differs depending on whether an in-effect is provided for the DS and whether the default selected button is determined statically or dynamically.
  • the TRANSFERDURATION also differs depending on the first display composition. If the first display composition is associated with an in-effect, the time taken for transferring the uncompressed graphics needed for the presentation of the in-effect is determined as the TRANSFERDURATION.
  • the in-effect defined by the DS n [ICS] is for introducing Page[ 0 ] and rendered in windows of the number that is indicated by the number_of_window field value.
  • the total size of graphics data needed for the first display composition of the in-effect is equal to the total size of the windows ( 0 ) through (number_of_windows- 1 ).
  • the total window size is divided by the transfer rate of the playback apparatus (128 Mbps).
  • the result is further multiplexed by 90 KHz.
  • the equation for calculating the EFFECTTD (DS n ) is shown in FIG. 35A .
  • the first display composition for presentation of the DS n [ICS] is Page[ 0 ] of the Multi-Page menu. If the default selected button is statically specified, Page[ 0 ] of the Multi-Page menu is presented as soon as uncompressed data needed for the selected state of the default selected button and the normal state of the other buttons is transferred from the object buffer to the graphics plane.
  • Page[ 0 ] of the Multi-Page menu associated with the DS n includes the buttons of the number indicated by the number_of button field value.
  • the data size needed for presentation of the first display composition is calculated as follows. First, the total size is calculated for the graphics data associated with the normal state of all the buttons on the page, which are specified by DS n (ICS).Page[ 0 ]. Buttons from ( 0 ) through (number_of_Button- 1 ).
  • the size of graphics data associated with the selected state of the default selected button of Page[ 0 ] (SBSIZE (DS n , DS n [ICS].PAGE[ 0 ].default_selected_button_id_ref)) is added. Further, from the resulting value, the size of graphics data associated with the normal state of the default selected button (NBSIZE(DS n , DS n [ICS] PAGE [ 0 ] default_selected-button-Id-ref)) is subtracted to obtain the data size needed for presentation of the first display composition.
  • FIG. 35B shows the equation for calculating the PAGDEFAULTTD(DS n ) in the case where no in-effect is provided and the default selected button is statically determined.
  • the equation shown in FIG. 35B is not applicable because it is not known which of the buttons on Page[ 0 ] is selected as default.
  • the TRANSFERDURATION needs to be calculated for the worst case.
  • the worst case is where the button with the largest size is selected as default among the buttons defined by the DS n (ICS).Page[ 0 ].
  • the button with the largest size is denoted as LRG ⁇ button: button ⁇ DS n [ICS].PAGE[ 0 ]. Button ⁇ .
  • the data size needed to be transfer in the above-mentioned worst case is calculated as follows. First, the total size of all the buttons on Page[ 0 ] (Page[ 0 ] through (number_of_Button- 1 )) is calculated. To the thus calculated total size, the size of the button LRG (BSIZE (DS n , LRG ⁇ button: button ⁇ DS n [ICS].PAGE[ 0 ].Button ⁇ )) is added. Further, from the resulting value, the size of graphics data associated with the normal state of the button LRG (NBSIZE (DS n , LRG ⁇ button: button ⁇ DS n [ICS] PAGE [ 0 ]. Button ⁇ )) is subtracted to obtain the maximum data size.
  • FIG. 35C shows the equation for calculating the PAGENODEFAULTTD (DS n ) in the case where no in-effect is provided and the default selected button is dynamically determined.
  • FIG. 36 is a view showing an algorithm for selectively determining one of the EFFECTTD, PAGEDEFAULTTD, and PAGENODEFAULTTD as the TRANSFERDURATION.
  • the EFFECTTD is determined as TRANSFERDURATION (return EFFECTTD(DS n )). If no in-effect is provided for introducing Page[ 0 ] and no default selected button is statically specified (DS n [ICS.
  • the PAGEDEFAULTTD is determined as the TRANSFERDURATION (return PAGENODEFAULTTD(DS n )). If no in-effect is provided for introducing Page[ 0 ] and the default selected button is statically determined, the PAGEDEFAULTTD is determined as the TRANSFERDURATION (return PAGEDEFAULTTD (DS n )).
  • the DTS and PTS values of an ICS are optimally determined in accordance with the data size of ODSs.
  • the interactive control is smoothly performed without delay.
  • FIG. 37 is a view showing the internal structure of the playback apparatus.
  • the playback apparatus according to the sixth embodiment is industrially manufactured based on the internal structure shown in FIG. 37 .
  • the playback apparatus is mainly composed up of two parts that are a system LSI and a drive device.
  • the playback apparatus can be industrially manufactured by mounting these parts on a cabinet and substrate of the apparatus.
  • the system LSI is an integrated circuit including various processing units for achieving the functions of the playback apparatus.
  • the playback apparatus manufactured in the above manner is composed of a BD drive 1 , a read buffer 2 , a demultiplexer (De-MUX) 3 , a video decoder 4 , a video plane 5 , P-Graphics decoder 6 , a presentation plane 7 , an adder 8 , an I-Graphics decoder 9 , an interactive graphics plane 10 , an adder 11 , an audio decoder 13 , CLUT (Color Lookup Table) units 14 and 15 , a PRS set 16 , a system clock 17 , a uses timeout timer 18 , a selection_timeout timer 19 , a compostion_timeout timer 20 , an effect_duration timer 21 , an operation receiving unit 22 , and a state control unit 23 .
  • the BD drive 1 performs loading, reading, and ejecting of the BD-ROM.
  • the BD drive 1 makes access to the BB-ROM 100 .
  • the read buffer is a FIFO (first-in first-out) memory. Accordingly, TS packets read from the BD-ROM are removed from the read buffer 2 in the same order as they arrive.
  • FIFO first-in first-out
  • the De-MUX 3 reads TS packets from the read buffer 2 and converts them to PES packets. Among the resulting PES-packets, PES packets having a predetermined stream PID are outputted to one of the video decoder 4 , the audio decoder 13 , the P-Graphics decoder 6 , and the I-Graphics decoder 9 .
  • the video decoder 4 decrypts PES packets received from the De-MUX 3 , and rendered the resulting uncompressed picture data to the video plane 5 .
  • the video plane 5 is a plane memory for storing uncompressed picture data.
  • plane refers to a memory area of a playback apparatus for storing one screen of pixel data. With provision of multiple planes, the playback apparatus can add the stored data on each plane pixel by pixel to produce one video output, whereby a plurality of graphics images are overlaid as one composite image.
  • the resolution of the video plane 5 is 1920 ⁇ 1080.
  • Picture data is stored on the video plane 5 as 16-bit pixel data expressed in YUV values.
  • the P-Graphics decoder 6 decodes a graphics stream read from the BD-ROM or HD, and renders the resulting uncompressed graphics to the presentation graphics plane 7 . As the graphics stream is decoded and rendered, subtitles appear on screen.
  • the presentation graphics plane 7 is a memory having an area for storing one screen of uncompressed graphics.
  • the resolution of this plane is 1920 ⁇ 1080, and uncompressed graphics are stored on the presentation graphics plane 7 as 8-bit pixel data expressed in index color.
  • the index color By converting the index color using the CLUT (Color Lookup Table), the uncompressed graphics stored on the presentation graphics plane 7 are presented.
  • the adder 8 overlays the uncompressed picture data (i) with the contents stored on the presentation graphics plane 7 .
  • the I-Graphics decoder 9 decodes an IG stream read from the BD-ROM or HD, and renders the resulting uncompressed graphics to the interactive graphics plane 10 .
  • uncompressed graphics data decoded by the I-Graphics decoder 9 is rendered.
  • the adder 11 overlays the data on the interactive graphics plane 10 with the output of the adder 8 (i.e. with the interim video data maid by overlaying uncompressed picture data and the contents stored on the presentation graphics plane 7 ) to make the final video output.
  • the audio decoder 13 decodes PES packets received from the De-MUX 3 to output uncompressed audio data.
  • the CLUT unit 14 converts index colors of the uncompressed graphics stored on the presentation graphics plane 7 to Y, Cr, and Cb values.
  • the CLUT unit 15 converts index colors of the uncompressed graphics stored on the interactive graphics plane 10 to Y, Cr, and Cb values.
  • the PSR set 16 is a set of registers provided within the playback apparatus.
  • the set of registers include 64 player status registers (PSRs) and 4,096 general purpose registers (GPRs).
  • the 64 player status registers represent the current status of the playback apparatus.
  • the PSRs 5 - 8 represent the current playback point.
  • the PRS 8 takes on a value ranging from 0 to 0xFFFFFFFF representing the current playback point (current PTM: Presentation Time) in 45 KHz.
  • the PSR 11 represents the page_id of the currently active page (current page).
  • the PSR 10 represents the button in the selected state (current button) on the current page.
  • the system clock 17 generates clock signals.
  • the De-MUX 3 performs the conversion of PES packet in synchronism with the clock signals.
  • the user_timeout timer 18 is set to the value of user_time_out_duration field at the time specified by the PTS (DS n [ICS]).
  • PTS DS n [ICS]
  • the timer 18 starts to count down toward zero in system clock ticks.
  • the timer 18 is reset back to the user_time_out_duration field value each time a user operation is received. If no user operation is received before the timer 18 reaches zero, the timer 18 times out.
  • the selection_timeout timer 19 is set to the value of selection_time_out_pts field at the time specified by the PTS (DS n [ICS]).
  • the timer 19 starts to count down toward zero in system clock ticks.
  • the timer 19 is reset back to the selection_time_out_pts field value each time a user operation is received. If no user operation is received before the timer 19 reaches zero, the timer 19 times out. When the timer 19 times out, it is known that the valid interaction period shown in FIG. 9 has ended.
  • the composition_timeout timer 20 is set to the value of compostion_time_out_pts field at the time specified by the DTS(DS n [ICS]).
  • the compostion_time_out_pts field value is set, the timer 20 starts to count down toward zero in system clock ticks. The timer 20 is not reset even if a user operation is received.
  • the timer 20 reaches to zero, i.e. the timer 20 times out, it is known that the Epoch END has been reached.
  • the effect_duration timer 21 is set to the value of effect_duration field at the time specified by the DTS (DS n [ICS]) if the DS n is an Epoch Start DS.
  • the effect_duration field value is set, the timer 21 starts to count down toward zero in system clock ticks. The timer 21 is not reset even if a user operation is received.
  • the timer 21 reaches zero, i.e. the timer 21 times out, it is known a next display composition of the animation effect needs to be displayed.
  • the operation receiving unit 22 receives a user operation made on the remote controller or the front panel of the playback apparatus, and outputs information representing the received user operation (the information is hereinafter referred to as “UO”) to the I-Graphics decoder 9 .
  • the state control unit 23 sets desired values to the PSRs according to the instructions from the I-Graphics decoder 9 .
  • Such instructions may be given by way of (i) direct addressing or (ii) indirect addressing.
  • direct addressing the I-Graphics decoder 9 outputs an immediate value to be set to an appropriate register of the PSR set 16 .
  • indirect addressing the I-Graphics decoder 9 outputs a notification of a change in the playback apparatus status or in the user preferences. On receiving such a notification, the status controller 23 determines the value to be set to reflect the change, and sets the thus determined value to an appropriate register of the PSR set 16 .
  • the status controller 23 determines the value by performing a “procedure when playback condition is changed” or “procedure when change is requested”.
  • a description is given to the procedures performed for setting the PSR 11 (current page) and the PSR 10 (current button).
  • FIG. 38A is a flowchart of the “procedure when playback condition is changed”, preformed with respect to the PSR 11 . This procedure is performed to set the first page_info structure within an ICS to the PSR 11 (step S 99 ).
  • FIG. 38B is a flowchart of the “procedure when change is requested”, preformed with respect to the PSR 11 .
  • a user operation is received and a page having the page number X is requested to be presented.
  • the status controller 23 judges whether X is a valid value (step S 100 ). If X is a valid value, the PSR 11 is set to the value of X (step S 101 ). If X is invalid, the PSR 11 remains unchanged (step S 102 ).
  • PSR 11 changes as described above. Turning now to the PSR 10 , a description is given to the “procedure when playback condition is changed” and to the “procedure when change is requested”.
  • FIG. 39A is a flowchart of the “procedure when playback condition is changed”, performed with respect to the PSR 10 .
  • the status controller 23 judges whether the default_selected_button_id_ref field value associated with the current page is valid (step S 111 ). If the judgment in the step S 111 results in YES, the default_selected_button_id_ref field value is set to the PSR 10 (step S 112 )
  • step S 113 If the default selected button_id_ref field value is invalid, it is then judged whether the PSR 10 holds a valid value (step S 113 ). If the PSR 10 is valid, the PSR 10 remains unchanged so that the currently held value is maintained (step S 114 ). On the other hand, if the PSR 10 is invalid, the PSR 10 is set to the button_id_ref field value identifying the first button_info structure of the current page (step S 115 ).
  • FIG. 39B is a flowchart of the “procedure when change is requested”, performed with respect to the PSR 10 .
  • a user operation is made and a button having the button number X is requested to be selected.
  • the status controller 23 judges Whether X is a valid button_id field value (step S 116 ). If X is a valid value, the PSR 10 is set to the value of X (step S 118 ). If X is an invalid value, the PSR 10 remains unchanged and thus the currently held value is maintained (step S 117 ). With the above-described procedures, the PSRs 10 and 11 are ensured to hold current values at all times. This concludes the description about the internal structure of playback apparatus.
  • the I-Graphics decoder is composed of a coded data buffer 33 , a stream graphics processor 34 , an object buffer 35 , a composition buffer 36 , and a graphics controller 37 .
  • the coded data buffer 33 is a buffer for temporally storing an ICS, a PDS, and an ODS along with their DTS and PTS.
  • the stream graphics processor 34 decodes an ODS and outputs the resulting uncompressed graphics objects to the object buffer 35 .
  • the object buffer 35 is a buffer for storing a plurality of uncompressed graphics objects (represented by the boxes in the figure), which are decoded by the stream graphics processor 34 .
  • the composition buffer 36 is a memory for storing an ICS.
  • the composition buffer 36 supplies to the graphics controller 37 , the page_info structures provided within the ICS and the button_info structures provided within each page_info structure.
  • the graphics controller 37 decodes an ICS stored in the composition buffer 37 and assembles a display composition based on the decoded ICS. That is, the graphics controller 37 reads graphics objects from the object buffer 15 with reference to the page_info structure identified by the value of PSR 11 (current page_info structure), and renders the read graphics objects to the interactive graphics plane 10 .
  • the graphics objects read herein are a sequence of graphics objects specified by the normal_start_object_id_ref field value through the normal_end_object_id_ref field value of each button_info structure associated with the current page.
  • buttons O-A, O-B, O-C, and O-D are read from the object buffer 15 .
  • arrows bg 1 , bg 2 , bg 3 , and bg 4 schematically represent the rendering of graphics objects performed by the graphics controller 37 .
  • a page including the buttons O-A, O-B, O-C, and O-D is stored on the interactive graphics plane 10 and overlaid with the video.
  • the graphics controller 37 updates the display composition in response to the changes in the value of PSR 10 or PSR 11 , the timeout of the timers 18 - 21 , or a user operation received by the composition timeout timer 20 .
  • the procedures performed by the graphics controller 37 are shown in FIGS. 41-47 in detail.
  • FIG. 41 is a flowchart of the main routine of the procedure performed by the graphics controller 37 .
  • steps S 1 -S 6 form a loop.
  • judgments are made as to whether specific events have occurred (steps S 1 -S 4 )
  • a display composition of the Multi-Page menu is updated to a subsequent display composition for animated presentation (step S 5 )
  • processing is performed according to a user operation (UO processing) (step S 6 ). If any of the judgments in the steps S 1 -S 4 results in YES, corresponding steps are performed and the processing returns to the main routine.
  • step S 1 the judgment is made as to whether the current playback point has reached the point specified by the PTS (DS n [ICS]). If the judgment in the step S 1 results in YES, the above described “procedure when playback condition is changed” is performed. As a result, Page[ 0 ], which is the 1 st page of the Multi-Page menu, is set as the current page (step S 7 ). Then, the user_timeout timer 18 , the selection_timeout timer 19 , and the composition_timeout timer 20 are all reset (step S 8 ), and the current page is presented on screen (step S 9 ). The processing then goes back to the step S 1 for the next loop iteration.
  • step S 2 the judgment is made as to whether the user_timeout timer 18 has timed out. If the judgment in the step S 2 results in YES, a loop of steps S 11 -S 14 is performed. In each iteration of the loop, an arbitrary Page[j] of the Multi-Page menu is removed from the screen (step S 13 ), followed by playback of an out-effect animation provided for Page[j] (step S 14 ). The loop is repeated for each page from Page[ 0 ] onward, defined within the ICS (steps S 11 and S 12 ).
  • step S 3 the judgment is made as to whether the selection_timeout timer 19 has timed out. If the judgment in the step S 3 results in YES, a button on the current page is automatically activated (auto-activate: steps S 1 ). Then, the processing returns the step S 1 for the next loop iteration.
  • step S 4 the judgment is made as to whether the composition timeout timer 20 has timed out. If the judgment in the step S 4 results in YES, Page[ 0 ], which is the first page of the Multi-Page menu, is removed from the screen (step S 15 ), followed by playback of the out-effect animation provided for Page[ 0 ] (step S 16 ). Then, the processing returns to the step S 1 for the next loop iteration.
  • the first presentation is carried out through steps S 15 -S 23 shown in FIG. 42 , after the state control information 23 performs the above-described “procedure when playback condition is changed” to set Page[ 0 ] as the current page.
  • the state control information 23 performs the above-described “procedure when playback condition is changed” to set Page[ 0 ] as the current page.
  • the in-effect animation for the current page is executed.
  • a current button setting procedure is performed.
  • the current button is specified by the PSR 10 .
  • the “procedure when playback condition is changed” is performed with respect to the PSR 10 , so that the current button is determined.
  • the processing moves onto the steps S 17 -S 22 .
  • buttons S 17 and S 22 form a loop which is repeated for each button_info structure provided in the current page (steps S 17 and S 18 ).
  • the button_info structure to be processed in the current iteration of the loop is designated as button_info (p).
  • step S 19 a judgment is made as to whether the button_info (p) is associated with the current button. If the judgment in the step S 19 results in YES, the step S 20 is performed next. If not, the step S 21 is performed next.
  • step S 21 among the graphics objects on the object buffer 35 , one specified by the normal_start_object_id_ref field value is designated as the graphics object (p).
  • step S 20 among the graphics objects on the object buffer 35 , one specified by the selected_start_object_id_ref field value is designated as the graphics object (p).
  • the graphics object (p) is designated in the step S 20 or S 21 , the graphics object (p) is rendered to the interactive graphics plane 10 at the position specified by the button_horizontal_position and button_vertical_position field values (step S 22 ).
  • the first graphics object is rendered for each button to the interactive graphics plane 10 .
  • the CLUT unit 15 is set so as to present the current page using the pallet data specified by the pallet_id_ref field value of the current page (step S 23 ). This concludes the description of the flowchart shown in FIG. 42 .
  • FIG. 43 is a flowchart of the procedure for playback of an in-effect animation.
  • a variable t denotes an arbitrary one of display compositions within the effect_sequence defined for the in-effect.
  • a variable u denotes an arbitrary one of a composition_object out of composition_objects used for the effect(t).
  • the variables t and u are initialized to “0” (step S 30 ).
  • the effect_duration_timer 21 is set to the value of the effect (t).effect_duration field (step S 31 ), and the CLUT units 14 and 15 are set so as to use the pallet data identified by the effect (t).pallet_id_ref field value for presentation (step S 32 ).
  • a loop of steps S 33 -S 40 is performed.
  • the loop is a two-level loop.
  • the first-level loop (steps S 33 -S 37 ) uses the variable u as a control variable
  • the second-level loop (step S 33 -S 40 ) uses the variable t as a control variable.
  • the cropped part of the graphics object is then rendered within the window identified by the window_id_ref field of the composition_object(u), at the position specified by the composition_object_horizontal_position and the composition_object_vertical_position fields within the composition_object(u) (step S 34 ). Then, the variable u is incremented by “1”. The first-level loop is repeated until the variable u is equal to the number_of_composition_object field value.
  • step S 38 the variable t is incremented by “1”, and the variable u is initialized to “0” (steps S 39 ).
  • step S 40 the number_of_effect
  • a current page is presented and the first display is updated for animated presentation.
  • the first display composition of the current page is presented by rendering to the interactive graphics plane 10 , the graphics objects associated with the normal_start_object_id_ref and the selected_start_object_id_ref field values provided in each button_info structure.
  • Each button is presented in animation by updating the button image on the interactive graphics plane 10 for each iteration of the loop in the main routine.
  • Each button image is updated by overwriting the interactive graphics plane 10 with an arbitrary one of a sequence of graphics objects (qth graphics object) associated with the button.
  • each button is presented in animation by rendering a sequence of graphics objects associated with the normal_state_info and selected_state_info fields provided within each button_info structure, one by one to the interactive graphics plane 10 .
  • the variable q is used to specify individual graphics objects associated with the normal_state_info and selected-state_info fields provided within each button_info structure.
  • variable q is set to the value of variable animation (p) corresponding to the button_info (p) structure.
  • the variable q represents the current frame number of the button_info(p).
  • step S 45 a judgment is made as to whether the button_info(p) corresponds to the button currently in the selected state (i.e. the current button).
  • the variable q is added to the normal_start_object_id_ref in the button_info(p) to obtain an ID(q) (step S 46 ).
  • step S 47 is performed.
  • step S 47 a judgment is made as to whether the current button is in the activated state. If the judgment in the step S 47 results in YES, the variable q is added to the activated_start_object_id_ref field value in the button_info(p) to obtain an ID(q) (step S 54 ). Then, one of the navigation commands associated with the button_info (p) is executed (step S 55 ).
  • variable q is added to the selected_start_object_id_ref field value in the button_info(p) to obtain an ID(q) (step S 48 ).
  • the one identified by the ID (q) is rendered to the interactive graphics plane 10 .
  • the rendering position on the interactive graphics plane 10 is specified by the button_horizontal_position and the button_vertical_position fields provided in the button info (p).
  • the q th one of the graphics objects associated with the selected (or activated) state of the current button is rendered to the interactive graphics plane 10 .
  • step S 50 a judgment is made as to whether the sum of the normal_start_object_id_ref field value and the variable q has reached the normal_end_object_id_ref field value. If the sum has not reached the normal_end_object_id_ref field value, the variable q is incremented by “1” and the incremented variable q is set as the variable animation (p) (step S 51 ). On the other hand, if the sum has reached the normal_end_object_id_ref field value, a judgment is made as to whether the repeat_flag field is set to “1” (step S 52 ).
  • variable animation (p) is initialized to “0” (step S 53 ). The above steps are repeated for each button_info structure within the ICS (steps S 42 and S 43 ), and then processing returns to the main routine.
  • each button image presented on screen is updated with a new graphics object each time the main routine is performed.
  • the buttons images are presented in animation.
  • the graphics controller 37 adjusts the duration for presenting each graphics object so as to maintain the animation_frame_rate_code. This concludes the description of the procedure for animated presentation.
  • FIG. 45 is a flowchart of the procedure for executing a navigation command. First, a judgment is made as to whether the navigation command is a SetButtonPageCommand (step S 56 ). If the navigation command is not a SetButtonPageCommand, the navigation command is simply executed (step S 57 ).
  • the navigation command is a SetButtonPageCommand
  • instructions are given to the state control unit 23 so as to designate the page number specified by the operand of the navigation command as the page number X as well as to set the button number specified by the operand of the navigation command to the PSR 10 (step S 58 ).
  • the state control unit 23 executes the “procedure when change is requested”, with respect to the PSR 11 (step S 59 ).
  • the PSR 11 holds a value representing the currently presented page (i.e. the current page).
  • the state control unit 23 performs the “procedure when playback condition is changed”, with respect to the PSR 10 (step S 60 ). This concludes the description of the flowchart shown in FIG. 45 .
  • FIG. 46 shows the procedure for processing a UO.
  • steps S 61 -S 64 judgments are made in steps S 61 -S 64 as to whether specific events have occurred. An occurrence of each event requires specific steps to be performed. The processing then returns to the main routine. Specifically, in the step S 61 , a judgment is made as to whether the UO_mask_table field is set to “1”. If the field is set to “1”, the processing returns to the main routine without performing any step.
  • step S 62 the judgment is made as to whether any of the Move Up, Down, Left, and Right keys on the remote controller is pushed.
  • the user_timeout timer 18 and the selection_timeout timer 19 are reset (step S 65 ).
  • another button is designated as the new current button (step S 66 ), followed by the judgment as to whether the auto_action_flag of the newly designated current button is set to “01” (step S 67 ). If the auto_action_flag is not set to “01”, the processing returns to the main routine. On the other hand, if the auto_action_flag is set to “01”, the current button(i) is activated (step S 69 ). Then, the variable animation(i) is set to “0” (step S 70 ).
  • step S 63 the judgment is made as to whether a numeric input is made. If a numeric input is made, the user_timeout timer 18 and the selection timeout timer 19 are reset (step S 71 ), and another button is designated as the new current button (step S 72 ). The processing then returns to the main routine.
  • step S 64 a judgment is made as to whether an activate key on the remote controller is pushed.
  • the user_timeout timer 18 and the selection_timeout timer 19 are reset (step S 68 ), and then the current button (i) is activated (step S 69 ).
  • the variable animation(i) is set to “0” (step S 70 ).
  • step S 66 of designating a new current button described above is performed by calling a subroutine shown in FIG. 47 . Now, a description is given to the subroutine, with reference to the figure.
  • FIG. 47 is a flowchart of the current button change procedure. First, it is determined which of upper_button_id_ref, lower_button_id_ref, left_button_id_ref, and right_button_id_ref fields within the neighbor_info associated with the current button corresponds to the pushed key (step S 75 ).
  • buttons Y and a button X denote the button specified by the one of upper-button_id_ref, lower_button_id_ref, left_button_id_ref, and right_button_id_ref fields (step S 76 ). Then, the state control unit 23 performs the “procedure when change is requested” with respect to the PSR 10 (step S 77 ). As a result, the PSR 10 is set to the value X.
  • variables animation (X) and animation(Y) are set to “0” (step S 78 ), and returns to the main routine.
  • a button is activated in response to a UO. Yet, the button state transition to the activated state is made also when the selection_time_out_pts times out.
  • FIG. 48 a description is given to the procedure for auto-activation of a button upon timeout of the selection_time_out_pts.
  • FIG. 48 is a flowchart of the procedure for auto-activation 20 of a button.
  • a judgment is made as to whether the default_activated_button_id_ref field is set to “00” or “FF” (step S 91 ). If the field is set to “00”, the processing returns to the main routine without performing any step. On the other hand, if the default_activated_button_id_ref is set to “FF”, the current button(i) is activated (step S 93 ). Thereafter, the variable animation(i) corresponding to the current button(i) is set to “0” (step S 94 ), and returns to the main routine.
  • the button specified by the default_activated_button_id_ref field value is determined to be the new current button (i) (step S 92 ).
  • the button specified by the default_activated_button_id_ref field value receives the activated state (step S 93 ).
  • the variable animation (i) corresponding to the current button(i) is set to “0” (step S 94 ), and returns to the main routine.
  • the button in the selected state is automatically put into the activated state at an end of a predetermined duration. This concludes the description of the flowchart shown in FIG. 47 .
  • the I-Graphics decoder 9 operates as follows. First, similarly to the Always-On U/I case, the I-Graphics decoder 9 performs the decoding to obtain decoded graphics objects on the object buffer 35 . The I-Graphics decoder 9 then renders the graphics objects to the interactive graphics plane 10 by the time shown by the PTS(DS n [ICS]). Up to this point, the I-Graphics decoder 9 operates in a similar manner to the case where Always-On U/I is used.
  • the processing performed thereafter differs in that the graphics controller 37 does not overlay the page stored on the interactive graphics plane 10 , and thus the page is not presented. That is to say, graphics controller 37 “turns off” the page by not outputting the data stored on the interactive graphics plane 10 to the CLUT unit 15 . With the page not presented on screen (i.e. turned off), the graphics controller 37 waits for a user operation for Pop-Up_on. In response to a user operation for Pop-Up_on, the graphics controller 37 outputs the graphics data stored on the interactive graphics plane 10 to the CLUT unit 15 where the graphics data is overlaid with the video data. As a result, the page is “turned on” or presented on screen. As above, the graphics controller 37 turns on and off the graphics display to realize Pop-Up display.
  • the graphics controller 37 tunes off all on-screen pages, thereby leaving no page presented on screen.
  • the user_timeout timer 18 counts down toward zero.
  • the graphics controller 37 also turns off all on-screen pages, thereby leaving no page presented on screen. The turn on and turn off procedures described above are performed to realize the state transition shown in FIG. 12A .
  • the sixth embodiment of the present invention realizes the industrial production of playback apparatuses supporting the BD-ROM having the data structures as described in the first to fifth embodiments. This helps for widespread use of the BD-ROMs according to the first to fifth embodiments.
  • a seventh embodiment of the present invention relates to a manufacturing process of the BD-ROM.
  • FIG. 49 is a flowchart of the manufacturing process of the BD-ROMs according to the first to fifth embodiments.
  • the manufacturing process includes a material production step of recording video, audio, and the like (step S 201 ), an authoring step of creating an application format using an authoring device (step S 202 ), and a pressing step of creating an original master of the BD-ROM and performing stamping and bonding to complete the BD-ROM (step S 203 ).
  • the BD-ROM authoring step includes the following steps S 204 -S 213 .
  • step S 240 control information, palette definition information, and graphics are generated.
  • step S 205 the control information, the palette definition information, and the graphics are converted to functional segments.
  • step S 206 a PTS of each ICS is set based on the display timing of a picture to be synchronized with.
  • step S 207 a DTS [ODS] and a PTS[ODS] are set based on the PTS[ICS].
  • step S 208 a DTS[ICS] and a PTS[PDS] are set based on the DTS[ODS].
  • step S 209 changes in occupancy of each buffer in the player model are graphed.
  • step S 210 a judgment is made as to whether the graphed changes satisfy the constraints of the player model. If the judgment results in the negative, the DTS and PTS of each functional segment are rewritten in the step S 211 . If the judgment results in the affirmative, a graphics stream is generated in the step S 212 , and the graphics stream is multiplexed with a video stream and an audio stream, which are separately generated, to form an AV Clip in the step S 213 . After this, the AV Clip is adapted to the Blue-ray Disc Read-Only Format, to complete the application format.
  • the present invention is not limited to those specific embodiments.
  • the present invention may be embodied with any of the modifications (A) to (L) below.
  • the invention recited in each of the claims of this application includes extension and generalization of the above embodiments and their modifications below. The degree of extension and generalization depends upon the state of the art in the technical field of the present invention at the time when the present invention was made.
  • Examples of such a recording medium include: an optical disc such as a DVD-ROM, a DVD-RAM, a DVD-RW, a DVD-R, a DVD+RW, a DVD+R, a CD-R, or a CD-RW; a magneto-optical disk such as a PD or an MO; a semiconductor memory card such as a CompactFlash card, a SmartMedia card, a Memory Stick card, a Multimediacard, or a PCMCIA card; a magnetic disk such as a flexible disk, SuperDisk, Zip, or Clik!; a removable hard disk drive such as ORB, Jaz, SparQ, SyJet, EZFley, or Microdrive, and a nonremovable hard disk drive.
  • an optical disc such as a DVD-ROM, a DVD-RAM, a DVD-RW, a DVD-R, a DVD+RW, a DVD+R, a CD-R, or a CD-RW
  • the playback apparatus decodes an AV Clip on the BD-ROM and outputs the decoded AV Clip to the television.
  • the playback apparatus may be equipped with only a BD drive, with the remaining construction elements being provided in the television.
  • the playback apparatus and the television can be incorporated in a home network connected with an IEEE 1394 connector.
  • the above embodiments describe the case where the playback apparatus is connected to the television, but the playback apparatus may instead be integrated with a display device.
  • the playback apparatus may include only the system LSI (integrated circuit) which carries out an essential part of the processing.
  • the playback apparatus and the integrated circuit are both an invention described in this specification.
  • an act of manufacturing a playback apparatus based on the internal structure of the playback apparatus described in the sixth embodiment is an act of practicing the present invention.
  • any act of assigning with charge (i.e. for sale) or without charge (i.e. as a gift), leasing, and importing the playback apparatus is an act of practicing the present invention.
  • an act of offering for assignment or lease of the reproduction apparatus using storefront displays, catalogs, or brochures is an act of practicing the present invention.
  • an extension header may be added to each TS packet in the AV Clip.
  • the extension header is called a TP_extra_header, includes an arrival_time_stamp and a copy_permission_indicator, and has a data length of 4 bytes.
  • TS packets with TP_extra_headers (hereafter “EX TS packets”) are grouped in units of 32 packets, and each group is written to three sectors.
  • the 32 EX TS packets contained in the three sectors are called an “Aligned Unit”.
  • the playback apparatus transmits an Aligned Unit in the following manner.
  • the reproduction apparatus removes a TP_extra_header from each of the 32 EX TS packets in the Aligned Unit, encrypts the body of each TS packet according to the DTCP Specification, and outputs the encrypted TS packets.
  • the playback apparatus inserts an isochronous packet between adjacent TS packets. A position where the isochronous packet is inserted is based on a time shown by an arrival_time_stamp of the TP_extra_header.
  • the playback apparatus outputs a DTCP_descriptor, as well as the TS packets.
  • the DTCP_descriptor corresponds to a copy_permission_indicator in the TP_extra_header. With the provision of the DTCP_descriptor indicating “copy prohibited”, it is possible to prevent, when using the TS packets in the home network connected via the IEEE 1394 connector, the TS packets from being recorded to other devices.
  • the above embodiments relate to the case where an AV Clip of the Blu-ray.
  • Disc Read-Only Format is used as a digital stream, but the present invention may also be embodied with a VOB (Video Object) of the DVD-Video Format or the DVD-Video Recording Format.
  • the VOB is a program stream that complies with the ISO/IEC 13818-1 Standard and is obtained by multiplexing a video stream and an audio stream.
  • the video stream in the AV Clip may be an MPEG4 video stream or a WMV video stream.
  • the audio stream in the AV Clip may be a Linear PCM audio stream, a Dolby AC-3 audio stream, an MP3 audio stream, an MPEG-AAC audio stream, or a dts audio stream.
  • the movie described in the above embodiments may be one obtained by encoding an analog image signal broadcast by analog broadcasting. Also, the movie may be stream data made up of a transport stream broadcast by digital broadcasting.
  • an analog/digital image signal recorded on a videotape may be encoded to obtain content.
  • an analog/digital image signal directly captured by a video camera may be encoded to obtain a content.
  • a digital work distributed by a distribution server is applicable too.
  • Graphics objects described in the above embodiments is run-length encoded raster data.
  • Run-length encoding is used for compression/encoding of graphics objects, because the run-length encoding is suitable for compression and decompression of subtitles.
  • Subtitles have the property that a continuous length of the same pixel value in a horizontal direction is relatively long. Therefore, by performing compression using run-length encoding, a high compression rate can be attained.
  • run-length encoding reduces a load for decompression, and is therefore suitable for realizing decoding by software.
  • the use of run-length encoding for graphics objects is not essential to the present invention.
  • graphics objects may be PNG data.
  • graphics objects may be vector data instead of raster data.
  • graphics objects may be transparent patterns.
  • the transfer rate Rc may be determined so that clearing and rendering of the graphics plane complete within a vertical blanking time. Suppose the vertical blanking time is 25% of 1/29.93 seconds. Then Rc is determined to be 1 Gbps. By determining Rc in this way, graphics can be displayed smoothly, which achieves a significant effect on the practical use.
  • writing in sync with line scan can be used together with writing within a vertical blanking time. This ensures smooth presentation if the transfer rate Rc is 256 Mbps.
  • the playback apparatus may include a line buffer for storing uncompressed pixels of one line. Since conversion to an image signal is performed for each horizontal row (line), conversion to an image signal can equally be performed with the line buffer.
  • the graphics plane described in the sixth embodiment preferably has a double buffer structure. If the graphics plane has a double buffer structure, even when large-sized graphics data needs to be rendered, which requires a time corresponding to several frames, each display composition may be instantaneously presented by switching between two buffers. Thus, a double buffer structure is effective when the menu of a full-screen size needs to be presented.
  • the playback apparatus according to the present invention is suitable for a personal use, for example in a home theater system. Since the above embodiments of the present invention disclose the internal structure, playback apparatuses having the internal structure can be produced on a large scale. Thus, the playback apparatus according to the present invention is industrially usable in its quality and has industrial applicability.

Abstract

A BD-ROM stores a video stream and a graphics stream. The video stream represents a moving picture. The graphics stream is used for overlaying a multi-page menu on the moving picture, and includes interactive control information (ICS) and graphics data (ODS) used for generating the multi page menu. A PTS attached to a PES packet containing the interactive control information shows timing for the first presentation of a main page the multi-page menu in accordance with the proceeding of the video stream playback. In addition, the interactive control information includes information (selection_time_out_pts) showing a timeout upon which a button on a page of the multi-page menu is automatically activated, and information (user_time_out_duration) showing a timeout upon which a sub-page of the multi-page menu is automatically removed.

Description

    TECHNICAL FIELD
  • The present invention relates to a field of user interface control. More particularly, the present invention relates to improvements in recording mediums used for consumer movie distribution, as well as in consumer playback apparatuses having a user interface for playback of the recording mediums.
  • BACKGROUND ART
  • A user interface using a menu structure is an essential function for receiving user operations, such as selection of languages and chapters for playback.
  • When a movie offers a number of user-selectable options such as chapters, there may be a case where all the options can not be presented on a single screen display. Accordingly, there is a demand for a multi-page menu with which multiple pages are sequentially presented in response to user operations, so that numerous options are appropriately presented on respective pages of the menu.
  • Unfortunately, however, if too many pages are displayed, those pages occupy a large portion of the screen, thereby interfering with viewing the main movie also displayed on the screen. Such a problem can be avoided if a user repeatedly makes operations for removing each page that is no longer of his interest. However, the need to make such troublesome operations over and over may be another factor bothering the user watching the movie.
  • In addition, the interactive control that a DVD provides has the following disadvantage. For each menu page to be displayed, an access needs to be made to the DVD to read page information. Because of the access, the menu page can not be displayed without interruption to video playback. The interruption makes it difficult for the user to concentrate on, and enjoy the movie.
  • DISCLOSURE OF THE INVENTION
  • The present invention aims to provide a recording medium supporting presentation of multi-page menus which will not interfere with the video playback.
  • To achieve the above aim, the present invention provides a recording medium having a video stream and a graphics stream recorded thereon. The video stream represents a moving picture made up of a plurality of pictures. The graphic stream is used for overlaying a multi-page menu on the moving picture, and includes interactive control information and graphics data. The graphics data is used for generating the multi-page menu. The interactive control information includes time information used for controlling behavior of the multi-page menu in accordance with a playback proceeding of the video stream.
  • According to the recording medium having the above structure, the interactive control information multiplexed in the graphics stream includes control information. The control information defines the multi-page menu behavior in accordance with the proceeding of the movie playback. This structure makes it possible to automatically present a number of pages during the time the movie playback is showing scenery, for example. In addition, it is also possible to automatically remove the pages from the screen when the movie playback reaches a climax scene. With such control, the multi-page menu is presented without interfering with viewing the movie playback. Consequently, the user can enjoy the convenience provided by the multi-page menus, while watching the movie playback without obstructive display.
  • In addition, the information used for controlling the multi-page menu behavior is included in the graphics stream, and the graphics stream is multiplexed with the video stream. This structure allows a playback apparatus to read the information needed for presentation of the multi-page menu, without interrupting the video stream reading. Thus, pages of the multi-page menu are smoothly presented in synchronization with the video frame rate.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a view showing a usage pattern of a recording medium according to the present invention;
  • FIG. 2 is a view showing an example structure of a BD-ROM;
  • FIG. 3 is a view schematically showing the structure of an AV Clip;
  • FIG. 4A is a view showing the structure of an interactive graphics stream;
  • FIG. 4B is a view showing PES packets carrying functional segments;
  • FIG. 5 is a view showing the logical structure defined by various types of functional segments;
  • FIG. 6 is a view showing the time axis of AV clip playback to which a DSn is allocated;
  • FIGS. 7A and 7B are views showing the relations between interactive_composition structures and ICSs;
  • FIG. 8 is a view showing the internal structure of an ICS;
  • FIG. 9 is a view showing the processing performed on an ICS in the case where a stream_model field shows a “Multiplexed” type;
  • FIG. 10 is a view showing the processing performed on an ICS in the case where the stream_model field shows a “Preload” type;
  • FIG. 11 is a timing chart illustrating the functions of selection_time_out_pts and composition_time_out_pts fields in an ICS belonging to the DSn in the case where the DSn is an Epoch Start DS (n=1);
  • FIG. 12A is a view showing the state transition of a Multi-page menu when a Pop-Up U/I is used;
  • FIG. 12B is a view showing the state transition of the Multi-page menu when an Always-On U/I is used;
  • FIG. 13 is the display scenario incorporating the Multi-Page menu;
  • FIGS. 14A-14C show a display example defined by the selection_time_out_pts;
  • FIGS. 15A-15D show a display example defined by the user_time_out_duration;
  • FIG. 16 is a view showing the internal structure of page information associated with an arbitrary one of multiple pages (xth page) of the Multi-Page Menu;
  • FIG. 17 is a view showing the state transition of buttons O-A, O-B, O-C, and O-D;
  • FIG. 18 is a view showing an example of the button_info description defining the button state transition shown in FIG. 17;
  • FIG. 19 is a view showing graphical images derived from ODSs 11-19;
  • FIG. 20 is a view showing an effect_sequence structure;
  • FIG. 21A is a view schematically illustrating how an effect_duration is determined;
  • FIG. 21B shows an equation for calculating the effect_duration;
  • FIG. 22 is a view showing a specific example of an in-effect animation;
  • FIG. 23 is a view showing an out-effect animation presented in accordance with the window_width, window_height, and effect_duration field values;
  • FIG. 24 is a view showing the internal structure of an arbitrary composition_object(i) in detail;
  • FIG. 25 is a view showing a specific example of an in-effect presenting a Right-Scroll animation effect;
  • FIG. 26 is a view showing an example of the composition_object (0) structure realizing the in-effect shown in FIG. 25;
  • FIG. 27 is a view showing an example of the composition_object(1) structure realizing the in-effect shown in FIG. 25;
  • FIG. 28 is a view showing an example of the composition_object(2) structure realizing the in-effect shown in FIG. 25;
  • FIG. 29 is a view showing an example of the composition_object(3) structure realizing the in-effect shown in FIG. 25;
  • FIG. 30 is a view showing the order of ODSs within the DSn;
  • FIG. 31 is a view showing how the order of ODSs within S-ODSs differs between the cases where the default_selected_button_id ref field is set to “00” and to the valid value specifying a button B;
  • FIG. 32A is a view showing a LASTODSPTS in the case where the default selected button is statically determined;
  • FIG. 32B is a view showing a LASTODSPTS in the case where the default selected button is dynamically determined;
  • FIG. 33A is a view showing a LASTODSPTS in the case where the DS includes no ODS associated with the selected state;
  • FIG. 33B is a view showing a LASTODSPTS in the case where the DS includes no ODS associated with the normal state;
  • FIG. 33C is a view showing a LASTODSPTS in the case where the DS includes no ODS at all;
  • FIG. 34A is a view showing an algorithm for obtaining the DECODEDURATION value from the value calculated by LASTODSPTS (DSn)−DTS (DSn[ICS]) and the value of PLANECLERATIME;
  • FIGS. 34B and 34C are views showing examples of how to calculate the DECODEDURATION value;
  • FIG. 35A is a view showing an equation for calculating the EFFECTTD;
  • FIG. 35B is a view showing an equation for calculating the PAGDEFAULTTD in the case where no in-effect is provided and the default selected button is statically determined;
  • FIG. 35C is a view showing an equation for calculating the PAGENODEFAULTTD in the case where no in-effect is provided and the default selected button is dynamically determined;
  • FIG. 36 is a view showing an algorithm for selectively determining one of EFFECTTD, PAGEDEFAULTTD, and PAGENODEFAULTTD as TRANSFERDURATION;
  • FIG. 37 is a view showing the internal structure of a playback apparatus according to the present invention;
  • FIG. 38A is a flowchart of “procedure when playback condition is changed”, preformed with respect to a PSR 11;
  • FIG. 38B is a flowchart of “procedure when change is requested”, preformed with respect to the PSR 11;
  • FIG. 39A is a flowchart of “procedure when playback condition is changed”, performed with respect to a PSR 10;
  • FIG. 39B is a flowchart of “procedure when change is requested”, performed with respect to the PSR 10;
  • FIG. 40 is a view illustrating the internal structure of an I-Graphics decoder 9;
  • FIG. 41 is a flowchart of the main routine of the procedure performed by a graphics controller 37;
  • FIG. 42 is a flowchart of the procedure for the first presentation;
  • FIG. 43 is a flowchart of the procedure for playback of an in-effect animation;
  • FIG. 44 is a flowchart of the display update procedure for presentation buttons in animation;
  • FIG. 45 is a flowchart of the procedure for executing a navigation command;
  • FIG. 46 shows the procedure for processing a UO;
  • FIG. 47 is a flowchart of the current button change procedure;
  • FIG. 48 is a flowchart of the procedure for auto-activation of a button; and
  • FIG. 49 is a flowchart of the manufacturing process of the BD-ROMs according to the first to fifth embodiments.
  • BEST MODE FOR CARRYING OUT THE INVENTION First Embodiment
  • Hereinafter, a description is given to an embodiment of a recording medium according to the present invention. First of all, a description is given to usage of the recording medium, which is one form of practicing the present invention. FIG. 1 shows a usage pattern of the recording medium according to the present invention. In the figure, a BD-ROM 100 is are cording medium according to the present invention. The BD-ROM 100 is used for providing movies to a home theater system composed of a playback apparatus 200, a remote controller 300, and a television 400.
  • This concludes the description of the usage of recording medium according to the present invention. Next, a description is given to production of the recording medium, which is another form of practicing the present invention. The recording medium can be embodied by making improvements to an application layer of a BD-ROM. FIG. 2 shows an example structure of the BD-ROM 100.
  • In the figure, the BD-ROM 100 is shown on the fourth level, and a BD-ROM's track is shown on the third level. The track is stretched out into a straight line in the figure, although the track in practice spirals outwards from the center of the BD-ROM. The track is composed of a lead-in area, a volume area, and a lead-out area. The volume area has a layer model of a physical layer, a file system layer, and an application layer. The first level shows, in a directory structure, a format of the application layer (application format) of the BD-ROM.
  • The first level shows the format of the application layer (application format). In the figure, the BD-ROM has a ROOT directory, and a BDMV directory under the ROOT directory, and a STREAM directory under the BDMV directory.
  • The STREAM directory contains an AV clip and Sub clips. The AV clip and Sub clips are contained in files in the STREAM directory, such as “00001.m2ts”, “00002.m2ts”, and “00003.m2ts”. Each file in the STREAM directory has a filename of a unified format, “xxxxx.m2ts” where x is an arbitrary integer.
  • <AV Clip Structure>
  • Now, a description is given to the AV clip. FIG. 3 is a view schematically showing the structure of the AV Clip.
  • In the drawing, the middle level shows the AV clip. This AV clip can be created as follows. The video stream made up of a plurality of video frames (pictures pj1, pj2, pj3, . . . ) and the audio stream made up of a plurality of audio frames both shown on the upper first level are separately converted to PES packets shown on the upper second level, and further converted to TS packets shown on the upper third level. Likewise, the presentation graphics stream (hereinafter, “PG stream”) and the interactive graphics stream (hereinafter, “IG stream”) both shown on the lower first level are separately converted to PES packets shown on the lower second level, and further converted to TS packets shown on the lower third level. These TS packets of the video, audio, PG and IG streams are multiplexed to form the AV clip.
  • The PG and IG streams are data in stream format used for compositing graphics with each picture of the video stream. An entity of data carried by the PG and IG streams for overlaying on a picture is referred to as a “graphics object”.
  • The PG stream is a graphics stream used for presentation of subtitles using the graphics objects.
  • The IG stream is a graphics stream used for interactive presentation of the graphics objects. This concludes the description of the AV clip structure. Next, a description is given to the AV clip and the Sub clips. The Sub clips are not multiplexed. Rather, the Sub clips are files exclusively containing the IG stream, audio stream, and text data. The IG stream stored as a Sub clip is loaded to a memory in advance for synchronous playback with an AV clip. This operation of loading a Sub clip to a memory in advance is referred to as “preloading”.
  • Next, a description is given to the IG stream. FIG. 4A is a view showing the structure of the IG stream. On the first level, a string of TS packets constituting the AV clip is shown. On the second level, a string of PES packets constituting the graphics stream is shown. The PES packets shown on the second level are formed by concatenating payloads of TS packets having a predetermined PID within the TS packet string shown on the first level. Note that no further description is given to the PG stream because it is not the gist of the present invention.
  • On the third level, the structure of graphics stream is shown. The graphics stream is composed of functional segments that include an ICS (Interactive Composition Segment), a PDS (Palette Definition Segment), an ODS (Object Definition Segment), and an END (End of Display Set Segment). Of these functional segments, the ICS is a display composition segment, whereas the PDS, ODS, and END are definition segments. Each functional segment is either in one-to-one or one-to-multiple correspondence with PES packets. That is, one functional segment is recorded on the BD-ROM 100 after being converted to a single PES packet, or fragmented and converted to a plurality of PES packets.
  • Hereinafter, a description is given one by one to those functional segments.
  • The ICS is a functional segment used to define an interactive display composition of graphics. One example of the interactive display composition is a Multi-Page menu.
  • The ODS is a run-length encoded graphics object composed of a plurality of pieces of run-length data. The run-length data expresses a pixel string using a Pixel Code which shows a pixel value and a continuous length of the pixel value. The Pixel Code is an 8-bit value showing one of the values from 1 to 255. Through the use of this Pixel Code, the run-length data sets arbitrary 256 pixel colors out of full color (16,777,216 colors).
  • The PDS is a functional segment storing pallet data. The pallet data shows combinations of Pixel Codes of 1 to 255 and pixel values. The pixel value used herein is made up of a red color difference component (Cr value), a blue color difference component (Cb value), a luminance component (Y value), and a transparency (T value). Substituting a Pixel Code of each piece of run-length data into a pixel value shown by the palette data produces a color.
  • The END provides an explicit indication that the transmission of functional segments is complete. The END is inserted into the graphics stream as the last segment. This concludes the description of each functional segment.
  • FIG. 4B is a view showing PES packets carrying functional segments. As shown in the figure, each PES packet is composed of a packet header and a payload. The payload is an entity carrying a functional segment, and the packet header carries a DTS and a PTS associated with the functional segment. Hereinafter, a DTS and a PTS in a packet header of a PES packet carrying a functional segment are simply described as a DTS and a PTS of that functional segment.
  • These various types of functional segments define a logical structure such as the one shown in FIG. 5. In the drawing, the first level shows Epochs, the second level shows DSs (Display Sets), and the third level shows types of the DSs. The functional segments shown on the third level in FIG. 4A is shown on the fourth level in this figure.
  • An Epoch on the first level refers to one time unit of continuous memory management on a time axis of the AV clip playback, and also refers to a group of data allocated to that time unit. Memory mentioned herein includes a graphics plane for storing one screen of graphics objects and an object buffer for storing uncompressed graphics data. Continuous memory management means that neither the graphics plane nor the object buffer is flushed throughout the Epoch, and that clearing and rendering of graphics are performed only within a predetermined rectangular area on the graphics plane (to flush means to clear the entire graphics plane or the entire object buffer). The size and position of this rectangular area are fixed throughout the Epoch. As long as clearing and rendering of graphics are performed within this fixed rectangular area on the graphics plane, seamless playback is guaranteed. In other words, the Epoch is a time unit of seamless playback on the time axis of the AV clip playback. To change the graphics clearing/rendering area on the graphics plane, it is necessary to define a point on the playback time axis at which the change is effected and to set a new Epoch from the point onward. In this case, the boundary between two Epochs is not seamless.
  • The seamless playback mentioned herein means that the clearing and rendering of graphics are each completed with a predetermined number of video frames. In the case of IG stream, the number of video frames is four to five. The number of video frames is determined based on the ratio of the fixed rectangular area to the entire graphics plane and the transfer rate between the object buffer and the graphics plane.
  • A DS on the second level is a group of functional segments included in the graphics stream and provides one complete display composition. In FIG. 5, dashed lines hk1 show to which Epoch the DSs on the second level belong. The DS1, DS2, DS3 . . . , and DSn are shown to belong to the first Epoch on the first level.
  • The third level shows the types of DSs. The first DS in an Epoch is an “Epoch Start” DS. Each DS other then the first one in the Epoch is either an “Acquisition Point” DS, a “Normal Case” DS, or an “Epoch Continue” DS. Though the DSs are shown in the order of Acquisition Point, Normal Case, and Epoch Continue, this order is merely an example and the DSs may be arranged in a different order.
  • The Epoch Start DS provides a display effect “new display”, and indicates a start of a new Epoch. Thus, the Epoch Start DS contains all functional segments necessary for the next screen composition. The Epoch Start DS is provided at a point to which a skip operation is likely to be made, such as a start of a chapter in a movie.
  • The Acquisition Point DS provides a display effect “display refresh”, and is identical in contents to the preceding Epoch Start DS. The Acquisition Point DS is not the start of the Epoch, but contains all functional segments necessary for the next screen composition. Therefore, graphics can be displayed reliably when playback is started from the Acquisition Point DS. That is to say, the Acquisition Point DS enables a display composition to be made from a midpoint in the Epoch.
  • The Acquisition Point DS is provided at a point to which a skip operation may be made, such as a point that may be designated by a time search. The time search is an operation of locating a playback point corresponding to a time input by a user in minutes/seconds. The time input is made in a relatively large unit such as ten minutes and ten seconds. Accordingly, the Acquisition Point DS is provided at a point that can be designated by a time search made in units of 10 minutes and 10 seconds. By providing the Acquisition Point DS at such a point that can be designated by a time search, playback of the graphics stream can be smoothly carried out when a time search is conducted.
  • The Normal Case DS provides a display effect “display update”, and contains only a difference from the immediately preceding display composition. For example, if a DSv is identical in contents to an immediately preceding DSu but different in display composition, the DSv is a Normal Case DS which contains only an ICS or an ODS. This eliminates the need to provide overlapping ODSs in DSs, so that the amount of data stored on the BD-ROM 100 is reduced. Since the Normal Case DS contains the difference without the overlapping data, no display composition can be displayed with the Normal Case DS alone.
  • The Epoch Continue DS indicates at the start of an AV clip that one Epoch continues from a different AV clip. For example, a DSn contains a composition state, field having a value indicating “Epoch Continue”. In this case, the DSn belongs to the same Epoch as an immediately preceding DSn−1 even if the two DSs are on different AV clips. Since the DSn and DSn−1 belong to the same Epoch, there will be no flushing of the graphics plane and of the object buffer even if there is a boundary of AV clips between the two DSs.
  • In FIG. 5, dashed lines kz1 show to which DS the functional segments on the fourth level belong. Since the functional segments shown in FIGS. 4A and 5 are the same, it is known that the functional segments shown in FIG. 4A belong to the Epoch Start DS. Functional segments belonging to the Acquisition Point DS are identical to those belonging to the Epoch Start DS. In addition, functional segments belonging to the Normal Case DS are identical to those belonging to the Epoch Start DS, except that some of the functional segments are omitted in the Normal Case DS.
  • This concludes the description of the logical structure defined by the functional segments. Next, a description is given to the allocation of DSs having such an ICS and ODSs, to the time axis of AV clip playback. Since the Epoch is a time unit of continuous memory management and includes one or more DSs, it is important how those DSs are allocated to the time axis of the AV clip playback. Note that the “time axis of the AV clip playback” refers to the time axis used for defining the decode timing and playback timing of individual pictures multiplexed in the AV clip. The decode timing and playback timing are expressed in the time accuracy of 90 KHz. The DTS and PTS of an ICS or of an ODS within a DS show the timing based on which the synchronous control is realized. DSs are allocated to the time axis of the AV clip playback by carrying out the synchronous control in accordance with each DTS and PTS attached to an ICS and to an ODS.
  • Suppose an arbitrary one of DSs within an Epoch is denoted as a DSn, and that the DSn is allocated to the time axis of the AV clip playback based on the DTS and PTS set as shown in FIG. 6.
  • FIG. 6 shows the time axis of AV clip playback to which the DSn is allocated. In the figure, the duration of DSn starts at the time shown by the value of DTS of the ICS belonging to the DSn (DTS (DSn[ICS])), and ends at the time shown by the value of PTS of the END belonging to the DSn (PTS(DSn[ICS])). The display timing for the first presentation of the DSn is shown by the value of PTS of the ICS(PTS(DSn[ICS])). By adjusting the PTS (DSn[ICS]) value to correspond to the display timing of a desired picture within the video stream, the first presentation of the DSn will be synchronized with the video stream.
  • The PTS(DSn[ICS]) value is the sum of the time taken for decoding the ODS (DECODEDURATION) and the time taken for transferring the decoded graphics object (TRANSFERDURATION).
  • The decoding of ODS needed for the first presentation is completed within the DECODEDURATION. The duration mc1 shown in FIG. 6 is the time required for decoding an ODSm, an arbitrary one of ODSs within the DSn. The start of the decode duration mc1 is shown by the value of DTS(ODSn[ODSm]), and the end of the decode duration mc1 is shown by the value of PTS (ODSn[ODSm]).
  • An Epoch is defined by allocating every ODS within the Epoch to the time axis AV playback as described above. This concludes the description of the allocation to the time axis of AV playback.
  • One feature of the first embodiment is to control the behavior of a Multi-page menu in accordance with the proceeding of AV playback. The novel structure realizing the feature resides in an interactive_composition structure provided in the ICS. Hereinafter, a description is given to the interactive_composition structure.
  • FIGS. 7A and 7B are views showing the relations between the interactive_composition structures and the ICSs. The interactive_composition structure and the ICS may be in one-to-one correspondence as shown in FIG. 7A, or one-to-multiple correspondence as shown in FIG. 7B.
  • The one-to-one correspondence is applicable to the case where the interactive_composition structure is small enough to be contained in one ICS.
  • The one-to-multiple correspondence is applicable to the case where the interactive_composition structure is too large in size to be carried by one ICS and thus fragmented to be stored in multiple ICSs. Since the interactive_composition structure can be fragmented to be stored in multiple ICSs, there is no restriction imposed on the size of interactive_composition structure. The interactive_composition structure may be as large as 512 Kbytes, 1 Mbyte, or any number of bytes. As above, one interactive_composition structure may correspond to multiple ICSs. For the simplicity sake, however, a description hereinafter is given on precondition that the ICS and interactive_composition structure is in one-to-one correspondence.
  • FIG. 8 is a view showing the internal structure of an ICS. The ICS contains one interactive_composition structure or a fragment of one interactive_composition structure. As shown in the left part of FIG. 8, the ICS is composed of the following fields: “segment descriptor” showing the segment type “ICS”; “video_descriptor” showing the width and height of the video plane in pixels and the frame rate of the associated video stream; “composition_descriptor” showing the composition_state, which is information showing the type of DS to which the ICS belongs is whether a Normal Case DS, an Acquisition Point DS, an Epoch Start DS, or an Effect_Sequence; and “interactive_composition data_fragment” showing that whether the ICS is carrying an entire interactive_composition structure or a fragment of interactive_composition structure.
  • As indicated by an arrow cu1, FIG. 8 shows the internal structure of an interactive_composition structure in detail. The interactive_composition structure is composed of the following fields: “interactive_composition_length”, “stream_model”, “user_interface_model”, “composition_time_out_pts”, “selection_time_out_pts”, “user_time_out_duration”, “page information (1), (2) . . . (i) . . . (number_of_page-1)” for each page of the Multi-Page menu available for presentation.
  • The interactive_composition_length field shows the length of the interactive_composition structure.
  • The “stream_model” field shows the type of stream model of the associated interactive_composition structure. The stream model shows how the interactive_composition structure is stored on the BD-ROM and how the data is to be processed by a buffer (referred to as a “composition buffer”) on a playback apparatus. Specifically, the stream model shows whether (i) the graphics stream is multiplexed with an AV clip and separated to be loaded to the composition buffer or (ii) a graphics stream is stored as a Sub clip in a separate file from the AV clip and preloaded to the composition buffer. The stream_model field is provided in the interactive_composition structure because the composition buffer needs to perform different processing depending on whether the interactive_composition structure is preloaded or multiplexed. FIG. 9 is a view showing the processing performed on the “Multiplexed” ICS. In the figure, the video stream multiplexed in the AV clip is shown on the fourth level, and the graphic stream multiplexed in the AV clip is shown on the third level. The graphics stream includes three DSs, which are a DS1, a DS2, and DS8. The second level shows the proceeding of AV playback. The first level shows contents stored in the composition buffer at different times of the AV playback. When the current playback point reaches the beginning of DS1 multiplexed in the AV clip, the DS1 is loaded to the composition buffer as indicated by an arrow pf1. As a result, the composition buffer stores DS1[ICS]. Then, when the current playback point reaches the beginning of DS2 multiplexed in the AV clip, the DS2 is loaded to the composition buffer as indicated by an arrow pf2. As a result, the composition buffer stores DS2[ICS]. That is, in the case where the stream_model is “Multiplexed”, there are more than one interactive_composition structures. The playback apparatus discards a currently stored interactive_composition structure and newly loads another interactive_composition structure residing at the current playback point. Thus, the composition buffer stores one interactive_composition structure at a time.
  • This discarding and loading operation is especially effective at the time of a skip operation. When a skip operation is executed, the interactive_composition structure associated with the current playback point is no longer necessary. Thus, the playback apparatus discards the interactive_composition structure currently on the composition buffer. As shown in FIG. 9, when a skip operation am1 is executed, the composition buffer discards the stored contents at the execution timing am2 of the skip operation. Then, the DS8[ICS] residing on the skip destination is loaded to the composition buffer as indicated by an arrow pf3.
  • As above, in the case of the “multiplexed” type stream model, the composition buffer needs to discard the stored interactive_composition structure as the AV playback proceeds. On the contrary, in the case of the “preload” type stream model, the composition buffer is not required to discard the stored interactive_composition structure at all. If the composition buffer discards the stored interactive_composition structure as the AV playback proceeds, the same interactive_composition structure as the discarded one needs to be loaded from the BD-ROM again. This useless operation of re-loading the same interactive_composition structure results in an extra burden. FIG. 10 is a view showing the contents of the composition buffer. FIG. 10 is similar to FIG. 9 with respect to what is shown in each of the first to fourth level. Yet, FIG. 10 is totally different from FIG. 9 with respect to the contents of the composition buffer stored at different times. Specifically, even if the current playback point reaches where the DS1 is multiplexed, the ICS of the DS1 is not loaded to the composition buffer (indicated by the mark “X” in the figure).
  • Similarly, even if the current playback point reaches where the DS2 is multiplexed, the ICS of the DS2 is not loaded to the composition buffer (indicated by the mark “X” in the figure). Similarly, the contents stored in the composition buffer are retained rather than discarded, even if a skip operation is executed. The interactive_composition structure is provided with the stream_model field for allowing a controlling subject (the graphics controller) to accurately judge whether to discard the stored interactive_composition structure as the proceeding of AV playback or to retain the stored interactive_composition structure irrespective of the proceeding of AV playback. This concludes the description of the stream_model field.
  • Referring back to FIG. 8, the “user_interface_model” shows the type of user interface model to be used for the interactive_composition structure. The user_interface_model shows either “Always-On U/I” or “Pop-Up U/I”. When the Always-On U/I is used, a menu appears on screen and disappears in accordance with the proceeding of AV clip playback. When the Pop-Up U/I is used, a menu appears on screen and disappears in response to a user operation.
  • The composition_time_out_pts field shows the end time of an Epoch (Epoch END) to which the ICS belongs. The interactive control defined by the ICS is no longer valid at the Epoch END. That is, a point of time shown by the composition_time_out_pts field is the end of valid interactive control period.
  • The “selection_time_out_pts” field shows the button selection termination time. At the time shown by the selection_time_out_pts value, the button that is currently ins the selected is automatically activated. Note that buttons are elements representing options in the multi-page menu. The selection_time out_pts field defines the valid interaction period for user interaction to activate the buttons.
  • The IF statement (if(stream_model==‘0b’)) shown in the figure means that the above described “composition_time_out_pts” and “selection_time out_pts” fields are optional information that are provided only when the “stream_model” field shows “Multiplexed” type. In the case of ICS of which stream model is “Preloaded” type, neither of the “composition_time_out_pts” and “selection_time_out_pts” fields is provided.
  • The user_time_out_duration field shows the timeout at which on-screen pages presented in response to a user operation are removed. When the Always-On U/I is used, pages from the second page onward (referred to as sub pages) are presented in response to user operations. Thus, after the timeout defined by the user_time_out_duration, all subpages are removed from the screen leaving the first page on-screen. When the Pop-Up I/O is used, not only the sub pages but also every page of the Multi-page menu is presented in response to user operations. Thus, after the user_time_out_duration, all on-screen pages are removed leaving no page remain presented (No Menu Display).
  • Next, a description is given to the meaning of the selection_time_out_pts and composition_time_out_pts fields in an Epoch.
  • FIG. 11 is a timing chart illustrating the functions of the selection_time_out_pts and composition_time_out_pts fields in an ICS belonging to a DSn in the case where the DSn is an Epoch Start DS (n=1). The timing chart shows that the composition_time_out_pts value specifies the end of Epoch (Epoch END). The interactive display is valid for a period from the PTS(DSn[ICS]) to the Epoch END. This period is referred to as the “Interactive Display Period”.
  • The selection_time_out_pts specifies a time preceding the Epoch End. User interaction may be made during a period from the PTS(DSn[ICS]) to the time specified by the selection_time_out_pts value. This period is referred to as the “valid interaction period”. As above, the selection_time_out_pts field defines the end time of a period which starts from the first presentation of the interactive display and during which the interactive display is available for user interaction. The composition_time_out_pts field defines a period starting from the first presentation of the interactive display during which the interactive display in an Epoch remains valid and thus presented. This concludes the description of the selection_time out_pts and composition_time_out_pts fields. Next, a description is given to the Multi-page menu state transition defined by the selection_time_out_pts, composition_time_out_pts, and user_time_out_duration fields.
  • FIG. 12A is a view showing the state transition of the Multi-page menu based on the selection_time_out_pts, composition_time_out_pts, and user_time_out_duration fields. FIG. 12A is a view showing the state transition of the Multi-Page menu when the Pop-Up U/I is used. The Multi-page menu shown in the figure has the following three states: No Menu Display, 1st Page Only, Multi-Page Exist.
  • The 1st page only is the state where only the first page of the Multi-Page menu is presented.
  • The Multi-Page exist is the state where pages from the second page onward (i.e. sub pages) are presented in addition to the first page.
  • An arrow jt1 represents the transition from the No Menu Display state to the 1st Page Only state. This state transition is triggered by an event “Pop-Up_On”. The “Pop-Up_On” is an event generated within the playback apparatus in response to a user operation for a menu call.
  • An arrow jt2 represents the transition from the 1st Page Only state to the No Menu Display state. This state transition is triggered by an event “Pop-Up_Off”. The “Pop-Up_Off” is an event generated within the playback apparatus in response to a user operation for a menu removal.
  • Arrows jt3 and jt4 represent the transition from the 1st Page Only state to the Multi-Page Exist state, and then from the Multi-Page Exist state to the 1st Page Only state. The state transition is triggered by an event “ChangeButtonPage” generated upon the timeout defined by the selection_time_out_pts. The ChangeButtonPage is an event generated within the playback apparatus in response to a user operation for switching a menu page. The event is generated on precondition that a user operation for switching a menu page is made. The selection_time_out_pts shows the time for forcefully invalidating a button for receiving a user operation for switching a menu page. Thus, the description of the selection_time_out_pts field realizes the state transition of “1st Page Only→Multi Page Exist→1st Page Only”.
  • An arrow jt5 represents the transition from the 1st Page Only state to the No Menu Display state. An arrow jt6 represents the transition from the Multi-Page Exist state to the No Menu Display state. The state transition jt6 is triggered by the timeout of the user_time out_duration.
  • In the figure, the state transitions to the No Menu Display state from the 1st Page Only state (jt5) and from the Multi-Page Exist state are defined by the user_time_out_duration field. That is to say, it is possible to define such control that on-screen menu pages are automatically removed when the AV playback reaches to a climax scene. With this control, it is avoided that the menu pages interferes with viewing the video stream playback.
  • FIG. 12B shows the state transition of Multi-page menu when the Always-On U/I is used. An arrow at1 represents the transition from the No Menu Display state to the 1st Page Only state. This state transition is triggered by an event “Epoch Start”. The “Epoch Start” is an event showing that the current playback point reaches the point corresponding to the PTS associated with the interactive_composition structure. It is shown that the 1st page is automatically presented as the video stream playback proceeds and without waiting for any user operation.
  • An arrow at2 shows the transition from the 1st Page Only state to the Multi-Page Exist state, and an arrow at3 shows the transition from the Multi-Page Exist state to the 1st Page Only state. These state transitions are also triggered by the ChangeButtonPage event generated upon a timeout of the valid interaction period defined by the selection_time_out_pts.
  • An arrow at4 represents the transition from the Multi-Page Exist state to the 1st Page Only state. This state transition is triggered by the timeout of the user_time_out_duration. As apparent also from this state transition, when the Always-On U/I is used, the state transition is made to the 1st Page Only state not to the No Menu Display state upon the time but of the user_time out_duration.
  • An arrow at5 represents the transition from the 1st Page Only state to the No Menu Display state. This state transition is triggered by the time out of the composition_time_out_pts. Note that the composition_time_out_pts shows the end of Epoch (EpochEND).
  • As apparent from FIGS. 12A and 12B, the Multi-Page menu behavior is defined by the selection_time_out_pts, user_time_out_duration, and composition_time_out_pts fields provided in the interactive_composition structure.
  • FIGS. 13-15 show the screen display presented under control defined as above. FIG. 13 shows the display scenario incorporating the Multi-Page menu. In this scenario, the Multi-Page menu is not presented on screen at first. The 1st page of the Multi-Page menu is presented when the current playback point reaches the point t1, and a Sub page is additionally presented at the point t2. When the current playback point reaches the point t3, the on-screen Sub page is removed leaving only the 1st page on screen. The 1st page is then removed at the point t4 leaving no menu page displayed on screen.
  • At the bottom of FIG. 13 is the interactive_composition structure described to achieve the above display scenario. The PTS(DSn[ICS]) is set to the value corresponding to the point t1, the selection_time_out_pts to the value corresponding to the point t2, the user_time_out_duration to the value corresponding to the point t3, and the composition_time_out_pts to the value corresponding to the point t4. With this setting, the Multi-Page menu undergoes the state transitions shown in FIG. 13.
  • Now, a description is given to the setting of selection_time_out_pts value in greater detail with reference to a specific example.
  • FIGS. 14A-14C show a display example defined by the selection_time_out_pts. As shown in FIG. 14A, at the point of t1, the first page of Multi-Page Menu is overlaid with a picture of the video stream. A picture presented at the point t2 is of a scenic shot and is of less importance in the story of the AV clip. Since the picture presents no valuable contents for users, the selection_time_out_pts value is set to automatically activate a button presented on the 1st Page at the point t2. Consequently, a Sub Page is overlaid with the scenic short shown in FIG. 14B, so that a composite image as shown in FIG. 14C appears on screen.
  • In a similar manner, a description is given to the setting of user_time_out_duration value with reference to a specific display example. FIGS. 15A-15D show a display example defined by the user_time_out_duration. In this example, the Multi-Page menu is in the Multi-Page Exist state_where multiple Sub Pages (2nd, 3rd, and 4th pages) are presented on screen as shown in FIG. 15A. In the Multi-Page Exist state, the pages of the Multi-Page menu are overlaid with a picture included in the video stream. Here, the picture to be presented at the point t3 is a character image as shown in FIG. 15B. In this case, the multiple Sub Pages overlaid with the picture end up covering most of the character image as shown in FIG. 15C. As a result, the facial expression of the character image can not be seen. To avoid this interference, the user_time_out_duration is set to the value corresponding to a time immediately before the point t3. With this setting, the Multi-Page menu undergoes the state transition from the Multi-Page Exist state to the 1st Page Only state. As a result, the composite image as shown in FIG. 14D is presented at the point t3. Here, the 1st page is solely presented on screen without blocking the character image too much. Furthermore, since the 1st page remains on screen rather than removing every page of the Multi-Page Menu, there is no need to call the menu again.
  • As described above, according to the first embodiment, the ICS multiplexed in the graphics stream includes control information defining behavior of a Multi-Page Menu in accordance with the proceeding of AV playback. This makes it possible to define such control that numerous pages are presented during the playback of less important scenes, such as scenic shorts, and that on-screen pages are removed as the AV playback approaches climax scenes. With such control, the menu pages are presented without interfering with the playback of the AV clip, which is the main content. Thus, the user can concentrate on viewing the AV playback while enjoying convenience of the Multi-Page Menu functions.
  • Second Embodiment
  • A second embodiment relates to display compositions for presenting pages of a Multi-Page Menu. According to the second embodiment, a display composition is designed to present a plurality of graphical elements called button on screen. In addition, an animation effect is presented on screen before introducing and/or after removing a menu page.
  • FIG. 16 is a view showing the internal structure of page information associated with an arbitrary one of multiple pages (xth page) of the Multi-Page Menu. As shown in the left part of the figure, page information (x) is composed of the following fields: “page_id” showing an identifier uniquely identifying the page(x); “UO_mask_table”; “in_effects” specifying an animation effect to be presented as introduction of the page x to the screen; “out_effects” specifying an animation effect to be presented after removal of the page(x) from the screen; “animation_frame_rate_code” showing an animation frame rate to be followed at the time of presenting animation effects for the page(x); “default_selected_button_id_ref”, “default_activated_button_id_ref”; “pallet_id_ref”; and “button_info (1), (2) . . . (number_of_buttons-1)” each associated with one of the buttons presented on the page(x).
  • The UO_mask_table field shows permissions and prohibitions of user operations on the page (x). If this field is set to a value showing “prohibited”, a corresponding user operation to the playback apparatus is invalid while the page (x) is active.
  • The default_selected_button_id_ref field shows whether a button to be selected as default upon starting the presentation of page (x) is determined dynamically or statically. When this field is set to “0xFF”, the default selected button is determined dynamically. In such a case, the value held in a Player Status Register (PSR) of the playback apparatus is interpreted with priority, and the button shown by the PSR is put into the selected state. On the other hand, when this field is set to the value other than “0xFF”, the default selected button is determined statically. In such a case, the PRS is overwritten with the default_selected_button_id_ref field value, and the button specified by this field is put into the selected state.
  • The default_activated_button_id_ref field shows a button to be automatically activated at the time defined by the selection_time_out_pts field value. When the default_activated_button_id_ref field value is set to “FF”, the button that is in the selected state at the time of the timeout is automatically activated. On the other hand, when this field is set to “00”, no button is automatically activated. When this field is set to the value that is neither “00” or “FF”, the value is interpreted as a valid button number statically specifying the button to be automatically activated.
  • The pallet_id_ref field shows the ID associated with a pallet to be set in a CLUT unit, which will be described later.
  • Each button_info structure is information defining a button to be presented on the page (x). With the above fields, each page of the Multi-Page menu is defined. Next, a description is given to the internal structure of button_info structure. In the description, an arbitrary one of the buttons on the page (x) is denoted as a button(i). In FIG. 16, arrows cx1 indicate that the internal structure of button_info (i) is excerpted to be shown in detail.
  • Each button presented on the page has three states which are a normal state, a selected state, and an activated state. When a button is in the normal state, the button is simply displayed. When a button is in the selected stated, the button is currently focused as a result of a user operation but not yet activated. When a button is in the activated state, the button has been activated. Since each button has these three states, the button_info(i) is provided with the following information.
  • The button_id field shows a value uniquely identifying the button(i) in the interactive_composition structure.
  • The button_numeric_select_value field shows a flag indicating whether the button(i) is numerically selectable.
  • The auto_action_flag field shows whether the button(i) is automatically activated. When the auto_action_flag field is set to ON (bit value of “1”), the button(i) is transferred not to the selected state but directly to the activated state when selected. On the other hand, when the auto_action_flag field is set to OFF (bit value of “0”), the button(i) is transferred not to the activated state but to the selected state when selected.
  • The button_horizontal_position field and button_vertical_position field respectively specify the horizontal and vertical positions of the top left pixel of the button(i) on the interactive display.
  • The neighbor_info structure shows buttons to receive the selected state in response to user operations made to move a focus in an up, down, right, and left direction when the button (i) is in the selected state. The neighbor_info structure is composed of upper_button_id_ref, lower_button_id_ref, left_button_id_ref, and right_button_id_ref fields.
  • The upper_button_id_ref field specifies the button number of a button to receive the selected state if a user operation instructing to move the focus in the up direction is made while the button (i) is in the selected state. The move up operation is executed at a push of a Move Up key of the remote controller. If this field is set to the same button number as that of the button(i), the user operation to the Move Up key is ignored.
  • Similarly, the lower_button_id_ref field, the left_button_id_ref field, the right_button_id_ref field each specify the button numbers of a button to receive the selected state if a user operation instructing to move the focus to the down, left, or right direction is made, while the button(i) is in the selected state. The respective operations are executed at a push of the Move Down key, Move Left key, and Move Right key of the remote controller. If these fields are set to the same button number as that of the button (i), the user operations to the those keys are ignored.
  • The normal_state_info structure defines the normal state of the button(i), and is composed of the normal_start_object_id_ref, normal_end_object_id_ref, and normal_repeat_flag fields.
  • The normal_start_object_id_ref field specifies the first one of the object_id values that are serially assigned to a sequence of ODSs used to present the normal state of button (i) in animation.
  • The normal_end_object_id_ref field specifies the last one of the object_id values that are serially assigned to the sequence of ODSs used to present the normal state of button(i) in animation. If the normal_end_object_id_ref field specifies the same ID value as that of the normal_start_object_id_ref, the static image of a graphics object identified by that ID is presented as the button(i).
  • The normal_repeat_flag field specifies whether the animation of the button(i) in the normal state is to be continuously repeated.
  • The selected_state_info structure defines the selected state of the button(i). The selected_state_info structure is composed of the selected_state_sound_id_ref, selected_start_object_id_ref, selected_end_object_id_ref, and selected_repeat_flag fields.
  • The selected_state_sound_id_ref field specifies sound data to be reproduced as a click sound when the button(i) is put to the selected state. The sound data is specified by the sound_id of a piece of sound data contained in the file called sound.bdmv. When this field is set to “0xFF”, no sound data is associated with the selected state of the button (i) and thus no click sound is reproduced when the button(i) is changed to the selected state.
  • The selected_start_object_id_ref field specifies the first one of the object_id values that are serially assigned to a sequence of ODSs used to present the selected state of button(i) in animation.
  • The selcted_end_object_id_ref field specifies the last one of the object_id values that are serially as signed to the sequence of ODSs used to present the normal state of button(i) in animation.
  • If the selected_end_object_id_ref field specifies the same ID value as that of the selected_start_object_id_ref, the static image of a graphics object identified by the ID is presented as the button(i).
  • The selected_repeat flag field specifies whether the animation of the button(i) in the selected state is to be continuously repeated. If the selcted_start_object_id_ref and selcted_end_object_id_ref fields have the same value, the selected_repeat_flag field is set to the value of “00”.
  • The activated_state_info structure defines the activated state of the button(i), and is composed of the activated_state_sound_id_ref, activated_start_object_id_ref, and activated_end_object_id_ref fields.
  • The activated_state_sound_id_ref field specifies sound data to be reproduced as a click sound when the button(i) is put to the activated state. The sound data is specified by the sound_id of a piece of sound data contained in the sound.bdmv file. When this field is set to “0xFF”, no sound data is associated with the activated state of the button(i) and thus no click sound is reproduced when the button(i) is changed to the activated state.
  • The activated_start_object_id_ref field specifies the first one of the object_id values that are serially assigned to a sequence of ODSs used to present the activated state of button(i) in animation.
  • The activated end_object_id_ref field specifies the last one of the object_id values that are serially assigned to the sequence of ODSs used to present the activated state of button (i) in animation.
  • The “navigation_command” structure shows a navigation command to be executed when the button(i) is activated. A representative example of the navigation command is, a SetButtonPage command. The SetButtonPage command instructs the playback apparatus to display a desired page of the Multi-Page menu with a desired one of buttons on the page in the selected state. With the use of such navigation commands, content creators can readily describe page transitions at the time of authoring.
  • This concludes the description of the button_info structure. Now, a reference is made to FIG. 18 showing an example of the button_info description defining the state transition of buttons O-A to O-D, as illustrated in FIG. 17. In FIG. 17, arrows hh1 and hh2 represent the state transitions defined by the neighbor_info ( ) structure of the button info(1), which is associated with the button O-A. In the button info(1), the lower_button_id_ref field is set to the value specifying the button O-C. Thus, if a user operation is made to the Move Down key (FIG. 17, up1) while the button O-A is in the selected state, the button O-C receives the selected state (sj1). The right_button_id_ref field in the button info(1) is set to the value specifying the button O-B. Thus, if a user operation is made to the Move Right key (up2) while the button O-A is in the selected state, the button O-B receives the selected state (sj2).
  • An arrow hh3 in FIG. 17 represents the button state transition defined by the neighbor_info( ) structure in the button_info(2), which is associated with the button O-C. In the button_info(2), the upper_button_id_ref field is set to the value specifying the button O-A. Thus, if a user operation is made to the Move Up key (up3) while the button O-C is in the selected state, the button O-A is put back into the selected state (sj3).
  • Now, a description is given to the graphical images of the buttons O-A, O-B, O-C, and O-D. A DSn provided with the ICS shown in FIG. 18 includes ODSs 11-19 corresponding to the graphical images shown in FIG. 19. The normal_start_object_id_ref and normal_end_object_id_ref fields in the button_info(0) specify the ODSs 11 and 13, respectively. Thus, the normal state of the button O-A is presented in animation using the sequence of the ODSs 11-13. Similarly, the selected_start_object_id_ref and selected_end_object_id_ref fields in the button_info(0) specify the ODSs 14 and 16, respectively. Thus, the selected state of the button O-A is presented in animation using the sequence of the ODSs 14-16. With this setting, when a user operation is made to put the button O-A to the selected state, the graphical image serving as the button O-A is changed from the one presented using the ODSs 11-13 to the one presented using the ODSs 14-16. Here, if the normal_repeat_flag field in the normal_state_info( ) structure is set to the value “1”, the animated presentation of the ODSs 11-13 is repeated as indicated by “→(A)” and “→(A)” in the figure. Similarly, if the selected_repeat_flag field in the selected_state_info( ) structure is set to the value “1”, the animated presentation of the ODSs 14-16 is repeated as indicated by “→(B)” and “→(B)” in the figure.
  • As above, a plurality of sets of ODSs that can be presented in animation are associated with the buttons O-A, O-B, O-C, and O-D. Thus, with the ICS referencing the ODSs, such a control is defined that a character image serving as a button changes its facial expression in response to user operations.
  • This concludes the description of the button_info structure. Next, a description is given to animation effects. As shown in FIG. 16, the in-effects field and the out-effects field each specify an effect_sequence structure. FIG. 20 is a view showing an effect_sequence structure. As shown in the left half of the figure, the effect_sequence structure is composed of: window_info(0), (1), (2) . . . (number_of_windows-1); and the effect_info(0), (1), (2) . . . (number_of_effects-1).
  • The animation effect is presented by updating display compositions at fixed intervals. Each effect_info structure is a piece of information defining one display composition.
  • Arrows ec1 indicate that the internal structure of effect_info is excerpted to be shown in detail. As shown in the figure, the effect_info structure is composed of: the effect_duration field showing the above-mentioned fixed interval, i.e. a time period for which the associated display composition is to be presented; the pallet_id_ref field specifying a pallet used for the associated display composition; and the composition_object (0), (1), (2) . . . (number_of_composition_object-1) specifying the details of the display composition.
  • Each window_info structure defines a window or an area on the graphics plane within which the display compositions are rendered. As indicated by arrows wc1, the window_info is excerpted to be shown in detail. As shown in the figure, the window_info is composed of the following fields: “window_id” uniquely identifying a widow on the graphics plane; “window_horizontal_position” specifying the horizontal position of the top left pixel of the window; “window_vertical_position” specifying the vertical position of the top left pixel of the window; “window_width” specifying the width of the window; and “window_height” specifying the height of the window.
  • Now, a description is given to values that may be set in each of the window_horizontal_position, window_vertical_position, window_width, and window_height fields. These fields are set to the value corresponding to the coordinates within the two-dimensional graphics plane having the height and width defined by the video_height and video_width fields.
  • Since the window_horizontal_position field shows the horizontal address of the pixel on the graphics plane, this field takes on a value from “1” to the video_width value. Similarly, since the window_vertical_position shows the vertical address of the pixel on the graphics plane, this field takes on a value from “1” to the video_height value.
  • Since the window_width field shows the width of a window on the graphics plane, this field takes on a value from “1” to the value calculated by subtracting the window_horizontal_position value from the video_width value. Similarly, since the window_height field shows the height of window on the graphics plane, this field takes on a value from “1” to the value calculated by subtracting the window_vertical_position value from the video height value.
  • As above, the window_info is provided with window_horizontal_position, window_vertical_position, window_width, window_height fields defining the size and postion of a window on the graphics plane.
  • Now, a description is given to the constraint on the window size for realizing the framework for presenting display compositions of graphics at a rate of 24 frames per second. This frame rate corresponds to the video frame rate applied to the video stream playback. With this framework, the graphics objects are presented in detailed synchronization with the video stream. To realize the framework, the window must be cleared and rendered at the video frame rate. Here, a consideration is given to the required transfer rate from the object buffer to the graphics plane.
  • First, a constraint on the window size is examined. Let Rc denote the transfer rate from the object buffer to the graphics plane. In the worst-case scenario, the clearing and rendering of the window need to be performed at Rc. In other words, the clearing and rendering need to be performed at half of Rc (Rc/2).
  • Thus, the following equation needs to be satisfied.

  • (Window Size)×(Frame Rate)≈Rc/2
  • When the frame rate is 24, then RC is expressed in the following equation.

  • Rc=(Window Size)×2×24
  • When the total number of pixels of the graphics plane is 1920×1080 and the bit length of an index per pixel is 8 bits, the total capacity of the graphics plane is 2 Mbytes (≈1920×1080×8).
  • When Rc is 128 Mbps and the window size is 1/A of the graphics plane, the following equation is satisfied.

  • 128,000,000=1920×1080×1/A×8×2×24

  • 1 /A≈0.16(=128,000,000/(1920×1080×8×2×24))
  • Thus, by restricting the window size to 16% of the entire graphics plane, the synchronization between animation effects with the video stream is guaranteed as long as the animation effects are rendered at the transfer rate of 128 Mbps.
  • Suppose, for example, the effect_duration showing the rate of clearing and rendering a window is ½ or ¼ of the video frame rate. In this case, the window size can be as large as 32% or 64% of the entire graphics plane even if Rc is the same. That is to say, the effect_sequence structure has a feature that the effect_duration is determined to be longer for a larger window size, and shorter for a smaller window size. This concludes the description of the window size. Next, a description is given to the position of window. As mentioned above, the position and the size of a window are fixed throughout an Epoch.
  • The position and the size of a window are fixed throughout the duration of an Epoch for the following reason. When the position or size of a window is changed, the write address of the graphics plane needs to be changed, which incurs overhead. The overhead inevitably reduces the transfer rate from the object buffer to the graphics plane.
  • There is a constraint on the number of graphics objects allowed to be rendered in one window at a time. The number of graphics objects is restricted in order to reduce overhead incurred at the time of transferring decoded graphics objects. To be more specific, the overhead referred herein is incurred at the time of setting addresses of edge parts of the graphics objects. This overhead increases if the number of edge parts is greater.
  • If there is no restriction on the number of graphics objects that can be displayed in one window, the overhead is incurred unlimitedly when transferring graphics objects to the graphics plane, which increases fluctuations in transfer load. On the other hand, if the number of graphics objects in one window is limited to two, the transfer rate Rc can be set on an assumption that the number of instances of overhead is at most four. Hence a minimum standard for the transfer rate Rc can be determined easily.
  • Turning now to the effect_info structure, the effect_duration is determined depending on the window size. Now, a description is given to how the effect_duration is determined.
  • FIG. 21A is a view schematically illustrating how the effect_duration is determined in the figure, the (j+1)th display composition is presented after a plurality of video frames from the presentation of the jth display composition The number of the plurality of video frames is the effect_duration. In order to calculate the effect_duration, the total size of data to be rendered for presentation of each display composition must be known. Here, the effect_sequence is provided with windows of the number that is indicated by the number_of_windows field value. Thus, the total size of data needed to be rendered for presentation of the jth display composition is equal to the total size of the windows (0) through (number_of_windows-1), as indicated by the brace br.
  • As above, the jth display composition is provided with the windows (0) through (number_of_windows-1). Thus, the total data size of the windows is divided by the transfer rate of the playback apparatus (128 Mbps) and further multiplied by the time accuracy of PTS (90 KHz), so that the result is expressed in 90 KHz. The resulting value is the duration for which the jth display composition of the effect_sequence is presented. FIG. 21B shows the equation for calculating the effect_duration.
  • Note that in this specification, the operator “ceil( )” represents a divide that returns a value rounded up to an integer.
  • FIG. 22 is a view showing a specific example of an in-effect animation. The in-effect animation presents an image such that a flash of light appears on the palm of hand of the character image on screen and a menu gradually rises up after the flash of light disappears. The size of graphics data is the largest within the in-effect at the time when the menu is fully shown on screen (effect (4)). Thus, the window_width and window_height of a window are determined based on the data size of the effect (4). Suppose, the window size determined herein is about 50% of the entire graphics plane. This window size is nearly three times as large as the above-mentioned window size of 16% of the graphics plane. Consequently, the effect_duration for this in-effect is determined to be longer in order to allow the display composition to be duly updated at the rate of 128 Mbps. Each effect_duration of the effects (0), (1), (2), and (3) is set to be relatively long. According to this setting, each display composition is rendered for every three video frames, so that it is ensured that the in-effect animation is seamlessly presented. The same scheme applies to an out-effect animation. FIG. 23 is a view showing the out-effect animation for which window_width, window_height, and effect_duration field values are determined in the manner described above.
  • As described above, according to the second embodiment, the window_info defines an area in the graphics plane for rendering, and the effect_info defines a duration at intervals of which the rendering area is updated. Thus, the duration may be adjusted to be longer when the rendering area is set to be larger, and shorter when the rendering area is set to be smaller. With this setting, the seamless presentation of animation effects is ensured as long as the playback apparatus transfers graphics data at a fixed transfer-rate.
  • Third Embodiment
  • A third embodiment of the present invention relates to an improvement on the presentation of animation effects, such as Scroll, Wipe-In/Out, and Cut-In/Out. The structure for achieving the animation effects lies in the composition_object structures shown in FIG. 20. FIG. 24 is a view showing the internal structure of an arbitrary composition_object(i) in detail. As shown in the figure, the composition_object(i) is composed of the following fields: “object_id_ref”; “window_id_ref”, “object_cropped_flag”; “composition_object_horizontal_position”; “composition_object vertical_position”, and “cropping_rectangle_info(1), (2) . . . (n)”.
  • The object_id_ref field shows a reference value for a graphics object identifier (object_id). The reference value corresponds to the object_id identifying the graphics object that is used for generating a display composition according to the composition_object(i).
  • The window_id_ref field shows a reference value for a window identifier (window_id). The reference value corresponds to the window_id identifying the window in which the display composition according to the composition_object(i) is rendered.
  • The object_cropped flag field shows a flag indicating whether the graphics object cropped on the object buffer is to be presented or not to be presented. When this field is set to “1”, the graphics object cropped on the object buffer is presented. On the other hand, when this field is set to “0”, the cropped graphics object is not presented.
  • The composition_object_horizontal_position field shows the horizontal postion of the top left pixel of the graphics object on the graphics plane.
  • The composition_object_vertical_position field shows the vertical position of the top left pixel of the graphics object on the graphics plane.
  • The cropping_rectangle structure is valid when the object_cropped_flag field is set to “1”. In FIG. 24, arrows wd2 indicate that the internal structure of cropping_rectangle is excerpted to be shown in detail. As shown in the figure, the cropping_rectangle structure is composed of the following fields: “object_cropping_horizontal_position”; “object_cropping_vertical_position”; “object_cropping_width”; and “object_cropping_height”.
  • The object_cropping_horizontal_position field shows the horizontal position of the top left corner of the cropping rectangle. The cropping rectangle defines the region, of the graphics object to be cropped.
  • The object_cropping_vertical_postion field shows the vertical postion of the top left corner of the cropping rectangle.
  • The object_cropping_width field shows the width of the cropping rectangle.
  • The object_cropping_height field shows the height of the cropping rectangle.
  • This concludes the description of the structure of composition_object. Next, a description is given to a specific example of the composition_object structure.
  • FIG. 25 is a view showing a specific example of an in-effect presenting a Right-Scroll animation effect. The in-effect animation of this example presents an image such that a language selection menu gradually appears from the right edge of the screen. For this in-effect animation, composition_object(0), composition_object(1), composition_object(2), and composition_object(3) are allocated to points t0, t1, t2, and t3, respectively. Furthermore, the ICS and the effect_info structure associated with each of the DSs are defined to include the following window_info structures and composition_object structures.
  • Now, a description is given to the setting of each composition_object structure. FIGS. 26-29 each illustrate an example setting of the composition_object structure. FIG. 26 is a view showing an example of the composition_object(0) structure.
  • In the figure, the window_horizontal_position and window_vertical_position fields specify the coordinates of the top left pixel LPO of a window on the graphics plane. The window_width and window_height fields specify the width and height of the window.
  • As shown in the figure, the object_cropping_horizontal_position and object_cropping vertical_position fields specify a reference point ST1 used for determining a region of the graphics object to be cropped. The reference point ST1 is addressed based on the coordinate system which has an origin at the top left corner of the graphics object on the object buffer. The region having the length and width specified by the object_cropping_width and object_cropping_height fields from the reference point ST1 is determined as the cropping rectangle (solid-line box in the figure). The cropped part of the graphics object is rendered on the graphics plane at the position having the top left pixel LPO specified by the composition_object_horizontal_position and composition_object_vertical_position fields. In this example, about ¼ of the language selection menu from the right is rendered within the window on the graphics plane. As a result, the right one-quarter of the language selection menu is presented as a composite image overlaid on the video.
  • FIGS. 27, 28, and 29 are views showing examples of the composition_object structure(1), composition_object structure(2), and composition_object structure(3), respectively. The window_info structure shown in each of the figures is identical with the one shown in FIG. 26. Thus, the description is omitted. The composition_object (1) structure shown in FIG. 27, however, differs from the one shown in FIG. 26. In FIG. 27, the object_cropping_horizontal_position and object_cropping_vertical_position fields specify the coordinates of the top left pixel of about a right half, rather than ¼, of the language selection menu stored on the object buffer. The object_cropping_height and object_cropping_width fields specify the height and width of the right half of the language selection menu. Similarly, in FIG. 28, the respective fields of the composition_object (2) structure together specify about ¾ of the language selection menu from the right. In FIG. 29, the respective fields of the composition_object(3) structure together specify the entire part of the language selection menu. By rendering the cropped part of the graphics object as shown in FIGS. 27, 28, and 29 to the graphics plane, the right half of the language selection menu, the ¾ of the language selection menu from the right, and the entire language selection menu are overlaid on the video.
  • As described above, according to the third embodiment of the present invention, with the composition_object structures, an animation effect is easily defined as desired. For example, it is easily defined that a graphics object stored on the object buffer is rendered so as to make a gradual appearance on screen or gradual disappearance from the screen.
  • Fourth Embodiment
  • A fourth embodiment of the present invention relates to an improvement made through optimizing the order of ODSs within a DS. As described above, the ODSs within a DS are referenced by the interactive_composition so as to present an in-effect or out-effect animation or one state of a button. The order of ODSs within a DS is determined depending on whether the ODSs are used for an in-effect animation, an out effect animation, or a button state.
  • Specifically, the ODSs within a DS are grouped into (1) ODSs used for presentation of an in-effect animation for Page[0], (2) ODSs used for presentation of the normal state of a button, (3) ODSs used for presentation of the selected state of a button, (4) ODSs used for presentation of the activated state of a button, (5) ODSs used for presentation of an in-effect for Page[0] and an in-effect and out-effect animations for Page[1] onward. That is to say, ODSs associated with the same presentation is put into the same group. The group of ODSs associated with the in-effect is referred to as an “in-effect group”. The group of ODSs associated with each button state is referred to as a “button-state group”. The groups of ODSs associated with the out-effect of Page[0], in-effects and out-effects of Page[1] onward are referred to as a “remain group”.
  • Those groups of ODSs are arranged in the following order: in-effect group—normal-state group→seleted-state group—activated-state group→remain group. In this way, the order of ODSs within a DS is determined based on the presentation targets with which the ODSs are associated.
  • Next, a description is given specifically to the order of ODSs within a DSn. FIG. 30 is a view showing the order of ODSs within the DSn.
  • In the figure, ODS1-ODSg belong to a group of ODSs associated with the in-effect for Page[0] (in-effect group).
  • ODSg+1-ODSh are a group of ODSs associated with the normal state of buttons (ODSs for Normal state).
  • ODSh+1-ODSi belong to a group of ODSs associated with the selected state of the buttons (ODSs for Selected state).
  • ODSi+1-ODSj belong to a group of ODSs associated with the activated state of the buttons (ODSs for Activated state).
  • ODSj+1-ODSk belong to a group of ODSs associated with the out-effect of Page[0] and ODSs associated with the in-effects and out-effects of pages from Page[1] onward (remain group).
  • The groups of ODSs are arranged in the following order: in-effect group→normal-state group→selected-state group→activated-state group→remain group. This order is determined so that ODSs needed for the first presentation of interactive display are read first, whereas ODSs needed only after a display update are read later.
  • Next, a description is given to the order of ODSs when there is a multiple-reference. The term “multiple-reference” used herein means that the same object_id is referenced by two or more pieces of normal_state_info, selected_state_info, and activated_state info within an ICS. For example, with the multiple references, a specific graphics object used to render a button in the normal state is commonly used to render another button in the selected state. In other words, the graphics object is shared, so that the number of ODSs can be reduced. Here, a problem arises as to which button-state group an ODS having the multiple-reference belongs.
  • Specifically, when one ODS is associated with one button in the normal state as well as another button in the selected state, it needs to be determined whether the ODS belongs to the button-state group for the normal state or the button-state group for the selected state.
  • In such a case, the ODS placed in the button-state group that first appears in the ICS.
  • For example, if an ODS is referenced by both the normal-state and selected-state groups, the ODS is placed in the normal-state group (N-ODSs) but not in the selected-state group (S-ODSs). Also, if an ODS is referenced by both the selected-state and activated-state groups, the ODS is placed in the selected-state group (S-ODSs) but not in the activated-state group (A-ODSs). In this manner, each ODS referenced by two or more different state_info structures, the ODS appears only once in the DS and placed in the first one of the button-state groups.
  • The same applies to the ODSs associated with animation effects. If a graphics object is commonly used in the in-effect and the out-effect for Page[0], the graphics object appears only once in the DS and placed in the in-effect group. Similarly, if a graphics object is commonly used in the in-effect for Page[0] and the in-effect or out-effect for Page[1] onward, the graphics object appears only once in the DS and placed in the in-effect group. This concludes the description of the order of ODSs where there is a multiple-reference to an ODS.
  • Because of possible multiple-references to ODSs as described above, the DSn may not contain all the ODS groups mentioned above. For example, the DSn may not contain the selected-state group ODSs due to a multiple-reference. Furthermore, the DSn may not contain the in-effect group ODSs. This is also because an in-effect is not essential to the presentation of interactive display.
  • Now, a description is given to the order of ODSs within the button-state group for the selected state (S-ODSs). Among the S-ODSs, which ODSs should be placed at the beginning depends on whether a default selected button is determined statically or dynamically. The default selected button is statically determined if the default_selected_button_id_ref field in the interactive_composition is set to a valid value other than “00”. This valid value specifies the button to be selected as default. In the case where the default_selected_button_id_ref field value is valid and that ODSs associated with the specified default selected button are not present in the N-ODSs, the ODSs are placed at the beginning of S-ODSs.
  • On the other hand, if the default_selected_button_id_ref field is set to “00”, a different button is dynamically selected as default depending on the state of playback apparatus.
  • The default_selected button_id_ref field is set to the value “00” in the case, for example, where an AV Clip into which the DS is multiplexed serves as a merge point of a plurality of playback paths. In such a case, a different button needs to be selected as default depending on which playback path has been taken. Thus, it is not necessary to place specific ODSs at the beginning of S-ODSs.
  • FIG. 31 is a view showing how the order of ODSs within S-ODSs differs between the cases where the default_selected_button_id_ref field is set to “00” and to the valid value specifying the button B. In the figure, arrows ss1 indicate the order of ODSs within S-ODSs in the case where the default_selected_button_id_ref field specifies the button_id identifying the button B. Arrows ss2 indicate the order of ODSs within 8-ODSs in the case where the default_selected_button_id_ref field is set to “00”. As shown in the figure, in the case where the default_selected_button_id_ref field specifies the button B, the ODSs associated with the selected state of button B (ODSs for Button B) are placed at the beginning of the S-ODSs, followed by the ODSs associated with the other buttons (ODSs for Buttons A, C, and D). On the other hand, in the case where the default_selected_button_id_ref field is set to “00”, the ODSs associated with the selected state of button A are placed at the beginning of the S-ODSs. As described above, the order of ODSs within S-ODSs differs depending on whether the default_selected_button_id_ref field value is valued or not.
  • According to the fourth embodiment, the ODSs within an DS are arranged in the order of appearance in the DS. Thus, the display composition containing a number of graphics elements can be smoothly presented.
  • Fifth Embodiment
  • In the first embodiment, how to allocate DSs to the time axis of AV Clip playback is described. In a fifth embodiment of the present invention, a description is given to how to set PTS and DTS values.
  • First of all, a description is given to a mechanism of effecting synchronous control based on the DTS and PTS within an ODS.
  • The DTS shows a time, with the time accuracy of 90 KHz, at which decoding of the associated ODS needs to be started. The PTS shows a deadline for completing the decoding.
  • The decoding of ODS can not be completed instantly, and takes a certain duration. In order to explicitly show the start and end times of decoding the ODS, the DTS and PTS show the decoding start time and decoding deadline.
  • The PTS value shows the deadline by which decoding of the associated ODS needs to be completed and the resulting uncompressed graphics object needs to be available on the object buffer of the playback apparatus.
  • The decoding start time of an arbitrary ODSj belonging to a DSn is shown by DTS (DSn[ODS]) with the time accuracy of 90 KHz. Thus, the decoding deadline of the ODSj is determined by a sum of the DTS (DSn[ODS]) value and a maximum time which may be taken for the decoding.
  • Here, let SIZE (ODSn[ODSj]) denotes the size of the ODSj after decompression, and Rd denotes the decoding rate for ODSj, then the maximum time taken for the decoding is ceil(SIZE (DSn[ODSj])).
  • By converting this maximum time to the accuracy of 90 KHz and adding the result to the DTS of the ODSj, the decoding deadline to be shown by the PTS is calculated with the accuracy of 90 KHz.
  • The PTS of the ODSj belonging to the DSn can be expressed by the following equation.

  • PTS(DS[ODSj])=DTS(DSn[ODSj])+90,000×ceil(SIZE(DSn[ODSj])/Rd)
  • In addition, the relation between two adjacent ODSs (ODSj and ODSj+1) needs to satisfy the following equation.

  • PTS(DSn[ODSj])≦DTS(DSn[ODSj+1])
  • This concludes the description of the PTS and the DTS of an ODS. Next, a description is given to the PTS within an ICS. The ICS needs to be loaded to the composition buffer of the playback apparatus before the decoding start time of the first ODS within the DSn (i.e. before the time shown by DTS(DSn[ODS1])) and also before the first presentation of the DS, becomes valid (i.e. the time shown by PTS (DSn[PDS1])). Thus, the following equations need to be satisfied.

  • DTS(DSn[ICS])≦DTS(DSn[ODS1])

  • DTS(DSn[ICS])≦PTS(DSn[PDS1])
  • Next, a description is given to the PTS within the ICS of the DSn. The PTS (DSn[ICS]) value is set to satisfy the following equation.

  • PTS(DSn[ICS])≧PTS(DSn[ICS])+DECODEDURATION+TRANSFERDURATION
  • The DECODEDURATION shows the time taken for decoding graphics objects needed for the first presentation of the DSn(ICS). The decoding start time is equal to the DTS (ICS) value at the earliest. Here, let LASTODSPTS denote the PTS associated with the graphics object of which decoding is completed at the latest among graphics objects needed for the first presentation of DSn(ICS), then the DECODEDURATION is equal to LASTODSPTS(DSn)−DTS(DSn[ICS]).
  • The LASTODSPTS value takes on a different value depending on the default_selected_button_id_ref setting and the order of ODSs within the DS. FIGS. 32A, 32B, 33A, 33B, and 33C show different ODS orders to illustrate LASTODSPTS setting using the DTS of an ODS.
  • FIG. 32A shows the case where the default selected button is statically determined, i.e. where the default_selected_button_id_ref field is set to a non-zero value. In this case, the LASTOBSPTS takes on the PTS value of the first ODS within the S-ODSs (i.e. the value of PTS(DSn[ODSh+1])). In the figure, the ODSh+1 is enclosed within the solid line box.
  • FIG. 32B shows the case where the default selected button is determined dynamically, i.e. where the default_selected_button_id_ref is set to “00”. In this case, the LASTODSPTS takes on the PTS value of the last ODS among the S-ODSs (i.e. the value of PTS (DSn[ODSi])). In the figure, the ODSi is enclosed within the solid line box.
  • FIG. 33A shows a DS that includes no ODS associated with the selected state. In this case, the LASTODSPTS takes on the PTS value of the last ODS among the N-ODSs (i.e. the value of PTS (DSn[ODSh])). In the figure, the ODSh is enclosed within the solid line box.
  • FIG. 33B shows a DS that includes no ODS associated with the normal state. In this case, the LASTODSPTS takes on the PTS value of the last ODS among the in-effect ODSs (i.e. the value of PTS(DSn[ODSg])) In the figure, the ODSg is enclosed within the solid line box.
  • FIG. 33C shows a DS that includes no ODS at all. In this case, the LASTODSPTS takes on the DTS value specified in the ICS of the DSn (i.e. the value of DTS(DSn[ICS])). In the figure, the ICS is enclosed within the solid line box.
  • Note that the above description applies to the case where the DSn is not an Epoch Start DS. If the DSn is an Epoch Start DSn a different description is applied.
  • In the case of an Epoch Start DS, the graphics plane may be cleared entirely. The time taken for clearing the graphics plane may be longer than the time calculated by LASTODSPTS (DSn)−DTS(DSn[ICS]). According to a target decoder model for the BD-ROM, the graphics decoding and the graphics plane clearing are performed by different processing units (the graphics processor performs the decoding). Thus, the graphics decoding and the plane clearing are performed in parallel. Even if the decoding is completed for graphics data necessary for presentation of the first display composition, the graphics date can not be rendered to the graphics plane unless the plane has been cleared. Thus, the DECODEDURATION needs to set to the value equal to the plane clear duration. Suppose the graphics plane has the width and height equal to the video_width and video_height field values, respectively. Furthermore, the transfer rate between the object buffer and the graphics plane is 128 Mbps. In order to clear the entire graphics plane, the playback apparatus is required to overwrite the graphics plane having the size of 8×video_width×video_height at the transfer rate of 128 Mbps. Expressing the duration taken for this overwriting in 90 KHz, the time taken for clearing the graphics plane (referred to as “PLANECLEARTIME”) is calculated in the following equation.

  • PLANECLEARTIME=ceil((90000×8×DSn[ICS].video_width×DSn[ICS].video_height)/128,000,000)
  • FIG. 34A shows an algorithm for obtaining the DECODEDURATION value from the value calculated by LASTODSPTS(DSn)−DTS(DSn[ICS]) and the value of PLANECLERATIME. In the algorithm, the larger one of the LASTODSPTS(DSn)−DTS(DSn[ICS]) value and the PLANECLERATIME value is determined as the DECODEDURATION value.
  • As shown in the figure, if the composition_state field of the DSn(ICS) indicates that the DSn is not an epoch start DS (if(DSn(ICS) composition_state==EPOCH_START) else), the value of LASTODSPTS(DSn)−DTS(DSn[ICS]) is determined as the DECODEDURATION value (return (LASTODSPTS (DSn)−DTS (DSn[ICS]))). On the other hand, if the composition_state field of the DSn(ICS) indicates that the DSn is an epoch start DS (if (DSn(ICS) composition_state==EPOCHSTART)), the larger one of the LASTODSPTS(DSn)−DTS(DSn[ICS]) value and the PLANECLERATIME value is determined as the DECODEDURATION value (return(max(LASTODSPTS (DSn)−DTS(DSn[ICS])))).
  • FIGS. 34B and 34C are views showing examples of DECODEDURATION setting depending on whether or not the PLANECLEARTIME value is larger than the LASTODSPTS(DSn)−DTS(DSn[ICS]) value. If the PLANECLEARTIME value is larger than the LASTODSPTS(DSn)−DTS(DSn[ICS]) value, the PLANECLEARTIME is determined as the DECODEDURATION value, as shown in FIG. 34B. On the other hand, if the PLANECLEARTIME value is smaller than the LASTODSPTS(DSn)−DTS(DSn[ICS]) value, the LASTODSPTS(DSn)−DTS(DSn[ICS]) value is determined as the DECODEDURATION value, as shown in FIG. 34C.
  • The TRANSFERDURATION shows the time taken for transferring graphics data needed for presentation of the first display composition from the object buffer to the graphics plane. The first display composition within a DS differs depending on whether an in-effect is provided for the DS and whether the default selected button is determined statically or dynamically. Thus, the TRANSFERDURATION also differs depending on the first display composition. If the first display composition is associated with an in-effect, the time taken for transferring the uncompressed graphics needed for the presentation of the in-effect is determined as the TRANSFERDURATION.
  • Here, the in-effect defined by the DSn[ICS] is for introducing Page[0] and rendered in windows of the number that is indicated by the number_of_window field value. Thus, the total size of graphics data needed for the first display composition of the in-effect is equal to the total size of the windows (0) through (number_of_windows-1). Thus, to calculate the time taken for transferring graphics data needed for presentation of the first display composition within the in-effect sequence (EFFECTTD), the total window size is divided by the transfer rate of the playback apparatus (128 Mbps). To express the result in the time accuracy of the PTS (90 KHz), the result is further multiplexed by 90 KHz. The equation for calculating the EFFECTTD (DSn) is shown in FIG. 35A.
  • On the other hand, if the DSn is not provided with any in-effect, the first display composition for presentation of the DSn[ICS] is Page[0] of the Multi-Page menu. If the default selected button is statically specified, Page[0] of the Multi-Page menu is presented as soon as uncompressed data needed for the selected state of the default selected button and the normal state of the other buttons is transferred from the object buffer to the graphics plane.
  • Here, Page[0] of the Multi-Page menu associated with the DSn(ICS) includes the buttons of the number indicated by the number_of button field value. Thus, the data size needed for presentation of the first display composition is calculated as follows. First, the total size is calculated for the graphics data associated with the normal state of all the buttons on the page, which are specified by DSn(ICS).Page[0]. Buttons from (0) through (number_of_Button-1). Then, to the thus calculated total size, the size of graphics data associated with the selected state of the default selected button of Page[0] (SBSIZE (DSn, DSn[ICS].PAGE[0].default_selected_button_id_ref)) is added. Further, from the resulting value, the size of graphics data associated with the normal state of the default selected button (NBSIZE(DSn, DSn[ICS] PAGE [0] default_selected-button-Id-ref)) is subtracted to obtain the data size needed for presentation of the first display composition.
  • The thus calculated data size needed for presentation of the first display composition is divided by the transfer rate of the playback apparatus (128 Mbps), and multiplied by 90 KHz for expressing the result in the time accuracy of PTS (90 KHz). Then, the transfer duration taken for presentation of the first page is calculated. FIG. 35B shows the equation for calculating the PAGDEFAULTTD(DSn) in the case where no in-effect is provided and the default selected button is statically determined.
  • In the case where the default selected button is dynamically determined, the equation shown in FIG. 35B is not applicable because it is not known which of the buttons on Page[0] is selected as default. Thus, the TRANSFERDURATION needs to be calculated for the worst case. The worst case is where the button with the largest size is selected as default among the buttons defined by the DSn(ICS).Page[0]. Among the buttons of DSn(ICS).Page[0], the button with the largest size is denoted as LRG{button: buttonεDSn[ICS].PAGE[0]. Button}.
  • Then, the data size needed to be transfer in the above-mentioned worst case (the maximum data size) is calculated as follows. First, the total size of all the buttons on Page[0] (Page[0].Buttons from (0) through (number_of_Button-1)) is calculated. To the thus calculated total size, the size of the button LRG (BSIZE (DSn, LRG{button: buttonεDSn[ICS].PAGE[0].Button})) is added. Further, from the resulting value, the size of graphics data associated with the normal state of the button LRG (NBSIZE (DSn, LRG{button: buttonεDSn[ICS] PAGE [0]. Button})) is subtracted to obtain the maximum data size.
  • The maximum data size calculated in the above manner is divided by the transfer rate of the playback apparatus (128 Mbps), and further multiplied by 90 KHz for expressing the result in the time accuracy of PTS (90 KHz). Then, the transfer duration taken for presentation of the first page of the Multi-Page menu is calculated. FIG. 35C shows the equation for calculating the PAGENODEFAULTTD (DSn) in the case where no in-effect is provided and the default selected button is dynamically determined.
  • FIG. 36 is a view showing an algorithm for selectively determining one of the EFFECTTD, PAGEDEFAULTTD, and PAGENODEFAULTTD as the TRANSFERDURATION. As shown in the figure, if at least one in-effect is provided (if (DSn[ICS].PAGE[0].IN EFFECTS.number_of effects!=0)), the EFFECTTD is determined as TRANSFERDURATION (return EFFECTTD(DSn)). If no in-effect is provided for introducing Page[0] and no default selected button is statically specified (DSn[ICS. PAGE[0].default_selected_button_id_ref==0xFFFF]) the PAGEDEFAULTTD is determined as the TRANSFERDURATION (return PAGENODEFAULTTD(DSn)). If no in-effect is provided for introducing Page[0] and the default selected button is statically determined, the PAGEDEFAULTTD is determined as the TRANSFERDURATION (return PAGEDEFAULTTD (DSn)).
  • As described above, according to the fifth embodiment of the present invention, the DTS and PTS values of an ICS are optimally determined in accordance with the data size of ODSs. Thus, the interactive control is smoothly performed without delay.
  • Sixth Embodiment
  • A sixth embodiment of the present invention relates to the playback apparatus used for playback of the BD-ROMs described in the first to fifth embodiments. FIG. 37 is a view showing the internal structure of the playback apparatus. The playback apparatus according to the sixth embodiment is industrially manufactured based on the internal structure shown in FIG. 37. The playback apparatus is mainly composed up of two parts that are a system LSI and a drive device. The playback apparatus can be industrially manufactured by mounting these parts on a cabinet and substrate of the apparatus. The system LSI is an integrated circuit including various processing units for achieving the functions of the playback apparatus.
  • The playback apparatus manufactured in the above manner is composed of a BD drive 1, a read buffer 2, a demultiplexer (De-MUX) 3, a video decoder 4, a video plane 5, P-Graphics decoder 6, a presentation plane 7, an adder 8, an I-Graphics decoder 9, an interactive graphics plane 10, an adder 11, an audio decoder 13, CLUT (Color Lookup Table) units 14 and 15, a PRS set 16, a system clock 17, a uses timeout timer 18, a selection_timeout timer 19, a compostion_timeout timer 20, an effect_duration timer 21, an operation receiving unit 22, and a state control unit 23.
  • The BD drive 1 performs loading, reading, and ejecting of the BD-ROM. The BD drive 1 makes access to the BB-ROM 100.
  • The read buffer is a FIFO (first-in first-out) memory. Accordingly, TS packets read from the BD-ROM are removed from the read buffer 2 in the same order as they arrive.
  • The De-MUX 3 reads TS packets from the read buffer 2 and converts them to PES packets. Among the resulting PES-packets, PES packets having a predetermined stream PID are outputted to one of the video decoder 4, the audio decoder 13, the P-Graphics decoder 6, and the I-Graphics decoder 9.
  • The video decoder 4 decrypts PES packets received from the De-MUX 3, and rendered the resulting uncompressed picture data to the video plane 5.
  • The video plane 5 is a plane memory for storing uncompressed picture data. The term “plane” refers to a memory area of a playback apparatus for storing one screen of pixel data. With provision of multiple planes, the playback apparatus can add the stored data on each plane pixel by pixel to produce one video output, whereby a plurality of graphics images are overlaid as one composite image. The resolution of the video plane 5 is 1920×1080. Picture data is stored on the video plane 5 as 16-bit pixel data expressed in YUV values.
  • The P-Graphics decoder 6 decodes a graphics stream read from the BD-ROM or HD, and renders the resulting uncompressed graphics to the presentation graphics plane 7. As the graphics stream is decoded and rendered, subtitles appear on screen.
  • The presentation graphics plane 7 is a memory having an area for storing one screen of uncompressed graphics. The resolution of this plane is 1920×1080, and uncompressed graphics are stored on the presentation graphics plane 7 as 8-bit pixel data expressed in index color. By converting the index color using the CLUT (Color Lookup Table), the uncompressed graphics stored on the presentation graphics plane 7 are presented.
  • The adder 8 overlays the uncompressed picture data (i) with the contents stored on the presentation graphics plane 7.
  • The I-Graphics decoder 9 decodes an IG stream read from the BD-ROM or HD, and renders the resulting uncompressed graphics to the interactive graphics plane 10.
  • To the interactive graphics plane 10, uncompressed graphics data decoded by the I-Graphics decoder 9 is rendered.
  • The adder 11 overlays the data on the interactive graphics plane 10 with the output of the adder 8 (i.e. with the interim video data maid by overlaying uncompressed picture data and the contents stored on the presentation graphics plane 7) to make the final video output.
  • The audio decoder 13 decodes PES packets received from the De-MUX 3 to output uncompressed audio data.
  • The CLUT unit 14 converts index colors of the uncompressed graphics stored on the presentation graphics plane 7 to Y, Cr, and Cb values.
  • The CLUT unit 15 converts index colors of the uncompressed graphics stored on the interactive graphics plane 10 to Y, Cr, and Cb values.
  • The PSR set 16 is a set of registers provided within the playback apparatus. The set of registers include 64 player status registers (PSRs) and 4,096 general purpose registers (GPRs). The 64 player status registers represent the current status of the playback apparatus. Specifically, the PSRs 5-8 represent the current playback point. The PRS 8 takes on a value ranging from 0 to 0xFFFFFFFF representing the current playback point (current PTM: Presentation Time) in 45 KHz. The PSR 11 represents the page_id of the currently active page (current page). The PSR 10 represents the button in the selected state (current button) on the current page.
  • The system clock 17 generates clock signals. The De-MUX 3 performs the conversion of PES packet in synchronism with the clock signals.
  • The user_timeout timer 18 is set to the value of user_time_out_duration field at the time specified by the PTS (DSn[ICS]). When the user_time out_duration field value is set, the timer 18 starts to count down toward zero in system clock ticks. The timer 18 is reset back to the user_time_out_duration field value each time a user operation is received. If no user operation is received before the timer 18 reaches zero, the timer 18 times out.
  • The selection_timeout timer 19 is set to the value of selection_time_out_pts field at the time specified by the PTS (DSn[ICS]). When the selection_time_out_pts field value is set, the timer 19 starts to count down toward zero in system clock ticks. The timer 19 is reset back to the selection_time_out_pts field value each time a user operation is received. If no user operation is received before the timer 19 reaches zero, the timer 19 times out. When the timer 19 times out, it is known that the valid interaction period shown in FIG. 9 has ended.
  • The composition_timeout timer 20 is set to the value of compostion_time_out_pts field at the time specified by the DTS(DSn[ICS]). When the compostion_time_out_pts field value is set, the timer 20 starts to count down toward zero in system clock ticks. The timer 20 is not reset even if a user operation is received. When the timer 20 reaches to zero, i.e. the timer 20 times out, it is known that the Epoch END has been reached.
  • The effect_duration timer 21 is set to the value of effect_duration field at the time specified by the DTS (DSn[ICS]) if the DSn is an Epoch Start DS. When the effect_duration field value is set, the timer 21 starts to count down toward zero in system clock ticks. The timer 21 is not reset even if a user operation is received. When the timer 21 reaches zero, i.e. the timer 21 times out, it is known a next display composition of the animation effect needs to be displayed.
  • The operation receiving unit 22 receives a user operation made on the remote controller or the front panel of the playback apparatus, and outputs information representing the received user operation (the information is hereinafter referred to as “UO”) to the I-Graphics decoder 9.
  • The state control unit 23 sets desired values to the PSRs according to the instructions from the I-Graphics decoder 9. Such instructions may be given by way of (i) direct addressing or (ii) indirect addressing. In the case of direct addressing, the I-Graphics decoder 9 outputs an immediate value to be set to an appropriate register of the PSR set 16. In the case of indirect addressing, the I-Graphics decoder 9 outputs a notification of a change in the playback apparatus status or in the user preferences. On receiving such a notification, the status controller 23 determines the value to be set to reflect the change, and sets the thus determined value to an appropriate register of the PSR set 16. The status controller 23 determines the value by performing a “procedure when playback condition is changed” or “procedure when change is requested”. Hereinafter, a description is given to the procedures performed for setting the PSR 11 (current page) and the PSR 10 (current button).
  • FIG. 38A is a flowchart of the “procedure when playback condition is changed”, preformed with respect to the PSR 11. This procedure is performed to set the first page_info structure within an ICS to the PSR 11 (step S99).
  • FIG. 38B is a flowchart of the “procedure when change is requested”, preformed with respect to the PSR 11. Suppose that a user operation is received and a page having the page number X is requested to be presented. In response to the user operation, the status controller 23 judges whether X is a valid value (step S100). If X is a valid value, the PSR 11 is set to the value of X (step S101). If X is invalid, the PSR 11 remains unchanged (step S102).
  • The value of PSR 11 changes as described above. Turning now to the PSR 10, a description is given to the “procedure when playback condition is changed” and to the “procedure when change is requested”.
  • FIG. 39A is a flowchart of the “procedure when playback condition is changed”, performed with respect to the PSR 10.
  • First, the status controller 23 judges whether the default_selected_button_id_ref field value associated with the current page is valid (step S111). If the judgment in the step S111 results in YES, the default_selected_button_id_ref field value is set to the PSR 10 (step S112)
  • If the default selected button_id_ref field value is invalid, it is then judged whether the PSR 10 holds a valid value (step S113). If the PSR 10 is valid, the PSR 10 remains unchanged so that the currently held value is maintained (step S114). On the other hand, if the PSR 10 is invalid, the PSR 10 is set to the button_id_ref field value identifying the first button_info structure of the current page (step S115).
  • FIG. 39B is a flowchart of the “procedure when change is requested”, performed with respect to the PSR 10. Suppose that a user operation is made and a button having the button number X is requested to be selected. In response to the user operation, the status controller 23 judges Whether X is a valid button_id field value (step S116). If X is a valid value, the PSR 10 is set to the value of X (step S118). If X is an invalid value, the PSR 10 remains unchanged and thus the currently held value is maintained (step S117). With the above-described procedures, the PSRs 10 and 11 are ensured to hold current values at all times. This concludes the description about the internal structure of playback apparatus.
  • <Internal Structure of I-Graphics Decoder 9>
  • Next, a description is given to the internal structure of I-Graphics decoder 9, with reference to FIG. 40. As shown in the figure, the I-Graphics decoder is composed of a coded data buffer 33, a stream graphics processor 34, an object buffer 35, a composition buffer 36, and a graphics controller 37.
  • The coded data buffer 33 is a buffer for temporally storing an ICS, a PDS, and an ODS along with their DTS and PTS.
  • The stream graphics processor 34 decodes an ODS and outputs the resulting uncompressed graphics objects to the object buffer 35.
  • The object buffer 35 is a buffer for storing a plurality of uncompressed graphics objects (represented by the boxes in the figure), which are decoded by the stream graphics processor 34.
  • The composition buffer 36 is a memory for storing an ICS. The composition buffer 36 supplies to the graphics controller 37, the page_info structures provided within the ICS and the button_info structures provided within each page_info structure.
  • The graphics controller 37 decodes an ICS stored in the composition buffer 37 and assembles a display composition based on the decoded ICS. That is, the graphics controller 37 reads graphics objects from the object buffer 15 with reference to the page_info structure identified by the value of PSR 11 (current page_info structure), and renders the read graphics objects to the interactive graphics plane 10. The graphics objects read herein are a sequence of graphics objects specified by the normal_start_object_id_ref field value through the normal_end_object_id_ref field value of each button_info structure associated with the current page. With respect to the button_info structure specified by the PSR 10, a sequence of graphics objects specified by the selected_start_object_id_ref field value through the selected_end_object_id_ref field value is read from the object buffer 15. In FIG. 40, arrows bg1, bg2, bg3, and bg4 schematically represent the rendering of graphics objects performed by the graphics controller 37. As a result of the rendering, a page including the buttons O-A, O-B, O-C, and O-D is stored on the interactive graphics plane 10 and overlaid with the video.
  • This concludes the rough description of procedures to be performed by the graphics controller 37. Furthermore, the graphics controller 37 updates the display composition in response to the changes in the value of PSR 10 or PSR 11, the timeout of the timers 18-21, or a user operation received by the composition timeout timer 20. The procedures performed by the graphics controller 37 are shown in FIGS. 41-47 in detail.
  • FIG. 41 is a flowchart of the main routine of the procedure performed by the graphics controller 37. In the main routine, steps S1-S6 form a loop. In each iteration of the loop, judgments are made as to whether specific events have occurred (steps S1-S4), a display composition of the Multi-Page menu is updated to a subsequent display composition for animated presentation (step S5), and processing is performed according to a user operation (UO processing) (step S6). If any of the judgments in the steps S1-S4 results in YES, corresponding steps are performed and the processing returns to the main routine.
  • In the step S1, the judgment is made as to whether the current playback point has reached the point specified by the PTS (DSn[ICS]). If the judgment in the step S1 results in YES, the above described “procedure when playback condition is changed” is performed. As a result, Page[0], which is the 1st page of the Multi-Page menu, is set as the current page (step S7). Then, the user_timeout timer 18, the selection_timeout timer 19, and the composition_timeout timer 20 are all reset (step S8), and the current page is presented on screen (step S9). The processing then goes back to the step S1 for the next loop iteration.
  • In the step S2, the judgment is made as to whether the user_timeout timer 18 has timed out. If the judgment in the step S2 results in YES, a loop of steps S11-S14 is performed. In each iteration of the loop, an arbitrary Page[j] of the Multi-Page menu is removed from the screen (step S13), followed by playback of an out-effect animation provided for Page[j] (step S14). The loop is repeated for each page from Page[0] onward, defined within the ICS (steps S11 and S12).
  • In the step S3, the judgment is made as to whether the selection_timeout timer 19 has timed out. If the judgment in the step S3 results in YES, a button on the current page is automatically activated (auto-activate: steps S1). Then, the processing returns the step S1 for the next loop iteration.
  • In the step S4, the judgment is made as to whether the composition timeout timer 20 has timed out. If the judgment in the step S4 results in YES, Page[0], which is the first page of the Multi-Page menu, is removed from the screen (step S15), followed by playback of the out-effect animation provided for Page[0] (step S16). Then, the processing returns to the step S1 for the next loop iteration.
  • Now, a description is given to the first presentation of the multi-page menu. The first presentation is carried out through steps S15-S23 shown in FIG. 42, after the state control information 23 performs the above-described “procedure when playback condition is changed” to set Page[0] as the current page. In the step S24, the in-effect animation for the current page is executed. In the step S25 that follows, a current button setting procedure is performed. The current button is specified by the PSR 10. Thus, in the step S25, the “procedure when playback condition is changed” is performed with respect to the PSR 10, so that the current button is determined. Then, the processing moves onto the steps S17-S22.
  • The steps S17 and S22 form a loop which is repeated for each button_info structure provided in the current page (steps S17 and S18). The button_info structure to be processed in the current iteration of the loop is designated as button_info (p).
  • In the step S19, a judgment is made as to whether the button_info (p) is associated with the current button. If the judgment in the step S19 results in YES, the step S20 is performed next. If not, the step S21 is performed next.
  • In the step S21, among the graphics objects on the object buffer 35, one specified by the normal_start_object_id_ref field value is designated as the graphics object (p).
  • In the step S20, among the graphics objects on the object buffer 35, one specified by the selected_start_object_id_ref field value is designated as the graphics object (p).
  • Once the graphics object (p) is designated in the step S20 or S21, the graphics object (p) is rendered to the interactive graphics plane 10 at the position specified by the button_horizontal_position and button_vertical_position field values (step S22). By repeating the above steps for each button_info structures provided in the current page, among a plurality of graphics objects associated with each button state, the first graphics object is rendered for each button to the interactive graphics plane 10. After repeating the above steps, the CLUT unit 15 is set so as to present the current page using the pallet data specified by the pallet_id_ref field value of the current page (step S23). This concludes the description of the flowchart shown in FIG. 42.
  • FIG. 43 is a flowchart of the procedure for playback of an in-effect animation. In the flowchart, a variable t denotes an arbitrary one of display compositions within the effect_sequence defined for the in-effect. In addition, a variable u denotes an arbitrary one of a composition_object out of composition_objects used for the effect(t). As shown in the flowchart, first of all, the variables t and u are initialized to “0” (step S30). Then, the effect_duration_timer 21 is set to the value of the effect (t).effect_duration field (step S31), and the CLUT units 14 and 15 are set so as to use the pallet data identified by the effect (t).pallet_id_ref field value for presentation (step S32). Then, a loop of steps S33-S40 is performed. The loop is a two-level loop. The first-level loop (steps S33-S37) uses the variable u as a control variable, whereas the second-level loop (step S33-S40) uses the variable t as a control variable.
  • In the first-level loop, a judgment is made as to whether the object_cropped_flag field provided within the composition_object (u) structure associated with the effect (t) is set to “0” (step S33). If the field is set to “0”, no graphics object is presented on screen for composition_object (u) (step S35). If the field is set to “1”, on the other hand, the object is cropped according to the object_cropping_horizontal_position, object_cropping_vertical_position, cropping width, and cropping_height fields of the composition_object (u). The cropped part of the graphics object is then rendered within the window identified by the window_id_ref field of the composition_object(u), at the position specified by the composition_object_horizontal_position and the composition_object_vertical_position fields within the composition_object(u) (step S34). Then, the variable u is incremented by “1”. The first-level loop is repeated until the variable u is equal to the number_of_composition_object field value.
  • In the second-level loop, for each iteration of the first-level loop (steps S33-S37), the following steps are performed. After the effect_duration times out (step S38), the variable t is incremented by “1”, and the variable u is initialized to “0” (steps S39). The second-level loop is repeated until the variable t is equal to the number_of_effect (t) field value (step S40). This concludes the description of the procedure for playback of In-effect. Note that no description is given to the procedure for playback of an out-effect, because it is basically the same to the above-described procedure.
  • Immediately after playback of the in-effect, a current page is presented and the first display is updated for animated presentation. The first display composition of the current page is presented by rendering to the interactive graphics plane 10, the graphics objects associated with the normal_start_object_id_ref and the selected_start_object_id_ref field values provided in each button_info structure. Each button is presented in animation by updating the button image on the interactive graphics plane 10 for each iteration of the loop in the main routine. Each button image is updated by overwriting the interactive graphics plane 10 with an arbitrary one of a sequence of graphics objects (qth graphics object) associated with the button. That is to say, each button is presented in animation by rendering a sequence of graphics objects associated with the normal_state_info and selected_state_info fields provided within each button_info structure, one by one to the interactive graphics plane 10. Note that the variable q is used to specify individual graphics objects associated with the normal_state_info and selected-state_info fields provided within each button_info structure.
  • Now, with reference to FIG. 44, a description is given to the procedure for display update for animated presentation.
  • In a step S41, a judgment is made as to whether the first display composition has been presented. If the first display composition is not yet presented, the processing returns to the main routine without performing any operation. If the first display composition has been presented, on the other hand, steps S42-S53 are performed. The steps S41-S55 form a loop in which the steps S44-S55 are repeated for each button_info structure in the ICS (steps S42 and S43).
  • In the step S44, the variable q is set to the value of variable animation (p) corresponding to the button_info (p) structure. As a result, the variable q represents the current frame number of the button_info(p).
  • In the step S45, a judgment is made as to whether the button_info(p) corresponds to the button currently in the selected state (i.e. the current button).
  • If the button_info(p) does not correspond to the current button, the variable q is added to the normal_start_object_id_ref in the button_info(p) to obtain an ID(q) (step S46).
  • If the button_info (p) corresponds to the current button, the step S47 is performed.
  • In the step S47, a judgment is made as to whether the current button is in the activated state. If the judgment in the step S47 results in YES, the variable q is added to the activated_start_object_id_ref field value in the button_info(p) to obtain an ID(q) (step S54). Then, one of the navigation commands associated with the button_info (p) is executed (step S55).
  • On the other hand, if the current button is not in the activated state, the variable q is added to the selected_start_object_id_ref field value in the button_info(p) to obtain an ID(q) (step S48).
  • Once the ID(q) is obtained, among the graphics objects stored on the object buffer 35, the one identified by the ID (q) is rendered to the interactive graphics plane 10. The rendering position on the interactive graphics plane 10 is specified by the button_horizontal_position and the button_vertical_position fields provided in the button info (p).
  • Through the loop iteration, the qth one of the graphics objects associated with the selected (or activated) state of the current button is rendered to the interactive graphics plane 10.
  • In the step S50, a judgment is made as to whether the sum of the normal_start_object_id_ref field value and the variable q has reached the normal_end_object_id_ref field value. If the sum has not reached the normal_end_object_id_ref field value, the variable q is incremented by “1” and the incremented variable q is set as the variable animation (p) (step S51). On the other hand, if the sum has reached the normal_end_object_id_ref field value, a judgment is made as to whether the repeat_flag field is set to “1” (step S52). If the repeat_flag field is set to “1”, the variable animation (p) is initialized to “0” (step S53). The above steps are repeated for each button_info structure within the ICS (steps S42 and S43), and then processing returns to the main routine.
  • Through the above-described steps of S41-S53, each button image presented on screen is updated with a new graphics object each time the main routine is performed. Thus, by repeating the main routine, the buttons images are presented in animation. When presenting the buttons in animation, the graphics controller 37 adjusts the duration for presenting each graphics object so as to maintain the animation_frame_rate_code. This concludes the description of the procedure for animated presentation.
  • In the step S47 described above, if it is judged that the current button is in the activated state, a navigations command associated with the current button needs to be executed in the step S55. FIG. 45 is a flowchart of the procedure for executing a navigation command. First, a judgment is made as to whether the navigation command is a SetButtonPageCommand (step S56). If the navigation command is not a SetButtonPageCommand, the navigation command is simply executed (step S57). On the other hand, if the navigation command is a SetButtonPageCommand, instructions are given to the state control unit 23 so as to designate the page number specified by the operand of the navigation command as the page number X as well as to set the button number specified by the operand of the navigation command to the PSR 10 (step S58). Then, the state control unit 23 executes the “procedure when change is requested”, with respect to the PSR 11 (step S59). As mentioned above, the PSR 11 holds a value representing the currently presented page (i.e. the current page). Thus, by performing the “procedure when change is requested” with respect to the PSR 11, the current page is determined. Then, the state control unit 23 performs the “procedure when playback condition is changed”, with respect to the PSR 10 (step S60). This concludes the description of the flowchart shown in FIG. 45.
  • This concludes the description of the procedure for executing a navigation command. As stated above, a navigation command is executed when an associated button is activated. The button state changes are presented through the following procedure in response to a user operation (UO). FIG. 46 shows the procedure for processing a UO. As shown in the flowchart, judgments are made in steps S61-S64 as to whether specific events have occurred. An occurrence of each event requires specific steps to be performed. The processing then returns to the main routine. Specifically, in the step S61, a judgment is made as to whether the UO_mask_table field is set to “1”. If the field is set to “1”, the processing returns to the main routine without performing any step.
  • In the step S62, the judgment is made as to whether any of the Move Up, Down, Left, and Right keys on the remote controller is pushed. At a push of any of the keys, the user_timeout timer 18 and the selection_timeout timer 19 are reset (step S65). Then, another button is designated as the new current button (step S66), followed by the judgment as to whether the auto_action_flag of the newly designated current button is set to “01” (step S67). If the auto_action_flag is not set to “01”, the processing returns to the main routine. On the other hand, if the auto_action_flag is set to “01”, the current button(i) is activated (step S69). Then, the variable animation(i) is set to “0” (step S70).
  • In the step S63, the judgment is made as to whether a numeric input is made. If a numeric input is made, the user_timeout timer 18 and the selection timeout timer 19 are reset (step S71), and another button is designated as the new current button (step S72). The processing then returns to the main routine.
  • In the step S64, a judgment is made as to whether an activate key on the remote controller is pushed. At a push of the activate key, the user_timeout timer 18 and the selection_timeout timer 19 are reset (step S68), and then the current button (i) is activated (step S69). After the button state transition, the variable animation(i) is set to “0” (step S70). Note that the step S66 of designating a new current button described above is performed by calling a subroutine shown in FIG. 47. Now, a description is given to the subroutine, with reference to the figure.
  • FIG. 47 is a flowchart of the current button change procedure. First, it is determined which of upper_button_id_ref, lower_button_id_ref, left_button_id_ref, and right_button_id_ref fields within the neighbor_info associated with the current button corresponds to the pushed key (step S75).
  • Here, let a button(Y) denote the current button, and let a button X denote the button specified by the one of upper-button_id_ref, lower_button_id_ref, left_button_id_ref, and right_button_id_ref fields (step S76). Then, the state control unit 23 performs the “procedure when change is requested” with respect to the PSR 10 (step S77). As a result, the PSR 10 is set to the value X.
  • Thereafter, the variables animation (X) and animation(Y) are set to “0” (step S78), and returns to the main routine.
  • As described above, in the UO processing, a button is activated in response to a UO. Yet, the button state transition to the activated state is made also when the selection_time_out_pts times out. Now, with reference to FIG. 48, a description is given to the procedure for auto-activation of a button upon timeout of the selection_time_out_pts.
  • FIG. 48 is a flowchart of the procedure for auto-activation 20 of a button. First, a judgment is made as to whether the default_activated_button_id_ref field is set to “00” or “FF” (step S91). If the field is set to “00”, the processing returns to the main routine without performing any step. On the other hand, if the default_activated_button_id_ref is set to “FF”, the current button(i) is activated (step S93). Thereafter, the variable animation(i) corresponding to the current button(i) is set to “0” (step S94), and returns to the main routine.
  • If the default_activated_button_id_ref field value is neither “00” nor “FF”, the button specified by the default_activated_button_id_ref field value is determined to be the new current button (i) (step S92). Thus, the button specified by the default_activated_button_id_ref field value receives the activated state (step S93). Thereafter, the variable animation (i) corresponding to the current button(i) is set to “0” (step S94), and returns to the main routine.
  • Through the above steps, the button in the selected state is automatically put into the activated state at an end of a predetermined duration. This concludes the description of the flowchart shown in FIG. 47.
  • <Turn On/Off of Graphics Display in Pop-Up U/I>
  • The description given above relates to the procedure performed by the I-Graphics decoder 9 in the case where the user_inderface_model shows the Always-On U/I. In the case where the Pop-Up U/I is used, the I-Graphics decoder 9 operates as follows. First, similarly to the Always-On U/I case, the I-Graphics decoder 9 performs the decoding to obtain decoded graphics objects on the object buffer 35. The I-Graphics decoder 9 then renders the graphics objects to the interactive graphics plane 10 by the time shown by the PTS(DSn[ICS]). Up to this point, the I-Graphics decoder 9 operates in a similar manner to the case where Always-On U/I is used. Yet, the processing performed thereafter differs in that the graphics controller 37 does not overlay the page stored on the interactive graphics plane 10, and thus the page is not presented. That is to say, graphics controller 37 “turns off” the page by not outputting the data stored on the interactive graphics plane 10 to the CLUT unit 15. With the page not presented on screen (i.e. turned off), the graphics controller 37 waits for a user operation for Pop-Up_on. In response to a user operation for Pop-Up_on, the graphics controller 37 outputs the graphics data stored on the interactive graphics plane 10 to the CLUT unit 15 where the graphics data is overlaid with the video data. As a result, the page is “turned on” or presented on screen. As above, the graphics controller 37 turns on and off the graphics display to realize Pop-Up display.
  • In the operations performed thereafter, there is no difference between the Always On U/I and the Pop-Up U/I. When any of the buttons on the main page is activated, a navigation command associated with the activated button is executed. When the selection_timeout timer 19 times out, the current button on the main page is automatically activated and a sub page is presented on screen.
  • When a user operation is made for Pop-Up_off, the graphics controller 37 tunes off all on-screen pages, thereby leaving no page presented on screen.
  • If no user operation is made for a certain period of time, the user_timeout timer 18 counts down toward zero. When the user_timeout timer 18 times out, the graphics controller 37 also turns off all on-screen pages, thereby leaving no page presented on screen. The turn on and turn off procedures described above are performed to realize the state transition shown in FIG. 12A.
  • As above, through the turn on and the turn off operations, Pop-Up U/I is implemented.
  • As described above, the sixth embodiment of the present invention realizes the industrial production of playback apparatuses supporting the BD-ROM having the data structures as described in the first to fifth embodiments. This helps for widespread use of the BD-ROMs according to the first to fifth embodiments.
  • Seventh Embodiment
  • A seventh embodiment of the present invention relates to a manufacturing process of the BD-ROM. FIG. 49 is a flowchart of the manufacturing process of the BD-ROMs according to the first to fifth embodiments.
  • The manufacturing process includes a material production step of recording video, audio, and the like (step S201), an authoring step of creating an application format using an authoring device (step S202), and a pressing step of creating an original master of the BD-ROM and performing stamping and bonding to complete the BD-ROM (step S203).
  • The BD-ROM authoring step includes the following steps S204-S213.
  • Now, a description is given to the steps S204-S213. In the step S240, control information, palette definition information, and graphics are generated. In the step S205, the control information, the palette definition information, and the graphics are converted to functional segments. In the step S206, a PTS of each ICS is set based on the display timing of a picture to be synchronized with. In the step S207, a DTS [ODS] and a PTS[ODS] are set based on the PTS[ICS]. In the step S208, a DTS[ICS] and a PTS[PDS] are set based on the DTS[ODS].
  • In the step S209, changes in occupancy of each buffer in the player model are graphed. In the step S210, a judgment is made as to whether the graphed changes satisfy the constraints of the player model. If the judgment results in the negative, the DTS and PTS of each functional segment are rewritten in the step S211. If the judgment results in the affirmative, a graphics stream is generated in the step S212, and the graphics stream is multiplexed with a video stream and an audio stream, which are separately generated, to form an AV Clip in the step S213. After this, the AV Clip is adapted to the Blue-ray Disc Read-Only Format, to complete the application format.
  • Modifications
  • Though the Present Invention has been Described by Way of the above embodiments, the present invention is not limited to those specific embodiments. The present invention may be embodied with any of the modifications (A) to (L) below. The invention recited in each of the claims of this application includes extension and generalization of the above embodiments and their modifications below. The degree of extension and generalization depends upon the state of the art in the technical field of the present invention at the time when the present invention was made.
  • (A) The above embodiments describe the case where the BD-ROM is used as the recording medium. Main features of the present invention, however, lie in a graphics stream recorded on the recording medium, which does not rely on physical characteristics of BD-ROMS. Therefore, the present invention is applicable to any recording medium that is capable of recording a graphics stream. Examples of such a recording medium include: an optical disc such as a DVD-ROM, a DVD-RAM, a DVD-RW, a DVD-R, a DVD+RW, a DVD+R, a CD-R, or a CD-RW; a magneto-optical disk such as a PD or an MO; a semiconductor memory card such as a CompactFlash card, a SmartMedia card, a Memory Stick card, a Multimediacard, or a PCMCIA card; a magnetic disk such as a flexible disk, SuperDisk, Zip, or Clik!; a removable hard disk drive such as ORB, Jaz, SparQ, SyJet, EZFley, or Microdrive, and a nonremovable hard disk drive.
  • (B) The above embodiments describe the case where the playback apparatus decodes an AV Clip on the BD-ROM and outputs the decoded AV Clip to the television. As an alternative, the playback apparatus may be equipped with only a BD drive, with the remaining construction elements being provided in the television. In this case, the playback apparatus and the television can be incorporated in a home network connected with an IEEE 1394 connector. The above embodiments describe the case where the playback apparatus is connected to the television, but the playback apparatus may instead be integrated with a display device. Also, the playback apparatus may include only the system LSI (integrated circuit) which carries out an essential part of the processing. The playback apparatus and the integrated circuit are both an invention described in this specification. Accordingly, regardless of whether the playback apparatus or the integrated circuit is concerned, an act of manufacturing a playback apparatus based on the internal structure of the playback apparatus described in the sixth embodiment is an act of practicing the present invention. Also, any act of assigning with charge (i.e. for sale) or without charge (i.e. as a gift), leasing, and importing the playback apparatus is an act of practicing the present invention. Likewise, an act of offering for assignment or lease of the reproduction apparatus using storefront displays, catalogs, or brochures is an act of practicing the present invention.
  • (C) Information processing using the programs shown in the flowcharts is actually realized using hardware resources. Accordingly, the programs which describe the procedure steps shown in the flowcharts are themselves an invention. The above embodiments all relate to the case where the programs are incorporated in the playback apparatus, but the programs can be used independently of the playback apparatus. Acts of practicing the programs include (1) an act of manufacturing, (2) an act of assigning with or without charge, (3) an act of leasing, (4) an act of importing, (5) an act of providing to the public via a bi-directional electronic communications network, and (6) an act of offering for assignment or lease using storefront displays, catalogs, or brochures.
  • (D) The time elements of the steps which are executed in a time series in each of the flowcharts can be regarded as the necessary elements of the present invention. This being so, a playback method shown by these flowcharts is an invention. If the processing shown in each flowchart is carried out by performing the steps in a time series so as to achieve the intended aim and the intended effect, it is considered as an act of practicing the playback method of the present invention.
  • (E) When recording an AV Clip on the BD-ROM, an extension header may be added to each TS packet in the AV Clip. The extension header is called a TP_extra_header, includes an arrival_time_stamp and a copy_permission_indicator, and has a data length of 4 bytes. TS packets with TP_extra_headers (hereafter “EX TS packets”) are grouped in units of 32 packets, and each group is written to three sectors. One group made up of 32EXTS packets has 6,144 bytes (=32×192), which is equivalent to a size of three sectors that is 6144 bytes (=2048×3). The 32 EX TS packets contained in the three sectors are called an “Aligned Unit”.
  • In a home network connected with an IEEE 1394 connector, the playback apparatus transmits an Aligned Unit in the following manner. The reproduction apparatus removes a TP_extra_header from each of the 32 EX TS packets in the Aligned Unit, encrypts the body of each TS packet according to the DTCP Specification, and outputs the encrypted TS packets. When outputting the TS packets, the playback apparatus inserts an isochronous packet between adjacent TS packets. A position where the isochronous packet is inserted is based on a time shown by an arrival_time_stamp of the TP_extra_header. The playback apparatus outputs a DTCP_descriptor, as well as the TS packets. The DTCP_descriptor corresponds to a copy_permission_indicator in the TP_extra_header. With the provision of the DTCP_descriptor indicating “copy prohibited”, it is possible to prevent, when using the TS packets in the home network connected via the IEEE 1394 connector, the TS packets from being recorded to other devices.
  • (F) The above embodiments relate to the case where an AV Clip of the Blu-ray. Disc Read-Only Format is used as a digital stream, but the present invention may also be embodied with a VOB (Video Object) of the DVD-Video Format or the DVD-Video Recording Format. The VOB is a program stream that complies with the ISO/IEC 13818-1 Standard and is obtained by multiplexing a video stream and an audio stream. Also, the video stream in the AV Clip may be an MPEG4 video stream or a WMV video stream. Further, the audio stream in the AV Clip may be a Linear PCM audio stream, a Dolby AC-3 audio stream, an MP3 audio stream, an MPEG-AAC audio stream, or a dts audio stream.
  • (G) The movie described in the above embodiments may be one obtained by encoding an analog image signal broadcast by analog broadcasting. Also, the movie may be stream data made up of a transport stream broadcast by digital broadcasting.
  • Alternatively, an analog/digital image signal recorded on a videotape may be encoded to obtain content. Also, an analog/digital image signal directly captured by a video camera may be encoded to obtain a content. A digital work distributed by a distribution server is applicable too.
  • (H) Graphics objects described in the above embodiments is run-length encoded raster data. Run-length encoding is used for compression/encoding of graphics objects, because the run-length encoding is suitable for compression and decompression of subtitles. Subtitles have the property that a continuous length of the same pixel value in a horizontal direction is relatively long. Therefore, by performing compression using run-length encoding, a high compression rate can be attained. In addition, run-length encoding reduces a load for decompression, and is therefore suitable for realizing decoding by software. Nevertheless, the use of run-length encoding for graphics objects is not essential to the present invention. For example, graphics objects may be PNG data. Also, graphics objects may be vector data instead of raster data. Further, graphics objects may be transparent patterns.
  • (I) The transfer rate Rc may be determined so that clearing and rendering of the graphics plane complete within a vertical blanking time. Suppose the vertical blanking time is 25% of 1/29.93 seconds. Then Rc is determined to be 1 Gbps. By determining Rc in this way, graphics can be displayed smoothly, which achieves a significant effect on the practical use.
  • Also, writing in sync with line scan can be used together with writing within a vertical blanking time. This ensures smooth presentation if the transfer rate Rc is 256 Mbps.
  • (J) The above embodiments relate to the case where the playback apparatus is provided with the graphics plane. Alternatively, the playback apparatus may include a line buffer for storing uncompressed pixels of one line. Since conversion to an image signal is performed for each horizontal row (line), conversion to an image signal can equally be performed with the line buffer.
  • (K) In the case of a merge point of multiple playback paths, a different button needs to be selected as default depending on a playback path having been taken. Thus, it is desirable to define such playback control in the static scenario that a value unique to each playback path is stored to a register of the playback apparatus at the time when the playback path is actually taken. The playback procedure may be set to put the button specified by the register value to the selected state. In this way, a different button is put to the selected state depending on a playback path having been taken.
  • (L) The graphics plane described in the sixth embodiment preferably has a double buffer structure. If the graphics plane has a double buffer structure, even when large-sized graphics data needs to be rendered, which requires a time corresponding to several frames, each display composition may be instantaneously presented by switching between two buffers. Thus, a double buffer structure is effective when the menu of a full-screen size needs to be presented.
  • INDUSTRIAL APPLICABILITY
  • The playback apparatus according to the present invention is suitable for a personal use, for example in a home theater system. Since the above embodiments of the present invention disclose the internal structure, playback apparatuses having the internal structure can be produced on a large scale. Thus, the playback apparatus according to the present invention is industrially usable in its quality and has industrial applicability.

Claims (11)

1. A recording medium having recorded thereon a video stream and a graphics stream, wherein:
the graphic stream includes one or more interactive composition segments and object definition segments, each object definition segment defining a graphics object;
each interactive composition segment includes a plurality of pieces of page information;
the plurality of pieces of page information define display compositions for pages of a multi-page menu, and each page information includes one or more pieces of button information;
each piece of button information is used to present a graphics object as one state of a button material, so that each page of the multi-page menu is presented as an interactive display;
each interactive composition segment is encapsulated in a packet that is attached with a presentation time stamp and a decoding time stamp;
the decoding time stamp shows, on a playback time axis of the video stream, a time for a playback apparatus to start decoding the interactive composition segment; and
the presentation time stamp shows, on the playback time axis of the video stream, a time for the playback apparatus to first present a first page of the multi-page menu.
2. The recording medium according to claim 1, wherein:
each interactive composition segment belongs to a display set DSn, which is an n-th display unit in the graphics stream;
PTS(DSn[ICS]) denotes a value of a presentation time stamp attached to a packet containing an interactive composition segment belonging to DSn;
DTS(DSn[ICS]) denotes a value of a decoding time stamp attached to the packet containing the interactive composition segment belonging to DSn; and
the following equation is satisfied,

PTS(DSn[ICS])≧PTS(DSn[ICS])+DECODEDURATION(DSn)+TRANSFERDURATION(DSn),
where a value of DECODEDURATION(DSn) shows a longer one of (i) a duration taken to decode all object definition segments included in DSn and (ii) a duration taken to clear a screen, and
where a value of TRANSFERDURATION(DSn) shows a duration taken to transfer graphics objects corresponding to all button materials to be presented on the first page.
3. The recording medium according to claim 1, wherein:
the first page includes a plurality of button materials,
a specific one of the button materials is presented in a selected state and the other button materials are presented in a normal state; and
the value of TRANSFERDURATION(DSn) is calculated by dividing (i) a total size of a graphics object used to present the selected state of the specific button material and graphics objects used present the normal state of the other button materials by (ii) a predetermined transfer rate of the playback apparatus.
4. The recording medium according to claim 1, wherein:
the value of TRANSFERDURATION(DSn) is calculated by dividing a size T of the set of uncompressed graphics by a transfer rate of the playback apparatus; and
the size T is calculated by a size TA+a size B−a size NB, where
the size TA is a total size of graphics data associated with the normal state of each button material included in the first page,
the size B is a size of a button material LRG,
the size NB is a size of uncompressed graphics associated with the normal state of the button material LRG, and
the button material LRG is a button material of which uncompressed graphics is largest in data size, out of all of the button materials included in the first page.
5. A playback apparatus for playing back a video stream and a graphics stream, the playback apparatus comprising:
a video decoder operable to decode the video stream to obtain a moving picture; and
a graphics decoder operable to decode the graphics stream to obtain a first page and a second page,
wherein:
the graphic stream includes one or more interactive composition segments and object definition segments, each object definition segment defining a graphics object;
each interactive composition segment includes a plurality of pieces of page information;
the plurality of pieces of page information define display compositions for pages of a multi-page menu, and each page information includes one or more pieces of button information;
each piece of button information is used to present a graphics object as one state of a button material, so that each page of the multi-page menu is presented as an interactive display;
each interactive composition segment is encapsulated in a packet that is attached with a presentation time stamp and a decoding time stamp;
the graphics decoder is operable to (i) start the decoding when a current playback point reaches a point corresponding to the time shown by the decoding time stamp attached to the packet, and (ii) present the first page when the current playback point reaches a point corresponding to the time shown by the presentation time stamp attached to the packet.
6. The playback apparatus according to claim 5, further comprising:
a graphics plane operable to store graphics to be overlaid on the moving picture,
wherein:
the graphics decoder includes
a processor operable to decode graphics data included in the graphics stream, and
an object buffer operable to store uncompressed graphics obtained by the decoding by the processor;
each interactive composition segment belongs to a display set DSn, which is an n-th display unit in the graphics stream;
PTS(DSn[ICS]) denotes a value of a presentation time stamp attached to a packet containing an interactive composition segment belonging to DSn;
DTS(DSn[ICS]) denotes a value of a decoding time stamp attached to the packet containing the interactive composition segment belonging to DSn; and
the following equation is satisfied,

PTS(DSn[ICS])≧PTS(DSn[ICS])+DECODEDURATION(DSn)+TRANSFERDURATION(DSn),
where a value of DECODEDURATION(DSn) shows a longer one of (i) a duration taken to decode all object definition segments included in DSn and (ii) a duration taken to clear the graphics plane, and
where a value of TRANSFERDURATION(DSn) shows a duration taken to transfer graphics objects corresponding to all button materials to be presented on the first page.
7. The playback apparatus according to claim 5, further comprising:
the first page includes a plurality of button materials,
a specific one of the button materials is presented in a selected state and the other button materials are presented in a normal state; and
the value of TRANSFERDURATION(DSn) is calculated by dividing (i) a total size of a graphics object used to present the selected state of the specific button material and graphics objects used present the normal state of the other button materials by (ii) a predetermined transfer rate of the playback apparatus.
8. The playback apparatus according to claim 5, further comprising:
the value of TRANSFERDURATION(DSn) is calculated by dividing a size T of the set of uncompressed graphics by a transfer rate of the playback apparatus; and
the size T is calculated by a size TA+a size B−a size NB, where
the size TA is a total size of graphics data associated with the normal state of each button material included in the first page,
the size B is a size of a button material LRG,
the size NB is a size of uncompressed graphics associated with the normal state of the button material LRG, and
the button material LRG is a button material of which uncompressed graphics is largest in data size, out of all of the button materials included in the first page.
9. A method for recording onto a recording medium, the method comprising:
generating application data; and
recording the application data onto the recording medium, wherein:
the application data includes a video stream and a graphics stream;
the graphic stream includes one or more interactive composition segments and object definition segments, each object definition segment defining a graphics object;
each interactive composition segment includes a plurality of pieces of page information;
the plurality of pieces of page information define display compositions for pages of a multi-page menu, and each page information includes one or more pieces of button information;
each piece of button information is used to present a graphics object as one state of a button material, so that each page of the multi-page menu is presented as an interactive display;
each interactive composition segment is encapsulated in a packet that is attached with a presentation time stamp and a decoding time stamp;
the decoding time stamp shows, on a playback time axis of the video stream, a time for a playback apparatus to start decoding the interactive composition segment; and
the presentation time stamp shows, on the playback time axis of the video stream, a time for the playback apparatus to first present a first page of the multi-page menu.
10. A computer program used for enabling a computer to play back a video stream and a graphics stream, comprising code operable to cause the computer to:
decode the video stream to obtain a moving picture; and
decode the graphics stream to obtain a first page and a second page,
wherein:
the graphic stream includes one or more interactive composition segments and object definition segments, each object definition segment defining a graphics object;
each interactive composition segment includes a plurality of pieces of page information;
the plurality of pieces of page information define display compositions for pages of a multi-page menu, and each page information includes one or more pieces of button information;
each piece of button information is used to present a graphics object as one state of a button material, so that each page of the multi-page menu is presented as an interactive display;
each interactive composition segment is encapsulated in a packet that is attached with a presentation time stamp and a decoding time stamp; and
in said decoding of the graphics stream, the decoding is started when a current playback point reaches a point corresponding to the time shown by the decoding time stamp attached to the packet, and the first page is presented when the current playback point reaches a point corresponding to the time shown by the presentation time stamp attached to the packet.
11. A method for playing back a video stream and a graphics stream, the method comprising:
decoding the video stream to obtain a moving picture; and
decoding the graphics stream to obtain a first page and a second page,
wherein:
the graphic stream includes one or more interactive composition segments and object definition segments, each object definition segment defining a graphics object;
each interactive composition segment includes a plurality of pieces of page information;
the plurality of pieces of page information define display compositions for pages of a multi-page menu, and each page information includes one or more pieces of button information;
each piece of button information is used to present a graphics object as one state of a button material, so that each page of the multi-page menu is presented as an interactive display;
each interactive composition segment is encapsulated in a packet that is attached with a presentation time stamp and a decoding time stamp; and
in said decoding of the graphics stream, the decoding is started when a current playback point reaches a point corresponding to the time shown by the decoding time stamp attached to the packet, and the first page is presented when the current playback point reaches a point corresponding to the time shown by the presentation time stamp attached to the packet.
US12/607,765 2003-11-12 2009-10-28 Recording medium, playback apparatus and method, recording method, and computer-readable program Abandoned US20100046932A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/607,765 US20100046932A1 (en) 2003-11-12 2009-10-28 Recording medium, playback apparatus and method, recording method, and computer-readable program

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US51931703P 2003-11-12 2003-11-12
US10/577,968 US7634739B2 (en) 2003-11-12 2004-11-12 Information recording medium, and apparatus and method for recording information to information recording medium
PCT/JP2004/017215 WO2005048261A1 (en) 2003-11-12 2004-11-12 Recording medium, playback apparatus and method, recording method, and computer-readable program
US12/607,765 US20100046932A1 (en) 2003-11-12 2009-10-28 Recording medium, playback apparatus and method, recording method, and computer-readable program

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
PCT/JP2004/017215 Division WO2005048261A1 (en) 2003-11-12 2004-11-12 Recording medium, playback apparatus and method, recording method, and computer-readable program
US10/577,968 Division US7634739B2 (en) 2003-11-12 2004-11-12 Information recording medium, and apparatus and method for recording information to information recording medium

Publications (1)

Publication Number Publication Date
US20100046932A1 true US20100046932A1 (en) 2010-02-25

Family

ID=34590392

Family Applications (4)

Application Number Title Priority Date Filing Date
US10/577,968 Active 2025-05-26 US7634739B2 (en) 2003-11-12 2004-11-12 Information recording medium, and apparatus and method for recording information to information recording medium
US12/607,751 Abandoned US20100050126A1 (en) 2003-11-12 2009-10-28 Recording medium, playback apparatus and method, recording method, and computer-readable program
US12/607,744 Active 2026-09-09 US8490017B2 (en) 2003-11-12 2009-10-28 Recording medium, playback apparatus and method, recording method, and computer-readable program implementing stream model information showing whether graphics stream is multiplexed or non-multiplexed with video system
US12/607,765 Abandoned US20100046932A1 (en) 2003-11-12 2009-10-28 Recording medium, playback apparatus and method, recording method, and computer-readable program

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US10/577,968 Active 2025-05-26 US7634739B2 (en) 2003-11-12 2004-11-12 Information recording medium, and apparatus and method for recording information to information recording medium
US12/607,751 Abandoned US20100050126A1 (en) 2003-11-12 2009-10-28 Recording medium, playback apparatus and method, recording method, and computer-readable program
US12/607,744 Active 2026-09-09 US8490017B2 (en) 2003-11-12 2009-10-28 Recording medium, playback apparatus and method, recording method, and computer-readable program implementing stream model information showing whether graphics stream is multiplexed or non-multiplexed with video system

Country Status (7)

Country Link
US (4) US7634739B2 (en)
EP (1) EP1683155A1 (en)
JP (4) JP4170362B2 (en)
KR (4) KR101129121B1 (en)
CN (4) CN101661767B (en)
TW (1) TWI382404B (en)
WO (1) WO2005048261A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060236218A1 (en) * 2003-06-30 2006-10-19 Hiroshi Yahata Recording medium, reproduction device, recording method, program, and reproduction method

Families Citing this family (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20050066265A (en) * 2003-12-26 2005-06-30 엘지전자 주식회사 Method for managing and reproducing a menu information of high density optical disc
KR20050066264A (en) * 2003-12-26 2005-06-30 엘지전자 주식회사 Method for managing and reproducing a menu information of high density optical disc
EP1713269B1 (en) 2004-01-13 2012-08-08 Panasonic Corporation Recording medium, reproduction device, recording method, program, and reproduction method
KR100782808B1 (en) * 2004-01-13 2007-12-06 삼성전자주식회사 Storage medium recording interactive graphic stream and reproducing apparatus thereof
US8190003B2 (en) * 2004-01-14 2012-05-29 Samsung Electronics Co., Ltd. Storage medium storing interactive graphics stream activated in response to user's command, and reproducing apparatus for reproducing from the same
KR100561417B1 (en) * 2004-02-09 2006-03-16 삼성전자주식회사 Information storage medium recorded interactive graphic stream for the transition of AV data reproducing state, and reproducing method and apparatus thereof
KR101135224B1 (en) * 2004-02-17 2012-04-23 파나소닉 주식회사 Recording medium, reproduction device, program, and reproduction method
US7852351B2 (en) * 2004-05-21 2010-12-14 Lsi Corporation Multiplexed button data system
KR100677129B1 (en) * 2004-09-03 2007-02-02 삼성전자주식회사 Storage medium recording interactive graphic stream and reproducing apparatus and method thereof
US20070006238A1 (en) * 2005-07-01 2007-01-04 Microsoft Corporation Managing application states in an interactive media environment
US8305398B2 (en) 2005-07-01 2012-11-06 Microsoft Corporation Rendering and compositing multiple applications in an interactive media environment
US7721308B2 (en) * 2005-07-01 2010-05-18 Microsoft Corproation Synchronization aspects of interactive multimedia presentation management
US8656268B2 (en) 2005-07-01 2014-02-18 Microsoft Corporation Queueing events in an interactive media environment
US20070006062A1 (en) * 2005-07-01 2007-01-04 Microsoft Corporation Synchronization aspects of interactive multimedia presentation management
US8799757B2 (en) 2005-07-01 2014-08-05 Microsoft Corporation Synchronization aspects of interactive multimedia presentation management
US7500175B2 (en) * 2005-07-01 2009-03-03 Microsoft Corporation Aspects of media content rendering
JP2007048348A (en) * 2005-08-08 2007-02-22 Toshiba Corp Information storage medium, information reproducing apparatus, information reproducing method, and information reproducing program
US20080238938A1 (en) * 2005-08-29 2008-10-02 Eklund Don Effects for interactive graphic data in disc authoring
US8650541B2 (en) * 2006-08-04 2014-02-11 Apple Inc. Graphical motion composition files and methods for formatting and organization thereof
JP4858059B2 (en) * 2006-10-02 2012-01-18 ソニー株式会社 Playback device, display control method, and display control program
RU2496157C2 (en) * 2008-09-30 2013-10-20 Панасоник Корпорэйшн Recording medium where 3d video is recorded, reproducing device for reproducing 3d video and system large-scale integrated circuit
US20100107090A1 (en) * 2008-10-27 2010-04-29 Camille Hearst Remote linking to media asset groups
US20110060993A1 (en) * 2009-09-08 2011-03-10 Classified Ventures, Llc Interactive Detailed Video Navigation System
US8605324B2 (en) * 2010-03-05 2013-12-10 Kabushiki Kaisha Toshiba Image processing system, image processing method, and computer readable recording medium storing program thereof
US8621351B2 (en) * 2010-08-31 2013-12-31 Blackberry Limited Methods and electronic devices for selecting and displaying thumbnails
KR101313592B1 (en) * 2012-04-13 2013-10-01 애니포인트 미디어 그룹 Computing device and method for streaming
US9423925B1 (en) * 2012-07-11 2016-08-23 Google Inc. Adaptive content control and display for internet media
KR20140031758A (en) * 2012-09-05 2014-03-13 삼성전자주식회사 Information storage medium recorded interactive graphic data for controlling menu of av data by using pointing device, and reproducing method and apparatus thereof
CN103870130B (en) * 2012-12-14 2017-03-22 北京新媒传信科技有限公司 Method and system for presenting interface in terminal
CN111899770B (en) * 2014-09-12 2022-07-08 松下电器(美国)知识产权公司 Non-transitory computer readable medium, reproduction apparatus, and reproduction method

Citations (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5241655A (en) * 1988-11-15 1993-08-31 Hitachi, Ltd. Menu presentation method and data processing apparatus
US5287182A (en) * 1992-07-02 1994-02-15 At&T Bell Laboratories Timing recovery for variable bit-rate video on asynchronous transfer mode (ATM) networks
US5600573A (en) * 1992-12-09 1997-02-04 Discovery Communications, Inc. Operations center with video storage for a television program packaging and delivery system
US5664133A (en) * 1993-12-13 1997-09-02 Microsoft Corporation Context sensitive menu system/menu behavior
US5680151A (en) * 1990-06-12 1997-10-21 Radius Inc. Method and apparatus for transmitting video, data over a computer bus using block transfers
US5907658A (en) * 1995-08-21 1999-05-25 Matsushita Electric Industrial Co., Ltd. Multimedia optical disk, reproduction apparatus and method for achieving variable scene development based on interactive control
US5918012A (en) * 1996-03-29 1999-06-29 British Telecommunications Public Limited Company Hyperlinking time-based data files
US5929850A (en) * 1996-07-01 1999-07-27 Thomson Consumer Electronices, Inc. Interactive television system and method having on-demand web-like navigational capabilities for displaying requested hyperlinked web-like still images associated with television content
US5984512A (en) * 1994-07-29 1999-11-16 Discovision Associates Method for storing video information
US5999225A (en) * 1995-08-02 1999-12-07 Sony Corporation Caption display method for using digital video system
US6067400A (en) * 1996-03-29 2000-05-23 Matsushita Electric Industrial Co., Ltd. Multimedia optical disc having improved interactive reproduction procedure, a reproduction apparatus and a method for such a disc
US6121968A (en) * 1998-06-17 2000-09-19 Microsoft Corporation Adaptive menus
US6240555B1 (en) * 1996-03-29 2001-05-29 Microsoft Corporation Interactive entertainment system for presenting supplemental interactive content together with continuous video programs
US6243086B1 (en) * 1997-05-29 2001-06-05 Matsushita Electric Industrial Co., Ltd. Reproduction apparatus
US6246402B1 (en) * 1996-11-07 2001-06-12 Sony Corporation Reproduction control data generating apparatus and method of same
US20010050920A1 (en) * 2000-03-29 2001-12-13 Hassell Joel Gerard Rate controlled insertion of asynchronous data into a synchronous stream
US20010055476A1 (en) * 2000-04-21 2001-12-27 Toshiya Takahashi Video processing method and video processing apparatus
US20020010715A1 (en) * 2001-07-26 2002-01-24 Garry Chinn System and method for browsing using a limited display device
US20020063797A1 (en) * 2000-10-27 2002-05-30 Shuntaro Aratani Television signal reception apparatus and method, and broadcast reception apparatus and method
US20020172496A1 (en) * 2000-09-29 2002-11-21 Masanori Gunji Recording/reproducing apparatus and picture recording reservation method of recording/reproducing apparatus
US6580442B1 (en) * 1999-12-01 2003-06-17 Ericsson Inc. Touch-based information processing device and method
US20030118243A1 (en) * 2001-09-18 2003-06-26 Ugur Sezer Largest magnitude indices selection for (run, level) encoding of a block coded picture
US20030138051A1 (en) * 2002-01-22 2003-07-24 Chen Sherman (Xuemin) System and method of transmission and reception of video using compressed differential time stamps
US6661438B1 (en) * 2000-01-18 2003-12-09 Seiko Epson Corporation Display apparatus and portable information processing apparatus
US20030229900A1 (en) * 2002-05-10 2003-12-11 Richard Reisman Method and apparatus for browsing using multiple coordinated device sets
US20040047588A1 (en) * 2002-03-27 2004-03-11 Tomoyuki Okada Package medium, reproduction apparatus, and reproduction method
US20040136698A1 (en) * 2002-07-10 2004-07-15 Mock Wayne E. DVD conversion for on demand
US20040160463A1 (en) * 2003-02-18 2004-08-19 Battles Amy E. System and method for displaying menu information in an electronic display
US20040165865A1 (en) * 2003-02-25 2004-08-26 Seo Kang Soo Method and apparatus for recording/reproducing graphic and subtitle data on/from a high-density recording medium
US20040168121A1 (en) * 2002-06-20 2004-08-26 Bellsouth Intellectual Property Corporation System and method for providing substitute content in place of blocked content
US20040184780A1 (en) * 2003-02-05 2004-09-23 Seo Kang Soo Recording medium having data structure for managing reproducton duration of still pictures recorded thereon and recording and reproducing methods and apparatuses
US20050091597A1 (en) * 2003-10-06 2005-04-28 Jonathan Ackley System and method of playback and feature control for video players
US7027516B2 (en) * 1998-06-29 2006-04-11 Pinnacle Systems, Inc. Method and apparatus for splicing
US7075994B2 (en) * 2000-04-28 2006-07-11 Sony Corporation Signal transmission method and signal transmission apparatus
US20060153532A1 (en) * 2003-07-03 2006-07-13 Mccrossan Joseph Recording medium, reporduction apparatus, recording method, integrated circuit, program and reporduction method
US20070098363A1 (en) * 2004-11-12 2007-05-03 Mccrossan Joseph Recording medium, playback apparatus and method, recording method, and computer-readable program
US7710396B2 (en) * 2003-05-01 2010-05-04 Thomson Licensing Multimedia user interface

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2116713A (en) * 1982-03-09 1983-09-28 Tecalemit Electronics Ltd Vibration monitoring apparatus
US5648813A (en) * 1993-10-20 1997-07-15 Matsushita Electric Industrial Co. Ltd. Graphical-interactive-screen display apparatus and peripheral units
JP3172643B2 (en) * 1994-11-14 2001-06-04 シャープ株式会社 Digital recording and playback device
AU698969B2 (en) * 1995-04-14 1998-11-12 Kabushiki Kaisha Toshiba Recording medium, device and method for recording data on the medium, and device and method for reproducing data from the medium
JPH0916665A (en) * 1995-07-04 1997-01-17 Funai Denki Kenkyusho:Kk Video cd ordering system
JPH09128408A (en) * 1995-08-25 1997-05-16 Hitachi Ltd Media for interactive recording and reproducing and reproducing device
DE19545882B4 (en) * 1995-12-08 2004-04-08 Yes International Ag System for displaying visual information
JP3345412B2 (en) * 1996-03-29 2002-11-18 松下電器産業株式会社 Recording method, reproducing apparatus, and reproducing method for recording medium with improved performance of interactive reproduction progress
JPH10105452A (en) * 1996-06-19 1998-04-24 Chishiki Joho Kenkyusho:Kk Method and device for executing software and software flow medium
EP0814419A2 (en) 1996-06-19 1997-12-29 Knowledge Information Research Institute Method of and device for executing a software and medium for distribution
JP3655433B2 (en) * 1997-06-20 2005-06-02 パイオニア株式会社 Computer-readable recording medium and information reproducing apparatus
JP4416846B2 (en) * 1997-08-22 2010-02-17 ソニー株式会社 Computer-readable recording medium recording menu control data, and menu control method and apparatus
US5949752A (en) 1997-10-30 1999-09-07 Wea Manufacturing Inc. Recording media and methods for display of graphic data, text, and images
JP3376265B2 (en) * 1997-12-25 2003-02-10 株式会社東芝 Object sharing system for multiple contents
US20030001880A1 (en) * 2001-04-18 2003-01-02 Parkervision, Inc. Method, system, and computer program product for producing and distributing enhanced media
KR20010086133A (en) 1999-10-21 2001-09-07 요트.게.아. 롤페즈 Playback apparatus and method for playback of a record carrier
FR2812237B1 (en) * 2000-07-26 2002-12-27 Secap DEVICE AND METHOD FOR CENTRALIZED PRINTING AND MAILING OF MAIL
WO2002079902A2 (en) * 2000-12-01 2002-10-10 Smartdisk Corporation Systems, methods and computer program products for managing multimedia content
JP3717880B2 (en) 2002-10-01 2005-11-16 パイオニア株式会社 Information recording medium, information recording apparatus and method, information reproducing apparatus and method, information recording / reproducing apparatus and method, computer program for recording or reproduction control, and data structure including control signal
DE602004031567D1 (en) * 2003-02-28 2011-04-07 Panasonic Corp Recording medium, reproducing apparatus, recording apparatus, program and reproduction method.
EP1465047A1 (en) 2003-04-03 2004-10-06 Deutsche Thomson-Brandt Gmbh Method for presenting menu buttons
CN1856075B (en) 2003-07-01 2011-06-08 日本先锋公司 Information recording apparatus and method, information reproducing apparatus and method
CN1833286B (en) 2003-07-01 2012-01-18 日本先锋公司 Information recording medium, information recording device and method, information reproduction device and method, information recording/reproduction device and method, computer program for recording
KR101102630B1 (en) 2003-11-12 2012-01-04 파나소닉 주식회사 Recording medium, playback apparatus and method, recording method, and computer-readable medium

Patent Citations (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5241655A (en) * 1988-11-15 1993-08-31 Hitachi, Ltd. Menu presentation method and data processing apparatus
US5680151A (en) * 1990-06-12 1997-10-21 Radius Inc. Method and apparatus for transmitting video, data over a computer bus using block transfers
US5287182A (en) * 1992-07-02 1994-02-15 At&T Bell Laboratories Timing recovery for variable bit-rate video on asynchronous transfer mode (ATM) networks
US5600573A (en) * 1992-12-09 1997-02-04 Discovery Communications, Inc. Operations center with video storage for a television program packaging and delivery system
US5664133A (en) * 1993-12-13 1997-09-02 Microsoft Corporation Context sensitive menu system/menu behavior
US5984512A (en) * 1994-07-29 1999-11-16 Discovision Associates Method for storing video information
US5999225A (en) * 1995-08-02 1999-12-07 Sony Corporation Caption display method for using digital video system
US5907658A (en) * 1995-08-21 1999-05-25 Matsushita Electric Industrial Co., Ltd. Multimedia optical disk, reproduction apparatus and method for achieving variable scene development based on interactive control
US6067400A (en) * 1996-03-29 2000-05-23 Matsushita Electric Industrial Co., Ltd. Multimedia optical disc having improved interactive reproduction procedure, a reproduction apparatus and a method for such a disc
US5918012A (en) * 1996-03-29 1999-06-29 British Telecommunications Public Limited Company Hyperlinking time-based data files
US6240555B1 (en) * 1996-03-29 2001-05-29 Microsoft Corporation Interactive entertainment system for presenting supplemental interactive content together with continuous video programs
US5929850A (en) * 1996-07-01 1999-07-27 Thomson Consumer Electronices, Inc. Interactive television system and method having on-demand web-like navigational capabilities for displaying requested hyperlinked web-like still images associated with television content
US6246402B1 (en) * 1996-11-07 2001-06-12 Sony Corporation Reproduction control data generating apparatus and method of same
US6243086B1 (en) * 1997-05-29 2001-06-05 Matsushita Electric Industrial Co., Ltd. Reproduction apparatus
US6121968A (en) * 1998-06-17 2000-09-19 Microsoft Corporation Adaptive menus
US7027516B2 (en) * 1998-06-29 2006-04-11 Pinnacle Systems, Inc. Method and apparatus for splicing
US6580442B1 (en) * 1999-12-01 2003-06-17 Ericsson Inc. Touch-based information processing device and method
US6661438B1 (en) * 2000-01-18 2003-12-09 Seiko Epson Corporation Display apparatus and portable information processing apparatus
US20010050920A1 (en) * 2000-03-29 2001-12-13 Hassell Joel Gerard Rate controlled insertion of asynchronous data into a synchronous stream
US20010055476A1 (en) * 2000-04-21 2001-12-27 Toshiya Takahashi Video processing method and video processing apparatus
US7075994B2 (en) * 2000-04-28 2006-07-11 Sony Corporation Signal transmission method and signal transmission apparatus
US20020172496A1 (en) * 2000-09-29 2002-11-21 Masanori Gunji Recording/reproducing apparatus and picture recording reservation method of recording/reproducing apparatus
US20020063797A1 (en) * 2000-10-27 2002-05-30 Shuntaro Aratani Television signal reception apparatus and method, and broadcast reception apparatus and method
US20070300282A1 (en) * 2000-10-27 2007-12-27 Canon Kabushiki Kaisha Television signal reception apparatus and method, and broadcast reception apparatus and method
US20060225122A1 (en) * 2000-10-27 2006-10-05 Canon Kabushiki Kaisha Television Signal Reception Apparatus and Method, and Broadcast Reception Apparatus and Method
US20020010715A1 (en) * 2001-07-26 2002-01-24 Garry Chinn System and method for browsing using a limited display device
US20030118243A1 (en) * 2001-09-18 2003-06-26 Ugur Sezer Largest magnitude indices selection for (run, level) encoding of a block coded picture
US20030138051A1 (en) * 2002-01-22 2003-07-24 Chen Sherman (Xuemin) System and method of transmission and reception of video using compressed differential time stamps
US20040047588A1 (en) * 2002-03-27 2004-03-11 Tomoyuki Okada Package medium, reproduction apparatus, and reproduction method
US20030229900A1 (en) * 2002-05-10 2003-12-11 Richard Reisman Method and apparatus for browsing using multiple coordinated device sets
US20040168121A1 (en) * 2002-06-20 2004-08-26 Bellsouth Intellectual Property Corporation System and method for providing substitute content in place of blocked content
US20040136698A1 (en) * 2002-07-10 2004-07-15 Mock Wayne E. DVD conversion for on demand
US20040184780A1 (en) * 2003-02-05 2004-09-23 Seo Kang Soo Recording medium having data structure for managing reproducton duration of still pictures recorded thereon and recording and reproducing methods and apparatuses
US20040160463A1 (en) * 2003-02-18 2004-08-19 Battles Amy E. System and method for displaying menu information in an electronic display
US20040165865A1 (en) * 2003-02-25 2004-08-26 Seo Kang Soo Method and apparatus for recording/reproducing graphic and subtitle data on/from a high-density recording medium
US7710396B2 (en) * 2003-05-01 2010-05-04 Thomson Licensing Multimedia user interface
US20060153532A1 (en) * 2003-07-03 2006-07-13 Mccrossan Joseph Recording medium, reporduction apparatus, recording method, integrated circuit, program and reporduction method
US20050091597A1 (en) * 2003-10-06 2005-04-28 Jonathan Ackley System and method of playback and feature control for video players
US20070098363A1 (en) * 2004-11-12 2007-05-03 Mccrossan Joseph Recording medium, playback apparatus and method, recording method, and computer-readable program

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060236218A1 (en) * 2003-06-30 2006-10-19 Hiroshi Yahata Recording medium, reproduction device, recording method, program, and reproduction method
US20060282775A1 (en) * 2003-06-30 2006-12-14 Hiroshi Yahata Recording medium, reproduction apparatus, recording method, program, and reproduction method
US20060288290A1 (en) * 2003-06-30 2006-12-21 Hiroshi Yahata Recording medium, reproduction apparatus, recording method, program, and reproduction method
US20060288302A1 (en) * 2003-06-30 2006-12-21 Hiroshi Yahata Recording medium, reproduction apparatus, recording method, program, and reproduction method
US20080126922A1 (en) * 2003-06-30 2008-05-29 Hiroshi Yahata Recording medium, reproduction apparatus, recording method, program and reproduction method
US7716584B2 (en) * 2003-06-30 2010-05-11 Panasonic Corporation Recording medium, reproduction device, recording method, program, and reproduction method
US7913169B2 (en) 2003-06-30 2011-03-22 Panasonic Corporation Recording medium, reproduction apparatus, recording method, program, and reproduction method
US8006173B2 (en) 2003-06-30 2011-08-23 Panasonic Corporation Recording medium, reproduction apparatus, recording method, program and reproduction method
US8010908B2 (en) 2003-06-30 2011-08-30 Panasonic Corporation Recording medium, reproduction apparatus, recording method, program, and reproduction method
US8020117B2 (en) 2003-06-30 2011-09-13 Panasonic Corporation Recording medium, reproduction apparatus, recording method, program, and reproduction method

Also Published As

Publication number Publication date
JP4476338B2 (en) 2010-06-09
US20100046608A1 (en) 2010-02-25
EP1683155A1 (en) 2006-07-26
CN101661766A (en) 2010-03-03
WO2005048261A1 (en) 2005-05-26
KR20090121407A (en) 2009-11-25
KR101102639B1 (en) 2012-01-04
KR20090121406A (en) 2009-11-25
JP4170362B2 (en) 2008-10-22
CN101661766B (en) 2011-07-27
JP2007515025A (en) 2007-06-07
JP4500859B2 (en) 2010-07-14
TWI382404B (en) 2013-01-11
JP4476339B2 (en) 2010-06-09
US20060291810A1 (en) 2006-12-28
TW200527388A (en) 2005-08-16
KR20090121408A (en) 2009-11-25
KR101102679B1 (en) 2012-01-05
KR101076130B1 (en) 2011-10-21
KR20060115749A (en) 2006-11-09
KR101129121B1 (en) 2012-03-27
CN100550178C (en) 2009-10-14
CN101661789A (en) 2010-03-03
US7634739B2 (en) 2009-12-15
US20100050126A1 (en) 2010-02-25
JP2008251153A (en) 2008-10-16
CN101661789B (en) 2011-07-27
CN1879170A (en) 2006-12-13
CN101661767B (en) 2011-07-27
JP2008257846A (en) 2008-10-23
JP2008262678A (en) 2008-10-30
CN101661767A (en) 2010-03-03
US8490017B2 (en) 2013-07-16

Similar Documents

Publication Publication Date Title
US8490017B2 (en) Recording medium, playback apparatus and method, recording method, and computer-readable program implementing stream model information showing whether graphics stream is multiplexed or non-multiplexed with video system
JP4496170B2 (en) Recording method, playback device, program, and playback method
US8515238B2 (en) Reproduction device, recording method, program, and reproduction method
EP1683349B1 (en) Recording medium, playback apparatus and method, recording method, and computer-readable program
US8000580B2 (en) Recording medium, playback apparatus and method, recording method, and computer-readable program

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION