US7844166B2 - Reproduction device, video decoding device, synchronization/reproduction method, program and recording medium - Google Patents
Reproduction device, video decoding device, synchronization/reproduction method, program and recording medium Download PDFInfo
- Publication number
- US7844166B2 US7844166B2 US11/922,942 US92294206A US7844166B2 US 7844166 B2 US7844166 B2 US 7844166B2 US 92294206 A US92294206 A US 92294206A US 7844166 B2 US7844166 B2 US 7844166B2
- Authority
- US
- United States
- Prior art keywords
- decoded
- head picture
- time stamp
- video
- picture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims description 17
- 230000001360 synchronised effect Effects 0.000 claims description 33
- 230000008859 change Effects 0.000 claims description 4
- 239000000872 buffer Substances 0.000 description 28
- 238000010586 diagram Methods 0.000 description 13
- 239000013598 vector Substances 0.000 description 7
- 230000006835 compression Effects 0.000 description 5
- 238000007906 compression Methods 0.000 description 5
- 230000008569 process Effects 0.000 description 5
- 230000006870 function Effects 0.000 description 4
- 238000013139 quantization Methods 0.000 description 4
- 238000004364 calculation method Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 230000000717 retained effect Effects 0.000 description 2
- 230000006399 behavior Effects 0.000 description 1
- 230000003139 buffering effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000000153 supplemental effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/2368—Multiplexing of audio and video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/4302—Content synchronisation processes, e.g. decoder synchronisation
- H04N21/4307—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
- H04N21/43072—Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of multiple content streams on the same device
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/434—Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
- H04N21/4341—Demultiplexing of audio and video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/488—Data services, e.g. news ticker
- H04N21/4884—Data services, e.g. news ticker for displaying subtitles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/8547—Content authoring involving timestamps for synchronizing content
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/16—Analogue secrecy systems; Analogue subscription systems
- H04N7/162—Authorising the user terminal, e.g. by paying; Registering the use of a subscription channel, e.g. billing
- H04N7/163—Authorising the user terminal, e.g. by paying; Registering the use of a subscription channel, e.g. billing by receiver means only
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2537—Optical discs
- G11B2220/2562—DVDs [digital versatile discs]; Digital video discs; MMCDs; HDCDs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/8042—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/806—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal
- H04N9/8063—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal using time division multiplex of the PCM audio and PCM video signals
Definitions
- the present invention relates to a reproduction apparatus, a video decoding apparatus and a synchronized reproduction method for decoding a stream in which video and audio are multiplexed.
- a media player and a personal computer that can playback high-capacity optical disc media such as a digital versatile disc (DVD) or the like are in widespread use and movie contents of high quality stored in DVDs or the like have been supplied.
- a DVD recorder is also widely used for recoding in DVDs the contents supplied via television broadcasts.
- a compression-encoding technology is essential for storing video and audio in a storage medium such as a DVD.
- a control mechanism for synchronized playback is necessary.
- MPEG Motion Picture Experts Group
- H.264/AVC Advanced Video Coding
- the video and audio coded streams are system-multiplexed with header information such as playback time information necessary for synchronized playback and then stored in a storage medium.
- a program stream of a MPEG-2 system is a standard of system multiplexing technology for the use of a storage medium.
- presentation time stamp (PTS) information is attached on a basis of an access unit that is a unit of playback of video and audio.
- PTS presentation time stamp
- a multiplexed stream such as a program stream of the MPEG-2 system is not playbacked from the head but randomly accessed to playback at any intermediate picture, the synchronized playback of the video and audio could not be sometimes properly performed. Since the video stream has been possibly subjected to interframe prediction encoding at the moving image encoding process layer in which a future frame is referenced, if the playback is done at any randomly accessed point, an order of decoding the frames and an order of displaying the same will not coincide. Therefore, the PTS information obtained at the system layer upon the random access cannot be used without change as a criterion time stamp for synchronized playback of the video and audio.
- the present invention has been achieved in view of the foregoing problems. It is thus general purpose of the present invention to provide a reproduction apparatus, a video decoding apparatus and a synchronized reproduction method for synchronizing video and audio, when a stream in which video and audio are multiplexed is randomly accessed to playback.
- a reproduction apparatus comprises: a demultiplexer which separates video data and audio data from a multiplexed stream in which a video stream and an audio stream are multiplexed; a video decoder which decodes the video data; an audio decoder which decodes the audio data; and a controller which supplies a synchronization signal for synchronizing with playback of the video to the audio decoder.
- the video decoder comprises a time stamp calculator which when the multiplexed stream is randomly accessed to playback at any picture, obtains from the video stream a playback time stamp to specify a playback time of a decoded head picture to be first decoded and then calculates a playback time stamp of a display head picture to be first displayed based on the playback time stamp of the decoded head picture.
- the controller supplies the calculated playback time stamp of the display head picture as the synchronization signal to the audio decoder.
- Another embodiment of the present invention also provides a reproduction apparatus.
- This apparatus comprises: a demultiplexer which separates video data, audio data and caption data from a multiplexed stream in which a video stream, an audio stream and a caption stream are multiplexed; a video decoder which decodes the video data; an audio decoder which decodes the audio data; a caption decoder which decodes the caption data; and a controller which supplies a synchronization signal for synchronizing with playback of the video to the audio decoder and the caption decoder.
- the video decoder comprises a time stamp calculator which when the multiplexed stream is randomly accessed to playback at any picture, obtains from the video stream a playback time stamp to specify a playback time of a decoded head picture to be first decoded and then calculates a playback time stamp of a display head picture to be first displayed based on the playback time stamp of the decoded head picture.
- the controller supplies the calculated playback time stamp of the display head picture as the synchronization signal to the audio decoder and the caption decoder.
- Yet another embodiment of the present invention provides a video encoding apparatus.
- This apparatus comprises: a decoder which decodes video data separated from a multiplexed stream in which a video stream and an audio stream; and a time stamp calculator which when the multiplexed stream is randomly accessed to playback at any picture, obtains from the video stream a playback time stamp to specify a playback time of a decoded head picture to be first decoded and then calculates a playback time stamp of a display head picture to be first displayed based on the playback time stamp of the decoded head picture.
- Yet another embodiment of the present invention provides a synchronized reproduction method.
- This method comprises: obtaining from a video stream a playback time stamp to specify a playback time of a decoded head picture to be first decoded, when a multiplexed stream in which the video stream and an audio stream are multiplexed is randomly accessed to playback at any picture; calculating a playback time stamp of a display head picture to be first displayed based on the playback time stamp of the decoded head picture; and decoding and playbacking the audio stream in synchronization with the calculated playback time stamp of the display head picture.
- Yet another embodiment of the present invention provides a program.
- This program enables a computer to execute: obtaining from a video stream a playback time stamp to specify a playback time of a decoded head picture to be first decoded, when a multiplexed stream in which the video stream and an audio stream are multiplexed is randomly accessed to playback at any picture; calculating a playback time stamp of a display head picture to be first displayed based on the playback time stamp of the decoded head picture; and outputting the calculated playback time stamp of the display head picture as a synchronization signal for synchronizing playback of the audio with playback of the video.
- This program may be provided as a component for an embedded device firmware for performing a basic control over hardware resources such as video and audio decoders.
- the firmware may be stored in a semiconductor memory such as a ROM or a flash memory inside the device.
- a computer readable storage medium having this program embodied therein may be provided for supplying the firmware or updating at least a part of the firmware.
- this program may be transmitted via a communication line.
- video and audio and so on can be accurately synchronized when a stream in which the video and the audio and so on are multiplexed is randomly accessed to playback.
- FIG. 1 is a block diagram of a multiplex encoding apparatus according to an embodiment.
- FIG. 2 is a block diagram of the video encoder of FIG. 1 .
- FIG. 3 is a diagram explaining a data structure of a program stream generated by the multiplex encoding apparatus of FIG. 1 .
- FIG. 4 is a block diagram of a synchronized reproduction apparatus according to an embodiment.
- FIG. 5 is a block diagram of the video decoder of FIG. 4 .
- FIG. 6 is a diagram explaining a procedure for calculating a display head PTS employed by the display head PTS calculator of FIG. 5 .
- FIG. 7 is a flowchart explaining the synchronized playback process employed by the synchronized reproduction apparatus of FIG. 4 .
- 10 storing buffer, 12 variable length decoder, 14 inverse quantizer, 16 inverse orthogonal transformer, 18 adder, 20 frame memory, 22 motion compensation unit, 24 frame sorting buffer, 26 D/A converter, 28 display head PTS calculator, 100 synchronized reproduction apparatus, 110 demultiplexer, 120 clock controller, 130 video decoder, 140 audio decoder, 150 caption decoder, 160 display output unit, 170 audio output unit, 200 multiplex encoding apparatus, 210 video encoder, 220 audio encoder, 230 caption encoder, 240 video packetizer, 250 audio packetizer, 260 caption packetizer, and 270 multiplexer.
- FIG. 1 is a block diagram of multiplex encoding apparatus 200 according to an embodiment.
- the figure shows a block diagram focused on functions. These function blocks may be realized in various forms such as hardware only, software only, or a combination thereof.
- the multiplex encoding apparatus 200 receives an input of video data, audio data and caption data.
- the audio and caption should be synchronized with the video.
- the video encoder 210 encodes the input video data and generates an encoded video stream according to, for instance, the H.264/AVC standard.
- the audio encoder 220 encodes the input audio data and outputs an encoded audio stream according to, for instance, the MPEG audio standard.
- the caption encoder 230 encodes the input caption data and outputs a coded caption stream.
- the coded streams thus encoded by the video encoder 210 , the audio encoder 220 and the caption encoder 230 are called elementary streams (ES).
- NAL Network Abstraction Layer
- VCL Video Coding Layer
- the respective elementary streams of the video, audio and caption are processed on a basis of a unit called NAL unit at the layer called NAL and then mapped to the system layer for performing the transmission and/or the storage of the information.
- the NAL units will be assembled into a unit called an access unit to enable an access by a unit of picture in the video stream.
- the audio stream and the caption stream will be configured so that an access can be done by an access unit which is composed of the NAL units.
- the multiplex encoding apparatus 200 of the present embodiment generates a multiplexed stream and stores it in a storage medium according to the program stream (PS) standard of a MPEG-2 system at the system layer.
- PS program stream
- the respective streams of the video, audio and caption are packetized for multiplexing.
- the video packetizer 240 packetizes the coded video stream output from the video encoder 210 into packetized elementary stream (PES) packets.
- the audio packetizer 250 packetizes the coded audio stream output from the audio encoder 220 into PES packets
- the caption packetizer 260 packetizes the coded caption stream output from the caption encoder 230 into PES packets.
- the multiplexer 270 multiplexes the PES packets of the video, audio and caption as a program stream of the MPEG-2 system and outputs the generated multiplexed stream.
- the generated multiplexed stream is recorded in a storage medium such as a DVD, a memory card or the like.
- FIG. 2 is a block diagram of the video encoder 210 .
- the video encoder 210 receives an input signal of an moving image by a unit of frame and encodes the frame images so as to output a coded stream of the moving image.
- An A/D converter 50 digitizes the analog input signal and supplies the digitalized data to a frame sorting buffer 52 .
- the frame sorting buffer 52 is a buffer for reordering and retaining the frames for interframe predictive coding. According to H.264/AVC, either a past frame or a future frame can be used as a reference image in interframe predictive coding. In addition, the number of frames available for reference images is not limited.
- the buffer is provided for storing the reordered frames and a reference image is appropriately supplied to a motion prediction/compensation unit 66 .
- the motion prediction/compensation unit 66 uses the past or future frame(s) stored in the frame sorting buffer as the reference image(s), the motion prediction/compensation unit 66 makes motion prediction on each of macroblocks of a P frame or a B frame, thereby generating motion vectors and a motion-compensated predicted image.
- the motion prediction/compensation unit 66 provides the motion vectors thus generated to a variable-length encoder 68 and provides the predicted image to a differentiator 54 .
- the differentiator 54 simply outputs the frame to an orthogonal transformer 56 . If the image frame is a P/B frame, the differentiator 54 calculates a difference from the predicted image supplied from the motion prediction/compensation unit 66 , and supplies the difference image to the orthogonal transformer 56 .
- the orthogonal transformer 56 performs discrete cosine transform (DCT) on the I frame or the difference image of the P/B frame and supplies the resultant DCT coefficients to a quantizer 58 .
- DCT discrete cosine transform
- the quantizer 58 quantizes the DCT coefficients and supplies the resultant to the variable-length encoder 68 .
- the quantizer 58 supplies the quantified DCT coefficients of the image frame to an inverse quantizer 60 .
- the inverse quantizer 60 inversely quantizes the supplied quantization data, and supplies the resultant to an orthogonal inverse transformer 62 .
- the orthogonal inverse transformer 62 performs inverse discrete cosine transform on the supplied inverse quantization data. This restores the coded image frame.
- the restored image frame is stored in a frame memory 64 .
- the motion prediction/compensation unit 66 references the reconstructed frame stored in the frame memory 64 as a frame to be encoded and makes motion compensation on the frame.
- variable-length encoder 68 performs variable-length coding on the DCT coefficients quantized by the quantizer 58 and stores the variable-length encoded and compressed image data in a storing buffer 70 .
- the variable-length encoder 68 performs variable-length coding on the quantized DCT coefficients of the difference image together with the motion vectors supplied from the motion prediction/compensation unit 66 .
- a rate controller 72 controls the frame rate of the frame images stored in the storing buffer 70 and adjusts the size of a quantization step to be employed in the quantizer 58 . Since the frame image is variable-length encoded, the amount of each frame data is not known until the frame is actually encoded. For this reason, the rate controller 72 detects the frame rate occurring when the frame images that have been stored in the storing buffer 70 are playbacked and adjusts to an appropriate amount of coded data by increasing or decreasing the quantization step size so that the coded data can be playbacked at a certain frame rate.
- the image compression information including the compressed frame images and motion vector information stored in the storing buffer 70 is output as a coded video stream and supplied to the video packetizer 240 .
- FIG. 3 is a diagram explaining a data structure of the program stream 300 generated by the multiplex encoding apparatus 200 .
- the program stream 300 of the MPEG-2 system is one in which the video stream, audio stream and caption stream are packetized into the PES packets and then multiplexed.
- a unit in which one or more PES packets are assembled is called a pack and the program stream 300 has a data structure in which the packs are sequentially arranged.
- a configuration is adopted in which one pack has one PES packet therein.
- the head pack includes a pack header 302 and a video PES packet 304 .
- the header of the video PES packet 304 includes PTS information 305 to provide a criterion of playback time of the video and the payload thereof includes the video data 306 .
- the PTS information 305 is included in the video PES packet of an access unit that is subject to random access. However, the PTS information is not included in the video PES packet of an access unit that is not subject to random access. Otherwise, the data size of the program stream 300 would grow. If it is acceptable that the data size of the program stream 300 would grow, the PTS information could be included in the header of the video packet of every access unit regardless of whether it is subject to random access or not.
- the second pack includes a pack header 308 and a video PES packet 310 .
- the third pack includes a pack header 312 and an audio PES packet 314 .
- the fourth pack includes a pack header 316 and a caption PES packet 318 .
- the headers of the audio PES packet 314 and the caption PES packet 318 also include, if appropriate, the PTS information to provide a criterion of playback time of the audio and the caption.
- the PTS information 305 is retrieved from the header of the video PES packet 304 of the randomly accessed access unit.
- the H.264/AVC standard defines a random access function using an instantaneous decoding refresh (IDR) picture.
- the IDR picture allows the subsequent pictures to be decoded without the information of pictures prior thereto and resets a buffer that has temporarily retained some reference pictures.
- the video encoder 210 of the multiplex encoding apparatus 200 records the PTS information for the IDR picture and obtains a PTS value from the IDR picture that is a decoded head picture, when the program stream 300 stored in a storage medium is randomly accessed.
- the PTS indicating playback time is added to an access unit that is a unit for playback for each of the video, audio and caption, in order to enable synchronized playback of the video, audio and caption. If the PTS of each access unit of the video, audio and caption is synchronized with each other, the synchronized playback of the video, audio and caption can be basically achieved.
- the synchronized reproduction apparatus 100 of the present embodiment internally calculates the PTS of the display head picture based on the PTS of the decoded head picture.
- FIG. 4 is a block diagram of a synchronized reproduction apparatus 100 according to an embodiment. These function blocks may also be realized in various forms such as hardware only, software only, or a combination thereof.
- the synchronized reproduction apparatus 100 receives an input of a stream in which the packets of the video, audio and caption are multiplexed and performs synchronized playback of the video, audio and caption.
- This multiplexed stream is one that has been generated and stored in a storage medium by the multiplexed encoding apparatus 200 .
- the multiplexed stream is randomly accessed and the packet data of the stream is read from the random access point and input to the synchronized reproduction apparatus 100 .
- a demultiplexer 110 separates the video packets, audio packets and caption packets from the input multiplexed stream and provides them to a video decoder 130 , an audio decoder 140 and a caption decoder 150 respectively.
- the demultiplexer 110 also extracts the PTS information from the video packet at the system layer and provides the obtained PTS to a clock controller 120 .
- This PTS is one obtained from the header of the access unit at the random access point of the video stream and it is a PTS of the decoded head picture (hereinafter, it is referred to as “decoded head PTS”).
- the clock controller 120 supplies a synchronization signal for synchronizing with the playback of the video to the audio decoder 140 and the caption decoder 150 and performs synchronization control. First, the clock controller 120 provides the decoded head PTS to the video decoder 130 .
- the video decoder 130 decodes the input video compression information and calculates a PTS of the display head picture (hereinafter, it is referred to as “display head PTS”) based on the decoded head PTS.
- the video decoder 130 supplies the display head PTS to the clock controller 120 .
- the video decoder 130 also supplies the decoded video data to a display output unit 160 .
- the clock controller 120 supplies the display head PTS given by the video decoder 130 as a synchronization signal to the audio decoder 140 and the caption decoder 150 .
- the audio decoder 140 and the caption decoder 150 decode an audio access unit and a caption access unit that have the same PTS value as the display head PTS. Then the audio decoder 140 provides the decoded audio data to an audio output unit 170 and the caption decoder 150 provides the decoded caption data to the display output unit 160 .
- the display output unit 160 displays the video data and the caption data on a display apparatus and the audio output unit 170 outputs the audio data from a speaker. Since the audio data and the caption data are reproduced in synchronization with the PTS of the display head picture, the audio and caption are properly synchronized with the video.
- FIG. 5 is a block diagram of the video decoder 130 .
- the video decoder 130 receives an input of image compression information including a coded video frame and motion vector information from the demultiplexer 110 and decodes the compressed image information so as to generate an output signal.
- a storing buffer 10 stores the image compression information input from the demultiplexer 110 .
- a variable-length decoder 12 performs variable decoding on the image compression information stored in the storing buffer 10 , supplies the decoded image data to an inverse quantizer 14 , and supplies motion vector information to a motion compensation unit 22 .
- the inverse quantizer 14 inversely quantizes the image data decoded by the variable-length decoder 12 , and supplies the resultant to an inverse orthogonal transformer 16 .
- the inverse orthogonal transformer 16 performs inverse discrete cosine transform (IDCT) on the DCT coefficients that are inversely quantized by the inverse quantizer 14 , thereby restoring the original image data.
- IDCT inverse discrete cosine transform
- the adder 18 simply stores the image data of the I frame into a frame sorting buffer 24 as well as stores it into a frame memory 20 as a reference image for generating a predicted image for a P/B frame.
- the adder 18 adds the difference image and the predicted image supplied from the motion compensation unit 22 .
- the adder 18 thereby reconstructs the original image data and stores it in the frame sorting buffer 24 .
- the motion compensation unit 22 generates a predicted image for the P/B frame by using the motion vector information supplied from the variable-length decoder 12 and the reference image(s) stored in the frame memory 20 .
- the generated predicted image is supplied to the adder 18 .
- the frame sorting buffer 24 is a buffer for reordering the stored frame images in the display order and the reordered frame images are supplied to a D/A converter 26 .
- the D/A converter converts the frame image to an analog signal and outputs it to the display output unit 160 .
- a display head PTS calculator 28 obtains the decoded head PTS from the clock controller 120 .
- the variable-length decoder 12 extracts the two parameters, a CPB removal delay time cpb_removal_delay and a DPB output delay time dpb_output_delay and provides them to the display head PTS calculator 28 .
- the display head PTS calculator 28 calculates the display head PTS based on the decoded head PTS, the CPB removal delay and the DPB output delay and provides the display head PTS to the clock controller 120 .
- the CPB removal delay time and the DPB output delay time are syntax parameters of the video stream encoded according to the H.264/AVC standard and they are information included in the picture timing supplemental enhancement information (SEI). These parameters are defined according to a pseudo decoder model adopted in the H.264/AVC. Hereinafter, this pseudo decoder model is explained.
- the video decoder 130 decodes the coded data while buffering the bit stream of the video into the storing buffer 10 and then playback the image. It is required herein that the bit stream should have been generated so as to avoid the storing buffer 10 from overflowing or underflowing. For this purpose, the behavior of the decoder is modeled using the pseudo decoder model in the H.264/AVC standard.
- the pseudo decoder model defines two buffers: a coded picture buffer (CPB) and a decoded picture buffer (DPB).
- the CPB is a buffer for storing a bit stream that is not yet subject to decoding processing.
- the CPB corresponds to the storing buffer 10 of the video decoder 130 of FIG. 5 .
- the DPB is a buffer for storing the pictures decoded by the decoder until they are displayed.
- the DPB corresponds to the frame sorting buffer 24 of the video decoder 130 .
- the data of the access units is first input to the CPB and then the data of each access unit is retrieved from the CPB at a time specified by a CPB removal time and decoded.
- the decoded picture is input to the DPB at the CPB removal time and stored.
- the picture stored in the DPB is output from the DPB at a time specified by a DPB output time and displayed.
- the data of the video stream is retrieved from the CPB, the data is decoded at the same instant and input to the DPB.
- the DPB is used for temporarily retaining reference pictures for motion compensation, for reordering the decoded pictures in the display order and for retaining the decoded pictures until their display output time.
- the CPB removal time and the DPB output time coincide, the picture that has been retrieved from the CPB and decoded will be output without being stored in the DPB. However, if the picture is a reference picture, it will be retained in the DPB so as to be referenced for motion compensation. If the DPB output time is greater than the CPB removal time, the picture that has been retrieved from the CPB and decoded will be stored in the DPB so as to be sorted in the display order and the output of the picture will be delayed until its DPB output time.
- the CPB removal time is a delay time from when the first picture starts to be decoded until when a picture is retrieved from the CPB.
- the DPB output delay time is a delay time from when a picture is retrieved from the CPB until the picture is output from the DPB.
- FIG. 6 is a diagram explaining a procedure for calculating the display head PTS employed by the display head PTS calculator 28 .
- the decoding order of the pictures is shown as the upper line and the display order of the pictures is shown as the lower line.
- IDR 2 is a decoded head picture that has been randomly accessed.
- the symbols of IDR, P and B indicate an IDR picture, P picture, B picture respectively and the number subsequent to each symbol indicates an order of display.
- the pictures are displayed in an order of B 0 , B 1 and IDR 2 .
- B 0 is a display head picture.
- PTS IDR2 is a decoded head PTS.
- dpb_output_delay IDR2 is a delay time from when the decoded head picture IDR 2 is retrieved from the CPB and starts to be decoded until when the decoded head picture IDR 2 is output from the DPB.
- cpb_removal_delay B0 is a delay time from when the decoded head picture IDR 2 is retrieved from the CPB and starts to be decoded until when the display head picture B 0 is retrieved from the CPB and starts to be decoded.
- dpb_output_delay B0 is a delay time from when the display head picture B 0 is retrieved from the CPB and starts to be decoded until when the display head picture B 0 is output from the DPB.
- the display head PTS calculator 28 calculates PTS B0 or the PTS value of the display head picture B 0 according to the following formula.
- PTS B0 PTS IDR2 ⁇ ( dpb _output_delay IDR2 ⁇ ( cpb _removal_delay B0 +dpb _output_delay B0 ))
- this formula shows that when dpb_output_delay IDR2 or the DPB output delay time of the decoded head picture IDR 2 is subtracted from PTS IDR2 of the decoded head picture IDR 2 and then cpb_removal_delay B0 or the CPB removal delay time of the display head picture B 0 and dpb_output_delay B0 or the DPB output delay time thereof are added to the result of the subtraction, PTS B0 of the display head picture B 0 will be obtained.
- the sum of cpb_removal_delay B0 or the CPB removal delay time of the display head picture B 0 and dpb_output_delay B0 or the DPB output delay time thereof is a delay time from when the decoded head picture IDR 2 is retrieved from the CPB and starts to be decoded until when the display head picture B 0 is output from the DPB.
- the PTS value of the display head picture does not need to be calculated. It is because the PTS value of the decoded head picture can be used without change. It cannot be determined until the decoding processing proceeds that the decoded head picture becomes a display head picture, however, if the first picture that has been displayed has a PTS value, it means that the picture is also a decoded head picture and the PTS value of the decoded head picture can be used without change. Thereby, if the display head picture coincides with the decoded head picture, the calculation of the display head PTS can be omitted and the processing cost can be reduced.
- FIG. 7 is a flowchart explaining the procedure for the synchronized playback process employed by the synchronized reproduction apparatus 100 .
- the clock controller 120 obtains the decoded head PTS from the packet header at the random access point in the coded video stream separated by the demultiplexer 110 and provides it to the video decoder 130 (S 10 ).
- the display head PTS calculator 28 obtains the DPB output delay time dpb_output_delay from the decoded head access unit of the video stream decoded by the variable-length decoder 12 (S 12 ).
- the display head PTS calculator 28 determines whether or not the decoded head access unit and the display head access unit coincide (S 14 ). If the decoded head access unit and the display head access unit coincide (Y of S 14 ), the value of the decoded head PTS is input into a variable for retaining the display head PTS (S 20 ).
- the display head PTS calculator 28 obtains the CPB removal delay time cpb_removal_delay and the DPB output delay time dpb_output_delay from the display head access unit (S 16 ).
- the display head PTS calculator 28 subtracts the DPB output delay time dpb_output_delay of the decoded head access unit from the decoded head PTS and adds the CPB removal delay time cpb_removal_delay and the DPB output delay time dpb_output_delay of the display head access unit to the result of the subtraction so as to calculate the display head PTS (S 18 ).
- the display head PTS calculator 28 outputs the display head PTS obtained at Step S 20 or Step S 18 and gives it to the clock controller 120 (S 22 ).
- the clock controller 120 provides the display head PTS obtained from the video decoder 130 to the audio decoder 140 and the caption decoder 150 , and the audio decoder 140 and the caption decoder 150 respectively decode the audio and the caption that synchronize with the display head PTS of the video (S 24 ).
- the PTS value of the decoded head picture can be obtained at a system layer and the PTS value of the display head picture can be obtained by calculation with reference to the CPB removal delay time and the DPB output delay time that are parameters of the video stream.
- the playback of the audio and caption can be performed by using the PTS value of the display head picture as a criterion and therefore the playback of the audio and caption can be accurately synchronized with the playback of the video.
- the PTS value of the display head picture can be obtained by calculation, all pictures do not have to have the PTS information therein and it will be sufficient that only pictures at the random access points have the PTS information. This reduces the overhead at the system layer and lessens a burden for synchronization control.
- the video stream and the audio stream are multiplexed at a system layer by using a program stream of the MPEG-2 system so as to generate a multiplexed stream for a storage use.
- any other standard than the MPEG-2 system can be used at the system layer.
- the video and audio may be multiplexed in accordance with H.264/AVC file formats.
- the moving image encoding is performed by H.264/AVC in the embodiment, however, any other standard than H.264/AVC can be used as long as the standard allows the parameters necessary for calculating the display head PTS based on the decoded head PTS to be obtained from the video stream.
- a procedure is explained for calculating the display head PTS based on the decoded head PTS of the video when a multiplexed stream stored in a storage medium is randomly accessed.
- This procedure can be also applied to a situation where a broadcast program is watched starting from a halfway point in a broadcast multiplexed stream. When the broadcast program is watched starting from a halfway point, it will happens that the decoded head picture does not coincide with the display head picture. Therefore, the audio and caption has to be synchronized with the calculated PTS of the display head picture.
- the synchronization control of the present invention can be thus applied not only to a multiplexed stream for a storage use but also to a multiplexed stream for the use of transmission media in which the stream is transmitted by a transport stream of the MPEG-2 system, real-time transport protocol (RTP) or the like.
- RTP real-time transport protocol
- the present invention is applicable to a filed of moving image reproduction processing.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Security & Cryptography (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Indexing, Searching, Synchronizing, And The Amount Of Synchronization Travel Of Record Carriers (AREA)
- Television Signal Processing For Recording (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
Abstract
Description
PTS B0 =PTS IDR2−(dpb_output_delayIDR2−(cpb_removal_delayB0 +dpb_output_delayB0))
Claims (13)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2005-194383 | 2005-07-01 | ||
JP2005194383A JP4311570B2 (en) | 2005-07-01 | 2005-07-01 | Playback apparatus, video decoding apparatus, and synchronous playback method |
PCT/JP2006/311921 WO2007004395A1 (en) | 2005-07-01 | 2006-06-14 | Reproduction device, video decoding device, synchronization/reproduction method, program, and recording medium |
Publications (2)
Publication Number | Publication Date |
---|---|
US20090214178A1 US20090214178A1 (en) | 2009-08-27 |
US7844166B2 true US7844166B2 (en) | 2010-11-30 |
Family
ID=37604270
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/922,942 Active US7844166B2 (en) | 2005-07-01 | 2006-06-14 | Reproduction device, video decoding device, synchronization/reproduction method, program and recording medium |
Country Status (4)
Country | Link |
---|---|
US (1) | US7844166B2 (en) |
EP (1) | EP1909278B1 (en) |
JP (1) | JP4311570B2 (en) |
WO (1) | WO2007004395A1 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140086341A1 (en) * | 2012-09-24 | 2014-03-27 | Qualcomm Incorporated | Coded picture buffer removal times signaled in picture and sub-picture timing supplemental enhancement information messages |
RU2646381C2 (en) * | 2012-09-21 | 2018-03-02 | Квэлкомм Инкорпорейтед | Indication and activation of sets of parameters for video encoding |
RU2646333C2 (en) * | 2012-08-09 | 2018-03-02 | Сан Пэтент Траст | Image decoding method, image encoding method, image decoding device, image encoding device and image encoding and decoding device |
RU2659748C2 (en) * | 2013-01-07 | 2018-07-03 | МАЙКРОСОФТ ТЕКНОЛОДЖИ ЛАЙСЕНСИНГ, ЭлЭлСи | Syntax and semantics for buffering information to simplify video concatenation |
US11361108B1 (en) | 2021-03-11 | 2022-06-14 | Tsj Technology, Inc. | Unidirectional communication system for public access to detained persons |
Families Citing this family (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8875199B2 (en) * | 2006-11-13 | 2014-10-28 | Cisco Technology, Inc. | Indicating picture usefulness for playback optimization |
US20090180546A1 (en) | 2008-01-09 | 2009-07-16 | Rodriguez Arturo A | Assistance for processing pictures in concatenated video streams |
US8416859B2 (en) | 2006-11-13 | 2013-04-09 | Cisco Technology, Inc. | Signalling and extraction in compressed video of pictures belonging to interdependency tiers |
US8873932B2 (en) | 2007-12-11 | 2014-10-28 | Cisco Technology, Inc. | Inferential processing to ascertain plural levels of picture interdependencies |
US8958486B2 (en) | 2007-07-31 | 2015-02-17 | Cisco Technology, Inc. | Simultaneous processing of media and redundancy streams for mitigating impairments |
US8804845B2 (en) | 2007-07-31 | 2014-08-12 | Cisco Technology, Inc. | Non-enhancing media redundancy coding for mitigating transmission impairments |
US8416858B2 (en) | 2008-02-29 | 2013-04-09 | Cisco Technology, Inc. | Signalling picture encoding schemes and associated picture properties |
US8886022B2 (en) | 2008-06-12 | 2014-11-11 | Cisco Technology, Inc. | Picture interdependencies signals in context of MMCO to assist stream manipulation |
US8705631B2 (en) | 2008-06-17 | 2014-04-22 | Cisco Technology, Inc. | Time-shifted transport of multi-latticed video for resiliency from burst-error effects |
US8699578B2 (en) | 2008-06-17 | 2014-04-15 | Cisco Technology, Inc. | Methods and systems for processing multi-latticed video streams |
US8971402B2 (en) | 2008-06-17 | 2015-03-03 | Cisco Technology, Inc. | Processing of impaired and incomplete multi-latticed video streams |
ES2546091T3 (en) | 2008-11-12 | 2015-09-18 | Cisco Technology, Inc. | Processing of a video program that has plural processed representations of a single video signal for reconstruction and broadcast |
CA2749064C (en) * | 2009-01-08 | 2015-02-17 | Lg Electronics Inc. | 3d caption signal transmission method and 3d caption display method |
US8326131B2 (en) | 2009-02-20 | 2012-12-04 | Cisco Technology, Inc. | Signalling of decodable sub-sequences |
US8782261B1 (en) | 2009-04-03 | 2014-07-15 | Cisco Technology, Inc. | System and method for authorization of segment boundary notifications |
US20120039582A1 (en) * | 2009-04-20 | 2012-02-16 | Koninklijke Philips Electronics N.V. | Verification and synchronization of files obtained separately from a video content |
US8949883B2 (en) | 2009-05-12 | 2015-02-03 | Cisco Technology, Inc. | Signalling buffer characteristics for splicing operations of video streams |
US8279926B2 (en) | 2009-06-18 | 2012-10-02 | Cisco Technology, Inc. | Dynamic streaming with latticed representations of video |
US20110222837A1 (en) * | 2010-03-11 | 2011-09-15 | Cisco Technology, Inc. | Management of picture referencing in video streams for plural playback modes |
ES2370218B1 (en) * | 2010-05-20 | 2012-10-18 | Universidad Carlos Iii De Madrid | PROCEDURE AND DEVICE FOR SYNCHRONIZING SUBTITLES WITH AUDIO IN DIRECT SUBTITLE. |
CN103081465B (en) * | 2010-08-18 | 2016-05-18 | 松下电器产业株式会社 | 3D encoding device |
US9131245B2 (en) | 2011-09-23 | 2015-09-08 | Qualcomm Incorporated | Reference picture list construction for video coding |
JP5791451B2 (en) * | 2011-09-29 | 2015-10-07 | キヤノン株式会社 | Encoding device and program |
CN102623032B (en) * | 2012-04-01 | 2014-09-17 | 深圳英飞拓科技股份有限公司 | Method for realizing synchronous playback of hard disk recorder |
BR112014026177B1 (en) * | 2012-04-23 | 2023-01-17 | Sun Patent Trust | IMAGE CODING METHOD, IMAGE DECODING METHOD, IMAGE CODING APPARATUS, IMAGE DECODING APPARATUS, AND IMAGE CODING AND DECODING APPARATUS |
JP2015226305A (en) * | 2014-05-30 | 2015-12-14 | 三菱電機株式会社 | Encoder |
WO2016132977A1 (en) * | 2015-02-20 | 2016-08-25 | ソニー株式会社 | Transmission device, transmission method, reception device, and reception method |
CN106211102B (en) * | 2016-09-14 | 2020-05-05 | Oppo广东移动通信有限公司 | Method and device for distinguishing and synchronizing chat information |
CN107864393A (en) * | 2017-11-17 | 2018-03-30 | 青岛海信电器股份有限公司 | The method and device that video is shown with captioning synchronization |
KR102437726B1 (en) * | 2020-10-12 | 2022-08-26 | 세종대학교 산학협력단 | Apparatus and method for storing snack culture contents |
CN114302215B (en) * | 2021-12-29 | 2023-09-29 | 北京奕斯伟计算技术股份有限公司 | Video data stream decoding system, method, electronic device and medium |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0668700A2 (en) | 1994-02-17 | 1995-08-23 | SANYO ELECTRIC Co., Ltd. | Audio/video decoding system |
JPH07284065A (en) | 1994-04-06 | 1995-10-27 | Sony Corp | Reproducing method for audio information and moving picture information |
JPH10271457A (en) | 1997-03-19 | 1998-10-09 | Sony Corp | Reproduction method and reproduction device for coded data |
JPH11149717A (en) | 1997-11-19 | 1999-06-02 | Toshiba Corp | Decoding processing method and apparatus |
GB2338383A (en) | 1998-06-12 | 1999-12-15 | Mitsubishi Electric Corp | Synchronising demultiplexed audio and video signals |
US6233695B1 (en) * | 1997-12-31 | 2001-05-15 | Lg Information & Communications, Ltd. | Data transmission control system in set top box |
JP2001238208A (en) | 2000-02-21 | 2001-08-31 | Matsushita Electric Ind Co Ltd | Moving image decoding method and moving image decoding device |
JP2001346166A (en) | 2000-03-29 | 2001-12-14 | Matsushita Electric Ind Co Ltd | Method and apparatus for reproducing compressed encoded data |
JP2002197808A (en) | 1999-02-18 | 2002-07-12 | Toshiba Corp | Stream information processing system |
US20030128294A1 (en) | 2002-01-04 | 2003-07-10 | James Lundblad | Method and apparatus for synchronizing audio and video data |
JP2004120155A (en) | 2002-09-25 | 2004-04-15 | Sony Corp | Information reproducer and method therefor, recording medium, and program |
US20070092224A1 (en) * | 2003-09-02 | 2007-04-26 | Sony Corporation | Content receiving apparatus, video/audio output timing control method, and content provision system |
-
2005
- 2005-07-01 JP JP2005194383A patent/JP4311570B2/en not_active Expired - Fee Related
-
2006
- 2006-06-14 EP EP20060766690 patent/EP1909278B1/en active Active
- 2006-06-14 US US11/922,942 patent/US7844166B2/en active Active
- 2006-06-14 WO PCT/JP2006/311921 patent/WO2007004395A1/en active Application Filing
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0668700A2 (en) | 1994-02-17 | 1995-08-23 | SANYO ELECTRIC Co., Ltd. | Audio/video decoding system |
JPH07284065A (en) | 1994-04-06 | 1995-10-27 | Sony Corp | Reproducing method for audio information and moving picture information |
JPH10271457A (en) | 1997-03-19 | 1998-10-09 | Sony Corp | Reproduction method and reproduction device for coded data |
JPH11149717A (en) | 1997-11-19 | 1999-06-02 | Toshiba Corp | Decoding processing method and apparatus |
US6233695B1 (en) * | 1997-12-31 | 2001-05-15 | Lg Information & Communications, Ltd. | Data transmission control system in set top box |
GB2338383A (en) | 1998-06-12 | 1999-12-15 | Mitsubishi Electric Corp | Synchronising demultiplexed audio and video signals |
JP2002197808A (en) | 1999-02-18 | 2002-07-12 | Toshiba Corp | Stream information processing system |
JP2001238208A (en) | 2000-02-21 | 2001-08-31 | Matsushita Electric Ind Co Ltd | Moving image decoding method and moving image decoding device |
JP2001346166A (en) | 2000-03-29 | 2001-12-14 | Matsushita Electric Ind Co Ltd | Method and apparatus for reproducing compressed encoded data |
US20030128294A1 (en) | 2002-01-04 | 2003-07-10 | James Lundblad | Method and apparatus for synchronizing audio and video data |
JP2004120155A (en) | 2002-09-25 | 2004-04-15 | Sony Corp | Information reproducer and method therefor, recording medium, and program |
US20070092224A1 (en) * | 2003-09-02 | 2007-04-26 | Sony Corporation | Content receiving apparatus, video/audio output timing control method, and content provision system |
Non-Patent Citations (2)
Title |
---|
Lu G J et al.: "Mechanisms of MPEG Stream Synchronization", Computer Communication Review, ACM , New York, NY, US, col. 24, No. 1, Jan. 1, 1994, pp. 57-67, XP000560498 ISSN: 0146-4833. |
Wiegand T: "Draft ITU-T Recommendation and Final Draft International Standard of Joint Video Specification (ITU-T Rec. H.264/ISO/IEC 14496-10 AVC)" Joint Video Team (JVT) of ISO/IEC MPEG & ITU-T VCEG (ISO-IEC JTC1/SC29/WG11 and ITU-T SG16 Q6), XX, XX, Apr. 2, 2003 page complete, XP002328607. |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
RU2646333C2 (en) * | 2012-08-09 | 2018-03-02 | Сан Пэтент Траст | Image decoding method, image encoding method, image decoding device, image encoding device and image encoding and decoding device |
RU2646381C2 (en) * | 2012-09-21 | 2018-03-02 | Квэлкомм Инкорпорейтед | Indication and activation of sets of parameters for video encoding |
RU2646382C2 (en) * | 2012-09-21 | 2018-03-02 | Квэлкомм Инкорпорейтед | Indication and activation of sets of parameters for video encoding |
US9479774B2 (en) | 2012-09-24 | 2016-10-25 | Qualcomm Incorporated | Buffering period and recovery point supplemental enhancement information messages |
US9654802B2 (en) | 2012-09-24 | 2017-05-16 | Qualcomm Incorporated | Sequence level flag for sub-picture level coded picture buffer parameters |
US9491456B2 (en) * | 2012-09-24 | 2016-11-08 | Qualcomm Incorporated | Coded picture buffer removal times signaled in picture and sub-picture timing supplemental enhancement information messages |
US9503753B2 (en) | 2012-09-24 | 2016-11-22 | Qualcomm Incorporated | Coded picture buffer arrival and nominal removal times in video coding |
TWI569626B (en) * | 2012-09-24 | 2017-02-01 | 高通公司 | Methods for decoding and encoding video data, video decoding and encoding devices and computer-storage media thereof |
TWI574549B (en) * | 2012-09-24 | 2017-03-11 | 高通公司 | Methods for decoding and encoding video data, video decoding and encoding devices and non-transitory computer-readable storage media for decoding and encoding video data |
US9648352B2 (en) | 2012-09-24 | 2017-05-09 | Qualcomm Incorporated | Expanded decoding unit definition |
US20140086341A1 (en) * | 2012-09-24 | 2014-03-27 | Qualcomm Incorporated | Coded picture buffer removal times signaled in picture and sub-picture timing supplemental enhancement information messages |
RU2627098C2 (en) * | 2012-09-24 | 2017-08-03 | Квэлкомм Инкорпорейтед | Independent from unit of access time of delete from buffer coded pictures at coding video |
US9479773B2 (en) * | 2012-09-24 | 2016-10-25 | Qualcomm Incorporated | Access unit independent coded picture buffer removal times in video coding |
CN104662916A (en) * | 2012-09-24 | 2015-05-27 | 高通股份有限公司 | Coded picture buffer removal times signaled in picture and sub-picture timing supplemental enhancement information messages |
US20140086332A1 (en) * | 2012-09-24 | 2014-03-27 | Qualcomm Incorporated | Access unit independent coded picture buffer removal times in video coding |
CN104662916B (en) * | 2012-09-24 | 2019-01-11 | 高通股份有限公司 | The method and apparatus for removing the time for decoded picture buffer |
RU2659748C2 (en) * | 2013-01-07 | 2018-07-03 | МАЙКРОСОФТ ТЕКНОЛОДЖИ ЛАЙСЕНСИНГ, ЭлЭлСи | Syntax and semantics for buffering information to simplify video concatenation |
US10313698B2 (en) | 2013-01-07 | 2019-06-04 | Microsoft Technology Licensing, Llc | Syntax and semantics for buffering information to simplify video splicing |
US11361108B1 (en) | 2021-03-11 | 2022-06-14 | Tsj Technology, Inc. | Unidirectional communication system for public access to detained persons |
Also Published As
Publication number | Publication date |
---|---|
EP1909278A1 (en) | 2008-04-09 |
WO2007004395A1 (en) | 2007-01-11 |
JP4311570B2 (en) | 2009-08-12 |
EP1909278B1 (en) | 2012-06-20 |
JP2007012218A (en) | 2007-01-18 |
US20090214178A1 (en) | 2009-08-27 |
EP1909278A4 (en) | 2009-11-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7844166B2 (en) | Reproduction device, video decoding device, synchronization/reproduction method, program and recording medium | |
US8503541B2 (en) | Method and apparatus for determining timing information from a bit stream | |
US6980594B2 (en) | Generation of MPEG slow motion playout | |
JP4769717B2 (en) | Image decoding method | |
KR20010022752A (en) | Trick play signal generation for a digital video recorder | |
US20160127728A1 (en) | Video compression apparatus, video playback apparatus and video delivery system | |
JP2004534484A (en) | Transform coding of video data stream | |
JP2005354361A (en) | Coding apparatus and coding method, decoding apparatus and decoding method, program and program recording medium, data recording medium and data structure, and reproducing apparatus | |
US20040103446A1 (en) | Audio-video multiplexed data generating apparatus, reproducing apparatus and moving video decoding apparatus | |
JP4092681B2 (en) | MPEG recording device, playback device, and transmission device | |
JP2000244872A (en) | Transcoder equipment | |
JP2009290768A (en) | Video processing apparatus and video processing method | |
KR100394013B1 (en) | Apparatus for transcoding video snap image | |
JP3469866B2 (en) | Method for changing the bit rate of a data stream of an encoded video picture | |
JP4569847B2 (en) | Data reconstruction device and data reconstruction method | |
JP4178521B2 (en) | Encoded video signal recording method and video signal encoding apparatus | |
JP2823806B2 (en) | Image decoding device | |
US9219930B1 (en) | Method and system for timing media stream modifications | |
JP2010062871A (en) | Method and device for converting compressed hd video of further high picture rate to compressed video of further low picture rate capable of responding to decoding by legacy hd video decoder using h.264 specification | |
JP2011078068A (en) | System for video transmission | |
KR20180019511A (en) | Systems and methods for inclusion of accompanying message data in a compressed video bitstream | |
JP2006074327A (en) | Re-coder for video data and video distribution system | |
JP2006311135A (en) | Moving image encoding device | |
JP2008010997A (en) | Information processing apparatus, information processing method, and semiconductor integrated circuit | |
Kou | Digital Video Compression Standards |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY COMPUTER ENTERTAINMENT INC., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TAKAHASHI, KUNIAKI;REEL/FRAME:022409/0151 Effective date: 20090220 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: SONY NETWORK ENTERTAINMENT PLATFORM INC., JAPAN Free format text: CHANGE OF NAME;ASSIGNOR:SONY COMPUTER ENTERTAINMENT INC.;REEL/FRAME:027445/0773 Effective date: 20100401 |
|
AS | Assignment |
Owner name: SONY COMPUTER ENTERTAINMENT INC., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SONY NETWORK ENTERTAINMENT PLATFORM INC.;REEL/FRAME:027449/0380 Effective date: 20100401 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552) Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |