[go: up one dir, main page]

CN106791764A - A kind of method and apparatus for realizing coding and decoding video - Google Patents

A kind of method and apparatus for realizing coding and decoding video Download PDF

Info

Publication number
CN106791764A
CN106791764A CN201611083136.9A CN201611083136A CN106791764A CN 106791764 A CN106791764 A CN 106791764A CN 201611083136 A CN201611083136 A CN 201611083136A CN 106791764 A CN106791764 A CN 106791764A
Authority
CN
China
Prior art keywords
video data
data stream
coded
deep
data flow
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201611083136.9A
Other languages
Chinese (zh)
Inventor
戴向东
王猛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nubia Technology Co Ltd
Original Assignee
Nubia Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nubia Technology Co Ltd filed Critical Nubia Technology Co Ltd
Priority to CN201611083136.9A priority Critical patent/CN106791764A/en
Publication of CN106791764A publication Critical patent/CN106791764A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/161Encoding, multiplexing or demultiplexing different image signal components
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/128Adjusting depth or disparity
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/40Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video transcoding, i.e. partial or full decoding of a coded input stream followed by re-encoding of the decoded output stream

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

A kind of method and apparatus for realizing coding and decoding video, the method includes:Original video data stream and deep video data flow are obtained by binocular platform;First coded treatment is carried out to original video data stream using two-dimensional video coding method and obtains the first coded video data stream;Second coded treatment is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream and obtains the second coded video data stream.By the scheme of the embodiment of the present invention, the first coded treatment is first carried out to original video data stream using two-dimensional video coding method and obtains the first coded video data stream, the second coded treatment is carried out to the first coded video data stream further according to deep video data flow and the first coded video data stream, so as to compatible two-dimensional video coded format, it is ensured that three-dimensional video-frequency can be decoded and played on two-dimensional video player.

Description

A kind of method and apparatus for realizing coding and decoding video
Technical field
Present document relates to but be not limited to video technique, espespecially a kind of method and apparatus for realizing coding and decoding video.
Background technology
Three-dimensional (3D) video technique is based on double pupil principle of parallax of people, and Same Scene but slightly difference are obtained by video camera Two images, show the left eye and right eye for giving people respectively, binocular parallax is formed, so as to make one to obtain the depth letter of scene Cease and experience third dimension.Three-dimensional video-frequency technology can truly reappear objective world scene, show the depth of scene Sense, stereovision and authenticity, are the important directions of current video techniques development.
Current three-dimensional video-frequency can only be played on specific player, and traditional two dimension (2D) video player then cannot Play three-dimensional video-frequency.
The content of the invention
The embodiment of the present invention proposes a kind of method and apparatus for realizing coding and decoding video, being capable of compatible two-dimensional video coding Form, it is ensured that three-dimensional video-frequency can be decoded and played on two-dimensional video player.
The embodiment of the present invention proposes a kind of device for realizing coding and decoding video, including:
Acquisition module, for obtaining original video data stream and deep video data flow by binocular platform;
First coding module, for carrying out the first coded treatment to original video data stream using two-dimensional video coding method Obtain the first coded video data stream;
Second coding module, for according to deep video data flow and the first coded video data stream to the first encoded video Data flow carries out the second coded treatment and obtains the second coded video data stream.
Optionally, the binocular platform includes binocular camera;The acquisition module specifically for:
First video data stream is obtained by the first camera of binocular camera, while by the second of the binocular camera Camera obtains the second video data stream;
Using first video data stream or second video data stream as the original video data stream;
The image of each frame first and second video data stream according to first video data stream is corresponding each The image of frame second calculates each frame depth image in the deep video data flow;By depth described in all range image integrations Video data stream.
Optionally, the binocular platform includes camera and the equipment for obtaining depth information;
The acquisition module specifically for:
The original video data stream is obtained by the camera, while by described for obtaining setting for depth information It is standby to obtain the deep video data flow.
Optionally, second coding module specifically for:
Obtain the key message in first coded video data stream;
The end that the deep video data flow and the key message write first coded video data stream is obtained To the 3rd coded video data stream;
Added for indicating whether the zone bit information with depth information at the 3rd coded video data stream end Obtain second coded video data stream.
Optionally, second coding module specifically for:
Obtain the key message in first coded video data stream;
The head that the deep video data flow and the key message write first coded video data stream is obtained To the 3rd coded video data stream;
Added for indicating whether the zone bit information with depth information in the 3rd coded video data stream head Obtain second coded video data stream.
Optionally, also include:
Second decoder module, it is corresponding with second coded treatment for being carried out to second coded video data stream Second decoding process obtains first coded video data stream and the deep video data flow;
First decoder module, for being carried out and first coded treatment corresponding first to first encoded video streams Decoding process obtains the original video data stream;
Synchronous processing module, for synchronizing place to the original video data stream and the deep video data flow Reason.
Optionally, second decoder module specifically for:
Read the zone bit information of the second coded video data stream end or head;
Judge that second coded video data stream, with depth information, is compiled according to described second according to zone bit information Key message in code video flowing parses first encoded video streams and the deep video data flow.
The embodiment of the present invention also proposed a kind of method for realizing coding and decoding video, including:
Original video data stream and deep video data flow are obtained by binocular platform;
First coded treatment is carried out to original video data stream using two-dimensional video coding method and obtains the first encoded video Data flow;
Second volume is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream Code treatment obtains the second coded video data stream.
Optionally, the binocular platform includes binocular camera;It is described by binocular platform obtain original video data stream and Deep video data flow includes:
First video data stream is obtained by the first camera of binocular camera, while by the second of the binocular camera Camera obtains the second video data stream;
Using first video data stream or second video data stream as the original video data stream;
The image of each frame first and second video data stream according to first video data stream is corresponding each The image of frame second calculates each frame depth image in the deep video data flow;By depth described in all range image integrations Video data stream.
Optionally, the binocular platform includes camera and the equipment for obtaining depth information;
It is described to be included by binocular platform acquisition original video data stream and deep video data flow:
The original video data stream is obtained by the camera, while by described for obtaining setting for depth information It is standby to obtain the deep video data flow.
Optionally, it is described according to deep video data flow and the first coded video data stream to the first coded video data stream Carry out the second coded treatment and obtain the second coded video data stream including:
Obtain the key message in first coded video data stream;
The end that the deep video data flow and the key message write first coded video data stream is obtained To the 3rd coded video data stream;
Added for indicating whether the zone bit information with depth information at the 3rd coded video data stream end Obtain second coded video data stream.
Optionally, it is described according to deep video data flow and the first coded video data stream to the first coded video data stream Carry out the second coded treatment and obtain the second coded video data stream including:
Obtain the key message in first coded video data stream;
The head that the deep video data flow and the key message write first coded video data stream is obtained To the 3rd coded video data stream;
Added for indicating whether the zone bit information with depth information in the 3rd coded video data stream head Obtain second coded video data stream.
Optionally, the method also includes:
The second decoding process corresponding with second coded treatment is carried out to second coded video data stream to obtain First coded video data stream and the deep video data flow;
First encoded video streams are carried out with the first decoding process corresponding with first coded treatment and obtains described Original video data stream;
Treatment is synchronized to the original video data stream and the deep video data flow.
Optionally, it is described the second decoding process corresponding with the second coded treatment is carried out to the second coded video data stream to obtain Include to the first coded video data stream and the deep video data flow:
Read the zone bit information of the second coded video data stream end or head;
Judge that second coded video data stream, with depth information, is compiled according to described second according to zone bit information Key message in code video flowing parses first encoded video streams and the deep video data flow.
Compared with correlation technique, the embodiment of the present invention includes:Original video data stream and depth are obtained by binocular platform Video data stream;First coded treatment carried out to original video data stream using two-dimensional video coding method obtain the first coding to regard Frequency data stream;The first coded video data stream is carried out at the second coding according to deep video data flow and original video data stream Reason obtains the second coded video data stream.By the scheme of the embodiment of the present invention, first using two-dimensional video coding method to original Video data stream carries out the first coded treatment and obtains the first coded video data stream, further according to deep video data flow and original regards Frequency data stream carries out the second coded treatment to the first coded video data stream, so that compatibility two-dimensional video coded format, it is ensured that Three-dimensional video-frequency can be decoded and played on two-dimensional video player.
Brief description of the drawings
The accompanying drawing in the embodiment of the present invention is illustrated below, the accompanying drawing in embodiment be for it is of the invention enter one Step understanding, is used to explain the present invention together with specification, does not constitute limiting the scope of the invention.
Fig. 1 is an optional hardware architecture diagram of the mobile terminal for realizing each embodiment of the invention;
Fig. 2 is the wireless communication system schematic diagram of mobile terminal as shown in Figure 1;
Fig. 3 realizes the flow chart of the method for coding and decoding video for first embodiment of the invention;
Fig. 4 (a)~Fig. 4 (d) is four kinds of form schematic diagrams of the coded video data stream of first embodiment of the invention second;
Fig. 5 realizes the structure composition schematic diagram of the device of coding and decoding video for second embodiment of the invention.
The realization of the object of the invention, functional characteristics and advantage will be described further referring to the drawings in conjunction with the embodiments.
Specific embodiment
For the ease of the understanding of those skilled in the art, the invention will be further described below in conjunction with the accompanying drawings, not Can be used for limiting the scope of the invention.It should be noted that in the case where not conflicting, embodiment and reality in the application The various modes applied in example can be mutually combined.
It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, it is not intended to limit the present invention.
The mobile terminal of each embodiment of the invention is realized referring now to Description of Drawings.In follow-up description, use For represent element such as " module ", " part " or " unit " suffix only for being conducive to explanation of the invention, itself Not specific meaning.Therefore, " module " can be used mixedly with " part ".
Mobile terminal can be implemented in a variety of manners.For example, the terminal described in the present invention can include such as moving Phone, smart phone, notebook computer, digit broadcasting receiver, PDA (personal digital assistant), PAD (panel computer), PMP The mobile terminal of (portable media player), guider etc. and such as numeral TV, desktop computer etc. are consolidated Determine terminal.Hereinafter it is assumed that terminal is mobile terminal.However, it will be understood by those skilled in the art that, except being used in particular for movement Outside the element of purpose, construction according to the embodiment of the present invention can also apply to the terminal of fixed type.
Fig. 1 is that an optional hardware configuration of the mobile terminal for realizing each embodiment of the invention is illustrated.
Mobile terminal 1 00 can include wireless communication unit 110, A/V (audio/video) input block 120, output unit 150th, memory 160, interface unit 170, controller 180 and power subsystem 190 etc..Fig. 1 is shown with various assemblies Mobile terminal, it should be understood that being not required for implementing all components for showing.Can alternatively implement more or less Component.The element of mobile terminal will be discussed in more detail below.
Wireless communication unit 110 generally includes one or more assemblies, and it allows mobile terminal 1 00 and wireless communication system Or the radio communication between network.For example, wireless communication unit can include mobile communication module 112.
Mobile communication module 112 sends radio signals to base station (for example, access point, node B etc.), exterior terminal And at least one of server and/or receive from it radio signal.Such radio signal can be logical including voice Words signal, video calling signal or the various types of data for sending and/or receiving according to text and/or Multimedia Message.
A/V input blocks 120 are used to receive audio or video signal.A/V input blocks 120 can include camera 121, phase Static images or the image of video that 121 pairs, machine is obtained in Video Capture pattern or image capture mode by image capture apparatus Data are processed.Picture frame after treatment may be displayed on display unit 151.Picture frame after being processed through camera 121 can It is transmitted in memory 160 (or other storage mediums) or via wireless communication unit 110 with storing, can be according to shifting The construction of dynamic terminal provides two or more cameras 121.
Interface unit 170 is connected the interface that can pass through with mobile terminal 1 00 as at least one external device (ED).For example, External device (ED) can include wired or wireless head-band earphone port, external power source (or battery charger) port, wired or nothing Line FPDP, memory card port, the port for connecting the device with identification module, audio input/output (I/O) end Mouth, video i/o port, ear port etc..Identification module can be that storage uses each of mobile terminal 1 00 for verifying user Kind of information and subscriber identification module (UIM), client identification module (SIM), Universal Subscriber identification module (USIM) can be included Etc..In addition, the device (hereinafter referred to as " identifying device ") with identification module can take the form of smart card, therefore, know Other device can be connected via port or other attachment means with mobile terminal 1 00.Interface unit 170 can be used for reception and come from The input (for example, data message, electric power etc.) of the external device (ED) and input that will be received is transferred in mobile terminal 1 00 One or more elements can be used for transmitting data between mobile terminal and external device (ED).
In addition, when mobile terminal 1 00 is connected with external base, interface unit 170 can serve as allowing by it by electricity Power provides to the path of mobile terminal 1 00 from base or can serve as allowing the various command signals being input into from base to pass through it It is transferred to the path of mobile terminal.Be can serve as recognizing that mobile terminal is from the various command signals or electric power of base input The no signal being accurately fitted within base.Output unit 150 is configured to provide defeated with vision, audio and/or tactile manner Go out signal (for example, audio signal, vision signal, alarm signal, vibration signal etc.).Output unit 150 can include display Unit 151 etc..
Display unit 151 may be displayed on the information processed in mobile terminal 1 00.For example, when mobile terminal 1 00 is in electricity During words call mode, display unit 151 can show and converse or other communicate (for example, text messaging, multimedia file Download etc.) related user interface (UI) or graphic user interface (GUI).When mobile terminal 1 00 is in video calling pattern Or during image capture mode, display unit 151 can show the image of capture and/or the image of reception, show video or figure UI or GUI of picture and correlation function etc..
Meanwhile, when display unit 151 and touch pad in the form of layer it is superposed on one another to form touch-screen when, display unit 151 can serve as input unit and output device.Display unit 151 can include liquid crystal display (LCD), thin film transistor (TFT) In LCD (TFT-LCD), Organic Light Emitting Diode (OLED) display, flexible display, three-dimensional (3D) display etc. at least It is a kind of.Some in these displays may be constructed such that transparence to allow user to be watched from outside, and this is properly termed as transparent Display, typical transparent display can be, for example, TOLED (transparent organic light emitting diode) display etc..According to specific Desired implementation method, mobile terminal 1 00 can include two or more display units (or other display devices), for example, moving Dynamic terminal can include outernal display unit (not shown) and inner display unit (not shown).Touch-screen can be used to detect touch Input pressure and touch input position and touch input area.
Memory 160 can store software program for the treatment and control operation performed by controller 180 etc., Huo Zheke Temporarily to store oneself data (for example, telephone directory, message, still image, video etc.) through exporting or will export.And And, memory 160 can store the vibration of various modes on being exported when touching and being applied to touch-screen and audio signal Data.
Memory 160 can include the storage medium of at least one type, and the storage medium includes flash memory, hard disk, many Media card, card-type memory (for example, SD or DX memories etc.), random access storage device (RAM), static random-access storage Device (SRAM), read-only storage (ROM), Electrically Erasable Read Only Memory (EEPROM), programmable read only memory (PROM), magnetic storage, disk, CD etc..And, mobile terminal 1 00 can perform memory with by network connection The network storage device cooperation of 160 store function.
The overall operation of the generally control mobile terminal of controller 180.For example, controller 180 is performed and voice call, data Communication, video calling etc. related control and treatment.Controller 180 can be with execution pattern identifying processing, will be in touch-screen The handwriting input of upper execution or picture draw input and are identified as character or image.
Power subsystem 190 receives external power or internal power under the control of controller 180 and provides operation each unit Appropriate electric power needed for part and component.
Various implementation methods described herein can be with use such as computer software, hardware or its any combination of calculating Machine computer-readable recording medium is implemented.Implement for hardware, implementation method described herein can be by using application-specific IC (ASIC), digital signal processor (DSP), digital signal processing device (DSPD), programmable logic device (PLD), scene can Programming gate array (FPGA), processor, controller, microcontroller, microprocessor, it is designed to perform function described herein At least one in electronic unit is implemented, and in some cases, such implementation method can be implemented in controller 180. For software implementation, the implementation method of such as process or function can with allow to perform the single of at least one function or operation Software module is implemented.Software code can be come by the software application (or program) write with any appropriate programming language Implement, software code can be stored in memory 160 and performed by controller 180.
So far, oneself according to its function through describing mobile terminal.Below, for the sake of brevity, will description such as folded form, Slide type mobile terminal in various types of mobile terminals of board-type, oscillating-type, slide type mobile terminal etc. is used as showing Example.Therefore, the present invention can be applied to any kind of mobile terminal, and be not limited to slide type mobile terminal.
Mobile terminal 1 00 as shown in Figure 1 may be constructed such that using via frame or packet transmission data it is all if any Line and wireless communication system and satellite-based communication system are operated.
The communication system that mobile terminal wherein of the invention can be operated is described referring now to Fig. 2.
Such communication system can use different air interface and/or physical layer.For example, used by communication system Air interface includes such as frequency division multiple access (FDMA), time division multiple acess (TDMA), CDMA (CDMA) and universal mobile communications system System (UMTS) (especially, Long Term Evolution (LTE)), global system for mobile communications (GSM) etc..As non-limiting example, under The description in face is related to cdma communication system, but such teaching is equally applicable to other types of system.
With reference to Fig. 2, cdma wireless communication system can include multiple mobile terminal 1s 00, multiple base station (BS) 270, base station Controller (BSC) 275 and mobile switching centre (MSC) 280.MSC280 is configured to and Public Switched Telephony Network (PSTN) 290 form interface.MSC280 is also structured to form interface with the BSC275 that can be couple to base station 270 via back haul link. If any one in the interface that back haul link can be known according to Ganji is constructed, the interface includes such as E1/T1, ATM, IP, PPP, frame relay, HDSL, ADSL or xDSL.It will be appreciated that system can include multiple as shown in Figure 2
BSC2750。
Each BS270 can service one or more subregions (or region), by multidirectional antenna or the day of sensing specific direction Each subregion of line covering is radially away from BS270.Or, each subregion can be by two or more for diversity reception Antenna is covered.Each BS270 may be constructed such that the multiple frequency distribution of support, and the distribution of each frequency has specific frequency spectrum (for example, 1.25MHz, 5MHz etc.).
What subregion and frequency were distributed intersects can be referred to as CDMA Channel.BS270 can also be referred to as base station transceiver System (BTS) or other equivalent terms.In this case, term " base station " can be used for broadly representing single BSC275 and at least one BS270.Base station can also be referred to as " cellular station ".Or, each subregion of specific BS270 can be claimed It is multiple cellular stations.
As shown in Figure 2, broadcast singal is sent to broadcsting transmitter (BT) 295 mobile terminal operated in system 100.Broadcasting reception module 111 as shown in Figure 1 is arranged at mobile terminal 1 00 to receive the broadcast sent by BT295 Signal.In fig. 2 it is shown that several global positioning system (GPS) satellites 300.Satellite 300 helps position multiple mobile terminals At least one of 100.
In fig. 2, multiple satellites 300 are depicted, it is understood that be, it is possible to use any number of satellite obtains useful Location information.GPS module 115 as shown in Figure 1 is generally configured to coordinate with satellite 300 to be believed with obtaining desired positioning Breath.Substitute GPS tracking techniques or outside GPS tracking techniques, it is possible to use other of the position of mobile terminal can be tracked Technology.In addition, at least one gps satellite 300 can optionally or additionally process satellite dmb transmission.
Used as a typical operation of wireless communication system, BS270 receives the reverse link from various mobile terminal 1s 00 Signal.Mobile terminal 1 00 generally participates in call, information receiving and transmitting and other types of communication.Each of the reception of certain base station 270 is anti- Processed in specific BS270 to link signal.The data of acquisition are forwarded to the BSC275 of correlation.BSC provides call Resource allocation and the mobile management function of the coordination including the soft switching process between BS270.The number that BSC275 will also be received According to MSC280 is routed to, it provides the extra route service for forming interface with PSTN290.Similarly, PSTN290 with MSC280 forms interface, and MSC and BSC275 form interface, and BSC275 correspondingly controls BS270 with by forward link signals It is sent to mobile terminal 1 00.
Based on above-mentioned mobile terminal hardware configuration and communication system, the inventive method each embodiment is proposed.
As shown in figure 3, first embodiment of the invention proposes a kind of method for realizing coding and decoding video, including:
Step 300, by binocular platform obtain original video data stream and deep video data flow.
In this step, binocular platform can include binocular camera, i.e., including two cameras;A shooting can also be included Head and the equipment (such as active laser device etc.) for being used to obtain depth information.
When binocular platform includes binocular camera;Original video data stream and deep video data are obtained by binocular platform Stream includes:
First video data stream is obtained by the first camera of binocular camera, while the second shooting for passing through binocular camera Head obtains the second video data stream;
Using the first video data stream or the second video data stream as original video data stream;
The corresponding figure of each frame second of the image of each frame first and the second video data stream according to the first video data stream As calculating each frame depth image in deep video data flow;By all range image integration deep video data flows.
Wherein, the first video data stream includes the image of multiframe first, and the second video data stream includes the image of multiframe second, former Beginning video data stream includes the image of multiframe first or the second image, and deep video data flow includes multiframe depth image.
First image, the second image and depth image are corresponded.
Wherein it is possible to each frame depth image during deep video data flow is calculated using Stereo Matching Algorithm.
When binocular platform is including camera and for the equipment for obtaining depth information;
Obtaining original video data stream and deep video data flow by binocular platform includes:
Original video data stream is obtained by camera, while being regarded by for obtaining the equipment acquisition depth of depth information Frequency data stream.
Wherein, original video data stream includes the image that multiframe camera is obtained, and deep video data flow is included for obtaining Take the multiframe depth image that the equipment of depth information is obtained.
The image and depth image that camera is obtained are corresponded.
Step 301, the first coded treatment is carried out to original video data stream using two-dimensional video coding method obtain first Coded video data stream.
In this step, two-dimensional video coding method can be any related two-dimensional video coding method, and the present invention is implemented Example is not construed as limiting to this.
Step 302, the first coded video data stream is entered according to deep video data flow and the first coded video data stream The coded treatment of row second obtains the second coded video data stream.
In this step, the first coded video data stream is entered according to deep video data flow and the first coded video data stream The coded treatment of row second obtains the second coded video data stream to be included:
Obtain the key message in the first coded video data stream;
The end that deep video data flow and key message write the first coded video data stream is obtained into the 3rd coding to regard Frequency data stream;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end Second coded video data stream.
Or, is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream Two coded treatments obtain the second coded video data stream to be included:
Obtain the key message in the first coded video data stream;
Deep video data flow is carried out after the first coded treatment obtains coded treatment using two-dimensional video coding method Deep video data flow;
The end that deep video data flow after coded treatment and key message write the first coded video data stream is obtained To the 3rd coded video data stream;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end Second coded video data stream.
Or, is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream Two coded treatments obtain the second coded video data stream to be included:
Obtain the key message in the first coded video data stream;
The head that deep video data flow and key message write the first coded video data stream is obtained into the 3rd coding to regard Frequency data stream;
Added for indicating whether that the zone bit information with depth information is obtained in the 3rd coded video data stream head Second coded video data stream.
Or, is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream Two coded treatments obtain the second coded video data stream to be included:
Obtain the key message in the first coded video data stream;
Deep video data flow is carried out after the first coded treatment obtains coded treatment using two-dimensional video coding method Deep video data flow;
The head that deep video data flow after coded treatment and key message write the first coded video data stream is obtained To the 3rd coded video data stream;
Added for indicating whether that the zone bit information with depth information is obtained in the 3rd coded video data stream head Second coded video data stream.
Wherein, key message includes following any one or more:Video width and video height, frame per second (i.e. unit The number of image frames included in time), the effective length of the first coded video data stream, the coding lattice of the first coded video data stream Formula etc..
When being first compressed to original video data stream before the first coded treatment is carried out to original video data stream or During amplification, key message includes video width and video height, high according to video width and video in follow-up decoding process Degree is remapped.
Specifically technological means can be known come in obtaining the first coded video data stream using those skilled in the art Key message, is not intended to limit the present invention the protection domain of embodiment, repeats no more here.
Wherein, zone bit information can be represented using a bit, it is for instance possible to use 0 represents without depth Information, represents with depth information using 1.Can certainly be represented using other modes, specific method for expressing sheet Inventive embodiments are not construed as limiting.
Wherein, Fig. 4 (a)~Fig. 4 (d) is four kinds of form schematic diagrams of the second coded video data stream.
As shown in Fig. 4 (a), the second coded video data stream includes the first coded video data stream, deep video number successively According to stream, key message and zone bit information.
As shown in Fig. 4 (b), after the second coded video data stream includes the first coded video data stream, coded treatment successively Deep video data flow, key message and zone bit information.
As shown in Fig. 4 (c), the second coded video data stream includes zone bit information, key message, deep video number successively According to stream and the first coded video data stream.
As shown in Fig. 4 (d), after the second coded video data stream includes zone bit information, key message, coded treatment successively Deep video data flow and the first coded video data stream.
Optionally, the method also includes:
The second decoding process corresponding with the second coded treatment is carried out to the second coded video data stream and obtains the first coding Video data stream and deep video data flow;
First encoded video streams are carried out with the first decoding process corresponding with the first coded treatment and obtains original video data Stream;
Treatment is synchronized to original video data stream and deep video data flow.
Wherein, the second decoding process corresponding with the second coded treatment is carried out to the second coded video data stream and obtains first Coded video data stream and deep video data flow include:
Read the zone bit information of the second coded video data stream end or head;
Second coded video data stream is judged with depth information according to zone bit information, according to the second encoded video streams In key message parse the first encoded video streams and deep video data flow.
Or, the second decoding process corresponding with the second coded treatment is carried out to the second coded video data stream and obtains first Coded video data stream and deep video data flow include:
Read the zone bit information of the second coded video data stream end or head;
Second coded video data stream is judged with depth information according to zone bit information, according to the second encoded video streams In key message parse the deep video data flow after the first encoded video streams and coded treatment;
The first decoding process corresponding with the first coded treatment is carried out to the deep video data flow after coded treatment to obtain Deep video data flow.
Wherein, the key message in the second encoded video streams parses the first encoded video streams and deep video data Stream includes:
The effective length of the first coded video data stream in key message parses the first encoded video streams and depth Degree video data stream.
Wherein, after the key message in the second encoded video streams parses the first encoded video streams and coded treatment Deep video data flow includes:
The effective length of the first coded video data stream in key message parses the first encoded video streams and volume Deep video data flow after code treatment.
Wherein, synchronizing treatment to original video data stream and deep video data flow includes:
Frame per second in key message synchronizes treatment to original video data stream and deep video data flow.Will Each frame of original video data stream and deep video data flow is corresponded.
When key message includes video width and video height, also needed to be regarded original before treatment is synchronized Frequency data stream be remapped to original size (i.e. with key message in video width and video height identical size).
Wherein, the first encoded video streams are carried out with the first decoding process corresponding with the first coded treatment and obtains original video Data flow includes:
The coded format of the first coded video data stream in key message carries out first to the first encoded video streams Decoding process obtains original video data stream.
Optionally, when the second coded video data stream is judged without depth information according to zone bit information, explanation The first encoded video streams and zone bit information are only included in second coded video data stream, then in key message first The coded format of coded video data stream carries out the first decoding process to the first encoded video streams.
By the scheme of the embodiment of the present invention, first first is carried out to original video data stream using two-dimensional video coding method Coded treatment obtains the first coded video data stream, is encoded to first further according to deep video data flow and original video data stream Video data stream carries out the second coded treatment, so that compatibility two-dimensional video coded format, it is ensured that three-dimensional video-frequency can be two Decoded and played on dimension video player.
Referring to Fig. 5, second embodiment of the invention proposes a kind of device for realizing coding and decoding video, including:
Acquisition module, for obtaining original video data stream and deep video data flow by binocular platform;
First coding module, for carrying out the first coded treatment to original video data stream using two-dimensional video coding method Obtain the first coded video data stream;
Second coding module, for according to deep video data flow and the first coded video data stream to the first encoded video Data flow carries out the second coded treatment and obtains the second coded video data stream.
Wherein, acquisition module can be realized by Fig. 1 cameras 121 and controller 180, the first coding module and second Coding module can be arranged in the controller 180 of Fig. 1.
Optionally, binocular platform includes binocular camera;Acquisition module specifically for:
First video data stream is obtained by the first camera of binocular camera, while the second shooting for passing through binocular camera Head obtains the second video data stream;
Using the first video data stream or the second video data stream as original video data stream;
The corresponding figure of each frame second of the image of each frame first and the second video data stream according to the first video data stream As calculating each frame depth image in deep video data flow;By all range image integration deep video data flows.
Wherein, the first video data stream includes the image of multiframe first, and the second video data stream includes the image of multiframe second, former Beginning video data stream includes the image of multiframe first or the second image, and deep video data flow includes multiframe depth image.
First image, the second image and depth image are corresponded.
Wherein, acquisition module can calculate each frame depth map in deep video data flow using Stereo Matching Algorithm Picture.
Optionally, binocular platform includes camera and the equipment for obtaining depth information;
Acquisition module specifically for:
Original video data stream is obtained by camera, while being regarded by for obtaining the equipment acquisition depth of depth information Frequency data stream.
Wherein, original video data stream includes the image that multiframe camera is obtained, and deep video data flow is included for obtaining Take the multiframe depth image that the equipment of depth information is obtained.
The image and depth image that camera is obtained are corresponded.
Optionally, the second coding module specifically for:
Obtain the key message in the first coded video data stream;
The end that deep video data flow and key message write the first coded video data stream is obtained into the 3rd coding to regard Frequency data stream;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end Second coded video data stream.
Optionally, the second coding module specifically for:
Obtain the key message in the first coded video data stream;
Deep video data flow is carried out after the first coded treatment obtains coded treatment using two-dimensional video coding method Deep video data flow;
The end that deep video data flow after coded treatment and key message write the first coded video data stream is obtained To the 3rd coded video data stream;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end Second coded video data stream.
Optionally, the second coding module specifically for:
Obtain the key message in the first coded video data stream;
The head that deep video data flow and key message write the first coded video data stream is obtained into the 3rd coding to regard Frequency data stream;
Added for indicating whether that the zone bit information with depth information is obtained in the 3rd coded video data stream head Second coded video data stream.
Optionally, the second coding module specifically for:
Obtain the key message in the first coded video data stream;
Deep video data flow is carried out after the first coded treatment obtains coded treatment using two-dimensional video coding method Deep video data flow;
The head that deep video data flow after coded treatment and key message write the first coded video data stream is obtained To the 3rd coded video data stream;
Added for indicating whether that the zone bit information with depth information is obtained in the 3rd coded video data stream head Second coded video data stream.
Wherein, key message includes following any one or more:Video width and video height, frame per second (i.e. unit The number of image frames included in time), the effective length of the first coded video data stream, the coding lattice of the first coded video data stream Formula etc..
When being first compressed to original video data stream before the first coded treatment is carried out to original video data stream or During amplification, key message includes video width and video height, high according to video width and video in follow-up decoding process Degree is remapped.
Specifically technological means can be known come in obtaining the first coded video data stream using those skilled in the art Key message, is not intended to limit the present invention the protection domain of embodiment, repeats no more here.
Wherein, zone bit information can be represented using a bit, it is for instance possible to use 0 represents without depth Information, represents with depth information using 1.Can certainly be represented using other modes, specific method for expressing sheet Inventive embodiments are not construed as limiting.
Wherein, Fig. 4 (a)~Fig. 4 (d) is four kinds of form schematic diagrams of the second coded video data stream.
As shown in Fig. 4 (a), the second coded video data stream includes the first coded video data stream, deep video number successively According to stream, key message and zone bit information.
As shown in Fig. 4 (b), after the second coded video data stream includes the first coded video data stream, coded treatment successively Deep video data flow, key message and zone bit information.
As shown in Fig. 4 (c), the second coded video data stream includes zone bit information, key message, deep video number successively According to stream and the first coded video data stream.
As shown in Fig. 4 (d), after the second coded video data stream includes zone bit information, key message, coded treatment successively Deep video data flow and the first coded video data stream.
Optionally, also include:
Second decoder module, for carrying out the second decoding corresponding with the second coded treatment to the second coded video data stream Treatment obtains the first coded video data stream and deep video data flow;
First decoder module, for the first encoded video streams to be carried out with the first decoding process corresponding with the first coded treatment Obtain original video data stream;
Synchronous processing module, for synchronizing treatment to original video data stream and deep video data flow.
Optionally, the second decoder module specifically for:
Read the zone bit information of the second coded video data stream end or head;
Second coded video data stream is judged with depth information according to zone bit information, according to the second encoded video streams In key message parse the first encoded video streams and deep video data flow.
Optionally, the second decoder module specifically for:
Read the zone bit information of the second coded video data stream end or head;
Second coded video data stream is judged with depth information according to zone bit information, according to the second encoded video streams In key message parse the deep video data flow after the first encoded video streams and coded treatment;
The first decoding process corresponding with the first coded treatment is carried out to the deep video data flow after coded treatment to obtain Deep video data flow.
It should be noted that herein, term " including ", "comprising" or its any other variant be intended to non-row His property is included, so that process, method, article or device including a series of key elements not only include those key elements, and And also include other key elements being not expressly set out, or also include for this process, method, article or device institute are intrinsic Key element.In the absence of more restrictions, the key element limited by sentence "including a ...", it is not excluded that including this Also there is other identical element in the process of key element, method, article or device.
The embodiments of the present invention are for illustration only, and the quality of embodiment is not represented.
Through the above description of the embodiments, those skilled in the art can be understood that above-described embodiment side Method can add the mode of required general hardware platform to realize by software, naturally it is also possible to by hardware, but in many cases The former is more preferably implementation method.Based on such understanding, technical scheme is substantially done to prior art in other words The part for going out contribution can be embodied in the form of software product, and the computer software product is stored in a storage medium In (such as ROM/RAM, magnetic disc, CD), including some instructions are used to so that a station terminal equipment (can be mobile phone, computer, clothes Business device, air-conditioner, or network equipment etc.) perform method described in each embodiment of the invention.
The preferred embodiments of the present invention are these are only, the scope of the claims of the invention is not thereby limited, it is every to utilize this hair Equivalent structure or equivalent flow conversion that bright specification and accompanying drawing content are made, or directly or indirectly it is used in other related skills Art field, is included within the scope of the present invention.

Claims (10)

1. a kind of device for realizing coding and decoding video, it is characterised in that including:
Acquisition module, for obtaining original video data stream and deep video data flow by binocular platform;
First coding module, obtains for carrying out the first coded treatment to original video data stream using two-dimensional video coding method First coded video data stream;
Second coding module, for according to deep video data flow and the first coded video data stream to the first coding video frequency data Stream carries out the second coded treatment and obtains the second coded video data stream.
2. device according to claim 1, it is characterised in that the binocular platform includes binocular camera;The acquisition mould Block specifically for:
First video data stream is obtained by the first camera of binocular camera, while the second shooting for passing through the binocular camera Head obtains the second video data stream;
Using first video data stream or second video data stream as the original video data stream;
The image of each frame first and the corresponding each frame of second video data stream according to first video data stream Two images calculate each frame depth image in the deep video data flow;By deep video described in all range image integrations Data flow.
3. device according to claim 1, it is characterised in that the binocular platform includes camera and for obtaining depth The equipment of information;
The acquisition module specifically for:
The original video data stream is obtained by the camera, while being obtained for obtaining the equipment of depth information by described Take the deep video data flow.
4. device according to claim 1, it is characterised in that second coding module specifically for:
Obtain the key message in first coded video data stream;
The end that the deep video data flow and the key message write first coded video data stream is obtained the Three coded video data streams;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end Second coded video data stream.
5. the device according to Claims 1 to 4 any one, it is characterised in that also include:
Second decoder module, for being carried out and second coded treatment corresponding second to second coded video data stream Decoding process obtains first coded video data stream and the deep video data flow;
First decoder module, for first encoded video streams to be carried out with the first decoding corresponding with first coded treatment Treatment obtains the original video data stream;
Synchronous processing module, for synchronizing treatment to the original video data stream and the deep video data flow.
6. a kind of method for realizing coding and decoding video, it is characterised in that including:
Original video data stream and deep video data flow are obtained by binocular platform;
First coded treatment is carried out to original video data stream using two-dimensional video coding method and obtains the first coding video frequency data Stream;
The first coded video data stream is carried out at the second coding according to deep video data flow and the first coded video data stream Reason obtains the second coded video data stream.
7. method according to claim 6, it is characterised in that the binocular platform includes binocular camera;It is described by double Head platform obtains original video data stream and deep video data flow to be included:
First video data stream is obtained by the first camera of binocular camera, while the second shooting for passing through the binocular camera Head obtains the second video data stream;
Using first video data stream or second video data stream as the original video data stream;
The image of each frame first and the corresponding each frame of second video data stream according to first video data stream Two images calculate each frame depth image in the deep video data flow;By deep video described in all range image integrations Data flow.
8. method according to claim 6, it is characterised in that the binocular platform includes camera and for obtaining depth The equipment of information;
It is described to be included by binocular platform acquisition original video data stream and deep video data flow:
The original video data stream is obtained by the camera, while being obtained for obtaining the equipment of depth information by described Take the deep video data flow.
9. method according to claim 6, it is characterised in that described according to deep video data flow and the first encoded video Data flow carries out the second coded treatment to the first coded video data stream and obtains the second coded video data stream including:
Obtain the key message in first coded video data stream;
The end that the deep video data flow and the key message write first coded video data stream is obtained the Three coded video data streams;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end Second coded video data stream.
10. the method according to claim 6~9 any one, it is characterised in that the method also includes:
The second decoding process corresponding with second coded treatment is carried out to second coded video data stream and obtains described First coded video data stream and the deep video data flow;
First encoded video streams are carried out with the first decoding process corresponding with first coded treatment and obtains described original Video data stream;
Treatment is synchronized to the original video data stream and the deep video data flow.
CN201611083136.9A 2016-11-30 2016-11-30 A kind of method and apparatus for realizing coding and decoding video Pending CN106791764A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611083136.9A CN106791764A (en) 2016-11-30 2016-11-30 A kind of method and apparatus for realizing coding and decoding video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611083136.9A CN106791764A (en) 2016-11-30 2016-11-30 A kind of method and apparatus for realizing coding and decoding video

Publications (1)

Publication Number Publication Date
CN106791764A true CN106791764A (en) 2017-05-31

Family

ID=58901606

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611083136.9A Pending CN106791764A (en) 2016-11-30 2016-11-30 A kind of method and apparatus for realizing coding and decoding video

Country Status (1)

Country Link
CN (1) CN106791764A (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1389020A1 (en) * 2002-08-07 2004-02-11 Electronics and Telecommunications Research Institute Method and apparatus for multiplexing multi-view three-dimensional moving picture
CN1954606A (en) * 2004-05-21 2007-04-25 韩国电子通信研究院 Apparatus and method for transmitting/receiving 3D stereoscopic digital broadcasting signal by using 3D stereoscopic video additional data
CN101459857A (en) * 2007-12-10 2009-06-17 深圳华为通信技术有限公司 Communication terminal and information system
CN102246529A (en) * 2008-12-15 2011-11-16 皇家飞利浦电子股份有限公司 Image based 3D video format
CN102326390A (en) * 2009-02-19 2012-01-18 汤姆逊许可证公司 3D video format
CN104205821A (en) * 2012-04-05 2014-12-10 皇家飞利浦有限公司 Depth helper data

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1389020A1 (en) * 2002-08-07 2004-02-11 Electronics and Telecommunications Research Institute Method and apparatus for multiplexing multi-view three-dimensional moving picture
CN1954606A (en) * 2004-05-21 2007-04-25 韩国电子通信研究院 Apparatus and method for transmitting/receiving 3D stereoscopic digital broadcasting signal by using 3D stereoscopic video additional data
CN101459857A (en) * 2007-12-10 2009-06-17 深圳华为通信技术有限公司 Communication terminal and information system
CN102246529A (en) * 2008-12-15 2011-11-16 皇家飞利浦电子股份有限公司 Image based 3D video format
CN102326390A (en) * 2009-02-19 2012-01-18 汤姆逊许可证公司 3D video format
CN104205821A (en) * 2012-04-05 2014-12-10 皇家飞利浦有限公司 Depth helper data

Similar Documents

Publication Publication Date Title
CN106485689B (en) A kind of image processing method and device
CN104866265B (en) Multi-media file display method and device
CN104735255B (en) Multi-screen display method and system
KR101708306B1 (en) Mobile twrminal and 3d image convergence method thereof
CN108111874B (en) file processing method, terminal and server
CN105828180A (en) Apparatus and method for caching video frames
CN105100673B (en) Desktop sharing method and device based on VoLTE
CN106713640B (en) A kind of brightness adjusting method and equipment
CN106155694A (en) A kind of application and the display packing and device applied of attending to anything else
CN106534590A (en) Photo processing method and apparatus, and terminal
CN106200924B (en) A kind of terminal and its method that virtual reality video is controlled
CN106250081A (en) A kind of display packing based on double screen terminal and device
CN105187911A (en) Method and device for displaying video image and image display method
CN106791119B (en) Photo processing method and device and terminal
CN104935822A (en) Method and device for processing images
CN105635837B (en) A kind of video broadcasting method and device
CN106303259A (en) A kind of method and apparatus realizing taking pictures
CN105049916B (en) A kind of video recording method and device
CN106230804A (en) Picture sharing method, terminal and server
CN106651773A (en) Picture processing method and device
CN106341687A (en) Method and device for locating video decoding abnormality
CN106657950A (en) Projection device management device, method and projection data sharing device
CN106060522A (en) Video image processing device and method
CN106534203A (en) Mobile terminal and communication method
CN106651762A (en) Photo processing method, device and terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20170531

RJ01 Rejection of invention patent application after publication