CN106791764A - A kind of method and apparatus for realizing coding and decoding video - Google Patents
A kind of method and apparatus for realizing coding and decoding video Download PDFInfo
- Publication number
- CN106791764A CN106791764A CN201611083136.9A CN201611083136A CN106791764A CN 106791764 A CN106791764 A CN 106791764A CN 201611083136 A CN201611083136 A CN 201611083136A CN 106791764 A CN106791764 A CN 106791764A
- Authority
- CN
- China
- Prior art keywords
- video data
- data stream
- coded
- deep
- data flow
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 80
- 238000011282 treatment Methods 0.000 claims abstract description 82
- 230000001360 synchronised effect Effects 0.000 claims description 7
- 230000010354 integration Effects 0.000 claims description 6
- 210000003128 head Anatomy 0.000 description 21
- 238000004891 communication Methods 0.000 description 18
- 230000015654 memory Effects 0.000 description 12
- 238000010586 diagram Methods 0.000 description 6
- 230000006870 function Effects 0.000 description 6
- 238000010295 mobile communication Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 230000003321 amplification Effects 0.000 description 2
- 230000000712 assembly Effects 0.000 description 2
- 238000000429 assembly Methods 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 230000005611 electricity Effects 0.000 description 2
- 238000003199 nucleic acid amplification method Methods 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 101150012579 ADSL gene Proteins 0.000 description 1
- 102100020775 Adenylosuccinate lyase Human genes 0.000 description 1
- 108700040193 Adenylosuccinate lyases Proteins 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000005314 correlation function Methods 0.000 description 1
- 239000009730 ganji Substances 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 210000001747 pupil Anatomy 0.000 description 1
- 238000013468 resource allocation Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/40—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video transcoding, i.e. partial or full decoding of a coded input stream followed by re-encoding of the decoded output stream
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
A kind of method and apparatus for realizing coding and decoding video, the method includes:Original video data stream and deep video data flow are obtained by binocular platform;First coded treatment is carried out to original video data stream using two-dimensional video coding method and obtains the first coded video data stream;Second coded treatment is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream and obtains the second coded video data stream.By the scheme of the embodiment of the present invention, the first coded treatment is first carried out to original video data stream using two-dimensional video coding method and obtains the first coded video data stream, the second coded treatment is carried out to the first coded video data stream further according to deep video data flow and the first coded video data stream, so as to compatible two-dimensional video coded format, it is ensured that three-dimensional video-frequency can be decoded and played on two-dimensional video player.
Description
Technical field
Present document relates to but be not limited to video technique, espespecially a kind of method and apparatus for realizing coding and decoding video.
Background technology
Three-dimensional (3D) video technique is based on double pupil principle of parallax of people, and Same Scene but slightly difference are obtained by video camera
Two images, show the left eye and right eye for giving people respectively, binocular parallax is formed, so as to make one to obtain the depth letter of scene
Cease and experience third dimension.Three-dimensional video-frequency technology can truly reappear objective world scene, show the depth of scene
Sense, stereovision and authenticity, are the important directions of current video techniques development.
Current three-dimensional video-frequency can only be played on specific player, and traditional two dimension (2D) video player then cannot
Play three-dimensional video-frequency.
The content of the invention
The embodiment of the present invention proposes a kind of method and apparatus for realizing coding and decoding video, being capable of compatible two-dimensional video coding
Form, it is ensured that three-dimensional video-frequency can be decoded and played on two-dimensional video player.
The embodiment of the present invention proposes a kind of device for realizing coding and decoding video, including:
Acquisition module, for obtaining original video data stream and deep video data flow by binocular platform;
First coding module, for carrying out the first coded treatment to original video data stream using two-dimensional video coding method
Obtain the first coded video data stream;
Second coding module, for according to deep video data flow and the first coded video data stream to the first encoded video
Data flow carries out the second coded treatment and obtains the second coded video data stream.
Optionally, the binocular platform includes binocular camera;The acquisition module specifically for:
First video data stream is obtained by the first camera of binocular camera, while by the second of the binocular camera
Camera obtains the second video data stream;
Using first video data stream or second video data stream as the original video data stream;
The image of each frame first and second video data stream according to first video data stream is corresponding each
The image of frame second calculates each frame depth image in the deep video data flow;By depth described in all range image integrations
Video data stream.
Optionally, the binocular platform includes camera and the equipment for obtaining depth information;
The acquisition module specifically for:
The original video data stream is obtained by the camera, while by described for obtaining setting for depth information
It is standby to obtain the deep video data flow.
Optionally, second coding module specifically for:
Obtain the key message in first coded video data stream;
The end that the deep video data flow and the key message write first coded video data stream is obtained
To the 3rd coded video data stream;
Added for indicating whether the zone bit information with depth information at the 3rd coded video data stream end
Obtain second coded video data stream.
Optionally, second coding module specifically for:
Obtain the key message in first coded video data stream;
The head that the deep video data flow and the key message write first coded video data stream is obtained
To the 3rd coded video data stream;
Added for indicating whether the zone bit information with depth information in the 3rd coded video data stream head
Obtain second coded video data stream.
Optionally, also include:
Second decoder module, it is corresponding with second coded treatment for being carried out to second coded video data stream
Second decoding process obtains first coded video data stream and the deep video data flow;
First decoder module, for being carried out and first coded treatment corresponding first to first encoded video streams
Decoding process obtains the original video data stream;
Synchronous processing module, for synchronizing place to the original video data stream and the deep video data flow
Reason.
Optionally, second decoder module specifically for:
Read the zone bit information of the second coded video data stream end or head;
Judge that second coded video data stream, with depth information, is compiled according to described second according to zone bit information
Key message in code video flowing parses first encoded video streams and the deep video data flow.
The embodiment of the present invention also proposed a kind of method for realizing coding and decoding video, including:
Original video data stream and deep video data flow are obtained by binocular platform;
First coded treatment is carried out to original video data stream using two-dimensional video coding method and obtains the first encoded video
Data flow;
Second volume is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream
Code treatment obtains the second coded video data stream.
Optionally, the binocular platform includes binocular camera;It is described by binocular platform obtain original video data stream and
Deep video data flow includes:
First video data stream is obtained by the first camera of binocular camera, while by the second of the binocular camera
Camera obtains the second video data stream;
Using first video data stream or second video data stream as the original video data stream;
The image of each frame first and second video data stream according to first video data stream is corresponding each
The image of frame second calculates each frame depth image in the deep video data flow;By depth described in all range image integrations
Video data stream.
Optionally, the binocular platform includes camera and the equipment for obtaining depth information;
It is described to be included by binocular platform acquisition original video data stream and deep video data flow:
The original video data stream is obtained by the camera, while by described for obtaining setting for depth information
It is standby to obtain the deep video data flow.
Optionally, it is described according to deep video data flow and the first coded video data stream to the first coded video data stream
Carry out the second coded treatment and obtain the second coded video data stream including:
Obtain the key message in first coded video data stream;
The end that the deep video data flow and the key message write first coded video data stream is obtained
To the 3rd coded video data stream;
Added for indicating whether the zone bit information with depth information at the 3rd coded video data stream end
Obtain second coded video data stream.
Optionally, it is described according to deep video data flow and the first coded video data stream to the first coded video data stream
Carry out the second coded treatment and obtain the second coded video data stream including:
Obtain the key message in first coded video data stream;
The head that the deep video data flow and the key message write first coded video data stream is obtained
To the 3rd coded video data stream;
Added for indicating whether the zone bit information with depth information in the 3rd coded video data stream head
Obtain second coded video data stream.
Optionally, the method also includes:
The second decoding process corresponding with second coded treatment is carried out to second coded video data stream to obtain
First coded video data stream and the deep video data flow;
First encoded video streams are carried out with the first decoding process corresponding with first coded treatment and obtains described
Original video data stream;
Treatment is synchronized to the original video data stream and the deep video data flow.
Optionally, it is described the second decoding process corresponding with the second coded treatment is carried out to the second coded video data stream to obtain
Include to the first coded video data stream and the deep video data flow:
Read the zone bit information of the second coded video data stream end or head;
Judge that second coded video data stream, with depth information, is compiled according to described second according to zone bit information
Key message in code video flowing parses first encoded video streams and the deep video data flow.
Compared with correlation technique, the embodiment of the present invention includes:Original video data stream and depth are obtained by binocular platform
Video data stream;First coded treatment carried out to original video data stream using two-dimensional video coding method obtain the first coding to regard
Frequency data stream;The first coded video data stream is carried out at the second coding according to deep video data flow and original video data stream
Reason obtains the second coded video data stream.By the scheme of the embodiment of the present invention, first using two-dimensional video coding method to original
Video data stream carries out the first coded treatment and obtains the first coded video data stream, further according to deep video data flow and original regards
Frequency data stream carries out the second coded treatment to the first coded video data stream, so that compatibility two-dimensional video coded format, it is ensured that
Three-dimensional video-frequency can be decoded and played on two-dimensional video player.
Brief description of the drawings
The accompanying drawing in the embodiment of the present invention is illustrated below, the accompanying drawing in embodiment be for it is of the invention enter one
Step understanding, is used to explain the present invention together with specification, does not constitute limiting the scope of the invention.
Fig. 1 is an optional hardware architecture diagram of the mobile terminal for realizing each embodiment of the invention;
Fig. 2 is the wireless communication system schematic diagram of mobile terminal as shown in Figure 1;
Fig. 3 realizes the flow chart of the method for coding and decoding video for first embodiment of the invention;
Fig. 4 (a)~Fig. 4 (d) is four kinds of form schematic diagrams of the coded video data stream of first embodiment of the invention second;
Fig. 5 realizes the structure composition schematic diagram of the device of coding and decoding video for second embodiment of the invention.
The realization of the object of the invention, functional characteristics and advantage will be described further referring to the drawings in conjunction with the embodiments.
Specific embodiment
For the ease of the understanding of those skilled in the art, the invention will be further described below in conjunction with the accompanying drawings, not
Can be used for limiting the scope of the invention.It should be noted that in the case where not conflicting, embodiment and reality in the application
The various modes applied in example can be mutually combined.
It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, it is not intended to limit the present invention.
The mobile terminal of each embodiment of the invention is realized referring now to Description of Drawings.In follow-up description, use
For represent element such as " module ", " part " or " unit " suffix only for being conducive to explanation of the invention, itself
Not specific meaning.Therefore, " module " can be used mixedly with " part ".
Mobile terminal can be implemented in a variety of manners.For example, the terminal described in the present invention can include such as moving
Phone, smart phone, notebook computer, digit broadcasting receiver, PDA (personal digital assistant), PAD (panel computer), PMP
The mobile terminal of (portable media player), guider etc. and such as numeral TV, desktop computer etc. are consolidated
Determine terminal.Hereinafter it is assumed that terminal is mobile terminal.However, it will be understood by those skilled in the art that, except being used in particular for movement
Outside the element of purpose, construction according to the embodiment of the present invention can also apply to the terminal of fixed type.
Fig. 1 is that an optional hardware configuration of the mobile terminal for realizing each embodiment of the invention is illustrated.
Mobile terminal 1 00 can include wireless communication unit 110, A/V (audio/video) input block 120, output unit
150th, memory 160, interface unit 170, controller 180 and power subsystem 190 etc..Fig. 1 is shown with various assemblies
Mobile terminal, it should be understood that being not required for implementing all components for showing.Can alternatively implement more or less
Component.The element of mobile terminal will be discussed in more detail below.
Wireless communication unit 110 generally includes one or more assemblies, and it allows mobile terminal 1 00 and wireless communication system
Or the radio communication between network.For example, wireless communication unit can include mobile communication module 112.
Mobile communication module 112 sends radio signals to base station (for example, access point, node B etc.), exterior terminal
And at least one of server and/or receive from it radio signal.Such radio signal can be logical including voice
Words signal, video calling signal or the various types of data for sending and/or receiving according to text and/or Multimedia Message.
A/V input blocks 120 are used to receive audio or video signal.A/V input blocks 120 can include camera 121, phase
Static images or the image of video that 121 pairs, machine is obtained in Video Capture pattern or image capture mode by image capture apparatus
Data are processed.Picture frame after treatment may be displayed on display unit 151.Picture frame after being processed through camera 121 can
It is transmitted in memory 160 (or other storage mediums) or via wireless communication unit 110 with storing, can be according to shifting
The construction of dynamic terminal provides two or more cameras 121.
Interface unit 170 is connected the interface that can pass through with mobile terminal 1 00 as at least one external device (ED).For example,
External device (ED) can include wired or wireless head-band earphone port, external power source (or battery charger) port, wired or nothing
Line FPDP, memory card port, the port for connecting the device with identification module, audio input/output (I/O) end
Mouth, video i/o port, ear port etc..Identification module can be that storage uses each of mobile terminal 1 00 for verifying user
Kind of information and subscriber identification module (UIM), client identification module (SIM), Universal Subscriber identification module (USIM) can be included
Etc..In addition, the device (hereinafter referred to as " identifying device ") with identification module can take the form of smart card, therefore, know
Other device can be connected via port or other attachment means with mobile terminal 1 00.Interface unit 170 can be used for reception and come from
The input (for example, data message, electric power etc.) of the external device (ED) and input that will be received is transferred in mobile terminal 1 00
One or more elements can be used for transmitting data between mobile terminal and external device (ED).
In addition, when mobile terminal 1 00 is connected with external base, interface unit 170 can serve as allowing by it by electricity
Power provides to the path of mobile terminal 1 00 from base or can serve as allowing the various command signals being input into from base to pass through it
It is transferred to the path of mobile terminal.Be can serve as recognizing that mobile terminal is from the various command signals or electric power of base input
The no signal being accurately fitted within base.Output unit 150 is configured to provide defeated with vision, audio and/or tactile manner
Go out signal (for example, audio signal, vision signal, alarm signal, vibration signal etc.).Output unit 150 can include display
Unit 151 etc..
Display unit 151 may be displayed on the information processed in mobile terminal 1 00.For example, when mobile terminal 1 00 is in electricity
During words call mode, display unit 151 can show and converse or other communicate (for example, text messaging, multimedia file
Download etc.) related user interface (UI) or graphic user interface (GUI).When mobile terminal 1 00 is in video calling pattern
Or during image capture mode, display unit 151 can show the image of capture and/or the image of reception, show video or figure
UI or GUI of picture and correlation function etc..
Meanwhile, when display unit 151 and touch pad in the form of layer it is superposed on one another to form touch-screen when, display unit
151 can serve as input unit and output device.Display unit 151 can include liquid crystal display (LCD), thin film transistor (TFT)
In LCD (TFT-LCD), Organic Light Emitting Diode (OLED) display, flexible display, three-dimensional (3D) display etc. at least
It is a kind of.Some in these displays may be constructed such that transparence to allow user to be watched from outside, and this is properly termed as transparent
Display, typical transparent display can be, for example, TOLED (transparent organic light emitting diode) display etc..According to specific
Desired implementation method, mobile terminal 1 00 can include two or more display units (or other display devices), for example, moving
Dynamic terminal can include outernal display unit (not shown) and inner display unit (not shown).Touch-screen can be used to detect touch
Input pressure and touch input position and touch input area.
Memory 160 can store software program for the treatment and control operation performed by controller 180 etc., Huo Zheke
Temporarily to store oneself data (for example, telephone directory, message, still image, video etc.) through exporting or will export.And
And, memory 160 can store the vibration of various modes on being exported when touching and being applied to touch-screen and audio signal
Data.
Memory 160 can include the storage medium of at least one type, and the storage medium includes flash memory, hard disk, many
Media card, card-type memory (for example, SD or DX memories etc.), random access storage device (RAM), static random-access storage
Device (SRAM), read-only storage (ROM), Electrically Erasable Read Only Memory (EEPROM), programmable read only memory
(PROM), magnetic storage, disk, CD etc..And, mobile terminal 1 00 can perform memory with by network connection
The network storage device cooperation of 160 store function.
The overall operation of the generally control mobile terminal of controller 180.For example, controller 180 is performed and voice call, data
Communication, video calling etc. related control and treatment.Controller 180 can be with execution pattern identifying processing, will be in touch-screen
The handwriting input of upper execution or picture draw input and are identified as character or image.
Power subsystem 190 receives external power or internal power under the control of controller 180 and provides operation each unit
Appropriate electric power needed for part and component.
Various implementation methods described herein can be with use such as computer software, hardware or its any combination of calculating
Machine computer-readable recording medium is implemented.Implement for hardware, implementation method described herein can be by using application-specific IC
(ASIC), digital signal processor (DSP), digital signal processing device (DSPD), programmable logic device (PLD), scene can
Programming gate array (FPGA), processor, controller, microcontroller, microprocessor, it is designed to perform function described herein
At least one in electronic unit is implemented, and in some cases, such implementation method can be implemented in controller 180.
For software implementation, the implementation method of such as process or function can with allow to perform the single of at least one function or operation
Software module is implemented.Software code can be come by the software application (or program) write with any appropriate programming language
Implement, software code can be stored in memory 160 and performed by controller 180.
So far, oneself according to its function through describing mobile terminal.Below, for the sake of brevity, will description such as folded form,
Slide type mobile terminal in various types of mobile terminals of board-type, oscillating-type, slide type mobile terminal etc. is used as showing
Example.Therefore, the present invention can be applied to any kind of mobile terminal, and be not limited to slide type mobile terminal.
Mobile terminal 1 00 as shown in Figure 1 may be constructed such that using via frame or packet transmission data it is all if any
Line and wireless communication system and satellite-based communication system are operated.
The communication system that mobile terminal wherein of the invention can be operated is described referring now to Fig. 2.
Such communication system can use different air interface and/or physical layer.For example, used by communication system
Air interface includes such as frequency division multiple access (FDMA), time division multiple acess (TDMA), CDMA (CDMA) and universal mobile communications system
System (UMTS) (especially, Long Term Evolution (LTE)), global system for mobile communications (GSM) etc..As non-limiting example, under
The description in face is related to cdma communication system, but such teaching is equally applicable to other types of system.
With reference to Fig. 2, cdma wireless communication system can include multiple mobile terminal 1s 00, multiple base station (BS) 270, base station
Controller (BSC) 275 and mobile switching centre (MSC) 280.MSC280 is configured to and Public Switched Telephony Network (PSTN)
290 form interface.MSC280 is also structured to form interface with the BSC275 that can be couple to base station 270 via back haul link.
If any one in the interface that back haul link can be known according to Ganji is constructed, the interface includes such as E1/T1, ATM, IP,
PPP, frame relay, HDSL, ADSL or xDSL.It will be appreciated that system can include multiple as shown in Figure 2
BSC2750。
Each BS270 can service one or more subregions (or region), by multidirectional antenna or the day of sensing specific direction
Each subregion of line covering is radially away from BS270.Or, each subregion can be by two or more for diversity reception
Antenna is covered.Each BS270 may be constructed such that the multiple frequency distribution of support, and the distribution of each frequency has specific frequency spectrum
(for example, 1.25MHz, 5MHz etc.).
What subregion and frequency were distributed intersects can be referred to as CDMA Channel.BS270 can also be referred to as base station transceiver
System (BTS) or other equivalent terms.In this case, term " base station " can be used for broadly representing single
BSC275 and at least one BS270.Base station can also be referred to as " cellular station ".Or, each subregion of specific BS270 can be claimed
It is multiple cellular stations.
As shown in Figure 2, broadcast singal is sent to broadcsting transmitter (BT) 295 mobile terminal operated in system
100.Broadcasting reception module 111 as shown in Figure 1 is arranged at mobile terminal 1 00 to receive the broadcast sent by BT295
Signal.In fig. 2 it is shown that several global positioning system (GPS) satellites 300.Satellite 300 helps position multiple mobile terminals
At least one of 100.
In fig. 2, multiple satellites 300 are depicted, it is understood that be, it is possible to use any number of satellite obtains useful
Location information.GPS module 115 as shown in Figure 1 is generally configured to coordinate with satellite 300 to be believed with obtaining desired positioning
Breath.Substitute GPS tracking techniques or outside GPS tracking techniques, it is possible to use other of the position of mobile terminal can be tracked
Technology.In addition, at least one gps satellite 300 can optionally or additionally process satellite dmb transmission.
Used as a typical operation of wireless communication system, BS270 receives the reverse link from various mobile terminal 1s 00
Signal.Mobile terminal 1 00 generally participates in call, information receiving and transmitting and other types of communication.Each of the reception of certain base station 270 is anti-
Processed in specific BS270 to link signal.The data of acquisition are forwarded to the BSC275 of correlation.BSC provides call
Resource allocation and the mobile management function of the coordination including the soft switching process between BS270.The number that BSC275 will also be received
According to MSC280 is routed to, it provides the extra route service for forming interface with PSTN290.Similarly, PSTN290 with
MSC280 forms interface, and MSC and BSC275 form interface, and BSC275 correspondingly controls BS270 with by forward link signals
It is sent to mobile terminal 1 00.
Based on above-mentioned mobile terminal hardware configuration and communication system, the inventive method each embodiment is proposed.
As shown in figure 3, first embodiment of the invention proposes a kind of method for realizing coding and decoding video, including:
Step 300, by binocular platform obtain original video data stream and deep video data flow.
In this step, binocular platform can include binocular camera, i.e., including two cameras;A shooting can also be included
Head and the equipment (such as active laser device etc.) for being used to obtain depth information.
When binocular platform includes binocular camera;Original video data stream and deep video data are obtained by binocular platform
Stream includes:
First video data stream is obtained by the first camera of binocular camera, while the second shooting for passing through binocular camera
Head obtains the second video data stream;
Using the first video data stream or the second video data stream as original video data stream;
The corresponding figure of each frame second of the image of each frame first and the second video data stream according to the first video data stream
As calculating each frame depth image in deep video data flow;By all range image integration deep video data flows.
Wherein, the first video data stream includes the image of multiframe first, and the second video data stream includes the image of multiframe second, former
Beginning video data stream includes the image of multiframe first or the second image, and deep video data flow includes multiframe depth image.
First image, the second image and depth image are corresponded.
Wherein it is possible to each frame depth image during deep video data flow is calculated using Stereo Matching Algorithm.
When binocular platform is including camera and for the equipment for obtaining depth information;
Obtaining original video data stream and deep video data flow by binocular platform includes:
Original video data stream is obtained by camera, while being regarded by for obtaining the equipment acquisition depth of depth information
Frequency data stream.
Wherein, original video data stream includes the image that multiframe camera is obtained, and deep video data flow is included for obtaining
Take the multiframe depth image that the equipment of depth information is obtained.
The image and depth image that camera is obtained are corresponded.
Step 301, the first coded treatment is carried out to original video data stream using two-dimensional video coding method obtain first
Coded video data stream.
In this step, two-dimensional video coding method can be any related two-dimensional video coding method, and the present invention is implemented
Example is not construed as limiting to this.
Step 302, the first coded video data stream is entered according to deep video data flow and the first coded video data stream
The coded treatment of row second obtains the second coded video data stream.
In this step, the first coded video data stream is entered according to deep video data flow and the first coded video data stream
The coded treatment of row second obtains the second coded video data stream to be included:
Obtain the key message in the first coded video data stream;
The end that deep video data flow and key message write the first coded video data stream is obtained into the 3rd coding to regard
Frequency data stream;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end
Second coded video data stream.
Or, is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream
Two coded treatments obtain the second coded video data stream to be included:
Obtain the key message in the first coded video data stream;
Deep video data flow is carried out after the first coded treatment obtains coded treatment using two-dimensional video coding method
Deep video data flow;
The end that deep video data flow after coded treatment and key message write the first coded video data stream is obtained
To the 3rd coded video data stream;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end
Second coded video data stream.
Or, is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream
Two coded treatments obtain the second coded video data stream to be included:
Obtain the key message in the first coded video data stream;
The head that deep video data flow and key message write the first coded video data stream is obtained into the 3rd coding to regard
Frequency data stream;
Added for indicating whether that the zone bit information with depth information is obtained in the 3rd coded video data stream head
Second coded video data stream.
Or, is carried out to the first coded video data stream according to deep video data flow and the first coded video data stream
Two coded treatments obtain the second coded video data stream to be included:
Obtain the key message in the first coded video data stream;
Deep video data flow is carried out after the first coded treatment obtains coded treatment using two-dimensional video coding method
Deep video data flow;
The head that deep video data flow after coded treatment and key message write the first coded video data stream is obtained
To the 3rd coded video data stream;
Added for indicating whether that the zone bit information with depth information is obtained in the 3rd coded video data stream head
Second coded video data stream.
Wherein, key message includes following any one or more:Video width and video height, frame per second (i.e. unit
The number of image frames included in time), the effective length of the first coded video data stream, the coding lattice of the first coded video data stream
Formula etc..
When being first compressed to original video data stream before the first coded treatment is carried out to original video data stream or
During amplification, key message includes video width and video height, high according to video width and video in follow-up decoding process
Degree is remapped.
Specifically technological means can be known come in obtaining the first coded video data stream using those skilled in the art
Key message, is not intended to limit the present invention the protection domain of embodiment, repeats no more here.
Wherein, zone bit information can be represented using a bit, it is for instance possible to use 0 represents without depth
Information, represents with depth information using 1.Can certainly be represented using other modes, specific method for expressing sheet
Inventive embodiments are not construed as limiting.
Wherein, Fig. 4 (a)~Fig. 4 (d) is four kinds of form schematic diagrams of the second coded video data stream.
As shown in Fig. 4 (a), the second coded video data stream includes the first coded video data stream, deep video number successively
According to stream, key message and zone bit information.
As shown in Fig. 4 (b), after the second coded video data stream includes the first coded video data stream, coded treatment successively
Deep video data flow, key message and zone bit information.
As shown in Fig. 4 (c), the second coded video data stream includes zone bit information, key message, deep video number successively
According to stream and the first coded video data stream.
As shown in Fig. 4 (d), after the second coded video data stream includes zone bit information, key message, coded treatment successively
Deep video data flow and the first coded video data stream.
Optionally, the method also includes:
The second decoding process corresponding with the second coded treatment is carried out to the second coded video data stream and obtains the first coding
Video data stream and deep video data flow;
First encoded video streams are carried out with the first decoding process corresponding with the first coded treatment and obtains original video data
Stream;
Treatment is synchronized to original video data stream and deep video data flow.
Wherein, the second decoding process corresponding with the second coded treatment is carried out to the second coded video data stream and obtains first
Coded video data stream and deep video data flow include:
Read the zone bit information of the second coded video data stream end or head;
Second coded video data stream is judged with depth information according to zone bit information, according to the second encoded video streams
In key message parse the first encoded video streams and deep video data flow.
Or, the second decoding process corresponding with the second coded treatment is carried out to the second coded video data stream and obtains first
Coded video data stream and deep video data flow include:
Read the zone bit information of the second coded video data stream end or head;
Second coded video data stream is judged with depth information according to zone bit information, according to the second encoded video streams
In key message parse the deep video data flow after the first encoded video streams and coded treatment;
The first decoding process corresponding with the first coded treatment is carried out to the deep video data flow after coded treatment to obtain
Deep video data flow.
Wherein, the key message in the second encoded video streams parses the first encoded video streams and deep video data
Stream includes:
The effective length of the first coded video data stream in key message parses the first encoded video streams and depth
Degree video data stream.
Wherein, after the key message in the second encoded video streams parses the first encoded video streams and coded treatment
Deep video data flow includes:
The effective length of the first coded video data stream in key message parses the first encoded video streams and volume
Deep video data flow after code treatment.
Wherein, synchronizing treatment to original video data stream and deep video data flow includes:
Frame per second in key message synchronizes treatment to original video data stream and deep video data flow.Will
Each frame of original video data stream and deep video data flow is corresponded.
When key message includes video width and video height, also needed to be regarded original before treatment is synchronized
Frequency data stream be remapped to original size (i.e. with key message in video width and video height identical size).
Wherein, the first encoded video streams are carried out with the first decoding process corresponding with the first coded treatment and obtains original video
Data flow includes:
The coded format of the first coded video data stream in key message carries out first to the first encoded video streams
Decoding process obtains original video data stream.
Optionally, when the second coded video data stream is judged without depth information according to zone bit information, explanation
The first encoded video streams and zone bit information are only included in second coded video data stream, then in key message first
The coded format of coded video data stream carries out the first decoding process to the first encoded video streams.
By the scheme of the embodiment of the present invention, first first is carried out to original video data stream using two-dimensional video coding method
Coded treatment obtains the first coded video data stream, is encoded to first further according to deep video data flow and original video data stream
Video data stream carries out the second coded treatment, so that compatibility two-dimensional video coded format, it is ensured that three-dimensional video-frequency can be two
Decoded and played on dimension video player.
Referring to Fig. 5, second embodiment of the invention proposes a kind of device for realizing coding and decoding video, including:
Acquisition module, for obtaining original video data stream and deep video data flow by binocular platform;
First coding module, for carrying out the first coded treatment to original video data stream using two-dimensional video coding method
Obtain the first coded video data stream;
Second coding module, for according to deep video data flow and the first coded video data stream to the first encoded video
Data flow carries out the second coded treatment and obtains the second coded video data stream.
Wherein, acquisition module can be realized by Fig. 1 cameras 121 and controller 180, the first coding module and second
Coding module can be arranged in the controller 180 of Fig. 1.
Optionally, binocular platform includes binocular camera;Acquisition module specifically for:
First video data stream is obtained by the first camera of binocular camera, while the second shooting for passing through binocular camera
Head obtains the second video data stream;
Using the first video data stream or the second video data stream as original video data stream;
The corresponding figure of each frame second of the image of each frame first and the second video data stream according to the first video data stream
As calculating each frame depth image in deep video data flow;By all range image integration deep video data flows.
Wherein, the first video data stream includes the image of multiframe first, and the second video data stream includes the image of multiframe second, former
Beginning video data stream includes the image of multiframe first or the second image, and deep video data flow includes multiframe depth image.
First image, the second image and depth image are corresponded.
Wherein, acquisition module can calculate each frame depth map in deep video data flow using Stereo Matching Algorithm
Picture.
Optionally, binocular platform includes camera and the equipment for obtaining depth information;
Acquisition module specifically for:
Original video data stream is obtained by camera, while being regarded by for obtaining the equipment acquisition depth of depth information
Frequency data stream.
Wherein, original video data stream includes the image that multiframe camera is obtained, and deep video data flow is included for obtaining
Take the multiframe depth image that the equipment of depth information is obtained.
The image and depth image that camera is obtained are corresponded.
Optionally, the second coding module specifically for:
Obtain the key message in the first coded video data stream;
The end that deep video data flow and key message write the first coded video data stream is obtained into the 3rd coding to regard
Frequency data stream;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end
Second coded video data stream.
Optionally, the second coding module specifically for:
Obtain the key message in the first coded video data stream;
Deep video data flow is carried out after the first coded treatment obtains coded treatment using two-dimensional video coding method
Deep video data flow;
The end that deep video data flow after coded treatment and key message write the first coded video data stream is obtained
To the 3rd coded video data stream;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end
Second coded video data stream.
Optionally, the second coding module specifically for:
Obtain the key message in the first coded video data stream;
The head that deep video data flow and key message write the first coded video data stream is obtained into the 3rd coding to regard
Frequency data stream;
Added for indicating whether that the zone bit information with depth information is obtained in the 3rd coded video data stream head
Second coded video data stream.
Optionally, the second coding module specifically for:
Obtain the key message in the first coded video data stream;
Deep video data flow is carried out after the first coded treatment obtains coded treatment using two-dimensional video coding method
Deep video data flow;
The head that deep video data flow after coded treatment and key message write the first coded video data stream is obtained
To the 3rd coded video data stream;
Added for indicating whether that the zone bit information with depth information is obtained in the 3rd coded video data stream head
Second coded video data stream.
Wherein, key message includes following any one or more:Video width and video height, frame per second (i.e. unit
The number of image frames included in time), the effective length of the first coded video data stream, the coding lattice of the first coded video data stream
Formula etc..
When being first compressed to original video data stream before the first coded treatment is carried out to original video data stream or
During amplification, key message includes video width and video height, high according to video width and video in follow-up decoding process
Degree is remapped.
Specifically technological means can be known come in obtaining the first coded video data stream using those skilled in the art
Key message, is not intended to limit the present invention the protection domain of embodiment, repeats no more here.
Wherein, zone bit information can be represented using a bit, it is for instance possible to use 0 represents without depth
Information, represents with depth information using 1.Can certainly be represented using other modes, specific method for expressing sheet
Inventive embodiments are not construed as limiting.
Wherein, Fig. 4 (a)~Fig. 4 (d) is four kinds of form schematic diagrams of the second coded video data stream.
As shown in Fig. 4 (a), the second coded video data stream includes the first coded video data stream, deep video number successively
According to stream, key message and zone bit information.
As shown in Fig. 4 (b), after the second coded video data stream includes the first coded video data stream, coded treatment successively
Deep video data flow, key message and zone bit information.
As shown in Fig. 4 (c), the second coded video data stream includes zone bit information, key message, deep video number successively
According to stream and the first coded video data stream.
As shown in Fig. 4 (d), after the second coded video data stream includes zone bit information, key message, coded treatment successively
Deep video data flow and the first coded video data stream.
Optionally, also include:
Second decoder module, for carrying out the second decoding corresponding with the second coded treatment to the second coded video data stream
Treatment obtains the first coded video data stream and deep video data flow;
First decoder module, for the first encoded video streams to be carried out with the first decoding process corresponding with the first coded treatment
Obtain original video data stream;
Synchronous processing module, for synchronizing treatment to original video data stream and deep video data flow.
Optionally, the second decoder module specifically for:
Read the zone bit information of the second coded video data stream end or head;
Second coded video data stream is judged with depth information according to zone bit information, according to the second encoded video streams
In key message parse the first encoded video streams and deep video data flow.
Optionally, the second decoder module specifically for:
Read the zone bit information of the second coded video data stream end or head;
Second coded video data stream is judged with depth information according to zone bit information, according to the second encoded video streams
In key message parse the deep video data flow after the first encoded video streams and coded treatment;
The first decoding process corresponding with the first coded treatment is carried out to the deep video data flow after coded treatment to obtain
Deep video data flow.
It should be noted that herein, term " including ", "comprising" or its any other variant be intended to non-row
His property is included, so that process, method, article or device including a series of key elements not only include those key elements, and
And also include other key elements being not expressly set out, or also include for this process, method, article or device institute are intrinsic
Key element.In the absence of more restrictions, the key element limited by sentence "including a ...", it is not excluded that including this
Also there is other identical element in the process of key element, method, article or device.
The embodiments of the present invention are for illustration only, and the quality of embodiment is not represented.
Through the above description of the embodiments, those skilled in the art can be understood that above-described embodiment side
Method can add the mode of required general hardware platform to realize by software, naturally it is also possible to by hardware, but in many cases
The former is more preferably implementation method.Based on such understanding, technical scheme is substantially done to prior art in other words
The part for going out contribution can be embodied in the form of software product, and the computer software product is stored in a storage medium
In (such as ROM/RAM, magnetic disc, CD), including some instructions are used to so that a station terminal equipment (can be mobile phone, computer, clothes
Business device, air-conditioner, or network equipment etc.) perform method described in each embodiment of the invention.
The preferred embodiments of the present invention are these are only, the scope of the claims of the invention is not thereby limited, it is every to utilize this hair
Equivalent structure or equivalent flow conversion that bright specification and accompanying drawing content are made, or directly or indirectly it is used in other related skills
Art field, is included within the scope of the present invention.
Claims (10)
1. a kind of device for realizing coding and decoding video, it is characterised in that including:
Acquisition module, for obtaining original video data stream and deep video data flow by binocular platform;
First coding module, obtains for carrying out the first coded treatment to original video data stream using two-dimensional video coding method
First coded video data stream;
Second coding module, for according to deep video data flow and the first coded video data stream to the first coding video frequency data
Stream carries out the second coded treatment and obtains the second coded video data stream.
2. device according to claim 1, it is characterised in that the binocular platform includes binocular camera;The acquisition mould
Block specifically for:
First video data stream is obtained by the first camera of binocular camera, while the second shooting for passing through the binocular camera
Head obtains the second video data stream;
Using first video data stream or second video data stream as the original video data stream;
The image of each frame first and the corresponding each frame of second video data stream according to first video data stream
Two images calculate each frame depth image in the deep video data flow;By deep video described in all range image integrations
Data flow.
3. device according to claim 1, it is characterised in that the binocular platform includes camera and for obtaining depth
The equipment of information;
The acquisition module specifically for:
The original video data stream is obtained by the camera, while being obtained for obtaining the equipment of depth information by described
Take the deep video data flow.
4. device according to claim 1, it is characterised in that second coding module specifically for:
Obtain the key message in first coded video data stream;
The end that the deep video data flow and the key message write first coded video data stream is obtained the
Three coded video data streams;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end
Second coded video data stream.
5. the device according to Claims 1 to 4 any one, it is characterised in that also include:
Second decoder module, for being carried out and second coded treatment corresponding second to second coded video data stream
Decoding process obtains first coded video data stream and the deep video data flow;
First decoder module, for first encoded video streams to be carried out with the first decoding corresponding with first coded treatment
Treatment obtains the original video data stream;
Synchronous processing module, for synchronizing treatment to the original video data stream and the deep video data flow.
6. a kind of method for realizing coding and decoding video, it is characterised in that including:
Original video data stream and deep video data flow are obtained by binocular platform;
First coded treatment is carried out to original video data stream using two-dimensional video coding method and obtains the first coding video frequency data
Stream;
The first coded video data stream is carried out at the second coding according to deep video data flow and the first coded video data stream
Reason obtains the second coded video data stream.
7. method according to claim 6, it is characterised in that the binocular platform includes binocular camera;It is described by double
Head platform obtains original video data stream and deep video data flow to be included:
First video data stream is obtained by the first camera of binocular camera, while the second shooting for passing through the binocular camera
Head obtains the second video data stream;
Using first video data stream or second video data stream as the original video data stream;
The image of each frame first and the corresponding each frame of second video data stream according to first video data stream
Two images calculate each frame depth image in the deep video data flow;By deep video described in all range image integrations
Data flow.
8. method according to claim 6, it is characterised in that the binocular platform includes camera and for obtaining depth
The equipment of information;
It is described to be included by binocular platform acquisition original video data stream and deep video data flow:
The original video data stream is obtained by the camera, while being obtained for obtaining the equipment of depth information by described
Take the deep video data flow.
9. method according to claim 6, it is characterised in that described according to deep video data flow and the first encoded video
Data flow carries out the second coded treatment to the first coded video data stream and obtains the second coded video data stream including:
Obtain the key message in first coded video data stream;
The end that the deep video data flow and the key message write first coded video data stream is obtained the
Three coded video data streams;
Added for indicating whether that the zone bit information with depth information is obtained at the 3rd coded video data stream end
Second coded video data stream.
10. the method according to claim 6~9 any one, it is characterised in that the method also includes:
The second decoding process corresponding with second coded treatment is carried out to second coded video data stream and obtains described
First coded video data stream and the deep video data flow;
First encoded video streams are carried out with the first decoding process corresponding with first coded treatment and obtains described original
Video data stream;
Treatment is synchronized to the original video data stream and the deep video data flow.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201611083136.9A CN106791764A (en) | 2016-11-30 | 2016-11-30 | A kind of method and apparatus for realizing coding and decoding video |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201611083136.9A CN106791764A (en) | 2016-11-30 | 2016-11-30 | A kind of method and apparatus for realizing coding and decoding video |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN106791764A true CN106791764A (en) | 2017-05-31 |
Family
ID=58901606
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201611083136.9A Pending CN106791764A (en) | 2016-11-30 | 2016-11-30 | A kind of method and apparatus for realizing coding and decoding video |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN106791764A (en) |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP1389020A1 (en) * | 2002-08-07 | 2004-02-11 | Electronics and Telecommunications Research Institute | Method and apparatus for multiplexing multi-view three-dimensional moving picture |
| CN1954606A (en) * | 2004-05-21 | 2007-04-25 | 韩国电子通信研究院 | Apparatus and method for transmitting/receiving 3D stereoscopic digital broadcasting signal by using 3D stereoscopic video additional data |
| CN101459857A (en) * | 2007-12-10 | 2009-06-17 | 深圳华为通信技术有限公司 | Communication terminal and information system |
| CN102246529A (en) * | 2008-12-15 | 2011-11-16 | 皇家飞利浦电子股份有限公司 | Image based 3D video format |
| CN102326390A (en) * | 2009-02-19 | 2012-01-18 | 汤姆逊许可证公司 | 3D video format |
| CN104205821A (en) * | 2012-04-05 | 2014-12-10 | 皇家飞利浦有限公司 | Depth helper data |
-
2016
- 2016-11-30 CN CN201611083136.9A patent/CN106791764A/en active Pending
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP1389020A1 (en) * | 2002-08-07 | 2004-02-11 | Electronics and Telecommunications Research Institute | Method and apparatus for multiplexing multi-view three-dimensional moving picture |
| CN1954606A (en) * | 2004-05-21 | 2007-04-25 | 韩国电子通信研究院 | Apparatus and method for transmitting/receiving 3D stereoscopic digital broadcasting signal by using 3D stereoscopic video additional data |
| CN101459857A (en) * | 2007-12-10 | 2009-06-17 | 深圳华为通信技术有限公司 | Communication terminal and information system |
| CN102246529A (en) * | 2008-12-15 | 2011-11-16 | 皇家飞利浦电子股份有限公司 | Image based 3D video format |
| CN102326390A (en) * | 2009-02-19 | 2012-01-18 | 汤姆逊许可证公司 | 3D video format |
| CN104205821A (en) * | 2012-04-05 | 2014-12-10 | 皇家飞利浦有限公司 | Depth helper data |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN106485689B (en) | A kind of image processing method and device | |
| CN104866265B (en) | Multi-media file display method and device | |
| CN104735255B (en) | Multi-screen display method and system | |
| KR101708306B1 (en) | Mobile twrminal and 3d image convergence method thereof | |
| CN108111874B (en) | file processing method, terminal and server | |
| CN105828180A (en) | Apparatus and method for caching video frames | |
| CN105100673B (en) | Desktop sharing method and device based on VoLTE | |
| CN106713640B (en) | A kind of brightness adjusting method and equipment | |
| CN106155694A (en) | A kind of application and the display packing and device applied of attending to anything else | |
| CN106534590A (en) | Photo processing method and apparatus, and terminal | |
| CN106200924B (en) | A kind of terminal and its method that virtual reality video is controlled | |
| CN106250081A (en) | A kind of display packing based on double screen terminal and device | |
| CN105187911A (en) | Method and device for displaying video image and image display method | |
| CN106791119B (en) | Photo processing method and device and terminal | |
| CN104935822A (en) | Method and device for processing images | |
| CN105635837B (en) | A kind of video broadcasting method and device | |
| CN106303259A (en) | A kind of method and apparatus realizing taking pictures | |
| CN105049916B (en) | A kind of video recording method and device | |
| CN106230804A (en) | Picture sharing method, terminal and server | |
| CN106651773A (en) | Picture processing method and device | |
| CN106341687A (en) | Method and device for locating video decoding abnormality | |
| CN106657950A (en) | Projection device management device, method and projection data sharing device | |
| CN106060522A (en) | Video image processing device and method | |
| CN106534203A (en) | Mobile terminal and communication method | |
| CN106651762A (en) | Photo processing method, device and terminal |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| RJ01 | Rejection of invention patent application after publication |
Application publication date: 20170531 |
|
| RJ01 | Rejection of invention patent application after publication |