[go: up one dir, main page]

CN107438180B - Depth perception quality evaluation method for 3D video - Google Patents

Depth perception quality evaluation method for 3D video Download PDF

Info

Publication number
CN107438180B
CN107438180B CN201710751972.8A CN201710751972A CN107438180B CN 107438180 B CN107438180 B CN 107438180B CN 201710751972 A CN201710751972 A CN 201710751972A CN 107438180 B CN107438180 B CN 107438180B
Authority
CN
China
Prior art keywords
depth
image
distortion
evaluated
left view
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710751972.8A
Other languages
Chinese (zh)
Other versions
CN107438180A (en
Inventor
张云
樊春玲
李娜
张欢
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Institute of Advanced Technology of CAS
Original Assignee
Shenzhen Institute of Advanced Technology of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Institute of Advanced Technology of CAS filed Critical Shenzhen Institute of Advanced Technology of CAS
Priority to CN201710751972.8A priority Critical patent/CN107438180B/en
Publication of CN107438180A publication Critical patent/CN107438180A/en
Application granted granted Critical
Publication of CN107438180B publication Critical patent/CN107438180B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

The present invention relates to a kind of depth perception quality evaluating methods of 3 D video, corresponding with image to be evaluated with reference to 3-D image by choosing;Key point matching is carried out by the left view of image to be evaluated and with reference to the left view of 3-D image, and calculates the depth distortion of image left view to be evaluated;Key point matching is carried out by the right view of image to be evaluated and with reference to the right view of 3-D image, and calculates the depth distortion of image right view to be evaluated;Critical point detection is carried out to image to be evaluated and with reference to 3-D image respectively, and calculates the binocular depth distortion of image to be evaluated;The depth distortion of left view, the depth distortion of right view and binocular depth distortion are merged, depth quality distortion is obtained;Depth quality distortion is converted into depth perception quality, it is thus possible to which distortion evaluation is carried out to image to be evaluated according to depth perception quality.

Description

The depth perception quality evaluating method of 3 D video
Technical field
The present invention relates to the distortion of 3 D video evaluations, more particularly to a kind of depth perception quality evaluation of 3 D video Method.
Background technique
3 D video can be supplied to viewer's more true 3D vision impression, in recent years, 3 D video increasingly by To liking for spectators.Since the data volume of 3 D video is huge, part has to reduce to save bandwidth when on-line normalization Image quality.Depth perception is one of most important feature of 3 D stereo video and the main distinction of itself and two-dimensional video.Its matter Amount, referred to as depth perception quality or depth quality are to directly affect 3 D video together with picture quality and visual comfort One of big factor of key three of Quality of experience.3 D video is needed to compress and be handled, and is compressed and introduced distortion in treatment process, These distortions will directly affect depth perception quality again.
Summary of the invention
Based on this, it is necessary to provide a kind of depth perception quality evaluating method of 3 D video.
A kind of depth perception quality evaluating method of 3 D video, comprising the following steps:
It chooses corresponding with image to be evaluated with reference to 3-D image;
Key point matching is carried out by the left view of image to be evaluated and with reference to the left view of 3-D image, and is calculated to be evaluated The depth of valence image left view is distorted;
Key point matching is carried out by the right view of image to be evaluated and with reference to the right view of 3-D image, and is calculated to be evaluated The depth of valence image right view is distorted;
Critical point detection is carried out to image to be evaluated and with reference to 3-D image respectively, and calculates the binocular of image to be evaluated Depth distortion;
The depth distortion of the left view, the depth distortion of the right view and binocular depth distortion are melted It closes, obtains depth quality distortion;
The depth quality distortion is converted into depth perception quality, and according to the depth perception quality to described to be evaluated Valence image carries out distortion evaluation.
The left view by image to be evaluated and the left view with reference to 3-D image carry out in one of the embodiments, Key point matching, and the step of calculating the depth distortion of left view includes:
Calculate the distortion of left view Depth cue;
Obtain the set of keypoints detected;
The lower decreasing concentration of the left view Depth cue is calculated according to the set of keypoints;
The depth distortion of the left view is judged according to the lower decreasing concentration of the left view Depth cue.
In one of the embodiments, it is described calculate left view Depth cue distortion the step of include:
Using formulaCalculate the left view Depth cue Distortion, wherein φ (I) be receptive field in image Ι monocular vision information, Ω () be comparison function,For figure to be evaluated The left view of picture,For the corresponding left view with reference to 3-D image.
It is described in one of the embodiments, to be calculated under the left view Depth cue according to the set of keypoints The step of decreasing concentration includes:
Using formulaCalculate the left view The lower decreasing concentration of figure Depth cue, wherein Num () is the number of element in set,For distortion 3-D image left view inspection The set of characteristic points measured,For the set of characteristic points detected with reference to 3-D image left view.
The right view by image to be evaluated and the right view with reference to 3-D image carry out in one of the embodiments, Key point matching, and the step of calculating the depth distortion of right view includes:
Calculate the distortion of right view Depth cue;
Obtain the set of keypoints detected;
The lower decreasing concentration of the right view Depth cue is calculated according to the set of keypoints;
The depth distortion of the right view is calculated according to the judgement of the lower decreasing concentration of the right view Depth cue.
In one of the embodiments, it is described calculate right view Depth cue distortion the step of include:
Using formulaCalculate the left view Depth cue Distortion, wherein φ (I) is image Ι monocular vision information in receptive field, and Ω () is comparison function,For image to be evaluated Left view,For the corresponding left view with reference to 3-D image.
It is described in one of the embodiments, to be calculated under the right view Depth cue according to the set of keypoints The step of decreasing concentration includes:
Using formulaCalculate the right view The lower decreasing concentration of figure Depth cue, wherein Num () is the number of element in set,For distortion 3-D image right view inspection The set of characteristic points measured,For the set of characteristic points detected with reference to 3-D image right view.
It is described in one of the embodiments, to carry out critical point detection to image to be evaluated and with reference to 3-D image respectively, And the step of calculating the binocular depth distortion of image to be evaluated, includes:
Calculate the distortion of the binocular depth clue of image to be evaluated;
Obtain the set of binocular ranging point pair;
The lower decreasing concentration of the binocular depth clue is calculated according to the set of the binocular ranging point pair;
The depth distortion of the image to be evaluated is judged according to the lower decreasing concentration of the binocular depth clue.
The step of distortion of the binocular depth clue for calculating image to be evaluated in one of the embodiments, includes:
Using formulaIt is deep to calculate the binocular Spend the distortion of clue, wherein Θ (IL, IR) it is receptive field neutral body image to (IL, IR) binocular vision information, Ω () be than Compared with function,For image pair to be evaluated,3-D image pair is referred to be corresponding.
The set according to the binocular ranging point pair calculates the binocular depth line in one of the embodiments, The step of lower decreasing concentration of rope includes:
Using formulaIt is deep to calculate the binocular Spend the lower decreasing concentration of clue, wherein Num () is the number of element in set, MDisThe characteristic point detected for distortion 3-D image Set, MRefFor the set of characteristic points detected with reference to 3-D image.
The depth perception quality evaluating method of above-mentioned 3 D video is corresponding with image to be evaluated with reference to three-dimensional by choosing Image;Key point matching is carried out by the left view of image to be evaluated and with reference to the left view of 3-D image, and is calculated to be evaluated The depth of image left view is distorted;Key point is carried out by the right view of image to be evaluated and with reference to the right view of 3-D image Match, and calculates the depth distortion of image right view to be evaluated;It is carried out respectively to image to be evaluated and with reference to 3-D image crucial Point detection, and calculate the binocular depth distortion of image to be evaluated;By the distortion of the depth of left view, the depth distortion of right view and Binocular depth distortion is merged, and depth quality distortion is obtained;Depth quality distortion is converted into depth perception quality, thus energy It is enough that distortion evaluation is carried out to image to be evaluated according to depth perception quality.
Detailed description of the invention
Fig. 1 is the flow chart of the depth perception quality evaluating method of 3 D video;
Fig. 2 is the schematic diagram of the depth perception quality evaluating method of 3 D video.
Specific embodiment
To facilitate the understanding of the present invention, a more comprehensive description of the invention is given in the following sections with reference to the relevant attached drawings.In attached drawing Give presently preferred embodiments of the present invention.But the invention can be realized in many different forms, however it is not limited to this paper institute The embodiment of description.On the contrary, purpose of providing these embodiments is keeps the understanding to the disclosure more thorough Comprehensively.
Unless otherwise defined, all technical and scientific terms used herein and belong to technical field of the invention The normally understood meaning of technical staff is identical.Term as used herein in the specification of the present invention is intended merely to description tool The purpose of the embodiment of body, it is not intended that the limitation present invention.Term as used herein "and/or" includes one or more related Listed item any and all combinations.
As shown in Figure 1, the flow chart of the depth perception quality evaluating method for 3 D video.
A kind of depth perception quality evaluating method of 3 D video, comprising the following steps:
Step S110 chooses corresponding with image to be evaluated with reference to 3-D image.
Due to being to carry out distortion computation processing to image to be evaluated, therefore, it is necessary to choose corresponding 3-D image work For references object.
Image fault will lead to loss in detail, so as to cause the reduction of monocular and binocular depth clue, i.e. human eye vision system The receptive field of system visual information relevant to depth perception reduces.Thus, it the monocular depth to image to be evaluated can lose respectively The distortion of true and binocular depth is calculated, wherein monocular depth distortion refers to the distortion of left view depth and the distortion of right view depth.
Step S120 carries out key point matching by the left view of image to be evaluated and with reference to the left view of 3-D image, and Calculate the depth distortion of image left view to be evaluated.
Specifically, step S120 includes:
1. calculating the distortion of left view Depth cue.
2. obtaining the set of keypoints detected.
3. calculating the lower decreasing concentration of the left view Depth cue according to the set of keypoints.
4. judging that the depth of the left view is distorted according to the lower decreasing concentration of the left view Depth cue.
Calculate left view Depth cue distortion the step of include:
Using formulaCalculate the left view Depth cue Distortion, wherein φ (I) be receptive field in image Ι monocular vision information, Ω () be comparison function,For figure to be evaluated The left view of picture,For the corresponding left view with reference to 3-D image.
The step of calculating the lower decreasing concentration of the left view Depth cue according to the set of keypoints include:
Using formulaCalculate the left view The lower decreasing concentration of figure Depth cue, wherein Num () is the number of element in set,For distortion 3-D image left view inspection The set of characteristic points measured,For the set of characteristic points detected with reference to 3-D image left view.
In the present embodiment, the number of key point, that is, monocular vision information is reduced with the aggravation of distortion level.
Step S130 carries out key point matching by the right view of image to be evaluated and with reference to the right view of 3-D image, and Calculate the depth distortion of image right view to be evaluated.
Specifically, step S130 includes:
1. calculating the distortion of right view Depth cue.
2. obtaining the set of keypoints detected.
3. calculating the lower decreasing concentration of the right view Depth cue according to the set of keypoints.
4. being distorted according to the depth that the judgement of the lower decreasing concentration of the right view Depth cue calculates the right view.
Calculate right view Depth cue distortion the step of include:
Using formulaCalculate the left view Depth cue Distortion, wherein φ (I) is image Ι monocular vision information in receptive field, and Ω () is comparison function,For image to be evaluated Left view,For the corresponding left view with reference to 3-D image.
The step of calculating the lower decreasing concentration of the right view Depth cue according to the set of keypoints include:
Using formulaCalculate the right view The lower decreasing concentration of figure Depth cue, wherein Num () is the number of element in set,For distortion 3-D image right view inspection The set of characteristic points measured,For the set of characteristic points detected with reference to 3-D image right view.
In the present embodiment, the number of key point, that is, monocular vision information is reduced with the aggravation of distortion level.
Step S140 carries out critical point detection and matching to image to be evaluated and with reference to 3-D image respectively, and calculates The binocular depth of image to be evaluated is distorted.
Specifically, step S140 includes:
1. calculating the distortion of the binocular depth clue of image to be evaluated.
2. obtaining the set of binocular ranging point pair.
3. calculating the lower decreasing concentration of the binocular depth clue according to the set of the binocular ranging point pair.
4. judging that the depth of the image to be evaluated is distorted according to the lower decreasing concentration of the binocular depth clue.
The step of calculating the distortion of the binocular depth clue of image to be evaluated include:
Using formulaIt is deep to calculate the binocular Spend the distortion of clue, wherein Θ (IL, IR) it is receptive field neutral body image to (IL, IR) binocular vision information, Ω () be than Compared with function,For image pair to be evaluated,3-D image pair is referred to be corresponding.
The step of calculating the lower decreasing concentration of the binocular depth clue according to the set of the binocular ranging point pair include:
Using formulaIt is deep to calculate the binocular Spend the lower decreasing concentration of clue, wherein Num () is the number of element in set, MDisThe characteristic point detected for distortion 3-D image Set, MRefFor the set of characteristic points detected with reference to 3-D image.
In the present embodiment, the number of key point, that is, monocular vision information is reduced with the aggravation of distortion level.
Step S150 loses the depth distortion of the left view, the depth distortion of the right view and the binocular depth It is really merged, obtains total depth quality distortion.
In the present embodiment, using formula Δ D=Ψ (Δ DM, Δ DB) calculate total depth quality distortion, wherein Δ DM Including Δ DMLWith Δ DMR, Ψ () is the fusion function for integrating monocular and binocular depth distortion.
In one embodiment, formula can also be usedCalculate total depth quality distortion, ω1、ω2And ω3To adjust The weight parameter of whole each element relative importance.
Based on above-mentioned all embodiments, it to be capable of measuring the depth perception distortion of video, key, which is to find, is able to reflect sense By the monocular of Yezhong stereo-picture and the suitable visual signature of binocular vision information, i.e. φ (I) and Θ (IL, IR).In the present invention It is middle using the receptive field for using Difference of Gaussian filter analog vision nerve, then with SIFT (Scale-invariant feature Transform, Scale invariant features transform) description sublist reach.
Since 3 D video is usually shot by horizontally arranged camera, it is therefore necessary to exclude wherein obvious displacement errors Matching double points.It is vertical that the present embodiment uses random sampling unification algorism (RANdom Sample Consensus, RANSAC) calculating Affine transformation matrices between body image pair are to exclude abnormal point pair.If the parallax value of certain matching double points has exceeded some Zone of reasonableness, then these points were corresponding to be excluded before calculating affine transformation matrix.The present embodiment hangs down matching double points Straight and horizontal parallax threshold value is set to TVAnd TH
The depth quality distortion is converted into depth perception quality by step S160, and according to the depth perception quality Distortion evaluation is carried out to the image to be evaluated.
Specifically, depth quality distortion Δ D is converted to depth perception quality DMOS using logistic functionp
The formula of use are as follows:
In the case where symmetrical distortion and asymmetric distortion, curve matching goodness (R2) it is respectively 0.826 and 0.661.Two Three-dimensional image the quality evaluation algorithm FI-PSNR and MD3DQA of the current mainstream of kind be used to carry out performance comparison, left and right view The mean value of the PSNE and SSIM of figure are also used as comparison benchmark.The criterion of evaluation algorithms performance is to calculate the algorithm to predict Pearson correlation coefficients (PLCC) between DMOSp and practical DMOS, Spearman rank correlation coefficient (SROCC) and root mean square Error (RMSE).PLCC, SROCC and the RMSE of the present embodiment algorithm on symmetrical distortion data set are respectively 0.909,0.904 With 0.038, asymmetric data and on PLCC, SROCC and RMSE be respectively 0.813,0.750 and 0.036, comparing result is shown in Table 1.Experiment shows that method disclosed by the invention can accurately pre- depth measurement on symmetrical distortion data set and asymmetric data collection Spend the decline of perceived quality.
The performance of 1. depth perception quality evaluation algorithm of table compares (optimum of each column is indicated using runic)
Incorporated by reference to Fig. 2.
Based on above-mentioned all embodiments, the specific workflow of the depth perception quality evaluating method of 3 D video is as follows: Selection is corresponding with image to be evaluated to refer to 3-D image, the left view of left view and image to be evaluated to reference 3-D image Key point matching is carried out, to calculate the depth distortion of left view.To the right view and image to be evaluated of reference 3-D image Right view carry out key point matching, thus calculate right view depth distortion.Individually calculate the depth distortion of left and right view Belong to and calculates monocular depth distortion.And critical point detection directly is carried out to reference 3-D image and image to be evaluated, it will be able to count Calculate binocular depth distortion.After being merged to monocular depth distortion and binocular depth distortion, image to be evaluated can be carried out Depth perception quality evaluation.
The depth perception quality evaluating method of above-mentioned 3 D video is corresponding with image to be evaluated with reference to three-dimensional by choosing Image;Key point matching is carried out by the left view of image to be evaluated and with reference to the left view of 3-D image, and is calculated to be evaluated The depth of image left view is distorted;Key point is carried out by the right view of image to be evaluated and with reference to the right view of 3-D image Match, and calculates the depth distortion of image right view to be evaluated;It is carried out respectively to image to be evaluated and with reference to 3-D image crucial Point detection, and calculate the binocular depth distortion of image to be evaluated;By the distortion of the depth of left view, the depth distortion of right view and Binocular depth distortion is merged, and depth quality distortion is obtained;Depth quality distortion is converted into depth perception quality, thus energy It is enough that distortion evaluation is carried out to image to be evaluated according to depth perception quality.
Each technical characteristic of embodiment described above can be combined arbitrarily, for simplicity of description, not to above-mentioned reality It applies all possible combination of each technical characteristic in example to be all described, as long as however, the combination of these technical characteristics is not deposited In contradiction, all should be considered as described in this specification.
The embodiments described above only express several embodiments of the present invention, and the description thereof is more specific and detailed, but simultaneously It cannot therefore be construed as limiting the scope of the patent.It should be pointed out that coming for those of ordinary skill in the art It says, without departing from the inventive concept of the premise, various modifications and improvements can be made, these belong to protection of the invention Range.Therefore, the scope of protection of the patent of the invention shall be subject to the appended claims.

Claims (10)

1. a kind of depth perception quality evaluating method of 3 D video, which comprises the following steps:
It chooses corresponding with image to be evaluated with reference to 3-D image;
Key point matching is carried out by the left view of image to be evaluated and with reference to the left view of 3-D image, and calculates figure to be evaluated As the depth of left view is distorted;
Key point matching is carried out by the right view of image to be evaluated and with reference to the right view of 3-D image, and calculates figure to be evaluated As the depth of right view is distorted;
Critical point detection and matching are carried out to image to be evaluated and with reference to 3-D image respectively, and calculates the double of image to be evaluated The distortion of mesh depth;
The depth distortion of the left view, the depth distortion of the right view and binocular depth distortion are merged, obtained Take depth quality distortion;
The depth quality distortion is converted into depth perception quality, and according to the depth perception quality to the figure to be evaluated As carrying out distortion evaluation.
2. the depth perception quality evaluating method of 3 D video according to claim 1, which is characterized in that it is described will be to be evaluated What the left view of valence image and the depth for carrying out key point matching with reference to the left view of 3-D image, and calculating left view were distorted Step includes:
Calculate the distortion of left view Depth cue;
Obtain the set of keypoints detected;
The lower decreasing concentration of the left view Depth cue is calculated according to the set of keypoints;
The depth distortion of the left view is judged according to the lower decreasing concentration of the left view Depth cue.
3. the depth perception quality evaluating method of 3 D video according to claim 2, which is characterized in that the calculating is left The step of distortion of view Depth cue includes:
Using formulaCalculate the distortion of the left view Depth cue, wherein φ It (I) is image Ι monocular vision information in receptive field, Ω () is comparison function,For the left view of image to be evaluated, For the corresponding left view with reference to 3-D image.
4. the depth perception quality evaluating method of 3 D video according to claim 2, which is characterized in that described according to institute Stating the step of set of keypoints calculates the lower decreasing concentration of the left view Depth cue includes:
Using formulaThe lower decreasing concentration of the left view Depth cue is calculated, In, Num () is the number of element in set,To be distorted the set of characteristic points that 3-D image left view detects, For the set of characteristic points detected with reference to 3-D image.
5. the depth perception quality evaluating method of 3 D video according to claim 1, which is characterized in that it is described will be to be evaluated What the right view of valence image and the depth for carrying out key point matching with reference to the right view of 3-D image, and calculating right view were distorted Step includes:
Calculate the distortion of right view Depth cue;
Obtain the set of keypoints detected;
The lower decreasing concentration of the right view Depth cue is calculated according to the set of keypoints;
The depth distortion of the right view is calculated according to the judgement of the lower decreasing concentration of the right view Depth cue.
6. the depth perception quality evaluating method of 3 D video according to claim 5, which is characterized in that the calculating is right The step of distortion of view Depth cue includes:
Using formulaCalculate the distortion of the right view Depth cue, wherein φ It (I) is image Ι monocular vision information in receptive field, Ω () is comparison function,For the right view of image to be evaluated, For the corresponding right view with reference to 3-D image.
7. the depth perception quality evaluating method of 3 D video according to claim 5, which is characterized in that described according to institute Stating the step of set of keypoints calculates the lower decreasing concentration of the right view Depth cue includes:
Using formulaThe lower decreasing concentration of the right view Depth cue is calculated, In, Num () is the number of element in set,To be distorted the set of characteristic points that 3-D image right view detects, For the set of characteristic points detected with reference to 3-D image right view.
8. the depth perception quality evaluating method of 3 D video according to claim 1, which is characterized in that described right respectively Image to be evaluated and reference 3-D image carry out critical point detection and matching, and calculate the binocular depth distortion of image to be evaluated The step of include:
Calculate the distortion of the binocular depth clue of image to be evaluated;
Obtain the set of binocular ranging point pair;
The lower decreasing concentration of the binocular depth clue is calculated according to the set of the binocular ranging point pair;
The depth distortion of the image to be evaluated is judged according to the lower decreasing concentration of the binocular depth clue.
9. the depth perception quality evaluating method of 3 D video according to claim 8, which is characterized in that it is described calculate to Evaluate image binocular depth clue distortion the step of include:
Using formulaThe distortion of the binocular depth clue is calculated, Wherein, Θ (IL, IR) it is receptive field neutral body image to (IL, IR) binocular vision information, Ω () be comparison function,For image pair to be evaluated,3-D image pair is referred to be corresponding.
10. the depth perception quality evaluating method of 3 D video according to claim 8, which is characterized in that the basis The step of set of the binocular ranging point pair calculates the lower decreasing concentration of the binocular depth clue include:
Using formulaCalculate the lower decreasing concentration of the binocular depth clue, wherein Num () is the number of element in set, MDisFor the set of characteristic points that distortion 3-D image detects, MRefFor with reference to three-dimensional figure As the set of characteristic points detected.
CN201710751972.8A 2017-08-28 2017-08-28 Depth perception quality evaluation method for 3D video Active CN107438180B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710751972.8A CN107438180B (en) 2017-08-28 2017-08-28 Depth perception quality evaluation method for 3D video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710751972.8A CN107438180B (en) 2017-08-28 2017-08-28 Depth perception quality evaluation method for 3D video

Publications (2)

Publication Number Publication Date
CN107438180A CN107438180A (en) 2017-12-05
CN107438180B true CN107438180B (en) 2019-02-22

Family

ID=60460710

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710751972.8A Active CN107438180B (en) 2017-08-28 2017-08-28 Depth perception quality evaluation method for 3D video

Country Status (1)

Country Link
CN (1) CN107438180B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114648482B (en) * 2020-12-19 2025-03-21 中国科学院深圳先进技术研究院 Quality evaluation method and system for stereoscopic panoramic images
CN114841960B (en) * 2022-05-05 2025-04-29 咪咕视讯科技有限公司 Three-dimensional video quality evaluation method, device, equipment and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102170581A (en) * 2011-05-05 2011-08-31 天津大学 Human-visual-system (HVS)-based structural similarity (SSIM) and characteristic matching three-dimensional image quality evaluation method
CN102333233A (en) * 2011-09-23 2012-01-25 宁波大学 An Objective Evaluation Method of Stereoscopic Image Quality Based on Visual Perception

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8963998B2 (en) * 2011-04-15 2015-02-24 Tektronix, Inc. Full reference system for predicting subjective quality of three-dimensional video

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102170581A (en) * 2011-05-05 2011-08-31 天津大学 Human-visual-system (HVS)-based structural similarity (SSIM) and characteristic matching three-dimensional image quality evaluation method
CN102333233A (en) * 2011-09-23 2012-01-25 宁波大学 An Objective Evaluation Method of Stereoscopic Image Quality Based on Visual Perception

Also Published As

Publication number Publication date
CN107438180A (en) 2017-12-05

Similar Documents

Publication Publication Date Title
CN105959684B (en) Stereo image quality evaluation method based on binocular fusion
CN103763552B (en) Stereoscopic image non-reference quality evaluation method based on visual perception characteristics
CN105744256B (en) Based on the significant objective evaluation method for quality of stereo images of collection of illustrative plates vision
CN109345502B (en) Stereo image quality evaluation method based on disparity map stereo structure information extraction
CN102665086B (en) Method for obtaining parallax by using region-based local stereo matching
CN110909693A (en) 3D face living body detection method and device, computer equipment and storage medium
CN107578403A (en) Stereo image quality assessment method based on gradient information to guide binocular view fusion
CN108257165A (en) Image solid matching method, binocular vision equipment
CN103873854B (en) The defining method of a kind of stereo-picture subjective assessment subject's quantity and experimental data
CN108596975A (en) A kind of Stereo Matching Algorithm for weak texture region
CN103780895B (en) A Stereoscopic Video Quality Evaluation Method
CN103856775B (en) Processing method for subjective evaluation result of stereo video quality
CN105654142A (en) Natural scene statistics-based non-reference stereo image quality evaluation method
CN107371016A (en) No-reference 3D Stereo Image Quality Evaluation Method Based on Asymmetric Distortion
CN107438180B (en) Depth perception quality evaluation method for 3D video
CN103618891B (en) Objective evaluation method of stereo camera microspur convergence shooting quality
JP6359985B2 (en) Depth estimation model generation device and depth estimation device
CN104394405B (en) A kind of method for evaluating objective quality based on full reference picture
Messai et al. Deep learning and cyclopean view for no-reference stereoscopic image quality assessment
CN105138979A (en) Method for detecting the head of moving human body based on stereo visual sense
Wang et al. A simple quality assessment index for stereoscopic images based on 3d gradient magnitude
Md et al. Multiscale-ssim index based stereoscopic image quality assessment
CN105631887B (en) Two-step Parallax Improvement Method and System Based on Adaptive Support Weight Matching Algorithm
Li et al. Full-reference quality assessment of stereoscopic images using disparity-gradient-phase similarity
CN105869172B (en) A kind of depth map reliability evaluation Measurement Method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant