CN108257089B - A method of the big visual field video panorama splicing based on iteration closest approach - Google Patents
A method of the big visual field video panorama splicing based on iteration closest approach Download PDFInfo
- Publication number
- CN108257089B CN108257089B CN201810028354.5A CN201810028354A CN108257089B CN 108257089 B CN108257089 B CN 108257089B CN 201810028354 A CN201810028354 A CN 201810028354A CN 108257089 B CN108257089 B CN 108257089B
- Authority
- CN
- China
- Prior art keywords
- point
- frame
- view
- views
- frame view
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/04—Indexing scheme for image data processing or generation, in general involving 3D image data
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
本发明涉及一种基于迭代最近点的大视场视频全景图拼接的方法,利用场景的三维信息,将个帧视图转化到同一图像平面下,实现场景拼接。具体地,对所有相邻两帧视图,执行如下操作:1.提取并匹配相邻两帧视图的特征;2.计算相邻两帧视图的相对位姿;3.在极线约束下,对相邻两帧视图进行稠密匹配;4.根据稠密匹配结果,计算相邻两帧视图重叠区域的三维模型;5.利用迭代最近点法,将步骤4得到的三维模型转化到第0帧视图的相机坐标系下;6.将转化后的三维模型投影到第0帧视图平面上,从而建立三维模型上各点与第0帧视图上位置的映射关系;7.将映射第0帧视图相同位置的点融合,完成拼接。本发明相较于传统基于单应映射拼接方法,更加真实可靠。
The invention relates to a method for video panorama stitching with a large field of view based on iterative closest points. The three-dimensional information of the scene is used to transform each frame view into the same image plane to realize scene stitching. Specifically, for all views of two adjacent frames, perform the following operations: 1. Extract and match the features of the views of the two adjacent frames; 2. Calculate the relative pose of the views of the two adjacent frames; 3. Under the epipolar constraints, for Dense matching of two adjacent frames of views; 4. According to the dense matching results, calculate the 3D model of the overlapping area of the adjacent two frames of views; 5. Use the iterative nearest point method to convert the 3D model obtained in step 4 to the 0th frame view. Under the camera coordinate system; 6. Project the converted 3D model onto the view plane of the 0th frame, so as to establish the mapping relationship between each point on the 3D model and the position on the 0th frame view; 7. Map the same position of the 0th frame view point fusion to complete the stitching. Compared with the traditional splicing method based on homography, the present invention is more real and reliable.
Description
技术领域technical field
本发明涉及一种基于迭代最近点的大视场视频全景图拼接的方法,通过估计各像素点的深度值,并将各像素映射到同一图像平面上,然后把映射到相同位置的点融合。属于计算机视觉领域。The invention relates to a method for stitching a large-view video panorama based on the iterative nearest point, by estimating the depth value of each pixel point, mapping each pixel to the same image plane, and then merging the points mapped to the same position. belongs to the field of computer vision.
背景技术Background technique
视频全景图拼接即将一段视频中,有重叠区域的各帧视图进行无缝拼接,得到一幅反映视频场景全貌的全景图像。视频全景图拼接技术针对的是普通成像设备获取的视频。数码摄相机和智能手机的普及,大大降低了全景图获取的成本,高质量的视频全景图拼接技术也爆发出了巨大的市场需求。此外,视频全景图拼接技术也广泛应用于虚拟现实、增强现实等领域,其发展前景十分广阔,有着极高的研究价值。Video panorama stitching is to seamlessly stitch each frame of view with overlapping areas in a video to obtain a panoramic image that reflects the overall picture of the video scene. Video panorama stitching technology is aimed at videos obtained by common imaging devices. The popularity of digital cameras and smartphones has greatly reduced the cost of panorama acquisition, and high-quality video panorama stitching technology has also exploded in huge market demand. In addition, video panorama stitching technology is also widely used in virtual reality, augmented reality and other fields, and its development prospects are very broad and have extremely high research value.
传统的视频全景拼接技术基于平面单应假设。具体地,首先对相邻两帧视图提取并匹配鲁棒性较强的特征,利用这些匹配得到的特征估计单应映射矩阵,然后通过估计得到的单应矩阵将其中一幅视图的像素点映射到另一幅视图中,并对灰度值进行融合(如果是彩色图像,则对RGB三个通道分别融合,下同),得到拼接结果。传统方法在两幅视图的拼接过程中,假定所有场景都位于同一平面上。而现实中,场景都位于同一平面上的假设显然是不成立的。当摄像机距离远大于场景自身深度的变化时,可以忽略场景自身深度的变化,近似认为场景是位于同一平面上,此时,传统的视频全景拼接方法的效果比较理想。当场景自身的深度变化不能忽略时,传统方法的拼接结果就会产生较大的失真,使得传统方法在实际应用中受限。Traditional video panorama stitching techniques are based on the assumption of plane homography. Specifically, firstly extract and match the features with strong robustness for the views of two adjacent frames, use the features obtained by these matching to estimate the homography mapping matrix, and then map the pixels of one of the views through the estimated homography matrix. Go to another view, and fuse the gray values (if it is a color image, fuse the three RGB channels respectively, the same below), and get the stitching result. Traditional methods assume that all scenes are on the same plane during the stitching process of two views. In reality, the assumption that the scenes are all on the same plane is obviously not valid. When the camera distance is much greater than the change of the depth of the scene itself, the change of the depth of the scene itself can be ignored, and it is approximately considered that the scene is located on the same plane. At this time, the effect of the traditional video panorama stitching method is ideal. When the depth change of the scene itself cannot be ignored, the stitching result of the traditional method will produce large distortion, which makes the traditional method limited in practical application.
发明内容SUMMARY OF THE INVENTION
本发明技术解决问题:克服现有技术的不足,提供一种基于迭代最近点的大视场视频全景拼接的方法,利用场景的深度信息,准确地将不同视图的像素映射到同一图像平面上,从根本上解决了由于平面单应假设造成的全景图失真。The technology of the present invention solves the problem: overcomes the deficiencies of the prior art, provides a method for panorama stitching of a large field of view video based on the iterative closest point, and uses the depth information of the scene to accurately map the pixels of different views onto the same image plane, The panorama distortion due to the planar homography assumption is fundamentally solved.
本发明技术解决方案:一种基于迭代最近点的大视场视频全景拼接的方法,实现步骤如下:Technical solution of the present invention: a method for panorama stitching of large field of view video based on iterative nearest point, the implementation steps are as follows:
(1)提取并匹配相邻两帧视图的特征;(1) Extract and match the features of two adjacent frame views;
(2)基于步骤(1)中得到的特征,计算相邻两帧视图的相对位姿;(2) Based on the features obtained in step (1), calculate the relative poses of the views of two adjacent frames;
(3)通过相对位姿计算得到极限约束,在极线约束下,对相邻两帧视图进行稠密匹配,得到稠密的匹配点对;(3) The limit constraint is obtained through the relative pose calculation, and under the epipolar constraint, dense matching is performed on the views of two adjacent frames to obtain dense matching point pairs;
(4)利用步骤(3)得到的稠密的匹配点对,计算相邻两帧视图重叠区域的三维模型;(4) using the dense matching point pairs obtained in step (3) to calculate the three-dimensional model of the overlapping area of two adjacent frames of views;
(5)采用迭代最近点法,将步骤(4)得到的三维模型转化到第0帧视图的相机坐标系下,即世界坐标系;(5) Using the iterative closest point method, the three-dimensional model obtained in step (4) is transformed into the camera coordinate system of the 0th frame view, that is, the world coordinate system;
(6)将在步骤(5)中转化后的三维模型,投影到第0帧视图中,建立三维模型上各点与第0帧视图上位置的映射关系;(6) project the three-dimensional model converted in step (5) into the 0th frame view, and establish the mapping relationship between each point on the three-dimensional model and the position on the 0th frame view;
(7)基于步骤(6)得到的映射关系,将映射第0帧视图相同位置的点融合,完成拼接。(7) Based on the mapping relationship obtained in step (6), the points at the same position of the view in the 0th frame are fused to complete the splicing.
在步骤(1)提取并匹配相邻两帧视图的特征的实现如下:The implementation of extracting and matching the features of two adjacent frame views in step (1) is as follows:
(1)对相邻两帧视图提取SIFT特征点,并计算每个特征点的描述子;(1) Extract SIFT feature points for two adjacent frame views, and calculate the descriptor of each feature point;
(2)基于各个特征的描述子,对相邻两帧视图提取到的特征进行匹配,从而得到相邻两帧视图间,若干匹配特征点对。(2) Based on the descriptors of each feature, the features extracted from the views of two adjacent frames are matched to obtain a number of matching feature point pairs between the views of the two adjacent frames.
在步骤(4)计算相邻两帧视图重叠区域的三维模型,实现如下:In step (4), the three-dimensional model of the overlapping area of two adjacent frames of views is calculated, and the implementation is as follows:
根据摄影测距原理,计算步骤(3)中得到的所有两帧视图间匹配点对对应的三维点,这些三维点共同组成了相邻两帧视图重叠区域的三维模型,三维模型为Mk 代表Mk中各点,h为Mk中点的数目,上标k代表Mk是由第k帧视图和第k+1帧视图的稠密匹配点对,Mk中各点的颜色即为计算三维点时,匹配点对颜色的均值。According to the principle of photographic ranging, the three-dimensional points corresponding to all matching point pairs between the two frames of views obtained in step (3) are calculated, and these three-dimensional points together form the three-dimensional model of the overlapping area of the adjacent two frames of views, and the three-dimensional model is M k represents each point in M k , h is the number of points in M k , and the superscript k represents that M k is a dense matching point pair composed of the kth frame view and the k+1th frame view, and the color of each point in M k is When calculating 3D points, match the mean of the point pair color.
在步骤(5)利用迭代最近点法,将步骤(4)得到的三维模型转化到第0帧视图的相机坐标系下的实现如下:In step (5), the iterative closest point method is used to transform the 3D model obtained in step (4) into the camera coordinate system of the 0th frame view. The implementation is as follows:
(11)第0帧视图和第1帧视图重叠区域对应的三维模型为M0通过迭代最近点方法,计算从Mk到M0的最佳刚体变换,最佳刚体变换通过转矩阵Rk和平移矢量Tk进行描述;(11) The 3D model corresponding to the overlapping area of the 0th frame view and the 1st frame view is M 0. Through the iterative nearest point method, the optimal rigid body transformation from M k to M 0 is calculated. The translation vector T k is described;
(12)将最佳刚体变换作用于Mk中各点,如式(1)所示,(12) The optimal rigid body transformation is applied to each point in M k , as shown in formula (1),
其中,为将变换到世界坐标系后的点,i用于索引Mk(或定义见下文)中各点;记 即为把Mk变换到第0帧视图的相机坐标系下的结果。in, for the The point transformed to the world coordinate system, i is used to index Mk (or defined below) at points; note It is the result of transforming M k to the camera coordinate system of the 0th frame view.
在步骤(6)建立三维模型上各点与第0帧视图上位置的映射关系,具体实现如下:In step (6), the mapping relationship between each point on the three-dimensional model and the position on the 0th frame view is established, and the specific implementation is as follows:
(21)将中各点投影至第0帧视图如式(2)所示:(21) will The projection of each point in the middle to the 0th frame view is shown in formula (2):
式(2)中,π(·)为透视投影映射,[u,v]T为在第0帧视图中投影位置的坐标,其中,u和v分别为在第0帧视图中投影的横坐标和纵坐标;In formula (2), π( ) is the perspective projection mapping, and [u, v] T is The coordinates of the projected position in the view at frame 0, where u and v are The abscissa and ordinate of the projection in the 0th frame view;
(22)结合式(2)和式(1),得到Mk中各点与第0帧视图的位置建立映射,如式(3)所示:(22) Combine formula (2) and formula (1) to obtain the mapping between each point in M k and the position of the 0th frame view, as shown in formula (3):
在步骤(7)将映射第0帧视图相同位置的点融合,具体实现如下:In step (7), the points that map the same position of the 0th frame view are fused, and the specific implementation is as follows:
对于第0帧视图的所有位置p,执行如下操作:For all positions p of the view at frame 0, do the following:
(31)找出所有与通过步骤(6)与位置p建立对应关系点,计算与位置p建立对应关系的点的颜色的平均值,记为C;(31) find out all the points that establish a corresponding relationship with position p through step (6), calculate the mean value of the color of the point that establishes a corresponding relationship with position p, and denote it as C;
(32)将位置p的颜色赋为C。(32) Assign the color of position p to C.
本发明与现有技术相比的优点在于:本发明在全景拼接时,考虑到了场景的三维信息。计算相邻两帧视图重叠区域的三维模型,并将三维模型转化到第0帧视图的相机坐标系下,而后建立三维模型上各点到第0帧视图中位置的准确映射,从而得到更加真实、自然的全景图。而传统方法假定所有场景都位于同一平面上,通过单应矩阵建立映射关系,忽略了场景的三维信息,从而得到不准确的映射关系。当景深变化较为剧烈时,传统方法的拼接过程会产生大量瑕疵,得到的质量较差的全景图。Compared with the prior art, the present invention has the advantage that: the present invention considers the three-dimensional information of the scene during panoramic stitching. Calculate the 3D model of the overlapping area of two adjacent frame views, convert the 3D model to the camera coordinate system of the 0th frame view, and then establish an accurate mapping of each point on the 3D model to the position in the 0th frame view, so as to obtain a more realistic , natural panorama. However, the traditional method assumes that all the scenes are located on the same plane, and establishes the mapping relationship through the homography matrix, ignoring the three-dimensional information of the scene, thereby obtaining an inaccurate mapping relationship. When the depth of field changes sharply, the stitching process of the traditional method will generate a large number of defects, resulting in a poor quality panorama.
附图说明Description of drawings
图1展示了基于迭代最近点的大规模视频全景图拼接方法的流程;Figure 1 shows the flow of a large-scale video panorama stitching method based on iterative closest points;
图2展示了本发明对某段视频进行全景图拼接的实验结果,(a)几帧截取自实验所用视频的视图;(b)本发明的拼接结果。FIG. 2 shows the experimental result of panorama stitching of a certain video according to the present invention, (a) several frames are intercepted from the video used in the experiment; (b) the stitching result of the present invention.
具体实施方案specific implementation
下面结合附图及实施例对本发明进行详细说明。为方便叙述,本发明使用符号k索引视频各帧视图,第k帧视图与第k+1帧视图为相邻视图。The present invention will be described in detail below with reference to the accompanying drawings and embodiments. For the convenience of description, the present invention uses the symbol k to index each frame view of the video, and the kth frame view and the k+1th frame view are adjacent views.
如图1所示,本发明具体实施如下:As shown in Figure 1, the present invention is implemented as follows:
1.提取并匹配相邻两帧视图的特征;1. Extract and match the features of two adjacent frame views;
图像特征是指数字图像中,某些具有一类特定性质的像素点。每个图像特征往往对应一个描述子(特征向量),作用是对特征进行描述。常见的图像特征有FAST、HOG、SURF、SIFT等。考虑到位姿解算对特征的鲁棒性要求较高,本发明选用SIFT特征。Image features refer to some pixels with a certain type of properties in a digital image. Each image feature often corresponds to a descriptor (feature vector), which is used to describe the feature. Common image features include FAST, HOG, SURF, SIFT, etc. Considering that the pose calculation has a high requirement on the robustness of the feature, the present invention selects the SIFT feature.
特征匹配的依据是特征的描述子,具体地,记和分别是第k和第k+1帧视图中提取得到的特征,其中n和m分别为第k和第k+1帧视图中特征的数目。记D(·)为描述子算子,则和的描述子分别为和如果第k帧视图中的特征(0≤l≤n-1)与第k+1帧视图中的特征是匹配特征,则和必满足公式(4)所示条件。The basis of feature matching is the descriptor of the feature, specifically, record and are the features extracted from the kth and k+1th frame views, respectively, where n and m are the number of features in the kth and k+1th frame views, respectively. Denote D( ) as the descriptor operator, then and The descriptors of are and If the feature in the kth frame view (0≤l≤n-1) and features in the k+1th frame view is the matching feature, then and The conditions shown in formula (4) must be satisfied.
式(4)中的||·||符号代表欧式距离算子,min(·)代表最小值算子。假设经过匹配后,可以得到s组匹配特征,统一记为(x0,x′0),(x1,x′1),...,(xs-1,x′s-1)。The ||·|| symbol in formula (4) represents the Euclidean distance operator, and min(·) represents the minimum value operator. Assuming that after matching, s groups of matching features can be obtained, which are uniformly recorded as (x 0 , x′ 0 ), (x 1 , x′ 1 ),...,(x s-1 ,x′ s-1 ).
2.计算相邻两帧视图的相对位姿2. Calculate the relative pose of two adjacent frame views
记第k帧视图相对于第k+1帧视图的基础矩阵为F,则步骤1中得到的匹配特征应满足极线约束方程x′t和xt均为齐次坐标,t=0,1,...,s-1,用于索引步骤1中得到的匹配特征。当s≥8时,可以通过奇异值分解的方法估计出F。由相机内参矩阵为K和估计得到的F计算本矩阵E,并对E奇异值分解,得到第k+1帧视图相对于第k帧视图的位姿,并通过旋转矩阵和平移矢量进行描述。Denote the fundamental matrix of the kth frame view relative to the k+1th frame view as F, then the matching feature obtained in step 1 should satisfy the epipolar constraint equation Both x′ t and x t are homogeneous coordinates, t=0,1,...,s-1, which are used to index the matching features obtained in step 1. When s≥8, F can be estimated by singular value decomposition. Calculate this matrix E from the camera internal parameter matrix K and the estimated F, and decompose the singular value of E to obtain the pose of the k+1th frame view relative to the kth frame view, which is described by a rotation matrix and a translation vector.
3.在极线约束下,对相邻两帧视图进行稠密匹配3. Under the epipolar constraint, densely match the views of two adjacent frames
稠密匹配的目的就是在满足极线约束的条件下,尽可能对第k帧视图中的像素点在第k+1帧视图中匹配得到对应的像素点。匹配依据的是像素点的特征。记和分别代表第k帧和第k+1帧视图中的像素点的齐次坐标,其中,和分别为在第k帧视图上的横坐标和纵坐标;和分别为在第k+1帧视图上的横坐标和纵坐标;i和j分别为第k和第k+1帧视图像素点的索引。像素点特征算子记为V(·)。对在位于k+1帧视图且满足极线约束的像素点中,搜索与特征最接近的像素点作为其匹配点,如式(5)所示:The purpose of dense matching is to match the pixels in the kth frame view as much as possible to obtain the corresponding pixels in the k+1th frame view under the condition of satisfying the epipolar constraints. The matching is based on the characteristics of the pixel points. remember and represent the homogeneous coordinates of the pixels in the kth frame and the k+1th frame view, respectively, where, and respectively The abscissa and ordinate on the kth frame view; and respectively The abscissa and ordinate on the view of the k+1th frame; i and j are the indices of the view pixels of the kth and k+1th frames, respectively. The pixel point feature operator is denoted as V(·). right In the pixels located in the k+1 frame view and satisfying the epipolar constraint, search for the The pixel with the closest feature As its matching point, as shown in formula (5):
式(5)中,argmin代表求取最小参数算子。式(5)的第二行即为极限约束,其几何意义为到极线的距离小于ε。其中,F代表基础矩阵,由步骤2中得到的位姿重新计算得到。In formula (5), argmin represents the operator to obtain the minimum parameter. The second row of formula (5) is the limit constraint, and its geometric meaning is The distance to the epipolar line is less than ε. Among them, F represents the fundamental matrix, which is recalculated from the pose obtained in step 2.
4.利用步骤3稠密匹配得到的像素点计算相邻两帧视图重叠区域的三维模型;4. Use the pixels obtained by dense matching in step 3 to calculate the three-dimensional model of the overlapping area of two adjacent frames of views;
通过第k帧视图中的像素点的坐标和相机参数,可以得到一条起始于第k帧视图光心,并指向的射线;同样地,根据步骤2获得的位姿,也可以得到一条起始于第k+1帧视图光心,指向匹配点的射线。通过求解这两条射线的交点即可得到对应的三维位置上述计算的原理称为摄影测距原理。的灰度(或颜色)记为(会在步骤7中用到),定义为与其匹配点灰度的平均值(如果处理是彩色视频,的RGB通道分别为与其匹配点RGB通道的平均值)。将第k帧视图中具有匹配点的像素点全部转化为对应的三维位置,可以得到第k帧和第k+1帧视图重叠场景的三维模型,记为其中h为Mk所包含点的数目)。Through the pixels in the kth frame view coordinates and camera parameters, you can get a line starting from the kth frame view optical center, and ray; similarly, according to the pose obtained in step 2, you can also get a view optical center starting from the k+1th frame, pointing to A ray of matching points. By solving the intersection of these two rays, we can get Corresponding 3D position The above calculation The principle is called the principle of photometric ranging. The grayscale (or color) of ( will be used in step 7), defined as the average of its matching point grayscales (if the processing is color video, The RGB channels are to match the average of the dot RGB channels). Converting all the pixels with matching points in the kth frame view into the corresponding 3D position, the 3D model of the overlapping scene of the kth frame and the k+1th frame view can be obtained, denoted as where h is the number of points contained in Mk ).
5.利用迭代最近点法,将步骤4得到的三维模型转化到同一坐标系下;5. Using the iterative closest point method, transform the 3D model obtained in step 4 into the same coordinate system;
为了方便描述,在后面的叙述中,记第0帧视图对应的相机坐标系为世界坐标系(即M0所在的坐标系),本步骤的目的就是将Mk转化到世界坐标系下。在本步骤和步骤6中,i代表Mk(或定义见下文)中各点的索引。使用迭代最近点法,将Mk配准至M0,得到旋转矩阵Rk和平移矢量Tk。Rk和Tk是将Mk中各点变换到世界坐标系下的刚体变换,如式(6)所示,For the convenience of description, in the following description, the camera coordinate system corresponding to the 0th frame view is referred to as the world coordinate system (that is, the coordinate system where M 0 is located). The purpose of this step is to convert M k to the world coordinate system. In this step and step 6, i represents M k (or The index of each point in the definition see below). Using the iterative closest point method, M k is registered to M 0 , resulting in a rotation matrix R k and a translation vector T k . R k and T k are rigid body transformations that transform each point in M k to the world coordinate system, as shown in equation (6),
其中,为将变换到世界坐标系后的点;记 即为把Mk变换到世界坐标系的结果。in, for the The point after transformation to the world coordinate system; note It is the result of transforming M k to the world coordinate system.
6.将在步骤5中转化后的三维模型,投影到第0帧视图中,从而建立三维模型上各点与第0帧视图上位置的映射关系。6. Project the three-dimensional model transformed in step 5 into the 0th frame view, thereby establishing a mapping relationship between each point on the 3D model and the position on the 0th frame view.
对将投影到第一幅视图中,有:right Will Projected into the first view, there are:
式(4)中,π(·)为透视投影映射,[u,v]T为在第0帧视图中投影位置的坐标,其中u和v分别为在第0帧视图中投影的横坐标和纵坐标。式(7)把映射到了第0帧视视图中,坐标为[u,v]T的位置。将式(6)带入(7),得到到第0帧视图坐标为[u,v]T位置的映射,如式(8)所示:In formula (4), π( ) is the perspective projection mapping, and [u,v] T is The coordinates of the projected position in the view at frame 0, where u and v are The abscissa and ordinate of the projection in the view at frame 0. Formula (7) put It is mapped to the position of [u, v] T in the view of frame 0. Substituting equation (6) into (7), we get The mapping to the position where the view coordinates of the 0th frame are [u, v] T , as shown in formula (8):
按照式(8),即建立Mk中各点到第0帧视图中位置的映射。According to formula (8), the mapping of each point in M k to the position in the 0th frame view is established.
7.融合所有映射到第0帧视图的像素点。7. Fuse all pixels mapped to the 0th frame view.
记视频的帧数为nframes,对k=1,2,...,nframes-1,执行步骤1~步骤6。记第0帧视图中某像素点坐标为[u,v]T,其中,u为横坐标,v为纵坐标。将所有映射到第0帧视图[u,v]T位置的点组成的集合称为[u,v]T的支持集,记为U(u,v)。记U(u,v)={X0,X1,...,Xr-1}(r为U(u,v)中,点的数目),如果U(u,v)非空(即r>0),则[u,v]T处的灰度(或颜色)为:Denote the number of video frames as n frames , and for k=1, 2, . . . , n frames -1, perform steps 1 to 6. Denote the coordinates of a pixel in the 0th frame view as [u,v] T , where u is the abscissa and v is the ordinate. The set of all points mapped to the position of [u, v] T in the 0th frame is called the support set of [u, v] T , denoted as U(u, v). Write U(u,v)={X 0 ,X 1 ,...,X r-1 } (r is the number of points in U(u,v)), if U(u,v) is not empty ( That is, r>0), then the grayscale (or color) at [u, v] T is:
其中,C(Xi)代表Xi的颜色(定义见步骤4),为第0帧视图[u,v]T处的颜色,按照式(9)的定义,即为C(Xi)的平均值。对第0帧视图中,所有支持集非空的位置,重新计算灰度,得到相应位置的灰度,实现了拼接,得到全景图。Among them, C(X i ) represents the color of Xi ( see step 4 for definition), is the color at the view [u, v] T of the 0th frame, according to the definition of formula (9), is the average value of C(X i ). For all positions where the support set is not empty in the 0th frame view, the grayscale is recalculated to obtain the grayscale of the corresponding position, stitching is realized, and the panorama is obtained.
图2所示为本发明的实验结,(a)是从实验用视频中,截取的四帧视图;(b)是对实验用视频采用本发明进行全景拼接的结果。可以看出,本发明方法拼接得到的全景图反映了场景的全貌,且过渡自然,真实度高。Figure 2 shows the experimental results of the present invention, (a) is a four-frame view intercepted from the experimental video; (b) is the result of panoramic stitching of the experimental video using the present invention. It can be seen that the panorama image obtained by splicing by the method of the present invention reflects the whole picture of the scene, and the transition is natural and the degree of authenticity is high.
提供以上实施例仅仅是为了描述本发明的目的,而并非要限制本发明的范围。本发明的范围由所附权利要求限定。不脱离本发明的精神和原理而做出的各种等同替换和修改,均应涵盖在本发明的范围之内。The above embodiments are provided for the purpose of describing the present invention only, and are not intended to limit the scope of the present invention. The scope of the invention is defined by the appended claims. Various equivalent replacements and modifications made without departing from the spirit and principle of the present invention should be included within the scope of the present invention.
Claims (5)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201810028354.5A CN108257089B (en) | 2018-01-12 | 2018-01-12 | A method of the big visual field video panorama splicing based on iteration closest approach |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201810028354.5A CN108257089B (en) | 2018-01-12 | 2018-01-12 | A method of the big visual field video panorama splicing based on iteration closest approach |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN108257089A CN108257089A (en) | 2018-07-06 |
| CN108257089B true CN108257089B (en) | 2019-01-08 |
Family
ID=62726564
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201810028354.5A Active CN108257089B (en) | 2018-01-12 | 2018-01-12 | A method of the big visual field video panorama splicing based on iteration closest approach |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN108257089B (en) |
Families Citing this family (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110874818B (en) * | 2018-08-31 | 2023-06-23 | 阿里巴巴集团控股有限公司 | Image processing and virtual space construction method, device, system and storage medium |
| US11055835B2 (en) | 2019-11-19 | 2021-07-06 | Ke.com (Beijing) Technology, Co., Ltd. | Method and device for generating virtual reality data |
| CN111105347B (en) * | 2019-11-19 | 2020-11-13 | 贝壳找房(北京)科技有限公司 | Method, device and storage medium for generating panoramic image with depth information |
| CN113052761B (en) * | 2019-12-26 | 2024-01-30 | 炬星科技(深圳)有限公司 | Laser point cloud map fusion method, device and computer readable storage medium |
| CN111242990B (en) * | 2020-01-06 | 2024-01-30 | 西南电子技术研究所(中国电子科技集团公司第十研究所) | 360-degree three-dimensional reconstruction optimization method based on continuous phase dense matching |
| CN114143528B (en) * | 2020-09-04 | 2024-06-28 | 北京大视景科技有限公司 | Multi-video stream fusion method, electronic device, and storage medium |
| CN113160053B (en) * | 2021-04-01 | 2022-06-14 | 华南理工大学 | An underwater video image restoration and stitching method based on pose information |
Family Cites Families (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9307165B2 (en) * | 2008-08-08 | 2016-04-05 | Qualcomm Technologies, Inc. | In-camera panorama image stitching assistance |
| US20150346115A1 (en) * | 2014-05-30 | 2015-12-03 | Eric J. Seibel | 3d optical metrology of internal surfaces |
| CN104484648B (en) * | 2014-11-27 | 2017-07-25 | 浙江工业大学 | Robot variable viewing angle obstacle detection method based on contour recognition |
| CN105279789B (en) * | 2015-11-18 | 2016-11-30 | 中国兵器工业计算机应用技术研究所 | A kind of three-dimensional rebuilding method based on image sequence |
| CN105374011B (en) * | 2015-12-09 | 2019-03-15 | 中电科信息产业有限公司 | Point-clouds Registration method and device based on panoramic picture |
| CN107169924B (en) * | 2017-06-14 | 2020-10-09 | 歌尔科技有限公司 | Method and system for establishing three-dimensional panoramic image |
-
2018
- 2018-01-12 CN CN201810028354.5A patent/CN108257089B/en active Active
Also Published As
| Publication number | Publication date |
|---|---|
| CN108257089A (en) | 2018-07-06 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN108257089B (en) | A method of the big visual field video panorama splicing based on iteration closest approach | |
| Zhang et al. | Multi-viewpoint panorama construction with wide-baseline images | |
| CN109685913B (en) | Augmented Reality Realization Method Based on Computer Vision Positioning | |
| CN110211043A (en) | A kind of method for registering based on grid optimization for Panorama Mosaic | |
| CN107154014B (en) | Real-time color and depth panoramic image splicing method | |
| CN111553939B (en) | An Image Registration Algorithm for Multi-camera Cameras | |
| CN103761721B (en) | One is applicable to space rope system machine human stereo vision fast image splicing method | |
| CN109685078B (en) | Infrared image identification method based on automatic annotation | |
| CN108537848A (en) | A kind of two-stage pose optimal estimating method rebuild towards indoor scene | |
| CN106920210B (en) | A real-time panoramic correction method for fisheye video based on spherical expansion model | |
| CN113538569B (en) | Weak texture object pose estimation method and system | |
| CN111553845B (en) | A Fast Image Stitching Method Based on Optimized 3D Reconstruction | |
| CN106534833B (en) | A kind of double vision point three-dimensional video-frequency antihunt means of joint space time shaft | |
| CN104952075A (en) | Laser scanning three-dimensional model-oriented multi-image automatic texture mapping method | |
| Xu et al. | Layout-guided novel view synthesis from a single indoor panorama | |
| CN108830925B (en) | Three-dimensional digital modeling method based on spherical screen video stream | |
| TW201635242A (en) | Generation method, device and system for indoor two-dimension plan | |
| CN107993197A (en) | The joining method and system of a kind of panorama camera | |
| Chen et al. | Autosweep: Recovering 3d editable objects from a single photograph | |
| CN115330594A (en) | A rapid target recognition and calibration method based on UAV oblique photography 3D model | |
| CN107330856B (en) | A Panoramic Imaging Method Based on Projective Transformation and Thin Plate Splines | |
| CN110136090A (en) | Robust Elastic Model UAV Image Stitching Method with Local Preserving Registration | |
| CN104036496B (en) | Self-calibration method for radial distortion of fish-eye lens camera | |
| CN119169288A (en) | A scene object classification method and system based on multi-view depth images | |
| CN103700082A (en) | Image Mosaic Method Based on Relative Orientation of Dual Quaternions |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |