[go: up one dir, main page]

CN118134972A - Target tracking method based on fusion of intensity image and point cloud data - Google Patents

Target tracking method based on fusion of intensity image and point cloud data Download PDF

Info

Publication number
CN118134972A
CN118134972A CN202410347052.XA CN202410347052A CN118134972A CN 118134972 A CN118134972 A CN 118134972A CN 202410347052 A CN202410347052 A CN 202410347052A CN 118134972 A CN118134972 A CN 118134972A
Authority
CN
China
Prior art keywords
target
tracking
point cloud
image
intensity image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202410347052.XA
Other languages
Chinese (zh)
Inventor
刘雪莲
贾麟晶
王春阳
周绪浪
庄瑞
卫烘州
龙超杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Technological University
Original Assignee
Xian Technological University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Technological University filed Critical Xian Technological University
Priority to CN202410347052.XA priority Critical patent/CN118134972A/en
Publication of CN118134972A publication Critical patent/CN118134972A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/277Analysis of motion involving stochastic approaches, e.g. using Kalman filters
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/766Arrangements for image or video recognition or understanding using pattern recognition or machine learning using regression, e.g. by projecting features on hyperplanes
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

A target tracking method based on fusion of an intensity image and point cloud data relates to the fields of computer graphics and three-dimensional point cloud target tracking, and solves the problems of reduced tracking precision and even tracking failure caused by shielding of an existing target by an obstacle. According to the method, a multi-feature fusion mode is adopted, and nuclear correlation filtering tracking is carried out on a target intensity image formed by the laser radar by fusing the HOG features and Fourier description sub-features of the target of interest; converting a target range profile formed by the laser radar into a three-dimensional point cloud, and predicting a track of a point cloud target by using a Kalman tracker; and effectively judging the target shielding state by combining PSR and ISS, providing an adaptive factor according to different target shielding conditions, and correcting the positions of the kernel correlation and the Kalman tracker by using the adaptive factor to obtain a final tracking position. The method can carry out steady tracking on the target, and the average time consumption of each frame of data is 39ms.

Description

基于强度像与点云数据融合的目标跟踪方法Target tracking method based on fusion of intensity image and point cloud data

技术领域Technical Field

本发明涉及计算机图形学和三维点云目标跟踪领域,具体涉及一种基于强度像与点云数据融合的目标跟踪方法。The present invention relates to the fields of computer graphics and three-dimensional point cloud target tracking, and in particular to a target tracking method based on the fusion of intensity image and point cloud data.

背景技术Background technique

随着科技的高速发展,无人战车已在战场崭露头角。跟踪系统作为无人战车的重要组成部分,直接影响着对敌方目标进行精准跟踪和精确打击能力,而跟踪算法是决定跟踪系统性能优良的关键。三维成像激光雷达是一种以激光波束为载体进行主动探测的传感器,可以获取目标的三维点云图像,拥有全天时、高精度的成像探测能力。使用盖革模式雪崩光电二极管(Geiger-ModeAvalanche PhotoDiode,GM-APD)探测器的激光雷达能够对远距离目标进行高灵敏度单光子成像,在现代军事战争中有着广泛的应用前景。由于战场环境复杂,敌方战车、坦克等目标易受建筑物、岩石等物体遮挡与干扰,造成激光雷达获取的目标信息缺失,加大目标跟踪任务难度。基于核相关滤波跟踪算法跟踪精度高、速度快,但对遮挡情况由于获取的目标特征信息较少会导致跟踪精度大幅度下降,目标被完全遮挡时,卡尔曼滤波跟踪算法可对目标位置进行预测。With the rapid development of science and technology, unmanned combat vehicles have emerged on the battlefield. As an important component of unmanned combat vehicles, the tracking system directly affects the ability to accurately track and accurately strike enemy targets, and the tracking algorithm is the key to determining the excellent performance of the tracking system. Three-dimensional imaging laser radar is a sensor that uses laser beams as carriers for active detection. It can obtain three-dimensional point cloud images of targets and has all-day, high-precision imaging detection capabilities. Laser radars using Geiger-Mode Avalanche PhotoDiode (GM-APD) detectors can perform high-sensitivity single-photon imaging of long-distance targets and have broad application prospects in modern military warfare. Due to the complex battlefield environment, enemy targets such as tanks and tanks are easily blocked and interfered by objects such as buildings and rocks, resulting in the loss of target information obtained by the laser radar, which increases the difficulty of target tracking tasks. The tracking algorithm based on the kernel correlation filter has high tracking accuracy and fast speed, but the tracking accuracy will be greatly reduced in the case of occlusion due to the less target feature information obtained. When the target is completely blocked, the Kalman filter tracking algorithm can predict the target position.

发明内容Summary of the invention

本发明要提供一种基于强度像与点云数据融合的目标跟踪方法,以解决目标受到障碍物遮挡导致跟踪精度下降甚至跟踪失败的问题。The present invention aims to provide a target tracking method based on the fusion of intensity image and point cloud data to solve the problem that the target is blocked by obstacles, resulting in reduced tracking accuracy or even tracking failure.

基于强度像与点云数据融合的目标跟踪方法,该方法由以下步骤实现:The target tracking method based on the fusion of intensity image and point cloud data is implemented by the following steps:

步骤一、对激光雷达所成目标强度像的HOG特征、傅里叶描述子特征进行融合,并进行核相关滤波跟踪;Step 1: Fuse the HOG features and Fourier descriptor features of the target intensity image formed by the laser radar, and perform kernel correlation filter tracking;

步骤二、将激光雷达所成目标距离像转换为三维点云数据,利用卡尔曼跟踪器对点云目标进行轨迹预测;Step 2: Convert the target range image formed by the laser radar into three-dimensional point cloud data, and use the Kalman tracker to predict the trajectory of the point cloud target;

步骤三、结合峰值旁瓣比PSR与内部形状描述子ISS对目标遮挡状态进行有效判断,根据目标遮挡情况的不同,提出自适应因子,利用自适应因子对所述核相关滤波跟踪与卡尔曼跟踪器预测位置进行修正得到最终跟踪位置。Step 3: Combine the peak sidelobe ratio PSR and the internal shape descriptor ISS to effectively judge the target occlusion state. According to the different target occlusion conditions, an adaptive factor is proposed. The kernel correlation filter tracking and the Kalman tracker predicted position are corrected using the adaptive factor to obtain the final tracking position.

本发明的有益效果:Beneficial effects of the present invention:

1、本发明将目标强度像的方向梯度直方图(Histogram of Oriented Gradient,HOG)特征与傅里叶描述子进行融合,提供更全面的目标描述,提高核相关滤波跟踪算法的鲁棒性。1. The present invention fuses the Histogram of Oriented Gradient (HOG) features of the target intensity image with the Fourier descriptor to provide a more comprehensive target description and improve the robustness of the kernel correlation filter tracking algorithm.

2、本发明利用强度像与距离像转换后的点云数据结合峰值旁瓣比与内部形状描述对目标遮挡状态进行有效判断,并提出一种自适应因子,根据目标遮挡情况的不同对核相关滤波与卡尔曼滤波的跟踪结果进行加权,得出最终目标位置信息。2. The present invention utilizes the point cloud data converted from intensity image and range image combined with the peak-to-sidelobe ratio and internal shape description to effectively judge the target occlusion state, and proposes an adaptive factor to weight the tracking results of kernel correlation filtering and Kalman filtering according to different target occlusion conditions to obtain the final target position information.

3、在KITTI数据集对本发明所提算法进行实验验证,本发明所提算法能对处于不同遮挡状态的目标进行准确跟踪,相较卡尔曼滤波算法跟踪精度提高21.67%,相较于扩展卡尔曼滤波精度提高7.94%,平均每帧处理时间51ms;在GM-APD激光雷达实采场景中,本发明所述方法能对目标进行稳健跟踪,平均每帧数据耗时39ms。3. The algorithm proposed in the present invention is experimentally verified on the KITTI dataset. The algorithm proposed in the present invention can accurately track targets in different occlusion states. The tracking accuracy is improved by 21.67% compared with the Kalman filter algorithm, and the accuracy is improved by 7.94% compared with the extended Kalman filter. The average processing time per frame is 51ms. In the actual sampling scene of the GM-APD lidar, the method of the present invention can robustly track the target, and the average processing time per frame of data is 39ms.

附图说明BRIEF DESCRIPTION OF THE DRAWINGS

图1为本发明所述的基于强度像与点云数据融合的目标跟踪方法的流程图;FIG1 is a flow chart of a target tracking method based on fusion of intensity image and point cloud data according to the present invention;

图2为卡尔曼滤波原理图;Figure 2 is a schematic diagram of the Kalman filter principle;

图3为目标峰值响应图,其中,(a)为目标未被遮挡时峰值响应图,(b)为目标被遮挡时峰值响应图;FIG3 is a target peak response diagram, wherein (a) is a peak response diagram when the target is not blocked, and (b) is a peak response diagram when the target is blocked;

图4为ISS关键点提取结果图;Figure 4 is a diagram of the ISS key point extraction results;

图5为遮挡阈值结果图;其中,(a)为峰值旁瓣比PSR=7.69,(b)为ISS关键点N=9;(c)为峰值旁瓣比PSR=5.62,(d)为ISS关键点N=6;(e)为峰值旁瓣比PSR=1.77,(f)ISS关键点N=0;Figure 5 is a diagram of the occlusion threshold result; (a) is the peak sidelobe ratio PSR = 7.69, (b) is the ISS key point N = 9; (c) is the peak sidelobe ratio PSR = 5.62, (d) is the ISS key point N = 6; (e) is the peak sidelobe ratio PSR = 1.77, (f) ISS key point N = 0;

图6为算法跟踪结果图;其中,(a)为强度像跟踪结果1,(b)为最终跟踪结果1,(c)为强度像跟踪结果2,(d)为最终跟踪结果2,(e)为强度像跟踪结果3,(f)为最终跟踪结果3,(g)强度像跟踪结果4,(h)最终跟踪结果4;FIG6 is a diagram of the algorithm tracking results; wherein (a) is the intensity image tracking result 1, (b) is the final tracking result 1, (c) is the intensity image tracking result 2, (d) is the final tracking result 2, (e) is the intensity image tracking result 3, (f) is the final tracking result 3, (g) is the intensity image tracking result 4, and (h) is the final tracking result 4;

图7为误差曲线与跟踪轨迹图;其中,(a)为中心位置误差曲线图,(b)为卡尔曼滤波跟踪轨迹图,(c)为扩展卡尔曼跟踪轨迹图,(d)为融合跟踪轨迹图;FIG7 is an error curve and a tracking trajectory diagram; wherein (a) is a center position error curve diagram, (b) is a Kalman filter tracking trajectory diagram, (c) is an extended Kalman tracking trajectory diagram, and (d) is a fusion tracking trajectory diagram;

图8为室外探测场景图;Figure 8 is a diagram of an outdoor detection scene;

图9为GM-APD激光雷达跟踪结果图;其中,(a)为强度像跟踪结果1,(b)为最终跟踪结果1,(c)为强度像跟踪结果2,(d)为最终跟踪结果2,(e)为强度像跟踪结果3,(f)为最终跟踪结果3,(g)为强度像跟踪结果4,(h)为最终跟踪结果4,(i)为强度像跟踪结果5,(j)为最终跟踪结果5。Figure 9 is a diagram of the GM-APD laser radar tracking results; wherein, (a) is the intensity image tracking result 1, (b) is the final tracking result 1, (c) is the intensity image tracking result 2, (d) is the final tracking result 2, (e) is the intensity image tracking result 3, (f) is the final tracking result 3, (g) is the intensity image tracking result 4, (h) is the final tracking result 4, (i) is the intensity image tracking result 5, and (j) is the final tracking result 5.

具体实施方式Detailed ways

为了使本发明的目的和优点更加清楚明白,下面结合附图和实施例对本发明作进一步描述;应当理解,此处所描述的具体实施例仅用于解释本发明,并不用于限定本发明。In order to make the objects and advantages of the present invention more clearly understood, the present invention is further described below in conjunction with the accompanying drawings and embodiments; it should be understood that the specific embodiments described herein are only used to explain the present invention and are not used to limit the present invention.

基于强度像与点云数据融合的目标跟踪方法,如图1所示,具体包括以下步骤:The target tracking method based on the fusion of intensity image and point cloud data, as shown in FIG1, specifically includes the following steps:

步骤S1、采用多特征融合的方式,对激光雷达所成目标强度像融合感兴趣目标HOG特征、傅里叶描述子特征,进行核相关滤波跟踪;Step S1, using a multi-feature fusion method, the intensity image of the target formed by the laser radar is fused with the HOG feature and the Fourier descriptor feature of the target of interest, and kernel correlation filtering tracking is performed;

步骤S11、利用岭回归函数训练核相关滤波跟踪算法的分类器模型;Step S11, using the ridge regression function to train the classifier model of the kernel correlation filter tracking algorithm;

假设有训练集{(a1,b1),…(ai,bi)},对于岭回归函数f(xi)=wTai,训练的目标是让回归函数在样本向量ai和其对应标签bi之间的均方误差最小,优化公式如下式:Assume that there is a training set {(a 1 ,b 1 ),…(a i ,b i )}, for the ridge regression function f( xi )= wTai , the training goal is to minimize the mean square error between the sample vector ai and its corresponding label bi . The optimization formula is as follows:

其中,λ为防止过拟合的正则化参数,w为列向量表示权重系数,||w||2代表w的2范数,控制过度拟合,公式(1)有唯一最优解,可以由下式计算得出:Among them, λ is the regularization parameter to prevent overfitting, w is a column vector representing the weight coefficient, ||w|| 2 represents the 2-norm of w, which controls overfitting. Formula (1) has a unique optimal solution, which can be calculated by the following formula:

w=(ATA+λI)-1ATb (2)w=( ATA +λI) -1ATb (2 )

将式(2)用其复数形式替代:Replace formula (2) with its plural form:

w=(AHA+λI)-1AHb (3)w=(A H A+λI) -1 A H b (3)

其中,A表示样本向量ai循环移位得到的循环矩阵,b中每一个元素代表一个样本标签bi,I表示单位矩阵,AH为A的埃尔米特转置,AH=(A*)T,A*为A的复共轭。Wherein, A represents the circulant matrix obtained by cyclic shift of sample vector a i , each element in b represents a sample label b i , I represents the identity matrix, A H is the Hermitian transpose of A, A H =(A * ) T , and A * is the complex conjugate of A.

步骤S12、通过循环矩阵扩充训练样本;获得训练后的核相关滤波跟踪模型;Step S12, expanding the training samples through the circulant matrix; obtaining the trained kernel correlation filter tracking model;

假设向量a=[a1,a2,…,an]T代表含有感兴趣目标的图像块特征,并以此作为基础样本。将基础样本作为正样本,通过与置换矩阵P左乘,对向量a进行移位得到负样本。置换矩阵P如(4)所示:Assume that the vector a = [a 1 , a 2 , …, a n ] T represents the image block features containing the object of interest, and use it as the basic sample. The basic sample is used as the positive sample, and the vector a is shifted to obtain the negative sample by left multiplication with the permutation matrix P. The permutation matrix P is shown in (4):

Pa=[aN-1,a0,a1,…,aN-2]是向量a向右平移一个单位得到的,这里用PNa表示向量a向右位移N个单位得到的向量,如果N取负值,则将向量a反方向位移。Pa=[a N-1 ,a 0 ,a 1 ,…,a N-2 ] is the vector obtained by translating vector a one unit to the right. Here, P N a is used to represent the vector obtained by shifting vector a N units to the right. If N takes a negative value, vector a is shifted in the opposite direction.

对向量a进行循环移位操作后可得循环矩阵A:After performing a circular shift operation on vector a, the circulant matrix A can be obtained:

循环矩阵可以通过对角矩阵(diag)进行离散傅里叶变换(DFT)得到,并且与基础向量a无关,通过下式可表达此性质:The circulant matrix can be obtained by performing a discrete Fourier transform (DFT) on the diagonal matrix (diag) and is independent of the basis vector a. This property can be expressed by the following formula:

其中,F表示与基础向量a无关的常量矩阵,被称作傅里叶矩阵,对于任意待检测向量z,可以通过得到其傅里叶变换,*是共轭操作符,/>是a的离散傅里叶变换,N代表N×N维样本矩阵。Among them, F represents a constant matrix independent of the basis vector a, which is called the Fourier matrix. For any vector z to be detected, it can be obtained by Get its Fourier transform, * is the conjugate operator, /> is the discrete Fourier transform of a, and N represents the N×N dimensional sample matrix.

如果训练数据都是由基础样本循环移位获得的,可以使用傅里叶矩阵对角化方法进行元素级运算得到最终结果。将对角化融入协方差矩阵可得:If the training data are all obtained by cyclic shift of the basic samples, the Fourier matrix diagonalization method can be used to perform element-level operations to obtain the final result. Incorporating the diagonalization into the covariance matrix yields:

由于FFH=I,将上式(8)简化为:Since FF H =I, the above formula (8) is simplified to:

因为对角矩阵仅在对角元素上进行运算,可得:Since the diagonal matrix is operated only on the diagonal elements, we can get:

其中,⊙表示按元素相乘(矩阵对应位置相乘),将式(7)、(10)代入式(3)中计算:Where ⊙ represents element-by-element multiplication (matrix corresponding positions multiplication). Substitute equations (7) and (10) into equation (3) to calculate:

对式(11)结果的两边同时进行傅里叶变换,得到结果如下:Performing Fourier transform on both sides of the result of equation (11) simultaneously, the results are as follows:

其中,是一个向量,λ为防止过拟合的正则化参数,根据对角矩阵性质将上式简化:in, is a vector, λ is a regularization parameter to prevent overfitting, and the above formula is simplified according to the properties of the diagonal matrix:

步骤S13、基于核相关滤波跟踪模型对目标进行快速检测及定位;Step S13, quickly detecting and locating the target based on the kernel correlation filter tracking model;

在核相关滤波跟踪算法中,为了引入核技巧,需要将待优化参数w由线性空间(原空间)转变到对偶空间α;In the kernel correlation filter tracking algorithm, in order to introduce the kernel technique, it is necessary to transform the parameter w to be optimized from the linear space (original space) to the dual space α;

设定样本X=[x1,x2,…,xn]T用于训练,Y=[y1,y2,…,yn]T为样本标签值。从样本空间到Hilbert特征空间的非线性变换为对应的核函数则优化问题的最优解为:Set the sample X = [x 1 ,x 2 ,…,x n ] T for training, and Y = [y 1 ,y 2 ,…,y n ] T as the sample label value. The nonlinear transformation from the sample space to the Hilbert feature space is The corresponding kernel function The optimal solution to the optimization problem is:

式中,分别为样本xi、x′到Hilbert特征空间的映射函数,x为训练样本,x′为待检测样本;其解w位于样本xi张成的子空间内,w可表示为:In the formula, are respectively the mapping functions from samples xi and x′ to the Hilbert feature space, x is the training sample, and x′ is the sample to be tested; its solution w is located in the subspace spanned by sample xi , and w can be expressed as:

此时所求的变量从w转为α,α称作原空间在对偶空间的解。At this time, the variable sought changes from w to α, which is called the solution of the original space in the dual space.

因此函数响应f(z)为:Therefore the function response f(z) is:

其中κ(xi,zi)表示高维空间中训练样本与/>的内积。where κ( xi , zi ) represents the training samples in high-dimensional space With/> The inner product of .

岭回归核化形式的闭式解如下式:The closed-form solution of the ridge regression kernelization form is as follows:

α=(K+λI)-1y (17)α=(K+λI) -1y (17)

其中,y表示一个样本标签值的通用含义,多项式核、线性核和高斯核对应的核矩阵都是循环矩阵,将核矩阵K表示为:Among them, y represents the general meaning of a sample label value. The kernel matrices corresponding to the polynomial kernel, linear kernel, and Gaussian kernel are all circulant matrices. The kernel matrix K is expressed as:

K=C(κxx) (18)K=C(κ xx ) (18)

式中,κxx表示核矩阵K的第一行。对公式(17)进行计算得:In the formula, κ xx represents the first row of the kernel matrix K. Calculating formula (17) yields:

为了实现对目标的检测,需要在候选区域对分类器进行训练,得到所需的回归函数。通过下式对候选区域进行模型构建:In order to detect the target, it is necessary to train the classifier in the candidate area to obtain the required regression function. The model of the candidate area is constructed by the following formula:

Kz=C(κxz) (20)K z =C(κ xz ) (20)

其中Kz是训练样本与候选区域所组成的核矩阵,κxz是一个互相关矩阵。由此可以得到候选区域的相关响应(经函数映射后的响应):Where Kz is the kernel matrix composed of training samples and candidate regions, and κxz is a cross-correlation matrix. From this, we can get the relevant response of the candidate region (the response after function mapping):

f(z)=Kzα (21)f(z)=K z α (21)

利用傅里叶矩阵对角化特性快速检测目标,待检测向量z与训练所得分类器响应值在频域表示为:The diagonalization characteristics of the Fourier matrix are used to quickly detect the target. The vector z to be detected and the response value of the trained classifier are expressed in the frequency domain as follows:

其中,最大值所对应的位置就是预测的目标位置。in, The position corresponding to the maximum value is the predicted target position.

步骤S14、多特征融合;Step S14: multi-feature fusion;

首先计算目标强度像的HOG特征,其具体的步骤如下:First, calculate the HOG features of the target intensity image. The specific steps are as follows:

a.截取跟踪区域图像,并将图像灰度化,减少计算量。a. Capture the image of the tracking area and convert it to grayscale to reduce the amount of calculation.

b.采用Gamma校正法对输入图像进行归一化处理,使得提取到的特征对光照和背景噪声具有更高的鲁棒性。归一化操作公式如(23)所示:b. Use the Gamma correction method to normalize the input image so that the extracted features are more robust to illumination and background noise. The normalization operation formula is shown in (23):

I(hx,hy)=I(hx,hy)Gamma (23)I(h x , hy )=I(h x , hy ) Gamma (23)

式中,I(hx,hy)为图像矩阵,通常情况下Gamma=5.0。Where I(h x , hy ) is the image matrix, and Gamma is usually 5.0.

c.计算像素点的梯度信息,获得目标的轮廓信息。首先使用[-1,0,1]梯度算子对原图像做卷积运算,得到水平方向的梯度分量,然后使用[-1,0,1]T梯度算子对原始图像做卷积运算,得到垂直方向上的梯度分量。其计算公式如(24)和(25)所示:c. Calculate the gradient information of the pixel points to obtain the contour information of the target. First, use the [-1, 0, 1] gradient operator to perform a convolution operation on the original image to obtain the gradient component in the horizontal direction, and then use the [-1, 0, 1] T gradient operator to perform a convolution operation on the original image to obtain the gradient component in the vertical direction. The calculation formulas are shown in (24) and (25):

Gx(hx,hy)=I(hx+1,hy)-I(hx-1,hy) (24) Gx ( hx , hy )=I(hx +1 , hy )-I(hx -1 , hy ) (24)

Gy(hx,hy)=I(hx,hy+1)-I(hx,hy-1) (25)G y ( h x , h y ) = I ( h x , h y + 1 ) - I ( h x , h y - 1 ) (25)

式中,Gx(hx,hy)表示图像I(hx,hy)水平方向上的梯度值,Gy(hx,hy)表示图像I(hx,hy)垂直方向的梯度值。因此图像(hx,hy)坐标处的梯度值和梯度方向可表示为:In the formula, Gx ( hx , hy ) represents the horizontal gradient value of the image I( hx , hy ) and Gy ( hx , hy ) represents the vertical gradient value of the image I( hx , hy ). Therefore, the gradient value and gradient direction at the image ( hx , hy ) coordinate can be expressed as:

d.将图像划分成很多小cell单元格,统计出每个cell的梯度信息。将每个cell的梯度方向分成9个区间(bin),即每个bin的间隔为20°,然后将梯度信息根据梯度方向映射到对应的bin上。d. Divide the image into many small cells and calculate the gradient information of each cell. Divide the gradient direction of each cell into 9 bins, i.e., the interval of each bin is 20°, and then map the gradient information to the corresponding bin according to the gradient direction.

e.统计由cell组成的块(block)中的梯度信息,最后将所有的block内的梯度直方图串联就得到了整个图像的梯度直方图特征。e. Count the gradient information in the block composed of cells, and finally concatenate the gradient histograms in all blocks to obtain the gradient histogram features of the entire image.

本实施方式中,计算傅里叶描述子特征,其计算步骤如下:In this implementation, the Fourier descriptor features are calculated, and the calculation steps are as follows:

a.首先对强度像灰度化处理并进行Gamma校正。a. First, grayscale the intensity image and perform gamma correction.

b.利用Canny边缘检测算法提取待跟踪目标的轮廓点集合{(xc,yc)|c=1,2,…,m}。b. Use the Canny edge detection algorithm to extract the contour point set {(x c ,y c )|c=1,2,…,m} of the target to be tracked.

c.计算轮廓点的中心坐标:c. Calculate the center coordinates of the contour points:

d.将轮廓点从直角坐标系转换为极坐标系。d. Convert the contour points from rectangular coordinates to polar coordinates.

为极坐标系的极点,将轮廓点的坐标{(xc,yc)|c=1,2,…,m}转换为对应的极坐标{(rcc)|c=1,2,…,m}by The coordinates of the contour point {(x c ,y c )|c=1,2,…,m} are converted to the corresponding polar coordinates {(r cc )|c=1,2,…,m}

式中rc为中心轮廓点距离。Where r c is the distance between the center contour points.

对rc进行升序排序,获得中心点轮廓点距离序列,记为D=[r1,r2,…,rm]。Sort r c in ascending order to obtain the center point contour point distance sequence, denoted as D = [r 1 , r 2 , …, r m ].

e.对中心轮廓点距离序列进行快速傅里叶变换。e. Perform fast Fourier transform on the center contour point distance sequence.

式中,j是傅里叶变换虚数单位,最终构建出的傅里叶描述子为:In the formula, j is the Fourier transform imaginary unit, and the final constructed Fourier descriptor is:

式中|·|表示傅里叶频谱。Here, |·| represents the Fourier spectrum.

最后将HOG特征与傅里叶描述子进行串行融合,可提高核相关滤波跟踪算法的鲁棒性。Finally, the HOG features are serially fused with the Fourier descriptor to improve the robustness of the kernel correlation filter tracking algorithm.

步骤S2、将激光雷达所成目标距离像转换为三维点云,利用卡尔曼跟踪器对点云目标进行轨迹预测;Step S2, converting the target range image formed by the laser radar into a three-dimensional point cloud, and using the Kalman tracker to predict the trajectory of the point cloud target;

卡尔曼滤波器分为预测与更新两个阶段,如图2所示,在预测阶段,使用t-1时刻目标状态的最优估计值计算t时刻目标状态的预估计值,由于在预测阶段忽略了系统噪声的影响,因此所得到的预计估计值与真实状态值存在偏差;在更新阶段基于t时刻目标状态的观测值及状态观测矩阵对预估计值进行修正,从而得到目标状态的最优估计。具体过程为:The Kalman filter is divided into two stages: prediction and update. As shown in Figure 2, in the prediction stage, the optimal estimate of the target state at time t-1 is used to calculate the estimated value of the target state at time t. Since the influence of system noise is ignored in the prediction stage, there is a deviation between the estimated value and the actual state value; in the update stage, the estimated value is corrected based on the observed value of the target state at time t and the state observation matrix, so as to obtain the optimal estimate of the target state. The specific process is:

步骤S21、状态预测;Step S21, state prediction;

任意t时刻的目标状态预测过程如下:The target state prediction process at any time t is as follows:

其中为前一时刻目标状态的最优估计,At为状态转移矩阵,/>为当前时刻目标状态的预估计,/>为前一时刻目标状态最优估计值对应的协方差矩阵,/>为当前时刻目标状态的预估计值对应的协方差矩阵,Qt是ωt对应的噪声协方差矩阵。in is the optimal estimate of the target state at the previous moment, At is the state transfer matrix, /> is the estimated target state at the current moment, /> is the covariance matrix corresponding to the optimal estimate of the target state at the previous moment,/> is the covariance matrix corresponding to the estimated value of the target state at the current moment, and Qt is the noise covariance matrix corresponding to ωt .

步骤S22、状态更新;Step S22: Status update;

任意t时刻的目标状态更新过程如下:The target state update process at any time t is as follows:

其中,Kt为获得目标状态最优估计值的卡尔曼增益矩阵,Ht为观测矩阵,Rt为测量噪声υt的协方差矩阵,是当前时刻目标状态的最优估计,/>是当前时刻目标状态的最优估计值对应的协方差矩阵,Zt为当前时刻目标状态的观测值,It为单位矩阵。Where Kt is the Kalman gain matrix for obtaining the optimal estimate of the target state, Ht is the observation matrix, Rt is the covariance matrix of the measurement noise υt , is the optimal estimate of the target state at the current moment, /> is the covariance matrix corresponding to the optimal estimate of the target state at the current moment, Z t is the observed value of the target state at the current moment, and I t is the unit matrix.

在本发明中,感兴趣目标表示为(x,y,z,θ,vx,vy,vz)组成的6维向量,其中(x,y,z)分别代表车辆中心点坐标,(vx,vy,vz)分别为车辆行驶过程中x,y,z方向上的速度分量,θ为车辆朝向。考虑到在相邻点云帧上进行车辆跟踪的过程中,朝向不发生明显变化,因此本发明忽略目标角速度,并采用匀速模型描述帧间车辆运动。记某个车辆目标的状态如式(35)所示:In the present invention, the target of interest is represented as a 6-dimensional vector composed of (x, y, z, θ, v x , vy , v z ), where (x, y, z) represent the coordinates of the center point of the vehicle, (v x , vy , v z ) are the velocity components in the x, y, and z directions during the vehicle's travel, and θ is the vehicle's orientation. Considering that the orientation does not change significantly during vehicle tracking on adjacent point cloud frames, the present invention ignores the target angular velocity and uses a uniform velocity model to describe the vehicle motion between frames. The state of a certain vehicle target is shown in formula (35):

X=[x,y,z,θ,vx,vy,vz]T (35)X=[x,y,z,θ, vx , vy , vz ] T (35)

已知车辆在t-1时刻的状态为Xt-1,根据运动学公式可推算出在t时刻的状态Xt,如式(36)所示:It is known that the state of the vehicle at time t-1 is X t-1 . The state X t at time t can be calculated according to the kinematic formula, as shown in formula (36):

其中,Xt-1为t-1时刻车辆目标状态,Xt为根据运动学模型计算所得的t时刻车辆目标状态,Δt为第t-1帧与第t帧点云的时间间隔,At为Xt-1至Xt的状态转移矩阵,ωt为运动方程中的噪声且服从零均值高斯分布。Among them, Xt -1 is the target state of the vehicle at time t-1, Xt is the target state of the vehicle at time t calculated according to the kinematic model, Δt is the time interval between the point clouds of the t-1th frame and the tth frame, At is the state transfer matrix from Xt-1 to Xt , and ωt is the noise in the motion equation and obeys the zero-mean Gaussian distribution.

本实施方式中使用三维车辆目标运动模型估计目标位置,将模型输出值作为状态观测值,可建立观测方程,如式(37)所示:In this embodiment, the target position is estimated using a three-dimensional vehicle target motion model, and the model output value is used as the state observation value to establish an observation equation, as shown in formula (37):

其中,Zt为t时刻的车辆状态观测值,Ht为Xt到Zt的观测矩阵,υt为t时刻的观测噪声,在本发明中为目标模型的观测误差。Wherein, Z t is the vehicle state observation value at time t, H t is the observation matrix from X t to Z t , and υ t is the observation noise at time t, which is the observation error of the target model in the present invention.

步骤S3、结合峰值旁瓣比(Peak Sidelobe Ratio,PSR)与内部形状描述子(Intrinsic Shape Signatures,ISS)对目标遮挡状态进行有效判断,根据目标遮挡情况的不同,提出自适应因子ε,利用自适应因子ε对核相关与卡尔曼跟踪器位置进行修正得到最终跟踪位置。Step S3, combining the Peak Sidelobe Ratio (PSR) and the Internal Shape Signatures (ISS) to effectively judge the target occlusion state, and according to different target occlusion situations, propose an adaptive factor ε, and use the adaptive factor ε to correct the kernel correlation and Kalman tracker position to obtain the final tracking position.

步骤S31、计算峰值旁瓣比;Step S31, calculating the peak sidelobe ratio;

峰值旁瓣比是一个描述主瓣相对旁瓣突出程度的物理量,可以用来评估相关运算的两个目标信号的匹配程度。定义为:The peak-to-sidelobe ratio is a physical quantity that describes the prominence of the main lobe relative to the side lobe, and can be used to evaluate the matching degree of two target signals in the correlation operation. It is defined as:

式中,max(f(z))为相关响应的峰值,μ、σ分别为除最大响应峰值外,其余旁瓣的均值和标准差。Where max(f(z)) is the peak value of the correlation response, μ and σ are the mean and standard deviation of the side lobes except the maximum response peak.

如图3所示,核相关滤波跟踪算法正常跟踪目标时,其响应图接近理想二维高斯分布,并且峰值明显;而在有干扰的情况下,尤其是遮挡情况下,如图3中的(b),目标特征被破坏,响应图会产生波动,不再遵循高斯分布,而是出现多峰,旁瓣与峰值间的对比度下降,如图3中的(a),当目标未被遮挡时,响应峰值明显达到最大值。As shown in Figure 3, when the kernel correlation filter tracking algorithm tracks the target normally, its response graph is close to the ideal two-dimensional Gaussian distribution and has an obvious peak. However, in the presence of interference, especially occlusion, as shown in Figure 3 (b), the target features are destroyed, the response graph fluctuates, no longer follows the Gaussian distribution, but has multiple peaks, and the contrast between the sidelobe and the peak decreases, as shown in Figure 3 (a). When the target is not occluded, the response peak obviously reaches the maximum value.

本实施方式预先计算目标未被遮挡前五帧的PSR平均值μPSR,如式(39),作为目标未被遮挡峰值响应标准,当目标处于完全遮挡时,峰值响应并不为0,经过实验测试,将0.3μPSR和0.8μPSR分别定义为遮挡检测机制的低阈值TL和高阈值THThis implementation pre-calculates the PSR average μ PSR of the first five frames when the target is not occluded, as shown in formula (39), as the peak response standard when the target is not occluded. When the target is completely occluded, the peak response is not 0. After experimental testing, 0.3μ PSR and 0.8μ PSR are defined as the low threshold TL and high threshold TH of the occlusion detection mechanism, respectively.

式中,n为图像总帧数,k为图像序列数。Where n is the total number of image frames, and k is the number of image sequences.

步骤S32、结合峰值旁瓣比自适应地更新核相关滤波模型;Step S32, adaptively updating the kernel correlation filter model in combination with the peak sidelobe ratio;

对于核相关滤波跟踪过程中,目标的所在区域信息会跟着帧数的不断推进发生改变,为了适应随着视频帧的改变而发生变换的目标外观特征,本实施方式结合峰值旁瓣比自适应地更新外观模型,αk、xk、αk-1、xk-1分别表示第k帧和第k-1帧的参数和目标模型,γk为每一帧中根据计算的PSR值自适应变化的学习率参数,β是一个恒定值,设置为0.025:In the kernel correlation filter tracking process, the target area information will change with the continuous advancement of the frame number. In order to adapt to the target appearance features that change with the change of video frames, this implementation method adaptively updates the appearance model in combination with the peak sidelobe ratio. α k , x k , α k-1 , and x k-1 represent the parameters and target models of the kth frame and the k-1th frame, respectively. γ k is a learning rate parameter that adaptively changes according to the calculated PSR value in each frame. β is a constant value set to 0.025:

步骤S33、根据遮挡情况的不同关键点数量多少进一步作为遮挡状态的评判标准;Step S33, further using the number of key points of different occlusion conditions as a criterion for judging the occlusion state;

具体ISS关键点提取流程如下,其结果图如图4所示:The specific ISS key point extraction process is as follows, and the result is shown in Figure 4:

设点云目标有Np个点,其任意一点pi坐标为(xi,yi,zi)。Suppose the point cloud target has Np points, and the coordinates of any point pi are (x i , y i , z i ).

1)将点云上的每个点pi定义一个局部坐标系,并给定每个点一个搜索半径rframe1) Define a local coordinate system for each point pi on the point cloud and give each point a search radius rframe ;

2)查询点云数据中每个点pi在半径rframe周围内的所有点,并计算其权值,即:2) Query all points within the radius r frame around each point p i in the point cloud data and calculate their weights, that is:

wij=1/|pi-pj|,|pi-pj|<rframe (41)w ij =1/|pi - pj |,|pi - pj |<r frame (41)

3)计算每个点pi的协方差矩阵:3) Calculate the covariance matrix of each point p i :

4)计算每个点pi的协方差矩阵cov(pi)的特征值并按从大到小的顺序排列/> 4) Calculate the eigenvalues of the covariance matrix cov( pi ) for each point p i And arrange them in order from largest to smallest/>

5)设置阈值ε1与ε2,满足式(43)的点则视为ISS特征点;5) Set the thresholds ε 1 and ε 2 , and the points satisfying equation (43) are regarded as ISS feature points;

6)重复上述步骤,直至完成所有的点;6) Repeat the above steps until all points are completed;

设点云目标遮挡前一刻提取到的关键点数量为Niss,同理将0.3Niss,0.8Niss分别定义为遮挡检测机制的低阈值(TL)和高阈值(TH)。Suppose the number of key points extracted just before the point cloud target is occluded is Niss . Similarly, 0.3Niss and 0.8Niss are defined as the low threshold ( TL ) and high threshold ( TH ) of the occlusion detection mechanism, respectively.

最终遮挡检测策略设计如下所示:The final occlusion detection strategy design is as follows:

步骤S34、根据目标遮挡情况的不同,本发明提出自适应因子ε,利用自适应因子ε对核相关与卡尔曼跟踪器位置进行修正;Step S34: According to different target occlusion conditions, the present invention proposes an adaptive factor ε, and uses the adaptive factor ε to correct the kernel correlation and Kalman tracker position;

自适应因子ε计算公式如下所示:The calculation formula of the adaptive factor ε is as follows:

式中,PSRk为第k帧图像的峰值旁瓣比,Nk iss为第k帧图像的关键点数量,μPSR为目标未被遮挡前五帧的PSR平均值,Niss为点云目标遮挡前一刻提取到的关键点数量;利用自适应因子ε对核相关与卡尔曼跟踪器位置进行修正得到最终跟踪位置:Where PSR k is the peak sidelobe ratio of the kth frame image, N k iss is the number of key points in the kth frame image, μ PSR is the average PSR of the first five frames before the target is occluded, and N iss is the number of key points extracted just before the point cloud target is occluded. The adaptive factor ε is used to correct the kernel correlation and Kalman tracker positions to obtain the final tracking position:

式中,Fop(x,y,z)为结合卡尔曼滤波和核相关滤波的最佳目标三维估计位置,FKCF(x,y)为核相关滤波器计算出的目标二维位置,FKM(x,y)为卡尔曼滤波器计算出的目标二维位置,F(x,y)为加权融合后的目标二维坐标位置,FKM(z)为卡尔曼滤波器计算出的目标第三维位置。由上式可知当目标处于无遮挡状态,ε更侧重核相关滤波器的跟踪结果;当目标处于完全遮挡状态,ε更侧重卡尔曼滤波器的跟踪结果。In the formula, F op (x, y, z) is the best target three-dimensional estimated position combined with Kalman filter and kernel correlation filter, F KCF (x, y) is the target two-dimensional position calculated by kernel correlation filter, F KM (x, y) is the target two-dimensional position calculated by Kalman filter, F(x, y) is the target two-dimensional coordinate position after weighted fusion, and F KM (z) is the target third-dimensional position calculated by Kalman filter. It can be seen from the above formula that when the target is in an unobstructed state, ε focuses more on the tracking result of kernel correlation filter; when the target is in a completely obstructed state, ε focuses more on the tracking result of Kalman filter.

为验证本发明所提目标跟踪算法的有效性,本发明在KITTI数据集上进行实验验证,将KITTI数据集中的彩色图像转换为灰度图像模拟GM-APD激光雷达所成强度像,然后分别利用核相关滤波与卡尔曼滤波器对图像与点云数据进行跟踪任务。采用中心位置误差(CLE)评价跟踪算法性能。CLE指的是在跟踪过程中,真实目标框中心位置和预测目标框中心位置的欧氏距离,欧氏距离越小跟踪精度越高。CLE计算公式如下:In order to verify the effectiveness of the target tracking algorithm proposed in the present invention, the present invention conducts experimental verification on the KITTI dataset, converts the color image in the KITTI dataset into a grayscale image to simulate the intensity image formed by the GM-APD laser radar, and then uses the kernel correlation filter and the Kalman filter to track the image and point cloud data respectively. The center position error (CLE) is used to evaluate the performance of the tracking algorithm. CLE refers to the Euclidean distance between the center position of the real target frame and the center position of the predicted target frame during the tracking process. The smaller the Euclidean distance, the higher the tracking accuracy. The CLE calculation formula is as follows:

其中CLE即中心误差值,p点是预测目标框中心位置,r点是真实目标框中心位置。Where CLE is the center error value, point p is the center position of the predicted target box, and point r is the center position of the real target box.

KITTI数据集跟踪实验:KITTI dataset tracking experiment:

跟踪场景为KITTI数据集跟踪0001场景,跟踪目标为86号白色轿车,选取轿车从无遮挡到被其它汽车遮挡再到无遮挡的运动状态序列。对目标进行跟踪时需要每次计算遮挡阈值,以便结合自适应因子有效融合强度像与点云跟踪位置信息。对部分运动序列的求取的遮挡阈值如图5所示。The tracking scene is the KITTI dataset tracking 0001 scene, and the tracking target is the No. 86 white car. The motion state sequence of the car from unobstructed to obstructed by other cars and then to unobstructed is selected. When tracking the target, it is necessary to calculate the occlusion threshold each time in order to effectively fuse the intensity image and point cloud tracking position information in combination with the adaptive factor. The occlusion threshold obtained for some motion sequences is shown in Figure 5.

从图5的(a)、(b)中可以看出,当目标未被遮挡时峰值旁瓣比PSR=7.69,ISS关键点数量为9;当目标处于部分遮挡,如(c)、(d),此时峰值旁瓣比值下降至5.62,ISS关键点数量减少至6,满足本发明所设置的遮挡检测区间;当目标处于完全遮挡时,如(e)、(f),峰值旁瓣比并不为0而是下降至1.77,此时ISS关键点数量为0,同样满足本发明设置的当目标完全遮挡时的阈值区间,证明本发明所设置的遮挡阈值有效,满足后续跟踪任务对目标遮挡状态的判断。It can be seen from (a) and (b) of Figure 5 that when the target is not obscured, the peak sidelobe ratio PSR = 7.69, and the number of ISS key points is 9; when the target is partially obscured, such as (c) and (d), the peak sidelobe ratio drops to 5.62, and the number of ISS key points is reduced to 6, which meets the occlusion detection interval set by the present invention; when the target is completely obscured, such as (e) and (f), the peak sidelobe ratio is not 0 but drops to 1.77, and the number of ISS key points is 0, which also meets the threshold interval set by the present invention when the target is completely obscured, proving that the occlusion threshold set by the present invention is effective and meets the judgment of the target occlusion state by subsequent tracking tasks.

图6为本发明的跟踪结果图,(a)、(c)、(e)、(g)分别为强度像跟踪结果;(b)、(d)、(f)、(h)分别为最终跟踪结果。其中红色框为目标真实位置,黄色框为算法跟踪位置,可以看出当目标处于无遮挡状态时,基于多特征融合的核相关滤波能有效跟踪感兴趣目标,当目标处于部分遮挡时,基于多特征融合核相关滤波跟踪结果出现偏差,卡尔曼跟踪算法能以较小误差预测目标位置。Figure 6 is a tracking result diagram of the present invention, (a), (c), (e), (g) are intensity image tracking results respectively; (b), (d), (f), (h) are final tracking results respectively. The red box is the real position of the target, and the yellow box is the algorithm tracking position. It can be seen that when the target is in an unobstructed state, the kernel correlation filter based on multi-feature fusion can effectively track the target of interest. When the target is partially obstructed, the tracking result based on the kernel correlation filter based on multi-feature fusion is biased, and the Kalman tracking algorithm can predict the target position with a small error.

从图7中可以看出,其中,(a)为中心位置误差曲线图,(b)为卡尔曼滤波跟踪轨迹图,(c)为扩展卡尔曼跟踪轨迹图,(d)为融合跟踪轨迹图;As can be seen from Figure 7, (a) is the center position error curve, (b) is the Kalman filter tracking trajectory, (c) is the extended Kalman tracking trajectory, and (d) is the fusion tracking trajectory;

扩展卡尔曼滤波算法的运行时间高于卡尔曼滤波算法,而通过结合核相关与卡尔曼跟踪器,在目标处于无遮挡状态是跟踪精度要优于卡尔曼滤波与扩展卡尔曼滤波,当目标处于部分以及完全遮挡,通过自适应因子调节,保持较高跟踪精度。从表1可看出,本发明所提算法平均中心位置误差相较于卡尔曼滤波精度提高21.67%,相较于扩展卡尔曼滤波精度提高7.94%,证明本发明方法的能对处于不同遮挡状态的目标进行准确跟踪。The running time of the extended Kalman filter algorithm is higher than that of the Kalman filter algorithm. By combining the kernel correlation and the Kalman tracker, the tracking accuracy is better than that of the Kalman filter and the extended Kalman filter when the target is in an unobstructed state. When the target is partially or completely obstructed, the adaptive factor is adjusted to maintain a high tracking accuracy. As can be seen from Table 1, the average center position error of the algorithm proposed in the present invention is 21.67% higher than that of the Kalman filter, and 7.94% higher than that of the extended Kalman filter, which proves that the method of the present invention can accurately track targets in different obstruction states.

表1Table 1

上述分析实验通过KITTI数据集验证了本发明所提算法的优越性,下面将通过实验进行验证。The above analysis experiment verifies the superiority of the algorithm proposed in the present invention through the KITTI dataset, which will be verified by experiments below.

激光雷达实采场景跟踪实验:LiDAR real scene tracking experiment:

固定GM-APD激光雷达探测视场对室外600远处十字口进行探测成像,对GM-APD激光雷达所成强度像与距离像转换后的点云数据进行跟踪实验,探测场景如图所示。The GM-APD laser radar detection field of view is fixed to detect and image the crossroad 600 meters away outdoors, and a tracking experiment is carried out on the point cloud data after the intensity image and distance image formed by the GM-APD laser radar are converted. The detection scene is shown in the figure.

从图9中可以看出,当目标在强度像中较为完整时,如图中的(a)、(c)所示,核相关滤波跟踪算法能有效跟踪感兴趣目标,随着跟踪目标逐渐驶离探测视场,如图中的(e)、(g)、(i)所示,基于强度像的跟踪结果逐渐失效,通过自适应因子调节,此时目标位置由卡尔曼滤波进行预测,从点云结果图中的(b)、(d)、(f)、(h)、(j)能看出,当目标从探测视场中由完整信息到逐渐消失时,本发明算法能对该目标进行有效跟踪,由于探测数据受噪点的影响,在该实验中本发明算法平均处理每帧时间为39ms。It can be seen from Figure 9 that when the target is relatively complete in the intensity image, as shown in (a) and (c) in the figure, the kernel correlation filter tracking algorithm can effectively track the target of interest. As the tracked target gradually moves away from the detection field of view, as shown in (e), (g), and (i) in the figure, the tracking result based on the intensity image gradually becomes invalid. The target position is predicted by the Kalman filter through adaptive factor adjustment. It can be seen from (b), (d), (f), (h), and (j) in the point cloud result diagram that when the target gradually disappears from the detection field of view from complete information, the algorithm of the present invention can effectively track the target. Since the detection data is affected by noise, the average processing time per frame of the algorithm of the present invention is 39ms in this experiment.

以上所述实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。The technical features of the above-described embodiments may be arbitrarily combined. To make the description concise, not all possible combinations of the technical features in the above-described embodiments are described. However, as long as there is no contradiction in the combination of these technical features, they should be considered to be within the scope of this specification.

以上所述实施例仅表达了本发明的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本发明构思的前提下,还可以做出若干变形和改进,这些都属于本发明的保护范围。因此,本发明专利的保护范围应以所附权利要求为准。The above-mentioned embodiments only express several implementation methods of the present invention, and the descriptions thereof are relatively specific and detailed, but they cannot be understood as limiting the scope of the invention patent. It should be pointed out that, for ordinary technicians in this field, several variations and improvements can be made without departing from the concept of the present invention, and these all belong to the protection scope of the present invention. Therefore, the protection scope of the patent of the present invention shall be subject to the attached claims.

Claims (7)

1.基于强度像与点云数据融合的目标跟踪方法,其特征是:该方法由以下步骤实现:1. A target tracking method based on the fusion of intensity image and point cloud data, characterized in that the method is implemented by the following steps: 步骤一、对激光雷达所成目标强度像的HOG特征、傅里叶描述子特征进行融合,并进行核相关滤波跟踪;Step 1: Fuse the HOG features and Fourier descriptor features of the target intensity image formed by the laser radar, and perform kernel correlation filter tracking; 步骤二、将激光雷达所成目标距离像转换为三维点云数据,利用卡尔曼跟踪器对点云目标进行轨迹预测;Step 2: Convert the target range image formed by the laser radar into three-dimensional point cloud data, and use the Kalman tracker to predict the trajectory of the point cloud target; 步骤三、结合峰值旁瓣比PSR与内部形状描述子ISS对目标遮挡状态进行有效判断,根据目标遮挡情况的不同,提出自适应因子,利用自适应因子对所述核相关滤波跟踪与卡尔曼跟踪器预测位置进行修正得到最终跟踪位置。Step 3: Combine the peak sidelobe ratio PSR and the internal shape descriptor ISS to effectively judge the target occlusion state. According to the different target occlusion conditions, an adaptive factor is proposed. The kernel correlation filter tracking and the Kalman tracker predicted position are corrected using the adaptive factor to obtain the final tracking position. 2.根据权利要求1所述的基于强度像与点云数据融合的目标跟踪方法,其特征在于:步骤一的具体过程为:2. The target tracking method based on the fusion of intensity image and point cloud data according to claim 1 is characterized in that: the specific process of step 1 is: 步骤一一、利用岭回归函数训练核相关滤波跟踪算法的分类器模型;Step 11: Use ridge regression function to train the classifier model of kernel correlation filter tracking algorithm; 步骤一二、通过循环矩阵扩充训练样本,获得训练后的核相关滤波跟踪模型;Step 1 and 2: Expand the training samples through the circulant matrix to obtain the trained kernel correlation filter tracking model; 步骤一三、采用所述核相关滤波跟踪模型对目标进行检测及定位。Step 13: Use the kernel correlation filter tracking model to detect and locate the target. 3.根据权利要求1所述的基于强度像与点云数据融合的目标跟踪方法,其特征在于:步骤一中,计算目标强度像的HOG特征的具体过程为:3. The target tracking method based on the fusion of intensity image and point cloud data according to claim 1 is characterized in that: in step 1, the specific process of calculating the HOG feature of the target intensity image is: A、截取跟踪区域图像,对所述图像进行灰度化,并采用Gamma校正法对输入图像进行归一化处理;A. intercepting the tracking area image, graying the image, and normalizing the input image using a gamma correction method; B、计算归一化后图像的像素点的梯度信息,获得目标的轮廓信息;B. Calculate the gradient information of the pixels of the normalized image to obtain the contour information of the target; C、将图像划分成多个cell单元格,统计出每个cell的梯度信息,将每个cell的梯度方向分成9个bin,即每个bin的间隔为20°,然后将梯度信息根据梯度方向映射到对应的bin上;C. Divide the image into multiple cells, count the gradient information of each cell, divide the gradient direction of each cell into 9 bins, that is, the interval of each bin is 20°, and then map the gradient information to the corresponding bin according to the gradient direction; D、统计由cell组成的block中的梯度信息,最后将所有的block内的梯度直方图串联获得整个图像的梯度直方图特征。D. Count the gradient information in the block composed of cells, and finally concatenate the gradient histograms in all blocks to obtain the gradient histogram features of the entire image. 4.根据权利要求1所述的基于强度像与点云数据融合的目标跟踪方法,其特征在于:步骤一中,计算傅里叶描述子特征的过程为:4. The target tracking method based on the fusion of intensity image and point cloud data according to claim 1 is characterized in that: in step 1, the process of calculating the Fourier descriptor feature is: a、对强度像灰度化处理并进行Gamma校正;a. Grayscale the intensity image and perform gamma correction; b、利用Canny边缘检测算法提取待跟踪目标的轮廓点集合;b. Use the Canny edge detection algorithm to extract the contour point set of the target to be tracked; c、计算轮廓点的中心坐标,将轮廓点从直角坐标系转换为极坐标系,获得中心点轮廓点距离序列;c. Calculate the center coordinates of the contour points, convert the contour points from the rectangular coordinate system to the polar coordinate system, and obtain the center point and contour point distance sequence; d、对中心轮廓点距离序列进行快速傅里叶变换,构建傅里叶描述子。d. Perform fast Fourier transform on the center contour point distance sequence to construct the Fourier descriptor. 5.根据权利要求1所述的基于强度像与点云数据融合的目标跟踪方法,其特征在于:步骤三的具体过程为:5. The target tracking method based on the fusion of intensity image and point cloud data according to claim 1 is characterized in that the specific process of step three is: 步骤三一、计算峰值旁瓣比;Step 31: Calculate the peak sidelobe ratio; 步骤三二、根据步骤三一计算的峰值旁瓣比自适应地更新核相关滤波跟踪模型;Step 32: adaptively updating the kernel correlation filter tracking model according to the peak sidelobe ratio calculated in step 31; 步骤三三、采用ISS检测算法对点云目标关键点进行提取,设置遮挡检测策略;Step 3. Use the ISS detection algorithm to extract the key points of the point cloud target and set the occlusion detection strategy; 步骤三四、根据目标遮挡情况的不同,采用自适应因子ε对核相关与卡尔曼跟踪器位置进行修正;Step 3 and 4: According to the different occlusion conditions of the target, the adaptive factor ε is used to correct the kernel correlation and Kalman tracker position; 所述自适应因子ε计算公式为:The calculation formula of the adaptive factor ε is: 式中,PSRk为第k帧图像的峰值旁瓣比,Nk iss为第k帧图像的关键点数量,μPSR为目标未被遮挡前五帧的PSR平均值,Niss为点云目标遮挡前一刻提取到的关键点数量;Where PSR k is the peak sidelobe ratio of the k-th frame image, N k iss is the number of key points in the k-th frame image, μ PSR is the average PSR of the first five frames before the target is not occluded, and N iss is the number of key points extracted just before the point cloud target is occluded; 利用自适应因子ε对核相关滤波跟踪器与卡尔曼跟踪器位置进行修正获得最终跟踪位置:The adaptive factor ε is used to correct the positions of the kernel correlation filter tracker and the Kalman tracker to obtain the final tracking position: F(x,y)=εFKCF(x,y)+(1-ε)FKM(x,y)F(x,y)=εF KCF (x,y)+(1-ε)F KM (x,y) Fop(x,y,z)=F(x,y,FKM(z))F op (x, y, z) = F (x, y, F KM (z)) 式中,Fop(x,y,z)为结合卡尔曼滤波和核相关滤波的最佳目标三维估计位置,FKCF(x,y)为核相关滤波器计算出的目标二维位置,FKM(x,y)为卡尔曼滤波器计算出的目标二维位置,F(x,y)为加权融合后的目标二维坐标位置,FKM(z)为卡尔曼滤波器计算出的目标第三维位置。由上式可知当目标处于无遮挡状态,ε侧重核相关滤波器的跟踪结果;当目标处于完全遮挡状态,ε侧重卡尔曼滤波器的跟踪结果。In the formula, F op (x, y, z) is the best target three-dimensional estimated position combined with Kalman filter and kernel correlation filter, F KCF (x, y) is the target two-dimensional position calculated by kernel correlation filter, F KM (x, y) is the target two-dimensional position calculated by Kalman filter, F(x, y) is the target two-dimensional coordinate position after weighted fusion, and F KM (z) is the target third-dimensional position calculated by Kalman filter. It can be seen from the above formula that when the target is in an unobstructed state, ε focuses on the tracking result of kernel correlation filter; when the target is in a completely obstructed state, ε focuses on the tracking result of Kalman filter. 6.根据权利要求5所述的基于强度像与点云数据融合的目标跟踪方法,其特征在于:步骤三二中,结合峰值旁瓣比自适应地更新核相滤波模型,设定αk、xk分别表示第k帧的参数和目标模型,αk-1、xk-1分别表示第第k-1帧的参数和目标模型,设定γk为每一帧中根据计算的PSR值自适应变化的学习率参数,则有下式:6. The target tracking method based on the fusion of intensity image and point cloud data according to claim 5 is characterized in that: in step 32, the kernel phase filter model is adaptively updated in combination with the peak sidelobe ratio, α k and x k are set to represent the parameters and target model of the kth frame respectively, α k-1 and x k-1 are set to represent the parameters and target model of the k-1th frame respectively, and γ k is set to be a learning rate parameter that adaptively changes according to the calculated PSR value in each frame, then the following formula is obtained: αk=(1-γkk-1kαk α k =(1-γ kk-1k α k xk=(1-γk)xk-1kxk x k =(1-γ k )x k-1k x k 式中,β为恒定值。In the formula, β is a constant value. 7.根据权利要求5所述的基于强度像与点云数据融合的目标跟踪方法,其特征在于:步骤三三中,采用ISS检测算法对点云目标关键点进行提取的具体过程为:7. The target tracking method based on the fusion of intensity image and point cloud data according to claim 5 is characterized in that: in step 33, the specific process of extracting the key points of the point cloud target using the ISS detection algorithm is: 设点云目标有Np个点,其任意一点pi坐标为(xi,yi,zi);Assume that the point cloud target has N p points, and the coordinates of any point p i are (x i ,y i ,z i ); 将点云目标的每个点pi定义一个局部坐标系,并给定每个点一个搜索半径rframeDefine a local coordinate system for each point p i of the point cloud target and give each point a search radius r frame ; 查询每个点pi在半径rframe周围内的所有点,并计算其权值;Query all points around each point p i in the frame of radius r and calculate their weights; 计算每个点pi的协方差矩阵,计算每个点pi的协方差矩阵的特征值并按从大到小的顺序排列/> Calculate the covariance matrix of each point p i and the eigenvalues of the covariance matrix of each point p i And arrange them in order from largest to smallest/> 设置阈值ε1与ε2,满足并且/>的点则视为ISS特征点;直至完成所有关键点提取;Set thresholds ε 1 and ε 2 to satisfy And/> The points are regarded as ISS feature points until all key points are extracted; 设点云目标遮挡前一刻提取的关键点数量为Niss,并将0.3Niss,0.8Niss分别定义为遮挡检测机制的低阈值TL和高阈值THAssume that the number of key points extracted before the point cloud target is occluded is Niss , and define 0.3Niss and 0.8Niss as the low threshold TL and high threshold TH of the occlusion detection mechanism respectively; 最终遮挡检测策略如下:The final occlusion detection strategy is as follows:
CN202410347052.XA 2024-03-26 2024-03-26 Target tracking method based on fusion of intensity image and point cloud data Pending CN118134972A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410347052.XA CN118134972A (en) 2024-03-26 2024-03-26 Target tracking method based on fusion of intensity image and point cloud data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410347052.XA CN118134972A (en) 2024-03-26 2024-03-26 Target tracking method based on fusion of intensity image and point cloud data

Publications (1)

Publication Number Publication Date
CN118134972A true CN118134972A (en) 2024-06-04

Family

ID=91235326

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202410347052.XA Pending CN118134972A (en) 2024-03-26 2024-03-26 Target tracking method based on fusion of intensity image and point cloud data

Country Status (1)

Country Link
CN (1) CN118134972A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119359801A (en) * 2024-09-25 2025-01-24 同济大学 A method for generating flexible object manipulation strategies based on multimodal fusion

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119359801A (en) * 2024-09-25 2025-01-24 同济大学 A method for generating flexible object manipulation strategies based on multimodal fusion

Similar Documents

Publication Publication Date Title
Nabati et al. Rrpn: Radar region proposal network for object detection in autonomous vehicles
US12072705B2 (en) Intelligent decision-making method and system for unmanned surface vehicle
Bai et al. Robust detection and tracking method for moving object based on radar and camera data fusion
Nie et al. A multimodality fusion deep neural network and safety test strategy for intelligent vehicles
CN114049382B (en) Target fusion tracking method, system and medium in intelligent network connection environment
Wang et al. YOLOv3-MT: A YOLOv3 using multi-target tracking for vehicle visual detection
CN107491731A (en) A kind of Ground moving target detection and recognition methods towards precision strike
CN105321189A (en) Complex environment target tracking method based on continuous adaptive mean shift multi-feature fusion
Hu et al. Face Detection based on SSD and CamShift
CN115546705B (en) Target identification method, terminal device and storage medium
CN109636834A (en) Video frequency vehicle target tracking algorism based on TLD innovatory algorithm
Chen et al. A graph-based track-before-detect algorithm for automotive radar target detection
Tian et al. Performance evaluation of deception against synthetic aperture radar based on multifeature fusion
Kopp et al. Tackling Clutter in Radar Data--Label Generation and Detection Using PointNet++
Liu et al. Correlation filter with motion detection for robust tracking of shape-deformed targets
Kahler et al. Predicted radar/optical feature fusion gains for target identification
Zhao et al. Nighttime pedestrian detection based on a fusion of visual information and millimeter-wave radar
CN120126307A (en) A method for cross-border tracking traffic warning recognition based on multi-array cameras
CN115471526A (en) Automatic driving target detection and tracking method based on multi-source heterogeneous information fusion
CN119064925A (en) A radar target detection method based on camera supervised feature enhancement
CN118134972A (en) Target tracking method based on fusion of intensity image and point cloud data
Ogunrinde Multi-sensor fusion for object detection and tracking under foggy weather conditions
CN115797794A (en) Satellite video multi-target tracking method based on knowledge distillation
Zhang et al. An efficient and flexible approach for multiple vehicle tracking in the aerial video sequence
CN113052871B (en) A target detection and automatic tracking method based on intelligent selection strategy

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination